{ "best_global_step": null, "best_metric": null, "best_model_checkpoint": null, "epoch": 5.0, "eval_steps": 500, "global_step": 5210, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0009601536245799327, "grad_norm": 4.10444855064962, "learning_rate": 0.0, "loss": 1.0452, "step": 1 }, { "epoch": 0.0019203072491598655, "grad_norm": 3.9618572065429634, "learning_rate": 1.9193857965451055e-08, "loss": 0.9371, "step": 2 }, { "epoch": 0.0028804608737397984, "grad_norm": 3.419070113612808, "learning_rate": 3.838771593090211e-08, "loss": 0.8672, "step": 3 }, { "epoch": 0.003840614498319731, "grad_norm": 4.504547011150056, "learning_rate": 5.758157389635317e-08, "loss": 0.9135, "step": 4 }, { "epoch": 0.004800768122899664, "grad_norm": 3.2967386081328036, "learning_rate": 7.677543186180422e-08, "loss": 0.9124, "step": 5 }, { "epoch": 0.005760921747479597, "grad_norm": 4.405246876588464, "learning_rate": 9.596928982725528e-08, "loss": 1.0376, "step": 6 }, { "epoch": 0.00672107537205953, "grad_norm": 4.9391091176922055, "learning_rate": 1.1516314779270635e-07, "loss": 0.8895, "step": 7 }, { "epoch": 0.007681228996639462, "grad_norm": 4.6313004015791694, "learning_rate": 1.343570057581574e-07, "loss": 0.9149, "step": 8 }, { "epoch": 0.008641382621219395, "grad_norm": 3.984493556236412, "learning_rate": 1.5355086372360844e-07, "loss": 0.8656, "step": 9 }, { "epoch": 0.009601536245799328, "grad_norm": 4.462584310971016, "learning_rate": 1.7274472168905953e-07, "loss": 0.9996, "step": 10 }, { "epoch": 0.01056168987037926, "grad_norm": 4.946483104014558, "learning_rate": 1.9193857965451055e-07, "loss": 0.8856, "step": 11 }, { "epoch": 0.011521843494959194, "grad_norm": 4.539379208096224, "learning_rate": 2.1113243761996164e-07, "loss": 1.1004, "step": 12 }, { "epoch": 0.012481997119539127, "grad_norm": 3.563643514638869, "learning_rate": 2.303262955854127e-07, "loss": 1.0574, "step": 13 }, { "epoch": 0.01344215074411906, "grad_norm": 4.300671203231014, "learning_rate": 2.4952015355086375e-07, "loss": 0.9869, "step": 14 }, { "epoch": 0.014402304368698993, "grad_norm": 4.137094807385917, "learning_rate": 2.687140115163148e-07, "loss": 0.987, "step": 15 }, { "epoch": 0.015362457993278924, "grad_norm": 3.8479659773850954, "learning_rate": 2.8790786948176586e-07, "loss": 0.9499, "step": 16 }, { "epoch": 0.01632261161785886, "grad_norm": 4.078585583081318, "learning_rate": 3.071017274472169e-07, "loss": 0.954, "step": 17 }, { "epoch": 0.01728276524243879, "grad_norm": 3.318497535833776, "learning_rate": 3.262955854126679e-07, "loss": 0.8825, "step": 18 }, { "epoch": 0.018242918867018725, "grad_norm": 2.9493015778475886, "learning_rate": 3.4548944337811905e-07, "loss": 1.0167, "step": 19 }, { "epoch": 0.019203072491598656, "grad_norm": 3.553502002331041, "learning_rate": 3.646833013435701e-07, "loss": 0.944, "step": 20 }, { "epoch": 0.020163226116178587, "grad_norm": 3.4081726405322854, "learning_rate": 3.838771593090211e-07, "loss": 0.855, "step": 21 }, { "epoch": 0.02112337974075852, "grad_norm": 2.87670565533856, "learning_rate": 4.0307101727447224e-07, "loss": 0.892, "step": 22 }, { "epoch": 0.022083533365338453, "grad_norm": 3.4510527426824087, "learning_rate": 4.2226487523992327e-07, "loss": 1.0082, "step": 23 }, { "epoch": 0.023043686989918388, "grad_norm": 3.738131349097318, "learning_rate": 4.414587332053743e-07, "loss": 0.948, "step": 24 }, { "epoch": 0.02400384061449832, "grad_norm": 3.2633271031520508, "learning_rate": 4.606525911708254e-07, "loss": 0.9681, "step": 25 }, { "epoch": 0.024963994239078253, "grad_norm": 3.458928003679058, "learning_rate": 4.798464491362765e-07, "loss": 0.8356, "step": 26 }, { "epoch": 0.025924147863658185, "grad_norm": 2.9414490652730585, "learning_rate": 4.990403071017275e-07, "loss": 1.0147, "step": 27 }, { "epoch": 0.02688430148823812, "grad_norm": 2.8282025991712603, "learning_rate": 5.182341650671785e-07, "loss": 0.9439, "step": 28 }, { "epoch": 0.02784445511281805, "grad_norm": 2.9564929653602174, "learning_rate": 5.374280230326296e-07, "loss": 0.9143, "step": 29 }, { "epoch": 0.028804608737397985, "grad_norm": 3.1655276997538473, "learning_rate": 5.566218809980807e-07, "loss": 0.897, "step": 30 }, { "epoch": 0.029764762361977917, "grad_norm": 3.2154482313705666, "learning_rate": 5.758157389635317e-07, "loss": 0.9446, "step": 31 }, { "epoch": 0.030724915986557848, "grad_norm": 2.113475487202004, "learning_rate": 5.950095969289827e-07, "loss": 1.0111, "step": 32 }, { "epoch": 0.03168506961113778, "grad_norm": 1.9644374588449889, "learning_rate": 6.142034548944338e-07, "loss": 0.9824, "step": 33 }, { "epoch": 0.03264522323571772, "grad_norm": 2.2536495629371873, "learning_rate": 6.333973128598848e-07, "loss": 0.964, "step": 34 }, { "epoch": 0.033605376860297645, "grad_norm": 2.0621232082523213, "learning_rate": 6.525911708253358e-07, "loss": 0.9925, "step": 35 }, { "epoch": 0.03456553048487758, "grad_norm": 2.4394849530250533, "learning_rate": 6.717850287907871e-07, "loss": 1.0279, "step": 36 }, { "epoch": 0.035525684109457514, "grad_norm": 1.6741022589129015, "learning_rate": 6.909788867562381e-07, "loss": 0.8821, "step": 37 }, { "epoch": 0.03648583773403745, "grad_norm": 2.3222552574990765, "learning_rate": 7.101727447216891e-07, "loss": 1.1046, "step": 38 }, { "epoch": 0.03744599135861738, "grad_norm": 1.9690958068895354, "learning_rate": 7.293666026871402e-07, "loss": 0.7766, "step": 39 }, { "epoch": 0.03840614498319731, "grad_norm": 1.9414084504593458, "learning_rate": 7.485604606525912e-07, "loss": 0.9705, "step": 40 }, { "epoch": 0.039366298607777246, "grad_norm": 3.245290590634154, "learning_rate": 7.677543186180422e-07, "loss": 1.0038, "step": 41 }, { "epoch": 0.040326452232357174, "grad_norm": 1.8595665959681709, "learning_rate": 7.869481765834934e-07, "loss": 0.9792, "step": 42 }, { "epoch": 0.04128660585693711, "grad_norm": 1.5774800398911444, "learning_rate": 8.061420345489445e-07, "loss": 0.9147, "step": 43 }, { "epoch": 0.04224675948151704, "grad_norm": 1.6857616573262868, "learning_rate": 8.253358925143955e-07, "loss": 0.8915, "step": 44 }, { "epoch": 0.04320691310609698, "grad_norm": 1.722471027921981, "learning_rate": 8.445297504798465e-07, "loss": 0.9164, "step": 45 }, { "epoch": 0.044167066730676906, "grad_norm": 1.892828306643539, "learning_rate": 8.637236084452976e-07, "loss": 0.8689, "step": 46 }, { "epoch": 0.04512722035525684, "grad_norm": 1.552617863695917, "learning_rate": 8.829174664107486e-07, "loss": 0.8007, "step": 47 }, { "epoch": 0.046087373979836775, "grad_norm": 1.8930164151379243, "learning_rate": 9.021113243761997e-07, "loss": 0.9919, "step": 48 }, { "epoch": 0.04704752760441671, "grad_norm": 1.7346341958776508, "learning_rate": 9.213051823416508e-07, "loss": 0.9317, "step": 49 }, { "epoch": 0.04800768122899664, "grad_norm": 1.6796946282263094, "learning_rate": 9.404990403071018e-07, "loss": 0.8578, "step": 50 }, { "epoch": 0.04896783485357657, "grad_norm": 1.4459845637698316, "learning_rate": 9.59692898272553e-07, "loss": 0.8367, "step": 51 }, { "epoch": 0.04992798847815651, "grad_norm": 1.6088132452607184, "learning_rate": 9.788867562380039e-07, "loss": 0.8785, "step": 52 }, { "epoch": 0.050888142102736435, "grad_norm": 1.7846652063512944, "learning_rate": 9.98080614203455e-07, "loss": 0.968, "step": 53 }, { "epoch": 0.05184829572731637, "grad_norm": 1.2943443396235055, "learning_rate": 1.0172744721689061e-06, "loss": 0.7964, "step": 54 }, { "epoch": 0.052808449351896304, "grad_norm": 1.7456405075737602, "learning_rate": 1.036468330134357e-06, "loss": 1.0337, "step": 55 }, { "epoch": 0.05376860297647624, "grad_norm": 1.6134102461852162, "learning_rate": 1.0556621880998082e-06, "loss": 1.0233, "step": 56 }, { "epoch": 0.05472875660105617, "grad_norm": 1.2553627482465546, "learning_rate": 1.074856046065259e-06, "loss": 0.8179, "step": 57 }, { "epoch": 0.0556889102256361, "grad_norm": 1.8531884297945367, "learning_rate": 1.0940499040307102e-06, "loss": 1.0597, "step": 58 }, { "epoch": 0.056649063850216036, "grad_norm": 1.6467062051297792, "learning_rate": 1.1132437619961614e-06, "loss": 0.8744, "step": 59 }, { "epoch": 0.05760921747479597, "grad_norm": 1.283676444866443, "learning_rate": 1.1324376199616125e-06, "loss": 0.9178, "step": 60 }, { "epoch": 0.0585693710993759, "grad_norm": 1.1279263431250381, "learning_rate": 1.1516314779270634e-06, "loss": 0.828, "step": 61 }, { "epoch": 0.05952952472395583, "grad_norm": 1.1442091020497502, "learning_rate": 1.1708253358925146e-06, "loss": 0.9064, "step": 62 }, { "epoch": 0.06048967834853577, "grad_norm": 1.201687643460303, "learning_rate": 1.1900191938579655e-06, "loss": 0.8777, "step": 63 }, { "epoch": 0.061449831973115696, "grad_norm": 1.2122758520764403, "learning_rate": 1.2092130518234166e-06, "loss": 0.9441, "step": 64 }, { "epoch": 0.06240998559769563, "grad_norm": 1.031817210002686, "learning_rate": 1.2284069097888675e-06, "loss": 0.7742, "step": 65 }, { "epoch": 0.06337013922227556, "grad_norm": 1.3134170454430059, "learning_rate": 1.2476007677543187e-06, "loss": 1.0127, "step": 66 }, { "epoch": 0.0643302928468555, "grad_norm": 1.0520178131880826, "learning_rate": 1.2667946257197696e-06, "loss": 0.8569, "step": 67 }, { "epoch": 0.06529044647143543, "grad_norm": 1.114672550184489, "learning_rate": 1.2859884836852207e-06, "loss": 0.8917, "step": 68 }, { "epoch": 0.06625060009601537, "grad_norm": 1.0124568717816598, "learning_rate": 1.3051823416506717e-06, "loss": 0.8223, "step": 69 }, { "epoch": 0.06721075372059529, "grad_norm": 1.2688655372994084, "learning_rate": 1.324376199616123e-06, "loss": 0.9627, "step": 70 }, { "epoch": 0.06817090734517522, "grad_norm": 0.9375020568067154, "learning_rate": 1.3435700575815741e-06, "loss": 0.8383, "step": 71 }, { "epoch": 0.06913106096975516, "grad_norm": 0.9778363521658088, "learning_rate": 1.362763915547025e-06, "loss": 0.7943, "step": 72 }, { "epoch": 0.0700912145943351, "grad_norm": 1.0855246260250955, "learning_rate": 1.3819577735124762e-06, "loss": 0.9235, "step": 73 }, { "epoch": 0.07105136821891503, "grad_norm": 0.941260012819113, "learning_rate": 1.4011516314779271e-06, "loss": 0.8378, "step": 74 }, { "epoch": 0.07201152184349496, "grad_norm": 1.306447101768563, "learning_rate": 1.4203454894433783e-06, "loss": 0.9868, "step": 75 }, { "epoch": 0.0729716754680749, "grad_norm": 1.1030356780156272, "learning_rate": 1.4395393474088292e-06, "loss": 0.7981, "step": 76 }, { "epoch": 0.07393182909265482, "grad_norm": 1.3168485541317252, "learning_rate": 1.4587332053742803e-06, "loss": 1.022, "step": 77 }, { "epoch": 0.07489198271723475, "grad_norm": 1.081165919814939, "learning_rate": 1.4779270633397312e-06, "loss": 0.8966, "step": 78 }, { "epoch": 0.07585213634181469, "grad_norm": 1.02325232681021, "learning_rate": 1.4971209213051824e-06, "loss": 0.9088, "step": 79 }, { "epoch": 0.07681228996639462, "grad_norm": 0.9141319894974212, "learning_rate": 1.5163147792706335e-06, "loss": 0.748, "step": 80 }, { "epoch": 0.07777244359097456, "grad_norm": 0.8990270852512517, "learning_rate": 1.5355086372360844e-06, "loss": 0.722, "step": 81 }, { "epoch": 0.07873259721555449, "grad_norm": 1.1062890605776887, "learning_rate": 1.5547024952015358e-06, "loss": 0.8408, "step": 82 }, { "epoch": 0.07969275084013443, "grad_norm": 0.9504998513577274, "learning_rate": 1.5738963531669867e-06, "loss": 0.7876, "step": 83 }, { "epoch": 0.08065290446471435, "grad_norm": 0.9108099147398749, "learning_rate": 1.5930902111324378e-06, "loss": 0.6678, "step": 84 }, { "epoch": 0.08161305808929428, "grad_norm": 0.8478291906324409, "learning_rate": 1.612284069097889e-06, "loss": 0.7622, "step": 85 }, { "epoch": 0.08257321171387422, "grad_norm": 0.9918980202618719, "learning_rate": 1.63147792706334e-06, "loss": 0.8953, "step": 86 }, { "epoch": 0.08353336533845415, "grad_norm": 1.063403145156319, "learning_rate": 1.650671785028791e-06, "loss": 0.9117, "step": 87 }, { "epoch": 0.08449351896303409, "grad_norm": 0.9486472580860418, "learning_rate": 1.669865642994242e-06, "loss": 0.7685, "step": 88 }, { "epoch": 0.08545367258761402, "grad_norm": 1.335137744930855, "learning_rate": 1.689059500959693e-06, "loss": 1.0255, "step": 89 }, { "epoch": 0.08641382621219396, "grad_norm": 0.9401566174684787, "learning_rate": 1.708253358925144e-06, "loss": 0.8968, "step": 90 }, { "epoch": 0.08737397983677389, "grad_norm": 1.3498781465009442, "learning_rate": 1.7274472168905951e-06, "loss": 0.9677, "step": 91 }, { "epoch": 0.08833413346135381, "grad_norm": 0.9660056525384505, "learning_rate": 1.746641074856046e-06, "loss": 0.7901, "step": 92 }, { "epoch": 0.08929428708593375, "grad_norm": 1.2676562950184382, "learning_rate": 1.7658349328214972e-06, "loss": 0.8744, "step": 93 }, { "epoch": 0.09025444071051368, "grad_norm": 1.3023436377480102, "learning_rate": 1.7850287907869481e-06, "loss": 0.7728, "step": 94 }, { "epoch": 0.09121459433509362, "grad_norm": 0.7923412562582911, "learning_rate": 1.8042226487523995e-06, "loss": 0.8049, "step": 95 }, { "epoch": 0.09217474795967355, "grad_norm": 1.2346601859811188, "learning_rate": 1.8234165067178506e-06, "loss": 0.885, "step": 96 }, { "epoch": 0.09313490158425348, "grad_norm": 1.000225725736867, "learning_rate": 1.8426103646833015e-06, "loss": 0.8206, "step": 97 }, { "epoch": 0.09409505520883342, "grad_norm": 0.758343228335248, "learning_rate": 1.8618042226487527e-06, "loss": 0.7521, "step": 98 }, { "epoch": 0.09505520883341334, "grad_norm": 1.1046237969893864, "learning_rate": 1.8809980806142036e-06, "loss": 0.8369, "step": 99 }, { "epoch": 0.09601536245799328, "grad_norm": 1.1338327005345359, "learning_rate": 1.9001919385796547e-06, "loss": 0.8931, "step": 100 }, { "epoch": 0.09697551608257321, "grad_norm": 0.9327248890043498, "learning_rate": 1.919385796545106e-06, "loss": 0.8061, "step": 101 }, { "epoch": 0.09793566970715314, "grad_norm": 0.8651894389465067, "learning_rate": 1.9385796545105568e-06, "loss": 0.7515, "step": 102 }, { "epoch": 0.09889582333173308, "grad_norm": 1.0131523836861378, "learning_rate": 1.9577735124760077e-06, "loss": 0.7936, "step": 103 }, { "epoch": 0.09985597695631301, "grad_norm": 1.3237433252824902, "learning_rate": 1.9769673704414586e-06, "loss": 0.9497, "step": 104 }, { "epoch": 0.10081613058089295, "grad_norm": 0.8294265321180141, "learning_rate": 1.99616122840691e-06, "loss": 0.696, "step": 105 }, { "epoch": 0.10177628420547287, "grad_norm": 1.0709415430890428, "learning_rate": 2.015355086372361e-06, "loss": 0.9035, "step": 106 }, { "epoch": 0.1027364378300528, "grad_norm": 1.1389486559157997, "learning_rate": 2.0345489443378122e-06, "loss": 0.9063, "step": 107 }, { "epoch": 0.10369659145463274, "grad_norm": 1.198066897099718, "learning_rate": 2.053742802303263e-06, "loss": 0.8781, "step": 108 }, { "epoch": 0.10465674507921267, "grad_norm": 0.9750735328201544, "learning_rate": 2.072936660268714e-06, "loss": 0.7101, "step": 109 }, { "epoch": 0.10561689870379261, "grad_norm": 1.0025479059904023, "learning_rate": 2.0921305182341654e-06, "loss": 0.8546, "step": 110 }, { "epoch": 0.10657705232837254, "grad_norm": 1.023153846829638, "learning_rate": 2.1113243761996164e-06, "loss": 0.766, "step": 111 }, { "epoch": 0.10753720595295248, "grad_norm": 1.1866992174392377, "learning_rate": 2.1305182341650673e-06, "loss": 0.9112, "step": 112 }, { "epoch": 0.1084973595775324, "grad_norm": 0.9202490222786923, "learning_rate": 2.149712092130518e-06, "loss": 0.8579, "step": 113 }, { "epoch": 0.10945751320211233, "grad_norm": 0.817007852872672, "learning_rate": 2.1689059500959696e-06, "loss": 0.6764, "step": 114 }, { "epoch": 0.11041766682669227, "grad_norm": 1.067753287607919, "learning_rate": 2.1880998080614205e-06, "loss": 0.8194, "step": 115 }, { "epoch": 0.1113778204512722, "grad_norm": 1.2639599676461073, "learning_rate": 2.2072936660268714e-06, "loss": 0.9191, "step": 116 }, { "epoch": 0.11233797407585214, "grad_norm": 0.8627272285722953, "learning_rate": 2.2264875239923228e-06, "loss": 0.7871, "step": 117 }, { "epoch": 0.11329812770043207, "grad_norm": 1.5435824386314636, "learning_rate": 2.2456813819577737e-06, "loss": 0.8058, "step": 118 }, { "epoch": 0.114258281325012, "grad_norm": 0.8737093779753992, "learning_rate": 2.264875239923225e-06, "loss": 0.7815, "step": 119 }, { "epoch": 0.11521843494959194, "grad_norm": 0.9911445122330073, "learning_rate": 2.284069097888676e-06, "loss": 0.7812, "step": 120 }, { "epoch": 0.11617858857417186, "grad_norm": 1.1276076624574065, "learning_rate": 2.303262955854127e-06, "loss": 0.8646, "step": 121 }, { "epoch": 0.1171387421987518, "grad_norm": 0.8819871946455501, "learning_rate": 2.322456813819578e-06, "loss": 0.7848, "step": 122 }, { "epoch": 0.11809889582333173, "grad_norm": 0.909557789076695, "learning_rate": 2.341650671785029e-06, "loss": 0.7442, "step": 123 }, { "epoch": 0.11905904944791167, "grad_norm": 0.8752500907233446, "learning_rate": 2.36084452975048e-06, "loss": 0.9037, "step": 124 }, { "epoch": 0.1200192030724916, "grad_norm": 1.0499196050669244, "learning_rate": 2.380038387715931e-06, "loss": 0.8083, "step": 125 }, { "epoch": 0.12097935669707154, "grad_norm": 0.9930471648167666, "learning_rate": 2.3992322456813823e-06, "loss": 0.7578, "step": 126 }, { "epoch": 0.12193951032165147, "grad_norm": 1.0067661322745958, "learning_rate": 2.4184261036468333e-06, "loss": 0.9388, "step": 127 }, { "epoch": 0.12289966394623139, "grad_norm": 1.0415341855708207, "learning_rate": 2.437619961612284e-06, "loss": 0.8657, "step": 128 }, { "epoch": 0.12385981757081133, "grad_norm": 0.9717585813494415, "learning_rate": 2.456813819577735e-06, "loss": 0.8449, "step": 129 }, { "epoch": 0.12481997119539126, "grad_norm": 1.0176462139327744, "learning_rate": 2.4760076775431864e-06, "loss": 0.8038, "step": 130 }, { "epoch": 0.1257801248199712, "grad_norm": 1.0445462536491283, "learning_rate": 2.4952015355086374e-06, "loss": 0.802, "step": 131 }, { "epoch": 0.12674027844455113, "grad_norm": 1.1181790259353253, "learning_rate": 2.5143953934740883e-06, "loss": 0.7495, "step": 132 }, { "epoch": 0.12770043206913106, "grad_norm": 0.8328753571370403, "learning_rate": 2.5335892514395392e-06, "loss": 0.7604, "step": 133 }, { "epoch": 0.128660585693711, "grad_norm": 0.9294683651143865, "learning_rate": 2.5527831094049906e-06, "loss": 0.8084, "step": 134 }, { "epoch": 0.12962073931829093, "grad_norm": 1.0356863934576512, "learning_rate": 2.5719769673704415e-06, "loss": 0.867, "step": 135 }, { "epoch": 0.13058089294287087, "grad_norm": 1.0398671555923786, "learning_rate": 2.5911708253358924e-06, "loss": 0.8469, "step": 136 }, { "epoch": 0.1315410465674508, "grad_norm": 0.9916990845664827, "learning_rate": 2.6103646833013433e-06, "loss": 0.7663, "step": 137 }, { "epoch": 0.13250120019203074, "grad_norm": 1.076726566442809, "learning_rate": 2.629558541266795e-06, "loss": 0.7961, "step": 138 }, { "epoch": 0.13346135381661065, "grad_norm": 0.9164339055871082, "learning_rate": 2.648752399232246e-06, "loss": 0.8182, "step": 139 }, { "epoch": 0.13442150744119058, "grad_norm": 0.9147352937586469, "learning_rate": 2.6679462571976974e-06, "loss": 0.8894, "step": 140 }, { "epoch": 0.13538166106577051, "grad_norm": 0.7883211297717718, "learning_rate": 2.6871401151631483e-06, "loss": 0.6967, "step": 141 }, { "epoch": 0.13634181469035045, "grad_norm": 1.3504494536810547, "learning_rate": 2.7063339731285992e-06, "loss": 0.8917, "step": 142 }, { "epoch": 0.13730196831493038, "grad_norm": 1.0844729957529637, "learning_rate": 2.72552783109405e-06, "loss": 0.867, "step": 143 }, { "epoch": 0.13826212193951032, "grad_norm": 1.0629922772991385, "learning_rate": 2.7447216890595015e-06, "loss": 0.7721, "step": 144 }, { "epoch": 0.13922227556409025, "grad_norm": 0.7759978929785781, "learning_rate": 2.7639155470249524e-06, "loss": 0.6737, "step": 145 }, { "epoch": 0.1401824291886702, "grad_norm": 0.9224741560861578, "learning_rate": 2.7831094049904033e-06, "loss": 0.8346, "step": 146 }, { "epoch": 0.14114258281325012, "grad_norm": 0.9382148166925125, "learning_rate": 2.8023032629558543e-06, "loss": 0.7803, "step": 147 }, { "epoch": 0.14210273643783006, "grad_norm": 0.8862406078423355, "learning_rate": 2.8214971209213056e-06, "loss": 0.7378, "step": 148 }, { "epoch": 0.14306289006241, "grad_norm": 0.7632730666301243, "learning_rate": 2.8406909788867565e-06, "loss": 0.7306, "step": 149 }, { "epoch": 0.14402304368698993, "grad_norm": 0.7115596047348328, "learning_rate": 2.8598848368522074e-06, "loss": 0.7514, "step": 150 }, { "epoch": 0.14498319731156986, "grad_norm": 0.8284467736078194, "learning_rate": 2.8790786948176584e-06, "loss": 0.8112, "step": 151 }, { "epoch": 0.1459433509361498, "grad_norm": 0.8423622253935678, "learning_rate": 2.8982725527831097e-06, "loss": 0.6642, "step": 152 }, { "epoch": 0.1469035045607297, "grad_norm": 0.7707397234782508, "learning_rate": 2.9174664107485606e-06, "loss": 0.8061, "step": 153 }, { "epoch": 0.14786365818530964, "grad_norm": 1.0357464605566897, "learning_rate": 2.9366602687140116e-06, "loss": 0.8056, "step": 154 }, { "epoch": 0.14882381180988957, "grad_norm": 0.9531579205536315, "learning_rate": 2.9558541266794625e-06, "loss": 0.7873, "step": 155 }, { "epoch": 0.1497839654344695, "grad_norm": 0.7750573733017515, "learning_rate": 2.975047984644914e-06, "loss": 0.776, "step": 156 }, { "epoch": 0.15074411905904944, "grad_norm": 1.4888666657011065, "learning_rate": 2.9942418426103648e-06, "loss": 0.9984, "step": 157 }, { "epoch": 0.15170427268362938, "grad_norm": 0.9702173414716171, "learning_rate": 3.0134357005758157e-06, "loss": 0.7023, "step": 158 }, { "epoch": 0.1526644263082093, "grad_norm": 0.8684716298870753, "learning_rate": 3.032629558541267e-06, "loss": 0.7891, "step": 159 }, { "epoch": 0.15362457993278925, "grad_norm": 0.9347776960963162, "learning_rate": 3.051823416506718e-06, "loss": 0.6662, "step": 160 }, { "epoch": 0.15458473355736918, "grad_norm": 0.8318965740811947, "learning_rate": 3.071017274472169e-06, "loss": 0.8223, "step": 161 }, { "epoch": 0.15554488718194912, "grad_norm": 0.9388239414416779, "learning_rate": 3.09021113243762e-06, "loss": 0.8159, "step": 162 }, { "epoch": 0.15650504080652905, "grad_norm": 0.818422427915953, "learning_rate": 3.1094049904030716e-06, "loss": 0.761, "step": 163 }, { "epoch": 0.15746519443110898, "grad_norm": 0.9426119895680675, "learning_rate": 3.1285988483685225e-06, "loss": 0.8185, "step": 164 }, { "epoch": 0.15842534805568892, "grad_norm": 0.9321241598448508, "learning_rate": 3.1477927063339734e-06, "loss": 0.766, "step": 165 }, { "epoch": 0.15938550168026885, "grad_norm": 1.047512477967633, "learning_rate": 3.1669865642994248e-06, "loss": 0.7572, "step": 166 }, { "epoch": 0.1603456553048488, "grad_norm": 0.6566424836405258, "learning_rate": 3.1861804222648757e-06, "loss": 0.6956, "step": 167 }, { "epoch": 0.1613058089294287, "grad_norm": 0.9240979019872778, "learning_rate": 3.2053742802303266e-06, "loss": 0.8215, "step": 168 }, { "epoch": 0.16226596255400863, "grad_norm": 1.0798605484455204, "learning_rate": 3.224568138195778e-06, "loss": 0.8858, "step": 169 }, { "epoch": 0.16322611617858857, "grad_norm": 0.9369618471346057, "learning_rate": 3.243761996161229e-06, "loss": 0.7996, "step": 170 }, { "epoch": 0.1641862698031685, "grad_norm": 1.3177828524262762, "learning_rate": 3.26295585412668e-06, "loss": 0.8439, "step": 171 }, { "epoch": 0.16514642342774843, "grad_norm": 0.7454002627673212, "learning_rate": 3.2821497120921307e-06, "loss": 0.7022, "step": 172 }, { "epoch": 0.16610657705232837, "grad_norm": 0.8426405117310581, "learning_rate": 3.301343570057582e-06, "loss": 0.6438, "step": 173 }, { "epoch": 0.1670667306769083, "grad_norm": 1.192071955722208, "learning_rate": 3.320537428023033e-06, "loss": 0.8616, "step": 174 }, { "epoch": 0.16802688430148824, "grad_norm": 1.05511870059076, "learning_rate": 3.339731285988484e-06, "loss": 0.7901, "step": 175 }, { "epoch": 0.16898703792606817, "grad_norm": 0.9410097443912633, "learning_rate": 3.358925143953935e-06, "loss": 0.7251, "step": 176 }, { "epoch": 0.1699471915506481, "grad_norm": 0.7908091207352825, "learning_rate": 3.378119001919386e-06, "loss": 0.7251, "step": 177 }, { "epoch": 0.17090734517522804, "grad_norm": 1.1585296794723008, "learning_rate": 3.397312859884837e-06, "loss": 0.8346, "step": 178 }, { "epoch": 0.17186749879980798, "grad_norm": 1.0643947937835683, "learning_rate": 3.416506717850288e-06, "loss": 0.8351, "step": 179 }, { "epoch": 0.1728276524243879, "grad_norm": 0.909993490636638, "learning_rate": 3.435700575815739e-06, "loss": 0.7659, "step": 180 }, { "epoch": 0.17378780604896785, "grad_norm": 0.9280799184660852, "learning_rate": 3.4548944337811903e-06, "loss": 0.8617, "step": 181 }, { "epoch": 0.17474795967354778, "grad_norm": 0.8676496032438762, "learning_rate": 3.4740882917466412e-06, "loss": 0.7806, "step": 182 }, { "epoch": 0.1757081132981277, "grad_norm": 0.8168199913522222, "learning_rate": 3.493282149712092e-06, "loss": 0.8537, "step": 183 }, { "epoch": 0.17666826692270762, "grad_norm": 1.2683989338857937, "learning_rate": 3.512476007677543e-06, "loss": 0.9864, "step": 184 }, { "epoch": 0.17762842054728756, "grad_norm": 0.7977245009097769, "learning_rate": 3.5316698656429944e-06, "loss": 0.7154, "step": 185 }, { "epoch": 0.1785885741718675, "grad_norm": 0.8678783443608111, "learning_rate": 3.5508637236084453e-06, "loss": 0.7847, "step": 186 }, { "epoch": 0.17954872779644743, "grad_norm": 0.8001534003123866, "learning_rate": 3.5700575815738963e-06, "loss": 0.75, "step": 187 }, { "epoch": 0.18050888142102736, "grad_norm": 1.2193105152546846, "learning_rate": 3.589251439539348e-06, "loss": 0.9408, "step": 188 }, { "epoch": 0.1814690350456073, "grad_norm": 1.2137697079042309, "learning_rate": 3.608445297504799e-06, "loss": 0.7233, "step": 189 }, { "epoch": 0.18242918867018723, "grad_norm": 0.7008359809261104, "learning_rate": 3.62763915547025e-06, "loss": 0.6309, "step": 190 }, { "epoch": 0.18338934229476717, "grad_norm": 1.0139398094041818, "learning_rate": 3.6468330134357012e-06, "loss": 0.848, "step": 191 }, { "epoch": 0.1843494959193471, "grad_norm": 0.872433976644939, "learning_rate": 3.666026871401152e-06, "loss": 0.7498, "step": 192 }, { "epoch": 0.18530964954392704, "grad_norm": 1.1592254364266, "learning_rate": 3.685220729366603e-06, "loss": 0.7995, "step": 193 }, { "epoch": 0.18626980316850697, "grad_norm": 0.7777006396236404, "learning_rate": 3.704414587332054e-06, "loss": 0.6151, "step": 194 }, { "epoch": 0.1872299567930869, "grad_norm": 1.0365441759632188, "learning_rate": 3.7236084452975053e-06, "loss": 0.8464, "step": 195 }, { "epoch": 0.18819011041766684, "grad_norm": 0.9028843645929712, "learning_rate": 3.7428023032629563e-06, "loss": 0.8218, "step": 196 }, { "epoch": 0.18915026404224675, "grad_norm": 0.8649657472604179, "learning_rate": 3.761996161228407e-06, "loss": 0.8317, "step": 197 }, { "epoch": 0.19011041766682668, "grad_norm": 1.0300087132237699, "learning_rate": 3.781190019193858e-06, "loss": 0.8223, "step": 198 }, { "epoch": 0.19107057129140662, "grad_norm": 0.7863906558907481, "learning_rate": 3.8003838771593095e-06, "loss": 0.6627, "step": 199 }, { "epoch": 0.19203072491598655, "grad_norm": 0.9423035683344614, "learning_rate": 3.81957773512476e-06, "loss": 0.8169, "step": 200 }, { "epoch": 0.19299087854056649, "grad_norm": 0.855598498978331, "learning_rate": 3.838771593090212e-06, "loss": 0.6755, "step": 201 }, { "epoch": 0.19395103216514642, "grad_norm": 0.9409828196221999, "learning_rate": 3.857965451055662e-06, "loss": 0.7169, "step": 202 }, { "epoch": 0.19491118578972635, "grad_norm": 0.8246655722845775, "learning_rate": 3.8771593090211136e-06, "loss": 0.7122, "step": 203 }, { "epoch": 0.1958713394143063, "grad_norm": 0.8326155164575071, "learning_rate": 3.896353166986565e-06, "loss": 0.723, "step": 204 }, { "epoch": 0.19683149303888622, "grad_norm": 0.7931464494221239, "learning_rate": 3.915547024952015e-06, "loss": 0.6057, "step": 205 }, { "epoch": 0.19779164666346616, "grad_norm": 0.7780995356254307, "learning_rate": 3.934740882917467e-06, "loss": 0.7403, "step": 206 }, { "epoch": 0.1987518002880461, "grad_norm": 1.3143551535136784, "learning_rate": 3.953934740882917e-06, "loss": 0.845, "step": 207 }, { "epoch": 0.19971195391262603, "grad_norm": 0.8413602983207709, "learning_rate": 3.973128598848369e-06, "loss": 0.6077, "step": 208 }, { "epoch": 0.20067210753720596, "grad_norm": 0.8648303707284405, "learning_rate": 3.99232245681382e-06, "loss": 0.7781, "step": 209 }, { "epoch": 0.2016322611617859, "grad_norm": 1.2691886552317635, "learning_rate": 4.0115163147792705e-06, "loss": 0.8335, "step": 210 }, { "epoch": 0.20259241478636583, "grad_norm": 0.826957967390477, "learning_rate": 4.030710172744722e-06, "loss": 0.6816, "step": 211 }, { "epoch": 0.20355256841094574, "grad_norm": 1.3222299051529318, "learning_rate": 4.049904030710173e-06, "loss": 0.7353, "step": 212 }, { "epoch": 0.20451272203552567, "grad_norm": 1.2132274116261752, "learning_rate": 4.0690978886756245e-06, "loss": 0.8867, "step": 213 }, { "epoch": 0.2054728756601056, "grad_norm": 0.9450794570746543, "learning_rate": 4.088291746641076e-06, "loss": 0.7141, "step": 214 }, { "epoch": 0.20643302928468554, "grad_norm": 1.2876863040113522, "learning_rate": 4.107485604606526e-06, "loss": 0.7882, "step": 215 }, { "epoch": 0.20739318290926548, "grad_norm": 0.9360362606273996, "learning_rate": 4.126679462571978e-06, "loss": 0.7836, "step": 216 }, { "epoch": 0.2083533365338454, "grad_norm": 0.9993279868710626, "learning_rate": 4.145873320537428e-06, "loss": 0.8082, "step": 217 }, { "epoch": 0.20931349015842535, "grad_norm": 1.235035850582973, "learning_rate": 4.1650671785028795e-06, "loss": 0.8004, "step": 218 }, { "epoch": 0.21027364378300528, "grad_norm": 0.9381179392568807, "learning_rate": 4.184261036468331e-06, "loss": 0.768, "step": 219 }, { "epoch": 0.21123379740758522, "grad_norm": 0.9910843644780881, "learning_rate": 4.203454894433781e-06, "loss": 0.7645, "step": 220 }, { "epoch": 0.21219395103216515, "grad_norm": 0.8682014386300204, "learning_rate": 4.222648752399233e-06, "loss": 0.8008, "step": 221 }, { "epoch": 0.21315410465674509, "grad_norm": 0.7996602061967457, "learning_rate": 4.241842610364684e-06, "loss": 0.6751, "step": 222 }, { "epoch": 0.21411425828132502, "grad_norm": 0.7643702838107727, "learning_rate": 4.2610364683301346e-06, "loss": 0.6308, "step": 223 }, { "epoch": 0.21507441190590496, "grad_norm": 0.7727580177547458, "learning_rate": 4.280230326295586e-06, "loss": 0.8039, "step": 224 }, { "epoch": 0.2160345655304849, "grad_norm": 0.781293954751679, "learning_rate": 4.299424184261036e-06, "loss": 0.7629, "step": 225 }, { "epoch": 0.2169947191550648, "grad_norm": 0.7522233335095655, "learning_rate": 4.318618042226488e-06, "loss": 0.642, "step": 226 }, { "epoch": 0.21795487277964473, "grad_norm": 0.8669878721583953, "learning_rate": 4.337811900191939e-06, "loss": 0.8282, "step": 227 }, { "epoch": 0.21891502640422467, "grad_norm": 0.8099107154934304, "learning_rate": 4.35700575815739e-06, "loss": 0.7028, "step": 228 }, { "epoch": 0.2198751800288046, "grad_norm": 0.9845100813661802, "learning_rate": 4.376199616122841e-06, "loss": 0.784, "step": 229 }, { "epoch": 0.22083533365338454, "grad_norm": 0.9113836995070514, "learning_rate": 4.395393474088292e-06, "loss": 0.8049, "step": 230 }, { "epoch": 0.22179548727796447, "grad_norm": 1.0758992008065693, "learning_rate": 4.414587332053743e-06, "loss": 0.6992, "step": 231 }, { "epoch": 0.2227556409025444, "grad_norm": 1.1843983331457841, "learning_rate": 4.433781190019194e-06, "loss": 0.6509, "step": 232 }, { "epoch": 0.22371579452712434, "grad_norm": 1.0007800534077709, "learning_rate": 4.4529750479846455e-06, "loss": 0.7226, "step": 233 }, { "epoch": 0.22467594815170427, "grad_norm": 0.9387246913062324, "learning_rate": 4.472168905950096e-06, "loss": 0.6871, "step": 234 }, { "epoch": 0.2256361017762842, "grad_norm": 0.6938370562080751, "learning_rate": 4.491362763915547e-06, "loss": 0.6215, "step": 235 }, { "epoch": 0.22659625540086414, "grad_norm": 0.950819416620061, "learning_rate": 4.510556621880998e-06, "loss": 0.7802, "step": 236 }, { "epoch": 0.22755640902544408, "grad_norm": 0.8017347335574332, "learning_rate": 4.52975047984645e-06, "loss": 0.6717, "step": 237 }, { "epoch": 0.228516562650024, "grad_norm": 0.8050428360307473, "learning_rate": 4.5489443378119005e-06, "loss": 0.6818, "step": 238 }, { "epoch": 0.22947671627460395, "grad_norm": 0.8821252924559155, "learning_rate": 4.568138195777352e-06, "loss": 0.7449, "step": 239 }, { "epoch": 0.23043686989918388, "grad_norm": 0.7856789312338511, "learning_rate": 4.587332053742803e-06, "loss": 0.6611, "step": 240 }, { "epoch": 0.2313970235237638, "grad_norm": 1.1156614311491966, "learning_rate": 4.606525911708254e-06, "loss": 0.876, "step": 241 }, { "epoch": 0.23235717714834372, "grad_norm": 0.989731508370998, "learning_rate": 4.625719769673705e-06, "loss": 0.7967, "step": 242 }, { "epoch": 0.23331733077292366, "grad_norm": 0.9015357846180558, "learning_rate": 4.644913627639156e-06, "loss": 0.7433, "step": 243 }, { "epoch": 0.2342774843975036, "grad_norm": 0.8786810687928809, "learning_rate": 4.664107485604607e-06, "loss": 0.7615, "step": 244 }, { "epoch": 0.23523763802208353, "grad_norm": 0.9345128153655058, "learning_rate": 4.683301343570058e-06, "loss": 0.8201, "step": 245 }, { "epoch": 0.23619779164666346, "grad_norm": 1.0208133998122781, "learning_rate": 4.702495201535509e-06, "loss": 0.8632, "step": 246 }, { "epoch": 0.2371579452712434, "grad_norm": 0.8501556488464239, "learning_rate": 4.72168905950096e-06, "loss": 0.6626, "step": 247 }, { "epoch": 0.23811809889582333, "grad_norm": 1.0652218568751168, "learning_rate": 4.7408829174664115e-06, "loss": 0.7735, "step": 248 }, { "epoch": 0.23907825252040327, "grad_norm": 0.9429293917908278, "learning_rate": 4.760076775431862e-06, "loss": 0.6908, "step": 249 }, { "epoch": 0.2400384061449832, "grad_norm": 1.3044634730089144, "learning_rate": 4.779270633397313e-06, "loss": 0.8315, "step": 250 }, { "epoch": 0.24099855976956314, "grad_norm": 0.9711856073142983, "learning_rate": 4.798464491362765e-06, "loss": 0.7651, "step": 251 }, { "epoch": 0.24195871339414307, "grad_norm": 0.9736694986014306, "learning_rate": 4.817658349328215e-06, "loss": 0.8393, "step": 252 }, { "epoch": 0.242918867018723, "grad_norm": 0.7774458485887504, "learning_rate": 4.8368522072936665e-06, "loss": 0.5797, "step": 253 }, { "epoch": 0.24387902064330294, "grad_norm": 1.0591775792188718, "learning_rate": 4.856046065259117e-06, "loss": 0.7421, "step": 254 }, { "epoch": 0.24483917426788285, "grad_norm": 0.9563369521829792, "learning_rate": 4.875239923224568e-06, "loss": 0.7936, "step": 255 }, { "epoch": 0.24579932789246278, "grad_norm": 0.9708903640858975, "learning_rate": 4.89443378119002e-06, "loss": 0.8621, "step": 256 }, { "epoch": 0.24675948151704272, "grad_norm": 1.2801579663778193, "learning_rate": 4.91362763915547e-06, "loss": 0.9548, "step": 257 }, { "epoch": 0.24771963514162265, "grad_norm": 1.1450405323562343, "learning_rate": 4.9328214971209215e-06, "loss": 0.8226, "step": 258 }, { "epoch": 0.2486797887662026, "grad_norm": 0.7859672246088835, "learning_rate": 4.952015355086373e-06, "loss": 0.7042, "step": 259 }, { "epoch": 0.24963994239078252, "grad_norm": 1.0326649266365506, "learning_rate": 4.971209213051823e-06, "loss": 0.6778, "step": 260 }, { "epoch": 0.2506000960153625, "grad_norm": 1.048222082023533, "learning_rate": 4.990403071017275e-06, "loss": 0.8508, "step": 261 }, { "epoch": 0.2515602496399424, "grad_norm": 0.8911576076025675, "learning_rate": 5.009596928982726e-06, "loss": 0.7549, "step": 262 }, { "epoch": 0.2525204032645223, "grad_norm": 0.916766123291802, "learning_rate": 5.028790786948177e-06, "loss": 0.7671, "step": 263 }, { "epoch": 0.25348055688910226, "grad_norm": 1.0631310721168257, "learning_rate": 5.047984644913628e-06, "loss": 0.8063, "step": 264 }, { "epoch": 0.25444071051368217, "grad_norm": 0.8725734386370761, "learning_rate": 5.0671785028790784e-06, "loss": 0.7131, "step": 265 }, { "epoch": 0.25540086413826213, "grad_norm": 0.8316536720489665, "learning_rate": 5.086372360844531e-06, "loss": 0.7265, "step": 266 }, { "epoch": 0.25636101776284204, "grad_norm": 0.9970748675835756, "learning_rate": 5.105566218809981e-06, "loss": 0.7454, "step": 267 }, { "epoch": 0.257321171387422, "grad_norm": 0.9535633771618619, "learning_rate": 5.1247600767754325e-06, "loss": 0.6571, "step": 268 }, { "epoch": 0.2582813250120019, "grad_norm": 0.8613538929648485, "learning_rate": 5.143953934740883e-06, "loss": 0.6365, "step": 269 }, { "epoch": 0.25924147863658187, "grad_norm": 0.7818720912793148, "learning_rate": 5.163147792706334e-06, "loss": 0.6689, "step": 270 }, { "epoch": 0.2602016322611618, "grad_norm": 1.1243668807255598, "learning_rate": 5.182341650671785e-06, "loss": 0.7093, "step": 271 }, { "epoch": 0.26116178588574174, "grad_norm": 0.8095388748784867, "learning_rate": 5.201535508637236e-06, "loss": 0.6491, "step": 272 }, { "epoch": 0.26212193951032164, "grad_norm": 0.8251123401259121, "learning_rate": 5.220729366602687e-06, "loss": 0.7029, "step": 273 }, { "epoch": 0.2630820931349016, "grad_norm": 1.0234225951327185, "learning_rate": 5.239923224568139e-06, "loss": 0.7787, "step": 274 }, { "epoch": 0.2640422467594815, "grad_norm": 0.8156395046260909, "learning_rate": 5.25911708253359e-06, "loss": 0.748, "step": 275 }, { "epoch": 0.2650024003840615, "grad_norm": 1.752619806906743, "learning_rate": 5.278310940499041e-06, "loss": 0.7011, "step": 276 }, { "epoch": 0.2659625540086414, "grad_norm": 1.2293508853562407, "learning_rate": 5.297504798464492e-06, "loss": 0.8511, "step": 277 }, { "epoch": 0.2669227076332213, "grad_norm": 1.129445196169881, "learning_rate": 5.3166986564299425e-06, "loss": 0.9215, "step": 278 }, { "epoch": 0.26788286125780125, "grad_norm": 0.9224240850409846, "learning_rate": 5.335892514395395e-06, "loss": 0.6751, "step": 279 }, { "epoch": 0.26884301488238116, "grad_norm": 0.9120723445639141, "learning_rate": 5.355086372360845e-06, "loss": 0.7545, "step": 280 }, { "epoch": 0.2698031685069611, "grad_norm": 0.928293365902081, "learning_rate": 5.374280230326297e-06, "loss": 0.6732, "step": 281 }, { "epoch": 0.27076332213154103, "grad_norm": 1.1353528466568101, "learning_rate": 5.393474088291747e-06, "loss": 0.7985, "step": 282 }, { "epoch": 0.271723475756121, "grad_norm": 0.7430753620315901, "learning_rate": 5.4126679462571984e-06, "loss": 0.6042, "step": 283 }, { "epoch": 0.2726836293807009, "grad_norm": 0.950211244085802, "learning_rate": 5.431861804222649e-06, "loss": 0.8025, "step": 284 }, { "epoch": 0.27364378300528086, "grad_norm": 0.937222529270597, "learning_rate": 5.4510556621881e-06, "loss": 0.7407, "step": 285 }, { "epoch": 0.27460393662986077, "grad_norm": 0.8050513093448161, "learning_rate": 5.470249520153551e-06, "loss": 0.7118, "step": 286 }, { "epoch": 0.27556409025444073, "grad_norm": 0.8864773942756776, "learning_rate": 5.489443378119003e-06, "loss": 0.8022, "step": 287 }, { "epoch": 0.27652424387902064, "grad_norm": 0.9738545523136333, "learning_rate": 5.5086372360844535e-06, "loss": 0.752, "step": 288 }, { "epoch": 0.2774843975036006, "grad_norm": 0.8411836328341493, "learning_rate": 5.527831094049905e-06, "loss": 0.7974, "step": 289 }, { "epoch": 0.2784445511281805, "grad_norm": 1.6181316367535141, "learning_rate": 5.547024952015355e-06, "loss": 0.9532, "step": 290 }, { "epoch": 0.27940470475276047, "grad_norm": 0.689439609812373, "learning_rate": 5.566218809980807e-06, "loss": 0.5648, "step": 291 }, { "epoch": 0.2803648583773404, "grad_norm": 0.8719705922054843, "learning_rate": 5.585412667946257e-06, "loss": 0.6899, "step": 292 }, { "epoch": 0.2813250120019203, "grad_norm": 1.0315205516034471, "learning_rate": 5.6046065259117085e-06, "loss": 0.742, "step": 293 }, { "epoch": 0.28228516562650025, "grad_norm": 1.0794723463522409, "learning_rate": 5.623800383877159e-06, "loss": 0.7762, "step": 294 }, { "epoch": 0.28324531925108015, "grad_norm": 1.4624702903960136, "learning_rate": 5.642994241842611e-06, "loss": 0.7658, "step": 295 }, { "epoch": 0.2842054728756601, "grad_norm": 1.0646870897936773, "learning_rate": 5.662188099808062e-06, "loss": 0.7729, "step": 296 }, { "epoch": 0.28516562650024, "grad_norm": 1.061479494714767, "learning_rate": 5.681381957773513e-06, "loss": 0.7771, "step": 297 }, { "epoch": 0.28612578012482, "grad_norm": 0.9006044793333582, "learning_rate": 5.7005758157389635e-06, "loss": 0.653, "step": 298 }, { "epoch": 0.2870859337493999, "grad_norm": 0.9710492104436653, "learning_rate": 5.719769673704415e-06, "loss": 0.8325, "step": 299 }, { "epoch": 0.28804608737397985, "grad_norm": 0.9502684335871433, "learning_rate": 5.738963531669866e-06, "loss": 0.7973, "step": 300 }, { "epoch": 0.28900624099855976, "grad_norm": 1.0168631582685015, "learning_rate": 5.758157389635317e-06, "loss": 0.7553, "step": 301 }, { "epoch": 0.2899663946231397, "grad_norm": 0.8849117174476667, "learning_rate": 5.777351247600769e-06, "loss": 0.6733, "step": 302 }, { "epoch": 0.29092654824771963, "grad_norm": 1.0278263331787785, "learning_rate": 5.7965451055662194e-06, "loss": 0.9032, "step": 303 }, { "epoch": 0.2918867018722996, "grad_norm": 0.8460976585048708, "learning_rate": 5.815738963531671e-06, "loss": 0.6759, "step": 304 }, { "epoch": 0.2928468554968795, "grad_norm": 0.7549584936079545, "learning_rate": 5.834932821497121e-06, "loss": 0.6553, "step": 305 }, { "epoch": 0.2938070091214594, "grad_norm": 1.019083797565957, "learning_rate": 5.854126679462573e-06, "loss": 0.7249, "step": 306 }, { "epoch": 0.29476716274603937, "grad_norm": 0.7513150256638818, "learning_rate": 5.873320537428023e-06, "loss": 0.6384, "step": 307 }, { "epoch": 0.2957273163706193, "grad_norm": 1.4179102015535212, "learning_rate": 5.892514395393475e-06, "loss": 0.8966, "step": 308 }, { "epoch": 0.29668746999519924, "grad_norm": 1.4728845662277128, "learning_rate": 5.911708253358925e-06, "loss": 0.7806, "step": 309 }, { "epoch": 0.29764762361977914, "grad_norm": 1.1040252414244298, "learning_rate": 5.930902111324377e-06, "loss": 0.7547, "step": 310 }, { "epoch": 0.2986077772443591, "grad_norm": 1.001446727779251, "learning_rate": 5.950095969289828e-06, "loss": 0.6046, "step": 311 }, { "epoch": 0.299567930868939, "grad_norm": 0.9269756583004459, "learning_rate": 5.969289827255279e-06, "loss": 0.6301, "step": 312 }, { "epoch": 0.300528084493519, "grad_norm": 0.7409498601503547, "learning_rate": 5.9884836852207295e-06, "loss": 0.6579, "step": 313 }, { "epoch": 0.3014882381180989, "grad_norm": 1.0721143550898384, "learning_rate": 6.007677543186181e-06, "loss": 0.7269, "step": 314 }, { "epoch": 0.30244839174267885, "grad_norm": 0.9699898418232433, "learning_rate": 6.026871401151631e-06, "loss": 0.6846, "step": 315 }, { "epoch": 0.30340854536725875, "grad_norm": 1.1651620261283724, "learning_rate": 6.0460652591170836e-06, "loss": 0.79, "step": 316 }, { "epoch": 0.3043686989918387, "grad_norm": 1.0355151310502482, "learning_rate": 6.065259117082534e-06, "loss": 0.8844, "step": 317 }, { "epoch": 0.3053288526164186, "grad_norm": 1.0985877795960834, "learning_rate": 6.084452975047985e-06, "loss": 0.705, "step": 318 }, { "epoch": 0.3062890062409986, "grad_norm": 0.9385179069979243, "learning_rate": 6.103646833013436e-06, "loss": 0.5915, "step": 319 }, { "epoch": 0.3072491598655785, "grad_norm": 1.2082990096498116, "learning_rate": 6.122840690978887e-06, "loss": 0.9059, "step": 320 }, { "epoch": 0.3082093134901584, "grad_norm": 1.36865024870494, "learning_rate": 6.142034548944338e-06, "loss": 0.762, "step": 321 }, { "epoch": 0.30916946711473836, "grad_norm": 1.0644866250680478, "learning_rate": 6.161228406909789e-06, "loss": 0.5954, "step": 322 }, { "epoch": 0.31012962073931827, "grad_norm": 0.8884721632848516, "learning_rate": 6.18042226487524e-06, "loss": 0.625, "step": 323 }, { "epoch": 0.31108977436389823, "grad_norm": 0.8818027321751728, "learning_rate": 6.199616122840692e-06, "loss": 0.7117, "step": 324 }, { "epoch": 0.31204992798847814, "grad_norm": 0.9021583229891907, "learning_rate": 6.218809980806143e-06, "loss": 0.7507, "step": 325 }, { "epoch": 0.3130100816130581, "grad_norm": 0.7938817273886359, "learning_rate": 6.238003838771594e-06, "loss": 0.6891, "step": 326 }, { "epoch": 0.313970235237638, "grad_norm": 0.8215889850024551, "learning_rate": 6.257197696737045e-06, "loss": 0.675, "step": 327 }, { "epoch": 0.31493038886221797, "grad_norm": 0.9151622038938523, "learning_rate": 6.2763915547024955e-06, "loss": 0.7367, "step": 328 }, { "epoch": 0.3158905424867979, "grad_norm": 1.1316871515580638, "learning_rate": 6.295585412667947e-06, "loss": 0.7499, "step": 329 }, { "epoch": 0.31685069611137784, "grad_norm": 1.2065073515652582, "learning_rate": 6.314779270633397e-06, "loss": 0.7957, "step": 330 }, { "epoch": 0.31781084973595775, "grad_norm": 0.9208041908047743, "learning_rate": 6.3339731285988495e-06, "loss": 0.8203, "step": 331 }, { "epoch": 0.3187710033605377, "grad_norm": 0.7791978510371851, "learning_rate": 6.3531669865643e-06, "loss": 0.6661, "step": 332 }, { "epoch": 0.3197311569851176, "grad_norm": 0.7418045471585164, "learning_rate": 6.372360844529751e-06, "loss": 0.6463, "step": 333 }, { "epoch": 0.3206913106096976, "grad_norm": 1.0565981177256327, "learning_rate": 6.391554702495202e-06, "loss": 0.7707, "step": 334 }, { "epoch": 0.3216514642342775, "grad_norm": 0.9031724116469804, "learning_rate": 6.410748560460653e-06, "loss": 0.6793, "step": 335 }, { "epoch": 0.3226116178588574, "grad_norm": 0.7781773001393425, "learning_rate": 6.429942418426104e-06, "loss": 0.6459, "step": 336 }, { "epoch": 0.32357177148343735, "grad_norm": 0.8593452025568374, "learning_rate": 6.449136276391556e-06, "loss": 0.6822, "step": 337 }, { "epoch": 0.32453192510801726, "grad_norm": 0.8186754592663917, "learning_rate": 6.4683301343570056e-06, "loss": 0.6564, "step": 338 }, { "epoch": 0.3254920787325972, "grad_norm": 1.0916092993734632, "learning_rate": 6.487523992322458e-06, "loss": 0.764, "step": 339 }, { "epoch": 0.32645223235717713, "grad_norm": 0.8710694559600531, "learning_rate": 6.506717850287908e-06, "loss": 0.6784, "step": 340 }, { "epoch": 0.3274123859817571, "grad_norm": 0.8110415704911602, "learning_rate": 6.52591170825336e-06, "loss": 0.7387, "step": 341 }, { "epoch": 0.328372539606337, "grad_norm": 0.8828556649437408, "learning_rate": 6.54510556621881e-06, "loss": 0.7287, "step": 342 }, { "epoch": 0.32933269323091696, "grad_norm": 1.2725246720932115, "learning_rate": 6.5642994241842614e-06, "loss": 0.811, "step": 343 }, { "epoch": 0.33029284685549687, "grad_norm": 0.7844255792123335, "learning_rate": 6.583493282149712e-06, "loss": 0.6574, "step": 344 }, { "epoch": 0.33125300048007683, "grad_norm": 1.0594898359055331, "learning_rate": 6.602687140115164e-06, "loss": 0.8161, "step": 345 }, { "epoch": 0.33221315410465674, "grad_norm": 0.9067816472002763, "learning_rate": 6.621880998080615e-06, "loss": 0.7432, "step": 346 }, { "epoch": 0.3331733077292367, "grad_norm": 0.9779323895342715, "learning_rate": 6.641074856046066e-06, "loss": 0.6776, "step": 347 }, { "epoch": 0.3341334613538166, "grad_norm": 1.0175978054891786, "learning_rate": 6.6602687140115165e-06, "loss": 0.8015, "step": 348 }, { "epoch": 0.33509361497839657, "grad_norm": 0.883872534032956, "learning_rate": 6.679462571976968e-06, "loss": 0.7344, "step": 349 }, { "epoch": 0.3360537686029765, "grad_norm": 1.0114539216295986, "learning_rate": 6.698656429942419e-06, "loss": 0.8707, "step": 350 }, { "epoch": 0.3370139222275564, "grad_norm": 1.5193095014953575, "learning_rate": 6.71785028790787e-06, "loss": 0.8816, "step": 351 }, { "epoch": 0.33797407585213635, "grad_norm": 1.4661463945122115, "learning_rate": 6.737044145873322e-06, "loss": 0.8298, "step": 352 }, { "epoch": 0.33893422947671625, "grad_norm": 1.312325532593007, "learning_rate": 6.756238003838772e-06, "loss": 0.8697, "step": 353 }, { "epoch": 0.3398943831012962, "grad_norm": 1.1016074517088699, "learning_rate": 6.775431861804224e-06, "loss": 0.8743, "step": 354 }, { "epoch": 0.3408545367258761, "grad_norm": 0.8276345105181889, "learning_rate": 6.794625719769674e-06, "loss": 0.6648, "step": 355 }, { "epoch": 0.3418146903504561, "grad_norm": 0.9267569574587438, "learning_rate": 6.8138195777351256e-06, "loss": 0.8232, "step": 356 }, { "epoch": 0.342774843975036, "grad_norm": 1.3826200787324738, "learning_rate": 6.833013435700576e-06, "loss": 0.9685, "step": 357 }, { "epoch": 0.34373499759961595, "grad_norm": 1.041348315569698, "learning_rate": 6.852207293666027e-06, "loss": 0.7208, "step": 358 }, { "epoch": 0.34469515122419586, "grad_norm": 0.8621643542976296, "learning_rate": 6.871401151631478e-06, "loss": 0.7213, "step": 359 }, { "epoch": 0.3456553048487758, "grad_norm": 0.7300608009668774, "learning_rate": 6.89059500959693e-06, "loss": 0.671, "step": 360 }, { "epoch": 0.34661545847335573, "grad_norm": 0.81633518410933, "learning_rate": 6.909788867562381e-06, "loss": 0.7039, "step": 361 }, { "epoch": 0.3475756120979357, "grad_norm": 1.1546732002616575, "learning_rate": 6.928982725527832e-06, "loss": 0.8167, "step": 362 }, { "epoch": 0.3485357657225156, "grad_norm": 0.8581519667886864, "learning_rate": 6.9481765834932824e-06, "loss": 0.788, "step": 363 }, { "epoch": 0.34949591934709556, "grad_norm": 0.8835627645581613, "learning_rate": 6.967370441458734e-06, "loss": 0.7079, "step": 364 }, { "epoch": 0.35045607297167547, "grad_norm": 0.8827779471795008, "learning_rate": 6.986564299424184e-06, "loss": 0.6996, "step": 365 }, { "epoch": 0.3514162265962554, "grad_norm": 0.8953555125042707, "learning_rate": 7.005758157389636e-06, "loss": 0.6893, "step": 366 }, { "epoch": 0.35237638022083534, "grad_norm": 0.7831569602650763, "learning_rate": 7.024952015355086e-06, "loss": 0.6336, "step": 367 }, { "epoch": 0.35333653384541525, "grad_norm": 0.9858149305192234, "learning_rate": 7.044145873320538e-06, "loss": 0.8315, "step": 368 }, { "epoch": 0.3542966874699952, "grad_norm": 0.7844848679088969, "learning_rate": 7.063339731285989e-06, "loss": 0.6895, "step": 369 }, { "epoch": 0.3552568410945751, "grad_norm": 1.0830862090205244, "learning_rate": 7.08253358925144e-06, "loss": 0.7906, "step": 370 }, { "epoch": 0.3562169947191551, "grad_norm": 0.9399204455650686, "learning_rate": 7.101727447216891e-06, "loss": 0.7711, "step": 371 }, { "epoch": 0.357177148343735, "grad_norm": 0.8813139222956403, "learning_rate": 7.120921305182342e-06, "loss": 0.7288, "step": 372 }, { "epoch": 0.35813730196831495, "grad_norm": 0.9282003618819955, "learning_rate": 7.1401151631477925e-06, "loss": 0.6549, "step": 373 }, { "epoch": 0.35909745559289485, "grad_norm": 1.3890146942157322, "learning_rate": 7.159309021113245e-06, "loss": 0.8178, "step": 374 }, { "epoch": 0.3600576092174748, "grad_norm": 0.8577123862128432, "learning_rate": 7.178502879078696e-06, "loss": 0.7455, "step": 375 }, { "epoch": 0.3610177628420547, "grad_norm": 0.7206137334122115, "learning_rate": 7.1976967370441466e-06, "loss": 0.6044, "step": 376 }, { "epoch": 0.3619779164666347, "grad_norm": 1.0877917831405064, "learning_rate": 7.216890595009598e-06, "loss": 0.7161, "step": 377 }, { "epoch": 0.3629380700912146, "grad_norm": 1.0685480024983598, "learning_rate": 7.236084452975048e-06, "loss": 0.7652, "step": 378 }, { "epoch": 0.3638982237157945, "grad_norm": 0.9038801960931119, "learning_rate": 7.2552783109405e-06, "loss": 0.609, "step": 379 }, { "epoch": 0.36485837734037446, "grad_norm": 1.0754122529850099, "learning_rate": 7.27447216890595e-06, "loss": 0.7068, "step": 380 }, { "epoch": 0.36581853096495437, "grad_norm": 0.8852596187614097, "learning_rate": 7.2936660268714024e-06, "loss": 0.7428, "step": 381 }, { "epoch": 0.36677868458953433, "grad_norm": 0.8096950183070505, "learning_rate": 7.312859884836853e-06, "loss": 0.7185, "step": 382 }, { "epoch": 0.36773883821411424, "grad_norm": 1.2536664344782649, "learning_rate": 7.332053742802304e-06, "loss": 0.7803, "step": 383 }, { "epoch": 0.3686989918386942, "grad_norm": 0.8897189512407644, "learning_rate": 7.351247600767755e-06, "loss": 0.7419, "step": 384 }, { "epoch": 0.3696591454632741, "grad_norm": 1.0270195200866892, "learning_rate": 7.370441458733206e-06, "loss": 0.7929, "step": 385 }, { "epoch": 0.37061929908785407, "grad_norm": 1.077427696139474, "learning_rate": 7.389635316698657e-06, "loss": 0.6708, "step": 386 }, { "epoch": 0.371579452712434, "grad_norm": 0.9660909674138548, "learning_rate": 7.408829174664108e-06, "loss": 0.822, "step": 387 }, { "epoch": 0.37253960633701394, "grad_norm": 0.9062012183412091, "learning_rate": 7.4280230326295585e-06, "loss": 0.6835, "step": 388 }, { "epoch": 0.37349975996159385, "grad_norm": 0.7675553053084399, "learning_rate": 7.447216890595011e-06, "loss": 0.683, "step": 389 }, { "epoch": 0.3744599135861738, "grad_norm": 0.9513214832183791, "learning_rate": 7.466410748560461e-06, "loss": 0.6587, "step": 390 }, { "epoch": 0.3754200672107537, "grad_norm": 0.9232056513132132, "learning_rate": 7.4856046065259125e-06, "loss": 0.7324, "step": 391 }, { "epoch": 0.3763802208353337, "grad_norm": 0.9703744118943737, "learning_rate": 7.504798464491363e-06, "loss": 0.7771, "step": 392 }, { "epoch": 0.3773403744599136, "grad_norm": 1.192960787035273, "learning_rate": 7.523992322456814e-06, "loss": 0.7994, "step": 393 }, { "epoch": 0.3783005280844935, "grad_norm": 1.0765839765016048, "learning_rate": 7.543186180422265e-06, "loss": 0.8324, "step": 394 }, { "epoch": 0.37926068170907346, "grad_norm": 0.795823099441076, "learning_rate": 7.562380038387716e-06, "loss": 0.6231, "step": 395 }, { "epoch": 0.38022083533365336, "grad_norm": 0.8522183546632062, "learning_rate": 7.581573896353167e-06, "loss": 0.6785, "step": 396 }, { "epoch": 0.3811809889582333, "grad_norm": 1.0443180047149523, "learning_rate": 7.600767754318619e-06, "loss": 0.7922, "step": 397 }, { "epoch": 0.38214114258281323, "grad_norm": 0.9937672995880015, "learning_rate": 7.61996161228407e-06, "loss": 0.6542, "step": 398 }, { "epoch": 0.3831012962073932, "grad_norm": 0.8209289093448964, "learning_rate": 7.63915547024952e-06, "loss": 0.7421, "step": 399 }, { "epoch": 0.3840614498319731, "grad_norm": 0.9313930520207219, "learning_rate": 7.658349328214972e-06, "loss": 0.663, "step": 400 }, { "epoch": 0.38502160345655306, "grad_norm": 0.8260016865429053, "learning_rate": 7.677543186180423e-06, "loss": 0.6584, "step": 401 }, { "epoch": 0.38598175708113297, "grad_norm": 1.2954834365039563, "learning_rate": 7.696737044145875e-06, "loss": 0.6643, "step": 402 }, { "epoch": 0.38694191070571293, "grad_norm": 0.9621881942442273, "learning_rate": 7.715930902111324e-06, "loss": 0.7686, "step": 403 }, { "epoch": 0.38790206433029284, "grad_norm": 0.9559803460218635, "learning_rate": 7.735124760076776e-06, "loss": 0.8068, "step": 404 }, { "epoch": 0.3888622179548728, "grad_norm": 0.9717273201693036, "learning_rate": 7.754318618042227e-06, "loss": 0.7319, "step": 405 }, { "epoch": 0.3898223715794527, "grad_norm": 0.8863545711018999, "learning_rate": 7.773512476007678e-06, "loss": 0.6917, "step": 406 }, { "epoch": 0.39078252520403267, "grad_norm": 1.1158016541929965, "learning_rate": 7.79270633397313e-06, "loss": 0.7673, "step": 407 }, { "epoch": 0.3917426788286126, "grad_norm": 0.9850554814938863, "learning_rate": 7.811900191938581e-06, "loss": 0.7988, "step": 408 }, { "epoch": 0.3927028324531925, "grad_norm": 1.1796646032785665, "learning_rate": 7.83109404990403e-06, "loss": 0.7526, "step": 409 }, { "epoch": 0.39366298607777245, "grad_norm": 0.992609456727221, "learning_rate": 7.850287907869482e-06, "loss": 0.7645, "step": 410 }, { "epoch": 0.39462313970235235, "grad_norm": 0.8459459681150058, "learning_rate": 7.869481765834934e-06, "loss": 0.7252, "step": 411 }, { "epoch": 0.3955832933269323, "grad_norm": 0.9822339528249302, "learning_rate": 7.888675623800385e-06, "loss": 0.7524, "step": 412 }, { "epoch": 0.3965434469515122, "grad_norm": 1.0076198323953003, "learning_rate": 7.907869481765835e-06, "loss": 0.7532, "step": 413 }, { "epoch": 0.3975036005760922, "grad_norm": 0.8991187882751343, "learning_rate": 7.927063339731288e-06, "loss": 0.6892, "step": 414 }, { "epoch": 0.3984637542006721, "grad_norm": 0.7648617718453893, "learning_rate": 7.946257197696737e-06, "loss": 0.7228, "step": 415 }, { "epoch": 0.39942390782525206, "grad_norm": 0.8250161879847473, "learning_rate": 7.965451055662189e-06, "loss": 0.6813, "step": 416 }, { "epoch": 0.40038406144983196, "grad_norm": 0.8018138278635785, "learning_rate": 7.98464491362764e-06, "loss": 0.7205, "step": 417 }, { "epoch": 0.4013442150744119, "grad_norm": 1.0017308303249142, "learning_rate": 8.003838771593091e-06, "loss": 0.79, "step": 418 }, { "epoch": 0.40230436869899183, "grad_norm": 1.4064292131790226, "learning_rate": 8.023032629558541e-06, "loss": 0.7854, "step": 419 }, { "epoch": 0.4032645223235718, "grad_norm": 0.9132005328162996, "learning_rate": 8.042226487523992e-06, "loss": 0.7789, "step": 420 }, { "epoch": 0.4042246759481517, "grad_norm": 1.047515813971491, "learning_rate": 8.061420345489444e-06, "loss": 0.7707, "step": 421 }, { "epoch": 0.40518482957273166, "grad_norm": 1.0132428150818273, "learning_rate": 8.080614203454895e-06, "loss": 0.7708, "step": 422 }, { "epoch": 0.40614498319731157, "grad_norm": 0.956006460656562, "learning_rate": 8.099808061420346e-06, "loss": 0.7072, "step": 423 }, { "epoch": 0.4071051368218915, "grad_norm": 1.2592841639619778, "learning_rate": 8.119001919385798e-06, "loss": 0.8154, "step": 424 }, { "epoch": 0.40806529044647144, "grad_norm": 1.132909194875298, "learning_rate": 8.138195777351249e-06, "loss": 0.7413, "step": 425 }, { "epoch": 0.40902544407105135, "grad_norm": 1.183838848987664, "learning_rate": 8.157389635316699e-06, "loss": 0.8277, "step": 426 }, { "epoch": 0.4099855976956313, "grad_norm": 1.1281680378286056, "learning_rate": 8.176583493282152e-06, "loss": 0.7577, "step": 427 }, { "epoch": 0.4109457513202112, "grad_norm": 1.4823563755915514, "learning_rate": 8.195777351247601e-06, "loss": 0.7844, "step": 428 }, { "epoch": 0.4119059049447912, "grad_norm": 1.441711998299161, "learning_rate": 8.214971209213053e-06, "loss": 0.7906, "step": 429 }, { "epoch": 0.4128660585693711, "grad_norm": 0.897185384055612, "learning_rate": 8.234165067178504e-06, "loss": 0.7383, "step": 430 }, { "epoch": 0.41382621219395105, "grad_norm": 0.9927968025474867, "learning_rate": 8.253358925143955e-06, "loss": 0.7383, "step": 431 }, { "epoch": 0.41478636581853096, "grad_norm": 0.8163391981379262, "learning_rate": 8.272552783109405e-06, "loss": 0.6529, "step": 432 }, { "epoch": 0.4157465194431109, "grad_norm": 0.9640906024231488, "learning_rate": 8.291746641074856e-06, "loss": 0.7183, "step": 433 }, { "epoch": 0.4167066730676908, "grad_norm": 1.0954342997959736, "learning_rate": 8.310940499040308e-06, "loss": 0.7856, "step": 434 }, { "epoch": 0.4176668266922708, "grad_norm": 0.7311765641356345, "learning_rate": 8.330134357005759e-06, "loss": 0.6412, "step": 435 }, { "epoch": 0.4186269803168507, "grad_norm": 0.8378715191753644, "learning_rate": 8.34932821497121e-06, "loss": 0.6886, "step": 436 }, { "epoch": 0.4195871339414306, "grad_norm": 1.0956620252926728, "learning_rate": 8.368522072936662e-06, "loss": 0.8398, "step": 437 }, { "epoch": 0.42054728756601056, "grad_norm": 0.9777789006849532, "learning_rate": 8.387715930902111e-06, "loss": 0.7395, "step": 438 }, { "epoch": 0.42150744119059047, "grad_norm": 0.7415710984238854, "learning_rate": 8.406909788867563e-06, "loss": 0.7155, "step": 439 }, { "epoch": 0.42246759481517043, "grad_norm": 1.3102459639515294, "learning_rate": 8.426103646833014e-06, "loss": 0.945, "step": 440 }, { "epoch": 0.42342774843975034, "grad_norm": 0.9867508174899979, "learning_rate": 8.445297504798465e-06, "loss": 0.7323, "step": 441 }, { "epoch": 0.4243879020643303, "grad_norm": 0.944913156375559, "learning_rate": 8.464491362763915e-06, "loss": 0.7212, "step": 442 }, { "epoch": 0.4253480556889102, "grad_norm": 1.1133736062026403, "learning_rate": 8.483685220729368e-06, "loss": 0.6782, "step": 443 }, { "epoch": 0.42630820931349017, "grad_norm": 1.262853549567799, "learning_rate": 8.502879078694818e-06, "loss": 0.8496, "step": 444 }, { "epoch": 0.4272683629380701, "grad_norm": 1.0852277508008055, "learning_rate": 8.522072936660269e-06, "loss": 0.7713, "step": 445 }, { "epoch": 0.42822851656265004, "grad_norm": 0.8767097917658855, "learning_rate": 8.54126679462572e-06, "loss": 0.7025, "step": 446 }, { "epoch": 0.42918867018722995, "grad_norm": 0.9639174467066827, "learning_rate": 8.560460652591172e-06, "loss": 0.7014, "step": 447 }, { "epoch": 0.4301488238118099, "grad_norm": 1.3366331100759272, "learning_rate": 8.579654510556623e-06, "loss": 0.8542, "step": 448 }, { "epoch": 0.4311089774363898, "grad_norm": 1.0604554032024327, "learning_rate": 8.598848368522073e-06, "loss": 0.7313, "step": 449 }, { "epoch": 0.4320691310609698, "grad_norm": 1.2494204227274819, "learning_rate": 8.618042226487526e-06, "loss": 0.749, "step": 450 }, { "epoch": 0.4330292846855497, "grad_norm": 0.8377361032093765, "learning_rate": 8.637236084452976e-06, "loss": 0.7571, "step": 451 }, { "epoch": 0.4339894383101296, "grad_norm": 0.8350444278318566, "learning_rate": 8.656429942418427e-06, "loss": 0.6788, "step": 452 }, { "epoch": 0.43494959193470956, "grad_norm": 0.861382102683186, "learning_rate": 8.675623800383878e-06, "loss": 0.6628, "step": 453 }, { "epoch": 0.43590974555928946, "grad_norm": 0.7832190165948423, "learning_rate": 8.69481765834933e-06, "loss": 0.6056, "step": 454 }, { "epoch": 0.4368698991838694, "grad_norm": 1.4114033601105525, "learning_rate": 8.71401151631478e-06, "loss": 0.7722, "step": 455 }, { "epoch": 0.43783005280844933, "grad_norm": 0.910201895873847, "learning_rate": 8.73320537428023e-06, "loss": 0.7198, "step": 456 }, { "epoch": 0.4387902064330293, "grad_norm": 0.9320564232758569, "learning_rate": 8.752399232245682e-06, "loss": 0.6717, "step": 457 }, { "epoch": 0.4397503600576092, "grad_norm": 1.1788320675973982, "learning_rate": 8.771593090211133e-06, "loss": 0.8048, "step": 458 }, { "epoch": 0.44071051368218916, "grad_norm": 1.2997473053732969, "learning_rate": 8.790786948176585e-06, "loss": 0.797, "step": 459 }, { "epoch": 0.44167066730676907, "grad_norm": 0.9225681681342591, "learning_rate": 8.809980806142036e-06, "loss": 0.6532, "step": 460 }, { "epoch": 0.44263082093134903, "grad_norm": 1.1312982826279123, "learning_rate": 8.829174664107486e-06, "loss": 0.7009, "step": 461 }, { "epoch": 0.44359097455592894, "grad_norm": 1.1294486711655973, "learning_rate": 8.848368522072937e-06, "loss": 0.8262, "step": 462 }, { "epoch": 0.4445511281805089, "grad_norm": 0.8659119599494991, "learning_rate": 8.867562380038388e-06, "loss": 0.6966, "step": 463 }, { "epoch": 0.4455112818050888, "grad_norm": 1.219132975832696, "learning_rate": 8.88675623800384e-06, "loss": 0.8124, "step": 464 }, { "epoch": 0.4464714354296688, "grad_norm": 1.0042876437346029, "learning_rate": 8.905950095969291e-06, "loss": 0.6861, "step": 465 }, { "epoch": 0.4474315890542487, "grad_norm": 0.9449265489542873, "learning_rate": 8.925143953934742e-06, "loss": 0.7913, "step": 466 }, { "epoch": 0.4483917426788286, "grad_norm": 0.8569995755306987, "learning_rate": 8.944337811900192e-06, "loss": 0.7178, "step": 467 }, { "epoch": 0.44935189630340855, "grad_norm": 1.1805639663539265, "learning_rate": 8.963531669865643e-06, "loss": 0.7936, "step": 468 }, { "epoch": 0.45031204992798846, "grad_norm": 0.8400011961357207, "learning_rate": 8.982725527831095e-06, "loss": 0.5689, "step": 469 }, { "epoch": 0.4512722035525684, "grad_norm": 0.8851024334525514, "learning_rate": 9.001919385796546e-06, "loss": 0.6402, "step": 470 }, { "epoch": 0.4522323571771483, "grad_norm": 0.7967004944764023, "learning_rate": 9.021113243761996e-06, "loss": 0.6614, "step": 471 }, { "epoch": 0.4531925108017283, "grad_norm": 0.9414275310200424, "learning_rate": 9.040307101727449e-06, "loss": 0.7202, "step": 472 }, { "epoch": 0.4541526644263082, "grad_norm": 0.9826403683073871, "learning_rate": 9.0595009596929e-06, "loss": 0.7692, "step": 473 }, { "epoch": 0.45511281805088816, "grad_norm": 1.0110193572222188, "learning_rate": 9.07869481765835e-06, "loss": 0.6667, "step": 474 }, { "epoch": 0.45607297167546806, "grad_norm": 1.2767578865155917, "learning_rate": 9.097888675623801e-06, "loss": 0.6725, "step": 475 }, { "epoch": 0.457033125300048, "grad_norm": 1.1732435528216898, "learning_rate": 9.117082533589252e-06, "loss": 0.7623, "step": 476 }, { "epoch": 0.45799327892462793, "grad_norm": 1.2441028273216965, "learning_rate": 9.136276391554704e-06, "loss": 0.7001, "step": 477 }, { "epoch": 0.4589534325492079, "grad_norm": 1.3316738835354138, "learning_rate": 9.155470249520153e-06, "loss": 0.7998, "step": 478 }, { "epoch": 0.4599135861737878, "grad_norm": 0.7686268265350286, "learning_rate": 9.174664107485606e-06, "loss": 0.61, "step": 479 }, { "epoch": 0.46087373979836777, "grad_norm": 0.8018850099742323, "learning_rate": 9.193857965451056e-06, "loss": 0.6489, "step": 480 }, { "epoch": 0.46183389342294767, "grad_norm": 0.8208388864723809, "learning_rate": 9.213051823416507e-06, "loss": 0.6934, "step": 481 }, { "epoch": 0.4627940470475276, "grad_norm": 0.9665426846366839, "learning_rate": 9.232245681381959e-06, "loss": 0.6792, "step": 482 }, { "epoch": 0.46375420067210754, "grad_norm": 1.007128820924409, "learning_rate": 9.25143953934741e-06, "loss": 0.724, "step": 483 }, { "epoch": 0.46471435429668745, "grad_norm": 1.0261997682975148, "learning_rate": 9.27063339731286e-06, "loss": 0.5946, "step": 484 }, { "epoch": 0.4656745079212674, "grad_norm": 1.333261080468342, "learning_rate": 9.289827255278311e-06, "loss": 0.8718, "step": 485 }, { "epoch": 0.4666346615458473, "grad_norm": 0.9671887707242365, "learning_rate": 9.309021113243763e-06, "loss": 0.7487, "step": 486 }, { "epoch": 0.4675948151704273, "grad_norm": 0.8497059991950121, "learning_rate": 9.328214971209214e-06, "loss": 0.7079, "step": 487 }, { "epoch": 0.4685549687950072, "grad_norm": 0.7249198242516967, "learning_rate": 9.347408829174665e-06, "loss": 0.6165, "step": 488 }, { "epoch": 0.46951512241958715, "grad_norm": 0.8229069834132591, "learning_rate": 9.366602687140117e-06, "loss": 0.6627, "step": 489 }, { "epoch": 0.47047527604416706, "grad_norm": 1.3385115084222283, "learning_rate": 9.385796545105566e-06, "loss": 0.8156, "step": 490 }, { "epoch": 0.471435429668747, "grad_norm": 0.974624649951756, "learning_rate": 9.404990403071018e-06, "loss": 0.7504, "step": 491 }, { "epoch": 0.4723955832933269, "grad_norm": 1.1080632677358204, "learning_rate": 9.424184261036469e-06, "loss": 0.8421, "step": 492 }, { "epoch": 0.4733557369179069, "grad_norm": 0.8556613743533831, "learning_rate": 9.44337811900192e-06, "loss": 0.692, "step": 493 }, { "epoch": 0.4743158905424868, "grad_norm": 0.9254785492516286, "learning_rate": 9.46257197696737e-06, "loss": 0.6998, "step": 494 }, { "epoch": 0.47527604416706676, "grad_norm": 0.86622092826884, "learning_rate": 9.481765834932823e-06, "loss": 0.7397, "step": 495 }, { "epoch": 0.47623619779164666, "grad_norm": 1.0019871046494062, "learning_rate": 9.500959692898273e-06, "loss": 0.85, "step": 496 }, { "epoch": 0.47719635141622657, "grad_norm": 0.8617958102504633, "learning_rate": 9.520153550863724e-06, "loss": 0.776, "step": 497 }, { "epoch": 0.47815650504080653, "grad_norm": 1.2562130137268408, "learning_rate": 9.539347408829175e-06, "loss": 0.8099, "step": 498 }, { "epoch": 0.47911665866538644, "grad_norm": 0.9821453309391149, "learning_rate": 9.558541266794627e-06, "loss": 0.6773, "step": 499 }, { "epoch": 0.4800768122899664, "grad_norm": 1.05828910179939, "learning_rate": 9.577735124760078e-06, "loss": 0.7168, "step": 500 }, { "epoch": 0.4810369659145463, "grad_norm": 0.9868640429141303, "learning_rate": 9.59692898272553e-06, "loss": 0.8114, "step": 501 }, { "epoch": 0.4819971195391263, "grad_norm": 1.1829143668484243, "learning_rate": 9.61612284069098e-06, "loss": 0.6872, "step": 502 }, { "epoch": 0.4829572731637062, "grad_norm": 0.8615635381514307, "learning_rate": 9.63531669865643e-06, "loss": 0.6738, "step": 503 }, { "epoch": 0.48391742678828614, "grad_norm": 0.8209379963363762, "learning_rate": 9.654510556621882e-06, "loss": 0.7174, "step": 504 }, { "epoch": 0.48487758041286605, "grad_norm": 0.9115540949646725, "learning_rate": 9.673704414587333e-06, "loss": 0.648, "step": 505 }, { "epoch": 0.485837734037446, "grad_norm": 1.2409256804151456, "learning_rate": 9.692898272552784e-06, "loss": 0.7622, "step": 506 }, { "epoch": 0.4867978876620259, "grad_norm": 1.1560942111046362, "learning_rate": 9.712092130518234e-06, "loss": 0.7796, "step": 507 }, { "epoch": 0.4877580412866059, "grad_norm": 0.8035299048529602, "learning_rate": 9.731285988483687e-06, "loss": 0.656, "step": 508 }, { "epoch": 0.4887181949111858, "grad_norm": 1.036342331605695, "learning_rate": 9.750479846449137e-06, "loss": 0.7831, "step": 509 }, { "epoch": 0.4896783485357657, "grad_norm": 1.1180093118376802, "learning_rate": 9.769673704414588e-06, "loss": 0.7805, "step": 510 }, { "epoch": 0.49063850216034566, "grad_norm": 1.0402475871781887, "learning_rate": 9.78886756238004e-06, "loss": 0.7002, "step": 511 }, { "epoch": 0.49159865578492556, "grad_norm": 1.1158613252315968, "learning_rate": 9.80806142034549e-06, "loss": 0.7189, "step": 512 }, { "epoch": 0.4925588094095055, "grad_norm": 1.0850717010763868, "learning_rate": 9.82725527831094e-06, "loss": 0.7988, "step": 513 }, { "epoch": 0.49351896303408543, "grad_norm": 0.8691112939655864, "learning_rate": 9.846449136276392e-06, "loss": 0.6439, "step": 514 }, { "epoch": 0.4944791166586654, "grad_norm": 0.8290716367877923, "learning_rate": 9.865642994241843e-06, "loss": 0.6416, "step": 515 }, { "epoch": 0.4954392702832453, "grad_norm": 1.1768877707596828, "learning_rate": 9.884836852207294e-06, "loss": 0.8045, "step": 516 }, { "epoch": 0.49639942390782527, "grad_norm": 0.9725899579746908, "learning_rate": 9.904030710172746e-06, "loss": 0.8027, "step": 517 }, { "epoch": 0.4973595775324052, "grad_norm": 1.04502114586642, "learning_rate": 9.923224568138197e-06, "loss": 0.7518, "step": 518 }, { "epoch": 0.49831973115698514, "grad_norm": 0.8376418958087454, "learning_rate": 9.942418426103647e-06, "loss": 0.6565, "step": 519 }, { "epoch": 0.49927988478156504, "grad_norm": 1.0166873465455275, "learning_rate": 9.961612284069098e-06, "loss": 0.7358, "step": 520 }, { "epoch": 0.500240038406145, "grad_norm": 1.076055157170506, "learning_rate": 9.98080614203455e-06, "loss": 0.8593, "step": 521 }, { "epoch": 0.501200192030725, "grad_norm": 0.7962440751442428, "learning_rate": 1e-05, "loss": 0.7254, "step": 522 }, { "epoch": 0.5021603456553049, "grad_norm": 1.4848304927760054, "learning_rate": 9.999998877776725e-06, "loss": 0.7766, "step": 523 }, { "epoch": 0.5031204992798848, "grad_norm": 1.051882510661786, "learning_rate": 9.999995511107399e-06, "loss": 0.6751, "step": 524 }, { "epoch": 0.5040806529044647, "grad_norm": 0.9011966164143667, "learning_rate": 9.999989899993536e-06, "loss": 0.6707, "step": 525 }, { "epoch": 0.5050408065290446, "grad_norm": 0.9932963992052919, "learning_rate": 9.999982044437652e-06, "loss": 0.7539, "step": 526 }, { "epoch": 0.5060009601536246, "grad_norm": 1.1707754760106204, "learning_rate": 9.999971944443276e-06, "loss": 0.7675, "step": 527 }, { "epoch": 0.5069611137782045, "grad_norm": 0.9703938608986256, "learning_rate": 9.999959600014942e-06, "loss": 0.7173, "step": 528 }, { "epoch": 0.5079212674027844, "grad_norm": 0.8678152896967117, "learning_rate": 9.999945011158189e-06, "loss": 0.674, "step": 529 }, { "epoch": 0.5088814210273643, "grad_norm": 0.9138129049823934, "learning_rate": 9.999928177879567e-06, "loss": 0.6669, "step": 530 }, { "epoch": 0.5098415746519444, "grad_norm": 1.1069091317003281, "learning_rate": 9.999909100186634e-06, "loss": 0.7769, "step": 531 }, { "epoch": 0.5108017282765243, "grad_norm": 0.8172041321496114, "learning_rate": 9.999887778087948e-06, "loss": 0.6888, "step": 532 }, { "epoch": 0.5117618819011042, "grad_norm": 0.9933746785176321, "learning_rate": 9.999864211593088e-06, "loss": 0.7026, "step": 533 }, { "epoch": 0.5127220355256841, "grad_norm": 0.9260724243942703, "learning_rate": 9.999838400712627e-06, "loss": 0.6456, "step": 534 }, { "epoch": 0.5136821891502641, "grad_norm": 0.8095758735512073, "learning_rate": 9.999810345458153e-06, "loss": 0.6543, "step": 535 }, { "epoch": 0.514642342774844, "grad_norm": 1.201538585554887, "learning_rate": 9.999780045842262e-06, "loss": 0.7661, "step": 536 }, { "epoch": 0.5156024963994239, "grad_norm": 0.8606811723246726, "learning_rate": 9.999747501878551e-06, "loss": 0.6622, "step": 537 }, { "epoch": 0.5165626500240038, "grad_norm": 0.9469642579979548, "learning_rate": 9.999712713581634e-06, "loss": 0.658, "step": 538 }, { "epoch": 0.5175228036485837, "grad_norm": 0.9072210871043881, "learning_rate": 9.99967568096712e-06, "loss": 0.6135, "step": 539 }, { "epoch": 0.5184829572731637, "grad_norm": 0.8692355067333526, "learning_rate": 9.999636404051638e-06, "loss": 0.6917, "step": 540 }, { "epoch": 0.5194431108977436, "grad_norm": 1.2263969115211968, "learning_rate": 9.999594882852817e-06, "loss": 0.8238, "step": 541 }, { "epoch": 0.5204032645223235, "grad_norm": 1.055948734732631, "learning_rate": 9.999551117389298e-06, "loss": 0.8145, "step": 542 }, { "epoch": 0.5213634181469035, "grad_norm": 1.111297406973271, "learning_rate": 9.999505107680722e-06, "loss": 0.6771, "step": 543 }, { "epoch": 0.5223235717714835, "grad_norm": 1.0687725484474766, "learning_rate": 9.999456853747744e-06, "loss": 0.6863, "step": 544 }, { "epoch": 0.5232837253960634, "grad_norm": 1.0215261647165164, "learning_rate": 9.999406355612029e-06, "loss": 0.8192, "step": 545 }, { "epoch": 0.5242438790206433, "grad_norm": 1.0504878099206951, "learning_rate": 9.99935361329624e-06, "loss": 0.7702, "step": 546 }, { "epoch": 0.5252040326452232, "grad_norm": 0.9247372870261737, "learning_rate": 9.999298626824053e-06, "loss": 0.6404, "step": 547 }, { "epoch": 0.5261641862698032, "grad_norm": 0.8257431544804358, "learning_rate": 9.999241396220152e-06, "loss": 0.5912, "step": 548 }, { "epoch": 0.5271243398943831, "grad_norm": 0.9337787918859838, "learning_rate": 9.999181921510229e-06, "loss": 0.7069, "step": 549 }, { "epoch": 0.528084493518963, "grad_norm": 0.8745720460385571, "learning_rate": 9.999120202720978e-06, "loss": 0.6839, "step": 550 }, { "epoch": 0.5290446471435429, "grad_norm": 0.8798743305220332, "learning_rate": 9.999056239880107e-06, "loss": 0.7488, "step": 551 }, { "epoch": 0.530004800768123, "grad_norm": 1.028456882603741, "learning_rate": 9.998990033016327e-06, "loss": 0.7319, "step": 552 }, { "epoch": 0.5309649543927029, "grad_norm": 1.124345207997282, "learning_rate": 9.998921582159357e-06, "loss": 0.6906, "step": 553 }, { "epoch": 0.5319251080172828, "grad_norm": 0.8138538166317938, "learning_rate": 9.998850887339923e-06, "loss": 0.6906, "step": 554 }, { "epoch": 0.5328852616418627, "grad_norm": 0.8061057546414215, "learning_rate": 9.998777948589762e-06, "loss": 0.6912, "step": 555 }, { "epoch": 0.5338454152664426, "grad_norm": 0.93777497762767, "learning_rate": 9.998702765941613e-06, "loss": 0.6524, "step": 556 }, { "epoch": 0.5348055688910226, "grad_norm": 1.004853308062429, "learning_rate": 9.998625339429227e-06, "loss": 0.6966, "step": 557 }, { "epoch": 0.5357657225156025, "grad_norm": 1.0128280668893683, "learning_rate": 9.998545669087357e-06, "loss": 0.7971, "step": 558 }, { "epoch": 0.5367258761401824, "grad_norm": 0.8684418560735583, "learning_rate": 9.998463754951767e-06, "loss": 0.6852, "step": 559 }, { "epoch": 0.5376860297647623, "grad_norm": 0.8520904134172057, "learning_rate": 9.998379597059228e-06, "loss": 0.6729, "step": 560 }, { "epoch": 0.5386461833893423, "grad_norm": 0.9937564704655384, "learning_rate": 9.99829319544752e-06, "loss": 0.7865, "step": 561 }, { "epoch": 0.5396063370139222, "grad_norm": 1.2675777390002032, "learning_rate": 9.998204550155424e-06, "loss": 0.7686, "step": 562 }, { "epoch": 0.5405664906385022, "grad_norm": 1.0280792189617642, "learning_rate": 9.998113661222732e-06, "loss": 0.8165, "step": 563 }, { "epoch": 0.5415266442630821, "grad_norm": 0.9943655043281197, "learning_rate": 9.998020528690247e-06, "loss": 0.7704, "step": 564 }, { "epoch": 0.5424867978876621, "grad_norm": 1.2859679603692513, "learning_rate": 9.99792515259977e-06, "loss": 0.8846, "step": 565 }, { "epoch": 0.543446951512242, "grad_norm": 1.0889138803260305, "learning_rate": 9.997827532994118e-06, "loss": 0.6975, "step": 566 }, { "epoch": 0.5444071051368219, "grad_norm": 0.7831773478259658, "learning_rate": 9.99772766991711e-06, "loss": 0.6087, "step": 567 }, { "epoch": 0.5453672587614018, "grad_norm": 1.0198594460216381, "learning_rate": 9.997625563413574e-06, "loss": 0.8182, "step": 568 }, { "epoch": 0.5463274123859817, "grad_norm": 0.9699062977786885, "learning_rate": 9.997521213529345e-06, "loss": 0.7708, "step": 569 }, { "epoch": 0.5472875660105617, "grad_norm": 1.2873131930121828, "learning_rate": 9.997414620311262e-06, "loss": 0.8361, "step": 570 }, { "epoch": 0.5482477196351416, "grad_norm": 0.917702578791434, "learning_rate": 9.997305783807176e-06, "loss": 0.698, "step": 571 }, { "epoch": 0.5492078732597215, "grad_norm": 0.9252721583574467, "learning_rate": 9.997194704065942e-06, "loss": 0.6329, "step": 572 }, { "epoch": 0.5501680268843014, "grad_norm": 0.8218303377348494, "learning_rate": 9.997081381137424e-06, "loss": 0.6508, "step": 573 }, { "epoch": 0.5511281805088815, "grad_norm": 1.1230486292567072, "learning_rate": 9.996965815072489e-06, "loss": 0.7474, "step": 574 }, { "epoch": 0.5520883341334614, "grad_norm": 0.900587114499074, "learning_rate": 9.996848005923012e-06, "loss": 0.6737, "step": 575 }, { "epoch": 0.5530484877580413, "grad_norm": 0.8449100515330694, "learning_rate": 9.996727953741881e-06, "loss": 0.6692, "step": 576 }, { "epoch": 0.5540086413826212, "grad_norm": 1.3025601393444597, "learning_rate": 9.996605658582982e-06, "loss": 0.8245, "step": 577 }, { "epoch": 0.5549687950072012, "grad_norm": 1.4474758756292687, "learning_rate": 9.996481120501214e-06, "loss": 0.853, "step": 578 }, { "epoch": 0.5559289486317811, "grad_norm": 0.8015975256712742, "learning_rate": 9.996354339552482e-06, "loss": 0.5864, "step": 579 }, { "epoch": 0.556889102256361, "grad_norm": 1.2600242780278368, "learning_rate": 9.996225315793692e-06, "loss": 0.7469, "step": 580 }, { "epoch": 0.5578492558809409, "grad_norm": 1.0255216035407804, "learning_rate": 9.996094049282768e-06, "loss": 0.7509, "step": 581 }, { "epoch": 0.5588094095055209, "grad_norm": 0.9186250675612276, "learning_rate": 9.995960540078627e-06, "loss": 0.7048, "step": 582 }, { "epoch": 0.5597695631301008, "grad_norm": 1.2659587150200031, "learning_rate": 9.995824788241206e-06, "loss": 0.737, "step": 583 }, { "epoch": 0.5607297167546808, "grad_norm": 0.995170699344741, "learning_rate": 9.995686793831438e-06, "loss": 0.7069, "step": 584 }, { "epoch": 0.5616898703792607, "grad_norm": 0.9087181029697331, "learning_rate": 9.995546556911271e-06, "loss": 0.698, "step": 585 }, { "epoch": 0.5626500240038406, "grad_norm": 1.0741467972809995, "learning_rate": 9.995404077543654e-06, "loss": 0.7598, "step": 586 }, { "epoch": 0.5636101776284206, "grad_norm": 1.0890235856206183, "learning_rate": 9.995259355792544e-06, "loss": 0.7544, "step": 587 }, { "epoch": 0.5645703312530005, "grad_norm": 1.4674437868826986, "learning_rate": 9.995112391722905e-06, "loss": 0.8856, "step": 588 }, { "epoch": 0.5655304848775804, "grad_norm": 1.0822658305254027, "learning_rate": 9.994963185400708e-06, "loss": 0.7374, "step": 589 }, { "epoch": 0.5664906385021603, "grad_norm": 0.9161138179461333, "learning_rate": 9.994811736892931e-06, "loss": 0.6486, "step": 590 }, { "epoch": 0.5674507921267403, "grad_norm": 1.0206439768941382, "learning_rate": 9.994658046267556e-06, "loss": 0.7465, "step": 591 }, { "epoch": 0.5684109457513202, "grad_norm": 1.2177648198687736, "learning_rate": 9.994502113593575e-06, "loss": 0.7058, "step": 592 }, { "epoch": 0.5693710993759001, "grad_norm": 0.9079943269120534, "learning_rate": 9.994343938940983e-06, "loss": 0.7031, "step": 593 }, { "epoch": 0.57033125300048, "grad_norm": 1.1490095124518267, "learning_rate": 9.994183522380784e-06, "loss": 0.8233, "step": 594 }, { "epoch": 0.5712914066250601, "grad_norm": 0.9689579471177395, "learning_rate": 9.994020863984986e-06, "loss": 0.7036, "step": 595 }, { "epoch": 0.57225156024964, "grad_norm": 0.9059642102245151, "learning_rate": 9.993855963826605e-06, "loss": 0.7637, "step": 596 }, { "epoch": 0.5732117138742199, "grad_norm": 0.826419975031267, "learning_rate": 9.993688821979664e-06, "loss": 0.6647, "step": 597 }, { "epoch": 0.5741718674987998, "grad_norm": 1.3799994059689251, "learning_rate": 9.993519438519191e-06, "loss": 0.8248, "step": 598 }, { "epoch": 0.5751320211233797, "grad_norm": 1.3329321428324197, "learning_rate": 9.993347813521219e-06, "loss": 0.7907, "step": 599 }, { "epoch": 0.5760921747479597, "grad_norm": 1.0762773043946197, "learning_rate": 9.993173947062789e-06, "loss": 0.6793, "step": 600 }, { "epoch": 0.5770523283725396, "grad_norm": 1.0883361407403807, "learning_rate": 9.992997839221947e-06, "loss": 0.7649, "step": 601 }, { "epoch": 0.5780124819971195, "grad_norm": 0.818419048632561, "learning_rate": 9.99281949007775e-06, "loss": 0.6014, "step": 602 }, { "epoch": 0.5789726356216994, "grad_norm": 0.925565708144328, "learning_rate": 9.992638899710252e-06, "loss": 0.8321, "step": 603 }, { "epoch": 0.5799327892462794, "grad_norm": 1.0480058726984378, "learning_rate": 9.99245606820052e-06, "loss": 0.7904, "step": 604 }, { "epoch": 0.5808929428708594, "grad_norm": 1.0493926578065629, "learning_rate": 9.992270995630628e-06, "loss": 0.7892, "step": 605 }, { "epoch": 0.5818530964954393, "grad_norm": 1.025271193273663, "learning_rate": 9.992083682083649e-06, "loss": 0.7424, "step": 606 }, { "epoch": 0.5828132501200192, "grad_norm": 0.8017715958987326, "learning_rate": 9.991894127643666e-06, "loss": 0.6201, "step": 607 }, { "epoch": 0.5837734037445992, "grad_norm": 1.02339663583353, "learning_rate": 9.991702332395771e-06, "loss": 0.8318, "step": 608 }, { "epoch": 0.5847335573691791, "grad_norm": 1.0008362602959673, "learning_rate": 9.991508296426057e-06, "loss": 0.8143, "step": 609 }, { "epoch": 0.585693710993759, "grad_norm": 1.0054383838783265, "learning_rate": 9.991312019821626e-06, "loss": 0.8355, "step": 610 }, { "epoch": 0.5866538646183389, "grad_norm": 1.079969251938482, "learning_rate": 9.991113502670583e-06, "loss": 0.7181, "step": 611 }, { "epoch": 0.5876140182429188, "grad_norm": 0.9675198309374698, "learning_rate": 9.99091274506204e-06, "loss": 0.6992, "step": 612 }, { "epoch": 0.5885741718674988, "grad_norm": 0.9265564846858307, "learning_rate": 9.990709747086118e-06, "loss": 0.7145, "step": 613 }, { "epoch": 0.5895343254920787, "grad_norm": 1.131396375601068, "learning_rate": 9.990504508833936e-06, "loss": 0.7706, "step": 614 }, { "epoch": 0.5904944791166586, "grad_norm": 0.9397541260657793, "learning_rate": 9.990297030397626e-06, "loss": 0.7461, "step": 615 }, { "epoch": 0.5914546327412386, "grad_norm": 1.0298415847700937, "learning_rate": 9.990087311870322e-06, "loss": 0.8031, "step": 616 }, { "epoch": 0.5924147863658186, "grad_norm": 1.085934518011205, "learning_rate": 9.989875353346164e-06, "loss": 0.6666, "step": 617 }, { "epoch": 0.5933749399903985, "grad_norm": 1.2617171840949917, "learning_rate": 9.9896611549203e-06, "loss": 0.7978, "step": 618 }, { "epoch": 0.5943350936149784, "grad_norm": 0.8937032826035056, "learning_rate": 9.98944471668888e-06, "loss": 0.5989, "step": 619 }, { "epoch": 0.5952952472395583, "grad_norm": 0.9566585964879852, "learning_rate": 9.98922603874906e-06, "loss": 0.63, "step": 620 }, { "epoch": 0.5962554008641383, "grad_norm": 1.0112528521193505, "learning_rate": 9.989005121199004e-06, "loss": 0.7202, "step": 621 }, { "epoch": 0.5972155544887182, "grad_norm": 0.9293648518783405, "learning_rate": 9.98878196413788e-06, "loss": 0.6323, "step": 622 }, { "epoch": 0.5981757081132981, "grad_norm": 0.956085125892744, "learning_rate": 9.988556567665858e-06, "loss": 0.7654, "step": 623 }, { "epoch": 0.599135861737878, "grad_norm": 0.9281633105708211, "learning_rate": 9.988328931884116e-06, "loss": 0.6375, "step": 624 }, { "epoch": 0.600096015362458, "grad_norm": 1.0226799229413066, "learning_rate": 9.98809905689484e-06, "loss": 0.7259, "step": 625 }, { "epoch": 0.601056168987038, "grad_norm": 0.8032693266799052, "learning_rate": 9.987866942801216e-06, "loss": 0.7133, "step": 626 }, { "epoch": 0.6020163226116179, "grad_norm": 1.4296147366295766, "learning_rate": 9.987632589707441e-06, "loss": 0.8303, "step": 627 }, { "epoch": 0.6029764762361978, "grad_norm": 0.9250172003922403, "learning_rate": 9.98739599771871e-06, "loss": 0.6024, "step": 628 }, { "epoch": 0.6039366298607777, "grad_norm": 1.1438700000950706, "learning_rate": 9.987157166941228e-06, "loss": 0.662, "step": 629 }, { "epoch": 0.6048967834853577, "grad_norm": 1.1220619914172665, "learning_rate": 9.986916097482204e-06, "loss": 0.6314, "step": 630 }, { "epoch": 0.6058569371099376, "grad_norm": 0.8789973193978474, "learning_rate": 9.98667278944985e-06, "loss": 0.7375, "step": 631 }, { "epoch": 0.6068170907345175, "grad_norm": 0.8057309508316937, "learning_rate": 9.986427242953387e-06, "loss": 0.5812, "step": 632 }, { "epoch": 0.6077772443590974, "grad_norm": 0.880095094559781, "learning_rate": 9.986179458103036e-06, "loss": 0.6881, "step": 633 }, { "epoch": 0.6087373979836774, "grad_norm": 0.907188961894006, "learning_rate": 9.985929435010024e-06, "loss": 0.7341, "step": 634 }, { "epoch": 0.6096975516082573, "grad_norm": 0.9364183363106064, "learning_rate": 9.985677173786587e-06, "loss": 0.6677, "step": 635 }, { "epoch": 0.6106577052328372, "grad_norm": 1.048054174818575, "learning_rate": 9.98542267454596e-06, "loss": 0.7312, "step": 636 }, { "epoch": 0.6116178588574172, "grad_norm": 0.8664078826971757, "learning_rate": 9.985165937402386e-06, "loss": 0.7028, "step": 637 }, { "epoch": 0.6125780124819972, "grad_norm": 0.9277961803830227, "learning_rate": 9.98490696247111e-06, "loss": 0.724, "step": 638 }, { "epoch": 0.6135381661065771, "grad_norm": 0.8396203807869509, "learning_rate": 9.984645749868386e-06, "loss": 0.6548, "step": 639 }, { "epoch": 0.614498319731157, "grad_norm": 0.9258214999290996, "learning_rate": 9.984382299711466e-06, "loss": 0.7083, "step": 640 }, { "epoch": 0.6154584733557369, "grad_norm": 1.1495576515720454, "learning_rate": 9.984116612118612e-06, "loss": 0.811, "step": 641 }, { "epoch": 0.6164186269803168, "grad_norm": 1.0811322677470414, "learning_rate": 9.98384868720909e-06, "loss": 0.7121, "step": 642 }, { "epoch": 0.6173787806048968, "grad_norm": 1.0678746875333789, "learning_rate": 9.983578525103166e-06, "loss": 0.7262, "step": 643 }, { "epoch": 0.6183389342294767, "grad_norm": 1.2941031888069452, "learning_rate": 9.983306125922111e-06, "loss": 0.7824, "step": 644 }, { "epoch": 0.6192990878540566, "grad_norm": 1.234808125187241, "learning_rate": 9.983031489788208e-06, "loss": 0.8897, "step": 645 }, { "epoch": 0.6202592414786365, "grad_norm": 0.9864889254556639, "learning_rate": 9.982754616824733e-06, "loss": 0.7088, "step": 646 }, { "epoch": 0.6212193951032166, "grad_norm": 1.4849103086902267, "learning_rate": 9.982475507155974e-06, "loss": 0.8115, "step": 647 }, { "epoch": 0.6221795487277965, "grad_norm": 0.9710713264165676, "learning_rate": 9.98219416090722e-06, "loss": 0.6689, "step": 648 }, { "epoch": 0.6231397023523764, "grad_norm": 0.9949565240659222, "learning_rate": 9.981910578204764e-06, "loss": 0.6529, "step": 649 }, { "epoch": 0.6240998559769563, "grad_norm": 1.1653599617265311, "learning_rate": 9.981624759175902e-06, "loss": 0.6667, "step": 650 }, { "epoch": 0.6250600096015363, "grad_norm": 0.857203882078485, "learning_rate": 9.981336703948936e-06, "loss": 0.629, "step": 651 }, { "epoch": 0.6260201632261162, "grad_norm": 1.2562325141148218, "learning_rate": 9.98104641265317e-06, "loss": 0.7819, "step": 652 }, { "epoch": 0.6269803168506961, "grad_norm": 0.8303609468780491, "learning_rate": 9.980753885418916e-06, "loss": 0.6556, "step": 653 }, { "epoch": 0.627940470475276, "grad_norm": 0.9312046752846731, "learning_rate": 9.980459122377484e-06, "loss": 0.7674, "step": 654 }, { "epoch": 0.628900624099856, "grad_norm": 1.0080799216405272, "learning_rate": 9.98016212366119e-06, "loss": 0.6948, "step": 655 }, { "epoch": 0.6298607777244359, "grad_norm": 0.889450298515072, "learning_rate": 9.979862889403353e-06, "loss": 0.7473, "step": 656 }, { "epoch": 0.6308209313490158, "grad_norm": 1.0751806902287246, "learning_rate": 9.979561419738297e-06, "loss": 0.7826, "step": 657 }, { "epoch": 0.6317810849735958, "grad_norm": 1.0132157137488043, "learning_rate": 9.979257714801348e-06, "loss": 0.6816, "step": 658 }, { "epoch": 0.6327412385981757, "grad_norm": 0.9285553108576658, "learning_rate": 9.978951774728837e-06, "loss": 0.7153, "step": 659 }, { "epoch": 0.6337013922227557, "grad_norm": 1.1167317501912706, "learning_rate": 9.978643599658095e-06, "loss": 0.6994, "step": 660 }, { "epoch": 0.6346615458473356, "grad_norm": 1.0694773166991753, "learning_rate": 9.97833318972746e-06, "loss": 0.7261, "step": 661 }, { "epoch": 0.6356216994719155, "grad_norm": 1.342658784147305, "learning_rate": 9.978020545076274e-06, "loss": 0.8008, "step": 662 }, { "epoch": 0.6365818530964954, "grad_norm": 0.7805478376685379, "learning_rate": 9.977705665844875e-06, "loss": 0.6128, "step": 663 }, { "epoch": 0.6375420067210754, "grad_norm": 1.264593357103444, "learning_rate": 9.977388552174611e-06, "loss": 0.6654, "step": 664 }, { "epoch": 0.6385021603456553, "grad_norm": 0.8483003856434131, "learning_rate": 9.977069204207832e-06, "loss": 0.6568, "step": 665 }, { "epoch": 0.6394623139702352, "grad_norm": 1.0970136752728121, "learning_rate": 9.97674762208789e-06, "loss": 0.8615, "step": 666 }, { "epoch": 0.6404224675948151, "grad_norm": 1.2113418902324684, "learning_rate": 9.976423805959138e-06, "loss": 0.7511, "step": 667 }, { "epoch": 0.6413826212193952, "grad_norm": 1.2559011225976917, "learning_rate": 9.976097755966935e-06, "loss": 0.7459, "step": 668 }, { "epoch": 0.6423427748439751, "grad_norm": 0.854583778245372, "learning_rate": 9.975769472257642e-06, "loss": 0.705, "step": 669 }, { "epoch": 0.643302928468555, "grad_norm": 1.0568288836586655, "learning_rate": 9.97543895497862e-06, "loss": 0.6874, "step": 670 }, { "epoch": 0.6442630820931349, "grad_norm": 0.8484186050307778, "learning_rate": 9.975106204278235e-06, "loss": 0.6628, "step": 671 }, { "epoch": 0.6452232357177148, "grad_norm": 0.9070346933523904, "learning_rate": 9.974771220305855e-06, "loss": 0.7382, "step": 672 }, { "epoch": 0.6461833893422948, "grad_norm": 0.7284210856928233, "learning_rate": 9.974434003211853e-06, "loss": 0.6041, "step": 673 }, { "epoch": 0.6471435429668747, "grad_norm": 0.7171914497438403, "learning_rate": 9.974094553147599e-06, "loss": 0.6665, "step": 674 }, { "epoch": 0.6481036965914546, "grad_norm": 1.2461717089932496, "learning_rate": 9.973752870265473e-06, "loss": 0.8077, "step": 675 }, { "epoch": 0.6490638502160345, "grad_norm": 0.9834858544940579, "learning_rate": 9.973408954718848e-06, "loss": 0.7468, "step": 676 }, { "epoch": 0.6500240038406145, "grad_norm": 0.9184294222706681, "learning_rate": 9.973062806662106e-06, "loss": 0.6513, "step": 677 }, { "epoch": 0.6509841574651944, "grad_norm": 1.151049284884669, "learning_rate": 9.97271442625063e-06, "loss": 0.7695, "step": 678 }, { "epoch": 0.6519443110897744, "grad_norm": 0.9192364173384852, "learning_rate": 9.972363813640804e-06, "loss": 0.7792, "step": 679 }, { "epoch": 0.6529044647143543, "grad_norm": 0.8165139967381659, "learning_rate": 9.972010968990012e-06, "loss": 0.641, "step": 680 }, { "epoch": 0.6538646183389343, "grad_norm": 0.8247204727188149, "learning_rate": 9.971655892456646e-06, "loss": 0.7002, "step": 681 }, { "epoch": 0.6548247719635142, "grad_norm": 0.869473466223658, "learning_rate": 9.971298584200093e-06, "loss": 0.7229, "step": 682 }, { "epoch": 0.6557849255880941, "grad_norm": 0.7594593477812601, "learning_rate": 9.970939044380743e-06, "loss": 0.636, "step": 683 }, { "epoch": 0.656745079212674, "grad_norm": 0.8192096111185594, "learning_rate": 9.970577273159995e-06, "loss": 0.6929, "step": 684 }, { "epoch": 0.6577052328372539, "grad_norm": 0.8801904400737564, "learning_rate": 9.970213270700241e-06, "loss": 0.6701, "step": 685 }, { "epoch": 0.6586653864618339, "grad_norm": 1.096484388156687, "learning_rate": 9.969847037164881e-06, "loss": 0.7032, "step": 686 }, { "epoch": 0.6596255400864138, "grad_norm": 0.9812909729332547, "learning_rate": 9.969478572718307e-06, "loss": 0.6972, "step": 687 }, { "epoch": 0.6605856937109937, "grad_norm": 0.7559810355927282, "learning_rate": 9.969107877525925e-06, "loss": 0.6444, "step": 688 }, { "epoch": 0.6615458473355736, "grad_norm": 0.7933216933607872, "learning_rate": 9.968734951754134e-06, "loss": 0.5916, "step": 689 }, { "epoch": 0.6625060009601537, "grad_norm": 0.7938458454913603, "learning_rate": 9.968359795570333e-06, "loss": 0.6294, "step": 690 }, { "epoch": 0.6634661545847336, "grad_norm": 0.8151175067214695, "learning_rate": 9.96798240914293e-06, "loss": 0.7498, "step": 691 }, { "epoch": 0.6644263082093135, "grad_norm": 0.8258470944924684, "learning_rate": 9.96760279264133e-06, "loss": 0.6179, "step": 692 }, { "epoch": 0.6653864618338934, "grad_norm": 0.8909674152309772, "learning_rate": 9.967220946235933e-06, "loss": 0.739, "step": 693 }, { "epoch": 0.6663466154584734, "grad_norm": 0.8108166845923671, "learning_rate": 9.966836870098153e-06, "loss": 0.6559, "step": 694 }, { "epoch": 0.6673067690830533, "grad_norm": 0.8502920180055921, "learning_rate": 9.966450564400396e-06, "loss": 0.6385, "step": 695 }, { "epoch": 0.6682669227076332, "grad_norm": 1.3308580264709484, "learning_rate": 9.966062029316067e-06, "loss": 0.8202, "step": 696 }, { "epoch": 0.6692270763322131, "grad_norm": 0.8176688991698405, "learning_rate": 9.965671265019576e-06, "loss": 0.7205, "step": 697 }, { "epoch": 0.6701872299567931, "grad_norm": 0.7160414893096833, "learning_rate": 9.965278271686335e-06, "loss": 0.5894, "step": 698 }, { "epoch": 0.671147383581373, "grad_norm": 0.8250604787361822, "learning_rate": 9.964883049492755e-06, "loss": 0.6939, "step": 699 }, { "epoch": 0.672107537205953, "grad_norm": 0.827635879711952, "learning_rate": 9.964485598616244e-06, "loss": 0.6265, "step": 700 }, { "epoch": 0.6730676908305329, "grad_norm": 0.9123753351721733, "learning_rate": 9.964085919235217e-06, "loss": 0.6449, "step": 701 }, { "epoch": 0.6740278444551128, "grad_norm": 1.0108317660636383, "learning_rate": 9.963684011529084e-06, "loss": 0.6533, "step": 702 }, { "epoch": 0.6749879980796928, "grad_norm": 0.9862069309281521, "learning_rate": 9.963279875678256e-06, "loss": 0.7086, "step": 703 }, { "epoch": 0.6759481517042727, "grad_norm": 0.9878216419373168, "learning_rate": 9.962873511864146e-06, "loss": 0.7262, "step": 704 }, { "epoch": 0.6769083053288526, "grad_norm": 0.978449601968194, "learning_rate": 9.962464920269167e-06, "loss": 0.8676, "step": 705 }, { "epoch": 0.6778684589534325, "grad_norm": 0.9107988538066738, "learning_rate": 9.962054101076733e-06, "loss": 0.7407, "step": 706 }, { "epoch": 0.6788286125780125, "grad_norm": 0.691940630730434, "learning_rate": 9.961641054471252e-06, "loss": 0.6434, "step": 707 }, { "epoch": 0.6797887662025924, "grad_norm": 1.0050777282215912, "learning_rate": 9.96122578063814e-06, "loss": 0.677, "step": 708 }, { "epoch": 0.6807489198271723, "grad_norm": 0.7640397564270462, "learning_rate": 9.960808279763807e-06, "loss": 0.6851, "step": 709 }, { "epoch": 0.6817090734517522, "grad_norm": 0.9737611671450259, "learning_rate": 9.960388552035666e-06, "loss": 0.587, "step": 710 }, { "epoch": 0.6826692270763323, "grad_norm": 0.7348175868192905, "learning_rate": 9.959966597642127e-06, "loss": 0.6121, "step": 711 }, { "epoch": 0.6836293807009122, "grad_norm": 0.7454408895351605, "learning_rate": 9.9595424167726e-06, "loss": 0.6383, "step": 712 }, { "epoch": 0.6845895343254921, "grad_norm": 1.1468710093910879, "learning_rate": 9.9591160096175e-06, "loss": 0.7406, "step": 713 }, { "epoch": 0.685549687950072, "grad_norm": 1.0615540483674522, "learning_rate": 9.95868737636823e-06, "loss": 0.7449, "step": 714 }, { "epoch": 0.6865098415746519, "grad_norm": 0.7713954726916293, "learning_rate": 9.958256517217204e-06, "loss": 0.5932, "step": 715 }, { "epoch": 0.6874699951992319, "grad_norm": 1.048663844718065, "learning_rate": 9.95782343235783e-06, "loss": 0.8067, "step": 716 }, { "epoch": 0.6884301488238118, "grad_norm": 0.9627910093841642, "learning_rate": 9.95738812198451e-06, "loss": 0.7208, "step": 717 }, { "epoch": 0.6893903024483917, "grad_norm": 0.9906607278101155, "learning_rate": 9.956950586292655e-06, "loss": 0.7105, "step": 718 }, { "epoch": 0.6903504560729716, "grad_norm": 0.7944942412706745, "learning_rate": 9.956510825478669e-06, "loss": 0.6577, "step": 719 }, { "epoch": 0.6913106096975516, "grad_norm": 1.1460416563571731, "learning_rate": 9.956068839739955e-06, "loss": 0.7193, "step": 720 }, { "epoch": 0.6922707633221316, "grad_norm": 0.8733037849866548, "learning_rate": 9.955624629274916e-06, "loss": 0.6614, "step": 721 }, { "epoch": 0.6932309169467115, "grad_norm": 0.9991602492012207, "learning_rate": 9.955178194282953e-06, "loss": 0.6815, "step": 722 }, { "epoch": 0.6941910705712914, "grad_norm": 0.8214274693195731, "learning_rate": 9.954729534964469e-06, "loss": 0.7527, "step": 723 }, { "epoch": 0.6951512241958714, "grad_norm": 0.9231615478789139, "learning_rate": 9.954278651520859e-06, "loss": 0.7285, "step": 724 }, { "epoch": 0.6961113778204513, "grad_norm": 1.2288355183170787, "learning_rate": 9.953825544154519e-06, "loss": 0.8585, "step": 725 }, { "epoch": 0.6970715314450312, "grad_norm": 1.156076980713657, "learning_rate": 9.953370213068847e-06, "loss": 0.7783, "step": 726 }, { "epoch": 0.6980316850696111, "grad_norm": 0.9162677367573459, "learning_rate": 9.952912658468235e-06, "loss": 0.7259, "step": 727 }, { "epoch": 0.6989918386941911, "grad_norm": 1.0142345842494351, "learning_rate": 9.952452880558073e-06, "loss": 0.7483, "step": 728 }, { "epoch": 0.699951992318771, "grad_norm": 0.9156658580169411, "learning_rate": 9.951990879544753e-06, "loss": 0.6737, "step": 729 }, { "epoch": 0.7009121459433509, "grad_norm": 0.9333699041046369, "learning_rate": 9.95152665563566e-06, "loss": 0.6228, "step": 730 }, { "epoch": 0.7018722995679308, "grad_norm": 0.7864784828695316, "learning_rate": 9.951060209039182e-06, "loss": 0.5757, "step": 731 }, { "epoch": 0.7028324531925108, "grad_norm": 0.9154756732438272, "learning_rate": 9.950591539964698e-06, "loss": 0.7202, "step": 732 }, { "epoch": 0.7037926068170908, "grad_norm": 0.8159133636197664, "learning_rate": 9.950120648622592e-06, "loss": 0.6835, "step": 733 }, { "epoch": 0.7047527604416707, "grad_norm": 0.7809064324079145, "learning_rate": 9.94964753522424e-06, "loss": 0.562, "step": 734 }, { "epoch": 0.7057129140662506, "grad_norm": 1.1456299994523809, "learning_rate": 9.949172199982019e-06, "loss": 0.6969, "step": 735 }, { "epoch": 0.7066730676908305, "grad_norm": 1.3353677080257012, "learning_rate": 9.9486946431093e-06, "loss": 0.8036, "step": 736 }, { "epoch": 0.7076332213154105, "grad_norm": 1.0155618463529212, "learning_rate": 9.948214864820455e-06, "loss": 0.7238, "step": 737 }, { "epoch": 0.7085933749399904, "grad_norm": 0.9588739753248952, "learning_rate": 9.94773286533085e-06, "loss": 0.7415, "step": 738 }, { "epoch": 0.7095535285645703, "grad_norm": 0.8891749290808151, "learning_rate": 9.947248644856852e-06, "loss": 0.6102, "step": 739 }, { "epoch": 0.7105136821891502, "grad_norm": 0.7510827119436133, "learning_rate": 9.946762203615818e-06, "loss": 0.6227, "step": 740 }, { "epoch": 0.7114738358137302, "grad_norm": 1.03977923096292, "learning_rate": 9.94627354182611e-06, "loss": 0.6601, "step": 741 }, { "epoch": 0.7124339894383102, "grad_norm": 0.7641684137626731, "learning_rate": 9.945782659707082e-06, "loss": 0.615, "step": 742 }, { "epoch": 0.7133941430628901, "grad_norm": 1.1836739514398755, "learning_rate": 9.945289557479086e-06, "loss": 0.8607, "step": 743 }, { "epoch": 0.71435429668747, "grad_norm": 1.0672694440645691, "learning_rate": 9.944794235363468e-06, "loss": 0.7822, "step": 744 }, { "epoch": 0.7153144503120499, "grad_norm": 0.8276856409752049, "learning_rate": 9.944296693582576e-06, "loss": 0.6953, "step": 745 }, { "epoch": 0.7162746039366299, "grad_norm": 1.227496536283197, "learning_rate": 9.94379693235975e-06, "loss": 0.7812, "step": 746 }, { "epoch": 0.7172347575612098, "grad_norm": 0.934069620370573, "learning_rate": 9.943294951919327e-06, "loss": 0.7458, "step": 747 }, { "epoch": 0.7181949111857897, "grad_norm": 0.9716004251098633, "learning_rate": 9.94279075248664e-06, "loss": 0.7342, "step": 748 }, { "epoch": 0.7191550648103696, "grad_norm": 1.0155927885160487, "learning_rate": 9.942284334288021e-06, "loss": 0.8186, "step": 749 }, { "epoch": 0.7201152184349496, "grad_norm": 0.7435028249036985, "learning_rate": 9.941775697550796e-06, "loss": 0.698, "step": 750 }, { "epoch": 0.7210753720595295, "grad_norm": 1.037064594720284, "learning_rate": 9.941264842503282e-06, "loss": 0.6872, "step": 751 }, { "epoch": 0.7220355256841094, "grad_norm": 0.9147032761045544, "learning_rate": 9.9407517693748e-06, "loss": 0.7122, "step": 752 }, { "epoch": 0.7229956793086894, "grad_norm": 0.8670736937294433, "learning_rate": 9.940236478395662e-06, "loss": 0.7099, "step": 753 }, { "epoch": 0.7239558329332694, "grad_norm": 0.6985402299699525, "learning_rate": 9.939718969797178e-06, "loss": 0.4975, "step": 754 }, { "epoch": 0.7249159865578493, "grad_norm": 1.147752522726644, "learning_rate": 9.939199243811653e-06, "loss": 0.7793, "step": 755 }, { "epoch": 0.7258761401824292, "grad_norm": 0.8979667876220845, "learning_rate": 9.938677300672382e-06, "loss": 0.7019, "step": 756 }, { "epoch": 0.7268362938070091, "grad_norm": 1.0471390027187375, "learning_rate": 9.938153140613662e-06, "loss": 0.8146, "step": 757 }, { "epoch": 0.727796447431589, "grad_norm": 1.0813293933485835, "learning_rate": 9.937626763870784e-06, "loss": 0.6765, "step": 758 }, { "epoch": 0.728756601056169, "grad_norm": 0.8776806341811652, "learning_rate": 9.937098170680034e-06, "loss": 0.7325, "step": 759 }, { "epoch": 0.7297167546807489, "grad_norm": 0.9747068741811712, "learning_rate": 9.936567361278687e-06, "loss": 0.6272, "step": 760 }, { "epoch": 0.7306769083053288, "grad_norm": 1.1019548285364111, "learning_rate": 9.93603433590502e-06, "loss": 0.7568, "step": 761 }, { "epoch": 0.7316370619299087, "grad_norm": 0.9939932523065834, "learning_rate": 9.935499094798305e-06, "loss": 0.7115, "step": 762 }, { "epoch": 0.7325972155544888, "grad_norm": 1.1156410921620932, "learning_rate": 9.934961638198804e-06, "loss": 0.8129, "step": 763 }, { "epoch": 0.7335573691790687, "grad_norm": 0.9769860898230756, "learning_rate": 9.934421966347773e-06, "loss": 0.7023, "step": 764 }, { "epoch": 0.7345175228036486, "grad_norm": 1.0902177322304396, "learning_rate": 9.93388007948747e-06, "loss": 0.7783, "step": 765 }, { "epoch": 0.7354776764282285, "grad_norm": 0.7372604029286459, "learning_rate": 9.933335977861138e-06, "loss": 0.6217, "step": 766 }, { "epoch": 0.7364378300528085, "grad_norm": 1.0142115675341663, "learning_rate": 9.93278966171302e-06, "loss": 0.7858, "step": 767 }, { "epoch": 0.7373979836773884, "grad_norm": 1.1136313073650677, "learning_rate": 9.932241131288352e-06, "loss": 0.8322, "step": 768 }, { "epoch": 0.7383581373019683, "grad_norm": 0.8292977925548107, "learning_rate": 9.931690386833363e-06, "loss": 0.6979, "step": 769 }, { "epoch": 0.7393182909265482, "grad_norm": 0.9730189524159811, "learning_rate": 9.931137428595276e-06, "loss": 0.7893, "step": 770 }, { "epoch": 0.7402784445511282, "grad_norm": 0.7940191342352438, "learning_rate": 9.930582256822307e-06, "loss": 0.6576, "step": 771 }, { "epoch": 0.7412385981757081, "grad_norm": 0.9666597698024856, "learning_rate": 9.93002487176367e-06, "loss": 0.7939, "step": 772 }, { "epoch": 0.742198751800288, "grad_norm": 0.936999804877027, "learning_rate": 9.929465273669567e-06, "loss": 0.7539, "step": 773 }, { "epoch": 0.743158905424868, "grad_norm": 0.9172212460073044, "learning_rate": 9.928903462791195e-06, "loss": 0.7207, "step": 774 }, { "epoch": 0.7441190590494479, "grad_norm": 0.8687429895741566, "learning_rate": 9.928339439380745e-06, "loss": 0.5919, "step": 775 }, { "epoch": 0.7450792126740279, "grad_norm": 0.8501185844318978, "learning_rate": 9.927773203691403e-06, "loss": 0.7479, "step": 776 }, { "epoch": 0.7460393662986078, "grad_norm": 1.2915173950859082, "learning_rate": 9.927204755977344e-06, "loss": 0.738, "step": 777 }, { "epoch": 0.7469995199231877, "grad_norm": 0.8664227663067325, "learning_rate": 9.926634096493738e-06, "loss": 0.6327, "step": 778 }, { "epoch": 0.7479596735477676, "grad_norm": 1.2380423772864184, "learning_rate": 9.92606122549675e-06, "loss": 0.692, "step": 779 }, { "epoch": 0.7489198271723476, "grad_norm": 1.568013927128809, "learning_rate": 9.925486143243534e-06, "loss": 0.8967, "step": 780 }, { "epoch": 0.7498799807969275, "grad_norm": 0.845781939448932, "learning_rate": 9.92490884999224e-06, "loss": 0.5935, "step": 781 }, { "epoch": 0.7508401344215074, "grad_norm": 0.7724652270764257, "learning_rate": 9.924329346002007e-06, "loss": 0.6346, "step": 782 }, { "epoch": 0.7518002880460873, "grad_norm": 1.204065151966616, "learning_rate": 9.923747631532968e-06, "loss": 0.8365, "step": 783 }, { "epoch": 0.7527604416706674, "grad_norm": 0.8440253022287962, "learning_rate": 9.92316370684625e-06, "loss": 0.7159, "step": 784 }, { "epoch": 0.7537205952952473, "grad_norm": 0.7841176229800747, "learning_rate": 9.922577572203969e-06, "loss": 0.6058, "step": 785 }, { "epoch": 0.7546807489198272, "grad_norm": 1.1108535651405738, "learning_rate": 9.921989227869236e-06, "loss": 0.787, "step": 786 }, { "epoch": 0.7556409025444071, "grad_norm": 0.7435096049011145, "learning_rate": 9.921398674106152e-06, "loss": 0.6444, "step": 787 }, { "epoch": 0.756601056168987, "grad_norm": 1.213652998444621, "learning_rate": 9.92080591117981e-06, "loss": 0.8066, "step": 788 }, { "epoch": 0.757561209793567, "grad_norm": 0.8042482080754291, "learning_rate": 9.920210939356294e-06, "loss": 0.589, "step": 789 }, { "epoch": 0.7585213634181469, "grad_norm": 0.9328289504245394, "learning_rate": 9.919613758902682e-06, "loss": 0.6762, "step": 790 }, { "epoch": 0.7594815170427268, "grad_norm": 1.1423203633068306, "learning_rate": 9.91901437008704e-06, "loss": 0.8467, "step": 791 }, { "epoch": 0.7604416706673067, "grad_norm": 0.8186246160052938, "learning_rate": 9.918412773178431e-06, "loss": 0.6557, "step": 792 }, { "epoch": 0.7614018242918867, "grad_norm": 0.955803163210281, "learning_rate": 9.917808968446903e-06, "loss": 0.661, "step": 793 }, { "epoch": 0.7623619779164666, "grad_norm": 0.8264688092238139, "learning_rate": 9.917202956163496e-06, "loss": 0.6274, "step": 794 }, { "epoch": 0.7633221315410466, "grad_norm": 0.9284939395885313, "learning_rate": 9.916594736600244e-06, "loss": 0.6903, "step": 795 }, { "epoch": 0.7642822851656265, "grad_norm": 1.3284567930888531, "learning_rate": 9.915984310030172e-06, "loss": 0.891, "step": 796 }, { "epoch": 0.7652424387902065, "grad_norm": 1.0025350705463052, "learning_rate": 9.91537167672729e-06, "loss": 0.7694, "step": 797 }, { "epoch": 0.7662025924147864, "grad_norm": 0.8861628999471639, "learning_rate": 9.914756836966604e-06, "loss": 0.7208, "step": 798 }, { "epoch": 0.7671627460393663, "grad_norm": 0.9752658748215958, "learning_rate": 9.914139791024112e-06, "loss": 0.7154, "step": 799 }, { "epoch": 0.7681228996639462, "grad_norm": 0.8201068509165711, "learning_rate": 9.913520539176797e-06, "loss": 0.6296, "step": 800 }, { "epoch": 0.7690830532885262, "grad_norm": 0.854380238784929, "learning_rate": 9.912899081702633e-06, "loss": 0.6608, "step": 801 }, { "epoch": 0.7700432069131061, "grad_norm": 0.9571683918229853, "learning_rate": 9.912275418880588e-06, "loss": 0.721, "step": 802 }, { "epoch": 0.771003360537686, "grad_norm": 0.9418784415018016, "learning_rate": 9.911649550990618e-06, "loss": 0.6827, "step": 803 }, { "epoch": 0.7719635141622659, "grad_norm": 1.2113269797866213, "learning_rate": 9.911021478313665e-06, "loss": 0.7688, "step": 804 }, { "epoch": 0.7729236677868458, "grad_norm": 1.2228209558022594, "learning_rate": 9.910391201131668e-06, "loss": 0.8139, "step": 805 }, { "epoch": 0.7738838214114259, "grad_norm": 1.0752944681261174, "learning_rate": 9.909758719727549e-06, "loss": 0.7968, "step": 806 }, { "epoch": 0.7748439750360058, "grad_norm": 1.1401225250337559, "learning_rate": 9.909124034385225e-06, "loss": 0.6945, "step": 807 }, { "epoch": 0.7758041286605857, "grad_norm": 0.8942684348442288, "learning_rate": 9.908487145389595e-06, "loss": 0.7326, "step": 808 }, { "epoch": 0.7767642822851656, "grad_norm": 0.953697929176355, "learning_rate": 9.907848053026554e-06, "loss": 0.6427, "step": 809 }, { "epoch": 0.7777244359097456, "grad_norm": 0.9285944035602395, "learning_rate": 9.907206757582987e-06, "loss": 0.6972, "step": 810 }, { "epoch": 0.7786845895343255, "grad_norm": 0.7638090267484249, "learning_rate": 9.90656325934676e-06, "loss": 0.6879, "step": 811 }, { "epoch": 0.7796447431589054, "grad_norm": 1.1685601140348312, "learning_rate": 9.905917558606734e-06, "loss": 0.6998, "step": 812 }, { "epoch": 0.7806048967834853, "grad_norm": 0.999228072829845, "learning_rate": 9.905269655652758e-06, "loss": 0.7612, "step": 813 }, { "epoch": 0.7815650504080653, "grad_norm": 1.0061463529746952, "learning_rate": 9.904619550775668e-06, "loss": 0.7944, "step": 814 }, { "epoch": 0.7825252040326452, "grad_norm": 0.9056842772570082, "learning_rate": 9.903967244267287e-06, "loss": 0.7354, "step": 815 }, { "epoch": 0.7834853576572252, "grad_norm": 1.478853524666308, "learning_rate": 9.903312736420433e-06, "loss": 0.917, "step": 816 }, { "epoch": 0.7844455112818051, "grad_norm": 1.4309348403618332, "learning_rate": 9.902656027528903e-06, "loss": 0.6921, "step": 817 }, { "epoch": 0.785405664906385, "grad_norm": 0.8082914277712884, "learning_rate": 9.90199711788749e-06, "loss": 0.5971, "step": 818 }, { "epoch": 0.786365818530965, "grad_norm": 0.7788440659193929, "learning_rate": 9.901336007791971e-06, "loss": 0.6062, "step": 819 }, { "epoch": 0.7873259721555449, "grad_norm": 0.9287350996778299, "learning_rate": 9.900672697539108e-06, "loss": 0.7207, "step": 820 }, { "epoch": 0.7882861257801248, "grad_norm": 0.791995989341828, "learning_rate": 9.900007187426659e-06, "loss": 0.6732, "step": 821 }, { "epoch": 0.7892462794047047, "grad_norm": 1.4775620297167587, "learning_rate": 9.89933947775336e-06, "loss": 0.8269, "step": 822 }, { "epoch": 0.7902064330292847, "grad_norm": 0.7898205056172806, "learning_rate": 9.898669568818942e-06, "loss": 0.58, "step": 823 }, { "epoch": 0.7911665866538646, "grad_norm": 0.9682106743679688, "learning_rate": 9.897997460924118e-06, "loss": 0.7425, "step": 824 }, { "epoch": 0.7921267402784445, "grad_norm": 0.7566417797167025, "learning_rate": 9.89732315437059e-06, "loss": 0.6578, "step": 825 }, { "epoch": 0.7930868939030244, "grad_norm": 1.0439738135585905, "learning_rate": 9.896646649461047e-06, "loss": 0.7722, "step": 826 }, { "epoch": 0.7940470475276045, "grad_norm": 0.7176023038358033, "learning_rate": 9.895967946499165e-06, "loss": 0.5635, "step": 827 }, { "epoch": 0.7950072011521844, "grad_norm": 0.9097535885813721, "learning_rate": 9.895287045789608e-06, "loss": 0.6544, "step": 828 }, { "epoch": 0.7959673547767643, "grad_norm": 0.9230366028387597, "learning_rate": 9.894603947638023e-06, "loss": 0.6428, "step": 829 }, { "epoch": 0.7969275084013442, "grad_norm": 1.0116104566672852, "learning_rate": 9.893918652351048e-06, "loss": 0.6934, "step": 830 }, { "epoch": 0.7978876620259241, "grad_norm": 0.8039174981071742, "learning_rate": 9.893231160236303e-06, "loss": 0.6662, "step": 831 }, { "epoch": 0.7988478156505041, "grad_norm": 0.8154076404450628, "learning_rate": 9.892541471602395e-06, "loss": 0.7025, "step": 832 }, { "epoch": 0.799807969275084, "grad_norm": 1.1804704478073578, "learning_rate": 9.891849586758918e-06, "loss": 0.7164, "step": 833 }, { "epoch": 0.8007681228996639, "grad_norm": 1.2849700375501172, "learning_rate": 9.891155506016452e-06, "loss": 0.7293, "step": 834 }, { "epoch": 0.8017282765242438, "grad_norm": 1.3472389914346394, "learning_rate": 9.890459229686564e-06, "loss": 0.7787, "step": 835 }, { "epoch": 0.8026884301488239, "grad_norm": 0.8904299453107525, "learning_rate": 9.889760758081803e-06, "loss": 0.6601, "step": 836 }, { "epoch": 0.8036485837734038, "grad_norm": 0.8609349931011556, "learning_rate": 9.889060091515707e-06, "loss": 0.6707, "step": 837 }, { "epoch": 0.8046087373979837, "grad_norm": 1.184320898105941, "learning_rate": 9.8883572303028e-06, "loss": 0.7671, "step": 838 }, { "epoch": 0.8055688910225636, "grad_norm": 0.9649280777152901, "learning_rate": 9.887652174758581e-06, "loss": 0.7635, "step": 839 }, { "epoch": 0.8065290446471436, "grad_norm": 0.9149126733007849, "learning_rate": 9.88694492519955e-06, "loss": 0.6651, "step": 840 }, { "epoch": 0.8074891982717235, "grad_norm": 1.1083506492838189, "learning_rate": 9.88623548194318e-06, "loss": 0.7192, "step": 841 }, { "epoch": 0.8084493518963034, "grad_norm": 0.7910033608312111, "learning_rate": 9.88552384530793e-06, "loss": 0.6453, "step": 842 }, { "epoch": 0.8094095055208833, "grad_norm": 0.8150352208068273, "learning_rate": 9.884810015613255e-06, "loss": 0.6077, "step": 843 }, { "epoch": 0.8103696591454633, "grad_norm": 0.7864208825515624, "learning_rate": 9.884093993179575e-06, "loss": 0.6374, "step": 844 }, { "epoch": 0.8113298127700432, "grad_norm": 0.8959183567788023, "learning_rate": 9.883375778328312e-06, "loss": 0.6593, "step": 845 }, { "epoch": 0.8122899663946231, "grad_norm": 1.0139478021722053, "learning_rate": 9.882655371381861e-06, "loss": 0.8, "step": 846 }, { "epoch": 0.813250120019203, "grad_norm": 0.9023123206682908, "learning_rate": 9.881932772663609e-06, "loss": 0.6779, "step": 847 }, { "epoch": 0.814210273643783, "grad_norm": 0.9622666786827915, "learning_rate": 9.881207982497917e-06, "loss": 0.7438, "step": 848 }, { "epoch": 0.815170427268363, "grad_norm": 0.7602304675839938, "learning_rate": 9.880481001210142e-06, "loss": 0.6648, "step": 849 }, { "epoch": 0.8161305808929429, "grad_norm": 0.7521205230443256, "learning_rate": 9.879751829126612e-06, "loss": 0.5361, "step": 850 }, { "epoch": 0.8170907345175228, "grad_norm": 0.9709115981480305, "learning_rate": 9.87902046657465e-06, "loss": 0.7385, "step": 851 }, { "epoch": 0.8180508881421027, "grad_norm": 0.6565874253176536, "learning_rate": 9.878286913882553e-06, "loss": 0.5719, "step": 852 }, { "epoch": 0.8190110417666827, "grad_norm": 1.295443326691402, "learning_rate": 9.877551171379606e-06, "loss": 0.8857, "step": 853 }, { "epoch": 0.8199711953912626, "grad_norm": 1.2461832345256898, "learning_rate": 9.876813239396079e-06, "loss": 0.6867, "step": 854 }, { "epoch": 0.8209313490158425, "grad_norm": 0.954909737324112, "learning_rate": 9.876073118263216e-06, "loss": 0.6105, "step": 855 }, { "epoch": 0.8218915026404224, "grad_norm": 0.9450097203587252, "learning_rate": 9.875330808313254e-06, "loss": 0.6602, "step": 856 }, { "epoch": 0.8228516562650025, "grad_norm": 0.8848389137759535, "learning_rate": 9.874586309879406e-06, "loss": 0.6188, "step": 857 }, { "epoch": 0.8238118098895824, "grad_norm": 0.8477829356555552, "learning_rate": 9.87383962329587e-06, "loss": 0.5798, "step": 858 }, { "epoch": 0.8247719635141623, "grad_norm": 0.8726627898566233, "learning_rate": 9.873090748897826e-06, "loss": 0.5439, "step": 859 }, { "epoch": 0.8257321171387422, "grad_norm": 0.9956760715485513, "learning_rate": 9.872339687021434e-06, "loss": 0.8595, "step": 860 }, { "epoch": 0.8266922707633221, "grad_norm": 1.1005980883063442, "learning_rate": 9.871586438003841e-06, "loss": 0.6637, "step": 861 }, { "epoch": 0.8276524243879021, "grad_norm": 0.9203453290274431, "learning_rate": 9.870831002183168e-06, "loss": 0.6644, "step": 862 }, { "epoch": 0.828612578012482, "grad_norm": 0.9029996018456606, "learning_rate": 9.870073379898526e-06, "loss": 0.6031, "step": 863 }, { "epoch": 0.8295727316370619, "grad_norm": 0.9765524419900932, "learning_rate": 9.869313571490002e-06, "loss": 0.6148, "step": 864 }, { "epoch": 0.8305328852616418, "grad_norm": 0.866107398072101, "learning_rate": 9.868551577298665e-06, "loss": 0.6821, "step": 865 }, { "epoch": 0.8314930388862218, "grad_norm": 0.9854234306165515, "learning_rate": 9.867787397666568e-06, "loss": 0.7228, "step": 866 }, { "epoch": 0.8324531925108017, "grad_norm": 0.8123425954393283, "learning_rate": 9.86702103293674e-06, "loss": 0.6783, "step": 867 }, { "epoch": 0.8334133461353816, "grad_norm": 0.9399100444654909, "learning_rate": 9.866252483453199e-06, "loss": 0.6867, "step": 868 }, { "epoch": 0.8343734997599616, "grad_norm": 0.8640459380647406, "learning_rate": 9.865481749560933e-06, "loss": 0.7011, "step": 869 }, { "epoch": 0.8353336533845416, "grad_norm": 0.8485072519746703, "learning_rate": 9.864708831605918e-06, "loss": 0.6631, "step": 870 }, { "epoch": 0.8362938070091215, "grad_norm": 0.9567775842749096, "learning_rate": 9.863933729935112e-06, "loss": 0.7182, "step": 871 }, { "epoch": 0.8372539606337014, "grad_norm": 1.0516381759578322, "learning_rate": 9.863156444896446e-06, "loss": 0.7079, "step": 872 }, { "epoch": 0.8382141142582813, "grad_norm": 0.9413927447733244, "learning_rate": 9.862376976838835e-06, "loss": 0.6884, "step": 873 }, { "epoch": 0.8391742678828612, "grad_norm": 0.9195205913319885, "learning_rate": 9.861595326112176e-06, "loss": 0.6712, "step": 874 }, { "epoch": 0.8401344215074412, "grad_norm": 1.1783287883587388, "learning_rate": 9.860811493067344e-06, "loss": 0.8669, "step": 875 }, { "epoch": 0.8410945751320211, "grad_norm": 0.8772907205474011, "learning_rate": 9.86002547805619e-06, "loss": 0.6561, "step": 876 }, { "epoch": 0.842054728756601, "grad_norm": 0.9754715934711046, "learning_rate": 9.85923728143155e-06, "loss": 0.6664, "step": 877 }, { "epoch": 0.8430148823811809, "grad_norm": 0.7871209727846218, "learning_rate": 9.858446903547239e-06, "loss": 0.5431, "step": 878 }, { "epoch": 0.843975036005761, "grad_norm": 1.1809738146459918, "learning_rate": 9.857654344758044e-06, "loss": 0.6913, "step": 879 }, { "epoch": 0.8449351896303409, "grad_norm": 0.8972240102728138, "learning_rate": 9.85685960541974e-06, "loss": 0.6584, "step": 880 }, { "epoch": 0.8458953432549208, "grad_norm": 1.2119601943426932, "learning_rate": 9.856062685889076e-06, "loss": 0.8084, "step": 881 }, { "epoch": 0.8468554968795007, "grad_norm": 0.8782401519864343, "learning_rate": 9.855263586523782e-06, "loss": 0.6934, "step": 882 }, { "epoch": 0.8478156505040807, "grad_norm": 1.1011422333215695, "learning_rate": 9.854462307682561e-06, "loss": 0.8057, "step": 883 }, { "epoch": 0.8487758041286606, "grad_norm": 0.870739816201533, "learning_rate": 9.853658849725105e-06, "loss": 0.6878, "step": 884 }, { "epoch": 0.8497359577532405, "grad_norm": 0.8889682628805293, "learning_rate": 9.852853213012072e-06, "loss": 0.7057, "step": 885 }, { "epoch": 0.8506961113778204, "grad_norm": 0.8928220364354131, "learning_rate": 9.852045397905106e-06, "loss": 0.6982, "step": 886 }, { "epoch": 0.8516562650024004, "grad_norm": 0.7685967917387115, "learning_rate": 9.851235404766827e-06, "loss": 0.6023, "step": 887 }, { "epoch": 0.8526164186269803, "grad_norm": 0.9419788019962784, "learning_rate": 9.850423233960831e-06, "loss": 0.7657, "step": 888 }, { "epoch": 0.8535765722515603, "grad_norm": 0.746014210751407, "learning_rate": 9.849608885851693e-06, "loss": 0.6511, "step": 889 }, { "epoch": 0.8545367258761402, "grad_norm": 0.9318182546273395, "learning_rate": 9.848792360804966e-06, "loss": 0.7648, "step": 890 }, { "epoch": 0.8554968795007201, "grad_norm": 0.8225579171099308, "learning_rate": 9.84797365918718e-06, "loss": 0.6408, "step": 891 }, { "epoch": 0.8564570331253001, "grad_norm": 0.9001477176865523, "learning_rate": 9.847152781365839e-06, "loss": 0.8032, "step": 892 }, { "epoch": 0.85741718674988, "grad_norm": 0.8208613585304696, "learning_rate": 9.846329727709429e-06, "loss": 0.6743, "step": 893 }, { "epoch": 0.8583773403744599, "grad_norm": 0.8524338800839776, "learning_rate": 9.845504498587406e-06, "loss": 0.7262, "step": 894 }, { "epoch": 0.8593374939990398, "grad_norm": 0.8346725616858679, "learning_rate": 9.844677094370212e-06, "loss": 0.7411, "step": 895 }, { "epoch": 0.8602976476236198, "grad_norm": 1.046399564666947, "learning_rate": 9.843847515429256e-06, "loss": 0.6454, "step": 896 }, { "epoch": 0.8612578012481997, "grad_norm": 0.8803702146937313, "learning_rate": 9.843015762136927e-06, "loss": 0.6747, "step": 897 }, { "epoch": 0.8622179548727796, "grad_norm": 1.0798459853586526, "learning_rate": 9.842181834866593e-06, "loss": 0.6195, "step": 898 }, { "epoch": 0.8631781084973595, "grad_norm": 1.0588860904838175, "learning_rate": 9.841345733992594e-06, "loss": 0.6548, "step": 899 }, { "epoch": 0.8641382621219396, "grad_norm": 0.9068423406183603, "learning_rate": 9.840507459890244e-06, "loss": 0.7016, "step": 900 }, { "epoch": 0.8650984157465195, "grad_norm": 0.8228043737201794, "learning_rate": 9.839667012935838e-06, "loss": 0.7098, "step": 901 }, { "epoch": 0.8660585693710994, "grad_norm": 0.952911120254865, "learning_rate": 9.838824393506644e-06, "loss": 0.7752, "step": 902 }, { "epoch": 0.8670187229956793, "grad_norm": 1.1598153756048137, "learning_rate": 9.837979601980902e-06, "loss": 0.8063, "step": 903 }, { "epoch": 0.8679788766202592, "grad_norm": 0.9386315769952516, "learning_rate": 9.837132638737834e-06, "loss": 0.68, "step": 904 }, { "epoch": 0.8689390302448392, "grad_norm": 1.087462847554996, "learning_rate": 9.83628350415763e-06, "loss": 0.7008, "step": 905 }, { "epoch": 0.8698991838694191, "grad_norm": 0.8411771694143324, "learning_rate": 9.835432198621457e-06, "loss": 0.726, "step": 906 }, { "epoch": 0.870859337493999, "grad_norm": 0.7553179773226595, "learning_rate": 9.83457872251146e-06, "loss": 0.566, "step": 907 }, { "epoch": 0.8718194911185789, "grad_norm": 0.8433064821555623, "learning_rate": 9.833723076210751e-06, "loss": 0.593, "step": 908 }, { "epoch": 0.8727796447431589, "grad_norm": 0.9530615935705777, "learning_rate": 9.832865260103423e-06, "loss": 0.7549, "step": 909 }, { "epoch": 0.8737397983677389, "grad_norm": 0.8515082031148432, "learning_rate": 9.832005274574541e-06, "loss": 0.72, "step": 910 }, { "epoch": 0.8746999519923188, "grad_norm": 0.9108880461763527, "learning_rate": 9.831143120010143e-06, "loss": 0.6999, "step": 911 }, { "epoch": 0.8756601056168987, "grad_norm": 0.91376489844038, "learning_rate": 9.830278796797239e-06, "loss": 0.6421, "step": 912 }, { "epoch": 0.8766202592414787, "grad_norm": 0.7740618403814947, "learning_rate": 9.829412305323818e-06, "loss": 0.6508, "step": 913 }, { "epoch": 0.8775804128660586, "grad_norm": 0.7543520755041994, "learning_rate": 9.828543645978835e-06, "loss": 0.7141, "step": 914 }, { "epoch": 0.8785405664906385, "grad_norm": 0.9000528487370836, "learning_rate": 9.827672819152223e-06, "loss": 0.703, "step": 915 }, { "epoch": 0.8795007201152184, "grad_norm": 0.8433161228792307, "learning_rate": 9.82679982523489e-06, "loss": 0.6853, "step": 916 }, { "epoch": 0.8804608737397984, "grad_norm": 0.8052708262643308, "learning_rate": 9.82592466461871e-06, "loss": 0.6944, "step": 917 }, { "epoch": 0.8814210273643783, "grad_norm": 0.7343744470176006, "learning_rate": 9.825047337696532e-06, "loss": 0.6551, "step": 918 }, { "epoch": 0.8823811809889582, "grad_norm": 0.7920120054193387, "learning_rate": 9.824167844862184e-06, "loss": 0.6621, "step": 919 }, { "epoch": 0.8833413346135381, "grad_norm": 1.1589241093425493, "learning_rate": 9.823286186510456e-06, "loss": 0.8307, "step": 920 }, { "epoch": 0.884301488238118, "grad_norm": 1.0308411276465357, "learning_rate": 9.822402363037118e-06, "loss": 0.7272, "step": 921 }, { "epoch": 0.8852616418626981, "grad_norm": 1.1822353207252856, "learning_rate": 9.82151637483891e-06, "loss": 0.7508, "step": 922 }, { "epoch": 0.886221795487278, "grad_norm": 0.7857884748804949, "learning_rate": 9.820628222313535e-06, "loss": 0.7076, "step": 923 }, { "epoch": 0.8871819491118579, "grad_norm": 0.9499440393131786, "learning_rate": 9.819737905859685e-06, "loss": 0.7271, "step": 924 }, { "epoch": 0.8881421027364378, "grad_norm": 0.9350469205321926, "learning_rate": 9.818845425877008e-06, "loss": 0.7614, "step": 925 }, { "epoch": 0.8891022563610178, "grad_norm": 0.9601261375109553, "learning_rate": 9.81795078276613e-06, "loss": 0.6873, "step": 926 }, { "epoch": 0.8900624099855977, "grad_norm": 0.7729371481310917, "learning_rate": 9.817053976928646e-06, "loss": 0.6482, "step": 927 }, { "epoch": 0.8910225636101776, "grad_norm": 0.9125764873861811, "learning_rate": 9.816155008767123e-06, "loss": 0.6429, "step": 928 }, { "epoch": 0.8919827172347575, "grad_norm": 0.8494292868618958, "learning_rate": 9.8152538786851e-06, "loss": 0.7187, "step": 929 }, { "epoch": 0.8929428708593375, "grad_norm": 1.0327035855245628, "learning_rate": 9.814350587087083e-06, "loss": 0.8386, "step": 930 }, { "epoch": 0.8939030244839175, "grad_norm": 1.25704632097803, "learning_rate": 9.813445134378548e-06, "loss": 0.6712, "step": 931 }, { "epoch": 0.8948631781084974, "grad_norm": 1.4071471853556032, "learning_rate": 9.812537520965946e-06, "loss": 0.7885, "step": 932 }, { "epoch": 0.8958233317330773, "grad_norm": 0.8047370709280061, "learning_rate": 9.811627747256694e-06, "loss": 0.6184, "step": 933 }, { "epoch": 0.8967834853576572, "grad_norm": 0.8961450450622589, "learning_rate": 9.81071581365918e-06, "loss": 0.7482, "step": 934 }, { "epoch": 0.8977436389822372, "grad_norm": 1.0721950111620535, "learning_rate": 9.809801720582761e-06, "loss": 0.7676, "step": 935 }, { "epoch": 0.8987037926068171, "grad_norm": 1.4637312897559682, "learning_rate": 9.808885468437764e-06, "loss": 0.857, "step": 936 }, { "epoch": 0.899663946231397, "grad_norm": 0.916791958511611, "learning_rate": 9.807967057635484e-06, "loss": 0.7157, "step": 937 }, { "epoch": 0.9006240998559769, "grad_norm": 0.9006178053517172, "learning_rate": 9.807046488588186e-06, "loss": 0.79, "step": 938 }, { "epoch": 0.9015842534805569, "grad_norm": 0.7734758019438522, "learning_rate": 9.806123761709103e-06, "loss": 0.6086, "step": 939 }, { "epoch": 0.9025444071051368, "grad_norm": 0.8565853824491059, "learning_rate": 9.80519887741244e-06, "loss": 0.7378, "step": 940 }, { "epoch": 0.9035045607297167, "grad_norm": 1.099152577132124, "learning_rate": 9.804271836113364e-06, "loss": 0.8379, "step": 941 }, { "epoch": 0.9044647143542967, "grad_norm": 1.0021886214024693, "learning_rate": 9.803342638228014e-06, "loss": 0.7335, "step": 942 }, { "epoch": 0.9054248679788767, "grad_norm": 0.7582353201401353, "learning_rate": 9.802411284173502e-06, "loss": 0.6416, "step": 943 }, { "epoch": 0.9063850216034566, "grad_norm": 0.9255846160567454, "learning_rate": 9.801477774367898e-06, "loss": 0.7566, "step": 944 }, { "epoch": 0.9073451752280365, "grad_norm": 1.012738557541491, "learning_rate": 9.800542109230247e-06, "loss": 0.6663, "step": 945 }, { "epoch": 0.9083053288526164, "grad_norm": 0.9937585511100733, "learning_rate": 9.799604289180556e-06, "loss": 0.8013, "step": 946 }, { "epoch": 0.9092654824771963, "grad_norm": 0.8613197675950497, "learning_rate": 9.798664314639805e-06, "loss": 0.6435, "step": 947 }, { "epoch": 0.9102256361017763, "grad_norm": 1.0101048757422355, "learning_rate": 9.79772218602994e-06, "loss": 0.7303, "step": 948 }, { "epoch": 0.9111857897263562, "grad_norm": 0.9053856499081823, "learning_rate": 9.796777903773868e-06, "loss": 0.7051, "step": 949 }, { "epoch": 0.9121459433509361, "grad_norm": 0.8294829420202751, "learning_rate": 9.795831468295472e-06, "loss": 0.7455, "step": 950 }, { "epoch": 0.913106096975516, "grad_norm": 1.0436987006919598, "learning_rate": 9.794882880019594e-06, "loss": 0.7973, "step": 951 }, { "epoch": 0.914066250600096, "grad_norm": 0.9167698848756176, "learning_rate": 9.793932139372045e-06, "loss": 0.7078, "step": 952 }, { "epoch": 0.915026404224676, "grad_norm": 1.1052562719758796, "learning_rate": 9.792979246779602e-06, "loss": 0.8434, "step": 953 }, { "epoch": 0.9159865578492559, "grad_norm": 0.9050969429807252, "learning_rate": 9.79202420267001e-06, "loss": 0.7025, "step": 954 }, { "epoch": 0.9169467114738358, "grad_norm": 0.7367349047104323, "learning_rate": 9.791067007471977e-06, "loss": 0.6527, "step": 955 }, { "epoch": 0.9179068650984158, "grad_norm": 0.846560830443399, "learning_rate": 9.790107661615179e-06, "loss": 0.7406, "step": 956 }, { "epoch": 0.9188670187229957, "grad_norm": 0.9029119912412141, "learning_rate": 9.789146165530255e-06, "loss": 0.7326, "step": 957 }, { "epoch": 0.9198271723475756, "grad_norm": 1.0029747596473106, "learning_rate": 9.78818251964881e-06, "loss": 0.7213, "step": 958 }, { "epoch": 0.9207873259721555, "grad_norm": 0.885570869841395, "learning_rate": 9.787216724403414e-06, "loss": 0.733, "step": 959 }, { "epoch": 0.9217474795967355, "grad_norm": 0.886645228108422, "learning_rate": 9.786248780227604e-06, "loss": 0.7216, "step": 960 }, { "epoch": 0.9227076332213154, "grad_norm": 0.7986871098154948, "learning_rate": 9.785278687555879e-06, "loss": 0.674, "step": 961 }, { "epoch": 0.9236677868458953, "grad_norm": 0.8986004495176464, "learning_rate": 9.784306446823702e-06, "loss": 0.6393, "step": 962 }, { "epoch": 0.9246279404704753, "grad_norm": 0.86181323967847, "learning_rate": 9.783332058467502e-06, "loss": 0.7279, "step": 963 }, { "epoch": 0.9255880940950552, "grad_norm": 0.7353375746103941, "learning_rate": 9.782355522924674e-06, "loss": 0.6476, "step": 964 }, { "epoch": 0.9265482477196352, "grad_norm": 0.9322836841869301, "learning_rate": 9.78137684063357e-06, "loss": 0.7358, "step": 965 }, { "epoch": 0.9275084013442151, "grad_norm": 1.1366894034596995, "learning_rate": 9.780396012033514e-06, "loss": 0.6432, "step": 966 }, { "epoch": 0.928468554968795, "grad_norm": 0.9113528026289865, "learning_rate": 9.779413037564787e-06, "loss": 0.6448, "step": 967 }, { "epoch": 0.9294287085933749, "grad_norm": 1.087468585048316, "learning_rate": 9.778427917668636e-06, "loss": 0.7209, "step": 968 }, { "epoch": 0.9303888622179549, "grad_norm": 1.0563667320505763, "learning_rate": 9.777440652787273e-06, "loss": 0.7589, "step": 969 }, { "epoch": 0.9313490158425348, "grad_norm": 0.7211554814732748, "learning_rate": 9.776451243363866e-06, "loss": 0.5118, "step": 970 }, { "epoch": 0.9323091694671147, "grad_norm": 1.230910824786631, "learning_rate": 9.775459689842558e-06, "loss": 0.7975, "step": 971 }, { "epoch": 0.9332693230916946, "grad_norm": 2.4519137812887766, "learning_rate": 9.774465992668438e-06, "loss": 0.6522, "step": 972 }, { "epoch": 0.9342294767162747, "grad_norm": 0.931230107570766, "learning_rate": 9.773470152287573e-06, "loss": 0.7015, "step": 973 }, { "epoch": 0.9351896303408546, "grad_norm": 0.7960692587293202, "learning_rate": 9.772472169146977e-06, "loss": 0.5862, "step": 974 }, { "epoch": 0.9361497839654345, "grad_norm": 1.149067562691375, "learning_rate": 9.771472043694645e-06, "loss": 0.7183, "step": 975 }, { "epoch": 0.9371099375900144, "grad_norm": 0.9003261740503984, "learning_rate": 9.770469776379514e-06, "loss": 0.6604, "step": 976 }, { "epoch": 0.9380700912145943, "grad_norm": 0.8865707124888934, "learning_rate": 9.769465367651494e-06, "loss": 0.7534, "step": 977 }, { "epoch": 0.9390302448391743, "grad_norm": 0.9029657074496114, "learning_rate": 9.768458817961455e-06, "loss": 0.6492, "step": 978 }, { "epoch": 0.9399903984637542, "grad_norm": 0.8735830280990771, "learning_rate": 9.767450127761223e-06, "loss": 0.705, "step": 979 }, { "epoch": 0.9409505520883341, "grad_norm": 0.8542518063327101, "learning_rate": 9.766439297503589e-06, "loss": 0.7168, "step": 980 }, { "epoch": 0.941910705712914, "grad_norm": 0.9514398633010328, "learning_rate": 9.765426327642306e-06, "loss": 0.7887, "step": 981 }, { "epoch": 0.942870859337494, "grad_norm": 0.9104882795807692, "learning_rate": 9.764411218632085e-06, "loss": 0.7036, "step": 982 }, { "epoch": 0.943831012962074, "grad_norm": 1.002912069440644, "learning_rate": 9.763393970928593e-06, "loss": 0.8065, "step": 983 }, { "epoch": 0.9447911665866539, "grad_norm": 0.8578149298877878, "learning_rate": 9.76237458498847e-06, "loss": 0.6088, "step": 984 }, { "epoch": 0.9457513202112338, "grad_norm": 0.9391133796085549, "learning_rate": 9.7613530612693e-06, "loss": 0.7695, "step": 985 }, { "epoch": 0.9467114738358138, "grad_norm": 0.8128391732144891, "learning_rate": 9.760329400229639e-06, "loss": 0.6483, "step": 986 }, { "epoch": 0.9476716274603937, "grad_norm": 1.3965529118863698, "learning_rate": 9.759303602328993e-06, "loss": 0.7482, "step": 987 }, { "epoch": 0.9486317810849736, "grad_norm": 1.1364272892255123, "learning_rate": 9.758275668027834e-06, "loss": 0.8119, "step": 988 }, { "epoch": 0.9495919347095535, "grad_norm": 0.7607919191381459, "learning_rate": 9.757245597787592e-06, "loss": 0.5055, "step": 989 }, { "epoch": 0.9505520883341335, "grad_norm": 1.3512683493926583, "learning_rate": 9.756213392070654e-06, "loss": 0.847, "step": 990 }, { "epoch": 0.9515122419587134, "grad_norm": 1.0055604817556336, "learning_rate": 9.755179051340365e-06, "loss": 0.696, "step": 991 }, { "epoch": 0.9524723955832933, "grad_norm": 0.8960308207260537, "learning_rate": 9.75414257606103e-06, "loss": 0.7052, "step": 992 }, { "epoch": 0.9534325492078732, "grad_norm": 1.6528396218056225, "learning_rate": 9.75310396669791e-06, "loss": 0.65, "step": 993 }, { "epoch": 0.9543927028324531, "grad_norm": 0.9504595309189022, "learning_rate": 9.752063223717227e-06, "loss": 0.701, "step": 994 }, { "epoch": 0.9553528564570332, "grad_norm": 0.8657221623098704, "learning_rate": 9.75102034758616e-06, "loss": 0.6853, "step": 995 }, { "epoch": 0.9563130100816131, "grad_norm": 1.4498758555939062, "learning_rate": 9.749975338772846e-06, "loss": 0.858, "step": 996 }, { "epoch": 0.957273163706193, "grad_norm": 0.8358272239904824, "learning_rate": 9.748928197746376e-06, "loss": 0.6797, "step": 997 }, { "epoch": 0.9582333173307729, "grad_norm": 0.7715505618594579, "learning_rate": 9.7478789249768e-06, "loss": 0.6262, "step": 998 }, { "epoch": 0.9591934709553529, "grad_norm": 1.0550577459287398, "learning_rate": 9.746827520935129e-06, "loss": 0.7722, "step": 999 }, { "epoch": 0.9601536245799328, "grad_norm": 0.9238651853761882, "learning_rate": 9.745773986093325e-06, "loss": 0.6213, "step": 1000 }, { "epoch": 0.9611137782045127, "grad_norm": 0.8270884165456949, "learning_rate": 9.744718320924307e-06, "loss": 0.6676, "step": 1001 }, { "epoch": 0.9620739318290926, "grad_norm": 0.9442368413928642, "learning_rate": 9.743660525901952e-06, "loss": 0.7282, "step": 1002 }, { "epoch": 0.9630340854536726, "grad_norm": 0.9494012195220293, "learning_rate": 9.742600601501095e-06, "loss": 0.6505, "step": 1003 }, { "epoch": 0.9639942390782525, "grad_norm": 0.746237711806975, "learning_rate": 9.741538548197522e-06, "loss": 0.5424, "step": 1004 }, { "epoch": 0.9649543927028325, "grad_norm": 0.7240689562290187, "learning_rate": 9.74047436646798e-06, "loss": 0.6064, "step": 1005 }, { "epoch": 0.9659145463274124, "grad_norm": 1.27040585629925, "learning_rate": 9.739408056790168e-06, "loss": 0.7296, "step": 1006 }, { "epoch": 0.9668746999519923, "grad_norm": 0.8569682496384651, "learning_rate": 9.73833961964274e-06, "loss": 0.6501, "step": 1007 }, { "epoch": 0.9678348535765723, "grad_norm": 0.7889182481402485, "learning_rate": 9.737269055505308e-06, "loss": 0.6484, "step": 1008 }, { "epoch": 0.9687950072011522, "grad_norm": 0.7183187204288622, "learning_rate": 9.736196364858435e-06, "loss": 0.6282, "step": 1009 }, { "epoch": 0.9697551608257321, "grad_norm": 0.8865414454394241, "learning_rate": 9.735121548183642e-06, "loss": 0.6628, "step": 1010 }, { "epoch": 0.970715314450312, "grad_norm": 1.119802007113434, "learning_rate": 9.7340446059634e-06, "loss": 0.7767, "step": 1011 }, { "epoch": 0.971675468074892, "grad_norm": 0.8485729754401015, "learning_rate": 9.732965538681139e-06, "loss": 0.6563, "step": 1012 }, { "epoch": 0.9726356216994719, "grad_norm": 0.9115309399968349, "learning_rate": 9.73188434682124e-06, "loss": 0.7344, "step": 1013 }, { "epoch": 0.9735957753240518, "grad_norm": 0.7930990074027986, "learning_rate": 9.73080103086904e-06, "loss": 0.6992, "step": 1014 }, { "epoch": 0.9745559289486317, "grad_norm": 1.0681026116953887, "learning_rate": 9.729715591310824e-06, "loss": 0.7526, "step": 1015 }, { "epoch": 0.9755160825732118, "grad_norm": 0.9295989305510667, "learning_rate": 9.728628028633839e-06, "loss": 0.7508, "step": 1016 }, { "epoch": 0.9764762361977917, "grad_norm": 1.0247971917022445, "learning_rate": 9.727538343326277e-06, "loss": 0.6792, "step": 1017 }, { "epoch": 0.9774363898223716, "grad_norm": 0.8156349543989204, "learning_rate": 9.72644653587729e-06, "loss": 0.6052, "step": 1018 }, { "epoch": 0.9783965434469515, "grad_norm": 1.267095702708427, "learning_rate": 9.725352606776973e-06, "loss": 0.7221, "step": 1019 }, { "epoch": 0.9793566970715314, "grad_norm": 1.1773844457535763, "learning_rate": 9.724256556516383e-06, "loss": 0.7575, "step": 1020 }, { "epoch": 0.9803168506961114, "grad_norm": 0.8759037979641289, "learning_rate": 9.723158385587525e-06, "loss": 0.6295, "step": 1021 }, { "epoch": 0.9812770043206913, "grad_norm": 0.7741309756826021, "learning_rate": 9.722058094483355e-06, "loss": 0.7066, "step": 1022 }, { "epoch": 0.9822371579452712, "grad_norm": 1.0779524387453767, "learning_rate": 9.72095568369778e-06, "loss": 0.6972, "step": 1023 }, { "epoch": 0.9831973115698511, "grad_norm": 0.8201702458678146, "learning_rate": 9.719851153725668e-06, "loss": 0.6189, "step": 1024 }, { "epoch": 0.9841574651944311, "grad_norm": 1.0241763879043673, "learning_rate": 9.718744505062822e-06, "loss": 0.7156, "step": 1025 }, { "epoch": 0.985117618819011, "grad_norm": 1.3487671995413677, "learning_rate": 9.717635738206007e-06, "loss": 0.8066, "step": 1026 }, { "epoch": 0.986077772443591, "grad_norm": 0.954718850746842, "learning_rate": 9.716524853652942e-06, "loss": 0.6691, "step": 1027 }, { "epoch": 0.9870379260681709, "grad_norm": 0.9882884208643541, "learning_rate": 9.715411851902284e-06, "loss": 0.8365, "step": 1028 }, { "epoch": 0.9879980796927509, "grad_norm": 0.8928585163177973, "learning_rate": 9.71429673345365e-06, "loss": 0.6603, "step": 1029 }, { "epoch": 0.9889582333173308, "grad_norm": 0.8602045132363018, "learning_rate": 9.713179498807607e-06, "loss": 0.7759, "step": 1030 }, { "epoch": 0.9899183869419107, "grad_norm": 1.0392955316068169, "learning_rate": 9.712060148465669e-06, "loss": 0.8069, "step": 1031 }, { "epoch": 0.9908785405664906, "grad_norm": 0.7653748755297636, "learning_rate": 9.710938682930298e-06, "loss": 0.5981, "step": 1032 }, { "epoch": 0.9918386941910706, "grad_norm": 0.8993194138537441, "learning_rate": 9.709815102704909e-06, "loss": 0.7391, "step": 1033 }, { "epoch": 0.9927988478156505, "grad_norm": 1.1663062425226172, "learning_rate": 9.708689408293866e-06, "loss": 0.7099, "step": 1034 }, { "epoch": 0.9937590014402304, "grad_norm": 1.025169170332603, "learning_rate": 9.707561600202481e-06, "loss": 0.7599, "step": 1035 }, { "epoch": 0.9947191550648103, "grad_norm": 0.8490868023035005, "learning_rate": 9.706431678937015e-06, "loss": 0.6197, "step": 1036 }, { "epoch": 0.9956793086893903, "grad_norm": 0.881590711971278, "learning_rate": 9.705299645004679e-06, "loss": 0.7121, "step": 1037 }, { "epoch": 0.9966394623139703, "grad_norm": 0.8226229058494607, "learning_rate": 9.704165498913624e-06, "loss": 0.6792, "step": 1038 }, { "epoch": 0.9975996159385502, "grad_norm": 0.8461723052514307, "learning_rate": 9.703029241172965e-06, "loss": 0.7096, "step": 1039 }, { "epoch": 0.9985597695631301, "grad_norm": 1.4493899513094717, "learning_rate": 9.701890872292753e-06, "loss": 0.8469, "step": 1040 }, { "epoch": 0.99951992318771, "grad_norm": 0.8212880794034917, "learning_rate": 9.700750392783986e-06, "loss": 0.6593, "step": 1041 }, { "epoch": 1.0, "grad_norm": 0.8212880794034917, "learning_rate": 9.699607803158618e-06, "loss": 0.9052, "step": 1042 }, { "epoch": 1.00096015362458, "grad_norm": 1.979183416616796, "learning_rate": 9.698463103929542e-06, "loss": 0.6698, "step": 1043 }, { "epoch": 1.0019203072491598, "grad_norm": 1.194841846390985, "learning_rate": 9.697316295610604e-06, "loss": 0.6337, "step": 1044 }, { "epoch": 1.0028804608737398, "grad_norm": 0.7305704485296048, "learning_rate": 9.696167378716591e-06, "loss": 0.6687, "step": 1045 }, { "epoch": 1.0038406144983196, "grad_norm": 0.9983308072508276, "learning_rate": 9.695016353763241e-06, "loss": 0.5879, "step": 1046 }, { "epoch": 1.0048007681228996, "grad_norm": 0.8991654186275129, "learning_rate": 9.693863221267238e-06, "loss": 0.6391, "step": 1047 }, { "epoch": 1.0057609217474797, "grad_norm": 1.0751680115912563, "learning_rate": 9.692707981746208e-06, "loss": 0.6165, "step": 1048 }, { "epoch": 1.0067210753720595, "grad_norm": 0.9632006490799951, "learning_rate": 9.691550635718729e-06, "loss": 0.5246, "step": 1049 }, { "epoch": 1.0076812289966395, "grad_norm": 0.9737967200662035, "learning_rate": 9.690391183704318e-06, "loss": 0.5746, "step": 1050 }, { "epoch": 1.0086413826212195, "grad_norm": 1.0941767083212883, "learning_rate": 9.689229626223441e-06, "loss": 0.6584, "step": 1051 }, { "epoch": 1.0096015362457993, "grad_norm": 1.1275259328083957, "learning_rate": 9.688065963797512e-06, "loss": 0.6538, "step": 1052 }, { "epoch": 1.0105616898703793, "grad_norm": 1.0177432950149377, "learning_rate": 9.686900196948886e-06, "loss": 0.7093, "step": 1053 }, { "epoch": 1.011521843494959, "grad_norm": 1.1758295486424089, "learning_rate": 9.685732326200859e-06, "loss": 0.6063, "step": 1054 }, { "epoch": 1.0124819971195391, "grad_norm": 0.8036535371131399, "learning_rate": 9.684562352077679e-06, "loss": 0.5458, "step": 1055 }, { "epoch": 1.0134421507441191, "grad_norm": 0.8523996887077252, "learning_rate": 9.683390275104533e-06, "loss": 0.5606, "step": 1056 }, { "epoch": 1.014402304368699, "grad_norm": 0.6886197269579926, "learning_rate": 9.682216095807559e-06, "loss": 0.6589, "step": 1057 }, { "epoch": 1.015362457993279, "grad_norm": 0.9199186343288881, "learning_rate": 9.681039814713827e-06, "loss": 0.5962, "step": 1058 }, { "epoch": 1.0163226116178588, "grad_norm": 0.8556150302431942, "learning_rate": 9.67986143235136e-06, "loss": 0.6507, "step": 1059 }, { "epoch": 1.0172827652424388, "grad_norm": 0.9465823364290421, "learning_rate": 9.67868094924912e-06, "loss": 0.6125, "step": 1060 }, { "epoch": 1.0182429188670188, "grad_norm": 1.044185258102178, "learning_rate": 9.677498365937017e-06, "loss": 0.6174, "step": 1061 }, { "epoch": 1.0192030724915986, "grad_norm": 0.658112961311754, "learning_rate": 9.676313682945896e-06, "loss": 0.4621, "step": 1062 }, { "epoch": 1.0201632261161786, "grad_norm": 1.0934825991930441, "learning_rate": 9.675126900807548e-06, "loss": 0.6556, "step": 1063 }, { "epoch": 1.0211233797407586, "grad_norm": 1.2092416883751174, "learning_rate": 9.673938020054709e-06, "loss": 0.6299, "step": 1064 }, { "epoch": 1.0220835333653384, "grad_norm": 0.9431979147437243, "learning_rate": 9.672747041221055e-06, "loss": 0.5121, "step": 1065 }, { "epoch": 1.0230436869899184, "grad_norm": 0.7591028563484827, "learning_rate": 9.671553964841206e-06, "loss": 0.6919, "step": 1066 }, { "epoch": 1.0240038406144982, "grad_norm": 1.012039783903709, "learning_rate": 9.670358791450713e-06, "loss": 0.6697, "step": 1067 }, { "epoch": 1.0249639942390782, "grad_norm": 0.927961175032116, "learning_rate": 9.669161521586086e-06, "loss": 0.697, "step": 1068 }, { "epoch": 1.0259241478636583, "grad_norm": 1.0235526352012247, "learning_rate": 9.66796215578476e-06, "loss": 0.6848, "step": 1069 }, { "epoch": 1.026884301488238, "grad_norm": 0.9662517812913155, "learning_rate": 9.666760694585122e-06, "loss": 0.6561, "step": 1070 }, { "epoch": 1.027844455112818, "grad_norm": 0.7741909714540829, "learning_rate": 9.665557138526492e-06, "loss": 0.606, "step": 1071 }, { "epoch": 1.0288046087373979, "grad_norm": 0.95939756208335, "learning_rate": 9.664351488149136e-06, "loss": 0.6796, "step": 1072 }, { "epoch": 1.029764762361978, "grad_norm": 0.7937556521270224, "learning_rate": 9.663143743994254e-06, "loss": 0.6183, "step": 1073 }, { "epoch": 1.030724915986558, "grad_norm": 1.188778091335553, "learning_rate": 9.661933906603994e-06, "loss": 0.6158, "step": 1074 }, { "epoch": 1.0316850696111377, "grad_norm": 0.7760575064382742, "learning_rate": 9.660721976521433e-06, "loss": 0.5965, "step": 1075 }, { "epoch": 1.0326452232357177, "grad_norm": 0.7771969993051077, "learning_rate": 9.659507954290599e-06, "loss": 0.6805, "step": 1076 }, { "epoch": 1.0336053768602977, "grad_norm": 1.0477643317113738, "learning_rate": 9.658291840456454e-06, "loss": 0.5891, "step": 1077 }, { "epoch": 1.0345655304848775, "grad_norm": 1.2387882313065417, "learning_rate": 9.657073635564893e-06, "loss": 0.659, "step": 1078 }, { "epoch": 1.0355256841094576, "grad_norm": 0.6482059068817437, "learning_rate": 9.65585334016276e-06, "loss": 0.6532, "step": 1079 }, { "epoch": 1.0364858377340374, "grad_norm": 1.0367620311586878, "learning_rate": 9.65463095479783e-06, "loss": 0.6562, "step": 1080 }, { "epoch": 1.0374459913586174, "grad_norm": 1.0243295375305939, "learning_rate": 9.653406480018819e-06, "loss": 0.658, "step": 1081 }, { "epoch": 1.0384061449831974, "grad_norm": 1.1934871091611878, "learning_rate": 9.652179916375382e-06, "loss": 0.6246, "step": 1082 }, { "epoch": 1.0393662986077772, "grad_norm": 0.9132653014708247, "learning_rate": 9.65095126441811e-06, "loss": 0.6321, "step": 1083 }, { "epoch": 1.0403264522323572, "grad_norm": 1.1519688997991087, "learning_rate": 9.649720524698532e-06, "loss": 0.7549, "step": 1084 }, { "epoch": 1.041286605856937, "grad_norm": 1.11655474357333, "learning_rate": 9.648487697769112e-06, "loss": 0.6974, "step": 1085 }, { "epoch": 1.042246759481517, "grad_norm": 0.8001045357547697, "learning_rate": 9.647252784183254e-06, "loss": 0.5838, "step": 1086 }, { "epoch": 1.043206913106097, "grad_norm": 0.9818246098222206, "learning_rate": 9.646015784495297e-06, "loss": 0.5733, "step": 1087 }, { "epoch": 1.0441670667306768, "grad_norm": 1.1020710671995395, "learning_rate": 9.64477669926052e-06, "loss": 0.6603, "step": 1088 }, { "epoch": 1.0451272203552568, "grad_norm": 0.9120011192920432, "learning_rate": 9.643535529035131e-06, "loss": 0.6657, "step": 1089 }, { "epoch": 1.0460873739798369, "grad_norm": 0.8251180365224347, "learning_rate": 9.64229227437628e-06, "loss": 0.6821, "step": 1090 }, { "epoch": 1.0470475276044167, "grad_norm": 1.0371004928464527, "learning_rate": 9.641046935842048e-06, "loss": 0.6696, "step": 1091 }, { "epoch": 1.0480076812289967, "grad_norm": 0.9172883027482215, "learning_rate": 9.63979951399146e-06, "loss": 0.7055, "step": 1092 }, { "epoch": 1.0489678348535765, "grad_norm": 1.07257660780464, "learning_rate": 9.638550009384467e-06, "loss": 0.6431, "step": 1093 }, { "epoch": 1.0499279884781565, "grad_norm": 0.9023486406802833, "learning_rate": 9.637298422581955e-06, "loss": 0.7009, "step": 1094 }, { "epoch": 1.0508881421027365, "grad_norm": 0.7659909309667764, "learning_rate": 9.636044754145753e-06, "loss": 0.6158, "step": 1095 }, { "epoch": 1.0518482957273163, "grad_norm": 0.6618032315624883, "learning_rate": 9.634789004638618e-06, "loss": 0.6709, "step": 1096 }, { "epoch": 1.0528084493518963, "grad_norm": 1.1654588358065938, "learning_rate": 9.63353117462424e-06, "loss": 0.7293, "step": 1097 }, { "epoch": 1.0537686029764763, "grad_norm": 1.348628078256719, "learning_rate": 9.632271264667249e-06, "loss": 0.7297, "step": 1098 }, { "epoch": 1.0547287566010561, "grad_norm": 1.1698650755575342, "learning_rate": 9.631009275333204e-06, "loss": 0.7558, "step": 1099 }, { "epoch": 1.0556889102256362, "grad_norm": 1.0453753317425176, "learning_rate": 9.629745207188596e-06, "loss": 0.5465, "step": 1100 }, { "epoch": 1.056649063850216, "grad_norm": 0.7766400270936995, "learning_rate": 9.628479060800855e-06, "loss": 0.5921, "step": 1101 }, { "epoch": 1.057609217474796, "grad_norm": 0.9929731330790875, "learning_rate": 9.627210836738338e-06, "loss": 0.6077, "step": 1102 }, { "epoch": 1.058569371099376, "grad_norm": 0.8850049618710202, "learning_rate": 9.62594053557034e-06, "loss": 0.6798, "step": 1103 }, { "epoch": 1.0595295247239558, "grad_norm": 1.1702962644155386, "learning_rate": 9.624668157867084e-06, "loss": 0.6917, "step": 1104 }, { "epoch": 1.0604896783485358, "grad_norm": 0.9982744771831288, "learning_rate": 9.623393704199727e-06, "loss": 0.5472, "step": 1105 }, { "epoch": 1.0614498319731156, "grad_norm": 0.9369894939457977, "learning_rate": 9.622117175140357e-06, "loss": 0.732, "step": 1106 }, { "epoch": 1.0624099855976956, "grad_norm": 1.1468877217951514, "learning_rate": 9.620838571261995e-06, "loss": 0.6447, "step": 1107 }, { "epoch": 1.0633701392222756, "grad_norm": 1.0100003185066795, "learning_rate": 9.619557893138593e-06, "loss": 0.597, "step": 1108 }, { "epoch": 1.0643302928468554, "grad_norm": 1.088320181681635, "learning_rate": 9.618275141345032e-06, "loss": 0.653, "step": 1109 }, { "epoch": 1.0652904464714354, "grad_norm": 1.0997609050325625, "learning_rate": 9.616990316457127e-06, "loss": 0.6024, "step": 1110 }, { "epoch": 1.0662506000960155, "grad_norm": 0.7495655889923801, "learning_rate": 9.615703419051622e-06, "loss": 0.4956, "step": 1111 }, { "epoch": 1.0672107537205953, "grad_norm": 0.8206871512520483, "learning_rate": 9.614414449706191e-06, "loss": 0.6762, "step": 1112 }, { "epoch": 1.0681709073451753, "grad_norm": 0.818610709940569, "learning_rate": 9.61312340899944e-06, "loss": 0.5018, "step": 1113 }, { "epoch": 1.069131060969755, "grad_norm": 0.9521835761642582, "learning_rate": 9.6118302975109e-06, "loss": 0.7194, "step": 1114 }, { "epoch": 1.070091214594335, "grad_norm": 0.9195037877209145, "learning_rate": 9.61053511582104e-06, "loss": 0.6628, "step": 1115 }, { "epoch": 1.0710513682189151, "grad_norm": 0.9365829319168445, "learning_rate": 9.609237864511249e-06, "loss": 0.6495, "step": 1116 }, { "epoch": 1.072011521843495, "grad_norm": 0.7664232074740698, "learning_rate": 9.607938544163851e-06, "loss": 0.6318, "step": 1117 }, { "epoch": 1.072971675468075, "grad_norm": 0.8530164083433751, "learning_rate": 9.606637155362095e-06, "loss": 0.6722, "step": 1118 }, { "epoch": 1.0739318290926547, "grad_norm": 1.258942796118102, "learning_rate": 9.605333698690164e-06, "loss": 0.6966, "step": 1119 }, { "epoch": 1.0748919827172347, "grad_norm": 1.2741648891857347, "learning_rate": 9.604028174733164e-06, "loss": 0.632, "step": 1120 }, { "epoch": 1.0758521363418148, "grad_norm": 1.3301851451352087, "learning_rate": 9.60272058407713e-06, "loss": 0.6193, "step": 1121 }, { "epoch": 1.0768122899663946, "grad_norm": 1.0894962489196836, "learning_rate": 9.601410927309026e-06, "loss": 0.7504, "step": 1122 }, { "epoch": 1.0777724435909746, "grad_norm": 1.2480245216166486, "learning_rate": 9.600099205016744e-06, "loss": 0.6121, "step": 1123 }, { "epoch": 1.0787325972155546, "grad_norm": 0.8739742543605165, "learning_rate": 9.5987854177891e-06, "loss": 0.7487, "step": 1124 }, { "epoch": 1.0796927508401344, "grad_norm": 1.147171829192483, "learning_rate": 9.597469566215841e-06, "loss": 0.6494, "step": 1125 }, { "epoch": 1.0806529044647144, "grad_norm": 1.1418731490827236, "learning_rate": 9.596151650887637e-06, "loss": 0.7185, "step": 1126 }, { "epoch": 1.0816130580892942, "grad_norm": 0.86114356154144, "learning_rate": 9.594831672396088e-06, "loss": 0.6517, "step": 1127 }, { "epoch": 1.0825732117138742, "grad_norm": 0.7785362768565863, "learning_rate": 9.593509631333717e-06, "loss": 0.6563, "step": 1128 }, { "epoch": 1.0835333653384542, "grad_norm": 1.1289462412116846, "learning_rate": 9.592185528293975e-06, "loss": 0.6713, "step": 1129 }, { "epoch": 1.084493518963034, "grad_norm": 1.3339183136502764, "learning_rate": 9.590859363871237e-06, "loss": 0.7291, "step": 1130 }, { "epoch": 1.085453672587614, "grad_norm": 1.1141408371248178, "learning_rate": 9.589531138660803e-06, "loss": 0.5129, "step": 1131 }, { "epoch": 1.086413826212194, "grad_norm": 0.8236651089692074, "learning_rate": 9.588200853258901e-06, "loss": 0.6256, "step": 1132 }, { "epoch": 1.0873739798367739, "grad_norm": 0.8502689453294225, "learning_rate": 9.586868508262681e-06, "loss": 0.6401, "step": 1133 }, { "epoch": 1.0883341334613539, "grad_norm": 0.7498645676338961, "learning_rate": 9.58553410427022e-06, "loss": 0.5073, "step": 1134 }, { "epoch": 1.0892942870859337, "grad_norm": 0.759820347717538, "learning_rate": 9.584197641880513e-06, "loss": 0.5486, "step": 1135 }, { "epoch": 1.0902544407105137, "grad_norm": 1.0981131924799012, "learning_rate": 9.582859121693488e-06, "loss": 0.6539, "step": 1136 }, { "epoch": 1.0912145943350937, "grad_norm": 0.8013599572061515, "learning_rate": 9.581518544309993e-06, "loss": 0.6782, "step": 1137 }, { "epoch": 1.0921747479596735, "grad_norm": 0.930038256036996, "learning_rate": 9.580175910331797e-06, "loss": 0.5533, "step": 1138 }, { "epoch": 1.0931349015842535, "grad_norm": 0.8798036765611653, "learning_rate": 9.578831220361593e-06, "loss": 0.6293, "step": 1139 }, { "epoch": 1.0940950552088333, "grad_norm": 0.985885749229286, "learning_rate": 9.577484475003e-06, "loss": 0.776, "step": 1140 }, { "epoch": 1.0950552088334133, "grad_norm": 0.8788974932951535, "learning_rate": 9.576135674860557e-06, "loss": 0.6545, "step": 1141 }, { "epoch": 1.0960153624579934, "grad_norm": 0.8973977163184313, "learning_rate": 9.574784820539725e-06, "loss": 0.613, "step": 1142 }, { "epoch": 1.0969755160825732, "grad_norm": 0.6857269315495494, "learning_rate": 9.573431912646888e-06, "loss": 0.5337, "step": 1143 }, { "epoch": 1.0979356697071532, "grad_norm": 0.7843364728134028, "learning_rate": 9.572076951789353e-06, "loss": 0.6356, "step": 1144 }, { "epoch": 1.0988958233317332, "grad_norm": 1.0106150963788247, "learning_rate": 9.570719938575349e-06, "loss": 0.6451, "step": 1145 }, { "epoch": 1.099855976956313, "grad_norm": 0.9969693060693495, "learning_rate": 9.56936087361402e-06, "loss": 0.6723, "step": 1146 }, { "epoch": 1.100816130580893, "grad_norm": 0.6350467867920836, "learning_rate": 9.56799975751544e-06, "loss": 0.6387, "step": 1147 }, { "epoch": 1.1017762842054728, "grad_norm": 0.8111596612716201, "learning_rate": 9.566636590890598e-06, "loss": 0.4873, "step": 1148 }, { "epoch": 1.1027364378300528, "grad_norm": 0.68008849991049, "learning_rate": 9.565271374351405e-06, "loss": 0.5706, "step": 1149 }, { "epoch": 1.1036965914546328, "grad_norm": 0.8622397472548321, "learning_rate": 9.563904108510693e-06, "loss": 0.6944, "step": 1150 }, { "epoch": 1.1046567450792126, "grad_norm": 0.9247172468344798, "learning_rate": 9.562534793982212e-06, "loss": 0.7633, "step": 1151 }, { "epoch": 1.1056168987037926, "grad_norm": 1.1270968543123132, "learning_rate": 9.56116343138063e-06, "loss": 0.6836, "step": 1152 }, { "epoch": 1.1065770523283724, "grad_norm": 0.9276872113446832, "learning_rate": 9.559790021321544e-06, "loss": 0.6687, "step": 1153 }, { "epoch": 1.1075372059529525, "grad_norm": 0.8167219985434703, "learning_rate": 9.558414564421456e-06, "loss": 0.6429, "step": 1154 }, { "epoch": 1.1084973595775325, "grad_norm": 0.9585175363566261, "learning_rate": 9.557037061297799e-06, "loss": 0.5642, "step": 1155 }, { "epoch": 1.1094575132021123, "grad_norm": 1.2224106829795809, "learning_rate": 9.555657512568916e-06, "loss": 0.6614, "step": 1156 }, { "epoch": 1.1104176668266923, "grad_norm": 0.9267787954366882, "learning_rate": 9.554275918854073e-06, "loss": 0.6259, "step": 1157 }, { "epoch": 1.1113778204512723, "grad_norm": 1.1493752525001002, "learning_rate": 9.55289228077345e-06, "loss": 0.5238, "step": 1158 }, { "epoch": 1.112337974075852, "grad_norm": 1.1584797522544703, "learning_rate": 9.551506598948152e-06, "loss": 0.6087, "step": 1159 }, { "epoch": 1.1132981277004321, "grad_norm": 1.1429968595638211, "learning_rate": 9.550118874000196e-06, "loss": 0.6434, "step": 1160 }, { "epoch": 1.114258281325012, "grad_norm": 0.7908463265202398, "learning_rate": 9.548729106552515e-06, "loss": 0.6269, "step": 1161 }, { "epoch": 1.115218434949592, "grad_norm": 0.8601908471652295, "learning_rate": 9.54733729722896e-06, "loss": 0.7062, "step": 1162 }, { "epoch": 1.116178588574172, "grad_norm": 1.3161185108961662, "learning_rate": 9.5459434466543e-06, "loss": 0.7185, "step": 1163 }, { "epoch": 1.1171387421987518, "grad_norm": 1.2609632672502693, "learning_rate": 9.544547555454222e-06, "loss": 0.6282, "step": 1164 }, { "epoch": 1.1180988958233318, "grad_norm": 1.0107282981700176, "learning_rate": 9.543149624255326e-06, "loss": 0.7142, "step": 1165 }, { "epoch": 1.1190590494479116, "grad_norm": 1.2933085701879472, "learning_rate": 9.541749653685124e-06, "loss": 0.6152, "step": 1166 }, { "epoch": 1.1200192030724916, "grad_norm": 0.9493485838509301, "learning_rate": 9.540347644372054e-06, "loss": 0.5639, "step": 1167 }, { "epoch": 1.1209793566970716, "grad_norm": 1.1699928407130014, "learning_rate": 9.538943596945459e-06, "loss": 0.6672, "step": 1168 }, { "epoch": 1.1219395103216514, "grad_norm": 1.054846191171314, "learning_rate": 9.5375375120356e-06, "loss": 0.5749, "step": 1169 }, { "epoch": 1.1228996639462314, "grad_norm": 0.9755461794647705, "learning_rate": 9.536129390273659e-06, "loss": 0.5947, "step": 1170 }, { "epoch": 1.1238598175708114, "grad_norm": 1.0952428139847787, "learning_rate": 9.53471923229172e-06, "loss": 0.6303, "step": 1171 }, { "epoch": 1.1248199711953912, "grad_norm": 0.8293316205843364, "learning_rate": 9.533307038722792e-06, "loss": 0.6388, "step": 1172 }, { "epoch": 1.1257801248199713, "grad_norm": 0.853044841870104, "learning_rate": 9.531892810200793e-06, "loss": 0.6796, "step": 1173 }, { "epoch": 1.126740278444551, "grad_norm": 0.9677329460829311, "learning_rate": 9.530476547360555e-06, "loss": 0.6259, "step": 1174 }, { "epoch": 1.127700432069131, "grad_norm": 0.9214410165496613, "learning_rate": 9.529058250837823e-06, "loss": 0.7027, "step": 1175 }, { "epoch": 1.128660585693711, "grad_norm": 1.0208371137778163, "learning_rate": 9.527637921269256e-06, "loss": 0.6555, "step": 1176 }, { "epoch": 1.1296207393182909, "grad_norm": 0.8577142249773448, "learning_rate": 9.526215559292423e-06, "loss": 0.6555, "step": 1177 }, { "epoch": 1.130580892942871, "grad_norm": 0.9949978857012636, "learning_rate": 9.524791165545808e-06, "loss": 0.6785, "step": 1178 }, { "epoch": 1.1315410465674507, "grad_norm": 0.8371581691784334, "learning_rate": 9.523364740668805e-06, "loss": 0.5837, "step": 1179 }, { "epoch": 1.1325012001920307, "grad_norm": 0.7721894052547845, "learning_rate": 9.521936285301723e-06, "loss": 0.5555, "step": 1180 }, { "epoch": 1.1334613538166107, "grad_norm": 1.1530392126775835, "learning_rate": 9.52050580008578e-06, "loss": 0.647, "step": 1181 }, { "epoch": 1.1344215074411905, "grad_norm": 0.8882584796986628, "learning_rate": 9.519073285663103e-06, "loss": 0.633, "step": 1182 }, { "epoch": 1.1353816610657705, "grad_norm": 0.6625066874414952, "learning_rate": 9.517638742676736e-06, "loss": 0.5108, "step": 1183 }, { "epoch": 1.1363418146903506, "grad_norm": 1.095126816439967, "learning_rate": 9.516202171770627e-06, "loss": 0.664, "step": 1184 }, { "epoch": 1.1373019683149304, "grad_norm": 0.9821853969454247, "learning_rate": 9.51476357358964e-06, "loss": 0.5535, "step": 1185 }, { "epoch": 1.1382621219395104, "grad_norm": 0.8230541433387345, "learning_rate": 9.513322948779544e-06, "loss": 0.5974, "step": 1186 }, { "epoch": 1.1392222755640902, "grad_norm": 1.0318747066486453, "learning_rate": 9.511880297987022e-06, "loss": 0.6015, "step": 1187 }, { "epoch": 1.1401824291886702, "grad_norm": 0.9283572975370646, "learning_rate": 9.510435621859662e-06, "loss": 0.5332, "step": 1188 }, { "epoch": 1.1411425828132502, "grad_norm": 1.1622214530553203, "learning_rate": 9.508988921045967e-06, "loss": 0.7067, "step": 1189 }, { "epoch": 1.14210273643783, "grad_norm": 1.0644823730514474, "learning_rate": 9.507540196195342e-06, "loss": 0.646, "step": 1190 }, { "epoch": 1.14306289006241, "grad_norm": 1.0652858819332374, "learning_rate": 9.506089447958108e-06, "loss": 0.6199, "step": 1191 }, { "epoch": 1.1440230436869898, "grad_norm": 0.8973435469403997, "learning_rate": 9.504636676985486e-06, "loss": 0.6758, "step": 1192 }, { "epoch": 1.1449831973115698, "grad_norm": 1.1024344939681154, "learning_rate": 9.503181883929613e-06, "loss": 0.6629, "step": 1193 }, { "epoch": 1.1459433509361499, "grad_norm": 0.823155620797796, "learning_rate": 9.501725069443528e-06, "loss": 0.5352, "step": 1194 }, { "epoch": 1.1469035045607296, "grad_norm": 0.920089277861052, "learning_rate": 9.50026623418118e-06, "loss": 0.7027, "step": 1195 }, { "epoch": 1.1478636581853097, "grad_norm": 0.8626526128244462, "learning_rate": 9.498805378797427e-06, "loss": 0.5911, "step": 1196 }, { "epoch": 1.1488238118098897, "grad_norm": 0.7579879005315907, "learning_rate": 9.497342503948027e-06, "loss": 0.5749, "step": 1197 }, { "epoch": 1.1497839654344695, "grad_norm": 1.0169235512650967, "learning_rate": 9.49587761028965e-06, "loss": 0.6787, "step": 1198 }, { "epoch": 1.1507441190590495, "grad_norm": 0.8523371895264148, "learning_rate": 9.494410698479873e-06, "loss": 0.7058, "step": 1199 }, { "epoch": 1.1517042726836293, "grad_norm": 1.2005178651937876, "learning_rate": 9.492941769177176e-06, "loss": 0.7328, "step": 1200 }, { "epoch": 1.1526644263082093, "grad_norm": 0.8832048553303321, "learning_rate": 9.491470823040944e-06, "loss": 0.5712, "step": 1201 }, { "epoch": 1.1536245799327893, "grad_norm": 0.8036017943248163, "learning_rate": 9.489997860731475e-06, "loss": 0.5606, "step": 1202 }, { "epoch": 1.1545847335573691, "grad_norm": 0.9504085825190297, "learning_rate": 9.488522882909959e-06, "loss": 0.5772, "step": 1203 }, { "epoch": 1.1555448871819491, "grad_norm": 1.1058255172730238, "learning_rate": 9.4870458902385e-06, "loss": 0.7414, "step": 1204 }, { "epoch": 1.156505040806529, "grad_norm": 1.0494360528330533, "learning_rate": 9.485566883380105e-06, "loss": 0.5511, "step": 1205 }, { "epoch": 1.157465194431109, "grad_norm": 0.8050395399943184, "learning_rate": 9.484085862998685e-06, "loss": 0.5705, "step": 1206 }, { "epoch": 1.158425348055689, "grad_norm": 1.3210545259209256, "learning_rate": 9.482602829759052e-06, "loss": 0.7273, "step": 1207 }, { "epoch": 1.1593855016802688, "grad_norm": 1.0670356390118088, "learning_rate": 9.481117784326927e-06, "loss": 0.6525, "step": 1208 }, { "epoch": 1.1603456553048488, "grad_norm": 1.0506469846147937, "learning_rate": 9.479630727368927e-06, "loss": 0.6605, "step": 1209 }, { "epoch": 1.1613058089294288, "grad_norm": 1.0990503953512065, "learning_rate": 9.47814165955258e-06, "loss": 0.6048, "step": 1210 }, { "epoch": 1.1622659625540086, "grad_norm": 0.886698122519042, "learning_rate": 9.47665058154631e-06, "loss": 0.58, "step": 1211 }, { "epoch": 1.1632261161785886, "grad_norm": 0.5767286919204323, "learning_rate": 9.475157494019445e-06, "loss": 0.47, "step": 1212 }, { "epoch": 1.1641862698031684, "grad_norm": 0.8148504895223551, "learning_rate": 9.473662397642219e-06, "loss": 0.6024, "step": 1213 }, { "epoch": 1.1651464234277484, "grad_norm": 0.7381575031851613, "learning_rate": 9.472165293085765e-06, "loss": 0.6445, "step": 1214 }, { "epoch": 1.1661065770523285, "grad_norm": 1.0634539502136668, "learning_rate": 9.470666181022114e-06, "loss": 0.6353, "step": 1215 }, { "epoch": 1.1670667306769082, "grad_norm": 0.7859877564059962, "learning_rate": 9.469165062124203e-06, "loss": 0.603, "step": 1216 }, { "epoch": 1.1680268843014883, "grad_norm": 1.179710079977145, "learning_rate": 9.467661937065869e-06, "loss": 0.6596, "step": 1217 }, { "epoch": 1.168987037926068, "grad_norm": 0.7782067848151587, "learning_rate": 9.466156806521847e-06, "loss": 0.6734, "step": 1218 }, { "epoch": 1.169947191550648, "grad_norm": 1.021594102930777, "learning_rate": 9.464649671167777e-06, "loss": 0.5854, "step": 1219 }, { "epoch": 1.170907345175228, "grad_norm": 0.993552950135849, "learning_rate": 9.463140531680193e-06, "loss": 0.5452, "step": 1220 }, { "epoch": 1.171867498799808, "grad_norm": 0.9772571294154769, "learning_rate": 9.461629388736534e-06, "loss": 0.8187, "step": 1221 }, { "epoch": 1.172827652424388, "grad_norm": 1.3208010613629735, "learning_rate": 9.460116243015131e-06, "loss": 0.6414, "step": 1222 }, { "epoch": 1.173787806048968, "grad_norm": 0.8768539222329068, "learning_rate": 9.458601095195225e-06, "loss": 0.4989, "step": 1223 }, { "epoch": 1.1747479596735477, "grad_norm": 1.0427846388801434, "learning_rate": 9.45708394595695e-06, "loss": 0.6273, "step": 1224 }, { "epoch": 1.1757081132981277, "grad_norm": 1.3149440147802898, "learning_rate": 9.455564795981331e-06, "loss": 0.6789, "step": 1225 }, { "epoch": 1.1766682669227075, "grad_norm": 1.0521780068213056, "learning_rate": 9.454043645950303e-06, "loss": 0.6583, "step": 1226 }, { "epoch": 1.1776284205472876, "grad_norm": 0.7898395729921143, "learning_rate": 9.452520496546693e-06, "loss": 0.6751, "step": 1227 }, { "epoch": 1.1785885741718676, "grad_norm": 0.862160500247702, "learning_rate": 9.450995348454228e-06, "loss": 0.5618, "step": 1228 }, { "epoch": 1.1795487277964474, "grad_norm": 1.0000955512065735, "learning_rate": 9.44946820235753e-06, "loss": 0.6087, "step": 1229 }, { "epoch": 1.1805088814210274, "grad_norm": 0.8562482354619966, "learning_rate": 9.447939058942119e-06, "loss": 0.5393, "step": 1230 }, { "epoch": 1.1814690350456072, "grad_norm": 1.023607374434533, "learning_rate": 9.446407918894408e-06, "loss": 0.629, "step": 1231 }, { "epoch": 1.1824291886701872, "grad_norm": 1.1995953948499138, "learning_rate": 9.444874782901713e-06, "loss": 0.6569, "step": 1232 }, { "epoch": 1.1833893422947672, "grad_norm": 0.9253134403087281, "learning_rate": 9.443339651652241e-06, "loss": 0.5805, "step": 1233 }, { "epoch": 1.184349495919347, "grad_norm": 0.9815597546656902, "learning_rate": 9.441802525835095e-06, "loss": 0.7282, "step": 1234 }, { "epoch": 1.185309649543927, "grad_norm": 1.0297119661202323, "learning_rate": 9.440263406140276e-06, "loss": 0.5816, "step": 1235 }, { "epoch": 1.186269803168507, "grad_norm": 0.9010807701981315, "learning_rate": 9.438722293258678e-06, "loss": 0.5699, "step": 1236 }, { "epoch": 1.1872299567930868, "grad_norm": 0.9580992874377011, "learning_rate": 9.43717918788209e-06, "loss": 0.6507, "step": 1237 }, { "epoch": 1.1881901104176669, "grad_norm": 0.8133919581297315, "learning_rate": 9.435634090703195e-06, "loss": 0.7249, "step": 1238 }, { "epoch": 1.1891502640422467, "grad_norm": 1.5118913352009353, "learning_rate": 9.434087002415571e-06, "loss": 0.6661, "step": 1239 }, { "epoch": 1.1901104176668267, "grad_norm": 1.0434414907372425, "learning_rate": 9.432537923713689e-06, "loss": 0.6114, "step": 1240 }, { "epoch": 1.1910705712914067, "grad_norm": 1.0911432701382817, "learning_rate": 9.430986855292916e-06, "loss": 0.6395, "step": 1241 }, { "epoch": 1.1920307249159865, "grad_norm": 0.965543518659571, "learning_rate": 9.429433797849508e-06, "loss": 0.6735, "step": 1242 }, { "epoch": 1.1929908785405665, "grad_norm": 0.6645288158765468, "learning_rate": 9.427878752080613e-06, "loss": 0.5808, "step": 1243 }, { "epoch": 1.1939510321651463, "grad_norm": 0.9778088547759095, "learning_rate": 9.426321718684282e-06, "loss": 0.5994, "step": 1244 }, { "epoch": 1.1949111857897263, "grad_norm": 0.9561136702953977, "learning_rate": 9.424762698359442e-06, "loss": 0.7142, "step": 1245 }, { "epoch": 1.1958713394143063, "grad_norm": 0.8889897999501921, "learning_rate": 9.423201691805927e-06, "loss": 0.5206, "step": 1246 }, { "epoch": 1.1968314930388861, "grad_norm": 0.7010747008302045, "learning_rate": 9.421638699724453e-06, "loss": 0.5299, "step": 1247 }, { "epoch": 1.1977916466634662, "grad_norm": 0.717207721063206, "learning_rate": 9.420073722816634e-06, "loss": 0.6244, "step": 1248 }, { "epoch": 1.1987518002880462, "grad_norm": 1.183723478316145, "learning_rate": 9.418506761784967e-06, "loss": 0.6246, "step": 1249 }, { "epoch": 1.199711953912626, "grad_norm": 1.0019663745871579, "learning_rate": 9.416937817332846e-06, "loss": 0.6901, "step": 1250 }, { "epoch": 1.200672107537206, "grad_norm": 1.0878769284811893, "learning_rate": 9.415366890164553e-06, "loss": 0.5843, "step": 1251 }, { "epoch": 1.201632261161786, "grad_norm": 0.9575501465840736, "learning_rate": 9.41379398098526e-06, "loss": 0.5382, "step": 1252 }, { "epoch": 1.2025924147863658, "grad_norm": 1.0959029184043156, "learning_rate": 9.41221909050103e-06, "loss": 0.639, "step": 1253 }, { "epoch": 1.2035525684109458, "grad_norm": 0.7314750135431832, "learning_rate": 9.410642219418816e-06, "loss": 0.488, "step": 1254 }, { "epoch": 1.2045127220355256, "grad_norm": 0.8880812200529163, "learning_rate": 9.409063368446454e-06, "loss": 0.6768, "step": 1255 }, { "epoch": 1.2054728756601056, "grad_norm": 0.9563856672257456, "learning_rate": 9.407482538292679e-06, "loss": 0.6338, "step": 1256 }, { "epoch": 1.2064330292846854, "grad_norm": 1.2580779734185406, "learning_rate": 9.405899729667105e-06, "loss": 0.7655, "step": 1257 }, { "epoch": 1.2073931829092654, "grad_norm": 1.4122245193335805, "learning_rate": 9.404314943280238e-06, "loss": 0.6293, "step": 1258 }, { "epoch": 1.2083533365338455, "grad_norm": 1.004456655960007, "learning_rate": 9.402728179843472e-06, "loss": 0.7194, "step": 1259 }, { "epoch": 1.2093134901584253, "grad_norm": 0.9718377769196084, "learning_rate": 9.40113944006909e-06, "loss": 0.6075, "step": 1260 }, { "epoch": 1.2102736437830053, "grad_norm": 0.8056410695854963, "learning_rate": 9.39954872467026e-06, "loss": 0.6727, "step": 1261 }, { "epoch": 1.2112337974075853, "grad_norm": 1.1395241081987777, "learning_rate": 9.397956034361033e-06, "loss": 0.6235, "step": 1262 }, { "epoch": 1.212193951032165, "grad_norm": 0.9099166044524822, "learning_rate": 9.396361369856356e-06, "loss": 0.5521, "step": 1263 }, { "epoch": 1.2131541046567451, "grad_norm": 0.9482670851732254, "learning_rate": 9.394764731872054e-06, "loss": 0.752, "step": 1264 }, { "epoch": 1.2141142582813251, "grad_norm": 1.097472724767249, "learning_rate": 9.393166121124841e-06, "loss": 0.609, "step": 1265 }, { "epoch": 1.215074411905905, "grad_norm": 0.7035412553629894, "learning_rate": 9.391565538332317e-06, "loss": 0.5578, "step": 1266 }, { "epoch": 1.216034565530485, "grad_norm": 0.8075501145482592, "learning_rate": 9.389962984212966e-06, "loss": 0.5988, "step": 1267 }, { "epoch": 1.2169947191550647, "grad_norm": 0.8480717941013729, "learning_rate": 9.388358459486156e-06, "loss": 0.5897, "step": 1268 }, { "epoch": 1.2179548727796448, "grad_norm": 0.7633861626624882, "learning_rate": 9.386751964872144e-06, "loss": 0.62, "step": 1269 }, { "epoch": 1.2189150264042246, "grad_norm": 1.2097663722724836, "learning_rate": 9.385143501092066e-06, "loss": 0.6743, "step": 1270 }, { "epoch": 1.2198751800288046, "grad_norm": 0.8002869326602273, "learning_rate": 9.383533068867947e-06, "loss": 0.5981, "step": 1271 }, { "epoch": 1.2208353336533846, "grad_norm": 0.7781618707043757, "learning_rate": 9.381920668922688e-06, "loss": 0.6106, "step": 1272 }, { "epoch": 1.2217954872779644, "grad_norm": 0.7374096944679192, "learning_rate": 9.380306301980082e-06, "loss": 0.5807, "step": 1273 }, { "epoch": 1.2227556409025444, "grad_norm": 0.9590685749532264, "learning_rate": 9.378689968764801e-06, "loss": 0.8072, "step": 1274 }, { "epoch": 1.2237157945271244, "grad_norm": 1.4283230099409416, "learning_rate": 9.377071670002398e-06, "loss": 0.8246, "step": 1275 }, { "epoch": 1.2246759481517042, "grad_norm": 1.2480777873955116, "learning_rate": 9.37545140641931e-06, "loss": 0.5812, "step": 1276 }, { "epoch": 1.2256361017762842, "grad_norm": 0.8972216937851393, "learning_rate": 9.373829178742857e-06, "loss": 0.6656, "step": 1277 }, { "epoch": 1.2265962554008643, "grad_norm": 0.912255807064162, "learning_rate": 9.37220498770124e-06, "loss": 0.5277, "step": 1278 }, { "epoch": 1.227556409025444, "grad_norm": 0.8256377004156846, "learning_rate": 9.370578834023539e-06, "loss": 0.5971, "step": 1279 }, { "epoch": 1.228516562650024, "grad_norm": 0.8436643631601126, "learning_rate": 9.368950718439718e-06, "loss": 0.6026, "step": 1280 }, { "epoch": 1.2294767162746039, "grad_norm": 1.1585366508182342, "learning_rate": 9.367320641680621e-06, "loss": 0.8957, "step": 1281 }, { "epoch": 1.2304368698991839, "grad_norm": 1.0034979036065206, "learning_rate": 9.365688604477974e-06, "loss": 0.7027, "step": 1282 }, { "epoch": 1.2313970235237637, "grad_norm": 1.0547819770205946, "learning_rate": 9.364054607564376e-06, "loss": 0.6757, "step": 1283 }, { "epoch": 1.2323571771483437, "grad_norm": 0.7935215714266847, "learning_rate": 9.362418651673317e-06, "loss": 0.5696, "step": 1284 }, { "epoch": 1.2333173307729237, "grad_norm": 0.8534906068903421, "learning_rate": 9.360780737539155e-06, "loss": 0.5885, "step": 1285 }, { "epoch": 1.2342774843975035, "grad_norm": 1.0773588382542443, "learning_rate": 9.359140865897136e-06, "loss": 0.6675, "step": 1286 }, { "epoch": 1.2352376380220835, "grad_norm": 0.6873721213285777, "learning_rate": 9.357499037483377e-06, "loss": 0.6097, "step": 1287 }, { "epoch": 1.2361977916466635, "grad_norm": 1.0694026391613127, "learning_rate": 9.35585525303488e-06, "loss": 0.5995, "step": 1288 }, { "epoch": 1.2371579452712433, "grad_norm": 0.8025251059334171, "learning_rate": 9.35420951328952e-06, "loss": 0.7176, "step": 1289 }, { "epoch": 1.2381180988958234, "grad_norm": 0.8317084825627649, "learning_rate": 9.352561818986056e-06, "loss": 0.7253, "step": 1290 }, { "epoch": 1.2390782525204034, "grad_norm": 1.0439249377910602, "learning_rate": 9.350912170864116e-06, "loss": 0.5944, "step": 1291 }, { "epoch": 1.2400384061449832, "grad_norm": 1.0000945886003398, "learning_rate": 9.349260569664212e-06, "loss": 0.5165, "step": 1292 }, { "epoch": 1.2409985597695632, "grad_norm": 0.7914438922296891, "learning_rate": 9.347607016127728e-06, "loss": 0.5874, "step": 1293 }, { "epoch": 1.241958713394143, "grad_norm": 0.9061501471232559, "learning_rate": 9.34595151099693e-06, "loss": 0.5848, "step": 1294 }, { "epoch": 1.242918867018723, "grad_norm": 0.8975053965738891, "learning_rate": 9.344294055014952e-06, "loss": 0.5942, "step": 1295 }, { "epoch": 1.243879020643303, "grad_norm": 0.8033397963647192, "learning_rate": 9.342634648925813e-06, "loss": 0.6939, "step": 1296 }, { "epoch": 1.2448391742678828, "grad_norm": 0.982875046351557, "learning_rate": 9.340973293474397e-06, "loss": 0.7445, "step": 1297 }, { "epoch": 1.2457993278924628, "grad_norm": 0.9895562388888833, "learning_rate": 9.339309989406474e-06, "loss": 0.7268, "step": 1298 }, { "epoch": 1.2467594815170426, "grad_norm": 1.103522095099442, "learning_rate": 9.337644737468682e-06, "loss": 0.5586, "step": 1299 }, { "epoch": 1.2477196351416227, "grad_norm": 1.0749127308634423, "learning_rate": 9.335977538408533e-06, "loss": 0.648, "step": 1300 }, { "epoch": 1.2486797887662027, "grad_norm": 0.8788011392673045, "learning_rate": 9.334308392974417e-06, "loss": 0.5287, "step": 1301 }, { "epoch": 1.2496399423907825, "grad_norm": 0.8302607058467603, "learning_rate": 9.332637301915594e-06, "loss": 0.6509, "step": 1302 }, { "epoch": 1.2506000960153625, "grad_norm": 1.218745246890944, "learning_rate": 9.3309642659822e-06, "loss": 0.5637, "step": 1303 }, { "epoch": 1.2515602496399425, "grad_norm": 0.9203241271347962, "learning_rate": 9.329289285925242e-06, "loss": 0.5608, "step": 1304 }, { "epoch": 1.2525204032645223, "grad_norm": 1.0666198188539213, "learning_rate": 9.327612362496601e-06, "loss": 0.6861, "step": 1305 }, { "epoch": 1.2534805568891023, "grad_norm": 1.0450663677083945, "learning_rate": 9.325933496449031e-06, "loss": 0.6448, "step": 1306 }, { "epoch": 1.2544407105136821, "grad_norm": 0.7941700859032684, "learning_rate": 9.324252688536156e-06, "loss": 0.6432, "step": 1307 }, { "epoch": 1.2554008641382621, "grad_norm": 0.989893050356933, "learning_rate": 9.322569939512472e-06, "loss": 0.5319, "step": 1308 }, { "epoch": 1.256361017762842, "grad_norm": 0.9051503548624601, "learning_rate": 9.320885250133347e-06, "loss": 0.6317, "step": 1309 }, { "epoch": 1.257321171387422, "grad_norm": 0.9473337161170687, "learning_rate": 9.319198621155022e-06, "loss": 0.7312, "step": 1310 }, { "epoch": 1.258281325012002, "grad_norm": 0.9674278845965549, "learning_rate": 9.317510053334604e-06, "loss": 0.7436, "step": 1311 }, { "epoch": 1.2592414786365818, "grad_norm": 0.9637315099687463, "learning_rate": 9.315819547430077e-06, "loss": 0.7631, "step": 1312 }, { "epoch": 1.2602016322611618, "grad_norm": 1.202596062371259, "learning_rate": 9.314127104200287e-06, "loss": 0.6169, "step": 1313 }, { "epoch": 1.2611617858857418, "grad_norm": 0.7788196680348278, "learning_rate": 9.312432724404957e-06, "loss": 0.5915, "step": 1314 }, { "epoch": 1.2621219395103216, "grad_norm": 1.0725579729351014, "learning_rate": 9.310736408804674e-06, "loss": 0.4886, "step": 1315 }, { "epoch": 1.2630820931349016, "grad_norm": 0.6646188347402221, "learning_rate": 9.309038158160896e-06, "loss": 0.5885, "step": 1316 }, { "epoch": 1.2640422467594816, "grad_norm": 0.9067786842412199, "learning_rate": 9.30733797323595e-06, "loss": 0.597, "step": 1317 }, { "epoch": 1.2650024003840614, "grad_norm": 0.9647091937794006, "learning_rate": 9.305635854793031e-06, "loss": 0.5991, "step": 1318 }, { "epoch": 1.2659625540086414, "grad_norm": 0.902764198074658, "learning_rate": 9.3039318035962e-06, "loss": 0.6497, "step": 1319 }, { "epoch": 1.2669227076332212, "grad_norm": 0.9243233127074196, "learning_rate": 9.30222582041039e-06, "loss": 0.7103, "step": 1320 }, { "epoch": 1.2678828612578013, "grad_norm": 1.0980648026135322, "learning_rate": 9.300517906001399e-06, "loss": 0.7933, "step": 1321 }, { "epoch": 1.268843014882381, "grad_norm": 1.2911252663594879, "learning_rate": 9.298808061135887e-06, "loss": 0.687, "step": 1322 }, { "epoch": 1.269803168506961, "grad_norm": 0.9824169811040547, "learning_rate": 9.29709628658139e-06, "loss": 0.6428, "step": 1323 }, { "epoch": 1.270763322131541, "grad_norm": 0.7849149978457542, "learning_rate": 9.295382583106303e-06, "loss": 0.6252, "step": 1324 }, { "epoch": 1.2717234757561209, "grad_norm": 1.2014232449565303, "learning_rate": 9.29366695147989e-06, "loss": 0.6014, "step": 1325 }, { "epoch": 1.272683629380701, "grad_norm": 1.0467353960404917, "learning_rate": 9.291949392472276e-06, "loss": 0.5731, "step": 1326 }, { "epoch": 1.273643783005281, "grad_norm": 0.9714412257880511, "learning_rate": 9.290229906854462e-06, "loss": 0.6064, "step": 1327 }, { "epoch": 1.2746039366298607, "grad_norm": 0.9084648933376147, "learning_rate": 9.288508495398301e-06, "loss": 0.6734, "step": 1328 }, { "epoch": 1.2755640902544407, "grad_norm": 0.969929198466789, "learning_rate": 9.28678515887652e-06, "loss": 0.6139, "step": 1329 }, { "epoch": 1.2765242438790207, "grad_norm": 0.763695492388788, "learning_rate": 9.285059898062701e-06, "loss": 0.5335, "step": 1330 }, { "epoch": 1.2774843975036005, "grad_norm": 0.9948681022515916, "learning_rate": 9.2833327137313e-06, "loss": 0.7919, "step": 1331 }, { "epoch": 1.2784445511281806, "grad_norm": 1.2527592242429801, "learning_rate": 9.281603606657632e-06, "loss": 0.7141, "step": 1332 }, { "epoch": 1.2794047047527606, "grad_norm": 1.0421136278331202, "learning_rate": 9.27987257761787e-06, "loss": 0.7451, "step": 1333 }, { "epoch": 1.2803648583773404, "grad_norm": 1.0212822449770629, "learning_rate": 9.27813962738906e-06, "loss": 0.7021, "step": 1334 }, { "epoch": 1.2813250120019202, "grad_norm": 1.2167945128001032, "learning_rate": 9.2764047567491e-06, "loss": 0.7802, "step": 1335 }, { "epoch": 1.2822851656265002, "grad_norm": 1.0001114229750037, "learning_rate": 9.274667966476757e-06, "loss": 0.6086, "step": 1336 }, { "epoch": 1.2832453192510802, "grad_norm": 1.3098352661842951, "learning_rate": 9.272929257351659e-06, "loss": 0.6298, "step": 1337 }, { "epoch": 1.28420547287566, "grad_norm": 1.026505705463564, "learning_rate": 9.27118863015429e-06, "loss": 0.6164, "step": 1338 }, { "epoch": 1.28516562650024, "grad_norm": 0.7642904094825214, "learning_rate": 9.269446085666002e-06, "loss": 0.6974, "step": 1339 }, { "epoch": 1.28612578012482, "grad_norm": 1.3421493457675686, "learning_rate": 9.267701624669006e-06, "loss": 0.7769, "step": 1340 }, { "epoch": 1.2870859337493998, "grad_norm": 1.2318256709169872, "learning_rate": 9.265955247946369e-06, "loss": 0.718, "step": 1341 }, { "epoch": 1.2880460873739799, "grad_norm": 0.5972951208606349, "learning_rate": 9.26420695628202e-06, "loss": 0.6342, "step": 1342 }, { "epoch": 1.2890062409985599, "grad_norm": 0.9414743543057292, "learning_rate": 9.262456750460754e-06, "loss": 0.6592, "step": 1343 }, { "epoch": 1.2899663946231397, "grad_norm": 0.8249795298129238, "learning_rate": 9.260704631268211e-06, "loss": 0.5613, "step": 1344 }, { "epoch": 1.2909265482477197, "grad_norm": 0.7180048402250736, "learning_rate": 9.258950599490906e-06, "loss": 0.6387, "step": 1345 }, { "epoch": 1.2918867018722997, "grad_norm": 0.8257585381014539, "learning_rate": 9.257194655916202e-06, "loss": 0.6699, "step": 1346 }, { "epoch": 1.2928468554968795, "grad_norm": 0.9671511028236428, "learning_rate": 9.255436801332324e-06, "loss": 0.6292, "step": 1347 }, { "epoch": 1.2938070091214593, "grad_norm": 1.1123065869046798, "learning_rate": 9.253677036528356e-06, "loss": 0.7272, "step": 1348 }, { "epoch": 1.2947671627460393, "grad_norm": 0.7336799640636064, "learning_rate": 9.251915362294233e-06, "loss": 0.5327, "step": 1349 }, { "epoch": 1.2957273163706193, "grad_norm": 0.7837908722296932, "learning_rate": 9.250151779420756e-06, "loss": 0.5286, "step": 1350 }, { "epoch": 1.2966874699951991, "grad_norm": 0.8078677018592116, "learning_rate": 9.248386288699575e-06, "loss": 0.5162, "step": 1351 }, { "epoch": 1.2976476236197791, "grad_norm": 0.744158916593893, "learning_rate": 9.246618890923203e-06, "loss": 0.5791, "step": 1352 }, { "epoch": 1.2986077772443592, "grad_norm": 0.8130007308759658, "learning_rate": 9.244849586885005e-06, "loss": 0.6414, "step": 1353 }, { "epoch": 1.299567930868939, "grad_norm": 0.9702810061771191, "learning_rate": 9.243078377379202e-06, "loss": 0.6326, "step": 1354 }, { "epoch": 1.300528084493519, "grad_norm": 0.9809300048216328, "learning_rate": 9.24130526320087e-06, "loss": 0.8607, "step": 1355 }, { "epoch": 1.301488238118099, "grad_norm": 1.0626302732269803, "learning_rate": 9.239530245145945e-06, "loss": 0.5625, "step": 1356 }, { "epoch": 1.3024483917426788, "grad_norm": 0.9231518523794499, "learning_rate": 9.237753324011208e-06, "loss": 0.6242, "step": 1357 }, { "epoch": 1.3034085453672588, "grad_norm": 0.9958648270298048, "learning_rate": 9.235974500594305e-06, "loss": 0.5809, "step": 1358 }, { "epoch": 1.3043686989918388, "grad_norm": 0.8665512289046211, "learning_rate": 9.234193775693728e-06, "loss": 0.5805, "step": 1359 }, { "epoch": 1.3053288526164186, "grad_norm": 1.0722328708613724, "learning_rate": 9.232411150108826e-06, "loss": 0.7264, "step": 1360 }, { "epoch": 1.3062890062409986, "grad_norm": 1.121754628980953, "learning_rate": 9.2306266246398e-06, "loss": 0.5769, "step": 1361 }, { "epoch": 1.3072491598655784, "grad_norm": 0.9907449030552085, "learning_rate": 9.228840200087708e-06, "loss": 0.6189, "step": 1362 }, { "epoch": 1.3082093134901585, "grad_norm": 0.8782081356851079, "learning_rate": 9.227051877254452e-06, "loss": 0.6949, "step": 1363 }, { "epoch": 1.3091694671147383, "grad_norm": 0.9773507759719392, "learning_rate": 9.225261656942792e-06, "loss": 0.5597, "step": 1364 }, { "epoch": 1.3101296207393183, "grad_norm": 1.0167471904702028, "learning_rate": 9.22346953995634e-06, "loss": 0.7175, "step": 1365 }, { "epoch": 1.3110897743638983, "grad_norm": 1.000698610555494, "learning_rate": 9.22167552709956e-06, "loss": 0.5981, "step": 1366 }, { "epoch": 1.312049927988478, "grad_norm": 0.8273446055707737, "learning_rate": 9.219879619177761e-06, "loss": 0.6155, "step": 1367 }, { "epoch": 1.313010081613058, "grad_norm": 0.7277499551879643, "learning_rate": 9.21808181699711e-06, "loss": 0.5128, "step": 1368 }, { "epoch": 1.3139702352376381, "grad_norm": 0.9728612854277003, "learning_rate": 9.21628212136462e-06, "loss": 0.6877, "step": 1369 }, { "epoch": 1.314930388862218, "grad_norm": 1.1236532937666404, "learning_rate": 9.214480533088154e-06, "loss": 0.7108, "step": 1370 }, { "epoch": 1.315890542486798, "grad_norm": 1.1263052348503229, "learning_rate": 9.212677052976428e-06, "loss": 0.6913, "step": 1371 }, { "epoch": 1.316850696111378, "grad_norm": 1.063228228941723, "learning_rate": 9.210871681839006e-06, "loss": 0.7189, "step": 1372 }, { "epoch": 1.3178108497359577, "grad_norm": 1.0576631942254184, "learning_rate": 9.209064420486296e-06, "loss": 0.5953, "step": 1373 }, { "epoch": 1.3187710033605378, "grad_norm": 1.0158290900319558, "learning_rate": 9.207255269729559e-06, "loss": 0.792, "step": 1374 }, { "epoch": 1.3197311569851176, "grad_norm": 1.1635253849512268, "learning_rate": 9.205444230380904e-06, "loss": 0.6794, "step": 1375 }, { "epoch": 1.3206913106096976, "grad_norm": 1.058758555030243, "learning_rate": 9.203631303253292e-06, "loss": 0.6537, "step": 1376 }, { "epoch": 1.3216514642342774, "grad_norm": 1.27060484146149, "learning_rate": 9.201816489160518e-06, "loss": 0.6021, "step": 1377 }, { "epoch": 1.3226116178588574, "grad_norm": 0.9353327169531234, "learning_rate": 9.199999788917238e-06, "loss": 0.732, "step": 1378 }, { "epoch": 1.3235717714834374, "grad_norm": 1.0638428010191134, "learning_rate": 9.198181203338949e-06, "loss": 0.6378, "step": 1379 }, { "epoch": 1.3245319251080172, "grad_norm": 0.7530957760983812, "learning_rate": 9.196360733241992e-06, "loss": 0.5532, "step": 1380 }, { "epoch": 1.3254920787325972, "grad_norm": 0.7643480629680425, "learning_rate": 9.194538379443561e-06, "loss": 0.5982, "step": 1381 }, { "epoch": 1.3264522323571772, "grad_norm": 1.2382800658767443, "learning_rate": 9.192714142761687e-06, "loss": 0.6872, "step": 1382 }, { "epoch": 1.327412385981757, "grad_norm": 0.9482773959208286, "learning_rate": 9.190888024015252e-06, "loss": 0.6267, "step": 1383 }, { "epoch": 1.328372539606337, "grad_norm": 0.8059482314626758, "learning_rate": 9.189060024023981e-06, "loss": 0.6299, "step": 1384 }, { "epoch": 1.329332693230917, "grad_norm": 0.7575566203082403, "learning_rate": 9.187230143608445e-06, "loss": 0.5727, "step": 1385 }, { "epoch": 1.3302928468554969, "grad_norm": 0.7505305360529138, "learning_rate": 9.185398383590056e-06, "loss": 0.5105, "step": 1386 }, { "epoch": 1.3312530004800769, "grad_norm": 1.0936453368809955, "learning_rate": 9.18356474479107e-06, "loss": 0.6544, "step": 1387 }, { "epoch": 1.3322131541046567, "grad_norm": 0.7727678698322751, "learning_rate": 9.181729228034593e-06, "loss": 0.6575, "step": 1388 }, { "epoch": 1.3331733077292367, "grad_norm": 0.888295644793819, "learning_rate": 9.179891834144565e-06, "loss": 0.5823, "step": 1389 }, { "epoch": 1.3341334613538165, "grad_norm": 1.30932038903827, "learning_rate": 9.178052563945773e-06, "loss": 0.7406, "step": 1390 }, { "epoch": 1.3350936149783965, "grad_norm": 1.1887921836450326, "learning_rate": 9.176211418263845e-06, "loss": 0.7399, "step": 1391 }, { "epoch": 1.3360537686029765, "grad_norm": 1.2295700189055145, "learning_rate": 9.174368397925254e-06, "loss": 0.7603, "step": 1392 }, { "epoch": 1.3370139222275563, "grad_norm": 1.207188268979778, "learning_rate": 9.17252350375731e-06, "loss": 0.7259, "step": 1393 }, { "epoch": 1.3379740758521363, "grad_norm": 1.1977278393110833, "learning_rate": 9.170676736588167e-06, "loss": 0.6249, "step": 1394 }, { "epoch": 1.3389342294767164, "grad_norm": 1.0441196854472323, "learning_rate": 9.168828097246819e-06, "loss": 0.5615, "step": 1395 }, { "epoch": 1.3398943831012962, "grad_norm": 0.813353438300825, "learning_rate": 9.1669775865631e-06, "loss": 0.609, "step": 1396 }, { "epoch": 1.3408545367258762, "grad_norm": 0.8817076304272241, "learning_rate": 9.165125205367685e-06, "loss": 0.6455, "step": 1397 }, { "epoch": 1.3418146903504562, "grad_norm": 1.1068259507552038, "learning_rate": 9.16327095449209e-06, "loss": 0.737, "step": 1398 }, { "epoch": 1.342774843975036, "grad_norm": 0.9579148367178544, "learning_rate": 9.161414834768662e-06, "loss": 0.7518, "step": 1399 }, { "epoch": 1.343734997599616, "grad_norm": 0.8983809547493272, "learning_rate": 9.159556847030602e-06, "loss": 0.5944, "step": 1400 }, { "epoch": 1.3446951512241958, "grad_norm": 1.2312405806674218, "learning_rate": 9.157696992111935e-06, "loss": 0.7589, "step": 1401 }, { "epoch": 1.3456553048487758, "grad_norm": 1.1660763998360677, "learning_rate": 9.15583527084753e-06, "loss": 0.5737, "step": 1402 }, { "epoch": 1.3466154584733556, "grad_norm": 0.7501036109095863, "learning_rate": 9.153971684073098e-06, "loss": 0.6108, "step": 1403 }, { "epoch": 1.3475756120979356, "grad_norm": 0.803222024216549, "learning_rate": 9.15210623262518e-06, "loss": 0.6926, "step": 1404 }, { "epoch": 1.3485357657225157, "grad_norm": 0.7279829090141787, "learning_rate": 9.150238917341156e-06, "loss": 0.5683, "step": 1405 }, { "epoch": 1.3494959193470955, "grad_norm": 0.7380807634356421, "learning_rate": 9.148369739059247e-06, "loss": 0.6192, "step": 1406 }, { "epoch": 1.3504560729716755, "grad_norm": 0.9402408606290709, "learning_rate": 9.146498698618507e-06, "loss": 0.5391, "step": 1407 }, { "epoch": 1.3514162265962555, "grad_norm": 0.678828293880698, "learning_rate": 9.14462579685882e-06, "loss": 0.6524, "step": 1408 }, { "epoch": 1.3523763802208353, "grad_norm": 0.8606651849257381, "learning_rate": 9.142751034620923e-06, "loss": 0.5487, "step": 1409 }, { "epoch": 1.3533365338454153, "grad_norm": 0.8031392506357385, "learning_rate": 9.140874412746365e-06, "loss": 0.6475, "step": 1410 }, { "epoch": 1.3542966874699953, "grad_norm": 0.7958580364065274, "learning_rate": 9.138995932077549e-06, "loss": 0.6099, "step": 1411 }, { "epoch": 1.3552568410945751, "grad_norm": 0.7816848881203385, "learning_rate": 9.137115593457702e-06, "loss": 0.5355, "step": 1412 }, { "epoch": 1.3562169947191551, "grad_norm": 0.8040814673338804, "learning_rate": 9.135233397730889e-06, "loss": 0.6422, "step": 1413 }, { "epoch": 1.357177148343735, "grad_norm": 1.0016786735508854, "learning_rate": 9.133349345742007e-06, "loss": 0.6653, "step": 1414 }, { "epoch": 1.358137301968315, "grad_norm": 0.9834474821637443, "learning_rate": 9.131463438336785e-06, "loss": 0.546, "step": 1415 }, { "epoch": 1.3590974555928947, "grad_norm": 0.9562821726643888, "learning_rate": 9.12957567636179e-06, "loss": 0.7264, "step": 1416 }, { "epoch": 1.3600576092174748, "grad_norm": 0.7850151447645083, "learning_rate": 9.127686060664414e-06, "loss": 0.618, "step": 1417 }, { "epoch": 1.3610177628420548, "grad_norm": 1.0157683134515758, "learning_rate": 9.125794592092891e-06, "loss": 0.6853, "step": 1418 }, { "epoch": 1.3619779164666346, "grad_norm": 0.8147582654689672, "learning_rate": 9.123901271496276e-06, "loss": 0.4517, "step": 1419 }, { "epoch": 1.3629380700912146, "grad_norm": 0.8728348627433652, "learning_rate": 9.122006099724463e-06, "loss": 0.5467, "step": 1420 }, { "epoch": 1.3638982237157946, "grad_norm": 0.830985119354945, "learning_rate": 9.120109077628175e-06, "loss": 0.6977, "step": 1421 }, { "epoch": 1.3648583773403744, "grad_norm": 1.0768851141434388, "learning_rate": 9.118210206058961e-06, "loss": 0.6508, "step": 1422 }, { "epoch": 1.3658185309649544, "grad_norm": 0.8219909761821469, "learning_rate": 9.116309485869207e-06, "loss": 0.712, "step": 1423 }, { "epoch": 1.3667786845895344, "grad_norm": 0.9343228816048347, "learning_rate": 9.114406917912129e-06, "loss": 0.5597, "step": 1424 }, { "epoch": 1.3677388382141142, "grad_norm": 1.103194060490155, "learning_rate": 9.112502503041763e-06, "loss": 0.6461, "step": 1425 }, { "epoch": 1.3686989918386943, "grad_norm": 0.8673752045166463, "learning_rate": 9.110596242112985e-06, "loss": 0.5569, "step": 1426 }, { "epoch": 1.369659145463274, "grad_norm": 0.8140942048369333, "learning_rate": 9.108688135981494e-06, "loss": 0.5932, "step": 1427 }, { "epoch": 1.370619299087854, "grad_norm": 0.868104373839757, "learning_rate": 9.106778185503816e-06, "loss": 0.6547, "step": 1428 }, { "epoch": 1.3715794527124339, "grad_norm": 1.0786799824961455, "learning_rate": 9.10486639153731e-06, "loss": 0.5088, "step": 1429 }, { "epoch": 1.3725396063370139, "grad_norm": 0.6143077094303164, "learning_rate": 9.102952754940162e-06, "loss": 0.5385, "step": 1430 }, { "epoch": 1.373499759961594, "grad_norm": 0.8738832289747133, "learning_rate": 9.101037276571378e-06, "loss": 0.6767, "step": 1431 }, { "epoch": 1.3744599135861737, "grad_norm": 0.7730336655965768, "learning_rate": 9.099119957290798e-06, "loss": 0.5741, "step": 1432 }, { "epoch": 1.3754200672107537, "grad_norm": 1.023951620923364, "learning_rate": 9.097200797959089e-06, "loss": 0.6605, "step": 1433 }, { "epoch": 1.3763802208353337, "grad_norm": 1.0391664088562176, "learning_rate": 9.095279799437737e-06, "loss": 0.7714, "step": 1434 }, { "epoch": 1.3773403744599135, "grad_norm": 1.0547322698929424, "learning_rate": 9.09335696258906e-06, "loss": 0.593, "step": 1435 }, { "epoch": 1.3783005280844935, "grad_norm": 1.142676487344128, "learning_rate": 9.091432288276196e-06, "loss": 0.6406, "step": 1436 }, { "epoch": 1.3792606817090736, "grad_norm": 0.8967762944390829, "learning_rate": 9.089505777363114e-06, "loss": 0.6722, "step": 1437 }, { "epoch": 1.3802208353336534, "grad_norm": 0.998580740634205, "learning_rate": 9.087577430714603e-06, "loss": 0.675, "step": 1438 }, { "epoch": 1.3811809889582334, "grad_norm": 0.9865036026565934, "learning_rate": 9.085647249196275e-06, "loss": 0.5955, "step": 1439 }, { "epoch": 1.3821411425828132, "grad_norm": 0.7442769122417944, "learning_rate": 9.083715233674572e-06, "loss": 0.6767, "step": 1440 }, { "epoch": 1.3831012962073932, "grad_norm": 0.9989539111862603, "learning_rate": 9.081781385016754e-06, "loss": 0.6475, "step": 1441 }, { "epoch": 1.384061449831973, "grad_norm": 0.9861589250575252, "learning_rate": 9.079845704090903e-06, "loss": 0.5465, "step": 1442 }, { "epoch": 1.385021603456553, "grad_norm": 0.7124406870156075, "learning_rate": 9.077908191765925e-06, "loss": 0.7427, "step": 1443 }, { "epoch": 1.385981757081133, "grad_norm": 1.2551832531934104, "learning_rate": 9.075968848911553e-06, "loss": 0.6059, "step": 1444 }, { "epoch": 1.3869419107057128, "grad_norm": 0.7466798550085931, "learning_rate": 9.074027676398333e-06, "loss": 0.6547, "step": 1445 }, { "epoch": 1.3879020643302928, "grad_norm": 0.9219961862275176, "learning_rate": 9.07208467509764e-06, "loss": 0.5897, "step": 1446 }, { "epoch": 1.3888622179548729, "grad_norm": 1.1209824943005342, "learning_rate": 9.07013984588166e-06, "loss": 0.7414, "step": 1447 }, { "epoch": 1.3898223715794527, "grad_norm": 0.6727368283246921, "learning_rate": 9.068193189623412e-06, "loss": 0.5725, "step": 1448 }, { "epoch": 1.3907825252040327, "grad_norm": 0.9852252391770612, "learning_rate": 9.06624470719673e-06, "loss": 0.6547, "step": 1449 }, { "epoch": 1.3917426788286127, "grad_norm": 0.8772608236465879, "learning_rate": 9.064294399476265e-06, "loss": 0.6941, "step": 1450 }, { "epoch": 1.3927028324531925, "grad_norm": 1.24117726110633, "learning_rate": 9.062342267337487e-06, "loss": 0.8008, "step": 1451 }, { "epoch": 1.3936629860777725, "grad_norm": 1.3109923864358044, "learning_rate": 9.060388311656689e-06, "loss": 0.6624, "step": 1452 }, { "epoch": 1.3946231397023523, "grad_norm": 0.6936106315884446, "learning_rate": 9.058432533310982e-06, "loss": 0.5926, "step": 1453 }, { "epoch": 1.3955832933269323, "grad_norm": 1.1316100565785907, "learning_rate": 9.056474933178293e-06, "loss": 0.672, "step": 1454 }, { "epoch": 1.3965434469515121, "grad_norm": 0.7554251908094519, "learning_rate": 9.054515512137367e-06, "loss": 0.6293, "step": 1455 }, { "epoch": 1.3975036005760921, "grad_norm": 0.7941169819975522, "learning_rate": 9.052554271067771e-06, "loss": 0.6735, "step": 1456 }, { "epoch": 1.3984637542006721, "grad_norm": 1.032834042116622, "learning_rate": 9.05059121084988e-06, "loss": 0.7033, "step": 1457 }, { "epoch": 1.399423907825252, "grad_norm": 1.172527971993584, "learning_rate": 9.048626332364891e-06, "loss": 0.5754, "step": 1458 }, { "epoch": 1.400384061449832, "grad_norm": 0.6649321466030876, "learning_rate": 9.046659636494821e-06, "loss": 0.6322, "step": 1459 }, { "epoch": 1.401344215074412, "grad_norm": 0.9243573873847559, "learning_rate": 9.044691124122496e-06, "loss": 0.5668, "step": 1460 }, { "epoch": 1.4023043686989918, "grad_norm": 1.1321743124254269, "learning_rate": 9.042720796131561e-06, "loss": 0.6248, "step": 1461 }, { "epoch": 1.4032645223235718, "grad_norm": 1.0273180078167037, "learning_rate": 9.040748653406475e-06, "loss": 0.6524, "step": 1462 }, { "epoch": 1.4042246759481518, "grad_norm": 0.6695779535151744, "learning_rate": 9.038774696832511e-06, "loss": 0.6255, "step": 1463 }, { "epoch": 1.4051848295727316, "grad_norm": 0.9300287705872182, "learning_rate": 9.036798927295757e-06, "loss": 0.5871, "step": 1464 }, { "epoch": 1.4061449831973116, "grad_norm": 1.0847235180206294, "learning_rate": 9.034821345683117e-06, "loss": 0.6343, "step": 1465 }, { "epoch": 1.4071051368218914, "grad_norm": 0.839933431134408, "learning_rate": 9.032841952882306e-06, "loss": 0.6229, "step": 1466 }, { "epoch": 1.4080652904464714, "grad_norm": 1.0314373508027352, "learning_rate": 9.030860749781848e-06, "loss": 0.5518, "step": 1467 }, { "epoch": 1.4090254440710512, "grad_norm": 0.656385911968562, "learning_rate": 9.028877737271089e-06, "loss": 0.707, "step": 1468 }, { "epoch": 1.4099855976956313, "grad_norm": 1.1596852681921057, "learning_rate": 9.02689291624018e-06, "loss": 0.5802, "step": 1469 }, { "epoch": 1.4109457513202113, "grad_norm": 0.9871978408614472, "learning_rate": 9.024906287580085e-06, "loss": 0.6344, "step": 1470 }, { "epoch": 1.411905904944791, "grad_norm": 0.8364003118035733, "learning_rate": 9.022917852182582e-06, "loss": 0.6419, "step": 1471 }, { "epoch": 1.412866058569371, "grad_norm": 0.7371427747940097, "learning_rate": 9.020927610940256e-06, "loss": 0.6119, "step": 1472 }, { "epoch": 1.413826212193951, "grad_norm": 0.6821302849171256, "learning_rate": 9.018935564746509e-06, "loss": 0.6082, "step": 1473 }, { "epoch": 1.414786365818531, "grad_norm": 1.1518958555462937, "learning_rate": 9.016941714495546e-06, "loss": 0.5702, "step": 1474 }, { "epoch": 1.415746519443111, "grad_norm": 0.6955236870832502, "learning_rate": 9.014946061082385e-06, "loss": 0.5963, "step": 1475 }, { "epoch": 1.416706673067691, "grad_norm": 0.9921066956045631, "learning_rate": 9.012948605402856e-06, "loss": 0.6455, "step": 1476 }, { "epoch": 1.4176668266922707, "grad_norm": 0.860987047239994, "learning_rate": 9.010949348353595e-06, "loss": 0.6121, "step": 1477 }, { "epoch": 1.4186269803168507, "grad_norm": 1.0915941593317662, "learning_rate": 9.008948290832046e-06, "loss": 0.6725, "step": 1478 }, { "epoch": 1.4195871339414305, "grad_norm": 0.6891343724709093, "learning_rate": 9.006945433736461e-06, "loss": 0.6214, "step": 1479 }, { "epoch": 1.4205472875660106, "grad_norm": 0.9676702014545276, "learning_rate": 9.004940777965903e-06, "loss": 0.5119, "step": 1480 }, { "epoch": 1.4215074411905904, "grad_norm": 0.9199872748223653, "learning_rate": 9.002934324420242e-06, "loss": 0.6397, "step": 1481 }, { "epoch": 1.4224675948151704, "grad_norm": 0.7438262809636571, "learning_rate": 9.000926074000149e-06, "loss": 0.587, "step": 1482 }, { "epoch": 1.4234277484397504, "grad_norm": 1.1010460072847796, "learning_rate": 8.998916027607111e-06, "loss": 0.6042, "step": 1483 }, { "epoch": 1.4243879020643302, "grad_norm": 0.8103941891420554, "learning_rate": 8.996904186143414e-06, "loss": 0.6255, "step": 1484 }, { "epoch": 1.4253480556889102, "grad_norm": 0.9624929342696545, "learning_rate": 8.994890550512152e-06, "loss": 0.7322, "step": 1485 }, { "epoch": 1.4263082093134902, "grad_norm": 0.9457497579476025, "learning_rate": 8.992875121617224e-06, "loss": 0.6424, "step": 1486 }, { "epoch": 1.42726836293807, "grad_norm": 1.0297773328597029, "learning_rate": 8.990857900363337e-06, "loss": 0.6636, "step": 1487 }, { "epoch": 1.42822851656265, "grad_norm": 1.001183716104272, "learning_rate": 8.988838887655997e-06, "loss": 0.6676, "step": 1488 }, { "epoch": 1.42918867018723, "grad_norm": 0.7864775833100192, "learning_rate": 8.98681808440152e-06, "loss": 0.5352, "step": 1489 }, { "epoch": 1.4301488238118099, "grad_norm": 0.9003046410798068, "learning_rate": 8.984795491507021e-06, "loss": 0.6813, "step": 1490 }, { "epoch": 1.4311089774363899, "grad_norm": 0.8877274333935607, "learning_rate": 8.982771109880421e-06, "loss": 0.5728, "step": 1491 }, { "epoch": 1.43206913106097, "grad_norm": 0.9995145044644966, "learning_rate": 8.980744940430443e-06, "loss": 0.5827, "step": 1492 }, { "epoch": 1.4330292846855497, "grad_norm": 0.7349056742885587, "learning_rate": 8.978716984066613e-06, "loss": 0.533, "step": 1493 }, { "epoch": 1.4339894383101295, "grad_norm": 0.9636052504334137, "learning_rate": 8.976687241699258e-06, "loss": 0.7187, "step": 1494 }, { "epoch": 1.4349495919347095, "grad_norm": 1.2609663135673943, "learning_rate": 8.97465571423951e-06, "loss": 0.6081, "step": 1495 }, { "epoch": 1.4359097455592895, "grad_norm": 1.0051747048414337, "learning_rate": 8.972622402599298e-06, "loss": 0.6185, "step": 1496 }, { "epoch": 1.4368698991838693, "grad_norm": 0.9342590027159998, "learning_rate": 8.970587307691357e-06, "loss": 0.6567, "step": 1497 }, { "epoch": 1.4378300528084493, "grad_norm": 0.7095723445059271, "learning_rate": 8.968550430429212e-06, "loss": 0.5903, "step": 1498 }, { "epoch": 1.4387902064330294, "grad_norm": 0.8844444681219761, "learning_rate": 8.966511771727205e-06, "loss": 0.6396, "step": 1499 }, { "epoch": 1.4397503600576091, "grad_norm": 0.9962469376720036, "learning_rate": 8.96447133250046e-06, "loss": 0.6723, "step": 1500 }, { "epoch": 1.4407105136821892, "grad_norm": 0.9301251412509197, "learning_rate": 8.962429113664913e-06, "loss": 0.6111, "step": 1501 }, { "epoch": 1.4416706673067692, "grad_norm": 0.9186204660267107, "learning_rate": 8.960385116137293e-06, "loss": 0.7698, "step": 1502 }, { "epoch": 1.442630820931349, "grad_norm": 0.950461158322485, "learning_rate": 8.958339340835128e-06, "loss": 0.6309, "step": 1503 }, { "epoch": 1.443590974555929, "grad_norm": 0.8878618938650225, "learning_rate": 8.956291788676746e-06, "loss": 0.663, "step": 1504 }, { "epoch": 1.444551128180509, "grad_norm": 0.8367991193090787, "learning_rate": 8.95424246058127e-06, "loss": 0.5578, "step": 1505 }, { "epoch": 1.4455112818050888, "grad_norm": 0.793909659240403, "learning_rate": 8.952191357468621e-06, "loss": 0.6611, "step": 1506 }, { "epoch": 1.4464714354296688, "grad_norm": 0.9794750714417029, "learning_rate": 8.950138480259519e-06, "loss": 0.6533, "step": 1507 }, { "epoch": 1.4474315890542486, "grad_norm": 0.6877313877935152, "learning_rate": 8.948083829875478e-06, "loss": 0.6215, "step": 1508 }, { "epoch": 1.4483917426788286, "grad_norm": 0.9219048124296251, "learning_rate": 8.946027407238809e-06, "loss": 0.7049, "step": 1509 }, { "epoch": 1.4493518963034084, "grad_norm": 0.8485337158623132, "learning_rate": 8.943969213272616e-06, "loss": 0.6658, "step": 1510 }, { "epoch": 1.4503120499279885, "grad_norm": 0.973953129587728, "learning_rate": 8.941909248900804e-06, "loss": 0.5716, "step": 1511 }, { "epoch": 1.4512722035525685, "grad_norm": 0.9383092024304013, "learning_rate": 8.939847515048066e-06, "loss": 0.7261, "step": 1512 }, { "epoch": 1.4522323571771483, "grad_norm": 1.1232696223416054, "learning_rate": 8.937784012639892e-06, "loss": 0.5115, "step": 1513 }, { "epoch": 1.4531925108017283, "grad_norm": 0.8190951902136869, "learning_rate": 8.935718742602568e-06, "loss": 0.6712, "step": 1514 }, { "epoch": 1.4541526644263083, "grad_norm": 0.8236277866157379, "learning_rate": 8.933651705863172e-06, "loss": 0.6494, "step": 1515 }, { "epoch": 1.455112818050888, "grad_norm": 1.157172493747725, "learning_rate": 8.931582903349571e-06, "loss": 0.7019, "step": 1516 }, { "epoch": 1.4560729716754681, "grad_norm": 0.8216153313496258, "learning_rate": 8.929512335990434e-06, "loss": 0.5331, "step": 1517 }, { "epoch": 1.4570331253000481, "grad_norm": 0.8858959836402756, "learning_rate": 8.92744000471521e-06, "loss": 0.6234, "step": 1518 }, { "epoch": 1.457993278924628, "grad_norm": 1.0016853122708034, "learning_rate": 8.925365910454153e-06, "loss": 0.6871, "step": 1519 }, { "epoch": 1.458953432549208, "grad_norm": 1.0861685850248648, "learning_rate": 8.923290054138298e-06, "loss": 0.6412, "step": 1520 }, { "epoch": 1.4599135861737877, "grad_norm": 0.9897418593237511, "learning_rate": 8.921212436699476e-06, "loss": 0.6486, "step": 1521 }, { "epoch": 1.4608737397983678, "grad_norm": 0.8866753949735757, "learning_rate": 8.919133059070307e-06, "loss": 0.6866, "step": 1522 }, { "epoch": 1.4618338934229476, "grad_norm": 1.0077349809039247, "learning_rate": 8.9170519221842e-06, "loss": 0.6684, "step": 1523 }, { "epoch": 1.4627940470475276, "grad_norm": 0.8893568499354665, "learning_rate": 8.914969026975354e-06, "loss": 0.6305, "step": 1524 }, { "epoch": 1.4637542006721076, "grad_norm": 0.7358342773488264, "learning_rate": 8.912884374378763e-06, "loss": 0.539, "step": 1525 }, { "epoch": 1.4647143542966874, "grad_norm": 0.7799595073101397, "learning_rate": 8.910797965330202e-06, "loss": 0.6242, "step": 1526 }, { "epoch": 1.4656745079212674, "grad_norm": 0.8730797712943911, "learning_rate": 8.908709800766237e-06, "loss": 0.6535, "step": 1527 }, { "epoch": 1.4666346615458474, "grad_norm": 0.9992116263322782, "learning_rate": 8.906619881624225e-06, "loss": 0.5091, "step": 1528 }, { "epoch": 1.4675948151704272, "grad_norm": 0.8099343780087479, "learning_rate": 8.90452820884231e-06, "loss": 0.7839, "step": 1529 }, { "epoch": 1.4685549687950072, "grad_norm": 1.0226061417681898, "learning_rate": 8.902434783359417e-06, "loss": 0.6338, "step": 1530 }, { "epoch": 1.4695151224195873, "grad_norm": 1.1887181806594445, "learning_rate": 8.900339606115264e-06, "loss": 0.647, "step": 1531 }, { "epoch": 1.470475276044167, "grad_norm": 1.1304323419040059, "learning_rate": 8.898242678050354e-06, "loss": 0.632, "step": 1532 }, { "epoch": 1.471435429668747, "grad_norm": 1.0415085203460248, "learning_rate": 8.896144000105979e-06, "loss": 0.6936, "step": 1533 }, { "epoch": 1.4723955832933269, "grad_norm": 0.7548373484936044, "learning_rate": 8.894043573224207e-06, "loss": 0.6568, "step": 1534 }, { "epoch": 1.4733557369179069, "grad_norm": 0.9700225761850124, "learning_rate": 8.8919413983479e-06, "loss": 0.6566, "step": 1535 }, { "epoch": 1.4743158905424867, "grad_norm": 1.2465398618336105, "learning_rate": 8.889837476420703e-06, "loss": 0.5246, "step": 1536 }, { "epoch": 1.4752760441670667, "grad_norm": 0.8747913619780309, "learning_rate": 8.887731808387043e-06, "loss": 0.6446, "step": 1537 }, { "epoch": 1.4762361977916467, "grad_norm": 0.8847777945413707, "learning_rate": 8.885624395192131e-06, "loss": 0.7218, "step": 1538 }, { "epoch": 1.4771963514162265, "grad_norm": 0.952948569948575, "learning_rate": 8.883515237781964e-06, "loss": 0.6685, "step": 1539 }, { "epoch": 1.4781565050408065, "grad_norm": 0.7202818532813761, "learning_rate": 8.881404337103316e-06, "loss": 0.6809, "step": 1540 }, { "epoch": 1.4791166586653866, "grad_norm": 0.9483495849825297, "learning_rate": 8.879291694103754e-06, "loss": 0.5672, "step": 1541 }, { "epoch": 1.4800768122899663, "grad_norm": 1.2786847225651765, "learning_rate": 8.877177309731618e-06, "loss": 0.7113, "step": 1542 }, { "epoch": 1.4810369659145464, "grad_norm": 0.8694645023686174, "learning_rate": 8.875061184936031e-06, "loss": 0.5664, "step": 1543 }, { "epoch": 1.4819971195391264, "grad_norm": 0.9285498074182232, "learning_rate": 8.872943320666901e-06, "loss": 0.6483, "step": 1544 }, { "epoch": 1.4829572731637062, "grad_norm": 0.9558419692040294, "learning_rate": 8.870823717874912e-06, "loss": 0.5938, "step": 1545 }, { "epoch": 1.4839174267882862, "grad_norm": 0.9697805531266147, "learning_rate": 8.868702377511535e-06, "loss": 0.5921, "step": 1546 }, { "epoch": 1.484877580412866, "grad_norm": 0.8243385046829244, "learning_rate": 8.866579300529014e-06, "loss": 0.6845, "step": 1547 }, { "epoch": 1.485837734037446, "grad_norm": 1.0109534428974758, "learning_rate": 8.864454487880377e-06, "loss": 0.6936, "step": 1548 }, { "epoch": 1.4867978876620258, "grad_norm": 1.0836991081006613, "learning_rate": 8.862327940519429e-06, "loss": 0.7155, "step": 1549 }, { "epoch": 1.4877580412866058, "grad_norm": 1.1149964850170677, "learning_rate": 8.860199659400754e-06, "loss": 0.6451, "step": 1550 }, { "epoch": 1.4887181949111858, "grad_norm": 1.1527321537730821, "learning_rate": 8.858069645479715e-06, "loss": 0.6907, "step": 1551 }, { "epoch": 1.4896783485357656, "grad_norm": 0.9629236460912923, "learning_rate": 8.855937899712454e-06, "loss": 0.6133, "step": 1552 }, { "epoch": 1.4906385021603457, "grad_norm": 0.87827258455031, "learning_rate": 8.853804423055887e-06, "loss": 0.6619, "step": 1553 }, { "epoch": 1.4915986557849257, "grad_norm": 1.0586814670985836, "learning_rate": 8.851669216467708e-06, "loss": 0.6661, "step": 1554 }, { "epoch": 1.4925588094095055, "grad_norm": 1.0600064501507696, "learning_rate": 8.849532280906391e-06, "loss": 0.5763, "step": 1555 }, { "epoch": 1.4935189630340855, "grad_norm": 0.8237840509538912, "learning_rate": 8.847393617331182e-06, "loss": 0.6255, "step": 1556 }, { "epoch": 1.4944791166586655, "grad_norm": 0.7119796250611321, "learning_rate": 8.845253226702103e-06, "loss": 0.6424, "step": 1557 }, { "epoch": 1.4954392702832453, "grad_norm": 1.0667038246349823, "learning_rate": 8.843111109979957e-06, "loss": 0.6609, "step": 1558 }, { "epoch": 1.4963994239078253, "grad_norm": 0.8511832361678332, "learning_rate": 8.840967268126313e-06, "loss": 0.5854, "step": 1559 }, { "epoch": 1.4973595775324051, "grad_norm": 0.8403139689771888, "learning_rate": 8.83882170210352e-06, "loss": 0.6174, "step": 1560 }, { "epoch": 1.4983197311569851, "grad_norm": 1.1020014833427683, "learning_rate": 8.836674412874699e-06, "loss": 0.6122, "step": 1561 }, { "epoch": 1.499279884781565, "grad_norm": 1.2263917353667657, "learning_rate": 8.834525401403747e-06, "loss": 0.6581, "step": 1562 }, { "epoch": 1.500240038406145, "grad_norm": 1.2083590026378617, "learning_rate": 8.83237466865533e-06, "loss": 0.7204, "step": 1563 }, { "epoch": 1.501200192030725, "grad_norm": 1.3714642388266567, "learning_rate": 8.83022221559489e-06, "loss": 0.6166, "step": 1564 }, { "epoch": 1.5021603456553048, "grad_norm": 0.9359610139033587, "learning_rate": 8.828068043188641e-06, "loss": 0.5484, "step": 1565 }, { "epoch": 1.5031204992798848, "grad_norm": 0.8132984828344566, "learning_rate": 8.825912152403568e-06, "loss": 0.5275, "step": 1566 }, { "epoch": 1.5040806529044648, "grad_norm": 1.00052987041177, "learning_rate": 8.823754544207426e-06, "loss": 0.6941, "step": 1567 }, { "epoch": 1.5050408065290446, "grad_norm": 1.1483681753800414, "learning_rate": 8.821595219568741e-06, "loss": 0.7489, "step": 1568 }, { "epoch": 1.5060009601536246, "grad_norm": 1.1797067342328238, "learning_rate": 8.819434179456814e-06, "loss": 0.633, "step": 1569 }, { "epoch": 1.5069611137782046, "grad_norm": 0.7706919684847922, "learning_rate": 8.817271424841711e-06, "loss": 0.5551, "step": 1570 }, { "epoch": 1.5079212674027844, "grad_norm": 1.1832944692898923, "learning_rate": 8.81510695669427e-06, "loss": 0.6549, "step": 1571 }, { "epoch": 1.5088814210273642, "grad_norm": 0.8382583677741545, "learning_rate": 8.812940775986098e-06, "loss": 0.6005, "step": 1572 }, { "epoch": 1.5098415746519445, "grad_norm": 0.4655278245395952, "learning_rate": 8.810772883689567e-06, "loss": 0.6482, "step": 1573 }, { "epoch": 1.5108017282765243, "grad_norm": 1.2552520993407983, "learning_rate": 8.808603280777826e-06, "loss": 0.6881, "step": 1574 }, { "epoch": 1.511761881901104, "grad_norm": 0.8498448464598783, "learning_rate": 8.806431968224784e-06, "loss": 0.6014, "step": 1575 }, { "epoch": 1.512722035525684, "grad_norm": 0.8687837430582558, "learning_rate": 8.80425894700512e-06, "loss": 0.6521, "step": 1576 }, { "epoch": 1.513682189150264, "grad_norm": 1.1560435596685281, "learning_rate": 8.802084218094278e-06, "loss": 0.6443, "step": 1577 }, { "epoch": 1.5146423427748439, "grad_norm": 1.0577421969393455, "learning_rate": 8.799907782468473e-06, "loss": 0.5494, "step": 1578 }, { "epoch": 1.515602496399424, "grad_norm": 1.0500034774866775, "learning_rate": 8.797729641104687e-06, "loss": 0.5953, "step": 1579 }, { "epoch": 1.516562650024004, "grad_norm": 0.8873313492152511, "learning_rate": 8.795549794980655e-06, "loss": 0.5828, "step": 1580 }, { "epoch": 1.5175228036485837, "grad_norm": 0.8221735479862696, "learning_rate": 8.793368245074896e-06, "loss": 0.7215, "step": 1581 }, { "epoch": 1.5184829572731637, "grad_norm": 0.9271481578177346, "learning_rate": 8.791184992366678e-06, "loss": 0.6023, "step": 1582 }, { "epoch": 1.5194431108977438, "grad_norm": 0.941078799241275, "learning_rate": 8.789000037836045e-06, "loss": 0.6438, "step": 1583 }, { "epoch": 1.5204032645223235, "grad_norm": 0.8157399770308056, "learning_rate": 8.786813382463796e-06, "loss": 0.6049, "step": 1584 }, { "epoch": 1.5213634181469033, "grad_norm": 1.003683606521832, "learning_rate": 8.7846250272315e-06, "loss": 0.6617, "step": 1585 }, { "epoch": 1.5223235717714836, "grad_norm": 0.7939879905546342, "learning_rate": 8.782434973121481e-06, "loss": 0.5418, "step": 1586 }, { "epoch": 1.5232837253960634, "grad_norm": 0.9368894876921604, "learning_rate": 8.780243221116838e-06, "loss": 0.586, "step": 1587 }, { "epoch": 1.5242438790206432, "grad_norm": 0.7572267609640707, "learning_rate": 8.77804977220142e-06, "loss": 0.5468, "step": 1588 }, { "epoch": 1.5252040326452232, "grad_norm": 0.7921683746335574, "learning_rate": 8.775854627359844e-06, "loss": 0.681, "step": 1589 }, { "epoch": 1.5261641862698032, "grad_norm": 0.9520389402577641, "learning_rate": 8.773657787577488e-06, "loss": 0.7448, "step": 1590 }, { "epoch": 1.527124339894383, "grad_norm": 1.28632514066174, "learning_rate": 8.771459253840489e-06, "loss": 0.7022, "step": 1591 }, { "epoch": 1.528084493518963, "grad_norm": 0.9891666260687788, "learning_rate": 8.769259027135746e-06, "loss": 0.6225, "step": 1592 }, { "epoch": 1.529044647143543, "grad_norm": 1.1205822386930178, "learning_rate": 8.767057108450919e-06, "loss": 0.7457, "step": 1593 }, { "epoch": 1.5300048007681228, "grad_norm": 0.9690295937195536, "learning_rate": 8.76485349877442e-06, "loss": 0.569, "step": 1594 }, { "epoch": 1.5309649543927029, "grad_norm": 0.8587611670728004, "learning_rate": 8.76264819909543e-06, "loss": 0.7037, "step": 1595 }, { "epoch": 1.5319251080172829, "grad_norm": 0.8796629682080109, "learning_rate": 8.760441210403886e-06, "loss": 0.6699, "step": 1596 }, { "epoch": 1.5328852616418627, "grad_norm": 1.1369280940952775, "learning_rate": 8.758232533690479e-06, "loss": 0.5926, "step": 1597 }, { "epoch": 1.5338454152664425, "grad_norm": 0.9501857096089256, "learning_rate": 8.75602216994666e-06, "loss": 0.7569, "step": 1598 }, { "epoch": 1.5348055688910227, "grad_norm": 1.166320176979183, "learning_rate": 8.753810120164639e-06, "loss": 0.6328, "step": 1599 }, { "epoch": 1.5357657225156025, "grad_norm": 1.024832407449645, "learning_rate": 8.75159638533738e-06, "loss": 0.6275, "step": 1600 }, { "epoch": 1.5367258761401823, "grad_norm": 0.982609903268487, "learning_rate": 8.749380966458608e-06, "loss": 0.5939, "step": 1601 }, { "epoch": 1.5376860297647623, "grad_norm": 0.8636483201417748, "learning_rate": 8.747163864522796e-06, "loss": 0.622, "step": 1602 }, { "epoch": 1.5386461833893423, "grad_norm": 1.0814603542881396, "learning_rate": 8.744945080525182e-06, "loss": 0.6714, "step": 1603 }, { "epoch": 1.5396063370139221, "grad_norm": 0.7122642622598232, "learning_rate": 8.742724615461753e-06, "loss": 0.553, "step": 1604 }, { "epoch": 1.5405664906385022, "grad_norm": 0.9354071014644667, "learning_rate": 8.740502470329251e-06, "loss": 0.709, "step": 1605 }, { "epoch": 1.5415266442630822, "grad_norm": 0.9327572125606929, "learning_rate": 8.738278646125174e-06, "loss": 0.6542, "step": 1606 }, { "epoch": 1.542486797887662, "grad_norm": 0.8913602778098478, "learning_rate": 8.736053143847774e-06, "loss": 0.6981, "step": 1607 }, { "epoch": 1.543446951512242, "grad_norm": 0.849464359808423, "learning_rate": 8.733825964496052e-06, "loss": 0.6084, "step": 1608 }, { "epoch": 1.544407105136822, "grad_norm": 1.0160707421384596, "learning_rate": 8.731597109069768e-06, "loss": 0.6343, "step": 1609 }, { "epoch": 1.5453672587614018, "grad_norm": 0.8615921606848755, "learning_rate": 8.72936657856943e-06, "loss": 0.5975, "step": 1610 }, { "epoch": 1.5463274123859816, "grad_norm": 1.0559334022810416, "learning_rate": 8.727134373996298e-06, "loss": 0.6067, "step": 1611 }, { "epoch": 1.5472875660105618, "grad_norm": 0.8293394675876014, "learning_rate": 8.724900496352388e-06, "loss": 0.6265, "step": 1612 }, { "epoch": 1.5482477196351416, "grad_norm": 0.907250302856252, "learning_rate": 8.722664946640463e-06, "loss": 0.6863, "step": 1613 }, { "epoch": 1.5492078732597214, "grad_norm": 0.7119434255501595, "learning_rate": 8.720427725864035e-06, "loss": 0.6131, "step": 1614 }, { "epoch": 1.5501680268843014, "grad_norm": 1.192498592506731, "learning_rate": 8.71818883502737e-06, "loss": 0.6949, "step": 1615 }, { "epoch": 1.5511281805088815, "grad_norm": 0.7366160734276367, "learning_rate": 8.715948275135482e-06, "loss": 0.661, "step": 1616 }, { "epoch": 1.5520883341334613, "grad_norm": 1.1956359283306153, "learning_rate": 8.713706047194135e-06, "loss": 0.615, "step": 1617 }, { "epoch": 1.5530484877580413, "grad_norm": 1.0402153887155807, "learning_rate": 8.711462152209843e-06, "loss": 0.6046, "step": 1618 }, { "epoch": 1.5540086413826213, "grad_norm": 0.908680302285862, "learning_rate": 8.709216591189861e-06, "loss": 0.5614, "step": 1619 }, { "epoch": 1.554968795007201, "grad_norm": 0.942533363246108, "learning_rate": 8.706969365142202e-06, "loss": 0.7296, "step": 1620 }, { "epoch": 1.555928948631781, "grad_norm": 1.1252751156625829, "learning_rate": 8.70472047507562e-06, "loss": 0.7129, "step": 1621 }, { "epoch": 1.5568891022563611, "grad_norm": 1.0777395520816588, "learning_rate": 8.702469921999617e-06, "loss": 0.6325, "step": 1622 }, { "epoch": 1.557849255880941, "grad_norm": 0.9436126251924869, "learning_rate": 8.700217706924445e-06, "loss": 0.6647, "step": 1623 }, { "epoch": 1.558809409505521, "grad_norm": 1.0056881318307205, "learning_rate": 8.697963830861095e-06, "loss": 0.7058, "step": 1624 }, { "epoch": 1.559769563130101, "grad_norm": 1.4089242805365316, "learning_rate": 8.695708294821314e-06, "loss": 0.6496, "step": 1625 }, { "epoch": 1.5607297167546808, "grad_norm": 0.9501052271968954, "learning_rate": 8.693451099817583e-06, "loss": 0.6807, "step": 1626 }, { "epoch": 1.5616898703792605, "grad_norm": 0.9663875041411656, "learning_rate": 8.691192246863133e-06, "loss": 0.5223, "step": 1627 }, { "epoch": 1.5626500240038406, "grad_norm": 0.8374439092137342, "learning_rate": 8.688931736971941e-06, "loss": 0.6421, "step": 1628 }, { "epoch": 1.5636101776284206, "grad_norm": 0.931570949708988, "learning_rate": 8.686669571158724e-06, "loss": 0.6185, "step": 1629 }, { "epoch": 1.5645703312530004, "grad_norm": 0.5984428599424221, "learning_rate": 8.684405750438944e-06, "loss": 0.6322, "step": 1630 }, { "epoch": 1.5655304848775804, "grad_norm": 1.0257688821957753, "learning_rate": 8.682140275828809e-06, "loss": 0.6945, "step": 1631 }, { "epoch": 1.5664906385021604, "grad_norm": 1.196131580644623, "learning_rate": 8.679873148345262e-06, "loss": 0.5409, "step": 1632 }, { "epoch": 1.5674507921267402, "grad_norm": 1.2811185793922626, "learning_rate": 8.677604369005996e-06, "loss": 0.7596, "step": 1633 }, { "epoch": 1.5684109457513202, "grad_norm": 1.1343741470950064, "learning_rate": 8.67533393882944e-06, "loss": 0.627, "step": 1634 }, { "epoch": 1.5693710993759002, "grad_norm": 0.6331831089553249, "learning_rate": 8.673061858834766e-06, "loss": 0.6399, "step": 1635 }, { "epoch": 1.57033125300048, "grad_norm": 0.7686690961099248, "learning_rate": 8.670788130041886e-06, "loss": 0.6088, "step": 1636 }, { "epoch": 1.57129140662506, "grad_norm": 0.7223798568201186, "learning_rate": 8.668512753471453e-06, "loss": 0.5037, "step": 1637 }, { "epoch": 1.57225156024964, "grad_norm": 0.9369718545913843, "learning_rate": 8.666235730144858e-06, "loss": 0.679, "step": 1638 }, { "epoch": 1.5732117138742199, "grad_norm": 1.0869798428341313, "learning_rate": 8.663957061084234e-06, "loss": 0.5811, "step": 1639 }, { "epoch": 1.5741718674987997, "grad_norm": 1.0058969369809379, "learning_rate": 8.661676747312453e-06, "loss": 0.6357, "step": 1640 }, { "epoch": 1.5751320211233797, "grad_norm": 0.9362585521947556, "learning_rate": 8.65939478985312e-06, "loss": 0.6004, "step": 1641 }, { "epoch": 1.5760921747479597, "grad_norm": 1.0460971975109028, "learning_rate": 8.657111189730583e-06, "loss": 0.7195, "step": 1642 }, { "epoch": 1.5770523283725395, "grad_norm": 1.1130068728535885, "learning_rate": 8.654825947969924e-06, "loss": 0.6748, "step": 1643 }, { "epoch": 1.5780124819971195, "grad_norm": 1.241656353877118, "learning_rate": 8.652539065596966e-06, "loss": 0.6644, "step": 1644 }, { "epoch": 1.5789726356216995, "grad_norm": 0.8104942319130414, "learning_rate": 8.650250543638264e-06, "loss": 0.7011, "step": 1645 }, { "epoch": 1.5799327892462793, "grad_norm": 1.0014088175082412, "learning_rate": 8.647960383121113e-06, "loss": 0.6781, "step": 1646 }, { "epoch": 1.5808929428708594, "grad_norm": 0.869064842504527, "learning_rate": 8.64566858507354e-06, "loss": 0.658, "step": 1647 }, { "epoch": 1.5818530964954394, "grad_norm": 1.005570902875047, "learning_rate": 8.643375150524308e-06, "loss": 0.6098, "step": 1648 }, { "epoch": 1.5828132501200192, "grad_norm": 0.6757201180090499, "learning_rate": 8.641080080502919e-06, "loss": 0.5933, "step": 1649 }, { "epoch": 1.5837734037445992, "grad_norm": 1.0152577987377107, "learning_rate": 8.6387833760396e-06, "loss": 0.7232, "step": 1650 }, { "epoch": 1.5847335573691792, "grad_norm": 0.8873273325494148, "learning_rate": 8.636485038165323e-06, "loss": 0.63, "step": 1651 }, { "epoch": 1.585693710993759, "grad_norm": 1.050017379280611, "learning_rate": 8.634185067911781e-06, "loss": 0.7688, "step": 1652 }, { "epoch": 1.5866538646183388, "grad_norm": 1.018558104160898, "learning_rate": 8.631883466311412e-06, "loss": 0.6475, "step": 1653 }, { "epoch": 1.5876140182429188, "grad_norm": 0.9537473637043541, "learning_rate": 8.629580234397377e-06, "loss": 0.6829, "step": 1654 }, { "epoch": 1.5885741718674988, "grad_norm": 0.9527070620051072, "learning_rate": 8.627275373203572e-06, "loss": 0.5321, "step": 1655 }, { "epoch": 1.5895343254920786, "grad_norm": 0.9114444745627942, "learning_rate": 8.624968883764626e-06, "loss": 0.7248, "step": 1656 }, { "epoch": 1.5904944791166586, "grad_norm": 1.1088219341387284, "learning_rate": 8.622660767115897e-06, "loss": 0.6017, "step": 1657 }, { "epoch": 1.5914546327412387, "grad_norm": 0.8746588567393767, "learning_rate": 8.620351024293475e-06, "loss": 0.6206, "step": 1658 }, { "epoch": 1.5924147863658185, "grad_norm": 0.8031457846255833, "learning_rate": 8.618039656334173e-06, "loss": 0.6446, "step": 1659 }, { "epoch": 1.5933749399903985, "grad_norm": 0.8467501060228921, "learning_rate": 8.615726664275547e-06, "loss": 0.5922, "step": 1660 }, { "epoch": 1.5943350936149785, "grad_norm": 0.6346717082288017, "learning_rate": 8.613412049155872e-06, "loss": 0.6365, "step": 1661 }, { "epoch": 1.5952952472395583, "grad_norm": 0.9516471520327936, "learning_rate": 8.611095812014155e-06, "loss": 0.6284, "step": 1662 }, { "epoch": 1.5962554008641383, "grad_norm": 0.717532469905148, "learning_rate": 8.608777953890125e-06, "loss": 0.6446, "step": 1663 }, { "epoch": 1.5972155544887183, "grad_norm": 0.8032062881708539, "learning_rate": 8.60645847582425e-06, "loss": 0.5609, "step": 1664 }, { "epoch": 1.5981757081132981, "grad_norm": 0.858816922098711, "learning_rate": 8.604137378857713e-06, "loss": 0.5955, "step": 1665 }, { "epoch": 1.599135861737878, "grad_norm": 0.9954793233405728, "learning_rate": 8.601814664032434e-06, "loss": 0.5619, "step": 1666 }, { "epoch": 1.6000960153624582, "grad_norm": 0.789199262691236, "learning_rate": 8.599490332391054e-06, "loss": 0.6496, "step": 1667 }, { "epoch": 1.601056168987038, "grad_norm": 1.0439145506960599, "learning_rate": 8.59716438497694e-06, "loss": 0.6176, "step": 1668 }, { "epoch": 1.6020163226116177, "grad_norm": 0.8059771387590645, "learning_rate": 8.594836822834185e-06, "loss": 0.6284, "step": 1669 }, { "epoch": 1.6029764762361978, "grad_norm": 1.1851892312436938, "learning_rate": 8.592507647007606e-06, "loss": 0.6378, "step": 1670 }, { "epoch": 1.6039366298607778, "grad_norm": 1.2095928439615917, "learning_rate": 8.590176858542748e-06, "loss": 0.64, "step": 1671 }, { "epoch": 1.6048967834853576, "grad_norm": 1.0059844290272728, "learning_rate": 8.587844458485874e-06, "loss": 0.6637, "step": 1672 }, { "epoch": 1.6058569371099376, "grad_norm": 0.786912271753416, "learning_rate": 8.585510447883975e-06, "loss": 0.6401, "step": 1673 }, { "epoch": 1.6068170907345176, "grad_norm": 1.1928987923578274, "learning_rate": 8.583174827784762e-06, "loss": 0.7124, "step": 1674 }, { "epoch": 1.6077772443590974, "grad_norm": 0.6921766550388435, "learning_rate": 8.580837599236673e-06, "loss": 0.6119, "step": 1675 }, { "epoch": 1.6087373979836774, "grad_norm": 0.8937408889585307, "learning_rate": 8.57849876328886e-06, "loss": 0.7035, "step": 1676 }, { "epoch": 1.6096975516082574, "grad_norm": 0.7997629673841826, "learning_rate": 8.576158320991205e-06, "loss": 0.6052, "step": 1677 }, { "epoch": 1.6106577052328372, "grad_norm": 0.6859241805879727, "learning_rate": 8.573816273394308e-06, "loss": 0.497, "step": 1678 }, { "epoch": 1.611617858857417, "grad_norm": 0.7612143820907462, "learning_rate": 8.571472621549488e-06, "loss": 0.6037, "step": 1679 }, { "epoch": 1.6125780124819973, "grad_norm": 0.8716035539558866, "learning_rate": 8.569127366508782e-06, "loss": 0.5782, "step": 1680 }, { "epoch": 1.613538166106577, "grad_norm": 1.479455500406792, "learning_rate": 8.566780509324956e-06, "loss": 0.5259, "step": 1681 }, { "epoch": 1.6144983197311569, "grad_norm": 1.0387738539967515, "learning_rate": 8.564432051051485e-06, "loss": 0.6555, "step": 1682 }, { "epoch": 1.615458473355737, "grad_norm": 0.9344638332956712, "learning_rate": 8.562081992742569e-06, "loss": 0.5852, "step": 1683 }, { "epoch": 1.616418626980317, "grad_norm": 0.9119449952114868, "learning_rate": 8.559730335453122e-06, "loss": 0.6724, "step": 1684 }, { "epoch": 1.6173787806048967, "grad_norm": 1.0426726264568933, "learning_rate": 8.557377080238778e-06, "loss": 0.5962, "step": 1685 }, { "epoch": 1.6183389342294767, "grad_norm": 1.1014139301424974, "learning_rate": 8.55502222815589e-06, "loss": 0.7017, "step": 1686 }, { "epoch": 1.6192990878540567, "grad_norm": 0.7451135327588433, "learning_rate": 8.552665780261526e-06, "loss": 0.622, "step": 1687 }, { "epoch": 1.6202592414786365, "grad_norm": 0.9028354851224325, "learning_rate": 8.550307737613468e-06, "loss": 0.5291, "step": 1688 }, { "epoch": 1.6212193951032166, "grad_norm": 0.5984659318804851, "learning_rate": 8.547948101270215e-06, "loss": 0.513, "step": 1689 }, { "epoch": 1.6221795487277966, "grad_norm": 0.7796210933352996, "learning_rate": 8.545586872290987e-06, "loss": 0.6636, "step": 1690 }, { "epoch": 1.6231397023523764, "grad_norm": 0.9037213590679422, "learning_rate": 8.543224051735714e-06, "loss": 0.6158, "step": 1691 }, { "epoch": 1.6240998559769562, "grad_norm": 0.7881237606601644, "learning_rate": 8.540859640665036e-06, "loss": 0.6708, "step": 1692 }, { "epoch": 1.6250600096015364, "grad_norm": 1.212103145610133, "learning_rate": 8.538493640140317e-06, "loss": 0.6857, "step": 1693 }, { "epoch": 1.6260201632261162, "grad_norm": 0.7132093470495673, "learning_rate": 8.536126051223627e-06, "loss": 0.5887, "step": 1694 }, { "epoch": 1.626980316850696, "grad_norm": 0.880159966537792, "learning_rate": 8.533756874977751e-06, "loss": 0.516, "step": 1695 }, { "epoch": 1.627940470475276, "grad_norm": 0.8266456531474442, "learning_rate": 8.53138611246619e-06, "loss": 0.6119, "step": 1696 }, { "epoch": 1.628900624099856, "grad_norm": 0.8866201192892907, "learning_rate": 8.529013764753147e-06, "loss": 0.7739, "step": 1697 }, { "epoch": 1.6298607777244358, "grad_norm": 0.8399290593408749, "learning_rate": 8.526639832903552e-06, "loss": 0.6594, "step": 1698 }, { "epoch": 1.6308209313490158, "grad_norm": 1.2673129534743999, "learning_rate": 8.524264317983032e-06, "loss": 0.6216, "step": 1699 }, { "epoch": 1.6317810849735959, "grad_norm": 0.8504152176537805, "learning_rate": 8.521887221057932e-06, "loss": 0.6842, "step": 1700 }, { "epoch": 1.6327412385981757, "grad_norm": 0.9159797426041865, "learning_rate": 8.519508543195304e-06, "loss": 0.6405, "step": 1701 }, { "epoch": 1.6337013922227557, "grad_norm": 0.8478800391317822, "learning_rate": 8.517128285462914e-06, "loss": 0.5788, "step": 1702 }, { "epoch": 1.6346615458473357, "grad_norm": 0.831220398474555, "learning_rate": 8.514746448929235e-06, "loss": 0.6795, "step": 1703 }, { "epoch": 1.6356216994719155, "grad_norm": 0.9058049311557352, "learning_rate": 8.512363034663441e-06, "loss": 0.6652, "step": 1704 }, { "epoch": 1.6365818530964953, "grad_norm": 0.6922601855046951, "learning_rate": 8.50997804373543e-06, "loss": 0.47, "step": 1705 }, { "epoch": 1.6375420067210755, "grad_norm": 0.8182936640890778, "learning_rate": 8.507591477215793e-06, "loss": 0.5471, "step": 1706 }, { "epoch": 1.6385021603456553, "grad_norm": 0.8224078410660354, "learning_rate": 8.505203336175836e-06, "loss": 0.5485, "step": 1707 }, { "epoch": 1.6394623139702351, "grad_norm": 1.0191010140622352, "learning_rate": 8.502813621687568e-06, "loss": 0.58, "step": 1708 }, { "epoch": 1.6404224675948151, "grad_norm": 0.8393738187507012, "learning_rate": 8.50042233482371e-06, "loss": 0.6666, "step": 1709 }, { "epoch": 1.6413826212193952, "grad_norm": 1.045676328935702, "learning_rate": 8.498029476657686e-06, "loss": 0.6981, "step": 1710 }, { "epoch": 1.642342774843975, "grad_norm": 1.0768724033036865, "learning_rate": 8.495635048263617e-06, "loss": 0.6795, "step": 1711 }, { "epoch": 1.643302928468555, "grad_norm": 0.8850029016796019, "learning_rate": 8.493239050716344e-06, "loss": 0.6494, "step": 1712 }, { "epoch": 1.644263082093135, "grad_norm": 0.69415221474156, "learning_rate": 8.4908414850914e-06, "loss": 0.6276, "step": 1713 }, { "epoch": 1.6452232357177148, "grad_norm": 0.7390577494348052, "learning_rate": 8.488442352465029e-06, "loss": 0.5946, "step": 1714 }, { "epoch": 1.6461833893422948, "grad_norm": 0.6802478255863765, "learning_rate": 8.486041653914177e-06, "loss": 0.5396, "step": 1715 }, { "epoch": 1.6471435429668748, "grad_norm": 0.9854613335079515, "learning_rate": 8.483639390516488e-06, "loss": 0.721, "step": 1716 }, { "epoch": 1.6481036965914546, "grad_norm": 1.0924418774225988, "learning_rate": 8.481235563350316e-06, "loss": 0.6932, "step": 1717 }, { "epoch": 1.6490638502160344, "grad_norm": 0.971223552574565, "learning_rate": 8.478830173494712e-06, "loss": 0.6718, "step": 1718 }, { "epoch": 1.6500240038406147, "grad_norm": 0.8001413402747101, "learning_rate": 8.47642322202943e-06, "loss": 0.6546, "step": 1719 }, { "epoch": 1.6509841574651944, "grad_norm": 0.6014228213066338, "learning_rate": 8.474014710034923e-06, "loss": 0.5596, "step": 1720 }, { "epoch": 1.6519443110897742, "grad_norm": 0.825449594353482, "learning_rate": 8.471604638592348e-06, "loss": 0.6496, "step": 1721 }, { "epoch": 1.6529044647143543, "grad_norm": 1.131501349772565, "learning_rate": 8.469193008783562e-06, "loss": 0.677, "step": 1722 }, { "epoch": 1.6538646183389343, "grad_norm": 0.6212897437779491, "learning_rate": 8.466779821691117e-06, "loss": 0.5287, "step": 1723 }, { "epoch": 1.654824771963514, "grad_norm": 0.7137811105653147, "learning_rate": 8.464365078398269e-06, "loss": 0.5815, "step": 1724 }, { "epoch": 1.655784925588094, "grad_norm": 0.7495418707615965, "learning_rate": 8.461948779988967e-06, "loss": 0.6839, "step": 1725 }, { "epoch": 1.656745079212674, "grad_norm": 0.8051162750903877, "learning_rate": 8.459530927547864e-06, "loss": 0.6724, "step": 1726 }, { "epoch": 1.657705232837254, "grad_norm": 1.1643506611162628, "learning_rate": 8.457111522160309e-06, "loss": 0.7266, "step": 1727 }, { "epoch": 1.658665386461834, "grad_norm": 1.2042083969785013, "learning_rate": 8.454690564912347e-06, "loss": 0.7078, "step": 1728 }, { "epoch": 1.659625540086414, "grad_norm": 0.9697854435461845, "learning_rate": 8.452268056890717e-06, "loss": 0.6642, "step": 1729 }, { "epoch": 1.6605856937109937, "grad_norm": 0.8731465903705319, "learning_rate": 8.44984399918286e-06, "loss": 0.6297, "step": 1730 }, { "epoch": 1.6615458473355735, "grad_norm": 0.7081064172422609, "learning_rate": 8.447418392876907e-06, "loss": 0.5655, "step": 1731 }, { "epoch": 1.6625060009601538, "grad_norm": 0.965756239146769, "learning_rate": 8.44499123906169e-06, "loss": 0.7803, "step": 1732 }, { "epoch": 1.6634661545847336, "grad_norm": 1.2718229447877165, "learning_rate": 8.44256253882673e-06, "loss": 0.6101, "step": 1733 }, { "epoch": 1.6644263082093134, "grad_norm": 0.8450205247875427, "learning_rate": 8.440132293262246e-06, "loss": 0.5284, "step": 1734 }, { "epoch": 1.6653864618338934, "grad_norm": 0.6274423906900737, "learning_rate": 8.437700503459149e-06, "loss": 0.576, "step": 1735 }, { "epoch": 1.6663466154584734, "grad_norm": 0.9238310397364994, "learning_rate": 8.435267170509044e-06, "loss": 0.6393, "step": 1736 }, { "epoch": 1.6673067690830532, "grad_norm": 1.0529531248873276, "learning_rate": 8.432832295504224e-06, "loss": 0.667, "step": 1737 }, { "epoch": 1.6682669227076332, "grad_norm": 1.01971048984101, "learning_rate": 8.430395879537684e-06, "loss": 0.5532, "step": 1738 }, { "epoch": 1.6692270763322132, "grad_norm": 0.7389955274407817, "learning_rate": 8.4279579237031e-06, "loss": 0.6488, "step": 1739 }, { "epoch": 1.670187229956793, "grad_norm": 0.8604756799998691, "learning_rate": 8.425518429094848e-06, "loss": 0.6542, "step": 1740 }, { "epoch": 1.671147383581373, "grad_norm": 0.7102635880062884, "learning_rate": 8.423077396807991e-06, "loss": 0.5541, "step": 1741 }, { "epoch": 1.672107537205953, "grad_norm": 0.7656071773304757, "learning_rate": 8.42063482793828e-06, "loss": 0.6632, "step": 1742 }, { "epoch": 1.6730676908305329, "grad_norm": 0.9669866319641147, "learning_rate": 8.41819072358216e-06, "loss": 0.7176, "step": 1743 }, { "epoch": 1.6740278444551127, "grad_norm": 0.8254889973083515, "learning_rate": 8.415745084836763e-06, "loss": 0.658, "step": 1744 }, { "epoch": 1.674987998079693, "grad_norm": 0.7864770782672674, "learning_rate": 8.413297912799909e-06, "loss": 0.6239, "step": 1745 }, { "epoch": 1.6759481517042727, "grad_norm": 0.8250929535110689, "learning_rate": 8.410849208570108e-06, "loss": 0.7024, "step": 1746 }, { "epoch": 1.6769083053288525, "grad_norm": 0.8911127011042981, "learning_rate": 8.408398973246557e-06, "loss": 0.5748, "step": 1747 }, { "epoch": 1.6778684589534325, "grad_norm": 0.9164390294595771, "learning_rate": 8.405947207929142e-06, "loss": 0.6149, "step": 1748 }, { "epoch": 1.6788286125780125, "grad_norm": 0.9779803132243893, "learning_rate": 8.40349391371843e-06, "loss": 0.6599, "step": 1749 }, { "epoch": 1.6797887662025923, "grad_norm": 0.7892276241002623, "learning_rate": 8.401039091715685e-06, "loss": 0.5924, "step": 1750 }, { "epoch": 1.6807489198271723, "grad_norm": 1.0549963826206246, "learning_rate": 8.398582743022845e-06, "loss": 0.6219, "step": 1751 }, { "epoch": 1.6817090734517524, "grad_norm": 1.004428256780341, "learning_rate": 8.39612486874254e-06, "loss": 0.6614, "step": 1752 }, { "epoch": 1.6826692270763322, "grad_norm": 0.9333148975679042, "learning_rate": 8.393665469978086e-06, "loss": 0.5838, "step": 1753 }, { "epoch": 1.6836293807009122, "grad_norm": 0.6653866080032843, "learning_rate": 8.391204547833478e-06, "loss": 0.5445, "step": 1754 }, { "epoch": 1.6845895343254922, "grad_norm": 0.8476365106978943, "learning_rate": 8.388742103413397e-06, "loss": 0.662, "step": 1755 }, { "epoch": 1.685549687950072, "grad_norm": 1.055620776423284, "learning_rate": 8.38627813782321e-06, "loss": 0.6186, "step": 1756 }, { "epoch": 1.6865098415746518, "grad_norm": 0.7649636494232778, "learning_rate": 8.383812652168966e-06, "loss": 0.5678, "step": 1757 }, { "epoch": 1.687469995199232, "grad_norm": 1.182514954985506, "learning_rate": 8.38134564755739e-06, "loss": 0.7046, "step": 1758 }, { "epoch": 1.6884301488238118, "grad_norm": 0.957534949750098, "learning_rate": 8.378877125095901e-06, "loss": 0.7367, "step": 1759 }, { "epoch": 1.6893903024483916, "grad_norm": 1.0444950031086497, "learning_rate": 8.376407085892586e-06, "loss": 0.6024, "step": 1760 }, { "epoch": 1.6903504560729716, "grad_norm": 0.8684232469017854, "learning_rate": 8.373935531056222e-06, "loss": 0.5922, "step": 1761 }, { "epoch": 1.6913106096975516, "grad_norm": 0.8112513764014687, "learning_rate": 8.371462461696267e-06, "loss": 0.6664, "step": 1762 }, { "epoch": 1.6922707633221314, "grad_norm": 0.9785249320564392, "learning_rate": 8.36898787892285e-06, "loss": 0.652, "step": 1763 }, { "epoch": 1.6932309169467115, "grad_norm": 0.8443299360222513, "learning_rate": 8.366511783846785e-06, "loss": 0.5717, "step": 1764 }, { "epoch": 1.6941910705712915, "grad_norm": 0.9492204280904003, "learning_rate": 8.364034177579568e-06, "loss": 0.5981, "step": 1765 }, { "epoch": 1.6951512241958713, "grad_norm": 0.8400696932879268, "learning_rate": 8.361555061233367e-06, "loss": 0.7236, "step": 1766 }, { "epoch": 1.6961113778204513, "grad_norm": 0.9964178347595156, "learning_rate": 8.359074435921032e-06, "loss": 0.5403, "step": 1767 }, { "epoch": 1.6970715314450313, "grad_norm": 0.8169536597668587, "learning_rate": 8.35659230275609e-06, "loss": 0.6782, "step": 1768 }, { "epoch": 1.698031685069611, "grad_norm": 0.884264320743019, "learning_rate": 8.354108662852743e-06, "loss": 0.5944, "step": 1769 }, { "epoch": 1.6989918386941911, "grad_norm": 0.8463368128339248, "learning_rate": 8.351623517325872e-06, "loss": 0.5946, "step": 1770 }, { "epoch": 1.6999519923187711, "grad_norm": 0.9215932917191383, "learning_rate": 8.349136867291029e-06, "loss": 0.7032, "step": 1771 }, { "epoch": 1.700912145943351, "grad_norm": 1.1469206375846037, "learning_rate": 8.346648713864447e-06, "loss": 0.5747, "step": 1772 }, { "epoch": 1.7018722995679307, "grad_norm": 1.0273096652624132, "learning_rate": 8.344159058163032e-06, "loss": 0.6926, "step": 1773 }, { "epoch": 1.7028324531925108, "grad_norm": 1.0468328735209944, "learning_rate": 8.341667901304362e-06, "loss": 0.6367, "step": 1774 }, { "epoch": 1.7037926068170908, "grad_norm": 0.8297971309312128, "learning_rate": 8.339175244406693e-06, "loss": 0.6454, "step": 1775 }, { "epoch": 1.7047527604416706, "grad_norm": 0.8887451018391321, "learning_rate": 8.336681088588949e-06, "loss": 0.6147, "step": 1776 }, { "epoch": 1.7057129140662506, "grad_norm": 1.1639755106668326, "learning_rate": 8.33418543497073e-06, "loss": 0.6049, "step": 1777 }, { "epoch": 1.7066730676908306, "grad_norm": 1.2401545697707748, "learning_rate": 8.331688284672313e-06, "loss": 0.6016, "step": 1778 }, { "epoch": 1.7076332213154104, "grad_norm": 1.0080578274653993, "learning_rate": 8.329189638814637e-06, "loss": 0.7265, "step": 1779 }, { "epoch": 1.7085933749399904, "grad_norm": 0.8004924985893785, "learning_rate": 8.32668949851932e-06, "loss": 0.6335, "step": 1780 }, { "epoch": 1.7095535285645704, "grad_norm": 0.9182332712080521, "learning_rate": 8.324187864908646e-06, "loss": 0.676, "step": 1781 }, { "epoch": 1.7105136821891502, "grad_norm": 0.9360447067867036, "learning_rate": 8.321684739105573e-06, "loss": 0.668, "step": 1782 }, { "epoch": 1.7114738358137302, "grad_norm": 1.0069299834749539, "learning_rate": 8.319180122233729e-06, "loss": 0.7045, "step": 1783 }, { "epoch": 1.7124339894383103, "grad_norm": 0.8505583505692816, "learning_rate": 8.316674015417407e-06, "loss": 0.5511, "step": 1784 }, { "epoch": 1.71339414306289, "grad_norm": 0.9392736856679617, "learning_rate": 8.314166419781574e-06, "loss": 0.7951, "step": 1785 }, { "epoch": 1.7143542966874699, "grad_norm": 0.949856227859823, "learning_rate": 8.311657336451862e-06, "loss": 0.5242, "step": 1786 }, { "epoch": 1.7153144503120499, "grad_norm": 0.6883931384060656, "learning_rate": 8.30914676655457e-06, "loss": 0.7528, "step": 1787 }, { "epoch": 1.71627460393663, "grad_norm": 0.98519948807859, "learning_rate": 8.306634711216668e-06, "loss": 0.7072, "step": 1788 }, { "epoch": 1.7172347575612097, "grad_norm": 0.6892030222639617, "learning_rate": 8.304121171565788e-06, "loss": 0.6482, "step": 1789 }, { "epoch": 1.7181949111857897, "grad_norm": 0.7125790856057465, "learning_rate": 8.301606148730237e-06, "loss": 0.589, "step": 1790 }, { "epoch": 1.7191550648103697, "grad_norm": 0.8795161676212587, "learning_rate": 8.299089643838976e-06, "loss": 0.5631, "step": 1791 }, { "epoch": 1.7201152184349495, "grad_norm": 0.7832570713936075, "learning_rate": 8.296571658021638e-06, "loss": 0.5743, "step": 1792 }, { "epoch": 1.7210753720595295, "grad_norm": 0.9650106853806348, "learning_rate": 8.294052192408522e-06, "loss": 0.6905, "step": 1793 }, { "epoch": 1.7220355256841096, "grad_norm": 1.0931369687467396, "learning_rate": 8.291531248130589e-06, "loss": 0.6278, "step": 1794 }, { "epoch": 1.7229956793086894, "grad_norm": 0.946043615792626, "learning_rate": 8.289008826319463e-06, "loss": 0.6494, "step": 1795 }, { "epoch": 1.7239558329332694, "grad_norm": 0.9868980319283444, "learning_rate": 8.286484928107431e-06, "loss": 0.708, "step": 1796 }, { "epoch": 1.7249159865578494, "grad_norm": 1.0797693558715593, "learning_rate": 8.283959554627448e-06, "loss": 0.5917, "step": 1797 }, { "epoch": 1.7258761401824292, "grad_norm": 0.8175884320745836, "learning_rate": 8.281432707013123e-06, "loss": 0.5866, "step": 1798 }, { "epoch": 1.726836293807009, "grad_norm": 0.758657841940049, "learning_rate": 8.278904386398733e-06, "loss": 0.6305, "step": 1799 }, { "epoch": 1.727796447431589, "grad_norm": 0.8844928299448719, "learning_rate": 8.276374593919213e-06, "loss": 0.5233, "step": 1800 }, { "epoch": 1.728756601056169, "grad_norm": 0.8754516695206522, "learning_rate": 8.27384333071016e-06, "loss": 0.6419, "step": 1801 }, { "epoch": 1.7297167546807488, "grad_norm": 0.7020093044763878, "learning_rate": 8.27131059790783e-06, "loss": 0.5518, "step": 1802 }, { "epoch": 1.7306769083053288, "grad_norm": 0.8436005987775961, "learning_rate": 8.268776396649145e-06, "loss": 0.6617, "step": 1803 }, { "epoch": 1.7316370619299088, "grad_norm": 0.9756857174009178, "learning_rate": 8.266240728071673e-06, "loss": 0.5188, "step": 1804 }, { "epoch": 1.7325972155544886, "grad_norm": 0.9740674905629048, "learning_rate": 8.263703593313654e-06, "loss": 0.6953, "step": 1805 }, { "epoch": 1.7335573691790687, "grad_norm": 1.1091692083756792, "learning_rate": 8.261164993513978e-06, "loss": 0.6721, "step": 1806 }, { "epoch": 1.7345175228036487, "grad_norm": 0.9193650277880261, "learning_rate": 8.258624929812198e-06, "loss": 0.5776, "step": 1807 }, { "epoch": 1.7354776764282285, "grad_norm": 0.9405642181917395, "learning_rate": 8.256083403348518e-06, "loss": 0.6197, "step": 1808 }, { "epoch": 1.7364378300528085, "grad_norm": 0.8842317597183708, "learning_rate": 8.253540415263805e-06, "loss": 0.6455, "step": 1809 }, { "epoch": 1.7373979836773885, "grad_norm": 1.2525556684637205, "learning_rate": 8.250995966699577e-06, "loss": 0.6471, "step": 1810 }, { "epoch": 1.7383581373019683, "grad_norm": 0.8325639057185652, "learning_rate": 8.248450058798012e-06, "loss": 0.674, "step": 1811 }, { "epoch": 1.739318290926548, "grad_norm": 0.9386093315069838, "learning_rate": 8.245902692701939e-06, "loss": 0.6246, "step": 1812 }, { "epoch": 1.7402784445511283, "grad_norm": 0.5773594453805849, "learning_rate": 8.243353869554845e-06, "loss": 0.5397, "step": 1813 }, { "epoch": 1.7412385981757081, "grad_norm": 0.7602910957230709, "learning_rate": 8.240803590500865e-06, "loss": 0.5898, "step": 1814 }, { "epoch": 1.742198751800288, "grad_norm": 0.783301796575417, "learning_rate": 8.238251856684801e-06, "loss": 0.5335, "step": 1815 }, { "epoch": 1.743158905424868, "grad_norm": 0.7626706880227063, "learning_rate": 8.23569866925209e-06, "loss": 0.6455, "step": 1816 }, { "epoch": 1.744119059049448, "grad_norm": 0.9586216242854271, "learning_rate": 8.233144029348834e-06, "loss": 0.6502, "step": 1817 }, { "epoch": 1.7450792126740278, "grad_norm": 0.9594272813029572, "learning_rate": 8.230587938121783e-06, "loss": 0.6568, "step": 1818 }, { "epoch": 1.7460393662986078, "grad_norm": 1.131333563125407, "learning_rate": 8.228030396718342e-06, "loss": 0.8442, "step": 1819 }, { "epoch": 1.7469995199231878, "grad_norm": 1.1464726422267197, "learning_rate": 8.22547140628656e-06, "loss": 0.6636, "step": 1820 }, { "epoch": 1.7479596735477676, "grad_norm": 1.0505886467580807, "learning_rate": 8.222910967975143e-06, "loss": 0.5444, "step": 1821 }, { "epoch": 1.7489198271723476, "grad_norm": 0.7279680836717375, "learning_rate": 8.220349082933444e-06, "loss": 0.5736, "step": 1822 }, { "epoch": 1.7498799807969276, "grad_norm": 0.9036044006856216, "learning_rate": 8.217785752311464e-06, "loss": 0.4857, "step": 1823 }, { "epoch": 1.7508401344215074, "grad_norm": 0.8512738427002833, "learning_rate": 8.215220977259855e-06, "loss": 0.6298, "step": 1824 }, { "epoch": 1.7518002880460872, "grad_norm": 0.9388663704618966, "learning_rate": 8.21265475892992e-06, "loss": 0.6022, "step": 1825 }, { "epoch": 1.7527604416706675, "grad_norm": 0.9100035363511046, "learning_rate": 8.210087098473605e-06, "loss": 0.5189, "step": 1826 }, { "epoch": 1.7537205952952473, "grad_norm": 0.7460995776603218, "learning_rate": 8.207517997043504e-06, "loss": 0.6247, "step": 1827 }, { "epoch": 1.754680748919827, "grad_norm": 1.1656831392930258, "learning_rate": 8.204947455792863e-06, "loss": 0.7464, "step": 1828 }, { "epoch": 1.755640902544407, "grad_norm": 1.0594239744777147, "learning_rate": 8.202375475875566e-06, "loss": 0.6781, "step": 1829 }, { "epoch": 1.756601056168987, "grad_norm": 1.0825972605660767, "learning_rate": 8.19980205844615e-06, "loss": 0.6312, "step": 1830 }, { "epoch": 1.757561209793567, "grad_norm": 1.0227766149189508, "learning_rate": 8.197227204659795e-06, "loss": 0.721, "step": 1831 }, { "epoch": 1.758521363418147, "grad_norm": 0.9900726595557175, "learning_rate": 8.194650915672323e-06, "loss": 0.6407, "step": 1832 }, { "epoch": 1.759481517042727, "grad_norm": 0.9139426217716017, "learning_rate": 8.192073192640204e-06, "loss": 0.5891, "step": 1833 }, { "epoch": 1.7604416706673067, "grad_norm": 0.789432374214935, "learning_rate": 8.18949403672055e-06, "loss": 0.5539, "step": 1834 }, { "epoch": 1.7614018242918867, "grad_norm": 0.9748139994446605, "learning_rate": 8.186913449071117e-06, "loss": 0.5802, "step": 1835 }, { "epoch": 1.7623619779164668, "grad_norm": 1.1911498155083795, "learning_rate": 8.184331430850304e-06, "loss": 0.784, "step": 1836 }, { "epoch": 1.7633221315410466, "grad_norm": 0.8550723981855377, "learning_rate": 8.181747983217148e-06, "loss": 0.5341, "step": 1837 }, { "epoch": 1.7642822851656264, "grad_norm": 0.8366746824777689, "learning_rate": 8.179163107331336e-06, "loss": 0.6751, "step": 1838 }, { "epoch": 1.7652424387902066, "grad_norm": 0.9682741322791867, "learning_rate": 8.176576804353186e-06, "loss": 0.6593, "step": 1839 }, { "epoch": 1.7662025924147864, "grad_norm": 0.8988874316714861, "learning_rate": 8.173989075443666e-06, "loss": 0.6012, "step": 1840 }, { "epoch": 1.7671627460393662, "grad_norm": 0.9968864502904072, "learning_rate": 8.171399921764377e-06, "loss": 0.577, "step": 1841 }, { "epoch": 1.7681228996639462, "grad_norm": 1.1870791523346336, "learning_rate": 8.168809344477564e-06, "loss": 0.7492, "step": 1842 }, { "epoch": 1.7690830532885262, "grad_norm": 1.2558374633231113, "learning_rate": 8.16621734474611e-06, "loss": 0.6316, "step": 1843 }, { "epoch": 1.770043206913106, "grad_norm": 0.795261535578717, "learning_rate": 8.163623923733533e-06, "loss": 0.6423, "step": 1844 }, { "epoch": 1.771003360537686, "grad_norm": 0.9006044094771959, "learning_rate": 8.161029082603994e-06, "loss": 0.6765, "step": 1845 }, { "epoch": 1.771963514162266, "grad_norm": 1.2086419689894607, "learning_rate": 8.158432822522291e-06, "loss": 0.7404, "step": 1846 }, { "epoch": 1.7729236677868458, "grad_norm": 1.1396658247783216, "learning_rate": 8.155835144653856e-06, "loss": 0.6884, "step": 1847 }, { "epoch": 1.7738838214114259, "grad_norm": 1.2577883414359539, "learning_rate": 8.153236050164757e-06, "loss": 0.5973, "step": 1848 }, { "epoch": 1.7748439750360059, "grad_norm": 0.6232475792129873, "learning_rate": 8.1506355402217e-06, "loss": 0.5939, "step": 1849 }, { "epoch": 1.7758041286605857, "grad_norm": 0.7023435310124047, "learning_rate": 8.14803361599203e-06, "loss": 0.6059, "step": 1850 }, { "epoch": 1.7767642822851655, "grad_norm": 0.980738032564808, "learning_rate": 8.14543027864372e-06, "loss": 0.6665, "step": 1851 }, { "epoch": 1.7777244359097457, "grad_norm": 0.8474597938523347, "learning_rate": 8.14282552934538e-06, "loss": 0.6967, "step": 1852 }, { "epoch": 1.7786845895343255, "grad_norm": 0.9156627877809086, "learning_rate": 8.140219369266255e-06, "loss": 0.5962, "step": 1853 }, { "epoch": 1.7796447431589053, "grad_norm": 0.8119120979679388, "learning_rate": 8.137611799576222e-06, "loss": 0.5631, "step": 1854 }, { "epoch": 1.7806048967834853, "grad_norm": 0.8779452959729206, "learning_rate": 8.135002821445793e-06, "loss": 0.5551, "step": 1855 }, { "epoch": 1.7815650504080653, "grad_norm": 0.8063279482259118, "learning_rate": 8.132392436046107e-06, "loss": 0.6687, "step": 1856 }, { "epoch": 1.7825252040326451, "grad_norm": 1.1041715221645594, "learning_rate": 8.129780644548938e-06, "loss": 0.6847, "step": 1857 }, { "epoch": 1.7834853576572252, "grad_norm": 1.2926463585673336, "learning_rate": 8.127167448126698e-06, "loss": 0.7124, "step": 1858 }, { "epoch": 1.7844455112818052, "grad_norm": 0.9561987943161869, "learning_rate": 8.124552847952416e-06, "loss": 0.593, "step": 1859 }, { "epoch": 1.785405664906385, "grad_norm": 0.6746653357323353, "learning_rate": 8.121936845199759e-06, "loss": 0.6615, "step": 1860 }, { "epoch": 1.786365818530965, "grad_norm": 0.9298128585366622, "learning_rate": 8.119319441043024e-06, "loss": 0.6629, "step": 1861 }, { "epoch": 1.787325972155545, "grad_norm": 1.0094533350578547, "learning_rate": 8.116700636657137e-06, "loss": 0.6152, "step": 1862 }, { "epoch": 1.7882861257801248, "grad_norm": 1.1333679417174083, "learning_rate": 8.11408043321765e-06, "loss": 0.6689, "step": 1863 }, { "epoch": 1.7892462794047046, "grad_norm": 0.8226567861960843, "learning_rate": 8.111458831900743e-06, "loss": 0.6844, "step": 1864 }, { "epoch": 1.7902064330292848, "grad_norm": 1.0430559242302762, "learning_rate": 8.108835833883227e-06, "loss": 0.6733, "step": 1865 }, { "epoch": 1.7911665866538646, "grad_norm": 0.5363604245655079, "learning_rate": 8.106211440342535e-06, "loss": 0.5633, "step": 1866 }, { "epoch": 1.7921267402784444, "grad_norm": 0.9042803005938134, "learning_rate": 8.103585652456732e-06, "loss": 0.5834, "step": 1867 }, { "epoch": 1.7930868939030244, "grad_norm": 0.8184305785746305, "learning_rate": 8.100958471404506e-06, "loss": 0.6242, "step": 1868 }, { "epoch": 1.7940470475276045, "grad_norm": 0.8185358767517741, "learning_rate": 8.098329898365168e-06, "loss": 0.7176, "step": 1869 }, { "epoch": 1.7950072011521843, "grad_norm": 1.1193409627111437, "learning_rate": 8.095699934518659e-06, "loss": 0.5741, "step": 1870 }, { "epoch": 1.7959673547767643, "grad_norm": 1.1160200513024263, "learning_rate": 8.093068581045538e-06, "loss": 0.6246, "step": 1871 }, { "epoch": 1.7969275084013443, "grad_norm": 0.5916221320227232, "learning_rate": 8.090435839126998e-06, "loss": 0.5423, "step": 1872 }, { "epoch": 1.797887662025924, "grad_norm": 0.5606194736779239, "learning_rate": 8.087801709944842e-06, "loss": 0.5945, "step": 1873 }, { "epoch": 1.7988478156505041, "grad_norm": 0.9683674899599346, "learning_rate": 8.085166194681508e-06, "loss": 0.6521, "step": 1874 }, { "epoch": 1.7998079692750841, "grad_norm": 0.7713014350229963, "learning_rate": 8.082529294520046e-06, "loss": 0.6196, "step": 1875 }, { "epoch": 1.800768122899664, "grad_norm": 0.9469779801745436, "learning_rate": 8.079891010644134e-06, "loss": 0.601, "step": 1876 }, { "epoch": 1.8017282765242437, "grad_norm": 0.9116338468554332, "learning_rate": 8.077251344238068e-06, "loss": 0.6513, "step": 1877 }, { "epoch": 1.802688430148824, "grad_norm": 0.857899271263292, "learning_rate": 8.074610296486771e-06, "loss": 0.4723, "step": 1878 }, { "epoch": 1.8036485837734038, "grad_norm": 1.180050972009161, "learning_rate": 8.071967868575775e-06, "loss": 0.6215, "step": 1879 }, { "epoch": 1.8046087373979836, "grad_norm": 1.4428598223874027, "learning_rate": 8.069324061691242e-06, "loss": 0.7219, "step": 1880 }, { "epoch": 1.8055688910225636, "grad_norm": 0.9946011093587612, "learning_rate": 8.066678877019945e-06, "loss": 0.6268, "step": 1881 }, { "epoch": 1.8065290446471436, "grad_norm": 0.9974575065019996, "learning_rate": 8.064032315749283e-06, "loss": 0.6693, "step": 1882 }, { "epoch": 1.8074891982717234, "grad_norm": 0.8813478510096936, "learning_rate": 8.061384379067266e-06, "loss": 0.6082, "step": 1883 }, { "epoch": 1.8084493518963034, "grad_norm": 0.7043390173827017, "learning_rate": 8.058735068162527e-06, "loss": 0.5528, "step": 1884 }, { "epoch": 1.8094095055208834, "grad_norm": 0.5971529075618898, "learning_rate": 8.05608438422431e-06, "loss": 0.5682, "step": 1885 }, { "epoch": 1.8103696591454632, "grad_norm": 0.8958811820578466, "learning_rate": 8.053432328442483e-06, "loss": 0.4565, "step": 1886 }, { "epoch": 1.8113298127700432, "grad_norm": 0.7637988324506779, "learning_rate": 8.05077890200752e-06, "loss": 0.4907, "step": 1887 }, { "epoch": 1.8122899663946233, "grad_norm": 0.7978233502432843, "learning_rate": 8.04812410611052e-06, "loss": 0.6344, "step": 1888 }, { "epoch": 1.813250120019203, "grad_norm": 0.9839785201829384, "learning_rate": 8.045467941943193e-06, "loss": 0.5795, "step": 1889 }, { "epoch": 1.8142102736437828, "grad_norm": 0.9070882244476499, "learning_rate": 8.042810410697861e-06, "loss": 0.6322, "step": 1890 }, { "epoch": 1.815170427268363, "grad_norm": 0.951366885737946, "learning_rate": 8.04015151356746e-06, "loss": 0.705, "step": 1891 }, { "epoch": 1.8161305808929429, "grad_norm": 0.679688557524127, "learning_rate": 8.037491251745543e-06, "loss": 0.6046, "step": 1892 }, { "epoch": 1.8170907345175227, "grad_norm": 0.885991999707496, "learning_rate": 8.034829626426273e-06, "loss": 0.5734, "step": 1893 }, { "epoch": 1.8180508881421027, "grad_norm": 0.9375278008842035, "learning_rate": 8.032166638804422e-06, "loss": 0.712, "step": 1894 }, { "epoch": 1.8190110417666827, "grad_norm": 0.9277234019351296, "learning_rate": 8.02950229007538e-06, "loss": 0.5181, "step": 1895 }, { "epoch": 1.8199711953912625, "grad_norm": 0.9468878890795973, "learning_rate": 8.026836581435142e-06, "loss": 0.6156, "step": 1896 }, { "epoch": 1.8209313490158425, "grad_norm": 1.067817900447423, "learning_rate": 8.02416951408032e-06, "loss": 0.6649, "step": 1897 }, { "epoch": 1.8218915026404225, "grad_norm": 1.1063403468434947, "learning_rate": 8.02150108920813e-06, "loss": 0.6269, "step": 1898 }, { "epoch": 1.8228516562650023, "grad_norm": 0.6981166409720289, "learning_rate": 8.018831308016398e-06, "loss": 0.6265, "step": 1899 }, { "epoch": 1.8238118098895824, "grad_norm": 1.1733263108044807, "learning_rate": 8.01616017170356e-06, "loss": 0.6906, "step": 1900 }, { "epoch": 1.8247719635141624, "grad_norm": 0.7279532784108341, "learning_rate": 8.013487681468663e-06, "loss": 0.5782, "step": 1901 }, { "epoch": 1.8257321171387422, "grad_norm": 0.867654357558926, "learning_rate": 8.010813838511357e-06, "loss": 0.6203, "step": 1902 }, { "epoch": 1.826692270763322, "grad_norm": 1.106222200844575, "learning_rate": 8.008138644031906e-06, "loss": 0.6895, "step": 1903 }, { "epoch": 1.8276524243879022, "grad_norm": 1.1044749332086878, "learning_rate": 8.005462099231171e-06, "loss": 0.562, "step": 1904 }, { "epoch": 1.828612578012482, "grad_norm": 0.7803213363397452, "learning_rate": 8.002784205310626e-06, "loss": 0.4915, "step": 1905 }, { "epoch": 1.8295727316370618, "grad_norm": 0.6641171921521835, "learning_rate": 8.00010496347235e-06, "loss": 0.5365, "step": 1906 }, { "epoch": 1.8305328852616418, "grad_norm": 0.9051879255865399, "learning_rate": 7.997424374919024e-06, "loss": 0.5847, "step": 1907 }, { "epoch": 1.8314930388862218, "grad_norm": 0.6297501189499224, "learning_rate": 7.994742440853937e-06, "loss": 0.6902, "step": 1908 }, { "epoch": 1.8324531925108016, "grad_norm": 0.7916605218313448, "learning_rate": 7.992059162480983e-06, "loss": 0.5556, "step": 1909 }, { "epoch": 1.8334133461353816, "grad_norm": 0.9371997953299792, "learning_rate": 7.98937454100465e-06, "loss": 0.6506, "step": 1910 }, { "epoch": 1.8343734997599617, "grad_norm": 0.9117346897366826, "learning_rate": 7.986688577630042e-06, "loss": 0.6122, "step": 1911 }, { "epoch": 1.8353336533845415, "grad_norm": 0.9737111581944593, "learning_rate": 7.984001273562858e-06, "loss": 0.734, "step": 1912 }, { "epoch": 1.8362938070091215, "grad_norm": 0.9717448307074938, "learning_rate": 7.981312630009398e-06, "loss": 0.6507, "step": 1913 }, { "epoch": 1.8372539606337015, "grad_norm": 0.8543047685744309, "learning_rate": 7.97862264817657e-06, "loss": 0.5686, "step": 1914 }, { "epoch": 1.8382141142582813, "grad_norm": 1.2713625794083543, "learning_rate": 7.975931329271874e-06, "loss": 0.5946, "step": 1915 }, { "epoch": 1.839174267882861, "grad_norm": 0.8567896442362564, "learning_rate": 7.973238674503412e-06, "loss": 0.6278, "step": 1916 }, { "epoch": 1.8401344215074413, "grad_norm": 1.0192957707937629, "learning_rate": 7.970544685079895e-06, "loss": 0.5583, "step": 1917 }, { "epoch": 1.8410945751320211, "grad_norm": 0.6816045121960193, "learning_rate": 7.96784936221062e-06, "loss": 0.5951, "step": 1918 }, { "epoch": 1.842054728756601, "grad_norm": 0.7315603901231943, "learning_rate": 7.965152707105491e-06, "loss": 0.6478, "step": 1919 }, { "epoch": 1.843014882381181, "grad_norm": 0.8348069045511103, "learning_rate": 7.962454720975008e-06, "loss": 0.5666, "step": 1920 }, { "epoch": 1.843975036005761, "grad_norm": 0.9613979817907081, "learning_rate": 7.959755405030269e-06, "loss": 0.5809, "step": 1921 }, { "epoch": 1.8449351896303408, "grad_norm": 0.5983789723626103, "learning_rate": 7.957054760482964e-06, "loss": 0.5527, "step": 1922 }, { "epoch": 1.8458953432549208, "grad_norm": 0.7797326611806045, "learning_rate": 7.954352788545388e-06, "loss": 0.6024, "step": 1923 }, { "epoch": 1.8468554968795008, "grad_norm": 0.7968965897086806, "learning_rate": 7.951649490430424e-06, "loss": 0.5997, "step": 1924 }, { "epoch": 1.8478156505040806, "grad_norm": 0.9123070187964238, "learning_rate": 7.948944867351556e-06, "loss": 0.7407, "step": 1925 }, { "epoch": 1.8487758041286606, "grad_norm": 1.1379925689401817, "learning_rate": 7.946238920522862e-06, "loss": 0.6603, "step": 1926 }, { "epoch": 1.8497359577532406, "grad_norm": 1.014160190277735, "learning_rate": 7.943531651159007e-06, "loss": 0.6286, "step": 1927 }, { "epoch": 1.8506961113778204, "grad_norm": 1.1169919116789842, "learning_rate": 7.940823060475257e-06, "loss": 0.6098, "step": 1928 }, { "epoch": 1.8516562650024004, "grad_norm": 1.0259075013326764, "learning_rate": 7.938113149687474e-06, "loss": 0.6853, "step": 1929 }, { "epoch": 1.8526164186269805, "grad_norm": 1.0760070089988947, "learning_rate": 7.935401920012105e-06, "loss": 0.7549, "step": 1930 }, { "epoch": 1.8535765722515603, "grad_norm": 0.9097336225931517, "learning_rate": 7.932689372666192e-06, "loss": 0.6023, "step": 1931 }, { "epoch": 1.85453672587614, "grad_norm": 0.9262780975815753, "learning_rate": 7.929975508867364e-06, "loss": 0.6557, "step": 1932 }, { "epoch": 1.85549687950072, "grad_norm": 0.9266881096305977, "learning_rate": 7.927260329833854e-06, "loss": 0.5082, "step": 1933 }, { "epoch": 1.8564570331253, "grad_norm": 0.9694415784089337, "learning_rate": 7.924543836784471e-06, "loss": 0.6744, "step": 1934 }, { "epoch": 1.8574171867498799, "grad_norm": 0.8978225260638939, "learning_rate": 7.921826030938623e-06, "loss": 0.5713, "step": 1935 }, { "epoch": 1.85837734037446, "grad_norm": 0.965327741001955, "learning_rate": 7.919106913516299e-06, "loss": 0.6636, "step": 1936 }, { "epoch": 1.85933749399904, "grad_norm": 1.1319750257749779, "learning_rate": 7.916386485738089e-06, "loss": 0.6917, "step": 1937 }, { "epoch": 1.8602976476236197, "grad_norm": 0.9972505613895731, "learning_rate": 7.913664748825156e-06, "loss": 0.6493, "step": 1938 }, { "epoch": 1.8612578012481997, "grad_norm": 1.086126238723035, "learning_rate": 7.910941703999265e-06, "loss": 0.6493, "step": 1939 }, { "epoch": 1.8622179548727797, "grad_norm": 1.1018213772479695, "learning_rate": 7.908217352482759e-06, "loss": 0.5303, "step": 1940 }, { "epoch": 1.8631781084973595, "grad_norm": 0.7387896382795471, "learning_rate": 7.905491695498571e-06, "loss": 0.5875, "step": 1941 }, { "epoch": 1.8641382621219396, "grad_norm": 0.8403837960484959, "learning_rate": 7.902764734270219e-06, "loss": 0.7191, "step": 1942 }, { "epoch": 1.8650984157465196, "grad_norm": 0.8417085404157442, "learning_rate": 7.900036470021805e-06, "loss": 0.6144, "step": 1943 }, { "epoch": 1.8660585693710994, "grad_norm": 0.9972705854564717, "learning_rate": 7.89730690397802e-06, "loss": 0.6593, "step": 1944 }, { "epoch": 1.8670187229956792, "grad_norm": 0.775286854270647, "learning_rate": 7.894576037364136e-06, "loss": 0.5562, "step": 1945 }, { "epoch": 1.8679788766202592, "grad_norm": 1.103950646667244, "learning_rate": 7.89184387140601e-06, "loss": 0.5981, "step": 1946 }, { "epoch": 1.8689390302448392, "grad_norm": 0.9437937381561052, "learning_rate": 7.889110407330083e-06, "loss": 0.6734, "step": 1947 }, { "epoch": 1.869899183869419, "grad_norm": 1.0072187390554648, "learning_rate": 7.886375646363378e-06, "loss": 0.5411, "step": 1948 }, { "epoch": 1.870859337493999, "grad_norm": 0.8946624874620194, "learning_rate": 7.883639589733498e-06, "loss": 0.6573, "step": 1949 }, { "epoch": 1.871819491118579, "grad_norm": 0.7876613824318013, "learning_rate": 7.880902238668631e-06, "loss": 0.6199, "step": 1950 }, { "epoch": 1.8727796447431588, "grad_norm": 0.7697798829637127, "learning_rate": 7.878163594397543e-06, "loss": 0.5179, "step": 1951 }, { "epoch": 1.8737397983677389, "grad_norm": 0.9784685725164869, "learning_rate": 7.875423658149583e-06, "loss": 0.8139, "step": 1952 }, { "epoch": 1.8746999519923189, "grad_norm": 0.8112472111933303, "learning_rate": 7.87268243115468e-06, "loss": 0.6765, "step": 1953 }, { "epoch": 1.8756601056168987, "grad_norm": 0.7583042242097884, "learning_rate": 7.869939914643343e-06, "loss": 0.6931, "step": 1954 }, { "epoch": 1.8766202592414787, "grad_norm": 1.3643125965604521, "learning_rate": 7.867196109846653e-06, "loss": 0.6702, "step": 1955 }, { "epoch": 1.8775804128660587, "grad_norm": 0.9036068809707236, "learning_rate": 7.86445101799628e-06, "loss": 0.6262, "step": 1956 }, { "epoch": 1.8785405664906385, "grad_norm": 0.9991879746183677, "learning_rate": 7.861704640324466e-06, "loss": 0.52, "step": 1957 }, { "epoch": 1.8795007201152183, "grad_norm": 1.1069514482006277, "learning_rate": 7.858956978064025e-06, "loss": 0.6885, "step": 1958 }, { "epoch": 1.8804608737397985, "grad_norm": 0.9971132864800663, "learning_rate": 7.85620803244836e-06, "loss": 0.6125, "step": 1959 }, { "epoch": 1.8814210273643783, "grad_norm": 1.127427131764366, "learning_rate": 7.85345780471144e-06, "loss": 0.7519, "step": 1960 }, { "epoch": 1.8823811809889581, "grad_norm": 1.0336059018391037, "learning_rate": 7.850706296087811e-06, "loss": 0.6632, "step": 1961 }, { "epoch": 1.8833413346135381, "grad_norm": 0.944466215197925, "learning_rate": 7.8479535078126e-06, "loss": 0.5523, "step": 1962 }, { "epoch": 1.8843014882381182, "grad_norm": 0.7539275211146399, "learning_rate": 7.8451994411215e-06, "loss": 0.6455, "step": 1963 }, { "epoch": 1.885261641862698, "grad_norm": 1.0916849726619389, "learning_rate": 7.842444097250787e-06, "loss": 0.5074, "step": 1964 }, { "epoch": 1.886221795487278, "grad_norm": 0.716147788914688, "learning_rate": 7.839687477437304e-06, "loss": 0.6245, "step": 1965 }, { "epoch": 1.887181949111858, "grad_norm": 1.5119243480157998, "learning_rate": 7.836929582918464e-06, "loss": 0.8264, "step": 1966 }, { "epoch": 1.8881421027364378, "grad_norm": 1.2081887736613088, "learning_rate": 7.834170414932259e-06, "loss": 0.5512, "step": 1967 }, { "epoch": 1.8891022563610178, "grad_norm": 0.8624149910772163, "learning_rate": 7.831409974717252e-06, "loss": 0.6319, "step": 1968 }, { "epoch": 1.8900624099855978, "grad_norm": 1.0796741809522628, "learning_rate": 7.828648263512573e-06, "loss": 0.5691, "step": 1969 }, { "epoch": 1.8910225636101776, "grad_norm": 1.4232109772277561, "learning_rate": 7.825885282557926e-06, "loss": 0.7521, "step": 1970 }, { "epoch": 1.8919827172347574, "grad_norm": 0.8900825279575512, "learning_rate": 7.823121033093581e-06, "loss": 0.5797, "step": 1971 }, { "epoch": 1.8929428708593377, "grad_norm": 1.0501921575619346, "learning_rate": 7.820355516360383e-06, "loss": 0.6384, "step": 1972 }, { "epoch": 1.8939030244839175, "grad_norm": 1.0646554306120644, "learning_rate": 7.817588733599742e-06, "loss": 0.8821, "step": 1973 }, { "epoch": 1.8948631781084972, "grad_norm": 1.0902286069859322, "learning_rate": 7.814820686053634e-06, "loss": 0.608, "step": 1974 }, { "epoch": 1.8958233317330773, "grad_norm": 0.8392640184524829, "learning_rate": 7.812051374964611e-06, "loss": 0.6382, "step": 1975 }, { "epoch": 1.8967834853576573, "grad_norm": 0.8755971457615961, "learning_rate": 7.809280801575784e-06, "loss": 0.6939, "step": 1976 }, { "epoch": 1.897743638982237, "grad_norm": 0.7698304515240041, "learning_rate": 7.806508967130838e-06, "loss": 0.6185, "step": 1977 }, { "epoch": 1.898703792606817, "grad_norm": 0.862118981305274, "learning_rate": 7.803735872874012e-06, "loss": 0.6503, "step": 1978 }, { "epoch": 1.8996639462313971, "grad_norm": 0.9943161585681429, "learning_rate": 7.800961520050125e-06, "loss": 0.6327, "step": 1979 }, { "epoch": 1.900624099855977, "grad_norm": 0.9722404968042407, "learning_rate": 7.798185909904552e-06, "loss": 0.6233, "step": 1980 }, { "epoch": 1.901584253480557, "grad_norm": 0.7821611612468815, "learning_rate": 7.795409043683237e-06, "loss": 0.4699, "step": 1981 }, { "epoch": 1.902544407105137, "grad_norm": 0.9928676848134905, "learning_rate": 7.792630922632682e-06, "loss": 0.6884, "step": 1982 }, { "epoch": 1.9035045607297167, "grad_norm": 1.1677849667603468, "learning_rate": 7.789851547999957e-06, "loss": 0.5316, "step": 1983 }, { "epoch": 1.9044647143542965, "grad_norm": 0.8717417066700277, "learning_rate": 7.787070921032693e-06, "loss": 0.5911, "step": 1984 }, { "epoch": 1.9054248679788768, "grad_norm": 0.974831819459773, "learning_rate": 7.784289042979085e-06, "loss": 0.7372, "step": 1985 }, { "epoch": 1.9063850216034566, "grad_norm": 0.8995543159935274, "learning_rate": 7.781505915087891e-06, "loss": 0.5861, "step": 1986 }, { "epoch": 1.9073451752280364, "grad_norm": 0.9020584981453867, "learning_rate": 7.778721538608422e-06, "loss": 0.645, "step": 1987 }, { "epoch": 1.9083053288526164, "grad_norm": 0.7822145307765778, "learning_rate": 7.775935914790557e-06, "loss": 0.6439, "step": 1988 }, { "epoch": 1.9092654824771964, "grad_norm": 0.9844618647840571, "learning_rate": 7.773149044884732e-06, "loss": 0.7282, "step": 1989 }, { "epoch": 1.9102256361017762, "grad_norm": 0.8088525262945115, "learning_rate": 7.770360930141946e-06, "loss": 0.5936, "step": 1990 }, { "epoch": 1.9111857897263562, "grad_norm": 0.7765415408440055, "learning_rate": 7.767571571813752e-06, "loss": 0.5991, "step": 1991 }, { "epoch": 1.9121459433509362, "grad_norm": 0.8539444303301146, "learning_rate": 7.764780971152262e-06, "loss": 0.5954, "step": 1992 }, { "epoch": 1.913106096975516, "grad_norm": 0.8354730009451035, "learning_rate": 7.761989129410149e-06, "loss": 0.6299, "step": 1993 }, { "epoch": 1.914066250600096, "grad_norm": 0.8461755148093407, "learning_rate": 7.759196047840639e-06, "loss": 0.6397, "step": 1994 }, { "epoch": 1.915026404224676, "grad_norm": 0.9274518629050003, "learning_rate": 7.756401727697518e-06, "loss": 0.5895, "step": 1995 }, { "epoch": 1.9159865578492559, "grad_norm": 0.7336020802199351, "learning_rate": 7.753606170235124e-06, "loss": 0.6632, "step": 1996 }, { "epoch": 1.9169467114738357, "grad_norm": 0.8641080532203183, "learning_rate": 7.750809376708357e-06, "loss": 0.6496, "step": 1997 }, { "epoch": 1.917906865098416, "grad_norm": 0.9354336217178706, "learning_rate": 7.748011348372663e-06, "loss": 0.6766, "step": 1998 }, { "epoch": 1.9188670187229957, "grad_norm": 0.88608228514393, "learning_rate": 7.74521208648405e-06, "loss": 0.7184, "step": 1999 }, { "epoch": 1.9198271723475755, "grad_norm": 1.1547907772281552, "learning_rate": 7.742411592299077e-06, "loss": 0.7307, "step": 2000 }, { "epoch": 1.9207873259721555, "grad_norm": 1.194141367887897, "learning_rate": 7.739609867074856e-06, "loss": 0.7077, "step": 2001 }, { "epoch": 1.9217474795967355, "grad_norm": 0.7205636583457048, "learning_rate": 7.736806912069048e-06, "loss": 0.7215, "step": 2002 }, { "epoch": 1.9227076332213153, "grad_norm": 1.073701391187594, "learning_rate": 7.734002728539873e-06, "loss": 0.596, "step": 2003 }, { "epoch": 1.9236677868458953, "grad_norm": 0.7467720393028628, "learning_rate": 7.731197317746099e-06, "loss": 0.5966, "step": 2004 }, { "epoch": 1.9246279404704754, "grad_norm": 0.9084478767415991, "learning_rate": 7.728390680947043e-06, "loss": 0.6649, "step": 2005 }, { "epoch": 1.9255880940950552, "grad_norm": 0.8001754230024235, "learning_rate": 7.725582819402575e-06, "loss": 0.8135, "step": 2006 }, { "epoch": 1.9265482477196352, "grad_norm": 1.2084571008514406, "learning_rate": 7.722773734373114e-06, "loss": 0.574, "step": 2007 }, { "epoch": 1.9275084013442152, "grad_norm": 0.864033603118713, "learning_rate": 7.71996342711963e-06, "loss": 0.5994, "step": 2008 }, { "epoch": 1.928468554968795, "grad_norm": 0.8807523423199151, "learning_rate": 7.717151898903636e-06, "loss": 0.6204, "step": 2009 }, { "epoch": 1.9294287085933748, "grad_norm": 1.3340442391516703, "learning_rate": 7.714339150987202e-06, "loss": 0.6254, "step": 2010 }, { "epoch": 1.930388862217955, "grad_norm": 1.1938257519062383, "learning_rate": 7.711525184632936e-06, "loss": 0.6389, "step": 2011 }, { "epoch": 1.9313490158425348, "grad_norm": 0.9541984975533442, "learning_rate": 7.708710001103998e-06, "loss": 0.6179, "step": 2012 }, { "epoch": 1.9323091694671146, "grad_norm": 0.8715333344297239, "learning_rate": 7.705893601664099e-06, "loss": 0.6423, "step": 2013 }, { "epoch": 1.9332693230916946, "grad_norm": 0.9037585733829242, "learning_rate": 7.703075987577483e-06, "loss": 0.659, "step": 2014 }, { "epoch": 1.9342294767162747, "grad_norm": 0.9962308566027565, "learning_rate": 7.700257160108953e-06, "loss": 0.6994, "step": 2015 }, { "epoch": 1.9351896303408544, "grad_norm": 0.875545879471792, "learning_rate": 7.697437120523845e-06, "loss": 0.6477, "step": 2016 }, { "epoch": 1.9361497839654345, "grad_norm": 1.326548370125689, "learning_rate": 7.694615870088051e-06, "loss": 0.7586, "step": 2017 }, { "epoch": 1.9371099375900145, "grad_norm": 1.3520343865933047, "learning_rate": 7.691793410067993e-06, "loss": 0.5693, "step": 2018 }, { "epoch": 1.9380700912145943, "grad_norm": 0.7898603774915143, "learning_rate": 7.68896974173065e-06, "loss": 0.7202, "step": 2019 }, { "epoch": 1.9390302448391743, "grad_norm": 1.0378568832955137, "learning_rate": 7.686144866343533e-06, "loss": 0.6088, "step": 2020 }, { "epoch": 1.9399903984637543, "grad_norm": 0.939495727483475, "learning_rate": 7.683318785174698e-06, "loss": 0.6548, "step": 2021 }, { "epoch": 1.9409505520883341, "grad_norm": 0.8311874955825205, "learning_rate": 7.680491499492743e-06, "loss": 0.5841, "step": 2022 }, { "epoch": 1.941910705712914, "grad_norm": 0.7247692000417247, "learning_rate": 7.677663010566807e-06, "loss": 0.6338, "step": 2023 }, { "epoch": 1.9428708593374941, "grad_norm": 0.7816594483672441, "learning_rate": 7.674833319666568e-06, "loss": 0.5791, "step": 2024 }, { "epoch": 1.943831012962074, "grad_norm": 0.7972117558416147, "learning_rate": 7.672002428062245e-06, "loss": 0.5359, "step": 2025 }, { "epoch": 1.9447911665866537, "grad_norm": 0.7624472098413432, "learning_rate": 7.669170337024593e-06, "loss": 0.6179, "step": 2026 }, { "epoch": 1.9457513202112338, "grad_norm": 0.8354118500292511, "learning_rate": 7.66633704782491e-06, "loss": 0.6042, "step": 2027 }, { "epoch": 1.9467114738358138, "grad_norm": 0.8058746313903365, "learning_rate": 7.663502561735028e-06, "loss": 0.6624, "step": 2028 }, { "epoch": 1.9476716274603936, "grad_norm": 1.1058308079759365, "learning_rate": 7.660666880027316e-06, "loss": 0.7179, "step": 2029 }, { "epoch": 1.9486317810849736, "grad_norm": 0.6946700613486193, "learning_rate": 7.657830003974684e-06, "loss": 0.6086, "step": 2030 }, { "epoch": 1.9495919347095536, "grad_norm": 1.0249793508439, "learning_rate": 7.654991934850573e-06, "loss": 0.6536, "step": 2031 }, { "epoch": 1.9505520883341334, "grad_norm": 1.1972118131298104, "learning_rate": 7.652152673928964e-06, "loss": 0.7746, "step": 2032 }, { "epoch": 1.9515122419587134, "grad_norm": 1.093746649207187, "learning_rate": 7.649312222484367e-06, "loss": 0.5994, "step": 2033 }, { "epoch": 1.9524723955832934, "grad_norm": 0.8347955957670684, "learning_rate": 7.646470581791834e-06, "loss": 0.5921, "step": 2034 }, { "epoch": 1.9534325492078732, "grad_norm": 1.1171003230733125, "learning_rate": 7.643627753126946e-06, "loss": 0.7207, "step": 2035 }, { "epoch": 1.954392702832453, "grad_norm": 1.0691047798970714, "learning_rate": 7.640783737765817e-06, "loss": 0.8462, "step": 2036 }, { "epoch": 1.9553528564570333, "grad_norm": 0.9492361124946996, "learning_rate": 7.6379385369851e-06, "loss": 0.5779, "step": 2037 }, { "epoch": 1.956313010081613, "grad_norm": 1.0689059434697887, "learning_rate": 7.63509215206197e-06, "loss": 0.6699, "step": 2038 }, { "epoch": 1.9572731637061929, "grad_norm": 1.0449624471396741, "learning_rate": 7.63224458427414e-06, "loss": 0.6444, "step": 2039 }, { "epoch": 1.9582333173307729, "grad_norm": 0.9986913099968208, "learning_rate": 7.629395834899852e-06, "loss": 0.5719, "step": 2040 }, { "epoch": 1.959193470955353, "grad_norm": 0.7407757798607245, "learning_rate": 7.626545905217884e-06, "loss": 0.6555, "step": 2041 }, { "epoch": 1.9601536245799327, "grad_norm": 0.8675699160062569, "learning_rate": 7.6236947965075325e-06, "loss": 0.6806, "step": 2042 }, { "epoch": 1.9611137782045127, "grad_norm": 0.9725871357209583, "learning_rate": 7.6208425100486334e-06, "loss": 0.5911, "step": 2043 }, { "epoch": 1.9620739318290927, "grad_norm": 1.0953363285597555, "learning_rate": 7.617989047121548e-06, "loss": 0.608, "step": 2044 }, { "epoch": 1.9630340854536725, "grad_norm": 0.7565680575020098, "learning_rate": 7.6151344090071624e-06, "loss": 0.6421, "step": 2045 }, { "epoch": 1.9639942390782525, "grad_norm": 0.6536329266687899, "learning_rate": 7.612278596986895e-06, "loss": 0.5916, "step": 2046 }, { "epoch": 1.9649543927028326, "grad_norm": 0.998585072562118, "learning_rate": 7.609421612342689e-06, "loss": 0.6575, "step": 2047 }, { "epoch": 1.9659145463274124, "grad_norm": 1.0033923712734447, "learning_rate": 7.606563456357016e-06, "loss": 0.7084, "step": 2048 }, { "epoch": 1.9668746999519922, "grad_norm": 0.729060675551785, "learning_rate": 7.603704130312868e-06, "loss": 0.5785, "step": 2049 }, { "epoch": 1.9678348535765724, "grad_norm": 0.727944179708888, "learning_rate": 7.6008436354937695e-06, "loss": 0.718, "step": 2050 }, { "epoch": 1.9687950072011522, "grad_norm": 0.8281756041082237, "learning_rate": 7.5979819731837635e-06, "loss": 0.7344, "step": 2051 }, { "epoch": 1.969755160825732, "grad_norm": 0.9329862446149164, "learning_rate": 7.595119144667421e-06, "loss": 0.7165, "step": 2052 }, { "epoch": 1.970715314450312, "grad_norm": 0.8828957169594952, "learning_rate": 7.5922551512298345e-06, "loss": 0.5483, "step": 2053 }, { "epoch": 1.971675468074892, "grad_norm": 0.7476165719234853, "learning_rate": 7.589389994156622e-06, "loss": 0.7144, "step": 2054 }, { "epoch": 1.9726356216994718, "grad_norm": 0.8395104715820254, "learning_rate": 7.586523674733918e-06, "loss": 0.602, "step": 2055 }, { "epoch": 1.9735957753240518, "grad_norm": 0.9170654839495498, "learning_rate": 7.583656194248386e-06, "loss": 0.5761, "step": 2056 }, { "epoch": 1.9745559289486319, "grad_norm": 0.9806413963243523, "learning_rate": 7.5807875539872054e-06, "loss": 0.7685, "step": 2057 }, { "epoch": 1.9755160825732117, "grad_norm": 1.4222531529964972, "learning_rate": 7.577917755238079e-06, "loss": 0.7399, "step": 2058 }, { "epoch": 1.9764762361977917, "grad_norm": 0.8699719422558561, "learning_rate": 7.57504679928923e-06, "loss": 0.6308, "step": 2059 }, { "epoch": 1.9774363898223717, "grad_norm": 0.7656224040199374, "learning_rate": 7.572174687429395e-06, "loss": 0.6589, "step": 2060 }, { "epoch": 1.9783965434469515, "grad_norm": 0.8745606441292654, "learning_rate": 7.569301420947841e-06, "loss": 0.5932, "step": 2061 }, { "epoch": 1.9793566970715313, "grad_norm": 0.8861310358287826, "learning_rate": 7.566427001134342e-06, "loss": 0.5885, "step": 2062 }, { "epoch": 1.9803168506961115, "grad_norm": 1.167063177067495, "learning_rate": 7.563551429279198e-06, "loss": 0.6818, "step": 2063 }, { "epoch": 1.9812770043206913, "grad_norm": 1.1967136474653906, "learning_rate": 7.560674706673219e-06, "loss": 0.6355, "step": 2064 }, { "epoch": 1.982237157945271, "grad_norm": 0.9226125556126549, "learning_rate": 7.557796834607736e-06, "loss": 0.5356, "step": 2065 }, { "epoch": 1.9831973115698511, "grad_norm": 1.0065419074311543, "learning_rate": 7.554917814374595e-06, "loss": 0.5442, "step": 2066 }, { "epoch": 1.9841574651944311, "grad_norm": 0.8282971542529773, "learning_rate": 7.552037647266157e-06, "loss": 0.58, "step": 2067 }, { "epoch": 1.985117618819011, "grad_norm": 0.8136645031757505, "learning_rate": 7.5491563345753e-06, "loss": 0.632, "step": 2068 }, { "epoch": 1.986077772443591, "grad_norm": 0.7885466477598058, "learning_rate": 7.546273877595413e-06, "loss": 0.5123, "step": 2069 }, { "epoch": 1.987037926068171, "grad_norm": 0.8151657594849312, "learning_rate": 7.5433902776204015e-06, "loss": 0.5897, "step": 2070 }, { "epoch": 1.9879980796927508, "grad_norm": 0.8626704980877897, "learning_rate": 7.54050553594468e-06, "loss": 0.5942, "step": 2071 }, { "epoch": 1.9889582333173308, "grad_norm": 1.171503040808102, "learning_rate": 7.537619653863182e-06, "loss": 0.7002, "step": 2072 }, { "epoch": 1.9899183869419108, "grad_norm": 0.6807404401066102, "learning_rate": 7.5347326326713454e-06, "loss": 0.6855, "step": 2073 }, { "epoch": 1.9908785405664906, "grad_norm": 0.8141339969784308, "learning_rate": 7.531844473665125e-06, "loss": 0.6646, "step": 2074 }, { "epoch": 1.9918386941910706, "grad_norm": 0.859304594143771, "learning_rate": 7.528955178140985e-06, "loss": 0.6052, "step": 2075 }, { "epoch": 1.9927988478156506, "grad_norm": 0.916913498471583, "learning_rate": 7.526064747395898e-06, "loss": 0.7115, "step": 2076 }, { "epoch": 1.9937590014402304, "grad_norm": 0.908702440069623, "learning_rate": 7.523173182727348e-06, "loss": 0.6183, "step": 2077 }, { "epoch": 1.9947191550648102, "grad_norm": 0.8333238576228097, "learning_rate": 7.520280485433327e-06, "loss": 0.524, "step": 2078 }, { "epoch": 1.9956793086893903, "grad_norm": 0.7647798238474386, "learning_rate": 7.517386656812338e-06, "loss": 0.7692, "step": 2079 }, { "epoch": 1.9966394623139703, "grad_norm": 1.1052015364387784, "learning_rate": 7.514491698163387e-06, "loss": 0.7164, "step": 2080 }, { "epoch": 1.99759961593855, "grad_norm": 0.7534590995194151, "learning_rate": 7.511595610785991e-06, "loss": 0.6177, "step": 2081 }, { "epoch": 1.99855976956313, "grad_norm": 0.9689163976377876, "learning_rate": 7.508698395980173e-06, "loss": 0.6793, "step": 2082 }, { "epoch": 1.99951992318771, "grad_norm": 1.3092023218880824, "learning_rate": 7.5058000550464615e-06, "loss": 0.6288, "step": 2083 }, { "epoch": 2.0, "grad_norm": 2.528018598416632, "learning_rate": 7.502900589285891e-06, "loss": 0.7635, "step": 2084 }, { "epoch": 2.00096015362458, "grad_norm": 1.0376674055411663, "learning_rate": 7.500000000000001e-06, "loss": 0.5936, "step": 2085 }, { "epoch": 2.00192030724916, "grad_norm": 0.8336702282355283, "learning_rate": 7.497098288490834e-06, "loss": 0.5635, "step": 2086 }, { "epoch": 2.00288046087374, "grad_norm": 1.0549103972681355, "learning_rate": 7.494195456060938e-06, "loss": 0.6077, "step": 2087 }, { "epoch": 2.0038406144983196, "grad_norm": 0.6864330254988222, "learning_rate": 7.491291504013363e-06, "loss": 0.5743, "step": 2088 }, { "epoch": 2.0048007681229, "grad_norm": 0.7293517236114241, "learning_rate": 7.488386433651662e-06, "loss": 0.5259, "step": 2089 }, { "epoch": 2.0057609217474797, "grad_norm": 0.8202235460415569, "learning_rate": 7.4854802462798905e-06, "loss": 0.56, "step": 2090 }, { "epoch": 2.0067210753720595, "grad_norm": 1.0775747423677804, "learning_rate": 7.482572943202604e-06, "loss": 0.573, "step": 2091 }, { "epoch": 2.0076812289966393, "grad_norm": 0.8799209537300825, "learning_rate": 7.4796645257248625e-06, "loss": 0.6101, "step": 2092 }, { "epoch": 2.0086413826212195, "grad_norm": 0.7423371690109478, "learning_rate": 7.47675499515222e-06, "loss": 0.5267, "step": 2093 }, { "epoch": 2.0096015362457993, "grad_norm": 1.0838415117152171, "learning_rate": 7.473844352790735e-06, "loss": 0.5448, "step": 2094 }, { "epoch": 2.010561689870379, "grad_norm": 0.8933840695074711, "learning_rate": 7.470932599946965e-06, "loss": 0.5477, "step": 2095 }, { "epoch": 2.0115218434949593, "grad_norm": 0.8369724498231511, "learning_rate": 7.468019737927963e-06, "loss": 0.5694, "step": 2096 }, { "epoch": 2.012481997119539, "grad_norm": 1.3796250530122531, "learning_rate": 7.465105768041283e-06, "loss": 0.6764, "step": 2097 }, { "epoch": 2.013442150744119, "grad_norm": 0.8473428633725187, "learning_rate": 7.462190691594973e-06, "loss": 0.6126, "step": 2098 }, { "epoch": 2.014402304368699, "grad_norm": 0.9606394007802624, "learning_rate": 7.459274509897583e-06, "loss": 0.5751, "step": 2099 }, { "epoch": 2.015362457993279, "grad_norm": 0.762188311753643, "learning_rate": 7.45635722425815e-06, "loss": 0.4548, "step": 2100 }, { "epoch": 2.0163226116178588, "grad_norm": 0.9884310302577289, "learning_rate": 7.45343883598622e-06, "loss": 0.5616, "step": 2101 }, { "epoch": 2.017282765242439, "grad_norm": 0.6932614811698198, "learning_rate": 7.45051934639182e-06, "loss": 0.5156, "step": 2102 }, { "epoch": 2.018242918867019, "grad_norm": 0.7992141718024329, "learning_rate": 7.447598756785482e-06, "loss": 0.5073, "step": 2103 }, { "epoch": 2.0192030724915986, "grad_norm": 0.7554886213442823, "learning_rate": 7.444677068478223e-06, "loss": 0.5456, "step": 2104 }, { "epoch": 2.0201632261161784, "grad_norm": 0.874554130978943, "learning_rate": 7.441754282781561e-06, "loss": 0.645, "step": 2105 }, { "epoch": 2.0211233797407586, "grad_norm": 1.3305492245204977, "learning_rate": 7.438830401007504e-06, "loss": 0.6014, "step": 2106 }, { "epoch": 2.0220835333653384, "grad_norm": 0.8491323388801935, "learning_rate": 7.435905424468548e-06, "loss": 0.5113, "step": 2107 }, { "epoch": 2.023043686989918, "grad_norm": 0.9504154422373571, "learning_rate": 7.432979354477686e-06, "loss": 0.5721, "step": 2108 }, { "epoch": 2.0240038406144985, "grad_norm": 0.8639001784206435, "learning_rate": 7.430052192348398e-06, "loss": 0.5147, "step": 2109 }, { "epoch": 2.0249639942390782, "grad_norm": 1.175646040665266, "learning_rate": 7.427123939394658e-06, "loss": 0.5791, "step": 2110 }, { "epoch": 2.025924147863658, "grad_norm": 1.1749759148561074, "learning_rate": 7.424194596930925e-06, "loss": 0.512, "step": 2111 }, { "epoch": 2.0268843014882383, "grad_norm": 0.7995089694872192, "learning_rate": 7.421264166272152e-06, "loss": 0.5429, "step": 2112 }, { "epoch": 2.027844455112818, "grad_norm": 1.2543459774815198, "learning_rate": 7.418332648733774e-06, "loss": 0.6604, "step": 2113 }, { "epoch": 2.028804608737398, "grad_norm": 0.8358309185931371, "learning_rate": 7.415400045631723e-06, "loss": 0.5589, "step": 2114 }, { "epoch": 2.029764762361978, "grad_norm": 1.0289735136890514, "learning_rate": 7.41246635828241e-06, "loss": 0.5534, "step": 2115 }, { "epoch": 2.030724915986558, "grad_norm": 0.9706802452152404, "learning_rate": 7.4095315880027365e-06, "loss": 0.4843, "step": 2116 }, { "epoch": 2.0316850696111377, "grad_norm": 0.762609295982512, "learning_rate": 7.406595736110089e-06, "loss": 0.5203, "step": 2117 }, { "epoch": 2.0326452232357175, "grad_norm": 1.1204003260732456, "learning_rate": 7.4036588039223415e-06, "loss": 0.4793, "step": 2118 }, { "epoch": 2.0336053768602977, "grad_norm": 0.9187701766770359, "learning_rate": 7.400720792757851e-06, "loss": 0.5674, "step": 2119 }, { "epoch": 2.0345655304848775, "grad_norm": 0.7638989434367952, "learning_rate": 7.397781703935458e-06, "loss": 0.5521, "step": 2120 }, { "epoch": 2.0355256841094573, "grad_norm": 0.934155309636016, "learning_rate": 7.39484153877449e-06, "loss": 0.609, "step": 2121 }, { "epoch": 2.0364858377340376, "grad_norm": 0.9615532920669503, "learning_rate": 7.391900298594752e-06, "loss": 0.4831, "step": 2122 }, { "epoch": 2.0374459913586174, "grad_norm": 1.070484881688937, "learning_rate": 7.388957984716541e-06, "loss": 0.5189, "step": 2123 }, { "epoch": 2.038406144983197, "grad_norm": 0.8952414389848654, "learning_rate": 7.386014598460626e-06, "loss": 0.4167, "step": 2124 }, { "epoch": 2.0393662986077774, "grad_norm": 0.9014216053127138, "learning_rate": 7.383070141148263e-06, "loss": 0.598, "step": 2125 }, { "epoch": 2.040326452232357, "grad_norm": 0.7752497140406029, "learning_rate": 7.380124614101185e-06, "loss": 0.4828, "step": 2126 }, { "epoch": 2.041286605856937, "grad_norm": 0.913460367200707, "learning_rate": 7.377178018641613e-06, "loss": 0.5273, "step": 2127 }, { "epoch": 2.0422467594815172, "grad_norm": 1.2602850511829389, "learning_rate": 7.374230356092238e-06, "loss": 0.5862, "step": 2128 }, { "epoch": 2.043206913106097, "grad_norm": 0.7799399374145033, "learning_rate": 7.371281627776237e-06, "loss": 0.5707, "step": 2129 }, { "epoch": 2.044167066730677, "grad_norm": 0.9488807497474782, "learning_rate": 7.368331835017257e-06, "loss": 0.6102, "step": 2130 }, { "epoch": 2.0451272203552566, "grad_norm": 0.9241995600050251, "learning_rate": 7.365380979139435e-06, "loss": 0.5709, "step": 2131 }, { "epoch": 2.046087373979837, "grad_norm": 0.7638532589317082, "learning_rate": 7.362429061467377e-06, "loss": 0.5767, "step": 2132 }, { "epoch": 2.0470475276044167, "grad_norm": 1.203345750377053, "learning_rate": 7.359476083326164e-06, "loss": 0.4705, "step": 2133 }, { "epoch": 2.0480076812289965, "grad_norm": 1.0440960157917591, "learning_rate": 7.35652204604136e-06, "loss": 0.6119, "step": 2134 }, { "epoch": 2.0489678348535767, "grad_norm": 0.9992897329228346, "learning_rate": 7.353566950938997e-06, "loss": 0.584, "step": 2135 }, { "epoch": 2.0499279884781565, "grad_norm": 0.9094704970456408, "learning_rate": 7.35061079934559e-06, "loss": 0.52, "step": 2136 }, { "epoch": 2.0508881421027363, "grad_norm": 1.548740651183021, "learning_rate": 7.347653592588122e-06, "loss": 0.6874, "step": 2137 }, { "epoch": 2.0518482957273165, "grad_norm": 0.6869405757745318, "learning_rate": 7.344695331994052e-06, "loss": 0.5257, "step": 2138 }, { "epoch": 2.0528084493518963, "grad_norm": 1.1047195124817815, "learning_rate": 7.341736018891308e-06, "loss": 0.6253, "step": 2139 }, { "epoch": 2.053768602976476, "grad_norm": 1.0089414296589687, "learning_rate": 7.338775654608299e-06, "loss": 0.662, "step": 2140 }, { "epoch": 2.0547287566010564, "grad_norm": 0.9555945309749319, "learning_rate": 7.335814240473899e-06, "loss": 0.5657, "step": 2141 }, { "epoch": 2.055688910225636, "grad_norm": 1.0297268883093358, "learning_rate": 7.332851777817453e-06, "loss": 0.6257, "step": 2142 }, { "epoch": 2.056649063850216, "grad_norm": 0.8358315888830901, "learning_rate": 7.3298882679687836e-06, "loss": 0.5298, "step": 2143 }, { "epoch": 2.0576092174747957, "grad_norm": 1.0272001049026258, "learning_rate": 7.326923712258173e-06, "loss": 0.5633, "step": 2144 }, { "epoch": 2.058569371099376, "grad_norm": 0.8064550638913437, "learning_rate": 7.323958112016383e-06, "loss": 0.5396, "step": 2145 }, { "epoch": 2.059529524723956, "grad_norm": 0.7346254900295881, "learning_rate": 7.320991468574637e-06, "loss": 0.5652, "step": 2146 }, { "epoch": 2.0604896783485356, "grad_norm": 1.050620498042139, "learning_rate": 7.318023783264633e-06, "loss": 0.5746, "step": 2147 }, { "epoch": 2.061449831973116, "grad_norm": 0.9001365533180316, "learning_rate": 7.315055057418528e-06, "loss": 0.6319, "step": 2148 }, { "epoch": 2.0624099855976956, "grad_norm": 0.9087868741118574, "learning_rate": 7.312085292368957e-06, "loss": 0.5782, "step": 2149 }, { "epoch": 2.0633701392222754, "grad_norm": 1.0032907798965676, "learning_rate": 7.309114489449012e-06, "loss": 0.5158, "step": 2150 }, { "epoch": 2.0643302928468557, "grad_norm": 0.8047333020743894, "learning_rate": 7.306142649992256e-06, "loss": 0.5238, "step": 2151 }, { "epoch": 2.0652904464714354, "grad_norm": 0.8923640961315871, "learning_rate": 7.3031697753327145e-06, "loss": 0.6236, "step": 2152 }, { "epoch": 2.0662506000960152, "grad_norm": 0.7410555047191457, "learning_rate": 7.300195866804881e-06, "loss": 0.5286, "step": 2153 }, { "epoch": 2.0672107537205955, "grad_norm": 0.9183193095221048, "learning_rate": 7.2972209257437125e-06, "loss": 0.5724, "step": 2154 }, { "epoch": 2.0681709073451753, "grad_norm": 0.9950013517507624, "learning_rate": 7.2942449534846234e-06, "loss": 0.581, "step": 2155 }, { "epoch": 2.069131060969755, "grad_norm": 1.2364710693887315, "learning_rate": 7.291267951363501e-06, "loss": 0.6224, "step": 2156 }, { "epoch": 2.070091214594335, "grad_norm": 0.8175580570495983, "learning_rate": 7.288289920716686e-06, "loss": 0.495, "step": 2157 }, { "epoch": 2.071051368218915, "grad_norm": 1.0044765546316299, "learning_rate": 7.285310862880987e-06, "loss": 0.634, "step": 2158 }, { "epoch": 2.072011521843495, "grad_norm": 0.8483715448465391, "learning_rate": 7.28233077919367e-06, "loss": 0.5729, "step": 2159 }, { "epoch": 2.0729716754680747, "grad_norm": 0.7295755591686811, "learning_rate": 7.279349670992464e-06, "loss": 0.4986, "step": 2160 }, { "epoch": 2.073931829092655, "grad_norm": 0.8105721076425058, "learning_rate": 7.2763675396155545e-06, "loss": 0.5624, "step": 2161 }, { "epoch": 2.0748919827172347, "grad_norm": 0.7719437566240688, "learning_rate": 7.27338438640159e-06, "loss": 0.5407, "step": 2162 }, { "epoch": 2.0758521363418145, "grad_norm": 0.7276808182296286, "learning_rate": 7.270400212689675e-06, "loss": 0.5719, "step": 2163 }, { "epoch": 2.0768122899663948, "grad_norm": 0.8077643012522563, "learning_rate": 7.2674150198193735e-06, "loss": 0.5558, "step": 2164 }, { "epoch": 2.0777724435909746, "grad_norm": 0.7334645505753414, "learning_rate": 7.264428809130708e-06, "loss": 0.5756, "step": 2165 }, { "epoch": 2.0787325972155544, "grad_norm": 0.9409190007913237, "learning_rate": 7.2614415819641535e-06, "loss": 0.5321, "step": 2166 }, { "epoch": 2.0796927508401346, "grad_norm": 0.7842425567068839, "learning_rate": 7.258453339660648e-06, "loss": 0.5005, "step": 2167 }, { "epoch": 2.0806529044647144, "grad_norm": 0.852211263607617, "learning_rate": 7.25546408356158e-06, "loss": 0.5437, "step": 2168 }, { "epoch": 2.081613058089294, "grad_norm": 0.9681592264916435, "learning_rate": 7.252473815008794e-06, "loss": 0.4967, "step": 2169 }, { "epoch": 2.082573211713874, "grad_norm": 1.238369578646625, "learning_rate": 7.24948253534459e-06, "loss": 0.5977, "step": 2170 }, { "epoch": 2.0835333653384542, "grad_norm": 0.7138585190550476, "learning_rate": 7.246490245911722e-06, "loss": 0.5259, "step": 2171 }, { "epoch": 2.084493518963034, "grad_norm": 1.4460819671851026, "learning_rate": 7.243496948053394e-06, "loss": 0.5983, "step": 2172 }, { "epoch": 2.085453672587614, "grad_norm": 1.0834667313308408, "learning_rate": 7.240502643113271e-06, "loss": 0.4862, "step": 2173 }, { "epoch": 2.086413826212194, "grad_norm": 1.07301148270314, "learning_rate": 7.237507332435459e-06, "loss": 0.5356, "step": 2174 }, { "epoch": 2.087373979836774, "grad_norm": 0.9143853292286057, "learning_rate": 7.234511017364524e-06, "loss": 0.5716, "step": 2175 }, { "epoch": 2.0883341334613537, "grad_norm": 0.9435461494442817, "learning_rate": 7.2315136992454794e-06, "loss": 0.6431, "step": 2176 }, { "epoch": 2.089294287085934, "grad_norm": 1.1098814345032535, "learning_rate": 7.228515379423787e-06, "loss": 0.6069, "step": 2177 }, { "epoch": 2.0902544407105137, "grad_norm": 0.9747418566314479, "learning_rate": 7.225516059245365e-06, "loss": 0.6089, "step": 2178 }, { "epoch": 2.0912145943350935, "grad_norm": 0.9060472082866442, "learning_rate": 7.222515740056571e-06, "loss": 0.5863, "step": 2179 }, { "epoch": 2.0921747479596737, "grad_norm": 0.8031372845761876, "learning_rate": 7.219514423204219e-06, "loss": 0.5061, "step": 2180 }, { "epoch": 2.0931349015842535, "grad_norm": 0.7583681973477342, "learning_rate": 7.216512110035567e-06, "loss": 0.5531, "step": 2181 }, { "epoch": 2.0940950552088333, "grad_norm": 0.8391143934722556, "learning_rate": 7.213508801898325e-06, "loss": 0.5635, "step": 2182 }, { "epoch": 2.095055208833413, "grad_norm": 0.8309698225759319, "learning_rate": 7.21050450014064e-06, "loss": 0.5551, "step": 2183 }, { "epoch": 2.0960153624579934, "grad_norm": 0.9775059977148546, "learning_rate": 7.2074992061111145e-06, "loss": 0.4316, "step": 2184 }, { "epoch": 2.096975516082573, "grad_norm": 0.7565835002187093, "learning_rate": 7.204492921158794e-06, "loss": 0.5473, "step": 2185 }, { "epoch": 2.097935669707153, "grad_norm": 0.7642644725454123, "learning_rate": 7.201485646633164e-06, "loss": 0.4926, "step": 2186 }, { "epoch": 2.098895823331733, "grad_norm": 0.8176121571861301, "learning_rate": 7.198477383884161e-06, "loss": 0.589, "step": 2187 }, { "epoch": 2.099855976956313, "grad_norm": 0.6743312867304472, "learning_rate": 7.195468134262159e-06, "loss": 0.5312, "step": 2188 }, { "epoch": 2.100816130580893, "grad_norm": 0.8876509700978312, "learning_rate": 7.1924578991179815e-06, "loss": 0.4494, "step": 2189 }, { "epoch": 2.101776284205473, "grad_norm": 1.2367780383817584, "learning_rate": 7.189446679802889e-06, "loss": 0.6107, "step": 2190 }, { "epoch": 2.102736437830053, "grad_norm": 0.7540829302569384, "learning_rate": 7.186434477668586e-06, "loss": 0.4923, "step": 2191 }, { "epoch": 2.1036965914546326, "grad_norm": 0.9725761921459787, "learning_rate": 7.1834212940672155e-06, "loss": 0.533, "step": 2192 }, { "epoch": 2.104656745079213, "grad_norm": 1.4399847595110091, "learning_rate": 7.180407130351368e-06, "loss": 0.5716, "step": 2193 }, { "epoch": 2.1056168987037926, "grad_norm": 1.0515949843662522, "learning_rate": 7.177391987874065e-06, "loss": 0.6781, "step": 2194 }, { "epoch": 2.1065770523283724, "grad_norm": 0.6866164632102819, "learning_rate": 7.174375867988776e-06, "loss": 0.43, "step": 2195 }, { "epoch": 2.1075372059529527, "grad_norm": 0.68922549112802, "learning_rate": 7.171358772049402e-06, "loss": 0.4617, "step": 2196 }, { "epoch": 2.1084973595775325, "grad_norm": 1.1355217122097951, "learning_rate": 7.168340701410284e-06, "loss": 0.5831, "step": 2197 }, { "epoch": 2.1094575132021123, "grad_norm": 0.9942782469931802, "learning_rate": 7.165321657426205e-06, "loss": 0.6216, "step": 2198 }, { "epoch": 2.110417666826692, "grad_norm": 1.0365017619580432, "learning_rate": 7.1623016414523795e-06, "loss": 0.4429, "step": 2199 }, { "epoch": 2.1113778204512723, "grad_norm": 0.7804924949473452, "learning_rate": 7.159280654844462e-06, "loss": 0.6233, "step": 2200 }, { "epoch": 2.112337974075852, "grad_norm": 0.7584154183792926, "learning_rate": 7.156258698958537e-06, "loss": 0.5898, "step": 2201 }, { "epoch": 2.113298127700432, "grad_norm": 0.8414522799597683, "learning_rate": 7.153235775151134e-06, "loss": 0.4837, "step": 2202 }, { "epoch": 2.114258281325012, "grad_norm": 1.4072361460444907, "learning_rate": 7.150211884779208e-06, "loss": 0.5033, "step": 2203 }, { "epoch": 2.115218434949592, "grad_norm": 1.24613679980461, "learning_rate": 7.147187029200151e-06, "loss": 0.5917, "step": 2204 }, { "epoch": 2.1161785885741717, "grad_norm": 0.9486449243722676, "learning_rate": 7.144161209771788e-06, "loss": 0.5184, "step": 2205 }, { "epoch": 2.117138742198752, "grad_norm": 1.1676778665080763, "learning_rate": 7.141134427852378e-06, "loss": 0.4725, "step": 2206 }, { "epoch": 2.1180988958233318, "grad_norm": 0.9383775341059482, "learning_rate": 7.138106684800612e-06, "loss": 0.521, "step": 2207 }, { "epoch": 2.1190590494479116, "grad_norm": 1.1961365444523733, "learning_rate": 7.1350779819756085e-06, "loss": 0.5319, "step": 2208 }, { "epoch": 2.120019203072492, "grad_norm": 1.5587243205281245, "learning_rate": 7.132048320736924e-06, "loss": 0.6321, "step": 2209 }, { "epoch": 2.1209793566970716, "grad_norm": 0.9344323423403369, "learning_rate": 7.129017702444536e-06, "loss": 0.5632, "step": 2210 }, { "epoch": 2.1219395103216514, "grad_norm": 1.037161083885332, "learning_rate": 7.12598612845886e-06, "loss": 0.4804, "step": 2211 }, { "epoch": 2.122899663946231, "grad_norm": 0.8307980264877128, "learning_rate": 7.1229536001407376e-06, "loss": 0.5894, "step": 2212 }, { "epoch": 2.1238598175708114, "grad_norm": 0.8485691517062898, "learning_rate": 7.119920118851438e-06, "loss": 0.548, "step": 2213 }, { "epoch": 2.1248199711953912, "grad_norm": 1.0356073846741651, "learning_rate": 7.116885685952655e-06, "loss": 0.5878, "step": 2214 }, { "epoch": 2.125780124819971, "grad_norm": 0.8040891477630738, "learning_rate": 7.113850302806518e-06, "loss": 0.5689, "step": 2215 }, { "epoch": 2.1267402784445513, "grad_norm": 0.831654781563631, "learning_rate": 7.110813970775574e-06, "loss": 0.5214, "step": 2216 }, { "epoch": 2.127700432069131, "grad_norm": 0.8682717933217237, "learning_rate": 7.107776691222803e-06, "loss": 0.4983, "step": 2217 }, { "epoch": 2.128660585693711, "grad_norm": 0.8432328954164495, "learning_rate": 7.104738465511606e-06, "loss": 0.6504, "step": 2218 }, { "epoch": 2.129620739318291, "grad_norm": 0.6725013563999659, "learning_rate": 7.1016992950058104e-06, "loss": 0.5132, "step": 2219 }, { "epoch": 2.130580892942871, "grad_norm": 0.8264027473295606, "learning_rate": 7.098659181069667e-06, "loss": 0.525, "step": 2220 }, { "epoch": 2.1315410465674507, "grad_norm": 1.0637577469559023, "learning_rate": 7.0956181250678514e-06, "loss": 0.6335, "step": 2221 }, { "epoch": 2.132501200192031, "grad_norm": 0.7923330372834642, "learning_rate": 7.09257612836546e-06, "loss": 0.5533, "step": 2222 }, { "epoch": 2.1334613538166107, "grad_norm": 0.9134243218547725, "learning_rate": 7.089533192328011e-06, "loss": 0.5064, "step": 2223 }, { "epoch": 2.1344215074411905, "grad_norm": 0.8093243124011293, "learning_rate": 7.08648931832145e-06, "loss": 0.6355, "step": 2224 }, { "epoch": 2.1353816610657703, "grad_norm": 0.8135020461770336, "learning_rate": 7.0834445077121374e-06, "loss": 0.5999, "step": 2225 }, { "epoch": 2.1363418146903506, "grad_norm": 0.8972325635326398, "learning_rate": 7.080398761866856e-06, "loss": 0.6237, "step": 2226 }, { "epoch": 2.1373019683149304, "grad_norm": 1.0127395212828405, "learning_rate": 7.077352082152809e-06, "loss": 0.56, "step": 2227 }, { "epoch": 2.13826212193951, "grad_norm": 0.8880618621862372, "learning_rate": 7.0743044699376165e-06, "loss": 0.557, "step": 2228 }, { "epoch": 2.1392222755640904, "grad_norm": 0.7843700263614602, "learning_rate": 7.0712559265893214e-06, "loss": 0.4966, "step": 2229 }, { "epoch": 2.14018242918867, "grad_norm": 0.8496837074237693, "learning_rate": 7.068206453476381e-06, "loss": 0.5005, "step": 2230 }, { "epoch": 2.14114258281325, "grad_norm": 1.0477200273618636, "learning_rate": 7.065156051967672e-06, "loss": 0.6208, "step": 2231 }, { "epoch": 2.1421027364378302, "grad_norm": 0.8052718978711361, "learning_rate": 7.062104723432485e-06, "loss": 0.5346, "step": 2232 }, { "epoch": 2.14306289006241, "grad_norm": 1.09607774238334, "learning_rate": 7.059052469240533e-06, "loss": 0.6506, "step": 2233 }, { "epoch": 2.14402304368699, "grad_norm": 0.8866539015198924, "learning_rate": 7.0559992907619346e-06, "loss": 0.5185, "step": 2234 }, { "epoch": 2.14498319731157, "grad_norm": 0.7316271301458286, "learning_rate": 7.0529451893672356e-06, "loss": 0.5091, "step": 2235 }, { "epoch": 2.14594335093615, "grad_norm": 1.0084017576316693, "learning_rate": 7.0498901664273825e-06, "loss": 0.5415, "step": 2236 }, { "epoch": 2.1469035045607296, "grad_norm": 0.7429490120788453, "learning_rate": 7.046834223313746e-06, "loss": 0.4289, "step": 2237 }, { "epoch": 2.1478636581853094, "grad_norm": 1.032338292868619, "learning_rate": 7.043777361398108e-06, "loss": 0.5856, "step": 2238 }, { "epoch": 2.1488238118098897, "grad_norm": 0.8695382804288025, "learning_rate": 7.040719582052659e-06, "loss": 0.5042, "step": 2239 }, { "epoch": 2.1497839654344695, "grad_norm": 1.1604333237105544, "learning_rate": 7.037660886650003e-06, "loss": 0.6335, "step": 2240 }, { "epoch": 2.1507441190590493, "grad_norm": 1.2853505719966218, "learning_rate": 7.034601276563155e-06, "loss": 0.6418, "step": 2241 }, { "epoch": 2.1517042726836295, "grad_norm": 0.7952218958324107, "learning_rate": 7.031540753165545e-06, "loss": 0.5186, "step": 2242 }, { "epoch": 2.1526644263082093, "grad_norm": 1.0186730490946365, "learning_rate": 7.028479317831006e-06, "loss": 0.5755, "step": 2243 }, { "epoch": 2.153624579932789, "grad_norm": 0.8345576795396213, "learning_rate": 7.0254169719337864e-06, "loss": 0.5947, "step": 2244 }, { "epoch": 2.1545847335573693, "grad_norm": 0.9005493405578582, "learning_rate": 7.022353716848535e-06, "loss": 0.6245, "step": 2245 }, { "epoch": 2.155544887181949, "grad_norm": 0.8343209885853908, "learning_rate": 7.0192895539503205e-06, "loss": 0.486, "step": 2246 }, { "epoch": 2.156505040806529, "grad_norm": 0.7527412898643225, "learning_rate": 7.016224484614609e-06, "loss": 0.5324, "step": 2247 }, { "epoch": 2.157465194431109, "grad_norm": 0.7557484868655571, "learning_rate": 7.01315851021728e-06, "loss": 0.5355, "step": 2248 }, { "epoch": 2.158425348055689, "grad_norm": 1.8594937523518051, "learning_rate": 7.010091632134612e-06, "loss": 0.6077, "step": 2249 }, { "epoch": 2.1593855016802688, "grad_norm": 1.280114213307682, "learning_rate": 7.0070238517433e-06, "loss": 0.6584, "step": 2250 }, { "epoch": 2.160345655304849, "grad_norm": 0.911424371293422, "learning_rate": 7.003955170420434e-06, "loss": 0.4691, "step": 2251 }, { "epoch": 2.161305808929429, "grad_norm": 0.8331514183524391, "learning_rate": 7.000885589543513e-06, "loss": 0.5717, "step": 2252 }, { "epoch": 2.1622659625540086, "grad_norm": 0.980269446069045, "learning_rate": 6.99781511049044e-06, "loss": 0.6059, "step": 2253 }, { "epoch": 2.1632261161785884, "grad_norm": 0.951182358611006, "learning_rate": 6.994743734639517e-06, "loss": 0.5473, "step": 2254 }, { "epoch": 2.1641862698031686, "grad_norm": 0.7547614092281096, "learning_rate": 6.991671463369457e-06, "loss": 0.4975, "step": 2255 }, { "epoch": 2.1651464234277484, "grad_norm": 0.8798739537035324, "learning_rate": 6.988598298059366e-06, "loss": 0.6039, "step": 2256 }, { "epoch": 2.1661065770523282, "grad_norm": 0.7885006969042282, "learning_rate": 6.985524240088756e-06, "loss": 0.5453, "step": 2257 }, { "epoch": 2.1670667306769085, "grad_norm": 0.8015460989184183, "learning_rate": 6.982449290837538e-06, "loss": 0.5484, "step": 2258 }, { "epoch": 2.1680268843014883, "grad_norm": 0.7432124678066201, "learning_rate": 6.979373451686025e-06, "loss": 0.5264, "step": 2259 }, { "epoch": 2.168987037926068, "grad_norm": 0.8091685289608697, "learning_rate": 6.976296724014927e-06, "loss": 0.6227, "step": 2260 }, { "epoch": 2.1699471915506483, "grad_norm": 1.335387769295398, "learning_rate": 6.973219109205356e-06, "loss": 0.5738, "step": 2261 }, { "epoch": 2.170907345175228, "grad_norm": 0.9569211125681906, "learning_rate": 6.970140608638818e-06, "loss": 0.5432, "step": 2262 }, { "epoch": 2.171867498799808, "grad_norm": 0.8670437632164834, "learning_rate": 6.967061223697223e-06, "loss": 0.6153, "step": 2263 }, { "epoch": 2.172827652424388, "grad_norm": 1.0183565110168489, "learning_rate": 6.963980955762869e-06, "loss": 0.4742, "step": 2264 }, { "epoch": 2.173787806048968, "grad_norm": 0.9391584496445098, "learning_rate": 6.960899806218458e-06, "loss": 0.5909, "step": 2265 }, { "epoch": 2.1747479596735477, "grad_norm": 0.7742353646791891, "learning_rate": 6.957817776447087e-06, "loss": 0.5048, "step": 2266 }, { "epoch": 2.1757081132981275, "grad_norm": 0.7286370836356276, "learning_rate": 6.954734867832241e-06, "loss": 0.5439, "step": 2267 }, { "epoch": 2.1766682669227078, "grad_norm": 0.856156592854698, "learning_rate": 6.9516510817578105e-06, "loss": 0.5311, "step": 2268 }, { "epoch": 2.1776284205472876, "grad_norm": 0.8457842115685432, "learning_rate": 6.948566419608069e-06, "loss": 0.5057, "step": 2269 }, { "epoch": 2.1785885741718674, "grad_norm": 0.8606064140531403, "learning_rate": 6.945480882767694e-06, "loss": 0.5949, "step": 2270 }, { "epoch": 2.1795487277964476, "grad_norm": 0.8367766838357528, "learning_rate": 6.942394472621744e-06, "loss": 0.502, "step": 2271 }, { "epoch": 2.1805088814210274, "grad_norm": 1.0044427296428011, "learning_rate": 6.939307190555679e-06, "loss": 0.6425, "step": 2272 }, { "epoch": 2.181469035045607, "grad_norm": 0.8795126406714183, "learning_rate": 6.936219037955347e-06, "loss": 0.5986, "step": 2273 }, { "epoch": 2.1824291886701874, "grad_norm": 0.6879521307770827, "learning_rate": 6.9331300162069846e-06, "loss": 0.5242, "step": 2274 }, { "epoch": 2.183389342294767, "grad_norm": 0.9622141507904436, "learning_rate": 6.930040126697224e-06, "loss": 0.5884, "step": 2275 }, { "epoch": 2.184349495919347, "grad_norm": 0.9041096138526464, "learning_rate": 6.926949370813079e-06, "loss": 0.616, "step": 2276 }, { "epoch": 2.1853096495439273, "grad_norm": 0.8257305024697136, "learning_rate": 6.92385774994196e-06, "loss": 0.609, "step": 2277 }, { "epoch": 2.186269803168507, "grad_norm": 0.9507735279083417, "learning_rate": 6.920765265471663e-06, "loss": 0.7154, "step": 2278 }, { "epoch": 2.187229956793087, "grad_norm": 1.061799167313272, "learning_rate": 6.91767191879037e-06, "loss": 0.5834, "step": 2279 }, { "epoch": 2.1881901104176666, "grad_norm": 0.8447277184608661, "learning_rate": 6.9145777112866495e-06, "loss": 0.6125, "step": 2280 }, { "epoch": 2.189150264042247, "grad_norm": 0.8071625094509312, "learning_rate": 6.911482644349462e-06, "loss": 0.5405, "step": 2281 }, { "epoch": 2.1901104176668267, "grad_norm": 0.8408147800268508, "learning_rate": 6.908386719368147e-06, "loss": 0.6234, "step": 2282 }, { "epoch": 2.1910705712914065, "grad_norm": 0.8341328878326141, "learning_rate": 6.9052899377324345e-06, "loss": 0.5395, "step": 2283 }, { "epoch": 2.1920307249159867, "grad_norm": 0.8461156675047746, "learning_rate": 6.902192300832435e-06, "loss": 0.5201, "step": 2284 }, { "epoch": 2.1929908785405665, "grad_norm": 0.7930594138040835, "learning_rate": 6.899093810058643e-06, "loss": 0.494, "step": 2285 }, { "epoch": 2.1939510321651463, "grad_norm": 1.1674138036999815, "learning_rate": 6.895994466801943e-06, "loss": 0.4914, "step": 2286 }, { "epoch": 2.1949111857897265, "grad_norm": 0.9111474894700454, "learning_rate": 6.892894272453592e-06, "loss": 0.5888, "step": 2287 }, { "epoch": 2.1958713394143063, "grad_norm": 1.2345130562808382, "learning_rate": 6.889793228405237e-06, "loss": 0.5731, "step": 2288 }, { "epoch": 2.196831493038886, "grad_norm": 0.7342428043692181, "learning_rate": 6.8866913360489005e-06, "loss": 0.5413, "step": 2289 }, { "epoch": 2.1977916466634664, "grad_norm": 0.7455163275316943, "learning_rate": 6.883588596776994e-06, "loss": 0.5458, "step": 2290 }, { "epoch": 2.198751800288046, "grad_norm": 0.9959968367861125, "learning_rate": 6.8804850119822986e-06, "loss": 0.5992, "step": 2291 }, { "epoch": 2.199711953912626, "grad_norm": 0.846547923941351, "learning_rate": 6.8773805830579845e-06, "loss": 0.6129, "step": 2292 }, { "epoch": 2.2006721075372058, "grad_norm": 0.9733445516928936, "learning_rate": 6.8742753113975925e-06, "loss": 0.59, "step": 2293 }, { "epoch": 2.201632261161786, "grad_norm": 1.1846148708418383, "learning_rate": 6.871169198395049e-06, "loss": 0.4928, "step": 2294 }, { "epoch": 2.202592414786366, "grad_norm": 0.6707959708534591, "learning_rate": 6.868062245444655e-06, "loss": 0.5621, "step": 2295 }, { "epoch": 2.2035525684109456, "grad_norm": 0.8403487016665677, "learning_rate": 6.864954453941086e-06, "loss": 0.5042, "step": 2296 }, { "epoch": 2.204512722035526, "grad_norm": 1.1227483295684828, "learning_rate": 6.861845825279399e-06, "loss": 0.583, "step": 2297 }, { "epoch": 2.2054728756601056, "grad_norm": 0.9137573073198969, "learning_rate": 6.85873636085502e-06, "loss": 0.6171, "step": 2298 }, { "epoch": 2.2064330292846854, "grad_norm": 0.7752319643247556, "learning_rate": 6.855626062063761e-06, "loss": 0.6168, "step": 2299 }, { "epoch": 2.2073931829092657, "grad_norm": 1.3283903784325413, "learning_rate": 6.852514930301797e-06, "loss": 0.6487, "step": 2300 }, { "epoch": 2.2083533365338455, "grad_norm": 1.3821254216782837, "learning_rate": 6.8494029669656845e-06, "loss": 0.6148, "step": 2301 }, { "epoch": 2.2093134901584253, "grad_norm": 0.9756323186742466, "learning_rate": 6.846290173452348e-06, "loss": 0.5285, "step": 2302 }, { "epoch": 2.2102736437830055, "grad_norm": 1.1152880996306052, "learning_rate": 6.843176551159087e-06, "loss": 0.5955, "step": 2303 }, { "epoch": 2.2112337974075853, "grad_norm": 0.8561168705143323, "learning_rate": 6.8400621014835775e-06, "loss": 0.6121, "step": 2304 }, { "epoch": 2.212193951032165, "grad_norm": 0.8413935791195933, "learning_rate": 6.836946825823858e-06, "loss": 0.6522, "step": 2305 }, { "epoch": 2.213154104656745, "grad_norm": 0.9566627885726511, "learning_rate": 6.833830725578346e-06, "loss": 0.5287, "step": 2306 }, { "epoch": 2.214114258281325, "grad_norm": 1.19693767783819, "learning_rate": 6.83071380214582e-06, "loss": 0.6434, "step": 2307 }, { "epoch": 2.215074411905905, "grad_norm": 0.9451639288119044, "learning_rate": 6.827596056925441e-06, "loss": 0.6391, "step": 2308 }, { "epoch": 2.2160345655304847, "grad_norm": 0.8389462942964453, "learning_rate": 6.824477491316725e-06, "loss": 0.4956, "step": 2309 }, { "epoch": 2.216994719155065, "grad_norm": 1.198730191946927, "learning_rate": 6.821358106719569e-06, "loss": 0.4456, "step": 2310 }, { "epoch": 2.2179548727796448, "grad_norm": 0.9124918452113085, "learning_rate": 6.818237904534226e-06, "loss": 0.497, "step": 2311 }, { "epoch": 2.2189150264042246, "grad_norm": 0.9458151034537332, "learning_rate": 6.815116886161324e-06, "loss": 0.622, "step": 2312 }, { "epoch": 2.219875180028805, "grad_norm": 1.104442523660407, "learning_rate": 6.811995053001854e-06, "loss": 0.5058, "step": 2313 }, { "epoch": 2.2208353336533846, "grad_norm": 1.2040035207615456, "learning_rate": 6.808872406457175e-06, "loss": 0.6454, "step": 2314 }, { "epoch": 2.2217954872779644, "grad_norm": 0.9139475384514517, "learning_rate": 6.8057489479290085e-06, "loss": 0.4771, "step": 2315 }, { "epoch": 2.2227556409025446, "grad_norm": 1.433747254400768, "learning_rate": 6.802624678819441e-06, "loss": 0.4405, "step": 2316 }, { "epoch": 2.2237157945271244, "grad_norm": 0.8800878770850824, "learning_rate": 6.799499600530925e-06, "loss": 0.564, "step": 2317 }, { "epoch": 2.224675948151704, "grad_norm": 0.8299337660277917, "learning_rate": 6.796373714466272e-06, "loss": 0.6565, "step": 2318 }, { "epoch": 2.225636101776284, "grad_norm": 0.8282795617419546, "learning_rate": 6.793247022028663e-06, "loss": 0.5249, "step": 2319 }, { "epoch": 2.2265962554008643, "grad_norm": 0.9750794300000941, "learning_rate": 6.790119524621632e-06, "loss": 0.6108, "step": 2320 }, { "epoch": 2.227556409025444, "grad_norm": 1.0755542404779788, "learning_rate": 6.786991223649084e-06, "loss": 0.5901, "step": 2321 }, { "epoch": 2.228516562650024, "grad_norm": 1.0907624608529571, "learning_rate": 6.783862120515277e-06, "loss": 0.6587, "step": 2322 }, { "epoch": 2.229476716274604, "grad_norm": 0.9178478846434833, "learning_rate": 6.780732216624833e-06, "loss": 0.5431, "step": 2323 }, { "epoch": 2.230436869899184, "grad_norm": 0.8433899639799032, "learning_rate": 6.77760151338273e-06, "loss": 0.5262, "step": 2324 }, { "epoch": 2.2313970235237637, "grad_norm": 1.03458563402818, "learning_rate": 6.7744700121943095e-06, "loss": 0.6198, "step": 2325 }, { "epoch": 2.232357177148344, "grad_norm": 0.7738440747278639, "learning_rate": 6.77133771446527e-06, "loss": 0.5588, "step": 2326 }, { "epoch": 2.2333173307729237, "grad_norm": 0.6683792601686276, "learning_rate": 6.768204621601665e-06, "loss": 0.5181, "step": 2327 }, { "epoch": 2.2342774843975035, "grad_norm": 0.9918437327785985, "learning_rate": 6.765070735009906e-06, "loss": 0.469, "step": 2328 }, { "epoch": 2.2352376380220838, "grad_norm": 0.9927723848164123, "learning_rate": 6.761936056096761e-06, "loss": 0.5557, "step": 2329 }, { "epoch": 2.2361977916466635, "grad_norm": 1.1233198065358125, "learning_rate": 6.758800586269355e-06, "loss": 0.5608, "step": 2330 }, { "epoch": 2.2371579452712433, "grad_norm": 0.7134104115079637, "learning_rate": 6.755664326935167e-06, "loss": 0.5083, "step": 2331 }, { "epoch": 2.238118098895823, "grad_norm": 0.8028697068865261, "learning_rate": 6.752527279502029e-06, "loss": 0.5285, "step": 2332 }, { "epoch": 2.2390782525204034, "grad_norm": 0.9536140012416103, "learning_rate": 6.749389445378127e-06, "loss": 0.5678, "step": 2333 }, { "epoch": 2.240038406144983, "grad_norm": 1.2261772787421286, "learning_rate": 6.746250825972005e-06, "loss": 0.5963, "step": 2334 }, { "epoch": 2.240998559769563, "grad_norm": 0.8197472611035069, "learning_rate": 6.743111422692553e-06, "loss": 0.5669, "step": 2335 }, { "epoch": 2.241958713394143, "grad_norm": 0.958213688226831, "learning_rate": 6.739971236949017e-06, "loss": 0.5398, "step": 2336 }, { "epoch": 2.242918867018723, "grad_norm": 0.8985653616689689, "learning_rate": 6.736830270150992e-06, "loss": 0.5419, "step": 2337 }, { "epoch": 2.243879020643303, "grad_norm": 0.7915850975182945, "learning_rate": 6.733688523708422e-06, "loss": 0.5182, "step": 2338 }, { "epoch": 2.244839174267883, "grad_norm": 0.8760492591181075, "learning_rate": 6.7305459990316094e-06, "loss": 0.6072, "step": 2339 }, { "epoch": 2.245799327892463, "grad_norm": 1.375554334849348, "learning_rate": 6.727402697531193e-06, "loss": 0.5638, "step": 2340 }, { "epoch": 2.2467594815170426, "grad_norm": 0.9094262723912077, "learning_rate": 6.7242586206181735e-06, "loss": 0.5128, "step": 2341 }, { "epoch": 2.247719635141623, "grad_norm": 0.9079755266630164, "learning_rate": 6.721113769703887e-06, "loss": 0.4929, "step": 2342 }, { "epoch": 2.2486797887662027, "grad_norm": 0.8742051689833612, "learning_rate": 6.71796814620003e-06, "loss": 0.5511, "step": 2343 }, { "epoch": 2.2496399423907825, "grad_norm": 0.7994236514778583, "learning_rate": 6.714821751518635e-06, "loss": 0.582, "step": 2344 }, { "epoch": 2.2506000960153623, "grad_norm": 0.6907272082167039, "learning_rate": 6.7116745870720875e-06, "loss": 0.4995, "step": 2345 }, { "epoch": 2.2515602496399425, "grad_norm": 1.3904084432816364, "learning_rate": 6.708526654273113e-06, "loss": 0.6483, "step": 2346 }, { "epoch": 2.2525204032645223, "grad_norm": 0.9011568313822741, "learning_rate": 6.7053779545347874e-06, "loss": 0.6974, "step": 2347 }, { "epoch": 2.253480556889102, "grad_norm": 0.9003692505899151, "learning_rate": 6.702228489270528e-06, "loss": 0.5972, "step": 2348 }, { "epoch": 2.2544407105136823, "grad_norm": 0.995739685460113, "learning_rate": 6.699078259894094e-06, "loss": 0.536, "step": 2349 }, { "epoch": 2.255400864138262, "grad_norm": 1.3684079622341496, "learning_rate": 6.695927267819593e-06, "loss": 0.5302, "step": 2350 }, { "epoch": 2.256361017762842, "grad_norm": 1.3967457471618019, "learning_rate": 6.692775514461468e-06, "loss": 0.5267, "step": 2351 }, { "epoch": 2.257321171387422, "grad_norm": 0.8149418767064418, "learning_rate": 6.689623001234511e-06, "loss": 0.5397, "step": 2352 }, { "epoch": 2.258281325012002, "grad_norm": 0.9181381323565266, "learning_rate": 6.686469729553849e-06, "loss": 0.6286, "step": 2353 }, { "epoch": 2.2592414786365818, "grad_norm": 0.9673734432915574, "learning_rate": 6.683315700834953e-06, "loss": 0.6397, "step": 2354 }, { "epoch": 2.260201632261162, "grad_norm": 0.7972037466658118, "learning_rate": 6.680160916493631e-06, "loss": 0.5192, "step": 2355 }, { "epoch": 2.261161785885742, "grad_norm": 0.8293596255133032, "learning_rate": 6.677005377946037e-06, "loss": 0.5019, "step": 2356 }, { "epoch": 2.2621219395103216, "grad_norm": 1.2078158374896895, "learning_rate": 6.6738490866086515e-06, "loss": 0.6062, "step": 2357 }, { "epoch": 2.2630820931349014, "grad_norm": 1.3917572096419468, "learning_rate": 6.670692043898306e-06, "loss": 0.5367, "step": 2358 }, { "epoch": 2.2640422467594816, "grad_norm": 0.8120940513373924, "learning_rate": 6.66753425123216e-06, "loss": 0.5197, "step": 2359 }, { "epoch": 2.2650024003840614, "grad_norm": 0.8567098392389515, "learning_rate": 6.664375710027713e-06, "loss": 0.5323, "step": 2360 }, { "epoch": 2.265962554008641, "grad_norm": 0.8319743218466963, "learning_rate": 6.661216421702801e-06, "loss": 0.4889, "step": 2361 }, { "epoch": 2.2669227076332215, "grad_norm": 0.8402243438512916, "learning_rate": 6.6580563876755945e-06, "loss": 0.5961, "step": 2362 }, { "epoch": 2.2678828612578013, "grad_norm": 0.9794132827397477, "learning_rate": 6.6548956093646e-06, "loss": 0.601, "step": 2363 }, { "epoch": 2.268843014882381, "grad_norm": 1.323673887706244, "learning_rate": 6.651734088188655e-06, "loss": 0.6335, "step": 2364 }, { "epoch": 2.2698031685069613, "grad_norm": 1.3706422423320719, "learning_rate": 6.648571825566936e-06, "loss": 0.6333, "step": 2365 }, { "epoch": 2.270763322131541, "grad_norm": 0.732048924637249, "learning_rate": 6.645408822918945e-06, "loss": 0.5583, "step": 2366 }, { "epoch": 2.271723475756121, "grad_norm": 1.3190287041926665, "learning_rate": 6.642245081664524e-06, "loss": 0.5864, "step": 2367 }, { "epoch": 2.272683629380701, "grad_norm": 1.100971894588111, "learning_rate": 6.6390806032238375e-06, "loss": 0.607, "step": 2368 }, { "epoch": 2.273643783005281, "grad_norm": 1.2228464897597497, "learning_rate": 6.63591538901739e-06, "loss": 0.4323, "step": 2369 }, { "epoch": 2.2746039366298607, "grad_norm": 1.1874837162765433, "learning_rate": 6.632749440466013e-06, "loss": 0.6293, "step": 2370 }, { "epoch": 2.2755640902544405, "grad_norm": 0.8657452059309796, "learning_rate": 6.629582758990864e-06, "loss": 0.5127, "step": 2371 }, { "epoch": 2.2765242438790207, "grad_norm": 1.023282598891758, "learning_rate": 6.626415346013434e-06, "loss": 0.5702, "step": 2372 }, { "epoch": 2.2774843975036005, "grad_norm": 1.381686378404994, "learning_rate": 6.623247202955538e-06, "loss": 0.6763, "step": 2373 }, { "epoch": 2.2784445511281803, "grad_norm": 0.8348802486272411, "learning_rate": 6.620078331239328e-06, "loss": 0.5883, "step": 2374 }, { "epoch": 2.2794047047527606, "grad_norm": 0.8714481005739935, "learning_rate": 6.61690873228727e-06, "loss": 0.6147, "step": 2375 }, { "epoch": 2.2803648583773404, "grad_norm": 0.7022962290162079, "learning_rate": 6.613738407522169e-06, "loss": 0.5183, "step": 2376 }, { "epoch": 2.28132501200192, "grad_norm": 0.7506343788614488, "learning_rate": 6.610567358367143e-06, "loss": 0.5218, "step": 2377 }, { "epoch": 2.2822851656265004, "grad_norm": 0.8916457776413744, "learning_rate": 6.607395586245648e-06, "loss": 0.5138, "step": 2378 }, { "epoch": 2.28324531925108, "grad_norm": 0.7287075619370645, "learning_rate": 6.6042230925814565e-06, "loss": 0.535, "step": 2379 }, { "epoch": 2.28420547287566, "grad_norm": 0.9269163974469863, "learning_rate": 6.601049878798667e-06, "loss": 0.5106, "step": 2380 }, { "epoch": 2.2851656265002402, "grad_norm": 0.7846018632056144, "learning_rate": 6.597875946321701e-06, "loss": 0.5516, "step": 2381 }, { "epoch": 2.28612578012482, "grad_norm": 0.7639195120607605, "learning_rate": 6.594701296575303e-06, "loss": 0.5456, "step": 2382 }, { "epoch": 2.2870859337494, "grad_norm": 0.7149634899104811, "learning_rate": 6.5915259309845404e-06, "loss": 0.48, "step": 2383 }, { "epoch": 2.2880460873739796, "grad_norm": 0.7614270591097891, "learning_rate": 6.588349850974798e-06, "loss": 0.528, "step": 2384 }, { "epoch": 2.28900624099856, "grad_norm": 0.813915213341834, "learning_rate": 6.585173057971787e-06, "loss": 0.5043, "step": 2385 }, { "epoch": 2.2899663946231397, "grad_norm": 0.8001717725976218, "learning_rate": 6.5819955534015335e-06, "loss": 0.5269, "step": 2386 }, { "epoch": 2.2909265482477195, "grad_norm": 1.0362319276425576, "learning_rate": 6.578817338690389e-06, "loss": 0.5557, "step": 2387 }, { "epoch": 2.2918867018722997, "grad_norm": 0.8629327316291197, "learning_rate": 6.575638415265016e-06, "loss": 0.622, "step": 2388 }, { "epoch": 2.2928468554968795, "grad_norm": 0.6801870996040628, "learning_rate": 6.5724587845524024e-06, "loss": 0.4203, "step": 2389 }, { "epoch": 2.2938070091214593, "grad_norm": 1.0729003687552232, "learning_rate": 6.569278447979848e-06, "loss": 0.5311, "step": 2390 }, { "epoch": 2.2947671627460395, "grad_norm": 0.806414241916509, "learning_rate": 6.566097406974972e-06, "loss": 0.5595, "step": 2391 }, { "epoch": 2.2957273163706193, "grad_norm": 0.8887959321740087, "learning_rate": 6.562915662965713e-06, "loss": 0.5138, "step": 2392 }, { "epoch": 2.296687469995199, "grad_norm": 0.8513037902264567, "learning_rate": 6.5597332173803175e-06, "loss": 0.5424, "step": 2393 }, { "epoch": 2.2976476236197794, "grad_norm": 0.76475289191088, "learning_rate": 6.556550071647355e-06, "loss": 0.4456, "step": 2394 }, { "epoch": 2.298607777244359, "grad_norm": 1.0461107912872947, "learning_rate": 6.553366227195702e-06, "loss": 0.5973, "step": 2395 }, { "epoch": 2.299567930868939, "grad_norm": 0.7644561672411381, "learning_rate": 6.550181685454556e-06, "loss": 0.5151, "step": 2396 }, { "epoch": 2.3005280844935188, "grad_norm": 0.8397687351865347, "learning_rate": 6.546996447853421e-06, "loss": 0.5889, "step": 2397 }, { "epoch": 2.301488238118099, "grad_norm": 0.8753118376079007, "learning_rate": 6.5438105158221175e-06, "loss": 0.4996, "step": 2398 }, { "epoch": 2.302448391742679, "grad_norm": 1.0763037684727408, "learning_rate": 6.5406238907907745e-06, "loss": 0.5746, "step": 2399 }, { "epoch": 2.3034085453672586, "grad_norm": 1.056492128784647, "learning_rate": 6.537436574189837e-06, "loss": 0.5325, "step": 2400 }, { "epoch": 2.304368698991839, "grad_norm": 1.0306538666279959, "learning_rate": 6.5342485674500554e-06, "loss": 0.5835, "step": 2401 }, { "epoch": 2.3053288526164186, "grad_norm": 0.9287846543184227, "learning_rate": 6.531059872002492e-06, "loss": 0.5288, "step": 2402 }, { "epoch": 2.3062890062409984, "grad_norm": 0.96723897870749, "learning_rate": 6.527870489278519e-06, "loss": 0.4826, "step": 2403 }, { "epoch": 2.3072491598655787, "grad_norm": 0.8296293541686203, "learning_rate": 6.524680420709813e-06, "loss": 0.5954, "step": 2404 }, { "epoch": 2.3082093134901585, "grad_norm": 1.0899170715903312, "learning_rate": 6.521489667728365e-06, "loss": 0.5669, "step": 2405 }, { "epoch": 2.3091694671147383, "grad_norm": 1.01988545111116, "learning_rate": 6.51829823176647e-06, "loss": 0.5473, "step": 2406 }, { "epoch": 2.3101296207393185, "grad_norm": 0.8513449140263619, "learning_rate": 6.5151061142567285e-06, "loss": 0.5208, "step": 2407 }, { "epoch": 2.3110897743638983, "grad_norm": 1.1497388221440954, "learning_rate": 6.511913316632046e-06, "loss": 0.5946, "step": 2408 }, { "epoch": 2.312049927988478, "grad_norm": 0.8200266385589348, "learning_rate": 6.508719840325639e-06, "loss": 0.5504, "step": 2409 }, { "epoch": 2.313010081613058, "grad_norm": 0.8306564593647182, "learning_rate": 6.505525686771021e-06, "loss": 0.4873, "step": 2410 }, { "epoch": 2.313970235237638, "grad_norm": 0.8894099636235254, "learning_rate": 6.502330857402018e-06, "loss": 0.6225, "step": 2411 }, { "epoch": 2.314930388862218, "grad_norm": 0.942464231931411, "learning_rate": 6.49913535365275e-06, "loss": 0.5492, "step": 2412 }, { "epoch": 2.3158905424867977, "grad_norm": 1.1799985010571399, "learning_rate": 6.495939176957648e-06, "loss": 0.5105, "step": 2413 }, { "epoch": 2.316850696111378, "grad_norm": 1.2412085911265123, "learning_rate": 6.49274232875144e-06, "loss": 0.6064, "step": 2414 }, { "epoch": 2.3178108497359577, "grad_norm": 1.1697708105620999, "learning_rate": 6.489544810469157e-06, "loss": 0.5975, "step": 2415 }, { "epoch": 2.3187710033605375, "grad_norm": 1.3722517851791642, "learning_rate": 6.4863466235461316e-06, "loss": 0.6057, "step": 2416 }, { "epoch": 2.319731156985118, "grad_norm": 0.7388606082874246, "learning_rate": 6.483147769417992e-06, "loss": 0.6057, "step": 2417 }, { "epoch": 2.3206913106096976, "grad_norm": 0.8864776934567533, "learning_rate": 6.479948249520676e-06, "loss": 0.5425, "step": 2418 }, { "epoch": 2.3216514642342774, "grad_norm": 0.9653520737353172, "learning_rate": 6.476748065290409e-06, "loss": 0.6065, "step": 2419 }, { "epoch": 2.3226116178588576, "grad_norm": 1.265711049964671, "learning_rate": 6.473547218163723e-06, "loss": 0.5729, "step": 2420 }, { "epoch": 2.3235717714834374, "grad_norm": 0.8736473346750879, "learning_rate": 6.4703457095774395e-06, "loss": 0.559, "step": 2421 }, { "epoch": 2.324531925108017, "grad_norm": 1.0587849166586871, "learning_rate": 6.467143540968684e-06, "loss": 0.6009, "step": 2422 }, { "epoch": 2.325492078732597, "grad_norm": 1.1086070907033692, "learning_rate": 6.4639407137748774e-06, "loss": 0.5638, "step": 2423 }, { "epoch": 2.3264522323571772, "grad_norm": 0.9228464310286084, "learning_rate": 6.4607372294337335e-06, "loss": 0.5396, "step": 2424 }, { "epoch": 2.327412385981757, "grad_norm": 1.21297242148716, "learning_rate": 6.457533089383261e-06, "loss": 0.5413, "step": 2425 }, { "epoch": 2.328372539606337, "grad_norm": 0.7687253779509216, "learning_rate": 6.454328295061764e-06, "loss": 0.5532, "step": 2426 }, { "epoch": 2.329332693230917, "grad_norm": 0.8500929278715281, "learning_rate": 6.451122847907843e-06, "loss": 0.4883, "step": 2427 }, { "epoch": 2.330292846855497, "grad_norm": 1.146326050844335, "learning_rate": 6.447916749360386e-06, "loss": 0.5447, "step": 2428 }, { "epoch": 2.3312530004800767, "grad_norm": 0.9691610441968542, "learning_rate": 6.444710000858578e-06, "loss": 0.5901, "step": 2429 }, { "epoch": 2.332213154104657, "grad_norm": 0.8691136088620439, "learning_rate": 6.441502603841892e-06, "loss": 0.4891, "step": 2430 }, { "epoch": 2.3331733077292367, "grad_norm": 0.9674606311463032, "learning_rate": 6.438294559750098e-06, "loss": 0.5079, "step": 2431 }, { "epoch": 2.3341334613538165, "grad_norm": 0.8465954190298834, "learning_rate": 6.43508587002325e-06, "loss": 0.5061, "step": 2432 }, { "epoch": 2.3350936149783967, "grad_norm": 1.6951524557101236, "learning_rate": 6.431876536101695e-06, "loss": 0.5339, "step": 2433 }, { "epoch": 2.3360537686029765, "grad_norm": 1.333297990781059, "learning_rate": 6.42866655942607e-06, "loss": 0.5949, "step": 2434 }, { "epoch": 2.3370139222275563, "grad_norm": 0.8286083152911221, "learning_rate": 6.425455941437297e-06, "loss": 0.5287, "step": 2435 }, { "epoch": 2.337974075852136, "grad_norm": 0.8854792013270212, "learning_rate": 6.422244683576589e-06, "loss": 0.6031, "step": 2436 }, { "epoch": 2.3389342294767164, "grad_norm": 0.8788562421924145, "learning_rate": 6.419032787285446e-06, "loss": 0.5749, "step": 2437 }, { "epoch": 2.339894383101296, "grad_norm": 0.726955778740556, "learning_rate": 6.4158202540056536e-06, "loss": 0.4718, "step": 2438 }, { "epoch": 2.340854536725876, "grad_norm": 0.8034131880705507, "learning_rate": 6.412607085179282e-06, "loss": 0.5144, "step": 2439 }, { "epoch": 2.341814690350456, "grad_norm": 0.9834005517281514, "learning_rate": 6.409393282248692e-06, "loss": 0.5299, "step": 2440 }, { "epoch": 2.342774843975036, "grad_norm": 1.226425557947381, "learning_rate": 6.406178846656521e-06, "loss": 0.5072, "step": 2441 }, { "epoch": 2.343734997599616, "grad_norm": 0.8448212199788026, "learning_rate": 6.402963779845699e-06, "loss": 0.5385, "step": 2442 }, { "epoch": 2.344695151224196, "grad_norm": 0.6581420462165278, "learning_rate": 6.399748083259432e-06, "loss": 0.4634, "step": 2443 }, { "epoch": 2.345655304848776, "grad_norm": 0.9176361755898933, "learning_rate": 6.3965317583412135e-06, "loss": 0.563, "step": 2444 }, { "epoch": 2.3466154584733556, "grad_norm": 0.802091294517168, "learning_rate": 6.393314806534815e-06, "loss": 0.6175, "step": 2445 }, { "epoch": 2.347575612097936, "grad_norm": 0.7660216141952608, "learning_rate": 6.390097229284296e-06, "loss": 0.4425, "step": 2446 }, { "epoch": 2.3485357657225157, "grad_norm": 0.9850481632568591, "learning_rate": 6.386879028033988e-06, "loss": 0.4883, "step": 2447 }, { "epoch": 2.3494959193470955, "grad_norm": 0.7087514494758076, "learning_rate": 6.3836602042285105e-06, "loss": 0.5123, "step": 2448 }, { "epoch": 2.3504560729716752, "grad_norm": 0.8475286136248829, "learning_rate": 6.3804407593127585e-06, "loss": 0.5874, "step": 2449 }, { "epoch": 2.3514162265962555, "grad_norm": 1.1073268046769165, "learning_rate": 6.377220694731904e-06, "loss": 0.5219, "step": 2450 }, { "epoch": 2.3523763802208353, "grad_norm": 0.8027164270948304, "learning_rate": 6.374000011931403e-06, "loss": 0.4508, "step": 2451 }, { "epoch": 2.353336533845415, "grad_norm": 0.8646452244196177, "learning_rate": 6.370778712356981e-06, "loss": 0.5981, "step": 2452 }, { "epoch": 2.3542966874699953, "grad_norm": 0.9773588709482676, "learning_rate": 6.3675567974546506e-06, "loss": 0.475, "step": 2453 }, { "epoch": 2.355256841094575, "grad_norm": 0.8224935455973518, "learning_rate": 6.364334268670691e-06, "loss": 0.5587, "step": 2454 }, { "epoch": 2.356216994719155, "grad_norm": 0.7071076688115546, "learning_rate": 6.361111127451664e-06, "loss": 0.5584, "step": 2455 }, { "epoch": 2.357177148343735, "grad_norm": 0.8987540591413895, "learning_rate": 6.357887375244398e-06, "loss": 0.5817, "step": 2456 }, { "epoch": 2.358137301968315, "grad_norm": 0.8203878066459614, "learning_rate": 6.354663013496006e-06, "loss": 0.5124, "step": 2457 }, { "epoch": 2.3590974555928947, "grad_norm": 1.2034459889800146, "learning_rate": 6.3514380436538695e-06, "loss": 0.6047, "step": 2458 }, { "epoch": 2.360057609217475, "grad_norm": 0.8427239689714581, "learning_rate": 6.348212467165639e-06, "loss": 0.5902, "step": 2459 }, { "epoch": 2.361017762842055, "grad_norm": 1.1064511975819271, "learning_rate": 6.344986285479244e-06, "loss": 0.5515, "step": 2460 }, { "epoch": 2.3619779164666346, "grad_norm": 1.139032482729697, "learning_rate": 6.341759500042882e-06, "loss": 0.5384, "step": 2461 }, { "epoch": 2.3629380700912144, "grad_norm": 0.7064992644131434, "learning_rate": 6.338532112305025e-06, "loss": 0.4848, "step": 2462 }, { "epoch": 2.3638982237157946, "grad_norm": 0.70009307519383, "learning_rate": 6.3353041237144096e-06, "loss": 0.5621, "step": 2463 }, { "epoch": 2.3648583773403744, "grad_norm": 0.8572942409318149, "learning_rate": 6.332075535720047e-06, "loss": 0.6082, "step": 2464 }, { "epoch": 2.365818530964954, "grad_norm": 0.8927649728762119, "learning_rate": 6.328846349771215e-06, "loss": 0.4726, "step": 2465 }, { "epoch": 2.3667786845895344, "grad_norm": 0.930532915630933, "learning_rate": 6.325616567317461e-06, "loss": 0.6445, "step": 2466 }, { "epoch": 2.3677388382141142, "grad_norm": 0.9871487681630368, "learning_rate": 6.3223861898086e-06, "loss": 0.4227, "step": 2467 }, { "epoch": 2.368698991838694, "grad_norm": 0.7780213210459453, "learning_rate": 6.319155218694716e-06, "loss": 0.5629, "step": 2468 }, { "epoch": 2.3696591454632743, "grad_norm": 1.0370493567002148, "learning_rate": 6.315923655426152e-06, "loss": 0.5945, "step": 2469 }, { "epoch": 2.370619299087854, "grad_norm": 0.9546004438421463, "learning_rate": 6.312691501453527e-06, "loss": 0.6415, "step": 2470 }, { "epoch": 2.371579452712434, "grad_norm": 0.764129590752954, "learning_rate": 6.30945875822772e-06, "loss": 0.4583, "step": 2471 }, { "epoch": 2.372539606337014, "grad_norm": 0.7541802410900051, "learning_rate": 6.306225427199873e-06, "loss": 0.5398, "step": 2472 }, { "epoch": 2.373499759961594, "grad_norm": 0.8201506484074852, "learning_rate": 6.302991509821396e-06, "loss": 0.5399, "step": 2473 }, { "epoch": 2.3744599135861737, "grad_norm": 0.8315135096924002, "learning_rate": 6.299757007543954e-06, "loss": 0.515, "step": 2474 }, { "epoch": 2.3754200672107535, "grad_norm": 0.8615694637022332, "learning_rate": 6.296521921819489e-06, "loss": 0.5884, "step": 2475 }, { "epoch": 2.3763802208353337, "grad_norm": 0.6816953141128513, "learning_rate": 6.29328625410019e-06, "loss": 0.5465, "step": 2476 }, { "epoch": 2.3773403744599135, "grad_norm": 0.867603978706374, "learning_rate": 6.290050005838519e-06, "loss": 0.6176, "step": 2477 }, { "epoch": 2.3783005280844933, "grad_norm": 0.6792054116097529, "learning_rate": 6.286813178487188e-06, "loss": 0.4953, "step": 2478 }, { "epoch": 2.3792606817090736, "grad_norm": 0.7762117899305697, "learning_rate": 6.283575773499176e-06, "loss": 0.5306, "step": 2479 }, { "epoch": 2.3802208353336534, "grad_norm": 0.716273301591552, "learning_rate": 6.280337792327722e-06, "loss": 0.5818, "step": 2480 }, { "epoch": 2.381180988958233, "grad_norm": 1.1282952815415381, "learning_rate": 6.277099236426317e-06, "loss": 0.6016, "step": 2481 }, { "epoch": 2.3821411425828134, "grad_norm": 1.1970007179244857, "learning_rate": 6.273860107248718e-06, "loss": 0.4602, "step": 2482 }, { "epoch": 2.383101296207393, "grad_norm": 1.4262464668716024, "learning_rate": 6.270620406248931e-06, "loss": 0.6178, "step": 2483 }, { "epoch": 2.384061449831973, "grad_norm": 0.8567573242846797, "learning_rate": 6.267380134881229e-06, "loss": 0.5672, "step": 2484 }, { "epoch": 2.3850216034565532, "grad_norm": 0.9955691175043392, "learning_rate": 6.26413929460013e-06, "loss": 0.5001, "step": 2485 }, { "epoch": 2.385981757081133, "grad_norm": 0.8884788256851751, "learning_rate": 6.260897886860415e-06, "loss": 0.6232, "step": 2486 }, { "epoch": 2.386941910705713, "grad_norm": 1.231424878469658, "learning_rate": 6.2576559131171174e-06, "loss": 0.5403, "step": 2487 }, { "epoch": 2.3879020643302926, "grad_norm": 1.0303594113625156, "learning_rate": 6.254413374825523e-06, "loss": 0.6433, "step": 2488 }, { "epoch": 2.388862217954873, "grad_norm": 0.8198035952622746, "learning_rate": 6.251170273441174e-06, "loss": 0.5591, "step": 2489 }, { "epoch": 2.3898223715794527, "grad_norm": 0.875013241677267, "learning_rate": 6.247926610419865e-06, "loss": 0.496, "step": 2490 }, { "epoch": 2.390782525204033, "grad_norm": 0.8228697373135622, "learning_rate": 6.244682387217638e-06, "loss": 0.6607, "step": 2491 }, { "epoch": 2.3917426788286127, "grad_norm": 0.8489269597084933, "learning_rate": 6.241437605290792e-06, "loss": 0.5399, "step": 2492 }, { "epoch": 2.3927028324531925, "grad_norm": 0.8836121201700263, "learning_rate": 6.238192266095877e-06, "loss": 0.5976, "step": 2493 }, { "epoch": 2.3936629860777723, "grad_norm": 0.9313532051070622, "learning_rate": 6.234946371089688e-06, "loss": 0.5651, "step": 2494 }, { "epoch": 2.3946231397023525, "grad_norm": 1.1865557129599402, "learning_rate": 6.231699921729275e-06, "loss": 0.6115, "step": 2495 }, { "epoch": 2.3955832933269323, "grad_norm": 0.749055594153711, "learning_rate": 6.228452919471933e-06, "loss": 0.4582, "step": 2496 }, { "epoch": 2.396543446951512, "grad_norm": 0.6508787658378696, "learning_rate": 6.225205365775206e-06, "loss": 0.5341, "step": 2497 }, { "epoch": 2.3975036005760924, "grad_norm": 0.7153576209712074, "learning_rate": 6.221957262096887e-06, "loss": 0.514, "step": 2498 }, { "epoch": 2.398463754200672, "grad_norm": 0.7564120613494117, "learning_rate": 6.218708609895017e-06, "loss": 0.5709, "step": 2499 }, { "epoch": 2.399423907825252, "grad_norm": 0.9656082405990071, "learning_rate": 6.2154594106278775e-06, "loss": 0.5771, "step": 2500 }, { "epoch": 2.4003840614498317, "grad_norm": 1.1729894495115094, "learning_rate": 6.212209665754001e-06, "loss": 0.6094, "step": 2501 }, { "epoch": 2.401344215074412, "grad_norm": 1.004398495336107, "learning_rate": 6.2089593767321645e-06, "loss": 0.5589, "step": 2502 }, { "epoch": 2.4023043686989918, "grad_norm": 0.8968695167983407, "learning_rate": 6.2057085450213864e-06, "loss": 0.6094, "step": 2503 }, { "epoch": 2.403264522323572, "grad_norm": 0.8810037900885908, "learning_rate": 6.202457172080931e-06, "loss": 0.5872, "step": 2504 }, { "epoch": 2.404224675948152, "grad_norm": 0.8637221435207222, "learning_rate": 6.199205259370304e-06, "loss": 0.5419, "step": 2505 }, { "epoch": 2.4051848295727316, "grad_norm": 0.8406035248988037, "learning_rate": 6.195952808349255e-06, "loss": 0.566, "step": 2506 }, { "epoch": 2.4061449831973114, "grad_norm": 0.8137554977904026, "learning_rate": 6.192699820477775e-06, "loss": 0.4903, "step": 2507 }, { "epoch": 2.4071051368218916, "grad_norm": 0.7326256392933991, "learning_rate": 6.189446297216094e-06, "loss": 0.4434, "step": 2508 }, { "epoch": 2.4080652904464714, "grad_norm": 0.7503948123675097, "learning_rate": 6.186192240024684e-06, "loss": 0.5267, "step": 2509 }, { "epoch": 2.4090254440710512, "grad_norm": 0.8895985290442829, "learning_rate": 6.182937650364258e-06, "loss": 0.4768, "step": 2510 }, { "epoch": 2.4099855976956315, "grad_norm": 0.7795654798586693, "learning_rate": 6.179682529695766e-06, "loss": 0.5337, "step": 2511 }, { "epoch": 2.4109457513202113, "grad_norm": 0.7137915216337009, "learning_rate": 6.176426879480396e-06, "loss": 0.5585, "step": 2512 }, { "epoch": 2.411905904944791, "grad_norm": 1.014550274813306, "learning_rate": 6.173170701179576e-06, "loss": 0.5247, "step": 2513 }, { "epoch": 2.412866058569371, "grad_norm": 1.004923043308982, "learning_rate": 6.169913996254967e-06, "loss": 0.5926, "step": 2514 }, { "epoch": 2.413826212193951, "grad_norm": 0.8761774323267347, "learning_rate": 6.166656766168473e-06, "loss": 0.6616, "step": 2515 }, { "epoch": 2.414786365818531, "grad_norm": 0.9101147475075029, "learning_rate": 6.1633990123822265e-06, "loss": 0.5076, "step": 2516 }, { "epoch": 2.415746519443111, "grad_norm": 1.115085736840635, "learning_rate": 6.1601407363586e-06, "loss": 0.6124, "step": 2517 }, { "epoch": 2.416706673067691, "grad_norm": 0.8628158278945284, "learning_rate": 6.156881939560196e-06, "loss": 0.5883, "step": 2518 }, { "epoch": 2.4176668266922707, "grad_norm": 0.8746007022022848, "learning_rate": 6.153622623449858e-06, "loss": 0.6868, "step": 2519 }, { "epoch": 2.4186269803168505, "grad_norm": 0.8839204838514031, "learning_rate": 6.150362789490654e-06, "loss": 0.5184, "step": 2520 }, { "epoch": 2.4195871339414308, "grad_norm": 1.373994888552197, "learning_rate": 6.147102439145892e-06, "loss": 0.5555, "step": 2521 }, { "epoch": 2.4205472875660106, "grad_norm": 0.9619935290704965, "learning_rate": 6.143841573879107e-06, "loss": 0.6336, "step": 2522 }, { "epoch": 2.4215074411905904, "grad_norm": 1.11773885960763, "learning_rate": 6.140580195154065e-06, "loss": 0.5841, "step": 2523 }, { "epoch": 2.4224675948151706, "grad_norm": 1.059483843769777, "learning_rate": 6.137318304434768e-06, "loss": 0.592, "step": 2524 }, { "epoch": 2.4234277484397504, "grad_norm": 0.8503180390892521, "learning_rate": 6.1340559031854385e-06, "loss": 0.535, "step": 2525 }, { "epoch": 2.42438790206433, "grad_norm": 0.803123500447364, "learning_rate": 6.130792992870539e-06, "loss": 0.5678, "step": 2526 }, { "epoch": 2.42534805568891, "grad_norm": 0.9915737337469129, "learning_rate": 6.127529574954749e-06, "loss": 0.6289, "step": 2527 }, { "epoch": 2.4263082093134902, "grad_norm": 0.7828766391267719, "learning_rate": 6.124265650902989e-06, "loss": 0.5563, "step": 2528 }, { "epoch": 2.42726836293807, "grad_norm": 0.9879948541204199, "learning_rate": 6.121001222180394e-06, "loss": 0.5638, "step": 2529 }, { "epoch": 2.4282285165626503, "grad_norm": 0.9097371784834887, "learning_rate": 6.117736290252335e-06, "loss": 0.4271, "step": 2530 }, { "epoch": 2.42918867018723, "grad_norm": 1.281697426538018, "learning_rate": 6.1144708565844026e-06, "loss": 0.57, "step": 2531 }, { "epoch": 2.43014882381181, "grad_norm": 0.8714043867376116, "learning_rate": 6.111204922642414e-06, "loss": 0.5406, "step": 2532 }, { "epoch": 2.4311089774363897, "grad_norm": 1.061008206976894, "learning_rate": 6.107938489892417e-06, "loss": 0.6128, "step": 2533 }, { "epoch": 2.43206913106097, "grad_norm": 1.001637746723317, "learning_rate": 6.104671559800672e-06, "loss": 0.5974, "step": 2534 }, { "epoch": 2.4330292846855497, "grad_norm": 0.8348521570810359, "learning_rate": 6.101404133833674e-06, "loss": 0.5742, "step": 2535 }, { "epoch": 2.4339894383101295, "grad_norm": 0.9246231157153547, "learning_rate": 6.098136213458131e-06, "loss": 0.5565, "step": 2536 }, { "epoch": 2.4349495919347097, "grad_norm": 0.8354033712071397, "learning_rate": 6.0948678001409835e-06, "loss": 0.6163, "step": 2537 }, { "epoch": 2.4359097455592895, "grad_norm": 0.8719959893745762, "learning_rate": 6.091598895349382e-06, "loss": 0.5259, "step": 2538 }, { "epoch": 2.4368698991838693, "grad_norm": 0.8491979666228137, "learning_rate": 6.088329500550706e-06, "loss": 0.5734, "step": 2539 }, { "epoch": 2.437830052808449, "grad_norm": 1.007206451627018, "learning_rate": 6.08505961721255e-06, "loss": 0.6179, "step": 2540 }, { "epoch": 2.4387902064330294, "grad_norm": 0.8869638372079026, "learning_rate": 6.081789246802731e-06, "loss": 0.5715, "step": 2541 }, { "epoch": 2.439750360057609, "grad_norm": 0.8681394993301735, "learning_rate": 6.078518390789282e-06, "loss": 0.5642, "step": 2542 }, { "epoch": 2.4407105136821894, "grad_norm": 0.8336984143049684, "learning_rate": 6.075247050640458e-06, "loss": 0.5518, "step": 2543 }, { "epoch": 2.441670667306769, "grad_norm": 0.7748458902760568, "learning_rate": 6.071975227824724e-06, "loss": 0.5812, "step": 2544 }, { "epoch": 2.442630820931349, "grad_norm": 0.9025048534408506, "learning_rate": 6.068702923810771e-06, "loss": 0.4981, "step": 2545 }, { "epoch": 2.4435909745559288, "grad_norm": 0.7727223854257009, "learning_rate": 6.0654301400675e-06, "loss": 0.5755, "step": 2546 }, { "epoch": 2.444551128180509, "grad_norm": 1.3840193207116622, "learning_rate": 6.062156878064026e-06, "loss": 0.5742, "step": 2547 }, { "epoch": 2.445511281805089, "grad_norm": 0.8581178522248938, "learning_rate": 6.058883139269685e-06, "loss": 0.5759, "step": 2548 }, { "epoch": 2.4464714354296686, "grad_norm": 0.8833180984734436, "learning_rate": 6.055608925154019e-06, "loss": 0.5777, "step": 2549 }, { "epoch": 2.447431589054249, "grad_norm": 1.2095491899248723, "learning_rate": 6.052334237186792e-06, "loss": 0.6632, "step": 2550 }, { "epoch": 2.4483917426788286, "grad_norm": 1.0142684368668384, "learning_rate": 6.049059076837974e-06, "loss": 0.6191, "step": 2551 }, { "epoch": 2.4493518963034084, "grad_norm": 0.9360442206064634, "learning_rate": 6.045783445577752e-06, "loss": 0.6151, "step": 2552 }, { "epoch": 2.4503120499279882, "grad_norm": 0.7499773626569256, "learning_rate": 6.042507344876518e-06, "loss": 0.5531, "step": 2553 }, { "epoch": 2.4512722035525685, "grad_norm": 0.8613655948382608, "learning_rate": 6.0392307762048816e-06, "loss": 0.4518, "step": 2554 }, { "epoch": 2.4522323571771483, "grad_norm": 0.8134235231724852, "learning_rate": 6.03595374103366e-06, "loss": 0.5467, "step": 2555 }, { "epoch": 2.4531925108017285, "grad_norm": 1.0045957303579967, "learning_rate": 6.032676240833876e-06, "loss": 0.4874, "step": 2556 }, { "epoch": 2.4541526644263083, "grad_norm": 1.190734142715777, "learning_rate": 6.029398277076766e-06, "loss": 0.5355, "step": 2557 }, { "epoch": 2.455112818050888, "grad_norm": 0.746681461909044, "learning_rate": 6.0261198512337725e-06, "loss": 0.5205, "step": 2558 }, { "epoch": 2.456072971675468, "grad_norm": 0.9039361814757505, "learning_rate": 6.022840964776548e-06, "loss": 0.6401, "step": 2559 }, { "epoch": 2.457033125300048, "grad_norm": 0.861550001592557, "learning_rate": 6.019561619176948e-06, "loss": 0.5859, "step": 2560 }, { "epoch": 2.457993278924628, "grad_norm": 0.731662253802547, "learning_rate": 6.016281815907036e-06, "loss": 0.4721, "step": 2561 }, { "epoch": 2.4589534325492077, "grad_norm": 1.0255197116607544, "learning_rate": 6.013001556439079e-06, "loss": 0.529, "step": 2562 }, { "epoch": 2.459913586173788, "grad_norm": 1.045722174014107, "learning_rate": 6.0097208422455526e-06, "loss": 0.5588, "step": 2563 }, { "epoch": 2.4608737397983678, "grad_norm": 1.2027156775233168, "learning_rate": 6.006439674799132e-06, "loss": 0.6064, "step": 2564 }, { "epoch": 2.4618338934229476, "grad_norm": 0.8019677548569398, "learning_rate": 6.0031580555727005e-06, "loss": 0.5469, "step": 2565 }, { "epoch": 2.4627940470475274, "grad_norm": 0.9552436988694889, "learning_rate": 5.999875986039341e-06, "loss": 0.6006, "step": 2566 }, { "epoch": 2.4637542006721076, "grad_norm": 0.6852721257050925, "learning_rate": 5.996593467672339e-06, "loss": 0.493, "step": 2567 }, { "epoch": 2.4647143542966874, "grad_norm": 1.2358174577731809, "learning_rate": 5.993310501945184e-06, "loss": 0.6306, "step": 2568 }, { "epoch": 2.4656745079212676, "grad_norm": 0.9384820519264684, "learning_rate": 5.990027090331562e-06, "loss": 0.4674, "step": 2569 }, { "epoch": 2.4666346615458474, "grad_norm": 1.0358535151773187, "learning_rate": 5.986743234305361e-06, "loss": 0.6154, "step": 2570 }, { "epoch": 2.4675948151704272, "grad_norm": 1.843805828537616, "learning_rate": 5.983458935340669e-06, "loss": 0.6204, "step": 2571 }, { "epoch": 2.468554968795007, "grad_norm": 1.074667060655214, "learning_rate": 5.980174194911776e-06, "loss": 0.4692, "step": 2572 }, { "epoch": 2.4695151224195873, "grad_norm": 0.837590881687899, "learning_rate": 5.9768890144931626e-06, "loss": 0.536, "step": 2573 }, { "epoch": 2.470475276044167, "grad_norm": 0.9015121664772828, "learning_rate": 5.973603395559514e-06, "loss": 0.5951, "step": 2574 }, { "epoch": 2.471435429668747, "grad_norm": 0.7900274970859336, "learning_rate": 5.970317339585707e-06, "loss": 0.5804, "step": 2575 }, { "epoch": 2.472395583293327, "grad_norm": 0.7927658189547767, "learning_rate": 5.96703084804682e-06, "loss": 0.5429, "step": 2576 }, { "epoch": 2.473355736917907, "grad_norm": 1.0296233603130494, "learning_rate": 5.963743922418122e-06, "loss": 0.5416, "step": 2577 }, { "epoch": 2.4743158905424867, "grad_norm": 1.2529306839066106, "learning_rate": 5.960456564175079e-06, "loss": 0.5378, "step": 2578 }, { "epoch": 2.475276044167067, "grad_norm": 0.8434627884960615, "learning_rate": 5.9571687747933505e-06, "loss": 0.4488, "step": 2579 }, { "epoch": 2.4762361977916467, "grad_norm": 0.8837413238650831, "learning_rate": 5.9538805557487896e-06, "loss": 0.5775, "step": 2580 }, { "epoch": 2.4771963514162265, "grad_norm": 1.0720987402630815, "learning_rate": 5.950591908517446e-06, "loss": 0.5457, "step": 2581 }, { "epoch": 2.4781565050408068, "grad_norm": 0.8338841744497562, "learning_rate": 5.9473028345755535e-06, "loss": 0.5344, "step": 2582 }, { "epoch": 2.4791166586653866, "grad_norm": 0.9053563741606727, "learning_rate": 5.944013335399546e-06, "loss": 0.5061, "step": 2583 }, { "epoch": 2.4800768122899663, "grad_norm": 0.7578553986408919, "learning_rate": 5.940723412466044e-06, "loss": 0.5757, "step": 2584 }, { "epoch": 2.481036965914546, "grad_norm": 0.7990499267927504, "learning_rate": 5.937433067251857e-06, "loss": 0.4988, "step": 2585 }, { "epoch": 2.4819971195391264, "grad_norm": 0.7692930367910844, "learning_rate": 5.934142301233986e-06, "loss": 0.4982, "step": 2586 }, { "epoch": 2.482957273163706, "grad_norm": 0.7394997564531093, "learning_rate": 5.9308511158896235e-06, "loss": 0.5494, "step": 2587 }, { "epoch": 2.483917426788286, "grad_norm": 0.7481841583324081, "learning_rate": 5.927559512696143e-06, "loss": 0.5444, "step": 2588 }, { "epoch": 2.484877580412866, "grad_norm": 1.0117079820603654, "learning_rate": 5.924267493131113e-06, "loss": 0.5162, "step": 2589 }, { "epoch": 2.485837734037446, "grad_norm": 0.9671270138698712, "learning_rate": 5.9209750586722855e-06, "loss": 0.5465, "step": 2590 }, { "epoch": 2.486797887662026, "grad_norm": 0.844150896272441, "learning_rate": 5.917682210797598e-06, "loss": 0.428, "step": 2591 }, { "epoch": 2.487758041286606, "grad_norm": 0.8324186458907575, "learning_rate": 5.9143889509851785e-06, "loss": 0.5656, "step": 2592 }, { "epoch": 2.488718194911186, "grad_norm": 0.8068088789351598, "learning_rate": 5.91109528071333e-06, "loss": 0.5255, "step": 2593 }, { "epoch": 2.4896783485357656, "grad_norm": 0.9573040261948073, "learning_rate": 5.9078012014605505e-06, "loss": 0.6264, "step": 2594 }, { "epoch": 2.490638502160346, "grad_norm": 0.975670019302661, "learning_rate": 5.904506714705514e-06, "loss": 0.6016, "step": 2595 }, { "epoch": 2.4915986557849257, "grad_norm": 1.0028977515026853, "learning_rate": 5.901211821927084e-06, "loss": 0.6226, "step": 2596 }, { "epoch": 2.4925588094095055, "grad_norm": 0.7335530088562207, "learning_rate": 5.897916524604299e-06, "loss": 0.5944, "step": 2597 }, { "epoch": 2.4935189630340853, "grad_norm": 1.0249901112431188, "learning_rate": 5.894620824216385e-06, "loss": 0.6544, "step": 2598 }, { "epoch": 2.4944791166586655, "grad_norm": 0.8776395104996917, "learning_rate": 5.8913247222427464e-06, "loss": 0.5682, "step": 2599 }, { "epoch": 2.4954392702832453, "grad_norm": 0.8322144939295961, "learning_rate": 5.888028220162967e-06, "loss": 0.5387, "step": 2600 }, { "epoch": 2.496399423907825, "grad_norm": 1.0723426314759015, "learning_rate": 5.884731319456812e-06, "loss": 0.6414, "step": 2601 }, { "epoch": 2.4973595775324053, "grad_norm": 1.1467626076802113, "learning_rate": 5.881434021604224e-06, "loss": 0.511, "step": 2602 }, { "epoch": 2.498319731156985, "grad_norm": 1.0175910563123598, "learning_rate": 5.878136328085328e-06, "loss": 0.6208, "step": 2603 }, { "epoch": 2.499279884781565, "grad_norm": 0.8859984607877526, "learning_rate": 5.874838240380419e-06, "loss": 0.6282, "step": 2604 }, { "epoch": 2.5002400384061447, "grad_norm": 0.7598588368539752, "learning_rate": 5.871539759969978e-06, "loss": 0.5311, "step": 2605 }, { "epoch": 2.501200192030725, "grad_norm": 0.7728756363649294, "learning_rate": 5.8682408883346535e-06, "loss": 0.5934, "step": 2606 }, { "epoch": 2.5021603456553048, "grad_norm": 0.9108755978318818, "learning_rate": 5.864941626955274e-06, "loss": 0.603, "step": 2607 }, { "epoch": 2.503120499279885, "grad_norm": 0.9664400470801189, "learning_rate": 5.861641977312844e-06, "loss": 0.5799, "step": 2608 }, { "epoch": 2.504080652904465, "grad_norm": 0.7496799717280929, "learning_rate": 5.858341940888543e-06, "loss": 0.6009, "step": 2609 }, { "epoch": 2.5050408065290446, "grad_norm": 0.7991058500704575, "learning_rate": 5.855041519163718e-06, "loss": 0.5118, "step": 2610 }, { "epoch": 2.5060009601536244, "grad_norm": 0.8591897392479123, "learning_rate": 5.851740713619896e-06, "loss": 0.5491, "step": 2611 }, { "epoch": 2.5069611137782046, "grad_norm": 1.4045043894763731, "learning_rate": 5.848439525738772e-06, "loss": 0.6468, "step": 2612 }, { "epoch": 2.5079212674027844, "grad_norm": 1.128728202132514, "learning_rate": 5.845137957002214e-06, "loss": 0.6078, "step": 2613 }, { "epoch": 2.5088814210273642, "grad_norm": 1.0011648076541617, "learning_rate": 5.841836008892262e-06, "loss": 0.5436, "step": 2614 }, { "epoch": 2.5098415746519445, "grad_norm": 0.6841576278180675, "learning_rate": 5.838533682891123e-06, "loss": 0.5193, "step": 2615 }, { "epoch": 2.5108017282765243, "grad_norm": 0.6517321855369012, "learning_rate": 5.835230980481177e-06, "loss": 0.4538, "step": 2616 }, { "epoch": 2.511761881901104, "grad_norm": 1.1239394361885358, "learning_rate": 5.831927903144972e-06, "loss": 0.6236, "step": 2617 }, { "epoch": 2.512722035525684, "grad_norm": 0.8135860013948432, "learning_rate": 5.828624452365226e-06, "loss": 0.5512, "step": 2618 }, { "epoch": 2.513682189150264, "grad_norm": 1.0856429342340905, "learning_rate": 5.82532062962482e-06, "loss": 0.5597, "step": 2619 }, { "epoch": 2.514642342774844, "grad_norm": 1.2626446061123067, "learning_rate": 5.822016436406805e-06, "loss": 0.5917, "step": 2620 }, { "epoch": 2.515602496399424, "grad_norm": 0.923211733935662, "learning_rate": 5.818711874194398e-06, "loss": 0.5948, "step": 2621 }, { "epoch": 2.516562650024004, "grad_norm": 0.7914316160549708, "learning_rate": 5.8154069444709825e-06, "loss": 0.6219, "step": 2622 }, { "epoch": 2.5175228036485837, "grad_norm": 0.8247520190030821, "learning_rate": 5.8121016487201074e-06, "loss": 0.5478, "step": 2623 }, { "epoch": 2.5184829572731635, "grad_norm": 0.9709722420976675, "learning_rate": 5.80879598842548e-06, "loss": 0.6298, "step": 2624 }, { "epoch": 2.5194431108977438, "grad_norm": 1.1148109731498659, "learning_rate": 5.805489965070983e-06, "loss": 0.6187, "step": 2625 }, { "epoch": 2.5204032645223235, "grad_norm": 0.8089415376446725, "learning_rate": 5.8021835801406475e-06, "loss": 0.5707, "step": 2626 }, { "epoch": 2.5213634181469033, "grad_norm": 0.7609824470785016, "learning_rate": 5.798876835118681e-06, "loss": 0.4851, "step": 2627 }, { "epoch": 2.5223235717714836, "grad_norm": 0.846255343006585, "learning_rate": 5.79556973148944e-06, "loss": 0.5694, "step": 2628 }, { "epoch": 2.5232837253960634, "grad_norm": 0.7859796954476888, "learning_rate": 5.792262270737451e-06, "loss": 0.5853, "step": 2629 }, { "epoch": 2.524243879020643, "grad_norm": 1.1270012337805615, "learning_rate": 5.788954454347397e-06, "loss": 0.6065, "step": 2630 }, { "epoch": 2.525204032645223, "grad_norm": 0.9543770230510515, "learning_rate": 5.7856462838041235e-06, "loss": 0.5269, "step": 2631 }, { "epoch": 2.526164186269803, "grad_norm": 1.1994834817238367, "learning_rate": 5.782337760592631e-06, "loss": 0.641, "step": 2632 }, { "epoch": 2.527124339894383, "grad_norm": 1.0273381531145531, "learning_rate": 5.7790288861980795e-06, "loss": 0.5362, "step": 2633 }, { "epoch": 2.5280844935189632, "grad_norm": 1.2609291575461208, "learning_rate": 5.77571966210579e-06, "loss": 0.6433, "step": 2634 }, { "epoch": 2.529044647143543, "grad_norm": 1.052665364085432, "learning_rate": 5.772410089801234e-06, "loss": 0.5961, "step": 2635 }, { "epoch": 2.530004800768123, "grad_norm": 1.4217153578717827, "learning_rate": 5.769100170770048e-06, "loss": 0.7077, "step": 2636 }, { "epoch": 2.5309649543927026, "grad_norm": 0.7793414735654217, "learning_rate": 5.765789906498015e-06, "loss": 0.5169, "step": 2637 }, { "epoch": 2.531925108017283, "grad_norm": 0.6024339195922678, "learning_rate": 5.762479298471078e-06, "loss": 0.4211, "step": 2638 }, { "epoch": 2.5328852616418627, "grad_norm": 0.730249104528258, "learning_rate": 5.759168348175335e-06, "loss": 0.4734, "step": 2639 }, { "epoch": 2.5338454152664425, "grad_norm": 0.7789224062839942, "learning_rate": 5.755857057097036e-06, "loss": 0.5116, "step": 2640 }, { "epoch": 2.5348055688910227, "grad_norm": 0.839481202402132, "learning_rate": 5.752545426722583e-06, "loss": 0.596, "step": 2641 }, { "epoch": 2.5357657225156025, "grad_norm": 0.8922822483173513, "learning_rate": 5.7492334585385325e-06, "loss": 0.5982, "step": 2642 }, { "epoch": 2.5367258761401823, "grad_norm": 0.7427354698019442, "learning_rate": 5.745921154031593e-06, "loss": 0.5706, "step": 2643 }, { "epoch": 2.537686029764762, "grad_norm": 1.018329243984356, "learning_rate": 5.742608514688619e-06, "loss": 0.5132, "step": 2644 }, { "epoch": 2.5386461833893423, "grad_norm": 0.7817794793368867, "learning_rate": 5.739295541996622e-06, "loss": 0.6084, "step": 2645 }, { "epoch": 2.539606337013922, "grad_norm": 0.7204751715287393, "learning_rate": 5.7359822374427565e-06, "loss": 0.5058, "step": 2646 }, { "epoch": 2.5405664906385024, "grad_norm": 0.906596515556493, "learning_rate": 5.732668602514334e-06, "loss": 0.5282, "step": 2647 }, { "epoch": 2.541526644263082, "grad_norm": 0.7470642815821269, "learning_rate": 5.729354638698806e-06, "loss": 0.4896, "step": 2648 }, { "epoch": 2.542486797887662, "grad_norm": 1.4399332415731452, "learning_rate": 5.726040347483779e-06, "loss": 0.6631, "step": 2649 }, { "epoch": 2.5434469515122418, "grad_norm": 0.8061506775507885, "learning_rate": 5.722725730357001e-06, "loss": 0.5427, "step": 2650 }, { "epoch": 2.544407105136822, "grad_norm": 1.571841054152445, "learning_rate": 5.719410788806366e-06, "loss": 0.5281, "step": 2651 }, { "epoch": 2.545367258761402, "grad_norm": 1.323248619419142, "learning_rate": 5.716095524319919e-06, "loss": 0.6307, "step": 2652 }, { "epoch": 2.5463274123859816, "grad_norm": 1.6089134240705978, "learning_rate": 5.712779938385846e-06, "loss": 0.5465, "step": 2653 }, { "epoch": 2.547287566010562, "grad_norm": 1.1759595402971565, "learning_rate": 5.709464032492478e-06, "loss": 0.6773, "step": 2654 }, { "epoch": 2.5482477196351416, "grad_norm": 0.9942254554398645, "learning_rate": 5.706147808128288e-06, "loss": 0.5585, "step": 2655 }, { "epoch": 2.5492078732597214, "grad_norm": 0.7581638227015663, "learning_rate": 5.702831266781898e-06, "loss": 0.4687, "step": 2656 }, { "epoch": 2.550168026884301, "grad_norm": 1.1662681515644344, "learning_rate": 5.699514409942064e-06, "loss": 0.6293, "step": 2657 }, { "epoch": 2.5511281805088815, "grad_norm": 0.7473171401123998, "learning_rate": 5.69619723909769e-06, "loss": 0.5761, "step": 2658 }, { "epoch": 2.5520883341334613, "grad_norm": 1.2094306679903335, "learning_rate": 5.6928797557378145e-06, "loss": 0.5919, "step": 2659 }, { "epoch": 2.5530484877580415, "grad_norm": 0.8863618256267835, "learning_rate": 5.689561961351624e-06, "loss": 0.622, "step": 2660 }, { "epoch": 2.5540086413826213, "grad_norm": 0.8019160317528049, "learning_rate": 5.68624385742844e-06, "loss": 0.5485, "step": 2661 }, { "epoch": 2.554968795007201, "grad_norm": 0.7555748906363645, "learning_rate": 5.682925445457726e-06, "loss": 0.5061, "step": 2662 }, { "epoch": 2.555928948631781, "grad_norm": 0.7672109932957579, "learning_rate": 5.679606726929076e-06, "loss": 0.5281, "step": 2663 }, { "epoch": 2.556889102256361, "grad_norm": 0.8745285476841749, "learning_rate": 5.676287703332234e-06, "loss": 0.617, "step": 2664 }, { "epoch": 2.557849255880941, "grad_norm": 1.2319105851776015, "learning_rate": 5.672968376157071e-06, "loss": 0.5691, "step": 2665 }, { "epoch": 2.558809409505521, "grad_norm": 0.8063481637100169, "learning_rate": 5.6696487468935955e-06, "loss": 0.5123, "step": 2666 }, { "epoch": 2.559769563130101, "grad_norm": 0.8223549515134773, "learning_rate": 5.666328817031958e-06, "loss": 0.5102, "step": 2667 }, { "epoch": 2.5607297167546808, "grad_norm": 0.8956171271001523, "learning_rate": 5.663008588062435e-06, "loss": 0.5328, "step": 2668 }, { "epoch": 2.5616898703792605, "grad_norm": 0.8707589746823646, "learning_rate": 5.659688061475445e-06, "loss": 0.503, "step": 2669 }, { "epoch": 2.5626500240038403, "grad_norm": 1.1161694417423138, "learning_rate": 5.656367238761536e-06, "loss": 0.5843, "step": 2670 }, { "epoch": 2.5636101776284206, "grad_norm": 0.9445312718013131, "learning_rate": 5.653046121411389e-06, "loss": 0.5609, "step": 2671 }, { "epoch": 2.5645703312530004, "grad_norm": 0.9213867346380659, "learning_rate": 5.649724710915819e-06, "loss": 0.4595, "step": 2672 }, { "epoch": 2.5655304848775806, "grad_norm": 0.7763219286376297, "learning_rate": 5.646403008765772e-06, "loss": 0.5759, "step": 2673 }, { "epoch": 2.5664906385021604, "grad_norm": 0.8549150574278276, "learning_rate": 5.643081016452323e-06, "loss": 0.5729, "step": 2674 }, { "epoch": 2.56745079212674, "grad_norm": 0.7202362734717502, "learning_rate": 5.639758735466682e-06, "loss": 0.532, "step": 2675 }, { "epoch": 2.56841094575132, "grad_norm": 0.9365044924057544, "learning_rate": 5.636436167300181e-06, "loss": 0.58, "step": 2676 }, { "epoch": 2.5693710993759002, "grad_norm": 0.8918098940140661, "learning_rate": 5.633113313444288e-06, "loss": 0.5631, "step": 2677 }, { "epoch": 2.57033125300048, "grad_norm": 0.7045271126850429, "learning_rate": 5.629790175390598e-06, "loss": 0.492, "step": 2678 }, { "epoch": 2.5712914066250603, "grad_norm": 0.8657740650712769, "learning_rate": 5.626466754630829e-06, "loss": 0.6414, "step": 2679 }, { "epoch": 2.57225156024964, "grad_norm": 0.7968385194071146, "learning_rate": 5.623143052656831e-06, "loss": 0.4635, "step": 2680 }, { "epoch": 2.57321171387422, "grad_norm": 0.8709965663424127, "learning_rate": 5.619819070960577e-06, "loss": 0.5723, "step": 2681 }, { "epoch": 2.5741718674987997, "grad_norm": 1.0275940412286309, "learning_rate": 5.616494811034168e-06, "loss": 0.6256, "step": 2682 }, { "epoch": 2.5751320211233795, "grad_norm": 1.1085391485666076, "learning_rate": 5.613170274369827e-06, "loss": 0.6451, "step": 2683 }, { "epoch": 2.5760921747479597, "grad_norm": 1.1792104760949869, "learning_rate": 5.609845462459907e-06, "loss": 0.6684, "step": 2684 }, { "epoch": 2.5770523283725395, "grad_norm": 1.1275406256243168, "learning_rate": 5.606520376796875e-06, "loss": 0.6007, "step": 2685 }, { "epoch": 2.5780124819971197, "grad_norm": 1.0493644066352186, "learning_rate": 5.603195018873329e-06, "loss": 0.647, "step": 2686 }, { "epoch": 2.5789726356216995, "grad_norm": 0.7442107213055987, "learning_rate": 5.59986939018199e-06, "loss": 0.5147, "step": 2687 }, { "epoch": 2.5799327892462793, "grad_norm": 1.310694205008311, "learning_rate": 5.596543492215691e-06, "loss": 0.5234, "step": 2688 }, { "epoch": 2.580892942870859, "grad_norm": 0.7497772618727491, "learning_rate": 5.5932173264673975e-06, "loss": 0.5589, "step": 2689 }, { "epoch": 2.5818530964954394, "grad_norm": 0.8531017612328001, "learning_rate": 5.589890894430186e-06, "loss": 0.5549, "step": 2690 }, { "epoch": 2.582813250120019, "grad_norm": 1.0935655007829221, "learning_rate": 5.586564197597256e-06, "loss": 0.5547, "step": 2691 }, { "epoch": 2.5837734037445994, "grad_norm": 0.9889732716558621, "learning_rate": 5.583237237461929e-06, "loss": 0.5424, "step": 2692 }, { "epoch": 2.584733557369179, "grad_norm": 0.9601540011510014, "learning_rate": 5.579910015517642e-06, "loss": 0.6234, "step": 2693 }, { "epoch": 2.585693710993759, "grad_norm": 0.9655667494394466, "learning_rate": 5.576582533257946e-06, "loss": 0.5804, "step": 2694 }, { "epoch": 2.586653864618339, "grad_norm": 1.1599500437681431, "learning_rate": 5.573254792176515e-06, "loss": 0.6142, "step": 2695 }, { "epoch": 2.5876140182429186, "grad_norm": 0.8055917761388314, "learning_rate": 5.569926793767136e-06, "loss": 0.543, "step": 2696 }, { "epoch": 2.588574171867499, "grad_norm": 0.7374608177418394, "learning_rate": 5.566598539523711e-06, "loss": 0.5088, "step": 2697 }, { "epoch": 2.5895343254920786, "grad_norm": 0.9798894273861061, "learning_rate": 5.563270030940258e-06, "loss": 0.5086, "step": 2698 }, { "epoch": 2.590494479116659, "grad_norm": 0.9868702866660403, "learning_rate": 5.559941269510908e-06, "loss": 0.5279, "step": 2699 }, { "epoch": 2.5914546327412387, "grad_norm": 0.7776511313122865, "learning_rate": 5.556612256729909e-06, "loss": 0.5615, "step": 2700 }, { "epoch": 2.5924147863658185, "grad_norm": 0.8681252166887508, "learning_rate": 5.553282994091617e-06, "loss": 0.5227, "step": 2701 }, { "epoch": 2.5933749399903983, "grad_norm": 1.0131904223656298, "learning_rate": 5.549953483090506e-06, "loss": 0.5075, "step": 2702 }, { "epoch": 2.5943350936149785, "grad_norm": 0.9350566059680241, "learning_rate": 5.546623725221152e-06, "loss": 0.6307, "step": 2703 }, { "epoch": 2.5952952472395583, "grad_norm": 0.6745678287075527, "learning_rate": 5.543293721978251e-06, "loss": 0.5206, "step": 2704 }, { "epoch": 2.5962554008641385, "grad_norm": 1.0502880949416522, "learning_rate": 5.539963474856607e-06, "loss": 0.6256, "step": 2705 }, { "epoch": 2.5972155544887183, "grad_norm": 0.832894817361201, "learning_rate": 5.536632985351132e-06, "loss": 0.5039, "step": 2706 }, { "epoch": 2.598175708113298, "grad_norm": 1.3748271283433557, "learning_rate": 5.533302254956843e-06, "loss": 0.663, "step": 2707 }, { "epoch": 2.599135861737878, "grad_norm": 0.9122895257060709, "learning_rate": 5.529971285168873e-06, "loss": 0.6461, "step": 2708 }, { "epoch": 2.600096015362458, "grad_norm": 0.7935118508547235, "learning_rate": 5.526640077482462e-06, "loss": 0.5719, "step": 2709 }, { "epoch": 2.601056168987038, "grad_norm": 0.820666926819397, "learning_rate": 5.523308633392945e-06, "loss": 0.5489, "step": 2710 }, { "epoch": 2.6020163226116177, "grad_norm": 0.6119570245788007, "learning_rate": 5.519976954395779e-06, "loss": 0.4311, "step": 2711 }, { "epoch": 2.602976476236198, "grad_norm": 0.6642432910573033, "learning_rate": 5.516645041986516e-06, "loss": 0.5313, "step": 2712 }, { "epoch": 2.603936629860778, "grad_norm": 0.7798264402271071, "learning_rate": 5.513312897660815e-06, "loss": 0.508, "step": 2713 }, { "epoch": 2.6048967834853576, "grad_norm": 1.0501805707089604, "learning_rate": 5.509980522914442e-06, "loss": 0.5575, "step": 2714 }, { "epoch": 2.6058569371099374, "grad_norm": 1.10882125497511, "learning_rate": 5.506647919243265e-06, "loss": 0.5763, "step": 2715 }, { "epoch": 2.6068170907345176, "grad_norm": 1.0574842713113857, "learning_rate": 5.503315088143251e-06, "loss": 0.6628, "step": 2716 }, { "epoch": 2.6077772443590974, "grad_norm": 1.033691929307902, "learning_rate": 5.499982031110475e-06, "loss": 0.6109, "step": 2717 }, { "epoch": 2.6087373979836777, "grad_norm": 0.8658000534252954, "learning_rate": 5.496648749641109e-06, "loss": 0.5987, "step": 2718 }, { "epoch": 2.6096975516082574, "grad_norm": 1.1001232546430748, "learning_rate": 5.493315245231429e-06, "loss": 0.5409, "step": 2719 }, { "epoch": 2.6106577052328372, "grad_norm": 0.7709702929507947, "learning_rate": 5.489981519377808e-06, "loss": 0.5762, "step": 2720 }, { "epoch": 2.611617858857417, "grad_norm": 0.9451624707929138, "learning_rate": 5.486647573576721e-06, "loss": 0.5377, "step": 2721 }, { "epoch": 2.6125780124819973, "grad_norm": 0.7829191321063325, "learning_rate": 5.483313409324741e-06, "loss": 0.4993, "step": 2722 }, { "epoch": 2.613538166106577, "grad_norm": 1.4287806029580128, "learning_rate": 5.479979028118536e-06, "loss": 0.5677, "step": 2723 }, { "epoch": 2.614498319731157, "grad_norm": 0.8398535649953068, "learning_rate": 5.476644431454878e-06, "loss": 0.5038, "step": 2724 }, { "epoch": 2.615458473355737, "grad_norm": 0.8586504283953494, "learning_rate": 5.473309620830628e-06, "loss": 0.5039, "step": 2725 }, { "epoch": 2.616418626980317, "grad_norm": 1.2785210517177203, "learning_rate": 5.469974597742749e-06, "loss": 0.7732, "step": 2726 }, { "epoch": 2.6173787806048967, "grad_norm": 0.8568345802087936, "learning_rate": 5.4666393636882965e-06, "loss": 0.5428, "step": 2727 }, { "epoch": 2.6183389342294765, "grad_norm": 0.7168965499539294, "learning_rate": 5.463303920164423e-06, "loss": 0.5535, "step": 2728 }, { "epoch": 2.6192990878540567, "grad_norm": 0.9636494192078505, "learning_rate": 5.459968268668372e-06, "loss": 0.635, "step": 2729 }, { "epoch": 2.6202592414786365, "grad_norm": 1.1609488437204594, "learning_rate": 5.456632410697481e-06, "loss": 0.5268, "step": 2730 }, { "epoch": 2.6212193951032168, "grad_norm": 1.368478092023691, "learning_rate": 5.453296347749183e-06, "loss": 0.6599, "step": 2731 }, { "epoch": 2.6221795487277966, "grad_norm": 0.8140437722564302, "learning_rate": 5.449960081321e-06, "loss": 0.6026, "step": 2732 }, { "epoch": 2.6231397023523764, "grad_norm": 1.0059517082220328, "learning_rate": 5.446623612910545e-06, "loss": 0.4367, "step": 2733 }, { "epoch": 2.624099855976956, "grad_norm": 1.016043341918016, "learning_rate": 5.443286944015525e-06, "loss": 0.5883, "step": 2734 }, { "epoch": 2.6250600096015364, "grad_norm": 1.2542362354423706, "learning_rate": 5.439950076133734e-06, "loss": 0.6515, "step": 2735 }, { "epoch": 2.626020163226116, "grad_norm": 0.8511725494861647, "learning_rate": 5.4366130107630566e-06, "loss": 0.6157, "step": 2736 }, { "epoch": 2.626980316850696, "grad_norm": 1.022036877644442, "learning_rate": 5.4332757494014675e-06, "loss": 0.517, "step": 2737 }, { "epoch": 2.6279404704752762, "grad_norm": 0.8140087477719233, "learning_rate": 5.429938293547024e-06, "loss": 0.6466, "step": 2738 }, { "epoch": 2.628900624099856, "grad_norm": 0.853088671792183, "learning_rate": 5.426600644697877e-06, "loss": 0.6042, "step": 2739 }, { "epoch": 2.629860777724436, "grad_norm": 0.7077748515913191, "learning_rate": 5.423262804352262e-06, "loss": 0.5113, "step": 2740 }, { "epoch": 2.6308209313490156, "grad_norm": 1.0123824987801378, "learning_rate": 5.419924774008496e-06, "loss": 0.542, "step": 2741 }, { "epoch": 2.631781084973596, "grad_norm": 0.8584828342485062, "learning_rate": 5.416586555164989e-06, "loss": 0.5973, "step": 2742 }, { "epoch": 2.6327412385981757, "grad_norm": 0.7021215501532243, "learning_rate": 5.4132481493202295e-06, "loss": 0.4831, "step": 2743 }, { "epoch": 2.633701392222756, "grad_norm": 1.5830116567629555, "learning_rate": 5.409909557972796e-06, "loss": 0.6203, "step": 2744 }, { "epoch": 2.6346615458473357, "grad_norm": 0.8141050374160547, "learning_rate": 5.406570782621341e-06, "loss": 0.5443, "step": 2745 }, { "epoch": 2.6356216994719155, "grad_norm": 0.8863683462116143, "learning_rate": 5.403231824764609e-06, "loss": 0.5788, "step": 2746 }, { "epoch": 2.6365818530964953, "grad_norm": 0.8573363286507053, "learning_rate": 5.39989268590142e-06, "loss": 0.581, "step": 2747 }, { "epoch": 2.6375420067210755, "grad_norm": 0.7358260550205783, "learning_rate": 5.396553367530679e-06, "loss": 0.6022, "step": 2748 }, { "epoch": 2.6385021603456553, "grad_norm": 1.0326139705413568, "learning_rate": 5.39321387115137e-06, "loss": 0.5372, "step": 2749 }, { "epoch": 2.639462313970235, "grad_norm": 1.1873928771758193, "learning_rate": 5.389874198262559e-06, "loss": 0.6017, "step": 2750 }, { "epoch": 2.6404224675948154, "grad_norm": 0.9950247948406685, "learning_rate": 5.386534350363387e-06, "loss": 0.5539, "step": 2751 }, { "epoch": 2.641382621219395, "grad_norm": 0.8881802730924259, "learning_rate": 5.3831943289530765e-06, "loss": 0.6061, "step": 2752 }, { "epoch": 2.642342774843975, "grad_norm": 0.9093838800837128, "learning_rate": 5.379854135530929e-06, "loss": 0.541, "step": 2753 }, { "epoch": 2.6433029284685547, "grad_norm": 0.8701487928266687, "learning_rate": 5.376513771596319e-06, "loss": 0.6597, "step": 2754 }, { "epoch": 2.644263082093135, "grad_norm": 0.9922308453190182, "learning_rate": 5.373173238648704e-06, "loss": 0.5064, "step": 2755 }, { "epoch": 2.645223235717715, "grad_norm": 1.0391642522533553, "learning_rate": 5.369832538187609e-06, "loss": 0.6382, "step": 2756 }, { "epoch": 2.646183389342295, "grad_norm": 0.967559031591773, "learning_rate": 5.366491671712641e-06, "loss": 0.6506, "step": 2757 }, { "epoch": 2.647143542966875, "grad_norm": 0.8762337889398031, "learning_rate": 5.363150640723479e-06, "loss": 0.469, "step": 2758 }, { "epoch": 2.6481036965914546, "grad_norm": 1.0737484874452934, "learning_rate": 5.359809446719878e-06, "loss": 0.5786, "step": 2759 }, { "epoch": 2.6490638502160344, "grad_norm": 0.7436875311005265, "learning_rate": 5.356468091201661e-06, "loss": 0.4945, "step": 2760 }, { "epoch": 2.6500240038406147, "grad_norm": 0.8990678108202222, "learning_rate": 5.353126575668728e-06, "loss": 0.5796, "step": 2761 }, { "epoch": 2.6509841574651944, "grad_norm": 1.025458749376796, "learning_rate": 5.3497849016210515e-06, "loss": 0.6299, "step": 2762 }, { "epoch": 2.6519443110897742, "grad_norm": 0.7294776151262756, "learning_rate": 5.3464430705586715e-06, "loss": 0.5325, "step": 2763 }, { "epoch": 2.6529044647143545, "grad_norm": 1.3741304384768624, "learning_rate": 5.3431010839817e-06, "loss": 0.5279, "step": 2764 }, { "epoch": 2.6538646183389343, "grad_norm": 0.8679058405591311, "learning_rate": 5.339758943390319e-06, "loss": 0.5525, "step": 2765 }, { "epoch": 2.654824771963514, "grad_norm": 0.8316851449997237, "learning_rate": 5.3364166502847796e-06, "loss": 0.5426, "step": 2766 }, { "epoch": 2.655784925588094, "grad_norm": 0.7841594719336223, "learning_rate": 5.3330742061654014e-06, "loss": 0.5671, "step": 2767 }, { "epoch": 2.656745079212674, "grad_norm": 1.092703644685108, "learning_rate": 5.329731612532575e-06, "loss": 0.606, "step": 2768 }, { "epoch": 2.657705232837254, "grad_norm": 0.7748639884304823, "learning_rate": 5.32638887088675e-06, "loss": 0.4692, "step": 2769 }, { "epoch": 2.658665386461834, "grad_norm": 1.0267798623742261, "learning_rate": 5.323045982728451e-06, "loss": 0.5892, "step": 2770 }, { "epoch": 2.659625540086414, "grad_norm": 0.936608940140023, "learning_rate": 5.319702949558263e-06, "loss": 0.5509, "step": 2771 }, { "epoch": 2.6605856937109937, "grad_norm": 1.584489419641407, "learning_rate": 5.316359772876839e-06, "loss": 0.6624, "step": 2772 }, { "epoch": 2.6615458473355735, "grad_norm": 0.9119548271963658, "learning_rate": 5.313016454184894e-06, "loss": 0.5032, "step": 2773 }, { "epoch": 2.6625060009601538, "grad_norm": 1.1859711312770393, "learning_rate": 5.309672994983209e-06, "loss": 0.5328, "step": 2774 }, { "epoch": 2.6634661545847336, "grad_norm": 0.7731627102146101, "learning_rate": 5.306329396772627e-06, "loss": 0.5275, "step": 2775 }, { "epoch": 2.6644263082093134, "grad_norm": 0.9005536177359821, "learning_rate": 5.302985661054053e-06, "loss": 0.5587, "step": 2776 }, { "epoch": 2.6653864618338936, "grad_norm": 0.8080157454306476, "learning_rate": 5.2996417893284556e-06, "loss": 0.5151, "step": 2777 }, { "epoch": 2.6663466154584734, "grad_norm": 1.231521324385888, "learning_rate": 5.296297783096861e-06, "loss": 0.5159, "step": 2778 }, { "epoch": 2.667306769083053, "grad_norm": 1.3836170636634706, "learning_rate": 5.292953643860358e-06, "loss": 0.6017, "step": 2779 }, { "epoch": 2.668266922707633, "grad_norm": 1.360320557141012, "learning_rate": 5.289609373120096e-06, "loss": 0.54, "step": 2780 }, { "epoch": 2.6692270763322132, "grad_norm": 1.1309304600249424, "learning_rate": 5.286264972377283e-06, "loss": 0.4844, "step": 2781 }, { "epoch": 2.670187229956793, "grad_norm": 0.9339209152689378, "learning_rate": 5.282920443133184e-06, "loss": 0.5409, "step": 2782 }, { "epoch": 2.6711473835813733, "grad_norm": 0.7518714623824867, "learning_rate": 5.279575786889121e-06, "loss": 0.5919, "step": 2783 }, { "epoch": 2.672107537205953, "grad_norm": 0.6388645744984656, "learning_rate": 5.2762310051464775e-06, "loss": 0.4473, "step": 2784 }, { "epoch": 2.673067690830533, "grad_norm": 0.7959384285057761, "learning_rate": 5.272886099406688e-06, "loss": 0.4812, "step": 2785 }, { "epoch": 2.6740278444551127, "grad_norm": 0.8462621375108285, "learning_rate": 5.269541071171246e-06, "loss": 0.5331, "step": 2786 }, { "epoch": 2.674987998079693, "grad_norm": 0.8089505061026713, "learning_rate": 5.266195921941697e-06, "loss": 0.4758, "step": 2787 }, { "epoch": 2.6759481517042727, "grad_norm": 0.8614506137422301, "learning_rate": 5.262850653219643e-06, "loss": 0.5546, "step": 2788 }, { "epoch": 2.6769083053288525, "grad_norm": 1.2418132588100932, "learning_rate": 5.259505266506742e-06, "loss": 0.5958, "step": 2789 }, { "epoch": 2.6778684589534327, "grad_norm": 0.8783716699712195, "learning_rate": 5.256159763304703e-06, "loss": 0.6268, "step": 2790 }, { "epoch": 2.6788286125780125, "grad_norm": 0.7108021668234139, "learning_rate": 5.252814145115281e-06, "loss": 0.498, "step": 2791 }, { "epoch": 2.6797887662025923, "grad_norm": 1.321024270809396, "learning_rate": 5.249468413440293e-06, "loss": 0.5373, "step": 2792 }, { "epoch": 2.680748919827172, "grad_norm": 1.0014282048044487, "learning_rate": 5.246122569781601e-06, "loss": 0.6011, "step": 2793 }, { "epoch": 2.6817090734517524, "grad_norm": 0.8133480626765764, "learning_rate": 5.242776615641119e-06, "loss": 0.5047, "step": 2794 }, { "epoch": 2.682669227076332, "grad_norm": 0.7376771903333289, "learning_rate": 5.2394305525208076e-06, "loss": 0.4946, "step": 2795 }, { "epoch": 2.6836293807009124, "grad_norm": 0.9734249475502849, "learning_rate": 5.236084381922681e-06, "loss": 0.4845, "step": 2796 }, { "epoch": 2.684589534325492, "grad_norm": 1.1032527322019747, "learning_rate": 5.232738105348801e-06, "loss": 0.4999, "step": 2797 }, { "epoch": 2.685549687950072, "grad_norm": 1.610613226653167, "learning_rate": 5.229391724301271e-06, "loss": 0.6577, "step": 2798 }, { "epoch": 2.686509841574652, "grad_norm": 1.2558578840013455, "learning_rate": 5.22604524028225e-06, "loss": 0.6595, "step": 2799 }, { "epoch": 2.687469995199232, "grad_norm": 0.9171232046997707, "learning_rate": 5.222698654793936e-06, "loss": 0.6092, "step": 2800 }, { "epoch": 2.688430148823812, "grad_norm": 0.7244679218257815, "learning_rate": 5.219351969338576e-06, "loss": 0.5639, "step": 2801 }, { "epoch": 2.6893903024483916, "grad_norm": 0.7332134964780334, "learning_rate": 5.2160051854184616e-06, "loss": 0.6096, "step": 2802 }, { "epoch": 2.690350456072972, "grad_norm": 1.119578211364165, "learning_rate": 5.212658304535929e-06, "loss": 0.5494, "step": 2803 }, { "epoch": 2.6913106096975516, "grad_norm": 0.8561629752940173, "learning_rate": 5.209311328193356e-06, "loss": 0.5242, "step": 2804 }, { "epoch": 2.6922707633221314, "grad_norm": 1.2462771012252933, "learning_rate": 5.205964257893166e-06, "loss": 0.6502, "step": 2805 }, { "epoch": 2.6932309169467112, "grad_norm": 1.0899673211575334, "learning_rate": 5.202617095137823e-06, "loss": 0.4467, "step": 2806 }, { "epoch": 2.6941910705712915, "grad_norm": 1.0471169896393497, "learning_rate": 5.199269841429831e-06, "loss": 0.6928, "step": 2807 }, { "epoch": 2.6951512241958713, "grad_norm": 0.7687890470289543, "learning_rate": 5.195922498271737e-06, "loss": 0.5577, "step": 2808 }, { "epoch": 2.6961113778204515, "grad_norm": 1.1421851653504669, "learning_rate": 5.192575067166126e-06, "loss": 0.5977, "step": 2809 }, { "epoch": 2.6970715314450313, "grad_norm": 0.9160958431058133, "learning_rate": 5.189227549615629e-06, "loss": 0.5549, "step": 2810 }, { "epoch": 2.698031685069611, "grad_norm": 0.8784042820906985, "learning_rate": 5.185879947122906e-06, "loss": 0.4855, "step": 2811 }, { "epoch": 2.698991838694191, "grad_norm": 0.907374351738876, "learning_rate": 5.182532261190663e-06, "loss": 0.5592, "step": 2812 }, { "epoch": 2.699951992318771, "grad_norm": 0.8617196191178247, "learning_rate": 5.179184493321637e-06, "loss": 0.584, "step": 2813 }, { "epoch": 2.700912145943351, "grad_norm": 0.9157360299005421, "learning_rate": 5.175836645018608e-06, "loss": 0.472, "step": 2814 }, { "epoch": 2.7018722995679307, "grad_norm": 0.8069890764183622, "learning_rate": 5.1724887177843894e-06, "loss": 0.5988, "step": 2815 }, { "epoch": 2.702832453192511, "grad_norm": 0.9141465325313881, "learning_rate": 5.1691407131218296e-06, "loss": 0.625, "step": 2816 }, { "epoch": 2.7037926068170908, "grad_norm": 0.9740405879378661, "learning_rate": 5.165792632533811e-06, "loss": 0.6689, "step": 2817 }, { "epoch": 2.7047527604416706, "grad_norm": 0.9257389502318615, "learning_rate": 5.162444477523251e-06, "loss": 0.5538, "step": 2818 }, { "epoch": 2.7057129140662504, "grad_norm": 0.8693174235136153, "learning_rate": 5.159096249593104e-06, "loss": 0.6013, "step": 2819 }, { "epoch": 2.7066730676908306, "grad_norm": 1.0151638871843358, "learning_rate": 5.1557479502463495e-06, "loss": 0.5903, "step": 2820 }, { "epoch": 2.7076332213154104, "grad_norm": 0.963733371189027, "learning_rate": 5.152399580986005e-06, "loss": 0.6123, "step": 2821 }, { "epoch": 2.7085933749399906, "grad_norm": 0.8928826681496986, "learning_rate": 5.149051143315118e-06, "loss": 0.645, "step": 2822 }, { "epoch": 2.7095535285645704, "grad_norm": 0.7814790861295348, "learning_rate": 5.1457026387367655e-06, "loss": 0.6192, "step": 2823 }, { "epoch": 2.7105136821891502, "grad_norm": 0.8035770300978268, "learning_rate": 5.142354068754057e-06, "loss": 0.5708, "step": 2824 }, { "epoch": 2.71147383581373, "grad_norm": 0.8409492785492145, "learning_rate": 5.139005434870128e-06, "loss": 0.5876, "step": 2825 }, { "epoch": 2.7124339894383103, "grad_norm": 0.9688326063208984, "learning_rate": 5.135656738588144e-06, "loss": 0.6079, "step": 2826 }, { "epoch": 2.71339414306289, "grad_norm": 0.6590105781103468, "learning_rate": 5.132307981411301e-06, "loss": 0.461, "step": 2827 }, { "epoch": 2.71435429668747, "grad_norm": 0.808921434153706, "learning_rate": 5.128959164842821e-06, "loss": 0.4944, "step": 2828 }, { "epoch": 2.71531445031205, "grad_norm": 0.7226595200570162, "learning_rate": 5.125610290385948e-06, "loss": 0.5775, "step": 2829 }, { "epoch": 2.71627460393663, "grad_norm": 0.7893229307062095, "learning_rate": 5.1222613595439605e-06, "loss": 0.5008, "step": 2830 }, { "epoch": 2.7172347575612097, "grad_norm": 0.8773165982372033, "learning_rate": 5.118912373820154e-06, "loss": 0.5326, "step": 2831 }, { "epoch": 2.7181949111857895, "grad_norm": 0.9320372861634709, "learning_rate": 5.1155633347178545e-06, "loss": 0.5692, "step": 2832 }, { "epoch": 2.7191550648103697, "grad_norm": 1.0053888472872883, "learning_rate": 5.1122142437404085e-06, "loss": 0.6575, "step": 2833 }, { "epoch": 2.7201152184349495, "grad_norm": 0.728102282019473, "learning_rate": 5.1088651023911905e-06, "loss": 0.5049, "step": 2834 }, { "epoch": 2.7210753720595298, "grad_norm": 1.3897214334927501, "learning_rate": 5.10551591217359e-06, "loss": 0.5359, "step": 2835 }, { "epoch": 2.7220355256841096, "grad_norm": 1.0217414671105471, "learning_rate": 5.102166674591024e-06, "loss": 0.5543, "step": 2836 }, { "epoch": 2.7229956793086894, "grad_norm": 1.195603261692095, "learning_rate": 5.09881739114693e-06, "loss": 0.663, "step": 2837 }, { "epoch": 2.723955832933269, "grad_norm": 1.1187204832228432, "learning_rate": 5.095468063344767e-06, "loss": 0.5988, "step": 2838 }, { "epoch": 2.7249159865578494, "grad_norm": 1.4888469570248484, "learning_rate": 5.09211869268801e-06, "loss": 0.4996, "step": 2839 }, { "epoch": 2.725876140182429, "grad_norm": 0.7868526371550016, "learning_rate": 5.088769280680156e-06, "loss": 0.5455, "step": 2840 }, { "epoch": 2.726836293807009, "grad_norm": 0.9019120240545876, "learning_rate": 5.08541982882472e-06, "loss": 0.559, "step": 2841 }, { "epoch": 2.727796447431589, "grad_norm": 0.8401818082603348, "learning_rate": 5.082070338625238e-06, "loss": 0.643, "step": 2842 }, { "epoch": 2.728756601056169, "grad_norm": 0.7851040289981921, "learning_rate": 5.078720811585257e-06, "loss": 0.5511, "step": 2843 }, { "epoch": 2.729716754680749, "grad_norm": 0.7092846800656999, "learning_rate": 5.0753712492083455e-06, "loss": 0.4185, "step": 2844 }, { "epoch": 2.7306769083053286, "grad_norm": 1.008126223633706, "learning_rate": 5.072021652998086e-06, "loss": 0.5307, "step": 2845 }, { "epoch": 2.731637061929909, "grad_norm": 1.105926114399835, "learning_rate": 5.068672024458076e-06, "loss": 0.5738, "step": 2846 }, { "epoch": 2.7325972155544886, "grad_norm": 1.397762428509244, "learning_rate": 5.065322365091928e-06, "loss": 0.4099, "step": 2847 }, { "epoch": 2.733557369179069, "grad_norm": 1.571633951496243, "learning_rate": 5.061972676403269e-06, "loss": 0.5911, "step": 2848 }, { "epoch": 2.7345175228036487, "grad_norm": 1.1694295244708224, "learning_rate": 5.058622959895738e-06, "loss": 0.6348, "step": 2849 }, { "epoch": 2.7354776764282285, "grad_norm": 0.7975148580250856, "learning_rate": 5.055273217072988e-06, "loss": 0.4748, "step": 2850 }, { "epoch": 2.7364378300528083, "grad_norm": 0.8321586651214703, "learning_rate": 5.051923449438679e-06, "loss": 0.5879, "step": 2851 }, { "epoch": 2.7373979836773885, "grad_norm": 0.8560176982837903, "learning_rate": 5.048573658496492e-06, "loss": 0.6089, "step": 2852 }, { "epoch": 2.7383581373019683, "grad_norm": 1.170534424254789, "learning_rate": 5.0452238457501065e-06, "loss": 0.5421, "step": 2853 }, { "epoch": 2.739318290926548, "grad_norm": 0.841248933858448, "learning_rate": 5.0418740127032204e-06, "loss": 0.5878, "step": 2854 }, { "epoch": 2.7402784445511283, "grad_norm": 0.7040702711550074, "learning_rate": 5.038524160859538e-06, "loss": 0.5947, "step": 2855 }, { "epoch": 2.741238598175708, "grad_norm": 0.7051123098241211, "learning_rate": 5.035174291722772e-06, "loss": 0.4224, "step": 2856 }, { "epoch": 2.742198751800288, "grad_norm": 0.9907746475705105, "learning_rate": 5.031824406796642e-06, "loss": 0.5938, "step": 2857 }, { "epoch": 2.7431589054248677, "grad_norm": 0.9696717802654373, "learning_rate": 5.028474507584875e-06, "loss": 0.5924, "step": 2858 }, { "epoch": 2.744119059049448, "grad_norm": 1.0183772449736015, "learning_rate": 5.025124595591206e-06, "loss": 0.5498, "step": 2859 }, { "epoch": 2.7450792126740278, "grad_norm": 0.960736807537049, "learning_rate": 5.0217746723193775e-06, "loss": 0.6512, "step": 2860 }, { "epoch": 2.746039366298608, "grad_norm": 0.8033860365899875, "learning_rate": 5.018424739273129e-06, "loss": 0.5928, "step": 2861 }, { "epoch": 2.746999519923188, "grad_norm": 0.8825194475992075, "learning_rate": 5.015074797956212e-06, "loss": 0.4962, "step": 2862 }, { "epoch": 2.7479596735477676, "grad_norm": 0.8895411647530103, "learning_rate": 5.01172484987238e-06, "loss": 0.5296, "step": 2863 }, { "epoch": 2.7489198271723474, "grad_norm": 0.7201569777669447, "learning_rate": 5.008374896525387e-06, "loss": 0.602, "step": 2864 }, { "epoch": 2.7498799807969276, "grad_norm": 0.9608874513018172, "learning_rate": 5.0050249394189945e-06, "loss": 0.5821, "step": 2865 }, { "epoch": 2.7508401344215074, "grad_norm": 0.8436832906613604, "learning_rate": 5.001674980056959e-06, "loss": 0.5954, "step": 2866 }, { "epoch": 2.7518002880460872, "grad_norm": 0.9651470867076956, "learning_rate": 4.998325019943044e-06, "loss": 0.577, "step": 2867 }, { "epoch": 2.7527604416706675, "grad_norm": 1.3543047348532826, "learning_rate": 4.994975060581007e-06, "loss": 0.5658, "step": 2868 }, { "epoch": 2.7537205952952473, "grad_norm": 0.7662673742866931, "learning_rate": 4.991625103474614e-06, "loss": 0.5551, "step": 2869 }, { "epoch": 2.754680748919827, "grad_norm": 0.7083295895901572, "learning_rate": 4.988275150127622e-06, "loss": 0.5643, "step": 2870 }, { "epoch": 2.755640902544407, "grad_norm": 0.7039324403647123, "learning_rate": 4.984925202043789e-06, "loss": 0.5624, "step": 2871 }, { "epoch": 2.756601056168987, "grad_norm": 0.7802702844927903, "learning_rate": 4.981575260726873e-06, "loss": 0.5711, "step": 2872 }, { "epoch": 2.757561209793567, "grad_norm": 0.9750816948420783, "learning_rate": 4.978225327680625e-06, "loss": 0.5754, "step": 2873 }, { "epoch": 2.758521363418147, "grad_norm": 0.7463700948211263, "learning_rate": 4.974875404408794e-06, "loss": 0.5038, "step": 2874 }, { "epoch": 2.759481517042727, "grad_norm": 0.8248414667964794, "learning_rate": 4.971525492415126e-06, "loss": 0.4985, "step": 2875 }, { "epoch": 2.7604416706673067, "grad_norm": 0.8664310597459299, "learning_rate": 4.96817559320336e-06, "loss": 0.5593, "step": 2876 }, { "epoch": 2.7614018242918865, "grad_norm": 0.934577900059333, "learning_rate": 4.9648257082772305e-06, "loss": 0.535, "step": 2877 }, { "epoch": 2.7623619779164668, "grad_norm": 0.8244692024345123, "learning_rate": 4.9614758391404634e-06, "loss": 0.5333, "step": 2878 }, { "epoch": 2.7633221315410466, "grad_norm": 0.9079560390376263, "learning_rate": 4.9581259872967795e-06, "loss": 0.4444, "step": 2879 }, { "epoch": 2.7642822851656264, "grad_norm": 1.1663016513743478, "learning_rate": 4.954776154249896e-06, "loss": 0.4927, "step": 2880 }, { "epoch": 2.7652424387902066, "grad_norm": 0.966892190898895, "learning_rate": 4.95142634150351e-06, "loss": 0.5422, "step": 2881 }, { "epoch": 2.7662025924147864, "grad_norm": 0.7895857354156512, "learning_rate": 4.94807655056132e-06, "loss": 0.5709, "step": 2882 }, { "epoch": 2.767162746039366, "grad_norm": 1.054363600101862, "learning_rate": 4.944726782927014e-06, "loss": 0.6346, "step": 2883 }, { "epoch": 2.768122899663946, "grad_norm": 1.0806228749514644, "learning_rate": 4.9413770401042625e-06, "loss": 0.5141, "step": 2884 }, { "epoch": 2.769083053288526, "grad_norm": 0.7890337250026997, "learning_rate": 4.938027323596732e-06, "loss": 0.5157, "step": 2885 }, { "epoch": 2.770043206913106, "grad_norm": 1.256800664292196, "learning_rate": 4.934677634908073e-06, "loss": 0.4694, "step": 2886 }, { "epoch": 2.7710033605376863, "grad_norm": 0.6632549902514399, "learning_rate": 4.931327975541925e-06, "loss": 0.4281, "step": 2887 }, { "epoch": 2.771963514162266, "grad_norm": 0.7308443075784864, "learning_rate": 4.927978347001916e-06, "loss": 0.5604, "step": 2888 }, { "epoch": 2.772923667786846, "grad_norm": 0.9276315487258568, "learning_rate": 4.924628750791656e-06, "loss": 0.6186, "step": 2889 }, { "epoch": 2.7738838214114256, "grad_norm": 1.0336478577295947, "learning_rate": 4.9212791884147456e-06, "loss": 0.5694, "step": 2890 }, { "epoch": 2.774843975036006, "grad_norm": 0.9145856284179044, "learning_rate": 4.917929661374764e-06, "loss": 0.566, "step": 2891 }, { "epoch": 2.7758041286605857, "grad_norm": 0.8007582977170373, "learning_rate": 4.9145801711752804e-06, "loss": 0.5652, "step": 2892 }, { "epoch": 2.7767642822851655, "grad_norm": 0.854934153287406, "learning_rate": 4.911230719319846e-06, "loss": 0.5532, "step": 2893 }, { "epoch": 2.7777244359097457, "grad_norm": 0.7359926955892475, "learning_rate": 4.907881307311992e-06, "loss": 0.5847, "step": 2894 }, { "epoch": 2.7786845895343255, "grad_norm": 0.9491844489088328, "learning_rate": 4.904531936655236e-06, "loss": 0.6029, "step": 2895 }, { "epoch": 2.7796447431589053, "grad_norm": 0.8945819467100704, "learning_rate": 4.901182608853071e-06, "loss": 0.5848, "step": 2896 }, { "epoch": 2.780604896783485, "grad_norm": 0.6374272856258161, "learning_rate": 4.897833325408977e-06, "loss": 0.5796, "step": 2897 }, { "epoch": 2.7815650504080653, "grad_norm": 0.8398457896728108, "learning_rate": 4.894484087826413e-06, "loss": 0.5966, "step": 2898 }, { "epoch": 2.782525204032645, "grad_norm": 0.6870030390150745, "learning_rate": 4.891134897608811e-06, "loss": 0.494, "step": 2899 }, { "epoch": 2.7834853576572254, "grad_norm": 1.365923647620089, "learning_rate": 4.887785756259591e-06, "loss": 0.5493, "step": 2900 }, { "epoch": 2.784445511281805, "grad_norm": 1.049676134162727, "learning_rate": 4.884436665282146e-06, "loss": 0.5492, "step": 2901 }, { "epoch": 2.785405664906385, "grad_norm": 0.610684269732182, "learning_rate": 4.881087626179847e-06, "loss": 0.4412, "step": 2902 }, { "epoch": 2.7863658185309648, "grad_norm": 0.6956048540784247, "learning_rate": 4.877738640456042e-06, "loss": 0.5203, "step": 2903 }, { "epoch": 2.787325972155545, "grad_norm": 1.0657536454316456, "learning_rate": 4.874389709614053e-06, "loss": 0.5431, "step": 2904 }, { "epoch": 2.788286125780125, "grad_norm": 0.8930332084538601, "learning_rate": 4.871040835157181e-06, "loss": 0.5242, "step": 2905 }, { "epoch": 2.7892462794047046, "grad_norm": 0.6404812373048175, "learning_rate": 4.8676920185887e-06, "loss": 0.5052, "step": 2906 }, { "epoch": 2.790206433029285, "grad_norm": 0.7799834033541755, "learning_rate": 4.8643432614118565e-06, "loss": 0.4823, "step": 2907 }, { "epoch": 2.7911665866538646, "grad_norm": 0.9515823534050494, "learning_rate": 4.860994565129875e-06, "loss": 0.5635, "step": 2908 }, { "epoch": 2.7921267402784444, "grad_norm": 0.8864932146273218, "learning_rate": 4.857645931245946e-06, "loss": 0.6517, "step": 2909 }, { "epoch": 2.7930868939030242, "grad_norm": 0.7414355794171075, "learning_rate": 4.854297361263235e-06, "loss": 0.5521, "step": 2910 }, { "epoch": 2.7940470475276045, "grad_norm": 1.1493624484777596, "learning_rate": 4.850948856684884e-06, "loss": 0.4942, "step": 2911 }, { "epoch": 2.7950072011521843, "grad_norm": 0.9410159562071284, "learning_rate": 4.8476004190139954e-06, "loss": 0.6257, "step": 2912 }, { "epoch": 2.7959673547767645, "grad_norm": 1.1885655438864364, "learning_rate": 4.844252049753651e-06, "loss": 0.5456, "step": 2913 }, { "epoch": 2.7969275084013443, "grad_norm": 0.7593134742405633, "learning_rate": 4.840903750406898e-06, "loss": 0.5062, "step": 2914 }, { "epoch": 2.797887662025924, "grad_norm": 0.6215229202332143, "learning_rate": 4.8375555224767486e-06, "loss": 0.4695, "step": 2915 }, { "epoch": 2.798847815650504, "grad_norm": 0.8049803631652971, "learning_rate": 4.83420736746619e-06, "loss": 0.5375, "step": 2916 }, { "epoch": 2.799807969275084, "grad_norm": 1.0645577336663277, "learning_rate": 4.830859286878172e-06, "loss": 0.5607, "step": 2917 }, { "epoch": 2.800768122899664, "grad_norm": 1.246747683297623, "learning_rate": 4.8275112822156105e-06, "loss": 0.6334, "step": 2918 }, { "epoch": 2.8017282765242437, "grad_norm": 0.9169780860354989, "learning_rate": 4.824163354981393e-06, "loss": 0.604, "step": 2919 }, { "epoch": 2.802688430148824, "grad_norm": 0.8145005139939441, "learning_rate": 4.820815506678364e-06, "loss": 0.5147, "step": 2920 }, { "epoch": 2.8036485837734038, "grad_norm": 0.9609885656816305, "learning_rate": 4.81746773880934e-06, "loss": 0.6058, "step": 2921 }, { "epoch": 2.8046087373979836, "grad_norm": 0.9756321656329406, "learning_rate": 4.8141200528770955e-06, "loss": 0.4774, "step": 2922 }, { "epoch": 2.8055688910225633, "grad_norm": 0.8651505171413988, "learning_rate": 4.810772450384372e-06, "loss": 0.5511, "step": 2923 }, { "epoch": 2.8065290446471436, "grad_norm": 0.7712815911405732, "learning_rate": 4.807424932833875e-06, "loss": 0.571, "step": 2924 }, { "epoch": 2.8074891982717234, "grad_norm": 0.8245462040797099, "learning_rate": 4.8040775017282644e-06, "loss": 0.5325, "step": 2925 }, { "epoch": 2.8084493518963036, "grad_norm": 0.790485271315528, "learning_rate": 4.8007301585701726e-06, "loss": 0.591, "step": 2926 }, { "epoch": 2.8094095055208834, "grad_norm": 1.2168603768739528, "learning_rate": 4.7973829048621786e-06, "loss": 0.6272, "step": 2927 }, { "epoch": 2.810369659145463, "grad_norm": 0.9680261008639863, "learning_rate": 4.794035742106834e-06, "loss": 0.6073, "step": 2928 }, { "epoch": 2.811329812770043, "grad_norm": 0.8368659693760014, "learning_rate": 4.7906886718066445e-06, "loss": 0.6069, "step": 2929 }, { "epoch": 2.8122899663946233, "grad_norm": 1.0342809004333435, "learning_rate": 4.7873416954640725e-06, "loss": 0.5349, "step": 2930 }, { "epoch": 2.813250120019203, "grad_norm": 0.9159945303057621, "learning_rate": 4.783994814581539e-06, "loss": 0.606, "step": 2931 }, { "epoch": 2.814210273643783, "grad_norm": 0.8451665905286943, "learning_rate": 4.7806480306614256e-06, "loss": 0.6032, "step": 2932 }, { "epoch": 2.815170427268363, "grad_norm": 0.8522032466271976, "learning_rate": 4.7773013452060654e-06, "loss": 0.5628, "step": 2933 }, { "epoch": 2.816130580892943, "grad_norm": 0.7265664307278741, "learning_rate": 4.773954759717752e-06, "loss": 0.5498, "step": 2934 }, { "epoch": 2.8170907345175227, "grad_norm": 0.9271089319535823, "learning_rate": 4.7706082756987295e-06, "loss": 0.5315, "step": 2935 }, { "epoch": 2.8180508881421025, "grad_norm": 0.9098842188979439, "learning_rate": 4.7672618946512e-06, "loss": 0.5581, "step": 2936 }, { "epoch": 2.8190110417666827, "grad_norm": 1.0052221727957513, "learning_rate": 4.76391561807732e-06, "loss": 0.569, "step": 2937 }, { "epoch": 2.8199711953912625, "grad_norm": 0.9875836515585031, "learning_rate": 4.760569447479193e-06, "loss": 0.5974, "step": 2938 }, { "epoch": 2.8209313490158427, "grad_norm": 1.2131092460051147, "learning_rate": 4.757223384358884e-06, "loss": 0.6087, "step": 2939 }, { "epoch": 2.8218915026404225, "grad_norm": 0.9007497448200027, "learning_rate": 4.753877430218401e-06, "loss": 0.4907, "step": 2940 }, { "epoch": 2.8228516562650023, "grad_norm": 0.707046479052174, "learning_rate": 4.750531586559707e-06, "loss": 0.5913, "step": 2941 }, { "epoch": 2.823811809889582, "grad_norm": 0.9528809539484266, "learning_rate": 4.7471858548847214e-06, "loss": 0.5105, "step": 2942 }, { "epoch": 2.8247719635141624, "grad_norm": 0.7577805298375715, "learning_rate": 4.743840236695299e-06, "loss": 0.5788, "step": 2943 }, { "epoch": 2.825732117138742, "grad_norm": 0.9244880757595562, "learning_rate": 4.740494733493258e-06, "loss": 0.4742, "step": 2944 }, { "epoch": 2.826692270763322, "grad_norm": 0.8279435815239258, "learning_rate": 4.737149346780358e-06, "loss": 0.4702, "step": 2945 }, { "epoch": 2.827652424387902, "grad_norm": 0.8539386777392587, "learning_rate": 4.733804078058305e-06, "loss": 0.5507, "step": 2946 }, { "epoch": 2.828612578012482, "grad_norm": 0.8504197947771941, "learning_rate": 4.730458928828757e-06, "loss": 0.5914, "step": 2947 }, { "epoch": 2.829572731637062, "grad_norm": 0.7556468513286173, "learning_rate": 4.7271139005933144e-06, "loss": 0.5095, "step": 2948 }, { "epoch": 2.8305328852616416, "grad_norm": 0.6642307594267663, "learning_rate": 4.723768994853523e-06, "loss": 0.4961, "step": 2949 }, { "epoch": 2.831493038886222, "grad_norm": 0.7022323970501738, "learning_rate": 4.72042421311088e-06, "loss": 0.4912, "step": 2950 }, { "epoch": 2.8324531925108016, "grad_norm": 1.21788947494834, "learning_rate": 4.717079556866818e-06, "loss": 0.5952, "step": 2951 }, { "epoch": 2.833413346135382, "grad_norm": 1.1035645221232722, "learning_rate": 4.713735027622719e-06, "loss": 0.5295, "step": 2952 }, { "epoch": 2.8343734997599617, "grad_norm": 1.1009874366071952, "learning_rate": 4.710390626879906e-06, "loss": 0.5366, "step": 2953 }, { "epoch": 2.8353336533845415, "grad_norm": 1.090504542234294, "learning_rate": 4.707046356139643e-06, "loss": 0.637, "step": 2954 }, { "epoch": 2.8362938070091213, "grad_norm": 0.7857443718395714, "learning_rate": 4.703702216903142e-06, "loss": 0.6318, "step": 2955 }, { "epoch": 2.8372539606337015, "grad_norm": 0.9359143981557753, "learning_rate": 4.700358210671545e-06, "loss": 0.504, "step": 2956 }, { "epoch": 2.8382141142582813, "grad_norm": 0.7154175788225354, "learning_rate": 4.697014338945947e-06, "loss": 0.6103, "step": 2957 }, { "epoch": 2.839174267882861, "grad_norm": 0.744301528871271, "learning_rate": 4.693670603227374e-06, "loss": 0.5205, "step": 2958 }, { "epoch": 2.8401344215074413, "grad_norm": 1.4775796658773779, "learning_rate": 4.690327005016792e-06, "loss": 0.6151, "step": 2959 }, { "epoch": 2.841094575132021, "grad_norm": 0.75395867813605, "learning_rate": 4.686983545815107e-06, "loss": 0.6039, "step": 2960 }, { "epoch": 2.842054728756601, "grad_norm": 0.7958031610315721, "learning_rate": 4.683640227123162e-06, "loss": 0.4877, "step": 2961 }, { "epoch": 2.8430148823811807, "grad_norm": 0.7135724954806881, "learning_rate": 4.680297050441737e-06, "loss": 0.478, "step": 2962 }, { "epoch": 2.843975036005761, "grad_norm": 1.0095077857080448, "learning_rate": 4.67695401727155e-06, "loss": 0.6961, "step": 2963 }, { "epoch": 2.8449351896303408, "grad_norm": 0.6675177883530261, "learning_rate": 4.6736111291132505e-06, "loss": 0.5518, "step": 2964 }, { "epoch": 2.845895343254921, "grad_norm": 0.7735348790192751, "learning_rate": 4.670268387467427e-06, "loss": 0.6028, "step": 2965 }, { "epoch": 2.846855496879501, "grad_norm": 1.2922958944779253, "learning_rate": 4.666925793834599e-06, "loss": 0.5735, "step": 2966 }, { "epoch": 2.8478156505040806, "grad_norm": 1.3090113202953138, "learning_rate": 4.663583349715222e-06, "loss": 0.6098, "step": 2967 }, { "epoch": 2.8487758041286604, "grad_norm": 0.8693267385867451, "learning_rate": 4.660241056609685e-06, "loss": 0.6087, "step": 2968 }, { "epoch": 2.8497359577532406, "grad_norm": 0.7626962950150373, "learning_rate": 4.656898916018302e-06, "loss": 0.5031, "step": 2969 }, { "epoch": 2.8506961113778204, "grad_norm": 0.9387472047391705, "learning_rate": 4.653556929441332e-06, "loss": 0.6107, "step": 2970 }, { "epoch": 2.8516562650024007, "grad_norm": 0.822134612182305, "learning_rate": 4.650215098378949e-06, "loss": 0.564, "step": 2971 }, { "epoch": 2.8526164186269805, "grad_norm": 1.0861807469186362, "learning_rate": 4.646873424331272e-06, "loss": 0.5668, "step": 2972 }, { "epoch": 2.8535765722515603, "grad_norm": 0.9387023863941588, "learning_rate": 4.64353190879834e-06, "loss": 0.4555, "step": 2973 }, { "epoch": 2.85453672587614, "grad_norm": 0.7423575997544507, "learning_rate": 4.640190553280123e-06, "loss": 0.5369, "step": 2974 }, { "epoch": 2.85549687950072, "grad_norm": 0.9139717909126149, "learning_rate": 4.636849359276521e-06, "loss": 0.5826, "step": 2975 }, { "epoch": 2.8564570331253, "grad_norm": 0.6618828469085111, "learning_rate": 4.6335083282873605e-06, "loss": 0.5485, "step": 2976 }, { "epoch": 2.85741718674988, "grad_norm": 0.8423553165191531, "learning_rate": 4.6301674618123924e-06, "loss": 0.5662, "step": 2977 }, { "epoch": 2.85837734037446, "grad_norm": 1.3116315111158992, "learning_rate": 4.626826761351299e-06, "loss": 0.5887, "step": 2978 }, { "epoch": 2.85933749399904, "grad_norm": 0.7216634211314839, "learning_rate": 4.623486228403682e-06, "loss": 0.5303, "step": 2979 }, { "epoch": 2.8602976476236197, "grad_norm": 1.7472563573004827, "learning_rate": 4.620145864469073e-06, "loss": 0.551, "step": 2980 }, { "epoch": 2.8612578012481995, "grad_norm": 0.9186089447804818, "learning_rate": 4.616805671046925e-06, "loss": 0.6446, "step": 2981 }, { "epoch": 2.8622179548727797, "grad_norm": 0.8418834914359357, "learning_rate": 4.613465649636615e-06, "loss": 0.623, "step": 2982 }, { "epoch": 2.8631781084973595, "grad_norm": 0.7588486159923138, "learning_rate": 4.610125801737443e-06, "loss": 0.6225, "step": 2983 }, { "epoch": 2.86413826212194, "grad_norm": 0.696507623662516, "learning_rate": 4.6067861288486315e-06, "loss": 0.521, "step": 2984 }, { "epoch": 2.8650984157465196, "grad_norm": 0.8988956072299208, "learning_rate": 4.603446632469322e-06, "loss": 0.5921, "step": 2985 }, { "epoch": 2.8660585693710994, "grad_norm": 0.8788639031902122, "learning_rate": 4.600107314098583e-06, "loss": 0.5654, "step": 2986 }, { "epoch": 2.867018722995679, "grad_norm": 0.9251783326665547, "learning_rate": 4.596768175235393e-06, "loss": 0.4927, "step": 2987 }, { "epoch": 2.867978876620259, "grad_norm": 0.9660593021315126, "learning_rate": 4.59342921737866e-06, "loss": 0.549, "step": 2988 }, { "epoch": 2.868939030244839, "grad_norm": 0.7790559281819555, "learning_rate": 4.590090442027206e-06, "loss": 0.569, "step": 2989 }, { "epoch": 2.869899183869419, "grad_norm": 1.107689572964066, "learning_rate": 4.5867518506797705e-06, "loss": 0.6251, "step": 2990 }, { "epoch": 2.8708593374939992, "grad_norm": 0.8992345552855489, "learning_rate": 4.583413444835013e-06, "loss": 0.5276, "step": 2991 }, { "epoch": 2.871819491118579, "grad_norm": 1.0630113479794046, "learning_rate": 4.580075225991505e-06, "loss": 0.5771, "step": 2992 }, { "epoch": 2.872779644743159, "grad_norm": 0.7625696012771623, "learning_rate": 4.576737195647741e-06, "loss": 0.5733, "step": 2993 }, { "epoch": 2.8737397983677386, "grad_norm": 1.696809968007704, "learning_rate": 4.573399355302124e-06, "loss": 0.5933, "step": 2994 }, { "epoch": 2.874699951992319, "grad_norm": 0.9657290911233783, "learning_rate": 4.5700617064529774e-06, "loss": 0.5904, "step": 2995 }, { "epoch": 2.8756601056168987, "grad_norm": 0.6908617075298776, "learning_rate": 4.566724250598535e-06, "loss": 0.5241, "step": 2996 }, { "epoch": 2.876620259241479, "grad_norm": 0.8748572009271128, "learning_rate": 4.563386989236944e-06, "loss": 0.6677, "step": 2997 }, { "epoch": 2.8775804128660587, "grad_norm": 0.8007042073031894, "learning_rate": 4.560049923866266e-06, "loss": 0.5498, "step": 2998 }, { "epoch": 2.8785405664906385, "grad_norm": 0.9645700821188413, "learning_rate": 4.556713055984477e-06, "loss": 0.5645, "step": 2999 }, { "epoch": 2.8795007201152183, "grad_norm": 1.0625933416796327, "learning_rate": 4.5533763870894555e-06, "loss": 0.5769, "step": 3000 }, { "epoch": 2.8804608737397985, "grad_norm": 1.3233424171655706, "learning_rate": 4.5500399186790034e-06, "loss": 0.5333, "step": 3001 }, { "epoch": 2.8814210273643783, "grad_norm": 0.7491406377776249, "learning_rate": 4.5467036522508185e-06, "loss": 0.5327, "step": 3002 }, { "epoch": 2.882381180988958, "grad_norm": 1.1545666124710796, "learning_rate": 4.54336758930252e-06, "loss": 0.5766, "step": 3003 }, { "epoch": 2.8833413346135384, "grad_norm": 1.0331207906432043, "learning_rate": 4.54003173133163e-06, "loss": 0.5242, "step": 3004 }, { "epoch": 2.884301488238118, "grad_norm": 0.7180428648578799, "learning_rate": 4.536696079835578e-06, "loss": 0.551, "step": 3005 }, { "epoch": 2.885261641862698, "grad_norm": 0.8230022663619527, "learning_rate": 4.5333606363117035e-06, "loss": 0.5341, "step": 3006 }, { "epoch": 2.8862217954872778, "grad_norm": 0.9287200437678712, "learning_rate": 4.530025402257252e-06, "loss": 0.5706, "step": 3007 }, { "epoch": 2.887181949111858, "grad_norm": 0.9488256200060076, "learning_rate": 4.526690379169374e-06, "loss": 0.6282, "step": 3008 }, { "epoch": 2.888142102736438, "grad_norm": 1.0621611700707652, "learning_rate": 4.5233555685451245e-06, "loss": 0.5177, "step": 3009 }, { "epoch": 2.889102256361018, "grad_norm": 0.9802841980729945, "learning_rate": 4.5200209718814656e-06, "loss": 0.5483, "step": 3010 }, { "epoch": 2.890062409985598, "grad_norm": 0.7664223538806171, "learning_rate": 4.516686590675261e-06, "loss": 0.5091, "step": 3011 }, { "epoch": 2.8910225636101776, "grad_norm": 0.8737850764786155, "learning_rate": 4.513352426423281e-06, "loss": 0.657, "step": 3012 }, { "epoch": 2.8919827172347574, "grad_norm": 0.8983917639565002, "learning_rate": 4.510018480622193e-06, "loss": 0.56, "step": 3013 }, { "epoch": 2.8929428708593377, "grad_norm": 0.8866208087910415, "learning_rate": 4.506684754768573e-06, "loss": 0.6134, "step": 3014 }, { "epoch": 2.8939030244839175, "grad_norm": 0.7830629473468743, "learning_rate": 4.503351250358893e-06, "loss": 0.4617, "step": 3015 }, { "epoch": 2.8948631781084972, "grad_norm": 0.8773571375587473, "learning_rate": 4.500017968889525e-06, "loss": 0.6343, "step": 3016 }, { "epoch": 2.8958233317330775, "grad_norm": 1.1866691845024195, "learning_rate": 4.4966849118567515e-06, "loss": 0.648, "step": 3017 }, { "epoch": 2.8967834853576573, "grad_norm": 0.8127023089763843, "learning_rate": 4.4933520807567365e-06, "loss": 0.4477, "step": 3018 }, { "epoch": 2.897743638982237, "grad_norm": 0.8609146103034423, "learning_rate": 4.490019477085558e-06, "loss": 0.5206, "step": 3019 }, { "epoch": 2.898703792606817, "grad_norm": 0.8419273466100244, "learning_rate": 4.4866871023391865e-06, "loss": 0.4827, "step": 3020 }, { "epoch": 2.899663946231397, "grad_norm": 0.9596251209657255, "learning_rate": 4.483354958013485e-06, "loss": 0.5657, "step": 3021 }, { "epoch": 2.900624099855977, "grad_norm": 1.1356733688319245, "learning_rate": 4.480023045604223e-06, "loss": 0.5793, "step": 3022 }, { "epoch": 2.901584253480557, "grad_norm": 0.7257535296057803, "learning_rate": 4.476691366607056e-06, "loss": 0.5274, "step": 3023 }, { "epoch": 2.902544407105137, "grad_norm": 0.9158928207818822, "learning_rate": 4.473359922517541e-06, "loss": 0.5827, "step": 3024 }, { "epoch": 2.9035045607297167, "grad_norm": 0.736440836511334, "learning_rate": 4.4700287148311275e-06, "loss": 0.5323, "step": 3025 }, { "epoch": 2.9044647143542965, "grad_norm": 0.8102827622366076, "learning_rate": 4.466697745043158e-06, "loss": 0.5203, "step": 3026 }, { "epoch": 2.905424867978877, "grad_norm": 0.9815044676421529, "learning_rate": 4.463367014648872e-06, "loss": 0.5699, "step": 3027 }, { "epoch": 2.9063850216034566, "grad_norm": 1.1984427963246427, "learning_rate": 4.460036525143395e-06, "loss": 0.5296, "step": 3028 }, { "epoch": 2.9073451752280364, "grad_norm": 0.7733046921924837, "learning_rate": 4.4567062780217486e-06, "loss": 0.546, "step": 3029 }, { "epoch": 2.9083053288526166, "grad_norm": 0.820213474691907, "learning_rate": 4.45337627477885e-06, "loss": 0.5485, "step": 3030 }, { "epoch": 2.9092654824771964, "grad_norm": 0.7822305392642656, "learning_rate": 4.450046516909496e-06, "loss": 0.5664, "step": 3031 }, { "epoch": 2.910225636101776, "grad_norm": 0.7827655786987602, "learning_rate": 4.446717005908382e-06, "loss": 0.488, "step": 3032 }, { "epoch": 2.911185789726356, "grad_norm": 0.942969704529837, "learning_rate": 4.4433877432700915e-06, "loss": 0.5989, "step": 3033 }, { "epoch": 2.9121459433509362, "grad_norm": 0.8821968416656397, "learning_rate": 4.440058730489092e-06, "loss": 0.5678, "step": 3034 }, { "epoch": 2.913106096975516, "grad_norm": 0.7842272400743141, "learning_rate": 4.436729969059745e-06, "loss": 0.5346, "step": 3035 }, { "epoch": 2.9140662506000963, "grad_norm": 0.9521504538228132, "learning_rate": 4.433401460476291e-06, "loss": 0.6054, "step": 3036 }, { "epoch": 2.915026404224676, "grad_norm": 0.8423067660586891, "learning_rate": 4.430073206232865e-06, "loss": 0.5776, "step": 3037 }, { "epoch": 2.915986557849256, "grad_norm": 0.854079996496443, "learning_rate": 4.426745207823486e-06, "loss": 0.5988, "step": 3038 }, { "epoch": 2.9169467114738357, "grad_norm": 0.7997218722434373, "learning_rate": 4.4234174667420546e-06, "loss": 0.5843, "step": 3039 }, { "epoch": 2.917906865098416, "grad_norm": 0.7689503999427638, "learning_rate": 4.42008998448236e-06, "loss": 0.4944, "step": 3040 }, { "epoch": 2.9188670187229957, "grad_norm": 0.7520498230350355, "learning_rate": 4.416762762538072e-06, "loss": 0.4381, "step": 3041 }, { "epoch": 2.9198271723475755, "grad_norm": 0.9192296083733013, "learning_rate": 4.413435802402744e-06, "loss": 0.4744, "step": 3042 }, { "epoch": 2.9207873259721557, "grad_norm": 0.8373064374482215, "learning_rate": 4.410109105569817e-06, "loss": 0.5177, "step": 3043 }, { "epoch": 2.9217474795967355, "grad_norm": 1.634991515511753, "learning_rate": 4.406782673532604e-06, "loss": 0.6704, "step": 3044 }, { "epoch": 2.9227076332213153, "grad_norm": 0.7616665811224211, "learning_rate": 4.403456507784311e-06, "loss": 0.5906, "step": 3045 }, { "epoch": 2.923667786845895, "grad_norm": 1.1806443013710728, "learning_rate": 4.400130609818012e-06, "loss": 0.5958, "step": 3046 }, { "epoch": 2.9246279404704754, "grad_norm": 1.0525250906131438, "learning_rate": 4.3968049811266705e-06, "loss": 0.7535, "step": 3047 }, { "epoch": 2.925588094095055, "grad_norm": 0.7861272528265395, "learning_rate": 4.393479623203126e-06, "loss": 0.5988, "step": 3048 }, { "epoch": 2.9265482477196354, "grad_norm": 0.9185543041431213, "learning_rate": 4.390154537540095e-06, "loss": 0.5514, "step": 3049 }, { "epoch": 2.927508401344215, "grad_norm": 0.7535227382010293, "learning_rate": 4.386829725630173e-06, "loss": 0.4276, "step": 3050 }, { "epoch": 2.928468554968795, "grad_norm": 0.7367060733379464, "learning_rate": 4.383505188965835e-06, "loss": 0.5135, "step": 3051 }, { "epoch": 2.929428708593375, "grad_norm": 0.9582746104583708, "learning_rate": 4.380180929039424e-06, "loss": 0.5585, "step": 3052 }, { "epoch": 2.930388862217955, "grad_norm": 0.7012295578138237, "learning_rate": 4.376856947343171e-06, "loss": 0.5109, "step": 3053 }, { "epoch": 2.931349015842535, "grad_norm": 1.118823337588974, "learning_rate": 4.3735332453691725e-06, "loss": 0.5081, "step": 3054 }, { "epoch": 2.9323091694671146, "grad_norm": 1.095235456846253, "learning_rate": 4.370209824609403e-06, "loss": 0.5053, "step": 3055 }, { "epoch": 2.933269323091695, "grad_norm": 0.910221621472935, "learning_rate": 4.366886686555713e-06, "loss": 0.5972, "step": 3056 }, { "epoch": 2.9342294767162747, "grad_norm": 1.0840211467470768, "learning_rate": 4.36356383269982e-06, "loss": 0.5692, "step": 3057 }, { "epoch": 2.9351896303408544, "grad_norm": 1.1359764068099254, "learning_rate": 4.360241264533321e-06, "loss": 0.5988, "step": 3058 }, { "epoch": 2.9361497839654342, "grad_norm": 0.848376172661195, "learning_rate": 4.356918983547678e-06, "loss": 0.5465, "step": 3059 }, { "epoch": 2.9371099375900145, "grad_norm": 0.9378347948271931, "learning_rate": 4.353596991234228e-06, "loss": 0.5223, "step": 3060 }, { "epoch": 2.9380700912145943, "grad_norm": 0.8389768433136485, "learning_rate": 4.3502752890841825e-06, "loss": 0.5422, "step": 3061 }, { "epoch": 2.9390302448391745, "grad_norm": 0.7271534617602236, "learning_rate": 4.3469538785886115e-06, "loss": 0.4755, "step": 3062 }, { "epoch": 2.9399903984637543, "grad_norm": 0.6580027897675849, "learning_rate": 4.343632761238465e-06, "loss": 0.4704, "step": 3063 }, { "epoch": 2.940950552088334, "grad_norm": 0.7406099944513708, "learning_rate": 4.340311938524556e-06, "loss": 0.545, "step": 3064 }, { "epoch": 2.941910705712914, "grad_norm": 0.8381179988535694, "learning_rate": 4.336991411937566e-06, "loss": 0.6257, "step": 3065 }, { "epoch": 2.942870859337494, "grad_norm": 0.9060017750846274, "learning_rate": 4.333671182968045e-06, "loss": 0.613, "step": 3066 }, { "epoch": 2.943831012962074, "grad_norm": 0.7239155922614993, "learning_rate": 4.330351253106405e-06, "loss": 0.5647, "step": 3067 }, { "epoch": 2.9447911665866537, "grad_norm": 1.0692845527099746, "learning_rate": 4.327031623842931e-06, "loss": 0.4921, "step": 3068 }, { "epoch": 2.945751320211234, "grad_norm": 1.0283924631238515, "learning_rate": 4.323712296667768e-06, "loss": 0.6386, "step": 3069 }, { "epoch": 2.9467114738358138, "grad_norm": 1.110624516674431, "learning_rate": 4.320393273070924e-06, "loss": 0.6223, "step": 3070 }, { "epoch": 2.9476716274603936, "grad_norm": 1.2561554891592648, "learning_rate": 4.317074554542277e-06, "loss": 0.5437, "step": 3071 }, { "epoch": 2.9486317810849734, "grad_norm": 0.9797258682159491, "learning_rate": 4.313756142571561e-06, "loss": 0.6658, "step": 3072 }, { "epoch": 2.9495919347095536, "grad_norm": 0.8856876070128129, "learning_rate": 4.310438038648376e-06, "loss": 0.6137, "step": 3073 }, { "epoch": 2.9505520883341334, "grad_norm": 1.48213932746423, "learning_rate": 4.307120244262189e-06, "loss": 0.6744, "step": 3074 }, { "epoch": 2.9515122419587136, "grad_norm": 0.9698841733385383, "learning_rate": 4.303802760902313e-06, "loss": 0.536, "step": 3075 }, { "epoch": 2.9524723955832934, "grad_norm": 0.8115982188628449, "learning_rate": 4.300485590057939e-06, "loss": 0.4497, "step": 3076 }, { "epoch": 2.9534325492078732, "grad_norm": 1.51167064020923, "learning_rate": 4.2971687332181035e-06, "loss": 0.6824, "step": 3077 }, { "epoch": 2.954392702832453, "grad_norm": 0.968265487404137, "learning_rate": 4.293852191871712e-06, "loss": 0.6123, "step": 3078 }, { "epoch": 2.9553528564570333, "grad_norm": 0.8711958553443204, "learning_rate": 4.290535967507524e-06, "loss": 0.493, "step": 3079 }, { "epoch": 2.956313010081613, "grad_norm": 0.7396035789083081, "learning_rate": 4.287220061614155e-06, "loss": 0.5815, "step": 3080 }, { "epoch": 2.957273163706193, "grad_norm": 1.2598183533041238, "learning_rate": 4.283904475680082e-06, "loss": 0.533, "step": 3081 }, { "epoch": 2.958233317330773, "grad_norm": 1.0410618782714973, "learning_rate": 4.280589211193635e-06, "loss": 0.5511, "step": 3082 }, { "epoch": 2.959193470955353, "grad_norm": 0.7722332288541551, "learning_rate": 4.277274269643002e-06, "loss": 0.4078, "step": 3083 }, { "epoch": 2.9601536245799327, "grad_norm": 1.0244536378051896, "learning_rate": 4.273959652516223e-06, "loss": 0.4599, "step": 3084 }, { "epoch": 2.9611137782045125, "grad_norm": 0.9430748120335686, "learning_rate": 4.270645361301195e-06, "loss": 0.5295, "step": 3085 }, { "epoch": 2.9620739318290927, "grad_norm": 0.7919772009541621, "learning_rate": 4.267331397485666e-06, "loss": 0.5861, "step": 3086 }, { "epoch": 2.9630340854536725, "grad_norm": 0.7286247520550703, "learning_rate": 4.264017762557245e-06, "loss": 0.5382, "step": 3087 }, { "epoch": 2.9639942390782528, "grad_norm": 1.0140044662429655, "learning_rate": 4.260704458003379e-06, "loss": 0.6296, "step": 3088 }, { "epoch": 2.9649543927028326, "grad_norm": 0.7602106731502799, "learning_rate": 4.257391485311384e-06, "loss": 0.5675, "step": 3089 }, { "epoch": 2.9659145463274124, "grad_norm": 0.8413732338677852, "learning_rate": 4.254078845968409e-06, "loss": 0.6508, "step": 3090 }, { "epoch": 2.966874699951992, "grad_norm": 0.8027996847827238, "learning_rate": 4.2507665414614675e-06, "loss": 0.6144, "step": 3091 }, { "epoch": 2.9678348535765724, "grad_norm": 1.4568027839208566, "learning_rate": 4.247454573277418e-06, "loss": 0.5807, "step": 3092 }, { "epoch": 2.968795007201152, "grad_norm": 1.0113821892746118, "learning_rate": 4.244142942902965e-06, "loss": 0.5808, "step": 3093 }, { "epoch": 2.969755160825732, "grad_norm": 1.2145499790869638, "learning_rate": 4.240831651824665e-06, "loss": 0.5034, "step": 3094 }, { "epoch": 2.9707153144503122, "grad_norm": 0.93731701641738, "learning_rate": 4.237520701528923e-06, "loss": 0.5095, "step": 3095 }, { "epoch": 2.971675468074892, "grad_norm": 0.7795700982337397, "learning_rate": 4.234210093501987e-06, "loss": 0.5753, "step": 3096 }, { "epoch": 2.972635621699472, "grad_norm": 0.9477866505935786, "learning_rate": 4.230899829229955e-06, "loss": 0.6203, "step": 3097 }, { "epoch": 2.9735957753240516, "grad_norm": 1.0507790316056922, "learning_rate": 4.2275899101987665e-06, "loss": 0.5305, "step": 3098 }, { "epoch": 2.974555928948632, "grad_norm": 0.9011922794872462, "learning_rate": 4.2242803378942115e-06, "loss": 0.6035, "step": 3099 }, { "epoch": 2.9755160825732117, "grad_norm": 0.9973969736986135, "learning_rate": 4.220971113801921e-06, "loss": 0.5325, "step": 3100 }, { "epoch": 2.976476236197792, "grad_norm": 0.7834989461625173, "learning_rate": 4.2176622394073705e-06, "loss": 0.5007, "step": 3101 }, { "epoch": 2.9774363898223717, "grad_norm": 0.7736470945485645, "learning_rate": 4.214353716195879e-06, "loss": 0.4558, "step": 3102 }, { "epoch": 2.9783965434469515, "grad_norm": 1.0824067058134101, "learning_rate": 4.211045545652604e-06, "loss": 0.5026, "step": 3103 }, { "epoch": 2.9793566970715313, "grad_norm": 1.341244268603158, "learning_rate": 4.207737729262549e-06, "loss": 0.5895, "step": 3104 }, { "epoch": 2.9803168506961115, "grad_norm": 0.8530386991797212, "learning_rate": 4.2044302685105635e-06, "loss": 0.5952, "step": 3105 }, { "epoch": 2.9812770043206913, "grad_norm": 0.7481330647138021, "learning_rate": 4.201123164881321e-06, "loss": 0.5358, "step": 3106 }, { "epoch": 2.982237157945271, "grad_norm": 0.9650026168361474, "learning_rate": 4.1978164198593524e-06, "loss": 0.5355, "step": 3107 }, { "epoch": 2.9831973115698514, "grad_norm": 0.9061924416961308, "learning_rate": 4.194510034929019e-06, "loss": 0.5019, "step": 3108 }, { "epoch": 2.984157465194431, "grad_norm": 0.8269048720466765, "learning_rate": 4.19120401157452e-06, "loss": 0.6431, "step": 3109 }, { "epoch": 2.985117618819011, "grad_norm": 0.802448757130247, "learning_rate": 4.187898351279895e-06, "loss": 0.5511, "step": 3110 }, { "epoch": 2.9860777724435907, "grad_norm": 0.9608564821936294, "learning_rate": 4.184593055529018e-06, "loss": 0.5663, "step": 3111 }, { "epoch": 2.987037926068171, "grad_norm": 0.8180127169444024, "learning_rate": 4.181288125805603e-06, "loss": 0.5261, "step": 3112 }, { "epoch": 2.9879980796927508, "grad_norm": 0.8770528578600988, "learning_rate": 4.177983563593198e-06, "loss": 0.4581, "step": 3113 }, { "epoch": 2.988958233317331, "grad_norm": 1.0550492786287162, "learning_rate": 4.174679370375182e-06, "loss": 0.5784, "step": 3114 }, { "epoch": 2.989918386941911, "grad_norm": 0.6880941476581485, "learning_rate": 4.1713755476347766e-06, "loss": 0.4943, "step": 3115 }, { "epoch": 2.9908785405664906, "grad_norm": 0.7741011933555306, "learning_rate": 4.168072096855029e-06, "loss": 0.6249, "step": 3116 }, { "epoch": 2.9918386941910704, "grad_norm": 0.7358290128452621, "learning_rate": 4.164769019518823e-06, "loss": 0.6109, "step": 3117 }, { "epoch": 2.9927988478156506, "grad_norm": 0.8280945416323906, "learning_rate": 4.16146631710888e-06, "loss": 0.6321, "step": 3118 }, { "epoch": 2.9937590014402304, "grad_norm": 0.752596119077195, "learning_rate": 4.15816399110774e-06, "loss": 0.5765, "step": 3119 }, { "epoch": 2.9947191550648102, "grad_norm": 1.0623686842066253, "learning_rate": 4.154862042997789e-06, "loss": 0.5858, "step": 3120 }, { "epoch": 2.9956793086893905, "grad_norm": 0.8222127031552767, "learning_rate": 4.151560474261229e-06, "loss": 0.5545, "step": 3121 }, { "epoch": 2.9966394623139703, "grad_norm": 0.7736843651288571, "learning_rate": 4.148259286380105e-06, "loss": 0.5163, "step": 3122 }, { "epoch": 2.99759961593855, "grad_norm": 0.7258048208362412, "learning_rate": 4.144958480836283e-06, "loss": 0.5214, "step": 3123 }, { "epoch": 2.99855976956313, "grad_norm": 0.8274237964104535, "learning_rate": 4.141658059111459e-06, "loss": 0.6528, "step": 3124 }, { "epoch": 2.99951992318771, "grad_norm": 0.8605124424013356, "learning_rate": 4.138358022687156e-06, "loss": 0.5522, "step": 3125 }, { "epoch": 3.0, "grad_norm": 0.8605124424013356, "learning_rate": 4.135058373044728e-06, "loss": 0.5906, "step": 3126 }, { "epoch": 3.00096015362458, "grad_norm": 1.9169795858236396, "learning_rate": 4.131759111665349e-06, "loss": 0.5982, "step": 3127 }, { "epoch": 3.00192030724916, "grad_norm": 0.8234763748982856, "learning_rate": 4.128460240030025e-06, "loss": 0.503, "step": 3128 }, { "epoch": 3.00288046087374, "grad_norm": 0.9132598069146752, "learning_rate": 4.125161759619582e-06, "loss": 0.5142, "step": 3129 }, { "epoch": 3.0038406144983196, "grad_norm": 0.9877313680791293, "learning_rate": 4.121863671914672e-06, "loss": 0.4848, "step": 3130 }, { "epoch": 3.0048007681229, "grad_norm": 0.8834126025921609, "learning_rate": 4.118565978395777e-06, "loss": 0.4769, "step": 3131 }, { "epoch": 3.0057609217474797, "grad_norm": 0.9717396227992421, "learning_rate": 4.115268680543189e-06, "loss": 0.4305, "step": 3132 }, { "epoch": 3.0067210753720595, "grad_norm": 0.983717107354649, "learning_rate": 4.111971779837036e-06, "loss": 0.5054, "step": 3133 }, { "epoch": 3.0076812289966393, "grad_norm": 1.2755306152298072, "learning_rate": 4.108675277757255e-06, "loss": 0.5432, "step": 3134 }, { "epoch": 3.0086413826212195, "grad_norm": 1.0579480393184095, "learning_rate": 4.105379175783616e-06, "loss": 0.4583, "step": 3135 }, { "epoch": 3.0096015362457993, "grad_norm": 0.6774800658105046, "learning_rate": 4.102083475395702e-06, "loss": 0.5749, "step": 3136 }, { "epoch": 3.010561689870379, "grad_norm": 1.0235902648381585, "learning_rate": 4.098788178072917e-06, "loss": 0.4755, "step": 3137 }, { "epoch": 3.0115218434949593, "grad_norm": 0.8488950285864946, "learning_rate": 4.095493285294486e-06, "loss": 0.4971, "step": 3138 }, { "epoch": 3.012481997119539, "grad_norm": 1.0226829465798226, "learning_rate": 4.092198798539452e-06, "loss": 0.4676, "step": 3139 }, { "epoch": 3.013442150744119, "grad_norm": 0.7580439497600977, "learning_rate": 4.088904719286671e-06, "loss": 0.5306, "step": 3140 }, { "epoch": 3.014402304368699, "grad_norm": 0.7325064956791109, "learning_rate": 4.085611049014825e-06, "loss": 0.4614, "step": 3141 }, { "epoch": 3.015362457993279, "grad_norm": 0.7880545414631117, "learning_rate": 4.082317789202403e-06, "loss": 0.4734, "step": 3142 }, { "epoch": 3.0163226116178588, "grad_norm": 1.5198896757301419, "learning_rate": 4.079024941327715e-06, "loss": 0.4449, "step": 3143 }, { "epoch": 3.017282765242439, "grad_norm": 1.3728449585516413, "learning_rate": 4.075732506868889e-06, "loss": 0.5225, "step": 3144 }, { "epoch": 3.018242918867019, "grad_norm": 1.9519198264839264, "learning_rate": 4.072440487303859e-06, "loss": 0.4949, "step": 3145 }, { "epoch": 3.0192030724915986, "grad_norm": 0.9981927328704007, "learning_rate": 4.06914888411038e-06, "loss": 0.5846, "step": 3146 }, { "epoch": 3.0201632261161784, "grad_norm": 1.0315141005252209, "learning_rate": 4.065857698766016e-06, "loss": 0.5178, "step": 3147 }, { "epoch": 3.0211233797407586, "grad_norm": 1.1047696257813364, "learning_rate": 4.062566932748143e-06, "loss": 0.496, "step": 3148 }, { "epoch": 3.0220835333653384, "grad_norm": 1.027697453158103, "learning_rate": 4.059276587533959e-06, "loss": 0.4692, "step": 3149 }, { "epoch": 3.023043686989918, "grad_norm": 0.7810296456544716, "learning_rate": 4.0559866646004546e-06, "loss": 0.4893, "step": 3150 }, { "epoch": 3.0240038406144985, "grad_norm": 1.3911121391551189, "learning_rate": 4.052697165424448e-06, "loss": 0.4213, "step": 3151 }, { "epoch": 3.0249639942390782, "grad_norm": 1.2167521167485869, "learning_rate": 4.049408091482556e-06, "loss": 0.4218, "step": 3152 }, { "epoch": 3.025924147863658, "grad_norm": 1.373671629160218, "learning_rate": 4.046119444251211e-06, "loss": 0.5274, "step": 3153 }, { "epoch": 3.0268843014882383, "grad_norm": 0.5880561546815666, "learning_rate": 4.042831225206651e-06, "loss": 0.5539, "step": 3154 }, { "epoch": 3.027844455112818, "grad_norm": 0.8866816069324285, "learning_rate": 4.039543435824923e-06, "loss": 0.5331, "step": 3155 }, { "epoch": 3.028804608737398, "grad_norm": 0.7644419531278153, "learning_rate": 4.036256077581879e-06, "loss": 0.4831, "step": 3156 }, { "epoch": 3.029764762361978, "grad_norm": 1.304587135898252, "learning_rate": 4.032969151953182e-06, "loss": 0.383, "step": 3157 }, { "epoch": 3.030724915986558, "grad_norm": 0.6874287365598377, "learning_rate": 4.029682660414294e-06, "loss": 0.561, "step": 3158 }, { "epoch": 3.0316850696111377, "grad_norm": 1.0844254587367563, "learning_rate": 4.0263966044404885e-06, "loss": 0.542, "step": 3159 }, { "epoch": 3.0326452232357175, "grad_norm": 0.6440487577894802, "learning_rate": 4.023110985506839e-06, "loss": 0.4558, "step": 3160 }, { "epoch": 3.0336053768602977, "grad_norm": 1.158689820492563, "learning_rate": 4.019825805088225e-06, "loss": 0.5843, "step": 3161 }, { "epoch": 3.0345655304848775, "grad_norm": 1.0567733723118062, "learning_rate": 4.0165410646593324e-06, "loss": 0.4257, "step": 3162 }, { "epoch": 3.0355256841094573, "grad_norm": 1.2754192048093624, "learning_rate": 4.01325676569464e-06, "loss": 0.4989, "step": 3163 }, { "epoch": 3.0364858377340376, "grad_norm": 0.7787148441463547, "learning_rate": 4.009972909668442e-06, "loss": 0.5082, "step": 3164 }, { "epoch": 3.0374459913586174, "grad_norm": 0.9407915392859775, "learning_rate": 4.006689498054817e-06, "loss": 0.5035, "step": 3165 }, { "epoch": 3.038406144983197, "grad_norm": 0.8110441725046216, "learning_rate": 4.00340653232766e-06, "loss": 0.5189, "step": 3166 }, { "epoch": 3.0393662986077774, "grad_norm": 0.7759335555186365, "learning_rate": 4.00012401396066e-06, "loss": 0.5553, "step": 3167 }, { "epoch": 3.040326452232357, "grad_norm": 0.8549094105858437, "learning_rate": 3.9968419444273e-06, "loss": 0.5087, "step": 3168 }, { "epoch": 3.041286605856937, "grad_norm": 0.8356292482552343, "learning_rate": 3.993560325200869e-06, "loss": 0.4866, "step": 3169 }, { "epoch": 3.0422467594815172, "grad_norm": 0.6154807792309783, "learning_rate": 3.99027915775445e-06, "loss": 0.4219, "step": 3170 }, { "epoch": 3.043206913106097, "grad_norm": 0.9120778538705216, "learning_rate": 3.986998443560923e-06, "loss": 0.4943, "step": 3171 }, { "epoch": 3.044167066730677, "grad_norm": 0.8628198508650559, "learning_rate": 3.9837181840929665e-06, "loss": 0.5496, "step": 3172 }, { "epoch": 3.0451272203552566, "grad_norm": 0.7323404134370787, "learning_rate": 3.980438380823053e-06, "loss": 0.5914, "step": 3173 }, { "epoch": 3.046087373979837, "grad_norm": 1.1739213191384075, "learning_rate": 3.977159035223451e-06, "loss": 0.4692, "step": 3174 }, { "epoch": 3.0470475276044167, "grad_norm": 0.7842979416470234, "learning_rate": 3.973880148766228e-06, "loss": 0.4972, "step": 3175 }, { "epoch": 3.0480076812289965, "grad_norm": 0.7107701355264762, "learning_rate": 3.970601722923234e-06, "loss": 0.4535, "step": 3176 }, { "epoch": 3.0489678348535767, "grad_norm": 0.6206520841024241, "learning_rate": 3.9673237591661265e-06, "loss": 0.4706, "step": 3177 }, { "epoch": 3.0499279884781565, "grad_norm": 0.9333058352080239, "learning_rate": 3.964046258966342e-06, "loss": 0.4312, "step": 3178 }, { "epoch": 3.0508881421027363, "grad_norm": 0.720193730965791, "learning_rate": 3.960769223795118e-06, "loss": 0.5135, "step": 3179 }, { "epoch": 3.0518482957273165, "grad_norm": 0.7534034635003475, "learning_rate": 3.957492655123483e-06, "loss": 0.5761, "step": 3180 }, { "epoch": 3.0528084493518963, "grad_norm": 0.9498647214690381, "learning_rate": 3.954216554422249e-06, "loss": 0.5113, "step": 3181 }, { "epoch": 3.053768602976476, "grad_norm": 0.9084098366846697, "learning_rate": 3.9509409231620264e-06, "loss": 0.5974, "step": 3182 }, { "epoch": 3.0547287566010564, "grad_norm": 0.7473948015301468, "learning_rate": 3.947665762813209e-06, "loss": 0.5167, "step": 3183 }, { "epoch": 3.055688910225636, "grad_norm": 0.7932595662934029, "learning_rate": 3.944391074845982e-06, "loss": 0.5045, "step": 3184 }, { "epoch": 3.056649063850216, "grad_norm": 0.8686168813603024, "learning_rate": 3.941116860730319e-06, "loss": 0.5519, "step": 3185 }, { "epoch": 3.0576092174747957, "grad_norm": 0.8760333325049292, "learning_rate": 3.9378431219359755e-06, "loss": 0.4213, "step": 3186 }, { "epoch": 3.058569371099376, "grad_norm": 0.924549084249337, "learning_rate": 3.934569859932502e-06, "loss": 0.5247, "step": 3187 }, { "epoch": 3.059529524723956, "grad_norm": 0.894358066083043, "learning_rate": 3.9312970761892305e-06, "loss": 0.4337, "step": 3188 }, { "epoch": 3.0604896783485356, "grad_norm": 0.8114684771021129, "learning_rate": 3.9280247721752765e-06, "loss": 0.4525, "step": 3189 }, { "epoch": 3.061449831973116, "grad_norm": 0.9767338895308353, "learning_rate": 3.924752949359545e-06, "loss": 0.5507, "step": 3190 }, { "epoch": 3.0624099855976956, "grad_norm": 1.1970325151119439, "learning_rate": 3.921481609210719e-06, "loss": 0.4865, "step": 3191 }, { "epoch": 3.0633701392222754, "grad_norm": 0.8369022969671507, "learning_rate": 3.918210753197269e-06, "loss": 0.5155, "step": 3192 }, { "epoch": 3.0643302928468557, "grad_norm": 1.1375598492560581, "learning_rate": 3.914940382787452e-06, "loss": 0.4623, "step": 3193 }, { "epoch": 3.0652904464714354, "grad_norm": 1.1444291295463653, "learning_rate": 3.911670499449295e-06, "loss": 0.5691, "step": 3194 }, { "epoch": 3.0662506000960152, "grad_norm": 1.1671394579840728, "learning_rate": 3.908401104650621e-06, "loss": 0.4155, "step": 3195 }, { "epoch": 3.0672107537205955, "grad_norm": 0.8707213058668786, "learning_rate": 3.905132199859018e-06, "loss": 0.4047, "step": 3196 }, { "epoch": 3.0681709073451753, "grad_norm": 0.921285033580212, "learning_rate": 3.901863786541869e-06, "loss": 0.4523, "step": 3197 }, { "epoch": 3.069131060969755, "grad_norm": 1.1172103992927527, "learning_rate": 3.898595866166329e-06, "loss": 0.3392, "step": 3198 }, { "epoch": 3.070091214594335, "grad_norm": 0.7800362774630242, "learning_rate": 3.895328440199329e-06, "loss": 0.5156, "step": 3199 }, { "epoch": 3.071051368218915, "grad_norm": 0.8129014830480317, "learning_rate": 3.892061510107586e-06, "loss": 0.421, "step": 3200 }, { "epoch": 3.072011521843495, "grad_norm": 0.7143336550807435, "learning_rate": 3.888795077357587e-06, "loss": 0.5512, "step": 3201 }, { "epoch": 3.0729716754680747, "grad_norm": 1.2973123764383567, "learning_rate": 3.885529143415599e-06, "loss": 0.4261, "step": 3202 }, { "epoch": 3.073931829092655, "grad_norm": 1.019807590630448, "learning_rate": 3.882263709747667e-06, "loss": 0.5001, "step": 3203 }, { "epoch": 3.0748919827172347, "grad_norm": 1.2029068676426171, "learning_rate": 3.878998777819607e-06, "loss": 0.4551, "step": 3204 }, { "epoch": 3.0758521363418145, "grad_norm": 0.8439217160866255, "learning_rate": 3.8757343490970115e-06, "loss": 0.5136, "step": 3205 }, { "epoch": 3.0768122899663948, "grad_norm": 1.0414262651947876, "learning_rate": 3.872470425045251e-06, "loss": 0.5845, "step": 3206 }, { "epoch": 3.0777724435909746, "grad_norm": 0.7762549499956453, "learning_rate": 3.869207007129463e-06, "loss": 0.4965, "step": 3207 }, { "epoch": 3.0787325972155544, "grad_norm": 0.9776114907430524, "learning_rate": 3.865944096814564e-06, "loss": 0.4881, "step": 3208 }, { "epoch": 3.0796927508401346, "grad_norm": 1.003207657833835, "learning_rate": 3.862681695565234e-06, "loss": 0.4733, "step": 3209 }, { "epoch": 3.0806529044647144, "grad_norm": 0.7737706053445809, "learning_rate": 3.859419804845935e-06, "loss": 0.5341, "step": 3210 }, { "epoch": 3.081613058089294, "grad_norm": 1.3916160879493469, "learning_rate": 3.856158426120895e-06, "loss": 0.5028, "step": 3211 }, { "epoch": 3.082573211713874, "grad_norm": 1.1935456532694246, "learning_rate": 3.8528975608541095e-06, "loss": 0.4821, "step": 3212 }, { "epoch": 3.0835333653384542, "grad_norm": 0.9403246819077359, "learning_rate": 3.849637210509346e-06, "loss": 0.4345, "step": 3213 }, { "epoch": 3.084493518963034, "grad_norm": 0.8510355209675033, "learning_rate": 3.846377376550144e-06, "loss": 0.5743, "step": 3214 }, { "epoch": 3.085453672587614, "grad_norm": 0.893607086911257, "learning_rate": 3.843118060439805e-06, "loss": 0.5099, "step": 3215 }, { "epoch": 3.086413826212194, "grad_norm": 0.9110959221717266, "learning_rate": 3.839859263641403e-06, "loss": 0.4129, "step": 3216 }, { "epoch": 3.087373979836774, "grad_norm": 0.7860055761608671, "learning_rate": 3.836600987617776e-06, "loss": 0.4547, "step": 3217 }, { "epoch": 3.0883341334613537, "grad_norm": 0.8172285178572343, "learning_rate": 3.833343233831529e-06, "loss": 0.5131, "step": 3218 }, { "epoch": 3.089294287085934, "grad_norm": 1.0228165540098848, "learning_rate": 3.830086003745034e-06, "loss": 0.5235, "step": 3219 }, { "epoch": 3.0902544407105137, "grad_norm": 1.0398136856374005, "learning_rate": 3.826829298820425e-06, "loss": 0.5435, "step": 3220 }, { "epoch": 3.0912145943350935, "grad_norm": 1.262893131526517, "learning_rate": 3.823573120519606e-06, "loss": 0.5443, "step": 3221 }, { "epoch": 3.0921747479596737, "grad_norm": 0.7012947828589995, "learning_rate": 3.8203174703042345e-06, "loss": 0.478, "step": 3222 }, { "epoch": 3.0931349015842535, "grad_norm": 1.0152559564347587, "learning_rate": 3.817062349635742e-06, "loss": 0.5709, "step": 3223 }, { "epoch": 3.0940950552088333, "grad_norm": 1.3331277997508544, "learning_rate": 3.813807759975317e-06, "loss": 0.5436, "step": 3224 }, { "epoch": 3.095055208833413, "grad_norm": 1.0115570335640371, "learning_rate": 3.8105537027839078e-06, "loss": 0.5566, "step": 3225 }, { "epoch": 3.0960153624579934, "grad_norm": 0.7396714322870829, "learning_rate": 3.8073001795222274e-06, "loss": 0.599, "step": 3226 }, { "epoch": 3.096975516082573, "grad_norm": 0.8784561973482617, "learning_rate": 3.8040471916507467e-06, "loss": 0.5121, "step": 3227 }, { "epoch": 3.097935669707153, "grad_norm": 0.7587833307989981, "learning_rate": 3.8007947406296975e-06, "loss": 0.5119, "step": 3228 }, { "epoch": 3.098895823331733, "grad_norm": 0.9296521097653087, "learning_rate": 3.797542827919071e-06, "loss": 0.4548, "step": 3229 }, { "epoch": 3.099855976956313, "grad_norm": 0.7978340033548006, "learning_rate": 3.7942914549786152e-06, "loss": 0.4836, "step": 3230 }, { "epoch": 3.100816130580893, "grad_norm": 1.2051625309590994, "learning_rate": 3.7910406232678363e-06, "loss": 0.4564, "step": 3231 }, { "epoch": 3.101776284205473, "grad_norm": 1.212650613718318, "learning_rate": 3.7877903342460002e-06, "loss": 0.5143, "step": 3232 }, { "epoch": 3.102736437830053, "grad_norm": 0.8552688753642984, "learning_rate": 3.7845405893721237e-06, "loss": 0.497, "step": 3233 }, { "epoch": 3.1036965914546326, "grad_norm": 0.8436288120340306, "learning_rate": 3.781291390104986e-06, "loss": 0.3858, "step": 3234 }, { "epoch": 3.104656745079213, "grad_norm": 0.8398727038249661, "learning_rate": 3.7780427379031137e-06, "loss": 0.5002, "step": 3235 }, { "epoch": 3.1056168987037926, "grad_norm": 0.8786927268881798, "learning_rate": 3.774794634224794e-06, "loss": 0.507, "step": 3236 }, { "epoch": 3.1065770523283724, "grad_norm": 0.7386001503007736, "learning_rate": 3.7715470805280695e-06, "loss": 0.5421, "step": 3237 }, { "epoch": 3.1075372059529527, "grad_norm": 0.9028304911180118, "learning_rate": 3.7683000782707258e-06, "loss": 0.4574, "step": 3238 }, { "epoch": 3.1084973595775325, "grad_norm": 0.7103108922510698, "learning_rate": 3.765053628910314e-06, "loss": 0.5032, "step": 3239 }, { "epoch": 3.1094575132021123, "grad_norm": 1.1093798014814258, "learning_rate": 3.7618077339041244e-06, "loss": 0.4272, "step": 3240 }, { "epoch": 3.110417666826692, "grad_norm": 0.7519472928015878, "learning_rate": 3.758562394709208e-06, "loss": 0.4457, "step": 3241 }, { "epoch": 3.1113778204512723, "grad_norm": 1.0679610606830545, "learning_rate": 3.7553176127823643e-06, "loss": 0.6241, "step": 3242 }, { "epoch": 3.112337974075852, "grad_norm": 0.7391838227355032, "learning_rate": 3.7520733895801375e-06, "loss": 0.4109, "step": 3243 }, { "epoch": 3.113298127700432, "grad_norm": 0.8296676667294207, "learning_rate": 3.748829726558827e-06, "loss": 0.4529, "step": 3244 }, { "epoch": 3.114258281325012, "grad_norm": 0.9800311840942424, "learning_rate": 3.745586625174479e-06, "loss": 0.4268, "step": 3245 }, { "epoch": 3.115218434949592, "grad_norm": 0.9311991987346983, "learning_rate": 3.7423440868828842e-06, "loss": 0.5198, "step": 3246 }, { "epoch": 3.1161785885741717, "grad_norm": 0.924655578978222, "learning_rate": 3.7391021131395866e-06, "loss": 0.5267, "step": 3247 }, { "epoch": 3.117138742198752, "grad_norm": 0.9019512970756918, "learning_rate": 3.7358607053998717e-06, "loss": 0.481, "step": 3248 }, { "epoch": 3.1180988958233318, "grad_norm": 0.9999319721657661, "learning_rate": 3.732619865118772e-06, "loss": 0.51, "step": 3249 }, { "epoch": 3.1190590494479116, "grad_norm": 1.137462961949247, "learning_rate": 3.7293795937510703e-06, "loss": 0.4488, "step": 3250 }, { "epoch": 3.120019203072492, "grad_norm": 1.0220194752632692, "learning_rate": 3.7261398927512836e-06, "loss": 0.5151, "step": 3251 }, { "epoch": 3.1209793566970716, "grad_norm": 1.1631003288676722, "learning_rate": 3.722900763573685e-06, "loss": 0.5833, "step": 3252 }, { "epoch": 3.1219395103216514, "grad_norm": 0.8565451299837139, "learning_rate": 3.71966220767228e-06, "loss": 0.4947, "step": 3253 }, { "epoch": 3.122899663946231, "grad_norm": 0.8780569564641136, "learning_rate": 3.7164242265008244e-06, "loss": 0.4069, "step": 3254 }, { "epoch": 3.1238598175708114, "grad_norm": 0.9686828727057526, "learning_rate": 3.7131868215128135e-06, "loss": 0.4878, "step": 3255 }, { "epoch": 3.1248199711953912, "grad_norm": 0.9739954157141907, "learning_rate": 3.7099499941614823e-06, "loss": 0.4369, "step": 3256 }, { "epoch": 3.125780124819971, "grad_norm": 0.8316563435124084, "learning_rate": 3.7067137458998092e-06, "loss": 0.4599, "step": 3257 }, { "epoch": 3.1267402784445513, "grad_norm": 1.0293812572988235, "learning_rate": 3.7034780781805126e-06, "loss": 0.4595, "step": 3258 }, { "epoch": 3.127700432069131, "grad_norm": 0.9186771517587491, "learning_rate": 3.7002429924560458e-06, "loss": 0.4007, "step": 3259 }, { "epoch": 3.128660585693711, "grad_norm": 1.0046240639485262, "learning_rate": 3.6970084901786075e-06, "loss": 0.4918, "step": 3260 }, { "epoch": 3.129620739318291, "grad_norm": 1.095007659821479, "learning_rate": 3.6937745728001283e-06, "loss": 0.4264, "step": 3261 }, { "epoch": 3.130580892942871, "grad_norm": 1.072363922624475, "learning_rate": 3.6905412417722807e-06, "loss": 0.405, "step": 3262 }, { "epoch": 3.1315410465674507, "grad_norm": 0.9014566470219302, "learning_rate": 3.6873084985464734e-06, "loss": 0.4148, "step": 3263 }, { "epoch": 3.132501200192031, "grad_norm": 0.8836091683107561, "learning_rate": 3.6840763445738485e-06, "loss": 0.5976, "step": 3264 }, { "epoch": 3.1334613538166107, "grad_norm": 1.1041263183842713, "learning_rate": 3.6808447813052873e-06, "loss": 0.471, "step": 3265 }, { "epoch": 3.1344215074411905, "grad_norm": 0.8761195246480008, "learning_rate": 3.677613810191401e-06, "loss": 0.4891, "step": 3266 }, { "epoch": 3.1353816610657703, "grad_norm": 0.8099433053080253, "learning_rate": 3.674383432682539e-06, "loss": 0.4912, "step": 3267 }, { "epoch": 3.1363418146903506, "grad_norm": 0.73717949113711, "learning_rate": 3.671153650228787e-06, "loss": 0.4937, "step": 3268 }, { "epoch": 3.1373019683149304, "grad_norm": 1.1705142760693534, "learning_rate": 3.6679244642799544e-06, "loss": 0.4492, "step": 3269 }, { "epoch": 3.13826212193951, "grad_norm": 1.3508039507421803, "learning_rate": 3.664695876285592e-06, "loss": 0.581, "step": 3270 }, { "epoch": 3.1392222755640904, "grad_norm": 1.3286975008093105, "learning_rate": 3.6614678876949767e-06, "loss": 0.5674, "step": 3271 }, { "epoch": 3.14018242918867, "grad_norm": 0.9086451924583134, "learning_rate": 3.658240499957118e-06, "loss": 0.4973, "step": 3272 }, { "epoch": 3.14114258281325, "grad_norm": 0.831767261159991, "learning_rate": 3.6550137145207577e-06, "loss": 0.5199, "step": 3273 }, { "epoch": 3.1421027364378302, "grad_norm": 1.064417244412824, "learning_rate": 3.651787532834362e-06, "loss": 0.4589, "step": 3274 }, { "epoch": 3.14306289006241, "grad_norm": 1.0647796780630696, "learning_rate": 3.648561956346132e-06, "loss": 0.3878, "step": 3275 }, { "epoch": 3.14402304368699, "grad_norm": 1.057143740696986, "learning_rate": 3.645336986503995e-06, "loss": 0.5194, "step": 3276 }, { "epoch": 3.14498319731157, "grad_norm": 0.9168739034126954, "learning_rate": 3.642112624755603e-06, "loss": 0.5405, "step": 3277 }, { "epoch": 3.14594335093615, "grad_norm": 0.9667152587190232, "learning_rate": 3.6388888725483394e-06, "loss": 0.5107, "step": 3278 }, { "epoch": 3.1469035045607296, "grad_norm": 1.139684799952788, "learning_rate": 3.6356657313293104e-06, "loss": 0.4347, "step": 3279 }, { "epoch": 3.1478636581853094, "grad_norm": 1.4124094758587822, "learning_rate": 3.6324432025453494e-06, "loss": 0.5424, "step": 3280 }, { "epoch": 3.1488238118098897, "grad_norm": 1.3079094954870663, "learning_rate": 3.62922128764302e-06, "loss": 0.4964, "step": 3281 }, { "epoch": 3.1497839654344695, "grad_norm": 0.775140049698065, "learning_rate": 3.6259999880685987e-06, "loss": 0.5378, "step": 3282 }, { "epoch": 3.1507441190590493, "grad_norm": 1.1203295868000605, "learning_rate": 3.6227793052680983e-06, "loss": 0.4099, "step": 3283 }, { "epoch": 3.1517042726836295, "grad_norm": 0.6701852100289418, "learning_rate": 3.6195592406872436e-06, "loss": 0.5025, "step": 3284 }, { "epoch": 3.1526644263082093, "grad_norm": 0.7624256543533684, "learning_rate": 3.6163397957714895e-06, "loss": 0.4835, "step": 3285 }, { "epoch": 3.153624579932789, "grad_norm": 1.2803997701296086, "learning_rate": 3.6131209719660133e-06, "loss": 0.4304, "step": 3286 }, { "epoch": 3.1545847335573693, "grad_norm": 0.7726330223083834, "learning_rate": 3.609902770715705e-06, "loss": 0.5127, "step": 3287 }, { "epoch": 3.155544887181949, "grad_norm": 0.7436701699060877, "learning_rate": 3.6066851934651847e-06, "loss": 0.4761, "step": 3288 }, { "epoch": 3.156505040806529, "grad_norm": 0.8170685102326626, "learning_rate": 3.6034682416587885e-06, "loss": 0.4768, "step": 3289 }, { "epoch": 3.157465194431109, "grad_norm": 0.9983083806987929, "learning_rate": 3.600251916740569e-06, "loss": 0.4407, "step": 3290 }, { "epoch": 3.158425348055689, "grad_norm": 0.7060103037218862, "learning_rate": 3.597036220154303e-06, "loss": 0.4855, "step": 3291 }, { "epoch": 3.1593855016802688, "grad_norm": 0.7716448877721321, "learning_rate": 3.5938211533434797e-06, "loss": 0.5451, "step": 3292 }, { "epoch": 3.160345655304849, "grad_norm": 0.857246156354271, "learning_rate": 3.590606717751308e-06, "loss": 0.4432, "step": 3293 }, { "epoch": 3.161305808929429, "grad_norm": 0.7524779206901041, "learning_rate": 3.5873929148207187e-06, "loss": 0.5412, "step": 3294 }, { "epoch": 3.1622659625540086, "grad_norm": 0.7970495278284874, "learning_rate": 3.5841797459943473e-06, "loss": 0.4551, "step": 3295 }, { "epoch": 3.1632261161785884, "grad_norm": 0.6330084430633233, "learning_rate": 3.5809672127145563e-06, "loss": 0.486, "step": 3296 }, { "epoch": 3.1641862698031686, "grad_norm": 0.7981104053247745, "learning_rate": 3.5777553164234114e-06, "loss": 0.5152, "step": 3297 }, { "epoch": 3.1651464234277484, "grad_norm": 0.6291210694694186, "learning_rate": 3.5745440585627032e-06, "loss": 0.5006, "step": 3298 }, { "epoch": 3.1661065770523282, "grad_norm": 0.7616876378193421, "learning_rate": 3.571333440573932e-06, "loss": 0.4251, "step": 3299 }, { "epoch": 3.1670667306769085, "grad_norm": 0.6590344736870869, "learning_rate": 3.5681234638983054e-06, "loss": 0.4607, "step": 3300 }, { "epoch": 3.1680268843014883, "grad_norm": 0.7857745639458096, "learning_rate": 3.5649141299767504e-06, "loss": 0.4997, "step": 3301 }, { "epoch": 3.168987037926068, "grad_norm": 0.9787306422262078, "learning_rate": 3.561705440249903e-06, "loss": 0.5089, "step": 3302 }, { "epoch": 3.1699471915506483, "grad_norm": 0.7945684956896528, "learning_rate": 3.558497396158108e-06, "loss": 0.4383, "step": 3303 }, { "epoch": 3.170907345175228, "grad_norm": 0.8856586630188269, "learning_rate": 3.5552899991414248e-06, "loss": 0.4996, "step": 3304 }, { "epoch": 3.171867498799808, "grad_norm": 0.9920654608136007, "learning_rate": 3.5520832506396156e-06, "loss": 0.5142, "step": 3305 }, { "epoch": 3.172827652424388, "grad_norm": 1.036672917321548, "learning_rate": 3.548877152092158e-06, "loss": 0.4778, "step": 3306 }, { "epoch": 3.173787806048968, "grad_norm": 0.7704823325682532, "learning_rate": 3.5456717049382373e-06, "loss": 0.5466, "step": 3307 }, { "epoch": 3.1747479596735477, "grad_norm": 0.7674978193072904, "learning_rate": 3.5424669106167408e-06, "loss": 0.5893, "step": 3308 }, { "epoch": 3.1757081132981275, "grad_norm": 0.9147989251426705, "learning_rate": 3.5392627705662686e-06, "loss": 0.4799, "step": 3309 }, { "epoch": 3.1766682669227078, "grad_norm": 0.829745422340152, "learning_rate": 3.5360592862251242e-06, "loss": 0.5024, "step": 3310 }, { "epoch": 3.1776284205472876, "grad_norm": 1.3932038384284977, "learning_rate": 3.5328564590313155e-06, "loss": 0.51, "step": 3311 }, { "epoch": 3.1785885741718674, "grad_norm": 1.1944416034613807, "learning_rate": 3.529654290422563e-06, "loss": 0.4583, "step": 3312 }, { "epoch": 3.1795487277964476, "grad_norm": 0.9311160616111546, "learning_rate": 3.526452781836279e-06, "loss": 0.5257, "step": 3313 }, { "epoch": 3.1805088814210274, "grad_norm": 1.1335819437696018, "learning_rate": 3.5232519347095935e-06, "loss": 0.5324, "step": 3314 }, { "epoch": 3.181469035045607, "grad_norm": 0.8227539798327813, "learning_rate": 3.5200517504793254e-06, "loss": 0.5493, "step": 3315 }, { "epoch": 3.1824291886701874, "grad_norm": 1.0182117107492998, "learning_rate": 3.516852230582008e-06, "loss": 0.4762, "step": 3316 }, { "epoch": 3.183389342294767, "grad_norm": 0.8065517323000255, "learning_rate": 3.513653376453872e-06, "loss": 0.4254, "step": 3317 }, { "epoch": 3.184349495919347, "grad_norm": 1.0060038285957233, "learning_rate": 3.510455189530845e-06, "loss": 0.4818, "step": 3318 }, { "epoch": 3.1853096495439273, "grad_norm": 1.2435234283220338, "learning_rate": 3.507257671248561e-06, "loss": 0.4558, "step": 3319 }, { "epoch": 3.186269803168507, "grad_norm": 1.2422603046857028, "learning_rate": 3.5040608230423533e-06, "loss": 0.4589, "step": 3320 }, { "epoch": 3.187229956793087, "grad_norm": 1.1702480360808647, "learning_rate": 3.5008646463472506e-06, "loss": 0.5521, "step": 3321 }, { "epoch": 3.1881901104176666, "grad_norm": 0.7593572796185951, "learning_rate": 3.497669142597984e-06, "loss": 0.4148, "step": 3322 }, { "epoch": 3.189150264042247, "grad_norm": 0.7436389844107684, "learning_rate": 3.4944743132289793e-06, "loss": 0.5932, "step": 3323 }, { "epoch": 3.1901104176668267, "grad_norm": 0.806498410846488, "learning_rate": 3.4912801596743615e-06, "loss": 0.5399, "step": 3324 }, { "epoch": 3.1910705712914065, "grad_norm": 0.8988648259369305, "learning_rate": 3.4880866833679555e-06, "loss": 0.4467, "step": 3325 }, { "epoch": 3.1920307249159867, "grad_norm": 0.7432484106001649, "learning_rate": 3.4848938857432727e-06, "loss": 0.4714, "step": 3326 }, { "epoch": 3.1929908785405665, "grad_norm": 0.7790570056056669, "learning_rate": 3.4817017682335326e-06, "loss": 0.4808, "step": 3327 }, { "epoch": 3.1939510321651463, "grad_norm": 0.8203655244777148, "learning_rate": 3.478510332271635e-06, "loss": 0.5071, "step": 3328 }, { "epoch": 3.1949111857897265, "grad_norm": 1.0429001947133285, "learning_rate": 3.475319579290187e-06, "loss": 0.5393, "step": 3329 }, { "epoch": 3.1958713394143063, "grad_norm": 0.8923611974291888, "learning_rate": 3.4721295107214835e-06, "loss": 0.4582, "step": 3330 }, { "epoch": 3.196831493038886, "grad_norm": 0.9227351456099845, "learning_rate": 3.4689401279975087e-06, "loss": 0.4635, "step": 3331 }, { "epoch": 3.1977916466634664, "grad_norm": 1.052040291631147, "learning_rate": 3.4657514325499454e-06, "loss": 0.5015, "step": 3332 }, { "epoch": 3.198751800288046, "grad_norm": 0.8260269474623133, "learning_rate": 3.4625634258101636e-06, "loss": 0.4896, "step": 3333 }, { "epoch": 3.199711953912626, "grad_norm": 0.9440547162238836, "learning_rate": 3.459376109209226e-06, "loss": 0.475, "step": 3334 }, { "epoch": 3.2006721075372058, "grad_norm": 1.256427799879319, "learning_rate": 3.456189484177884e-06, "loss": 0.4399, "step": 3335 }, { "epoch": 3.201632261161786, "grad_norm": 0.6804558716406403, "learning_rate": 3.4530035521465803e-06, "loss": 0.4398, "step": 3336 }, { "epoch": 3.202592414786366, "grad_norm": 0.9038048126946161, "learning_rate": 3.4498183145454445e-06, "loss": 0.5258, "step": 3337 }, { "epoch": 3.2035525684109456, "grad_norm": 0.8839135647847125, "learning_rate": 3.446633772804299e-06, "loss": 0.4472, "step": 3338 }, { "epoch": 3.204512722035526, "grad_norm": 0.724847699263036, "learning_rate": 3.443449928352646e-06, "loss": 0.5149, "step": 3339 }, { "epoch": 3.2054728756601056, "grad_norm": 0.6856464428379478, "learning_rate": 3.4402667826196845e-06, "loss": 0.4873, "step": 3340 }, { "epoch": 3.2064330292846854, "grad_norm": 0.9728977896963593, "learning_rate": 3.4370843370342887e-06, "loss": 0.54, "step": 3341 }, { "epoch": 3.2073931829092657, "grad_norm": 0.6004953746931939, "learning_rate": 3.433902593025028e-06, "loss": 0.4435, "step": 3342 }, { "epoch": 3.2083533365338455, "grad_norm": 0.7561939003984248, "learning_rate": 3.430721552020154e-06, "loss": 0.591, "step": 3343 }, { "epoch": 3.2093134901584253, "grad_norm": 0.9633766696887327, "learning_rate": 3.4275412154475996e-06, "loss": 0.5252, "step": 3344 }, { "epoch": 3.2102736437830055, "grad_norm": 0.7525555590853745, "learning_rate": 3.4243615847349855e-06, "loss": 0.4594, "step": 3345 }, { "epoch": 3.2112337974075853, "grad_norm": 0.8941137810124361, "learning_rate": 3.4211826613096135e-06, "loss": 0.5436, "step": 3346 }, { "epoch": 3.212193951032165, "grad_norm": 1.346696062878902, "learning_rate": 3.418004446598467e-06, "loss": 0.5131, "step": 3347 }, { "epoch": 3.213154104656745, "grad_norm": 1.1283066238824602, "learning_rate": 3.4148269420282155e-06, "loss": 0.617, "step": 3348 }, { "epoch": 3.214114258281325, "grad_norm": 0.8088687382246346, "learning_rate": 3.4116501490252037e-06, "loss": 0.4534, "step": 3349 }, { "epoch": 3.215074411905905, "grad_norm": 0.7861195313254172, "learning_rate": 3.4084740690154617e-06, "loss": 0.4904, "step": 3350 }, { "epoch": 3.2160345655304847, "grad_norm": 0.8421193498521694, "learning_rate": 3.405298703424699e-06, "loss": 0.6605, "step": 3351 }, { "epoch": 3.216994719155065, "grad_norm": 0.8936995912989472, "learning_rate": 3.4021240536783e-06, "loss": 0.4803, "step": 3352 }, { "epoch": 3.2179548727796448, "grad_norm": 0.7834657070207709, "learning_rate": 3.3989501212013344e-06, "loss": 0.4308, "step": 3353 }, { "epoch": 3.2189150264042246, "grad_norm": 1.3892325718713698, "learning_rate": 3.3957769074185443e-06, "loss": 0.5451, "step": 3354 }, { "epoch": 3.219875180028805, "grad_norm": 1.0307430231022843, "learning_rate": 3.392604413754351e-06, "loss": 0.4858, "step": 3355 }, { "epoch": 3.2208353336533846, "grad_norm": 0.7629141708916664, "learning_rate": 3.3894326416328575e-06, "loss": 0.4494, "step": 3356 }, { "epoch": 3.2217954872779644, "grad_norm": 0.7993507535236091, "learning_rate": 3.3862615924778326e-06, "loss": 0.5496, "step": 3357 }, { "epoch": 3.2227556409025446, "grad_norm": 0.8765811111849761, "learning_rate": 3.3830912677127313e-06, "loss": 0.4966, "step": 3358 }, { "epoch": 3.2237157945271244, "grad_norm": 0.9638208494812566, "learning_rate": 3.3799216687606727e-06, "loss": 0.4255, "step": 3359 }, { "epoch": 3.224675948151704, "grad_norm": 0.9326855937975593, "learning_rate": 3.3767527970444614e-06, "loss": 0.5102, "step": 3360 }, { "epoch": 3.225636101776284, "grad_norm": 0.9260654682029714, "learning_rate": 3.3735846539865684e-06, "loss": 0.4293, "step": 3361 }, { "epoch": 3.2265962554008643, "grad_norm": 0.8435601770039308, "learning_rate": 3.370417241009138e-06, "loss": 0.5438, "step": 3362 }, { "epoch": 3.227556409025444, "grad_norm": 0.8390337220757541, "learning_rate": 3.367250559533989e-06, "loss": 0.4868, "step": 3363 }, { "epoch": 3.228516562650024, "grad_norm": 0.9848248488842075, "learning_rate": 3.3640846109826107e-06, "loss": 0.4834, "step": 3364 }, { "epoch": 3.229476716274604, "grad_norm": 1.106566723900973, "learning_rate": 3.3609193967761637e-06, "loss": 0.4086, "step": 3365 }, { "epoch": 3.230436869899184, "grad_norm": 1.031338516976276, "learning_rate": 3.3577549183354797e-06, "loss": 0.507, "step": 3366 }, { "epoch": 3.2313970235237637, "grad_norm": 0.8745496349562417, "learning_rate": 3.3545911770810565e-06, "loss": 0.4248, "step": 3367 }, { "epoch": 3.232357177148344, "grad_norm": 0.6670870257615991, "learning_rate": 3.3514281744330644e-06, "loss": 0.4378, "step": 3368 }, { "epoch": 3.2333173307729237, "grad_norm": 0.7175258442033999, "learning_rate": 3.348265911811346e-06, "loss": 0.5637, "step": 3369 }, { "epoch": 3.2342774843975035, "grad_norm": 0.8711641918108166, "learning_rate": 3.345104390635401e-06, "loss": 0.4653, "step": 3370 }, { "epoch": 3.2352376380220838, "grad_norm": 0.8241559772447349, "learning_rate": 3.3419436123244076e-06, "loss": 0.609, "step": 3371 }, { "epoch": 3.2361977916466635, "grad_norm": 1.0170265247902244, "learning_rate": 3.3387835782972e-06, "loss": 0.6001, "step": 3372 }, { "epoch": 3.2371579452712433, "grad_norm": 0.9951208819612899, "learning_rate": 3.3356242899722877e-06, "loss": 0.5067, "step": 3373 }, { "epoch": 3.238118098895823, "grad_norm": 0.9677527849392464, "learning_rate": 3.3324657487678423e-06, "loss": 0.5155, "step": 3374 }, { "epoch": 3.2390782525204034, "grad_norm": 0.9256983921424601, "learning_rate": 3.3293079561016957e-06, "loss": 0.4847, "step": 3375 }, { "epoch": 3.240038406144983, "grad_norm": 0.7421974124483168, "learning_rate": 3.3261509133913484e-06, "loss": 0.4159, "step": 3376 }, { "epoch": 3.240998559769563, "grad_norm": 1.0041269289615014, "learning_rate": 3.3229946220539657e-06, "loss": 0.515, "step": 3377 }, { "epoch": 3.241958713394143, "grad_norm": 1.2869229996929354, "learning_rate": 3.3198390835063685e-06, "loss": 0.4579, "step": 3378 }, { "epoch": 3.242918867018723, "grad_norm": 1.1029276639265235, "learning_rate": 3.3166842991650493e-06, "loss": 0.4553, "step": 3379 }, { "epoch": 3.243879020643303, "grad_norm": 1.00299934384802, "learning_rate": 3.3135302704461526e-06, "loss": 0.4962, "step": 3380 }, { "epoch": 3.244839174267883, "grad_norm": 0.8453627793234577, "learning_rate": 3.31037699876549e-06, "loss": 0.4933, "step": 3381 }, { "epoch": 3.245799327892463, "grad_norm": 1.135419385821641, "learning_rate": 3.307224485538534e-06, "loss": 0.4321, "step": 3382 }, { "epoch": 3.2467594815170426, "grad_norm": 0.8799649361672868, "learning_rate": 3.3040727321804083e-06, "loss": 0.4896, "step": 3383 }, { "epoch": 3.247719635141623, "grad_norm": 0.9054679680619622, "learning_rate": 3.300921740105908e-06, "loss": 0.5068, "step": 3384 }, { "epoch": 3.2486797887662027, "grad_norm": 0.9942474169925896, "learning_rate": 3.297771510729474e-06, "loss": 0.4736, "step": 3385 }, { "epoch": 3.2496399423907825, "grad_norm": 0.7572038775458307, "learning_rate": 3.2946220454652134e-06, "loss": 0.5707, "step": 3386 }, { "epoch": 3.2506000960153623, "grad_norm": 1.0486089912718168, "learning_rate": 3.2914733457268876e-06, "loss": 0.5364, "step": 3387 }, { "epoch": 3.2515602496399425, "grad_norm": 1.0162630937238362, "learning_rate": 3.288325412927914e-06, "loss": 0.4293, "step": 3388 }, { "epoch": 3.2525204032645223, "grad_norm": 0.8106734525240374, "learning_rate": 3.285178248481366e-06, "loss": 0.4897, "step": 3389 }, { "epoch": 3.253480556889102, "grad_norm": 0.8062464149909391, "learning_rate": 3.282031853799971e-06, "loss": 0.4738, "step": 3390 }, { "epoch": 3.2544407105136823, "grad_norm": 0.7668558313645275, "learning_rate": 3.278886230296113e-06, "loss": 0.4773, "step": 3391 }, { "epoch": 3.255400864138262, "grad_norm": 0.7674767051127245, "learning_rate": 3.2757413793818294e-06, "loss": 0.5147, "step": 3392 }, { "epoch": 3.256361017762842, "grad_norm": 0.7817380459592904, "learning_rate": 3.2725973024688073e-06, "loss": 0.5806, "step": 3393 }, { "epoch": 3.257321171387422, "grad_norm": 0.9615406322808412, "learning_rate": 3.2694540009683926e-06, "loss": 0.4621, "step": 3394 }, { "epoch": 3.258281325012002, "grad_norm": 1.0397307900735513, "learning_rate": 3.266311476291579e-06, "loss": 0.4739, "step": 3395 }, { "epoch": 3.2592414786365818, "grad_norm": 1.3089384314599368, "learning_rate": 3.2631697298490105e-06, "loss": 0.3833, "step": 3396 }, { "epoch": 3.260201632261162, "grad_norm": 1.1003631091197446, "learning_rate": 3.2600287630509854e-06, "loss": 0.4339, "step": 3397 }, { "epoch": 3.261161785885742, "grad_norm": 1.0278904467127565, "learning_rate": 3.256888577307448e-06, "loss": 0.495, "step": 3398 }, { "epoch": 3.2621219395103216, "grad_norm": 0.861877654907048, "learning_rate": 3.2537491740279946e-06, "loss": 0.5031, "step": 3399 }, { "epoch": 3.2630820931349014, "grad_norm": 1.0462265286464387, "learning_rate": 3.250610554621875e-06, "loss": 0.5278, "step": 3400 }, { "epoch": 3.2640422467594816, "grad_norm": 0.9017216398786566, "learning_rate": 3.2474727204979728e-06, "loss": 0.4671, "step": 3401 }, { "epoch": 3.2650024003840614, "grad_norm": 0.9150836037586261, "learning_rate": 3.2443356730648363e-06, "loss": 0.476, "step": 3402 }, { "epoch": 3.265962554008641, "grad_norm": 0.8048075163096247, "learning_rate": 3.2411994137306463e-06, "loss": 0.5498, "step": 3403 }, { "epoch": 3.2669227076332215, "grad_norm": 0.7883219432764031, "learning_rate": 3.23806394390324e-06, "loss": 0.5768, "step": 3404 }, { "epoch": 3.2678828612578013, "grad_norm": 1.1095353748985768, "learning_rate": 3.234929264990096e-06, "loss": 0.4862, "step": 3405 }, { "epoch": 3.268843014882381, "grad_norm": 0.8090144660006823, "learning_rate": 3.231795378398337e-06, "loss": 0.4224, "step": 3406 }, { "epoch": 3.2698031685069613, "grad_norm": 1.1088040381122468, "learning_rate": 3.2286622855347306e-06, "loss": 0.4324, "step": 3407 }, { "epoch": 3.270763322131541, "grad_norm": 1.0882096359441187, "learning_rate": 3.2255299878056913e-06, "loss": 0.6203, "step": 3408 }, { "epoch": 3.271723475756121, "grad_norm": 0.8470443377324077, "learning_rate": 3.2223984866172715e-06, "loss": 0.5098, "step": 3409 }, { "epoch": 3.272683629380701, "grad_norm": 0.8839347730105269, "learning_rate": 3.21926778337517e-06, "loss": 0.4445, "step": 3410 }, { "epoch": 3.273643783005281, "grad_norm": 0.9292901575964017, "learning_rate": 3.2161378794847255e-06, "loss": 0.5701, "step": 3411 }, { "epoch": 3.2746039366298607, "grad_norm": 1.0280423965840295, "learning_rate": 3.2130087763509165e-06, "loss": 0.4546, "step": 3412 }, { "epoch": 3.2755640902544405, "grad_norm": 1.0449348322428489, "learning_rate": 3.2098804753783696e-06, "loss": 0.4439, "step": 3413 }, { "epoch": 3.2765242438790207, "grad_norm": 1.4117582468518854, "learning_rate": 3.2067529779713384e-06, "loss": 0.5417, "step": 3414 }, { "epoch": 3.2774843975036005, "grad_norm": 0.834545741563065, "learning_rate": 3.2036262855337304e-06, "loss": 0.5023, "step": 3415 }, { "epoch": 3.2784445511281803, "grad_norm": 0.8775806214271853, "learning_rate": 3.200500399469077e-06, "loss": 0.4738, "step": 3416 }, { "epoch": 3.2794047047527606, "grad_norm": 0.6488458311892228, "learning_rate": 3.1973753211805597e-06, "loss": 0.4826, "step": 3417 }, { "epoch": 3.2803648583773404, "grad_norm": 1.3920183244966282, "learning_rate": 3.1942510520709928e-06, "loss": 0.4847, "step": 3418 }, { "epoch": 3.28132501200192, "grad_norm": 1.10889977673127, "learning_rate": 3.1911275935428256e-06, "loss": 0.4589, "step": 3419 }, { "epoch": 3.2822851656265004, "grad_norm": 0.8811237101294713, "learning_rate": 3.1880049469981468e-06, "loss": 0.54, "step": 3420 }, { "epoch": 3.28324531925108, "grad_norm": 0.9711858804561512, "learning_rate": 3.1848831138386772e-06, "loss": 0.5092, "step": 3421 }, { "epoch": 3.28420547287566, "grad_norm": 0.8794407625185534, "learning_rate": 3.1817620954657745e-06, "loss": 0.4877, "step": 3422 }, { "epoch": 3.2851656265002402, "grad_norm": 1.1402744200822665, "learning_rate": 3.178641893280433e-06, "loss": 0.5026, "step": 3423 }, { "epoch": 3.28612578012482, "grad_norm": 0.9062220933384991, "learning_rate": 3.1755225086832755e-06, "loss": 0.5382, "step": 3424 }, { "epoch": 3.2870859337494, "grad_norm": 0.9348221746091709, "learning_rate": 3.17240394307456e-06, "loss": 0.5365, "step": 3425 }, { "epoch": 3.2880460873739796, "grad_norm": 0.7974155831285559, "learning_rate": 3.169286197854181e-06, "loss": 0.5257, "step": 3426 }, { "epoch": 3.28900624099856, "grad_norm": 0.8549797936911163, "learning_rate": 3.1661692744216564e-06, "loss": 0.5336, "step": 3427 }, { "epoch": 3.2899663946231397, "grad_norm": 1.2075076328010979, "learning_rate": 3.163053174176145e-06, "loss": 0.5039, "step": 3428 }, { "epoch": 3.2909265482477195, "grad_norm": 1.0100587689743639, "learning_rate": 3.159937898516424e-06, "loss": 0.4953, "step": 3429 }, { "epoch": 3.2918867018722997, "grad_norm": 0.9848564484585245, "learning_rate": 3.1568234488409126e-06, "loss": 0.5622, "step": 3430 }, { "epoch": 3.2928468554968795, "grad_norm": 0.708640345897404, "learning_rate": 3.153709826547654e-06, "loss": 0.5303, "step": 3431 }, { "epoch": 3.2938070091214593, "grad_norm": 0.8810797727518063, "learning_rate": 3.1505970330343172e-06, "loss": 0.4382, "step": 3432 }, { "epoch": 3.2947671627460395, "grad_norm": 1.010395167841745, "learning_rate": 3.147485069698204e-06, "loss": 0.5386, "step": 3433 }, { "epoch": 3.2957273163706193, "grad_norm": 0.787228438476513, "learning_rate": 3.1443739379362396e-06, "loss": 0.5732, "step": 3434 }, { "epoch": 3.296687469995199, "grad_norm": 0.9897838093482716, "learning_rate": 3.1412636391449792e-06, "loss": 0.5347, "step": 3435 }, { "epoch": 3.2976476236197794, "grad_norm": 0.824423878330104, "learning_rate": 3.1381541747206034e-06, "loss": 0.478, "step": 3436 }, { "epoch": 3.298607777244359, "grad_norm": 0.7972459360842241, "learning_rate": 3.1350455460589157e-06, "loss": 0.457, "step": 3437 }, { "epoch": 3.299567930868939, "grad_norm": 0.9033481436205597, "learning_rate": 3.131937754555347e-06, "loss": 0.5695, "step": 3438 }, { "epoch": 3.3005280844935188, "grad_norm": 0.8463989704009459, "learning_rate": 3.128830801604953e-06, "loss": 0.5112, "step": 3439 }, { "epoch": 3.301488238118099, "grad_norm": 1.2160165041937496, "learning_rate": 3.125724688602409e-06, "loss": 0.5973, "step": 3440 }, { "epoch": 3.302448391742679, "grad_norm": 1.0603869994535533, "learning_rate": 3.122619416942019e-06, "loss": 0.5171, "step": 3441 }, { "epoch": 3.3034085453672586, "grad_norm": 0.9508428446663035, "learning_rate": 3.1195149880177027e-06, "loss": 0.5339, "step": 3442 }, { "epoch": 3.304368698991839, "grad_norm": 0.7105059621216405, "learning_rate": 3.1164114032230063e-06, "loss": 0.4813, "step": 3443 }, { "epoch": 3.3053288526164186, "grad_norm": 0.7368926800515069, "learning_rate": 3.1133086639511e-06, "loss": 0.4374, "step": 3444 }, { "epoch": 3.3062890062409984, "grad_norm": 0.7335110505804923, "learning_rate": 3.1102067715947637e-06, "loss": 0.564, "step": 3445 }, { "epoch": 3.3072491598655787, "grad_norm": 0.8872100917309773, "learning_rate": 3.10710572754641e-06, "loss": 0.5604, "step": 3446 }, { "epoch": 3.3082093134901585, "grad_norm": 0.9010731216986351, "learning_rate": 3.104005533198058e-06, "loss": 0.5782, "step": 3447 }, { "epoch": 3.3091694671147383, "grad_norm": 0.8802407506035985, "learning_rate": 3.100906189941356e-06, "loss": 0.651, "step": 3448 }, { "epoch": 3.3101296207393185, "grad_norm": 0.7781000936659725, "learning_rate": 3.0978076991675667e-06, "loss": 0.4995, "step": 3449 }, { "epoch": 3.3110897743638983, "grad_norm": 1.1236271415267984, "learning_rate": 3.0947100622675663e-06, "loss": 0.584, "step": 3450 }, { "epoch": 3.312049927988478, "grad_norm": 1.4324256623316132, "learning_rate": 3.091613280631853e-06, "loss": 0.5404, "step": 3451 }, { "epoch": 3.313010081613058, "grad_norm": 0.7869612731341731, "learning_rate": 3.08851735565054e-06, "loss": 0.5279, "step": 3452 }, { "epoch": 3.313970235237638, "grad_norm": 1.1346494459186027, "learning_rate": 3.0854222887133513e-06, "loss": 0.4303, "step": 3453 }, { "epoch": 3.314930388862218, "grad_norm": 0.866000571213409, "learning_rate": 3.082328081209633e-06, "loss": 0.5014, "step": 3454 }, { "epoch": 3.3158905424867977, "grad_norm": 0.7739511946641434, "learning_rate": 3.079234734528339e-06, "loss": 0.4463, "step": 3455 }, { "epoch": 3.316850696111378, "grad_norm": 1.1891617422451994, "learning_rate": 3.07614225005804e-06, "loss": 0.4767, "step": 3456 }, { "epoch": 3.3178108497359577, "grad_norm": 1.2357238497602845, "learning_rate": 3.0730506291869235e-06, "loss": 0.5276, "step": 3457 }, { "epoch": 3.3187710033605375, "grad_norm": 1.0884670257917874, "learning_rate": 3.069959873302778e-06, "loss": 0.4931, "step": 3458 }, { "epoch": 3.319731156985118, "grad_norm": 1.1290403640710958, "learning_rate": 3.0668699837930175e-06, "loss": 0.5119, "step": 3459 }, { "epoch": 3.3206913106096976, "grad_norm": 0.7494316857954071, "learning_rate": 3.0637809620446547e-06, "loss": 0.4734, "step": 3460 }, { "epoch": 3.3216514642342774, "grad_norm": 0.8324038477963301, "learning_rate": 3.0606928094443216e-06, "loss": 0.4288, "step": 3461 }, { "epoch": 3.3226116178588576, "grad_norm": 0.803008454071545, "learning_rate": 3.0576055273782574e-06, "loss": 0.4915, "step": 3462 }, { "epoch": 3.3235717714834374, "grad_norm": 0.9791300369344389, "learning_rate": 3.0545191172323076e-06, "loss": 0.5258, "step": 3463 }, { "epoch": 3.324531925108017, "grad_norm": 1.3169749963934823, "learning_rate": 3.0514335803919314e-06, "loss": 0.5656, "step": 3464 }, { "epoch": 3.325492078732597, "grad_norm": 0.9263434360773872, "learning_rate": 3.048348918242191e-06, "loss": 0.508, "step": 3465 }, { "epoch": 3.3264522323571772, "grad_norm": 0.8099158165746592, "learning_rate": 3.045265132167759e-06, "loss": 0.4764, "step": 3466 }, { "epoch": 3.327412385981757, "grad_norm": 0.9694820107541914, "learning_rate": 3.0421822235529154e-06, "loss": 0.5803, "step": 3467 }, { "epoch": 3.328372539606337, "grad_norm": 1.0244558086346123, "learning_rate": 3.0391001937815425e-06, "loss": 0.4776, "step": 3468 }, { "epoch": 3.329332693230917, "grad_norm": 0.9743805959643078, "learning_rate": 3.036019044237132e-06, "loss": 0.511, "step": 3469 }, { "epoch": 3.330292846855497, "grad_norm": 0.9588948858909689, "learning_rate": 3.0329387763027794e-06, "loss": 0.5312, "step": 3470 }, { "epoch": 3.3312530004800767, "grad_norm": 1.04468763108387, "learning_rate": 3.029859391361183e-06, "loss": 0.3894, "step": 3471 }, { "epoch": 3.332213154104657, "grad_norm": 0.8344698714899678, "learning_rate": 3.026780890794647e-06, "loss": 0.4876, "step": 3472 }, { "epoch": 3.3331733077292367, "grad_norm": 0.8060462744129175, "learning_rate": 3.0237032759850736e-06, "loss": 0.5248, "step": 3473 }, { "epoch": 3.3341334613538165, "grad_norm": 1.1630613506963554, "learning_rate": 3.020626548313976e-06, "loss": 0.467, "step": 3474 }, { "epoch": 3.3350936149783967, "grad_norm": 1.1468264053151076, "learning_rate": 3.017550709162464e-06, "loss": 0.5106, "step": 3475 }, { "epoch": 3.3360537686029765, "grad_norm": 1.0996049440267166, "learning_rate": 3.014475759911246e-06, "loss": 0.4738, "step": 3476 }, { "epoch": 3.3370139222275563, "grad_norm": 0.8339825682542231, "learning_rate": 3.0114017019406362e-06, "loss": 0.4712, "step": 3477 }, { "epoch": 3.337974075852136, "grad_norm": 1.2428681006821267, "learning_rate": 3.0083285366305447e-06, "loss": 0.3957, "step": 3478 }, { "epoch": 3.3389342294767164, "grad_norm": 0.8407989625072088, "learning_rate": 3.0052562653604827e-06, "loss": 0.5162, "step": 3479 }, { "epoch": 3.339894383101296, "grad_norm": 0.7798190179378429, "learning_rate": 3.002184889509562e-06, "loss": 0.383, "step": 3480 }, { "epoch": 3.340854536725876, "grad_norm": 0.9966949025318544, "learning_rate": 2.9991144104564883e-06, "loss": 0.5816, "step": 3481 }, { "epoch": 3.341814690350456, "grad_norm": 1.372045678978103, "learning_rate": 2.9960448295795667e-06, "loss": 0.4558, "step": 3482 }, { "epoch": 3.342774843975036, "grad_norm": 0.7843669148907011, "learning_rate": 2.9929761482567017e-06, "loss": 0.4639, "step": 3483 }, { "epoch": 3.343734997599616, "grad_norm": 0.8906113118464071, "learning_rate": 2.9899083678653885e-06, "loss": 0.5738, "step": 3484 }, { "epoch": 3.344695151224196, "grad_norm": 1.0823862372577724, "learning_rate": 2.9868414897827236e-06, "loss": 0.4484, "step": 3485 }, { "epoch": 3.345655304848776, "grad_norm": 0.9938859427650673, "learning_rate": 2.983775515385393e-06, "loss": 0.4138, "step": 3486 }, { "epoch": 3.3466154584733556, "grad_norm": 0.6917132827808095, "learning_rate": 2.9807104460496807e-06, "loss": 0.4833, "step": 3487 }, { "epoch": 3.347575612097936, "grad_norm": 0.7507381454570533, "learning_rate": 2.9776462831514664e-06, "loss": 0.5167, "step": 3488 }, { "epoch": 3.3485357657225157, "grad_norm": 0.8774975982861176, "learning_rate": 2.9745830280662157e-06, "loss": 0.557, "step": 3489 }, { "epoch": 3.3494959193470955, "grad_norm": 0.801406883183827, "learning_rate": 2.971520682168996e-06, "loss": 0.4493, "step": 3490 }, { "epoch": 3.3504560729716752, "grad_norm": 0.9055969978986355, "learning_rate": 2.968459246834455e-06, "loss": 0.5028, "step": 3491 }, { "epoch": 3.3514162265962555, "grad_norm": 0.9619519347504092, "learning_rate": 2.9653987234368443e-06, "loss": 0.605, "step": 3492 }, { "epoch": 3.3523763802208353, "grad_norm": 1.1390227692959662, "learning_rate": 2.9623391133499986e-06, "loss": 0.4684, "step": 3493 }, { "epoch": 3.353336533845415, "grad_norm": 0.8954179409138489, "learning_rate": 2.959280417947342e-06, "loss": 0.4406, "step": 3494 }, { "epoch": 3.3542966874699953, "grad_norm": 1.0618138639370571, "learning_rate": 2.9562226386018935e-06, "loss": 0.5656, "step": 3495 }, { "epoch": 3.355256841094575, "grad_norm": 0.8162687750926433, "learning_rate": 2.953165776686254e-06, "loss": 0.4683, "step": 3496 }, { "epoch": 3.356216994719155, "grad_norm": 0.991219242964635, "learning_rate": 2.9501098335726188e-06, "loss": 0.5581, "step": 3497 }, { "epoch": 3.357177148343735, "grad_norm": 1.5102520699850404, "learning_rate": 2.947054810632768e-06, "loss": 0.6087, "step": 3498 }, { "epoch": 3.358137301968315, "grad_norm": 1.0329819613889368, "learning_rate": 2.9440007092380663e-06, "loss": 0.4926, "step": 3499 }, { "epoch": 3.3590974555928947, "grad_norm": 1.1704983502592026, "learning_rate": 2.9409475307594683e-06, "loss": 0.3774, "step": 3500 }, { "epoch": 3.360057609217475, "grad_norm": 0.7612149703473756, "learning_rate": 2.937895276567516e-06, "loss": 0.5432, "step": 3501 }, { "epoch": 3.361017762842055, "grad_norm": 0.9354501084293928, "learning_rate": 2.934843948032329e-06, "loss": 0.5464, "step": 3502 }, { "epoch": 3.3619779164666346, "grad_norm": 0.8553016927067424, "learning_rate": 2.931793546523622e-06, "loss": 0.4754, "step": 3503 }, { "epoch": 3.3629380700912144, "grad_norm": 1.003999011888384, "learning_rate": 2.9287440734106802e-06, "loss": 0.513, "step": 3504 }, { "epoch": 3.3638982237157946, "grad_norm": 1.0013968479821462, "learning_rate": 2.9256955300623847e-06, "loss": 0.4875, "step": 3505 }, { "epoch": 3.3648583773403744, "grad_norm": 1.1732074138946595, "learning_rate": 2.9226479178471933e-06, "loss": 0.485, "step": 3506 }, { "epoch": 3.365818530964954, "grad_norm": 1.121287845214766, "learning_rate": 2.919601238133145e-06, "loss": 0.5386, "step": 3507 }, { "epoch": 3.3667786845895344, "grad_norm": 1.0628503583607112, "learning_rate": 2.916555492287865e-06, "loss": 0.3922, "step": 3508 }, { "epoch": 3.3677388382141142, "grad_norm": 1.0906115949500657, "learning_rate": 2.913510681678552e-06, "loss": 0.4885, "step": 3509 }, { "epoch": 3.368698991838694, "grad_norm": 1.0379872532763625, "learning_rate": 2.9104668076719876e-06, "loss": 0.6301, "step": 3510 }, { "epoch": 3.3696591454632743, "grad_norm": 0.7025046702797064, "learning_rate": 2.9074238716345426e-06, "loss": 0.515, "step": 3511 }, { "epoch": 3.370619299087854, "grad_norm": 0.6747414006265559, "learning_rate": 2.9043818749321506e-06, "loss": 0.4436, "step": 3512 }, { "epoch": 3.371579452712434, "grad_norm": 0.7597401851239275, "learning_rate": 2.9013408189303337e-06, "loss": 0.5209, "step": 3513 }, { "epoch": 3.372539606337014, "grad_norm": 0.8328299902396756, "learning_rate": 2.8983007049941904e-06, "loss": 0.5243, "step": 3514 }, { "epoch": 3.373499759961594, "grad_norm": 0.9898176968989575, "learning_rate": 2.8952615344883938e-06, "loss": 0.5185, "step": 3515 }, { "epoch": 3.3744599135861737, "grad_norm": 1.2995668154478905, "learning_rate": 2.8922233087771988e-06, "loss": 0.4762, "step": 3516 }, { "epoch": 3.3754200672107535, "grad_norm": 1.051770788359885, "learning_rate": 2.889186029224428e-06, "loss": 0.5861, "step": 3517 }, { "epoch": 3.3763802208353337, "grad_norm": 1.0683045130477125, "learning_rate": 2.8861496971934843e-06, "loss": 0.5549, "step": 3518 }, { "epoch": 3.3773403744599135, "grad_norm": 1.03705708764033, "learning_rate": 2.883114314047346e-06, "loss": 0.4918, "step": 3519 }, { "epoch": 3.3783005280844933, "grad_norm": 0.9055031211017993, "learning_rate": 2.880079881148564e-06, "loss": 0.4463, "step": 3520 }, { "epoch": 3.3792606817090736, "grad_norm": 0.9628099206557446, "learning_rate": 2.8770463998592645e-06, "loss": 0.6289, "step": 3521 }, { "epoch": 3.3802208353336534, "grad_norm": 1.1778203454236105, "learning_rate": 2.8740138715411413e-06, "loss": 0.5063, "step": 3522 }, { "epoch": 3.381180988958233, "grad_norm": 0.7951462641477786, "learning_rate": 2.8709822975554636e-06, "loss": 0.5078, "step": 3523 }, { "epoch": 3.3821411425828134, "grad_norm": 0.9501310886361978, "learning_rate": 2.8679516792630783e-06, "loss": 0.5397, "step": 3524 }, { "epoch": 3.383101296207393, "grad_norm": 0.7893678110012344, "learning_rate": 2.864922018024392e-06, "loss": 0.5037, "step": 3525 }, { "epoch": 3.384061449831973, "grad_norm": 0.9687577274749454, "learning_rate": 2.8618933151993895e-06, "loss": 0.5258, "step": 3526 }, { "epoch": 3.3850216034565532, "grad_norm": 1.2967041835797555, "learning_rate": 2.858865572147622e-06, "loss": 0.4543, "step": 3527 }, { "epoch": 3.385981757081133, "grad_norm": 0.7123119487250497, "learning_rate": 2.8558387902282124e-06, "loss": 0.4826, "step": 3528 }, { "epoch": 3.386941910705713, "grad_norm": 0.8582607367754344, "learning_rate": 2.852812970799852e-06, "loss": 0.5114, "step": 3529 }, { "epoch": 3.3879020643302926, "grad_norm": 0.9078713179237395, "learning_rate": 2.849788115220794e-06, "loss": 0.5536, "step": 3530 }, { "epoch": 3.388862217954873, "grad_norm": 0.9964774488844711, "learning_rate": 2.846764224848867e-06, "loss": 0.4196, "step": 3531 }, { "epoch": 3.3898223715794527, "grad_norm": 0.972734226798294, "learning_rate": 2.843741301041463e-06, "loss": 0.535, "step": 3532 }, { "epoch": 3.390782525204033, "grad_norm": 1.1082068948267962, "learning_rate": 2.8407193451555398e-06, "loss": 0.5253, "step": 3533 }, { "epoch": 3.3917426788286127, "grad_norm": 0.8041409202881534, "learning_rate": 2.8376983585476226e-06, "loss": 0.4578, "step": 3534 }, { "epoch": 3.3927028324531925, "grad_norm": 1.0924559266471119, "learning_rate": 2.834678342573797e-06, "loss": 0.4879, "step": 3535 }, { "epoch": 3.3936629860777723, "grad_norm": 1.3189711531627193, "learning_rate": 2.8316592985897173e-06, "loss": 0.4968, "step": 3536 }, { "epoch": 3.3946231397023525, "grad_norm": 0.9183817317696228, "learning_rate": 2.8286412279506006e-06, "loss": 0.5231, "step": 3537 }, { "epoch": 3.3955832933269323, "grad_norm": 0.9250457828363284, "learning_rate": 2.8256241320112253e-06, "loss": 0.4564, "step": 3538 }, { "epoch": 3.396543446951512, "grad_norm": 0.8946937208703398, "learning_rate": 2.822608012125937e-06, "loss": 0.491, "step": 3539 }, { "epoch": 3.3975036005760924, "grad_norm": 0.928900284540597, "learning_rate": 2.8195928696486345e-06, "loss": 0.4693, "step": 3540 }, { "epoch": 3.398463754200672, "grad_norm": 0.6434655953839171, "learning_rate": 2.8165787059327844e-06, "loss": 0.476, "step": 3541 }, { "epoch": 3.399423907825252, "grad_norm": 0.9090254151391515, "learning_rate": 2.8135655223314166e-06, "loss": 0.4866, "step": 3542 }, { "epoch": 3.4003840614498317, "grad_norm": 0.8675341376350035, "learning_rate": 2.810553320197113e-06, "loss": 0.5299, "step": 3543 }, { "epoch": 3.401344215074412, "grad_norm": 1.0919457783938986, "learning_rate": 2.8075421008820197e-06, "loss": 0.4356, "step": 3544 }, { "epoch": 3.4023043686989918, "grad_norm": 1.4686255143991713, "learning_rate": 2.804531865737842e-06, "loss": 0.539, "step": 3545 }, { "epoch": 3.403264522323572, "grad_norm": 0.9642791154972505, "learning_rate": 2.8015226161158403e-06, "loss": 0.568, "step": 3546 }, { "epoch": 3.404224675948152, "grad_norm": 0.774950373577753, "learning_rate": 2.798514353366838e-06, "loss": 0.5452, "step": 3547 }, { "epoch": 3.4051848295727316, "grad_norm": 0.7507442713077807, "learning_rate": 2.7955070788412085e-06, "loss": 0.4824, "step": 3548 }, { "epoch": 3.4061449831973114, "grad_norm": 0.9693902732008557, "learning_rate": 2.7925007938888867e-06, "loss": 0.3972, "step": 3549 }, { "epoch": 3.4071051368218916, "grad_norm": 1.087767790796374, "learning_rate": 2.789495499859361e-06, "loss": 0.5441, "step": 3550 }, { "epoch": 3.4080652904464714, "grad_norm": 1.226269885639581, "learning_rate": 2.786491198101677e-06, "loss": 0.362, "step": 3551 }, { "epoch": 3.4090254440710512, "grad_norm": 1.1426292963184659, "learning_rate": 2.783487889964435e-06, "loss": 0.5135, "step": 3552 }, { "epoch": 3.4099855976956315, "grad_norm": 0.9254092168452664, "learning_rate": 2.7804855767957833e-06, "loss": 0.4793, "step": 3553 }, { "epoch": 3.4109457513202113, "grad_norm": 1.2000833118531833, "learning_rate": 2.7774842599434294e-06, "loss": 0.4707, "step": 3554 }, { "epoch": 3.411905904944791, "grad_norm": 1.8360732010017804, "learning_rate": 2.7744839407546374e-06, "loss": 0.4787, "step": 3555 }, { "epoch": 3.412866058569371, "grad_norm": 1.1781130920829623, "learning_rate": 2.7714846205762134e-06, "loss": 0.5167, "step": 3556 }, { "epoch": 3.413826212193951, "grad_norm": 1.0943789209021804, "learning_rate": 2.7684863007545227e-06, "loss": 0.5715, "step": 3557 }, { "epoch": 3.414786365818531, "grad_norm": 1.1769085438831943, "learning_rate": 2.765488982635477e-06, "loss": 0.4655, "step": 3558 }, { "epoch": 3.415746519443111, "grad_norm": 0.7708889152563162, "learning_rate": 2.7624926675645415e-06, "loss": 0.4906, "step": 3559 }, { "epoch": 3.416706673067691, "grad_norm": 0.8794457592122289, "learning_rate": 2.759497356886731e-06, "loss": 0.4412, "step": 3560 }, { "epoch": 3.4176668266922707, "grad_norm": 1.08393941123188, "learning_rate": 2.7565030519466063e-06, "loss": 0.555, "step": 3561 }, { "epoch": 3.4186269803168505, "grad_norm": 0.9557100910995658, "learning_rate": 2.75350975408828e-06, "loss": 0.4719, "step": 3562 }, { "epoch": 3.4195871339414308, "grad_norm": 0.858680048828106, "learning_rate": 2.7505174646554112e-06, "loss": 0.5145, "step": 3563 }, { "epoch": 3.4205472875660106, "grad_norm": 0.8267738833381453, "learning_rate": 2.7475261849912073e-06, "loss": 0.5228, "step": 3564 }, { "epoch": 3.4215074411905904, "grad_norm": 0.9868786943640668, "learning_rate": 2.744535916438423e-06, "loss": 0.568, "step": 3565 }, { "epoch": 3.4224675948151706, "grad_norm": 0.6523838230635919, "learning_rate": 2.7415466603393538e-06, "loss": 0.4789, "step": 3566 }, { "epoch": 3.4234277484397504, "grad_norm": 1.141538687612932, "learning_rate": 2.738558418035846e-06, "loss": 0.4101, "step": 3567 }, { "epoch": 3.42438790206433, "grad_norm": 0.9141652354921685, "learning_rate": 2.7355711908692942e-06, "loss": 0.4743, "step": 3568 }, { "epoch": 3.42534805568891, "grad_norm": 0.9206877882281599, "learning_rate": 2.732584980180628e-06, "loss": 0.3911, "step": 3569 }, { "epoch": 3.4263082093134902, "grad_norm": 0.949703531055531, "learning_rate": 2.7295997873103286e-06, "loss": 0.4514, "step": 3570 }, { "epoch": 3.42726836293807, "grad_norm": 1.1261801738681143, "learning_rate": 2.726615613598411e-06, "loss": 0.5567, "step": 3571 }, { "epoch": 3.4282285165626503, "grad_norm": 0.9348033207183948, "learning_rate": 2.7236324603844467e-06, "loss": 0.4248, "step": 3572 }, { "epoch": 3.42918867018723, "grad_norm": 1.0304768823374868, "learning_rate": 2.7206503290075388e-06, "loss": 0.5099, "step": 3573 }, { "epoch": 3.43014882381181, "grad_norm": 0.8919440176091973, "learning_rate": 2.7176692208063317e-06, "loss": 0.4612, "step": 3574 }, { "epoch": 3.4311089774363897, "grad_norm": 0.8473778141273708, "learning_rate": 2.7146891371190144e-06, "loss": 0.5476, "step": 3575 }, { "epoch": 3.43206913106097, "grad_norm": 0.8433465704959358, "learning_rate": 2.7117100792833144e-06, "loss": 0.4358, "step": 3576 }, { "epoch": 3.4330292846855497, "grad_norm": 0.8592006703929673, "learning_rate": 2.7087320486364997e-06, "loss": 0.4982, "step": 3577 }, { "epoch": 3.4339894383101295, "grad_norm": 1.046126872885925, "learning_rate": 2.7057550465153782e-06, "loss": 0.5312, "step": 3578 }, { "epoch": 3.4349495919347097, "grad_norm": 0.8571370192963214, "learning_rate": 2.7027790742562905e-06, "loss": 0.4488, "step": 3579 }, { "epoch": 3.4359097455592895, "grad_norm": 0.9115964560149349, "learning_rate": 2.6998041331951196e-06, "loss": 0.4789, "step": 3580 }, { "epoch": 3.4368698991838693, "grad_norm": 0.7330117158315955, "learning_rate": 2.6968302246672863e-06, "loss": 0.5012, "step": 3581 }, { "epoch": 3.437830052808449, "grad_norm": 0.9744826082073551, "learning_rate": 2.693857350007746e-06, "loss": 0.4933, "step": 3582 }, { "epoch": 3.4387902064330294, "grad_norm": 0.9864728731226643, "learning_rate": 2.6908855105509912e-06, "loss": 0.5557, "step": 3583 }, { "epoch": 3.439750360057609, "grad_norm": 0.9165032936806508, "learning_rate": 2.6879147076310457e-06, "loss": 0.5159, "step": 3584 }, { "epoch": 3.4407105136821894, "grad_norm": 1.089019087899688, "learning_rate": 2.6849449425814713e-06, "loss": 0.5153, "step": 3585 }, { "epoch": 3.441670667306769, "grad_norm": 1.0383891080198955, "learning_rate": 2.6819762167353695e-06, "loss": 0.4362, "step": 3586 }, { "epoch": 3.442630820931349, "grad_norm": 0.862639009771432, "learning_rate": 2.6790085314253645e-06, "loss": 0.4164, "step": 3587 }, { "epoch": 3.4435909745559288, "grad_norm": 1.1704511470211356, "learning_rate": 2.6760418879836186e-06, "loss": 0.4829, "step": 3588 }, { "epoch": 3.444551128180509, "grad_norm": 0.7396280120905687, "learning_rate": 2.673076287741828e-06, "loss": 0.4597, "step": 3589 }, { "epoch": 3.445511281805089, "grad_norm": 0.9843797963340886, "learning_rate": 2.670111732031218e-06, "loss": 0.4401, "step": 3590 }, { "epoch": 3.4464714354296686, "grad_norm": 1.496496668474666, "learning_rate": 2.6671482221825484e-06, "loss": 0.4537, "step": 3591 }, { "epoch": 3.447431589054249, "grad_norm": 1.1074123179278998, "learning_rate": 2.6641857595261035e-06, "loss": 0.5136, "step": 3592 }, { "epoch": 3.4483917426788286, "grad_norm": 1.2173440440399963, "learning_rate": 2.661224345391703e-06, "loss": 0.5903, "step": 3593 }, { "epoch": 3.4493518963034084, "grad_norm": 0.9114873874260104, "learning_rate": 2.658263981108693e-06, "loss": 0.4453, "step": 3594 }, { "epoch": 3.4503120499279882, "grad_norm": 1.0901662633089784, "learning_rate": 2.65530466800595e-06, "loss": 0.5469, "step": 3595 }, { "epoch": 3.4512722035525685, "grad_norm": 0.7460966495868957, "learning_rate": 2.652346407411881e-06, "loss": 0.4857, "step": 3596 }, { "epoch": 3.4522323571771483, "grad_norm": 0.9067843612208649, "learning_rate": 2.649389200654412e-06, "loss": 0.5002, "step": 3597 }, { "epoch": 3.4531925108017285, "grad_norm": 0.7988963519236255, "learning_rate": 2.6464330490610026e-06, "loss": 0.5202, "step": 3598 }, { "epoch": 3.4541526644263083, "grad_norm": 0.990773342686563, "learning_rate": 2.6434779539586432e-06, "loss": 0.5925, "step": 3599 }, { "epoch": 3.455112818050888, "grad_norm": 0.8271593226319985, "learning_rate": 2.640523916673838e-06, "loss": 0.4771, "step": 3600 }, { "epoch": 3.456072971675468, "grad_norm": 0.6726025249255629, "learning_rate": 2.6375709385326255e-06, "loss": 0.5148, "step": 3601 }, { "epoch": 3.457033125300048, "grad_norm": 0.6929105689731146, "learning_rate": 2.6346190208605653e-06, "loss": 0.511, "step": 3602 }, { "epoch": 3.457993278924628, "grad_norm": 1.0799907133220887, "learning_rate": 2.6316681649827423e-06, "loss": 0.5159, "step": 3603 }, { "epoch": 3.4589534325492077, "grad_norm": 1.297470680682533, "learning_rate": 2.6287183722237663e-06, "loss": 0.379, "step": 3604 }, { "epoch": 3.459913586173788, "grad_norm": 0.8392311442726766, "learning_rate": 2.625769643907763e-06, "loss": 0.5158, "step": 3605 }, { "epoch": 3.4608737397983678, "grad_norm": 0.8618813840745123, "learning_rate": 2.622821981358388e-06, "loss": 0.5042, "step": 3606 }, { "epoch": 3.4618338934229476, "grad_norm": 0.904060056344071, "learning_rate": 2.6198753858988147e-06, "loss": 0.5106, "step": 3607 }, { "epoch": 3.4627940470475274, "grad_norm": 0.7870882079117728, "learning_rate": 2.616929858851739e-06, "loss": 0.5015, "step": 3608 }, { "epoch": 3.4637542006721076, "grad_norm": 0.9083527183750432, "learning_rate": 2.613985401539377e-06, "loss": 0.4907, "step": 3609 }, { "epoch": 3.4647143542966874, "grad_norm": 0.826722773095111, "learning_rate": 2.6110420152834614e-06, "loss": 0.5438, "step": 3610 }, { "epoch": 3.4656745079212676, "grad_norm": 0.9313867515638103, "learning_rate": 2.6080997014052474e-06, "loss": 0.5021, "step": 3611 }, { "epoch": 3.4666346615458474, "grad_norm": 1.099547362538891, "learning_rate": 2.6051584612255132e-06, "loss": 0.5175, "step": 3612 }, { "epoch": 3.4675948151704272, "grad_norm": 0.9473636164914084, "learning_rate": 2.6022182960645436e-06, "loss": 0.5067, "step": 3613 }, { "epoch": 3.468554968795007, "grad_norm": 0.9795547787007426, "learning_rate": 2.5992792072421526e-06, "loss": 0.5005, "step": 3614 }, { "epoch": 3.4695151224195873, "grad_norm": 0.7467632091162553, "learning_rate": 2.596341196077659e-06, "loss": 0.4356, "step": 3615 }, { "epoch": 3.470475276044167, "grad_norm": 0.8069536366760434, "learning_rate": 2.593404263889911e-06, "loss": 0.546, "step": 3616 }, { "epoch": 3.471435429668747, "grad_norm": 1.3129090475304142, "learning_rate": 2.5904684119972656e-06, "loss": 0.4679, "step": 3617 }, { "epoch": 3.472395583293327, "grad_norm": 1.561317962287694, "learning_rate": 2.587533641717592e-06, "loss": 0.5035, "step": 3618 }, { "epoch": 3.473355736917907, "grad_norm": 1.0076157696794383, "learning_rate": 2.5845999543682787e-06, "loss": 0.5174, "step": 3619 }, { "epoch": 3.4743158905424867, "grad_norm": 0.7993871771763448, "learning_rate": 2.581667351266226e-06, "loss": 0.4556, "step": 3620 }, { "epoch": 3.475276044167067, "grad_norm": 0.7815899459142782, "learning_rate": 2.5787358337278493e-06, "loss": 0.4875, "step": 3621 }, { "epoch": 3.4762361977916467, "grad_norm": 0.9735299744085145, "learning_rate": 2.575805403069077e-06, "loss": 0.4797, "step": 3622 }, { "epoch": 3.4771963514162265, "grad_norm": 0.9531645006274613, "learning_rate": 2.5728760606053443e-06, "loss": 0.61, "step": 3623 }, { "epoch": 3.4781565050408068, "grad_norm": 1.0465496207142517, "learning_rate": 2.5699478076516034e-06, "loss": 0.5143, "step": 3624 }, { "epoch": 3.4791166586653866, "grad_norm": 0.9092585450951955, "learning_rate": 2.5670206455223156e-06, "loss": 0.5546, "step": 3625 }, { "epoch": 3.4800768122899663, "grad_norm": 1.1024426759677541, "learning_rate": 2.5640945755314527e-06, "loss": 0.4566, "step": 3626 }, { "epoch": 3.481036965914546, "grad_norm": 1.028538450766313, "learning_rate": 2.5611695989924988e-06, "loss": 0.4934, "step": 3627 }, { "epoch": 3.4819971195391264, "grad_norm": 0.9615083667716328, "learning_rate": 2.5582457172184405e-06, "loss": 0.49, "step": 3628 }, { "epoch": 3.482957273163706, "grad_norm": 0.8787041980163668, "learning_rate": 2.5553229315217763e-06, "loss": 0.5455, "step": 3629 }, { "epoch": 3.483917426788286, "grad_norm": 0.8768296735086528, "learning_rate": 2.5524012432145203e-06, "loss": 0.4884, "step": 3630 }, { "epoch": 3.484877580412866, "grad_norm": 1.144190129916985, "learning_rate": 2.5494806536081806e-06, "loss": 0.4878, "step": 3631 }, { "epoch": 3.485837734037446, "grad_norm": 1.0996324560300688, "learning_rate": 2.5465611640137813e-06, "loss": 0.5037, "step": 3632 }, { "epoch": 3.486797887662026, "grad_norm": 1.0487896080181291, "learning_rate": 2.543642775741849e-06, "loss": 0.5009, "step": 3633 }, { "epoch": 3.487758041286606, "grad_norm": 1.0532871371100234, "learning_rate": 2.540725490102418e-06, "loss": 0.5542, "step": 3634 }, { "epoch": 3.488718194911186, "grad_norm": 0.9076711651948381, "learning_rate": 2.5378093084050287e-06, "loss": 0.4673, "step": 3635 }, { "epoch": 3.4896783485357656, "grad_norm": 1.0993184513791185, "learning_rate": 2.534894231958719e-06, "loss": 0.3361, "step": 3636 }, { "epoch": 3.490638502160346, "grad_norm": 1.1405723224696063, "learning_rate": 2.531980262072038e-06, "loss": 0.4883, "step": 3637 }, { "epoch": 3.4915986557849257, "grad_norm": 0.8156937274472742, "learning_rate": 2.5290674000530365e-06, "loss": 0.4071, "step": 3638 }, { "epoch": 3.4925588094095055, "grad_norm": 1.0011636587227253, "learning_rate": 2.526155647209265e-06, "loss": 0.6067, "step": 3639 }, { "epoch": 3.4935189630340853, "grad_norm": 1.0362949696936357, "learning_rate": 2.523245004847783e-06, "loss": 0.397, "step": 3640 }, { "epoch": 3.4944791166586655, "grad_norm": 0.646231391378274, "learning_rate": 2.520335474275141e-06, "loss": 0.4854, "step": 3641 }, { "epoch": 3.4954392702832453, "grad_norm": 0.8590995840255768, "learning_rate": 2.5174270567973957e-06, "loss": 0.4192, "step": 3642 }, { "epoch": 3.496399423907825, "grad_norm": 0.9077502478099381, "learning_rate": 2.5145197537201116e-06, "loss": 0.4973, "step": 3643 }, { "epoch": 3.4973595775324053, "grad_norm": 0.7941655430453408, "learning_rate": 2.51161356634834e-06, "loss": 0.5867, "step": 3644 }, { "epoch": 3.498319731156985, "grad_norm": 0.9906138165029826, "learning_rate": 2.5087084959866403e-06, "loss": 0.5672, "step": 3645 }, { "epoch": 3.499279884781565, "grad_norm": 0.7530891833094846, "learning_rate": 2.5058045439390634e-06, "loss": 0.5292, "step": 3646 }, { "epoch": 3.5002400384061447, "grad_norm": 1.0294900239387923, "learning_rate": 2.502901711509167e-06, "loss": 0.5642, "step": 3647 }, { "epoch": 3.501200192030725, "grad_norm": 0.7428205334341484, "learning_rate": 2.5000000000000015e-06, "loss": 0.4816, "step": 3648 }, { "epoch": 3.5021603456553048, "grad_norm": 0.7846560947005258, "learning_rate": 2.4970994107141106e-06, "loss": 0.5001, "step": 3649 }, { "epoch": 3.503120499279885, "grad_norm": 0.8587619325057322, "learning_rate": 2.4941999449535393e-06, "loss": 0.4565, "step": 3650 }, { "epoch": 3.504080652904465, "grad_norm": 0.7396319731279559, "learning_rate": 2.491301604019828e-06, "loss": 0.5037, "step": 3651 }, { "epoch": 3.5050408065290446, "grad_norm": 0.9757839015697863, "learning_rate": 2.4884043892140096e-06, "loss": 0.5627, "step": 3652 }, { "epoch": 3.5060009601536244, "grad_norm": 0.8060554194187629, "learning_rate": 2.4855083018366154e-06, "loss": 0.5034, "step": 3653 }, { "epoch": 3.5069611137782046, "grad_norm": 0.7671278529191324, "learning_rate": 2.4826133431876644e-06, "loss": 0.4751, "step": 3654 }, { "epoch": 3.5079212674027844, "grad_norm": 0.7342166549929332, "learning_rate": 2.479719514566672e-06, "loss": 0.5671, "step": 3655 }, { "epoch": 3.5088814210273642, "grad_norm": 0.8208041477033728, "learning_rate": 2.4768268172726535e-06, "loss": 0.5184, "step": 3656 }, { "epoch": 3.5098415746519445, "grad_norm": 0.7431816858197855, "learning_rate": 2.473935252604103e-06, "loss": 0.4418, "step": 3657 }, { "epoch": 3.5108017282765243, "grad_norm": 0.670283614869046, "learning_rate": 2.4710448218590176e-06, "loss": 0.5131, "step": 3658 }, { "epoch": 3.511761881901104, "grad_norm": 0.7973939452560215, "learning_rate": 2.468155526334875e-06, "loss": 0.5176, "step": 3659 }, { "epoch": 3.512722035525684, "grad_norm": 1.1305015039514879, "learning_rate": 2.465267367328655e-06, "loss": 0.5518, "step": 3660 }, { "epoch": 3.513682189150264, "grad_norm": 0.9585158983543841, "learning_rate": 2.46238034613682e-06, "loss": 0.4941, "step": 3661 }, { "epoch": 3.514642342774844, "grad_norm": 0.7783983304565283, "learning_rate": 2.4594944640553207e-06, "loss": 0.5236, "step": 3662 }, { "epoch": 3.515602496399424, "grad_norm": 0.8376117894802452, "learning_rate": 2.4566097223795997e-06, "loss": 0.4871, "step": 3663 }, { "epoch": 3.516562650024004, "grad_norm": 0.8248787798812816, "learning_rate": 2.453726122404587e-06, "loss": 0.6036, "step": 3664 }, { "epoch": 3.5175228036485837, "grad_norm": 1.0145337433832753, "learning_rate": 2.4508436654247004e-06, "loss": 0.5682, "step": 3665 }, { "epoch": 3.5184829572731635, "grad_norm": 1.1922058550220567, "learning_rate": 2.4479623527338447e-06, "loss": 0.4382, "step": 3666 }, { "epoch": 3.5194431108977438, "grad_norm": 1.0414414928144626, "learning_rate": 2.4450821856254074e-06, "loss": 0.5855, "step": 3667 }, { "epoch": 3.5204032645223235, "grad_norm": 1.4925971714649038, "learning_rate": 2.442203165392266e-06, "loss": 0.4564, "step": 3668 }, { "epoch": 3.5213634181469033, "grad_norm": 1.0663024554702885, "learning_rate": 2.439325293326783e-06, "loss": 0.5819, "step": 3669 }, { "epoch": 3.5223235717714836, "grad_norm": 1.3201821991744356, "learning_rate": 2.4364485707208037e-06, "loss": 0.3885, "step": 3670 }, { "epoch": 3.5232837253960634, "grad_norm": 1.3750795884191374, "learning_rate": 2.433572998865659e-06, "loss": 0.5331, "step": 3671 }, { "epoch": 3.524243879020643, "grad_norm": 0.8961201123802133, "learning_rate": 2.430698579052161e-06, "loss": 0.5854, "step": 3672 }, { "epoch": 3.525204032645223, "grad_norm": 1.2227624013547649, "learning_rate": 2.4278253125706044e-06, "loss": 0.5469, "step": 3673 }, { "epoch": 3.526164186269803, "grad_norm": 1.13182430292685, "learning_rate": 2.424953200710773e-06, "loss": 0.4942, "step": 3674 }, { "epoch": 3.527124339894383, "grad_norm": 1.1428638346812647, "learning_rate": 2.4220822447619223e-06, "loss": 0.5834, "step": 3675 }, { "epoch": 3.5280844935189632, "grad_norm": 0.9691992713750508, "learning_rate": 2.419212446012796e-06, "loss": 0.4607, "step": 3676 }, { "epoch": 3.529044647143543, "grad_norm": 0.9453446827040339, "learning_rate": 2.416343805751615e-06, "loss": 0.469, "step": 3677 }, { "epoch": 3.530004800768123, "grad_norm": 0.9498049268135695, "learning_rate": 2.4134763252660825e-06, "loss": 0.4793, "step": 3678 }, { "epoch": 3.5309649543927026, "grad_norm": 0.9194704353828093, "learning_rate": 2.4106100058433814e-06, "loss": 0.4383, "step": 3679 }, { "epoch": 3.531925108017283, "grad_norm": 0.6685207373507656, "learning_rate": 2.4077448487701667e-06, "loss": 0.4146, "step": 3680 }, { "epoch": 3.5328852616418627, "grad_norm": 0.8108009910561466, "learning_rate": 2.4048808553325804e-06, "loss": 0.5438, "step": 3681 }, { "epoch": 3.5338454152664425, "grad_norm": 0.8424271386901869, "learning_rate": 2.4020180268162374e-06, "loss": 0.5319, "step": 3682 }, { "epoch": 3.5348055688910227, "grad_norm": 0.9988786130557512, "learning_rate": 2.3991563645062317e-06, "loss": 0.5053, "step": 3683 }, { "epoch": 3.5357657225156025, "grad_norm": 0.613493078732792, "learning_rate": 2.3962958696871343e-06, "loss": 0.4564, "step": 3684 }, { "epoch": 3.5367258761401823, "grad_norm": 0.9696474409944819, "learning_rate": 2.393436543642987e-06, "loss": 0.5142, "step": 3685 }, { "epoch": 3.537686029764762, "grad_norm": 1.0544318715383136, "learning_rate": 2.3905783876573108e-06, "loss": 0.5233, "step": 3686 }, { "epoch": 3.5386461833893423, "grad_norm": 0.804378397042713, "learning_rate": 2.387721403013107e-06, "loss": 0.4796, "step": 3687 }, { "epoch": 3.539606337013922, "grad_norm": 0.913945117465564, "learning_rate": 2.3848655909928392e-06, "loss": 0.5096, "step": 3688 }, { "epoch": 3.5405664906385024, "grad_norm": 0.7994175146067617, "learning_rate": 2.382010952878456e-06, "loss": 0.5124, "step": 3689 }, { "epoch": 3.541526644263082, "grad_norm": 0.8610783544101467, "learning_rate": 2.379157489951367e-06, "loss": 0.5442, "step": 3690 }, { "epoch": 3.542486797887662, "grad_norm": 0.9264745234469997, "learning_rate": 2.376305203492468e-06, "loss": 0.536, "step": 3691 }, { "epoch": 3.5434469515122418, "grad_norm": 0.8529543018879439, "learning_rate": 2.373454094782119e-06, "loss": 0.5518, "step": 3692 }, { "epoch": 3.544407105136822, "grad_norm": 1.4374566016247357, "learning_rate": 2.370604165100148e-06, "loss": 0.5061, "step": 3693 }, { "epoch": 3.545367258761402, "grad_norm": 0.9341954677429397, "learning_rate": 2.367755415725862e-06, "loss": 0.5644, "step": 3694 }, { "epoch": 3.5463274123859816, "grad_norm": 0.5867763348605962, "learning_rate": 2.3649078479380316e-06, "loss": 0.4166, "step": 3695 }, { "epoch": 3.547287566010562, "grad_norm": 0.7843259524016734, "learning_rate": 2.3620614630149013e-06, "loss": 0.4167, "step": 3696 }, { "epoch": 3.5482477196351416, "grad_norm": 1.0225543399068884, "learning_rate": 2.359216262234184e-06, "loss": 0.5077, "step": 3697 }, { "epoch": 3.5492078732597214, "grad_norm": 1.4566743368469166, "learning_rate": 2.3563722468730567e-06, "loss": 0.4542, "step": 3698 }, { "epoch": 3.550168026884301, "grad_norm": 0.7720271583106578, "learning_rate": 2.3535294182081666e-06, "loss": 0.494, "step": 3699 }, { "epoch": 3.5511281805088815, "grad_norm": 0.7944729829187055, "learning_rate": 2.350687777515635e-06, "loss": 0.5025, "step": 3700 }, { "epoch": 3.5520883341334613, "grad_norm": 1.095405784097052, "learning_rate": 2.3478473260710387e-06, "loss": 0.5073, "step": 3701 }, { "epoch": 3.5530484877580415, "grad_norm": 1.0492361474905922, "learning_rate": 2.34500806514943e-06, "loss": 0.6049, "step": 3702 }, { "epoch": 3.5540086413826213, "grad_norm": 1.0510102520089994, "learning_rate": 2.3421699960253168e-06, "loss": 0.4478, "step": 3703 }, { "epoch": 3.554968795007201, "grad_norm": 1.0028775028817514, "learning_rate": 2.3393331199726837e-06, "loss": 0.4982, "step": 3704 }, { "epoch": 3.555928948631781, "grad_norm": 0.7138604788010771, "learning_rate": 2.3364974382649737e-06, "loss": 0.5122, "step": 3705 }, { "epoch": 3.556889102256361, "grad_norm": 1.1564355804555448, "learning_rate": 2.3336629521750902e-06, "loss": 0.4848, "step": 3706 }, { "epoch": 3.557849255880941, "grad_norm": 1.3145785660005114, "learning_rate": 2.3308296629754067e-06, "loss": 0.4891, "step": 3707 }, { "epoch": 3.558809409505521, "grad_norm": 1.2982312235701792, "learning_rate": 2.327997571937755e-06, "loss": 0.4431, "step": 3708 }, { "epoch": 3.559769563130101, "grad_norm": 0.8232093419026618, "learning_rate": 2.3251666803334316e-06, "loss": 0.4931, "step": 3709 }, { "epoch": 3.5607297167546808, "grad_norm": 0.8845397878208552, "learning_rate": 2.3223369894331947e-06, "loss": 0.5214, "step": 3710 }, { "epoch": 3.5616898703792605, "grad_norm": 1.0350311393087623, "learning_rate": 2.3195085005072587e-06, "loss": 0.4235, "step": 3711 }, { "epoch": 3.5626500240038403, "grad_norm": 0.8614605236224139, "learning_rate": 2.3166812148253045e-06, "loss": 0.5085, "step": 3712 }, { "epoch": 3.5636101776284206, "grad_norm": 0.9085009622227533, "learning_rate": 2.313855133656469e-06, "loss": 0.4731, "step": 3713 }, { "epoch": 3.5645703312530004, "grad_norm": 0.8282918213137883, "learning_rate": 2.3110302582693512e-06, "loss": 0.4669, "step": 3714 }, { "epoch": 3.5655304848775806, "grad_norm": 0.8170562075422441, "learning_rate": 2.3082065899320084e-06, "loss": 0.4789, "step": 3715 }, { "epoch": 3.5664906385021604, "grad_norm": 0.8251462613589396, "learning_rate": 2.3053841299119524e-06, "loss": 0.497, "step": 3716 }, { "epoch": 3.56745079212674, "grad_norm": 0.999742261776659, "learning_rate": 2.3025628794761544e-06, "loss": 0.4937, "step": 3717 }, { "epoch": 3.56841094575132, "grad_norm": 0.793195680324484, "learning_rate": 2.2997428398910497e-06, "loss": 0.4962, "step": 3718 }, { "epoch": 3.5693710993759002, "grad_norm": 0.7734379161122867, "learning_rate": 2.2969240124225177e-06, "loss": 0.4612, "step": 3719 }, { "epoch": 3.57033125300048, "grad_norm": 1.0892681259800918, "learning_rate": 2.294106398335905e-06, "loss": 0.4682, "step": 3720 }, { "epoch": 3.5712914066250603, "grad_norm": 0.9067502719360504, "learning_rate": 2.2912899988960015e-06, "loss": 0.5504, "step": 3721 }, { "epoch": 3.57225156024964, "grad_norm": 1.1358278711826326, "learning_rate": 2.288474815367065e-06, "loss": 0.5321, "step": 3722 }, { "epoch": 3.57321171387422, "grad_norm": 0.8021731537536497, "learning_rate": 2.2856608490128005e-06, "loss": 0.5574, "step": 3723 }, { "epoch": 3.5741718674987997, "grad_norm": 0.8367682666477645, "learning_rate": 2.2828481010963647e-06, "loss": 0.5687, "step": 3724 }, { "epoch": 3.5751320211233795, "grad_norm": 0.9701483090588082, "learning_rate": 2.280036572880372e-06, "loss": 0.5141, "step": 3725 }, { "epoch": 3.5760921747479597, "grad_norm": 0.7405689055150886, "learning_rate": 2.277226265626887e-06, "loss": 0.498, "step": 3726 }, { "epoch": 3.5770523283725395, "grad_norm": 1.0607640619735226, "learning_rate": 2.2744171805974257e-06, "loss": 0.5437, "step": 3727 }, { "epoch": 3.5780124819971197, "grad_norm": 1.3256536106280572, "learning_rate": 2.27160931905296e-06, "loss": 0.5217, "step": 3728 }, { "epoch": 3.5789726356216995, "grad_norm": 0.8499638714565552, "learning_rate": 2.268802682253903e-06, "loss": 0.5438, "step": 3729 }, { "epoch": 3.5799327892462793, "grad_norm": 1.0428843837392756, "learning_rate": 2.2659972714601264e-06, "loss": 0.4809, "step": 3730 }, { "epoch": 3.580892942870859, "grad_norm": 1.0558274897180953, "learning_rate": 2.2631930879309525e-06, "loss": 0.5024, "step": 3731 }, { "epoch": 3.5818530964954394, "grad_norm": 0.9989907353518815, "learning_rate": 2.2603901329251452e-06, "loss": 0.4865, "step": 3732 }, { "epoch": 3.582813250120019, "grad_norm": 0.9969480803787961, "learning_rate": 2.2575884077009246e-06, "loss": 0.5383, "step": 3733 }, { "epoch": 3.5837734037445994, "grad_norm": 1.0463867639889124, "learning_rate": 2.2547879135159495e-06, "loss": 0.4363, "step": 3734 }, { "epoch": 3.584733557369179, "grad_norm": 0.7839250915618612, "learning_rate": 2.2519886516273365e-06, "loss": 0.5509, "step": 3735 }, { "epoch": 3.585693710993759, "grad_norm": 0.9750669029336699, "learning_rate": 2.2491906232916457e-06, "loss": 0.5629, "step": 3736 }, { "epoch": 3.586653864618339, "grad_norm": 0.923178926434716, "learning_rate": 2.246393829764877e-06, "loss": 0.5187, "step": 3737 }, { "epoch": 3.5876140182429186, "grad_norm": 0.6258244342806871, "learning_rate": 2.243598272302484e-06, "loss": 0.4103, "step": 3738 }, { "epoch": 3.588574171867499, "grad_norm": 0.6794865483468701, "learning_rate": 2.240803952159362e-06, "loss": 0.4594, "step": 3739 }, { "epoch": 3.5895343254920786, "grad_norm": 0.8285395852090254, "learning_rate": 2.238010870589852e-06, "loss": 0.502, "step": 3740 }, { "epoch": 3.590494479116659, "grad_norm": 0.864140122558366, "learning_rate": 2.23521902884774e-06, "loss": 0.4703, "step": 3741 }, { "epoch": 3.5914546327412387, "grad_norm": 1.2018771279628955, "learning_rate": 2.23242842818625e-06, "loss": 0.491, "step": 3742 }, { "epoch": 3.5924147863658185, "grad_norm": 0.921511528035369, "learning_rate": 2.2296390698580558e-06, "loss": 0.5042, "step": 3743 }, { "epoch": 3.5933749399903983, "grad_norm": 1.268619241198752, "learning_rate": 2.2268509551152684e-06, "loss": 0.4884, "step": 3744 }, { "epoch": 3.5943350936149785, "grad_norm": 0.8326295584839796, "learning_rate": 2.2240640852094446e-06, "loss": 0.4909, "step": 3745 }, { "epoch": 3.5952952472395583, "grad_norm": 1.1142338429530596, "learning_rate": 2.2212784613915816e-06, "loss": 0.4257, "step": 3746 }, { "epoch": 3.5962554008641385, "grad_norm": 0.9280008416327914, "learning_rate": 2.2184940849121123e-06, "loss": 0.4623, "step": 3747 }, { "epoch": 3.5972155544887183, "grad_norm": 1.339624311278489, "learning_rate": 2.215710957020914e-06, "loss": 0.4489, "step": 3748 }, { "epoch": 3.598175708113298, "grad_norm": 1.2447357407185704, "learning_rate": 2.2129290789673085e-06, "loss": 0.4174, "step": 3749 }, { "epoch": 3.599135861737878, "grad_norm": 0.8174623074174056, "learning_rate": 2.2101484520000456e-06, "loss": 0.4961, "step": 3750 }, { "epoch": 3.600096015362458, "grad_norm": 0.8813397315741608, "learning_rate": 2.2073690773673203e-06, "loss": 0.5879, "step": 3751 }, { "epoch": 3.601056168987038, "grad_norm": 0.9858830688756335, "learning_rate": 2.2045909563167643e-06, "loss": 0.5454, "step": 3752 }, { "epoch": 3.6020163226116177, "grad_norm": 1.071825634764588, "learning_rate": 2.201814090095448e-06, "loss": 0.4343, "step": 3753 }, { "epoch": 3.602976476236198, "grad_norm": 1.2559037488410367, "learning_rate": 2.1990384799498767e-06, "loss": 0.477, "step": 3754 }, { "epoch": 3.603936629860778, "grad_norm": 1.0732322054440497, "learning_rate": 2.196264127125989e-06, "loss": 0.4152, "step": 3755 }, { "epoch": 3.6048967834853576, "grad_norm": 0.8833088127981783, "learning_rate": 2.193491032869165e-06, "loss": 0.5496, "step": 3756 }, { "epoch": 3.6058569371099374, "grad_norm": 0.9373903238770231, "learning_rate": 2.190719198424216e-06, "loss": 0.5028, "step": 3757 }, { "epoch": 3.6068170907345176, "grad_norm": 0.944221245075924, "learning_rate": 2.1879486250353894e-06, "loss": 0.565, "step": 3758 }, { "epoch": 3.6077772443590974, "grad_norm": 0.7762321908174875, "learning_rate": 2.185179313946368e-06, "loss": 0.451, "step": 3759 }, { "epoch": 3.6087373979836777, "grad_norm": 0.9884961295715916, "learning_rate": 2.1824112664002612e-06, "loss": 0.4139, "step": 3760 }, { "epoch": 3.6096975516082574, "grad_norm": 0.8602140654319029, "learning_rate": 2.1796444836396174e-06, "loss": 0.5996, "step": 3761 }, { "epoch": 3.6106577052328372, "grad_norm": 0.7522838564489276, "learning_rate": 2.1768789669064205e-06, "loss": 0.5096, "step": 3762 }, { "epoch": 3.611617858857417, "grad_norm": 0.865929050790382, "learning_rate": 2.174114717442076e-06, "loss": 0.429, "step": 3763 }, { "epoch": 3.6125780124819973, "grad_norm": 1.0329563213718749, "learning_rate": 2.1713517364874297e-06, "loss": 0.5358, "step": 3764 }, { "epoch": 3.613538166106577, "grad_norm": 0.9268408785709851, "learning_rate": 2.1685900252827485e-06, "loss": 0.4882, "step": 3765 }, { "epoch": 3.614498319731157, "grad_norm": 0.7930876984124821, "learning_rate": 2.1658295850677408e-06, "loss": 0.5102, "step": 3766 }, { "epoch": 3.615458473355737, "grad_norm": 0.8080380746302102, "learning_rate": 2.1630704170815386e-06, "loss": 0.4922, "step": 3767 }, { "epoch": 3.616418626980317, "grad_norm": 0.8897966813749479, "learning_rate": 2.160312522562699e-06, "loss": 0.5415, "step": 3768 }, { "epoch": 3.6173787806048967, "grad_norm": 1.2506724525784365, "learning_rate": 2.1575559027492136e-06, "loss": 0.5716, "step": 3769 }, { "epoch": 3.6183389342294765, "grad_norm": 1.1789110341249605, "learning_rate": 2.1548005588785e-06, "loss": 0.4554, "step": 3770 }, { "epoch": 3.6192990878540567, "grad_norm": 1.0806530889541532, "learning_rate": 2.1520464921874014e-06, "loss": 0.4932, "step": 3771 }, { "epoch": 3.6202592414786365, "grad_norm": 0.8575609104140245, "learning_rate": 2.1492937039121914e-06, "loss": 0.458, "step": 3772 }, { "epoch": 3.6212193951032168, "grad_norm": 0.7025828329685111, "learning_rate": 2.1465421952885633e-06, "loss": 0.4966, "step": 3773 }, { "epoch": 3.6221795487277966, "grad_norm": 1.0435178446943953, "learning_rate": 2.1437919675516407e-06, "loss": 0.3375, "step": 3774 }, { "epoch": 3.6231397023523764, "grad_norm": 0.9311801568883682, "learning_rate": 2.1410430219359757e-06, "loss": 0.5872, "step": 3775 }, { "epoch": 3.624099855976956, "grad_norm": 0.8543696803419472, "learning_rate": 2.1382953596755364e-06, "loss": 0.4807, "step": 3776 }, { "epoch": 3.6250600096015364, "grad_norm": 0.6916361837633496, "learning_rate": 2.1355489820037216e-06, "loss": 0.4973, "step": 3777 }, { "epoch": 3.626020163226116, "grad_norm": 0.9532827470322741, "learning_rate": 2.132803890153346e-06, "loss": 0.463, "step": 3778 }, { "epoch": 3.626980316850696, "grad_norm": 0.8046263747412048, "learning_rate": 2.1300600853566577e-06, "loss": 0.5301, "step": 3779 }, { "epoch": 3.6279404704752762, "grad_norm": 0.8000239556285753, "learning_rate": 2.12731756884532e-06, "loss": 0.4908, "step": 3780 }, { "epoch": 3.628900624099856, "grad_norm": 0.7662083945062137, "learning_rate": 2.1245763418504173e-06, "loss": 0.5297, "step": 3781 }, { "epoch": 3.629860777724436, "grad_norm": 0.941786873033903, "learning_rate": 2.121836405602458e-06, "loss": 0.4988, "step": 3782 }, { "epoch": 3.6308209313490156, "grad_norm": 0.8989732964236349, "learning_rate": 2.1190977613313707e-06, "loss": 0.4814, "step": 3783 }, { "epoch": 3.631781084973596, "grad_norm": 0.7133662201381121, "learning_rate": 2.1163604102665023e-06, "loss": 0.5928, "step": 3784 }, { "epoch": 3.6327412385981757, "grad_norm": 1.119860069470823, "learning_rate": 2.113624353636624e-06, "loss": 0.5072, "step": 3785 }, { "epoch": 3.633701392222756, "grad_norm": 1.0525805397708063, "learning_rate": 2.110889592669918e-06, "loss": 0.5356, "step": 3786 }, { "epoch": 3.6346615458473357, "grad_norm": 1.0145001690730728, "learning_rate": 2.1081561285939904e-06, "loss": 0.5104, "step": 3787 }, { "epoch": 3.6356216994719155, "grad_norm": 0.7998860661726392, "learning_rate": 2.1054239626358648e-06, "loss": 0.487, "step": 3788 }, { "epoch": 3.6365818530964953, "grad_norm": 0.9074747210286014, "learning_rate": 2.102693096021981e-06, "loss": 0.4887, "step": 3789 }, { "epoch": 3.6375420067210755, "grad_norm": 0.86384478643989, "learning_rate": 2.0999635299781976e-06, "loss": 0.439, "step": 3790 }, { "epoch": 3.6385021603456553, "grad_norm": 0.9958558351679027, "learning_rate": 2.0972352657297845e-06, "loss": 0.525, "step": 3791 }, { "epoch": 3.639462313970235, "grad_norm": 0.9578893523076633, "learning_rate": 2.0945083045014296e-06, "loss": 0.5214, "step": 3792 }, { "epoch": 3.6404224675948154, "grad_norm": 0.9295285040593804, "learning_rate": 2.0917826475172427e-06, "loss": 0.4854, "step": 3793 }, { "epoch": 3.641382621219395, "grad_norm": 1.0226542329271635, "learning_rate": 2.089058296000736e-06, "loss": 0.5503, "step": 3794 }, { "epoch": 3.642342774843975, "grad_norm": 1.1423323995818417, "learning_rate": 2.0863352511748466e-06, "loss": 0.4928, "step": 3795 }, { "epoch": 3.6433029284685547, "grad_norm": 1.03533052920242, "learning_rate": 2.083613514261913e-06, "loss": 0.5024, "step": 3796 }, { "epoch": 3.644263082093135, "grad_norm": 0.7966661608393908, "learning_rate": 2.0808930864837007e-06, "loss": 0.6208, "step": 3797 }, { "epoch": 3.645223235717715, "grad_norm": 0.7883334448460687, "learning_rate": 2.07817396906138e-06, "loss": 0.4264, "step": 3798 }, { "epoch": 3.646183389342295, "grad_norm": 0.5758708409771067, "learning_rate": 2.0754561632155307e-06, "loss": 0.4441, "step": 3799 }, { "epoch": 3.647143542966875, "grad_norm": 0.8178823372497325, "learning_rate": 2.072739670166148e-06, "loss": 0.4318, "step": 3800 }, { "epoch": 3.6481036965914546, "grad_norm": 0.926084233904893, "learning_rate": 2.070024491132636e-06, "loss": 0.4599, "step": 3801 }, { "epoch": 3.6490638502160344, "grad_norm": 1.08663159537345, "learning_rate": 2.067310627333811e-06, "loss": 0.4268, "step": 3802 }, { "epoch": 3.6500240038406147, "grad_norm": 0.8427337129058597, "learning_rate": 2.064598079987898e-06, "loss": 0.5099, "step": 3803 }, { "epoch": 3.6509841574651944, "grad_norm": 0.8179603742132566, "learning_rate": 2.0618868503125273e-06, "loss": 0.5954, "step": 3804 }, { "epoch": 3.6519443110897742, "grad_norm": 0.7560249473168097, "learning_rate": 2.059176939524741e-06, "loss": 0.5672, "step": 3805 }, { "epoch": 3.6529044647143545, "grad_norm": 1.031656492262692, "learning_rate": 2.056468348840995e-06, "loss": 0.5254, "step": 3806 }, { "epoch": 3.6538646183389343, "grad_norm": 0.995269264687381, "learning_rate": 2.0537610794771405e-06, "loss": 0.4514, "step": 3807 }, { "epoch": 3.654824771963514, "grad_norm": 0.9803150355911217, "learning_rate": 2.0510551326484457e-06, "loss": 0.4248, "step": 3808 }, { "epoch": 3.655784925588094, "grad_norm": 1.0222193066111003, "learning_rate": 2.0483505095695755e-06, "loss": 0.5002, "step": 3809 }, { "epoch": 3.656745079212674, "grad_norm": 0.9737729068926451, "learning_rate": 2.045647211454613e-06, "loss": 0.4691, "step": 3810 }, { "epoch": 3.657705232837254, "grad_norm": 1.0341692563727112, "learning_rate": 2.0429452395170372e-06, "loss": 0.4879, "step": 3811 }, { "epoch": 3.658665386461834, "grad_norm": 1.0379264387272915, "learning_rate": 2.0402445949697335e-06, "loss": 0.5148, "step": 3812 }, { "epoch": 3.659625540086414, "grad_norm": 0.7472698330961901, "learning_rate": 2.037545279024993e-06, "loss": 0.5545, "step": 3813 }, { "epoch": 3.6605856937109937, "grad_norm": 0.8919794058321632, "learning_rate": 2.0348472928945096e-06, "loss": 0.4766, "step": 3814 }, { "epoch": 3.6615458473355735, "grad_norm": 1.062392202332295, "learning_rate": 2.0321506377893807e-06, "loss": 0.5221, "step": 3815 }, { "epoch": 3.6625060009601538, "grad_norm": 0.8921050375491522, "learning_rate": 2.029455314920108e-06, "loss": 0.5038, "step": 3816 }, { "epoch": 3.6634661545847336, "grad_norm": 0.7802226070072077, "learning_rate": 2.026761325496589e-06, "loss": 0.511, "step": 3817 }, { "epoch": 3.6644263082093134, "grad_norm": 0.645453865491683, "learning_rate": 2.024068670728127e-06, "loss": 0.5026, "step": 3818 }, { "epoch": 3.6653864618338936, "grad_norm": 1.3946724699486888, "learning_rate": 2.021377351823432e-06, "loss": 0.4341, "step": 3819 }, { "epoch": 3.6663466154584734, "grad_norm": 1.8986763601403547, "learning_rate": 2.0186873699906017e-06, "loss": 0.545, "step": 3820 }, { "epoch": 3.667306769083053, "grad_norm": 0.9354063498424608, "learning_rate": 2.0159987264371445e-06, "loss": 0.5067, "step": 3821 }, { "epoch": 3.668266922707633, "grad_norm": 0.8701475253703482, "learning_rate": 2.0133114223699577e-06, "loss": 0.5577, "step": 3822 }, { "epoch": 3.6692270763322132, "grad_norm": 0.7750802536799141, "learning_rate": 2.01062545899535e-06, "loss": 0.4922, "step": 3823 }, { "epoch": 3.670187229956793, "grad_norm": 0.8723712586301335, "learning_rate": 2.007940837519021e-06, "loss": 0.4411, "step": 3824 }, { "epoch": 3.6711473835813733, "grad_norm": 1.0178903933025611, "learning_rate": 2.0052575591460636e-06, "loss": 0.4691, "step": 3825 }, { "epoch": 3.672107537205953, "grad_norm": 0.8826170186132017, "learning_rate": 2.002575625080977e-06, "loss": 0.5039, "step": 3826 }, { "epoch": 3.673067690830533, "grad_norm": 0.9869206414792273, "learning_rate": 1.9998950365276514e-06, "loss": 0.5201, "step": 3827 }, { "epoch": 3.6740278444551127, "grad_norm": 0.8560684597124795, "learning_rate": 1.997215794689375e-06, "loss": 0.5303, "step": 3828 }, { "epoch": 3.674987998079693, "grad_norm": 0.7383127909867832, "learning_rate": 1.9945379007688314e-06, "loss": 0.4928, "step": 3829 }, { "epoch": 3.6759481517042727, "grad_norm": 0.7468076027525046, "learning_rate": 1.991861355968096e-06, "loss": 0.5117, "step": 3830 }, { "epoch": 3.6769083053288525, "grad_norm": 0.8463043712303153, "learning_rate": 1.989186161488643e-06, "loss": 0.4848, "step": 3831 }, { "epoch": 3.6778684589534327, "grad_norm": 1.3024063440870273, "learning_rate": 1.986512318531338e-06, "loss": 0.4711, "step": 3832 }, { "epoch": 3.6788286125780125, "grad_norm": 0.9451838459506192, "learning_rate": 1.9838398282964412e-06, "loss": 0.5723, "step": 3833 }, { "epoch": 3.6797887662025923, "grad_norm": 0.8984380295977038, "learning_rate": 1.9811686919836054e-06, "loss": 0.5105, "step": 3834 }, { "epoch": 3.680748919827172, "grad_norm": 0.9644010017002285, "learning_rate": 1.9784989107918736e-06, "loss": 0.5156, "step": 3835 }, { "epoch": 3.6817090734517524, "grad_norm": 0.8869514781495743, "learning_rate": 1.97583048591968e-06, "loss": 0.4587, "step": 3836 }, { "epoch": 3.682669227076332, "grad_norm": 0.9684046292656198, "learning_rate": 1.9731634185648584e-06, "loss": 0.4465, "step": 3837 }, { "epoch": 3.6836293807009124, "grad_norm": 0.8799628084194613, "learning_rate": 1.970497709924622e-06, "loss": 0.4281, "step": 3838 }, { "epoch": 3.684589534325492, "grad_norm": 0.8535431548427477, "learning_rate": 1.9678333611955814e-06, "loss": 0.5537, "step": 3839 }, { "epoch": 3.685549687950072, "grad_norm": 0.903273864584626, "learning_rate": 1.965170373573729e-06, "loss": 0.4548, "step": 3840 }, { "epoch": 3.686509841574652, "grad_norm": 0.7350155536781002, "learning_rate": 1.9625087482544574e-06, "loss": 0.446, "step": 3841 }, { "epoch": 3.687469995199232, "grad_norm": 0.9795259354745636, "learning_rate": 1.959848486432542e-06, "loss": 0.4574, "step": 3842 }, { "epoch": 3.688430148823812, "grad_norm": 1.0993709057301138, "learning_rate": 1.957189589302141e-06, "loss": 0.4898, "step": 3843 }, { "epoch": 3.6893903024483916, "grad_norm": 0.7866786762507819, "learning_rate": 1.9545320580568077e-06, "loss": 0.5275, "step": 3844 }, { "epoch": 3.690350456072972, "grad_norm": 1.1158189243778438, "learning_rate": 1.95187589388948e-06, "loss": 0.5285, "step": 3845 }, { "epoch": 3.6913106096975516, "grad_norm": 1.1284373206097664, "learning_rate": 1.9492210979924805e-06, "loss": 0.5449, "step": 3846 }, { "epoch": 3.6922707633221314, "grad_norm": 0.7979221116671086, "learning_rate": 1.9465676715575203e-06, "loss": 0.5389, "step": 3847 }, { "epoch": 3.6932309169467112, "grad_norm": 1.2522482985777512, "learning_rate": 1.9439156157756917e-06, "loss": 0.4553, "step": 3848 }, { "epoch": 3.6941910705712915, "grad_norm": 1.25820112236808, "learning_rate": 1.9412649318374737e-06, "loss": 0.582, "step": 3849 }, { "epoch": 3.6951512241958713, "grad_norm": 0.8550605787518847, "learning_rate": 1.938615620932735e-06, "loss": 0.5143, "step": 3850 }, { "epoch": 3.6961113778204515, "grad_norm": 0.8137929552165107, "learning_rate": 1.9359676842507186e-06, "loss": 0.5665, "step": 3851 }, { "epoch": 3.6970715314450313, "grad_norm": 0.8441896784066916, "learning_rate": 1.933321122980057e-06, "loss": 0.5128, "step": 3852 }, { "epoch": 3.698031685069611, "grad_norm": 0.7850347983158845, "learning_rate": 1.9306759383087593e-06, "loss": 0.5643, "step": 3853 }, { "epoch": 3.698991838694191, "grad_norm": 1.0328012354856644, "learning_rate": 1.9280321314242252e-06, "loss": 0.4548, "step": 3854 }, { "epoch": 3.699951992318771, "grad_norm": 1.1001603937557543, "learning_rate": 1.925389703513232e-06, "loss": 0.396, "step": 3855 }, { "epoch": 3.700912145943351, "grad_norm": 0.8130512509029146, "learning_rate": 1.922748655761932e-06, "loss": 0.5294, "step": 3856 }, { "epoch": 3.7018722995679307, "grad_norm": 0.873188303261054, "learning_rate": 1.9201089893558677e-06, "loss": 0.5102, "step": 3857 }, { "epoch": 3.702832453192511, "grad_norm": 0.8535674373436822, "learning_rate": 1.9174707054799558e-06, "loss": 0.5369, "step": 3858 }, { "epoch": 3.7037926068170908, "grad_norm": 0.8480645324149666, "learning_rate": 1.9148338053184933e-06, "loss": 0.4198, "step": 3859 }, { "epoch": 3.7047527604416706, "grad_norm": 0.9971841488675032, "learning_rate": 1.9121982900551592e-06, "loss": 0.5152, "step": 3860 }, { "epoch": 3.7057129140662504, "grad_norm": 1.0445150084921866, "learning_rate": 1.909564160873004e-06, "loss": 0.5425, "step": 3861 }, { "epoch": 3.7066730676908306, "grad_norm": 1.1685086658623076, "learning_rate": 1.9069314189544602e-06, "loss": 0.4755, "step": 3862 }, { "epoch": 3.7076332213154104, "grad_norm": 0.8938301048032731, "learning_rate": 1.9043000654813433e-06, "loss": 0.465, "step": 3863 }, { "epoch": 3.7085933749399906, "grad_norm": 0.7113685751482485, "learning_rate": 1.9016701016348338e-06, "loss": 0.5509, "step": 3864 }, { "epoch": 3.7095535285645704, "grad_norm": 0.9320261965778083, "learning_rate": 1.8990415285954976e-06, "loss": 0.4377, "step": 3865 }, { "epoch": 3.7105136821891502, "grad_norm": 1.016872807058208, "learning_rate": 1.896414347543269e-06, "loss": 0.4787, "step": 3866 }, { "epoch": 3.71147383581373, "grad_norm": 0.806459322414318, "learning_rate": 1.8937885596574656e-06, "loss": 0.4871, "step": 3867 }, { "epoch": 3.7124339894383103, "grad_norm": 0.8753822355773663, "learning_rate": 1.8911641661167762e-06, "loss": 0.4839, "step": 3868 }, { "epoch": 3.71339414306289, "grad_norm": 1.164006424217447, "learning_rate": 1.8885411680992594e-06, "loss": 0.4774, "step": 3869 }, { "epoch": 3.71435429668747, "grad_norm": 0.6826735911891645, "learning_rate": 1.885919566782352e-06, "loss": 0.4419, "step": 3870 }, { "epoch": 3.71531445031205, "grad_norm": 0.8401645929392563, "learning_rate": 1.8832993633428643e-06, "loss": 0.4651, "step": 3871 }, { "epoch": 3.71627460393663, "grad_norm": 0.8863737697174425, "learning_rate": 1.880680558956976e-06, "loss": 0.5187, "step": 3872 }, { "epoch": 3.7172347575612097, "grad_norm": 1.1399252865377636, "learning_rate": 1.878063154800243e-06, "loss": 0.4085, "step": 3873 }, { "epoch": 3.7181949111857895, "grad_norm": 0.8536011919083282, "learning_rate": 1.8754471520475863e-06, "loss": 0.5728, "step": 3874 }, { "epoch": 3.7191550648103697, "grad_norm": 0.8730799840461532, "learning_rate": 1.8728325518733038e-06, "loss": 0.5084, "step": 3875 }, { "epoch": 3.7201152184349495, "grad_norm": 0.8665762940548026, "learning_rate": 1.870219355451061e-06, "loss": 0.4712, "step": 3876 }, { "epoch": 3.7210753720595298, "grad_norm": 0.8135934841114831, "learning_rate": 1.8676075639538943e-06, "loss": 0.5593, "step": 3877 }, { "epoch": 3.7220355256841096, "grad_norm": 0.9187047313785206, "learning_rate": 1.8649971785542097e-06, "loss": 0.4908, "step": 3878 }, { "epoch": 3.7229956793086894, "grad_norm": 0.992810890430203, "learning_rate": 1.8623882004237792e-06, "loss": 0.4667, "step": 3879 }, { "epoch": 3.723955832933269, "grad_norm": 1.3142674375909322, "learning_rate": 1.8597806307337446e-06, "loss": 0.6233, "step": 3880 }, { "epoch": 3.7249159865578494, "grad_norm": 0.9945921542726417, "learning_rate": 1.8571744706546213e-06, "loss": 0.4547, "step": 3881 }, { "epoch": 3.725876140182429, "grad_norm": 0.8803041157921664, "learning_rate": 1.8545697213562813e-06, "loss": 0.5941, "step": 3882 }, { "epoch": 3.726836293807009, "grad_norm": 0.9212313053559301, "learning_rate": 1.8519663840079721e-06, "loss": 0.4633, "step": 3883 }, { "epoch": 3.727796447431589, "grad_norm": 0.7888584787173604, "learning_rate": 1.8493644597782996e-06, "loss": 0.5648, "step": 3884 }, { "epoch": 3.728756601056169, "grad_norm": 0.8168582761873627, "learning_rate": 1.8467639498352442e-06, "loss": 0.4733, "step": 3885 }, { "epoch": 3.729716754680749, "grad_norm": 0.9275825383941655, "learning_rate": 1.8441648553461467e-06, "loss": 0.4361, "step": 3886 }, { "epoch": 3.7306769083053286, "grad_norm": 0.9560093747825662, "learning_rate": 1.8415671774777106e-06, "loss": 0.5131, "step": 3887 }, { "epoch": 3.731637061929909, "grad_norm": 0.8314953661469073, "learning_rate": 1.8389709173960063e-06, "loss": 0.5364, "step": 3888 }, { "epoch": 3.7325972155544886, "grad_norm": 1.5202721906714411, "learning_rate": 1.8363760762664678e-06, "loss": 0.4436, "step": 3889 }, { "epoch": 3.733557369179069, "grad_norm": 0.9866282014324954, "learning_rate": 1.8337826552538911e-06, "loss": 0.6272, "step": 3890 }, { "epoch": 3.7345175228036487, "grad_norm": 1.0009477035872365, "learning_rate": 1.8311906555224379e-06, "loss": 0.4887, "step": 3891 }, { "epoch": 3.7354776764282285, "grad_norm": 1.0092720448657926, "learning_rate": 1.8286000782356245e-06, "loss": 0.3975, "step": 3892 }, { "epoch": 3.7364378300528083, "grad_norm": 0.985125275406632, "learning_rate": 1.826010924556334e-06, "loss": 0.4364, "step": 3893 }, { "epoch": 3.7373979836773885, "grad_norm": 0.8278741831856571, "learning_rate": 1.823423195646815e-06, "loss": 0.5928, "step": 3894 }, { "epoch": 3.7383581373019683, "grad_norm": 1.0532816264050582, "learning_rate": 1.820836892668666e-06, "loss": 0.524, "step": 3895 }, { "epoch": 3.739318290926548, "grad_norm": 0.7513111122640018, "learning_rate": 1.8182520167828537e-06, "loss": 0.5707, "step": 3896 }, { "epoch": 3.7402784445511283, "grad_norm": 0.7947209252643154, "learning_rate": 1.8156685691496972e-06, "loss": 0.5718, "step": 3897 }, { "epoch": 3.741238598175708, "grad_norm": 0.8573046475617356, "learning_rate": 1.813086550928883e-06, "loss": 0.5228, "step": 3898 }, { "epoch": 3.742198751800288, "grad_norm": 0.9589644209666265, "learning_rate": 1.8105059632794514e-06, "loss": 0.4166, "step": 3899 }, { "epoch": 3.7431589054248677, "grad_norm": 1.1980927099822072, "learning_rate": 1.8079268073597978e-06, "loss": 0.5017, "step": 3900 }, { "epoch": 3.744119059049448, "grad_norm": 0.812548666481886, "learning_rate": 1.8053490843276788e-06, "loss": 0.5476, "step": 3901 }, { "epoch": 3.7450792126740278, "grad_norm": 0.8464961303252727, "learning_rate": 1.802772795340207e-06, "loss": 0.5346, "step": 3902 }, { "epoch": 3.746039366298608, "grad_norm": 0.9044411925586132, "learning_rate": 1.8001979415538507e-06, "loss": 0.5077, "step": 3903 }, { "epoch": 3.746999519923188, "grad_norm": 0.8693616242002253, "learning_rate": 1.797624524124436e-06, "loss": 0.4959, "step": 3904 }, { "epoch": 3.7479596735477676, "grad_norm": 1.022863475962555, "learning_rate": 1.7950525442071394e-06, "loss": 0.4943, "step": 3905 }, { "epoch": 3.7489198271723474, "grad_norm": 0.9943400776870912, "learning_rate": 1.792482002956495e-06, "loss": 0.5886, "step": 3906 }, { "epoch": 3.7498799807969276, "grad_norm": 1.369537914655778, "learning_rate": 1.7899129015263966e-06, "loss": 0.4073, "step": 3907 }, { "epoch": 3.7508401344215074, "grad_norm": 1.4046007181501785, "learning_rate": 1.7873452410700809e-06, "loss": 0.3702, "step": 3908 }, { "epoch": 3.7518002880460872, "grad_norm": 0.9899861861387333, "learning_rate": 1.7847790227401469e-06, "loss": 0.402, "step": 3909 }, { "epoch": 3.7527604416706675, "grad_norm": 1.194455910041424, "learning_rate": 1.7822142476885374e-06, "loss": 0.5171, "step": 3910 }, { "epoch": 3.7537205952952473, "grad_norm": 1.3871438350158458, "learning_rate": 1.7796509170665571e-06, "loss": 0.573, "step": 3911 }, { "epoch": 3.754680748919827, "grad_norm": 0.7856498868311365, "learning_rate": 1.777089032024859e-06, "loss": 0.4671, "step": 3912 }, { "epoch": 3.755640902544407, "grad_norm": 1.0138729481315152, "learning_rate": 1.774528593713441e-06, "loss": 0.5252, "step": 3913 }, { "epoch": 3.756601056168987, "grad_norm": 0.91317929655522, "learning_rate": 1.771969603281659e-06, "loss": 0.4878, "step": 3914 }, { "epoch": 3.757561209793567, "grad_norm": 1.0126046874664627, "learning_rate": 1.7694120618782169e-06, "loss": 0.5552, "step": 3915 }, { "epoch": 3.758521363418147, "grad_norm": 1.173012882698823, "learning_rate": 1.7668559706511667e-06, "loss": 0.4993, "step": 3916 }, { "epoch": 3.759481517042727, "grad_norm": 0.9646322024264772, "learning_rate": 1.7643013307479129e-06, "loss": 0.5414, "step": 3917 }, { "epoch": 3.7604416706673067, "grad_norm": 0.9305808795966658, "learning_rate": 1.7617481433152022e-06, "loss": 0.4831, "step": 3918 }, { "epoch": 3.7614018242918865, "grad_norm": 0.9697038775269641, "learning_rate": 1.759196409499135e-06, "loss": 0.4873, "step": 3919 }, { "epoch": 3.7623619779164668, "grad_norm": 0.8305581089570826, "learning_rate": 1.756646130445157e-06, "loss": 0.5645, "step": 3920 }, { "epoch": 3.7633221315410466, "grad_norm": 0.7868142638420784, "learning_rate": 1.7540973072980621e-06, "loss": 0.4732, "step": 3921 }, { "epoch": 3.7642822851656264, "grad_norm": 0.7829320199735725, "learning_rate": 1.7515499412019903e-06, "loss": 0.4408, "step": 3922 }, { "epoch": 3.7652424387902066, "grad_norm": 0.6912706635247561, "learning_rate": 1.7490040333004248e-06, "loss": 0.4526, "step": 3923 }, { "epoch": 3.7662025924147864, "grad_norm": 0.9070540133599901, "learning_rate": 1.7464595847361954e-06, "loss": 0.4647, "step": 3924 }, { "epoch": 3.767162746039366, "grad_norm": 0.8231270437646648, "learning_rate": 1.7439165966514831e-06, "loss": 0.5231, "step": 3925 }, { "epoch": 3.768122899663946, "grad_norm": 0.9203359413336057, "learning_rate": 1.741375070187804e-06, "loss": 0.5661, "step": 3926 }, { "epoch": 3.769083053288526, "grad_norm": 1.2729879009594947, "learning_rate": 1.7388350064860242e-06, "loss": 0.4624, "step": 3927 }, { "epoch": 3.770043206913106, "grad_norm": 0.8224558026770968, "learning_rate": 1.7362964066863469e-06, "loss": 0.5093, "step": 3928 }, { "epoch": 3.7710033605376863, "grad_norm": 1.1331052873317395, "learning_rate": 1.7337592719283276e-06, "loss": 0.414, "step": 3929 }, { "epoch": 3.771963514162266, "grad_norm": 1.1455539980262865, "learning_rate": 1.7312236033508584e-06, "loss": 0.4717, "step": 3930 }, { "epoch": 3.772923667786846, "grad_norm": 0.7987557405214618, "learning_rate": 1.7286894020921707e-06, "loss": 0.4329, "step": 3931 }, { "epoch": 3.7738838214114256, "grad_norm": 1.052994567979851, "learning_rate": 1.726156669289842e-06, "loss": 0.427, "step": 3932 }, { "epoch": 3.774843975036006, "grad_norm": 0.8040422840852289, "learning_rate": 1.7236254060807889e-06, "loss": 0.4827, "step": 3933 }, { "epoch": 3.7758041286605857, "grad_norm": 1.3099973233046538, "learning_rate": 1.7210956136012684e-06, "loss": 0.5269, "step": 3934 }, { "epoch": 3.7767642822851655, "grad_norm": 1.4628513184134746, "learning_rate": 1.718567292986879e-06, "loss": 0.5717, "step": 3935 }, { "epoch": 3.7777244359097457, "grad_norm": 0.8041894718454052, "learning_rate": 1.7160404453725543e-06, "loss": 0.4266, "step": 3936 }, { "epoch": 3.7786845895343255, "grad_norm": 0.9130413708113886, "learning_rate": 1.7135150718925674e-06, "loss": 0.4916, "step": 3937 }, { "epoch": 3.7796447431589053, "grad_norm": 0.7290900789987668, "learning_rate": 1.7109911736805385e-06, "loss": 0.4333, "step": 3938 }, { "epoch": 3.780604896783485, "grad_norm": 0.7169291784096596, "learning_rate": 1.708468751869412e-06, "loss": 0.546, "step": 3939 }, { "epoch": 3.7815650504080653, "grad_norm": 1.1096573417081423, "learning_rate": 1.7059478075914799e-06, "loss": 0.4567, "step": 3940 }, { "epoch": 3.782525204032645, "grad_norm": 0.8487093821359619, "learning_rate": 1.703428341978362e-06, "loss": 0.3654, "step": 3941 }, { "epoch": 3.7834853576572254, "grad_norm": 0.8008857619661854, "learning_rate": 1.7009103561610252e-06, "loss": 0.4744, "step": 3942 }, { "epoch": 3.784445511281805, "grad_norm": 0.8769721792355868, "learning_rate": 1.6983938512697655e-06, "loss": 0.5168, "step": 3943 }, { "epoch": 3.785405664906385, "grad_norm": 1.0537531886676172, "learning_rate": 1.695878828434212e-06, "loss": 0.5467, "step": 3944 }, { "epoch": 3.7863658185309648, "grad_norm": 1.327611291774154, "learning_rate": 1.6933652887833336e-06, "loss": 0.4432, "step": 3945 }, { "epoch": 3.787325972155545, "grad_norm": 1.0819815750984187, "learning_rate": 1.6908532334454308e-06, "loss": 0.4399, "step": 3946 }, { "epoch": 3.788286125780125, "grad_norm": 0.8517312791415846, "learning_rate": 1.6883426635481392e-06, "loss": 0.4373, "step": 3947 }, { "epoch": 3.7892462794047046, "grad_norm": 0.9145605945493591, "learning_rate": 1.6858335802184278e-06, "loss": 0.5742, "step": 3948 }, { "epoch": 3.790206433029285, "grad_norm": 0.9746489536168906, "learning_rate": 1.683325984582594e-06, "loss": 0.5827, "step": 3949 }, { "epoch": 3.7911665866538646, "grad_norm": 0.8151805710085742, "learning_rate": 1.680819877766272e-06, "loss": 0.6162, "step": 3950 }, { "epoch": 3.7921267402784444, "grad_norm": 0.9051821352600812, "learning_rate": 1.678315260894427e-06, "loss": 0.4655, "step": 3951 }, { "epoch": 3.7930868939030242, "grad_norm": 1.1961267517039198, "learning_rate": 1.6758121350913548e-06, "loss": 0.5374, "step": 3952 }, { "epoch": 3.7940470475276045, "grad_norm": 1.4289426772600347, "learning_rate": 1.6733105014806834e-06, "loss": 0.5505, "step": 3953 }, { "epoch": 3.7950072011521843, "grad_norm": 0.9637283242361324, "learning_rate": 1.6708103611853638e-06, "loss": 0.5219, "step": 3954 }, { "epoch": 3.7959673547767645, "grad_norm": 1.0571707038057643, "learning_rate": 1.6683117153276878e-06, "loss": 0.4816, "step": 3955 }, { "epoch": 3.7969275084013443, "grad_norm": 0.8921399441354533, "learning_rate": 1.6658145650292707e-06, "loss": 0.4689, "step": 3956 }, { "epoch": 3.797887662025924, "grad_norm": 1.1053075050969325, "learning_rate": 1.6633189114110531e-06, "loss": 0.4676, "step": 3957 }, { "epoch": 3.798847815650504, "grad_norm": 0.7620582148910336, "learning_rate": 1.6608247555933093e-06, "loss": 0.5017, "step": 3958 }, { "epoch": 3.799807969275084, "grad_norm": 0.9490577362828869, "learning_rate": 1.6583320986956386e-06, "loss": 0.4914, "step": 3959 }, { "epoch": 3.800768122899664, "grad_norm": 1.1112887817523307, "learning_rate": 1.6558409418369686e-06, "loss": 0.4855, "step": 3960 }, { "epoch": 3.8017282765242437, "grad_norm": 1.0793623596886386, "learning_rate": 1.6533512861355543e-06, "loss": 0.5527, "step": 3961 }, { "epoch": 3.802688430148824, "grad_norm": 1.11374748290245, "learning_rate": 1.6508631327089725e-06, "loss": 0.5119, "step": 3962 }, { "epoch": 3.8036485837734038, "grad_norm": 0.9390890414622188, "learning_rate": 1.64837648267413e-06, "loss": 0.5509, "step": 3963 }, { "epoch": 3.8046087373979836, "grad_norm": 0.8362692973648147, "learning_rate": 1.6458913371472573e-06, "loss": 0.5755, "step": 3964 }, { "epoch": 3.8055688910225633, "grad_norm": 0.8400659053712342, "learning_rate": 1.6434076972439106e-06, "loss": 0.4359, "step": 3965 }, { "epoch": 3.8065290446471436, "grad_norm": 0.9194437083092062, "learning_rate": 1.6409255640789695e-06, "loss": 0.4853, "step": 3966 }, { "epoch": 3.8074891982717234, "grad_norm": 0.8544348122528072, "learning_rate": 1.6384449387666351e-06, "loss": 0.5881, "step": 3967 }, { "epoch": 3.8084493518963036, "grad_norm": 0.9341065707062399, "learning_rate": 1.6359658224204328e-06, "loss": 0.539, "step": 3968 }, { "epoch": 3.8094095055208834, "grad_norm": 0.9295277765657625, "learning_rate": 1.6334882161532161e-06, "loss": 0.4507, "step": 3969 }, { "epoch": 3.810369659145463, "grad_norm": 0.8290416673990906, "learning_rate": 1.6310121210771524e-06, "loss": 0.5887, "step": 3970 }, { "epoch": 3.811329812770043, "grad_norm": 1.0045675402392988, "learning_rate": 1.628537538303736e-06, "loss": 0.4607, "step": 3971 }, { "epoch": 3.8122899663946233, "grad_norm": 0.8509987273130786, "learning_rate": 1.6260644689437767e-06, "loss": 0.4079, "step": 3972 }, { "epoch": 3.813250120019203, "grad_norm": 0.8050324597963481, "learning_rate": 1.6235929141074137e-06, "loss": 0.5108, "step": 3973 }, { "epoch": 3.814210273643783, "grad_norm": 0.9001162269689036, "learning_rate": 1.6211228749041007e-06, "loss": 0.5329, "step": 3974 }, { "epoch": 3.815170427268363, "grad_norm": 0.9999019411511203, "learning_rate": 1.6186543524426101e-06, "loss": 0.4956, "step": 3975 }, { "epoch": 3.816130580892943, "grad_norm": 1.1079573120839563, "learning_rate": 1.616187347831036e-06, "loss": 0.4543, "step": 3976 }, { "epoch": 3.8170907345175227, "grad_norm": 1.4442363355141827, "learning_rate": 1.6137218621767902e-06, "loss": 0.5907, "step": 3977 }, { "epoch": 3.8180508881421025, "grad_norm": 0.8103572053249724, "learning_rate": 1.6112578965866039e-06, "loss": 0.4794, "step": 3978 }, { "epoch": 3.8190110417666827, "grad_norm": 1.3293409860105172, "learning_rate": 1.608795452166525e-06, "loss": 0.4897, "step": 3979 }, { "epoch": 3.8199711953912625, "grad_norm": 1.1196197128883192, "learning_rate": 1.6063345300219162e-06, "loss": 0.5508, "step": 3980 }, { "epoch": 3.8209313490158427, "grad_norm": 1.1010593349142082, "learning_rate": 1.6038751312574592e-06, "loss": 0.5194, "step": 3981 }, { "epoch": 3.8218915026404225, "grad_norm": 0.9405254780045041, "learning_rate": 1.6014172569771564e-06, "loss": 0.362, "step": 3982 }, { "epoch": 3.8228516562650023, "grad_norm": 1.1618545891680916, "learning_rate": 1.598960908284316e-06, "loss": 0.6136, "step": 3983 }, { "epoch": 3.823811809889582, "grad_norm": 0.9558435771728494, "learning_rate": 1.5965060862815713e-06, "loss": 0.5488, "step": 3984 }, { "epoch": 3.8247719635141624, "grad_norm": 1.0675963057706244, "learning_rate": 1.5940527920708594e-06, "loss": 0.5649, "step": 3985 }, { "epoch": 3.825732117138742, "grad_norm": 0.6679846842705971, "learning_rate": 1.5916010267534433e-06, "loss": 0.5888, "step": 3986 }, { "epoch": 3.826692270763322, "grad_norm": 0.6167945281119668, "learning_rate": 1.5891507914298936e-06, "loss": 0.4764, "step": 3987 }, { "epoch": 3.827652424387902, "grad_norm": 0.9020946629486801, "learning_rate": 1.5867020872000927e-06, "loss": 0.4759, "step": 3988 }, { "epoch": 3.828612578012482, "grad_norm": 1.0321723091671204, "learning_rate": 1.5842549151632386e-06, "loss": 0.5604, "step": 3989 }, { "epoch": 3.829572731637062, "grad_norm": 1.272567950833854, "learning_rate": 1.5818092764178405e-06, "loss": 0.4288, "step": 3990 }, { "epoch": 3.8305328852616416, "grad_norm": 1.0085824423577194, "learning_rate": 1.5793651720617203e-06, "loss": 0.5255, "step": 3991 }, { "epoch": 3.831493038886222, "grad_norm": 0.7840594889785698, "learning_rate": 1.576922603192011e-06, "loss": 0.4571, "step": 3992 }, { "epoch": 3.8324531925108016, "grad_norm": 1.409729430104256, "learning_rate": 1.5744815709051532e-06, "loss": 0.5425, "step": 3993 }, { "epoch": 3.833413346135382, "grad_norm": 0.9655086730493586, "learning_rate": 1.5720420762969012e-06, "loss": 0.4534, "step": 3994 }, { "epoch": 3.8343734997599617, "grad_norm": 0.8210889717042036, "learning_rate": 1.5696041204623185e-06, "loss": 0.4426, "step": 3995 }, { "epoch": 3.8353336533845415, "grad_norm": 0.8441696785523906, "learning_rate": 1.5671677044957772e-06, "loss": 0.539, "step": 3996 }, { "epoch": 3.8362938070091213, "grad_norm": 0.6582017044533567, "learning_rate": 1.5647328294909598e-06, "loss": 0.513, "step": 3997 }, { "epoch": 3.8372539606337015, "grad_norm": 0.7568822979606367, "learning_rate": 1.5622994965408527e-06, "loss": 0.5163, "step": 3998 }, { "epoch": 3.8382141142582813, "grad_norm": 0.8127076783702565, "learning_rate": 1.5598677067377533e-06, "loss": 0.4661, "step": 3999 }, { "epoch": 3.839174267882861, "grad_norm": 0.7802066795763206, "learning_rate": 1.55743746117327e-06, "loss": 0.4876, "step": 4000 }, { "epoch": 3.8401344215074413, "grad_norm": 0.903354913849064, "learning_rate": 1.5550087609383102e-06, "loss": 0.43, "step": 4001 }, { "epoch": 3.841094575132021, "grad_norm": 0.8036141734601326, "learning_rate": 1.5525816071230932e-06, "loss": 0.4716, "step": 4002 }, { "epoch": 3.842054728756601, "grad_norm": 1.1587313410873723, "learning_rate": 1.5501560008171412e-06, "loss": 0.4544, "step": 4003 }, { "epoch": 3.8430148823811807, "grad_norm": 1.3464943371351896, "learning_rate": 1.5477319431092835e-06, "loss": 0.5132, "step": 4004 }, { "epoch": 3.843975036005761, "grad_norm": 0.6265898804608044, "learning_rate": 1.5453094350876563e-06, "loss": 0.4477, "step": 4005 }, { "epoch": 3.8449351896303408, "grad_norm": 0.9446916122857235, "learning_rate": 1.5428884778396929e-06, "loss": 0.4961, "step": 4006 }, { "epoch": 3.845895343254921, "grad_norm": 0.8814582563911034, "learning_rate": 1.540469072452137e-06, "loss": 0.3979, "step": 4007 }, { "epoch": 3.846855496879501, "grad_norm": 0.8777803423370264, "learning_rate": 1.5380512200110347e-06, "loss": 0.4239, "step": 4008 }, { "epoch": 3.8478156505040806, "grad_norm": 0.7624372154387893, "learning_rate": 1.5356349216017336e-06, "loss": 0.5122, "step": 4009 }, { "epoch": 3.8487758041286604, "grad_norm": 0.9267448911318271, "learning_rate": 1.5332201783088857e-06, "loss": 0.5678, "step": 4010 }, { "epoch": 3.8497359577532406, "grad_norm": 0.9886026327569047, "learning_rate": 1.5308069912164403e-06, "loss": 0.5485, "step": 4011 }, { "epoch": 3.8506961113778204, "grad_norm": 0.607462093167763, "learning_rate": 1.5283953614076514e-06, "loss": 0.446, "step": 4012 }, { "epoch": 3.8516562650024007, "grad_norm": 1.1368592046566988, "learning_rate": 1.5259852899650784e-06, "loss": 0.5596, "step": 4013 }, { "epoch": 3.8526164186269805, "grad_norm": 1.114966359161061, "learning_rate": 1.5235767779705717e-06, "loss": 0.5026, "step": 4014 }, { "epoch": 3.8535765722515603, "grad_norm": 0.8832283012536191, "learning_rate": 1.5211698265052905e-06, "loss": 0.4709, "step": 4015 }, { "epoch": 3.85453672587614, "grad_norm": 0.890974628808223, "learning_rate": 1.518764436649684e-06, "loss": 0.4858, "step": 4016 }, { "epoch": 3.85549687950072, "grad_norm": 0.8943417886449168, "learning_rate": 1.5163606094835115e-06, "loss": 0.4979, "step": 4017 }, { "epoch": 3.8564570331253, "grad_norm": 1.0963895124068708, "learning_rate": 1.5139583460858249e-06, "loss": 0.4354, "step": 4018 }, { "epoch": 3.85741718674988, "grad_norm": 0.9014673889989345, "learning_rate": 1.5115576475349715e-06, "loss": 0.5288, "step": 4019 }, { "epoch": 3.85837734037446, "grad_norm": 0.9219847846070074, "learning_rate": 1.5091585149086008e-06, "loss": 0.394, "step": 4020 }, { "epoch": 3.85933749399904, "grad_norm": 0.8895841162792735, "learning_rate": 1.5067609492836572e-06, "loss": 0.5092, "step": 4021 }, { "epoch": 3.8602976476236197, "grad_norm": 0.825488896963854, "learning_rate": 1.5043649517363835e-06, "loss": 0.5105, "step": 4022 }, { "epoch": 3.8612578012481995, "grad_norm": 0.7563483587562947, "learning_rate": 1.5019705233423177e-06, "loss": 0.517, "step": 4023 }, { "epoch": 3.8622179548727797, "grad_norm": 1.1118724278887282, "learning_rate": 1.4995776651762906e-06, "loss": 0.4072, "step": 4024 }, { "epoch": 3.8631781084973595, "grad_norm": 1.0306071874438925, "learning_rate": 1.4971863783124308e-06, "loss": 0.5127, "step": 4025 }, { "epoch": 3.86413826212194, "grad_norm": 0.8007556821579654, "learning_rate": 1.4947966638241662e-06, "loss": 0.4844, "step": 4026 }, { "epoch": 3.8650984157465196, "grad_norm": 0.7858284431034764, "learning_rate": 1.4924085227842084e-06, "loss": 0.5415, "step": 4027 }, { "epoch": 3.8660585693710994, "grad_norm": 0.6578999753793716, "learning_rate": 1.490021956264573e-06, "loss": 0.4836, "step": 4028 }, { "epoch": 3.867018722995679, "grad_norm": 0.8041491494771016, "learning_rate": 1.487636965336558e-06, "loss": 0.4633, "step": 4029 }, { "epoch": 3.867978876620259, "grad_norm": 0.8696932214180055, "learning_rate": 1.4852535510707662e-06, "loss": 0.5487, "step": 4030 }, { "epoch": 3.868939030244839, "grad_norm": 0.7141875387248724, "learning_rate": 1.4828717145370863e-06, "loss": 0.5793, "step": 4031 }, { "epoch": 3.869899183869419, "grad_norm": 0.9307610355594869, "learning_rate": 1.4804914568046957e-06, "loss": 0.4364, "step": 4032 }, { "epoch": 3.8708593374939992, "grad_norm": 0.8286171227172064, "learning_rate": 1.478112778942069e-06, "loss": 0.421, "step": 4033 }, { "epoch": 3.871819491118579, "grad_norm": 0.6368811100437496, "learning_rate": 1.4757356820169694e-06, "loss": 0.5105, "step": 4034 }, { "epoch": 3.872779644743159, "grad_norm": 0.7486523700611957, "learning_rate": 1.4733601670964493e-06, "loss": 0.4584, "step": 4035 }, { "epoch": 3.8737397983677386, "grad_norm": 1.034831253195014, "learning_rate": 1.4709862352468541e-06, "loss": 0.3794, "step": 4036 }, { "epoch": 3.874699951992319, "grad_norm": 0.8639164712719245, "learning_rate": 1.4686138875338136e-06, "loss": 0.4969, "step": 4037 }, { "epoch": 3.8756601056168987, "grad_norm": 1.1798226123992228, "learning_rate": 1.4662431250222504e-06, "loss": 0.5301, "step": 4038 }, { "epoch": 3.876620259241479, "grad_norm": 1.5611214955960346, "learning_rate": 1.4638739487763743e-06, "loss": 0.5717, "step": 4039 }, { "epoch": 3.8775804128660587, "grad_norm": 0.9158718482254138, "learning_rate": 1.4615063598596834e-06, "loss": 0.5324, "step": 4040 }, { "epoch": 3.8785405664906385, "grad_norm": 1.0159462881546517, "learning_rate": 1.4591403593349652e-06, "loss": 0.4908, "step": 4041 }, { "epoch": 3.8795007201152183, "grad_norm": 0.8188702441070949, "learning_rate": 1.4567759482642884e-06, "loss": 0.5225, "step": 4042 }, { "epoch": 3.8804608737397985, "grad_norm": 0.7480174298387952, "learning_rate": 1.4544131277090118e-06, "loss": 0.4845, "step": 4043 }, { "epoch": 3.8814210273643783, "grad_norm": 1.0563109548634233, "learning_rate": 1.4520518987297854e-06, "loss": 0.506, "step": 4044 }, { "epoch": 3.882381180988958, "grad_norm": 0.9220318408234932, "learning_rate": 1.449692262386534e-06, "loss": 0.5741, "step": 4045 }, { "epoch": 3.8833413346135384, "grad_norm": 0.7273071371308207, "learning_rate": 1.447334219738477e-06, "loss": 0.3956, "step": 4046 }, { "epoch": 3.884301488238118, "grad_norm": 1.2521962745810373, "learning_rate": 1.4449777718441106e-06, "loss": 0.55, "step": 4047 }, { "epoch": 3.885261641862698, "grad_norm": 1.027569054525299, "learning_rate": 1.4426229197612217e-06, "loss": 0.3821, "step": 4048 }, { "epoch": 3.8862217954872778, "grad_norm": 0.8098049890844544, "learning_rate": 1.44026966454688e-06, "loss": 0.3376, "step": 4049 }, { "epoch": 3.887181949111858, "grad_norm": 0.9688105832005914, "learning_rate": 1.4379180072574335e-06, "loss": 0.4287, "step": 4050 }, { "epoch": 3.888142102736438, "grad_norm": 0.9212506065046874, "learning_rate": 1.4355679489485158e-06, "loss": 0.4466, "step": 4051 }, { "epoch": 3.889102256361018, "grad_norm": 0.8382345034531979, "learning_rate": 1.433219490675045e-06, "loss": 0.4459, "step": 4052 }, { "epoch": 3.890062409985598, "grad_norm": 0.8185186215987766, "learning_rate": 1.4308726334912176e-06, "loss": 0.5471, "step": 4053 }, { "epoch": 3.8910225636101776, "grad_norm": 1.0445853626808408, "learning_rate": 1.4285273784505154e-06, "loss": 0.5757, "step": 4054 }, { "epoch": 3.8919827172347574, "grad_norm": 1.0068476914341893, "learning_rate": 1.4261837266056938e-06, "loss": 0.4723, "step": 4055 }, { "epoch": 3.8929428708593377, "grad_norm": 0.9988946498129965, "learning_rate": 1.4238416790087945e-06, "loss": 0.4636, "step": 4056 }, { "epoch": 3.8939030244839175, "grad_norm": 0.9511009253492032, "learning_rate": 1.4215012367111408e-06, "loss": 0.5572, "step": 4057 }, { "epoch": 3.8948631781084972, "grad_norm": 0.8208154700704717, "learning_rate": 1.4191624007633286e-06, "loss": 0.5047, "step": 4058 }, { "epoch": 3.8958233317330775, "grad_norm": 1.3936185477462202, "learning_rate": 1.41682517221524e-06, "loss": 0.5433, "step": 4059 }, { "epoch": 3.8967834853576573, "grad_norm": 1.612221147102768, "learning_rate": 1.4144895521160257e-06, "loss": 0.5244, "step": 4060 }, { "epoch": 3.897743638982237, "grad_norm": 0.9665811740298895, "learning_rate": 1.412155541514126e-06, "loss": 0.51, "step": 4061 }, { "epoch": 3.898703792606817, "grad_norm": 1.0852012095944785, "learning_rate": 1.4098231414572534e-06, "loss": 0.4703, "step": 4062 }, { "epoch": 3.899663946231397, "grad_norm": 0.891137152389279, "learning_rate": 1.4074923529923945e-06, "loss": 0.5287, "step": 4063 }, { "epoch": 3.900624099855977, "grad_norm": 0.8161137189996271, "learning_rate": 1.4051631771658159e-06, "loss": 0.5548, "step": 4064 }, { "epoch": 3.901584253480557, "grad_norm": 0.7022112983070327, "learning_rate": 1.402835615023061e-06, "loss": 0.5694, "step": 4065 }, { "epoch": 3.902544407105137, "grad_norm": 0.8867492636768103, "learning_rate": 1.400509667608947e-06, "loss": 0.455, "step": 4066 }, { "epoch": 3.9035045607297167, "grad_norm": 1.3021704440293786, "learning_rate": 1.398185335967568e-06, "loss": 0.5003, "step": 4067 }, { "epoch": 3.9044647143542965, "grad_norm": 1.5665114427313156, "learning_rate": 1.3958626211422887e-06, "loss": 0.4655, "step": 4068 }, { "epoch": 3.905424867978877, "grad_norm": 1.1887108859851676, "learning_rate": 1.3935415241757517e-06, "loss": 0.5419, "step": 4069 }, { "epoch": 3.9063850216034566, "grad_norm": 1.0119765240233674, "learning_rate": 1.3912220461098763e-06, "loss": 0.5449, "step": 4070 }, { "epoch": 3.9073451752280364, "grad_norm": 1.0349605000909057, "learning_rate": 1.3889041879858472e-06, "loss": 0.4879, "step": 4071 }, { "epoch": 3.9083053288526166, "grad_norm": 0.8031233014933145, "learning_rate": 1.3865879508441294e-06, "loss": 0.4133, "step": 4072 }, { "epoch": 3.9092654824771964, "grad_norm": 1.1065129095529487, "learning_rate": 1.384273335724452e-06, "loss": 0.5182, "step": 4073 }, { "epoch": 3.910225636101776, "grad_norm": 0.6943145482995982, "learning_rate": 1.3819603436658258e-06, "loss": 0.5068, "step": 4074 }, { "epoch": 3.911185789726356, "grad_norm": 1.141672546964942, "learning_rate": 1.3796489757065284e-06, "loss": 0.5065, "step": 4075 }, { "epoch": 3.9121459433509362, "grad_norm": 0.9611857820568376, "learning_rate": 1.3773392328841046e-06, "loss": 0.4467, "step": 4076 }, { "epoch": 3.913106096975516, "grad_norm": 0.8302982784968235, "learning_rate": 1.3750311162353752e-06, "loss": 0.5108, "step": 4077 }, { "epoch": 3.9140662506000963, "grad_norm": 0.856760490651798, "learning_rate": 1.3727246267964288e-06, "loss": 0.4345, "step": 4078 }, { "epoch": 3.915026404224676, "grad_norm": 0.8762754908871317, "learning_rate": 1.3704197656026242e-06, "loss": 0.5175, "step": 4079 }, { "epoch": 3.915986557849256, "grad_norm": 0.7149318512506908, "learning_rate": 1.36811653368859e-06, "loss": 0.5109, "step": 4080 }, { "epoch": 3.9169467114738357, "grad_norm": 1.035312655422887, "learning_rate": 1.3658149320882196e-06, "loss": 0.5038, "step": 4081 }, { "epoch": 3.917906865098416, "grad_norm": 0.870218905327093, "learning_rate": 1.3635149618346793e-06, "loss": 0.4618, "step": 4082 }, { "epoch": 3.9188670187229957, "grad_norm": 0.8358803660759263, "learning_rate": 1.3612166239604003e-06, "loss": 0.4671, "step": 4083 }, { "epoch": 3.9198271723475755, "grad_norm": 0.8919096105191371, "learning_rate": 1.3589199194970825e-06, "loss": 0.4867, "step": 4084 }, { "epoch": 3.9207873259721557, "grad_norm": 0.858890821958506, "learning_rate": 1.3566248494756934e-06, "loss": 0.4698, "step": 4085 }, { "epoch": 3.9217474795967355, "grad_norm": 0.9443161467946802, "learning_rate": 1.3543314149264625e-06, "loss": 0.4699, "step": 4086 }, { "epoch": 3.9227076332213153, "grad_norm": 0.7592184754767289, "learning_rate": 1.352039616878888e-06, "loss": 0.4756, "step": 4087 }, { "epoch": 3.923667786845895, "grad_norm": 0.8700350521442153, "learning_rate": 1.3497494563617375e-06, "loss": 0.4323, "step": 4088 }, { "epoch": 3.9246279404704754, "grad_norm": 0.988920977787674, "learning_rate": 1.347460934403036e-06, "loss": 0.5547, "step": 4089 }, { "epoch": 3.925588094095055, "grad_norm": 0.9435843357436853, "learning_rate": 1.3451740520300787e-06, "loss": 0.47, "step": 4090 }, { "epoch": 3.9265482477196354, "grad_norm": 0.7806248542133531, "learning_rate": 1.3428888102694187e-06, "loss": 0.5734, "step": 4091 }, { "epoch": 3.927508401344215, "grad_norm": 1.0348866547313655, "learning_rate": 1.3406052101468808e-06, "loss": 0.6038, "step": 4092 }, { "epoch": 3.928468554968795, "grad_norm": 1.050758920526436, "learning_rate": 1.338323252687549e-06, "loss": 0.4974, "step": 4093 }, { "epoch": 3.929428708593375, "grad_norm": 0.7927204434836473, "learning_rate": 1.3360429389157664e-06, "loss": 0.4849, "step": 4094 }, { "epoch": 3.930388862217955, "grad_norm": 0.7406220891742337, "learning_rate": 1.3337642698551428e-06, "loss": 0.478, "step": 4095 }, { "epoch": 3.931349015842535, "grad_norm": 0.8459990822147183, "learning_rate": 1.331487246528549e-06, "loss": 0.4387, "step": 4096 }, { "epoch": 3.9323091694671146, "grad_norm": 0.8172030753506038, "learning_rate": 1.3292118699581153e-06, "loss": 0.5039, "step": 4097 }, { "epoch": 3.933269323091695, "grad_norm": 0.8573790051241101, "learning_rate": 1.3269381411652366e-06, "loss": 0.5423, "step": 4098 }, { "epoch": 3.9342294767162747, "grad_norm": 0.9999519619658027, "learning_rate": 1.324666061170562e-06, "loss": 0.5412, "step": 4099 }, { "epoch": 3.9351896303408544, "grad_norm": 0.6505622426144994, "learning_rate": 1.3223956309940038e-06, "loss": 0.454, "step": 4100 }, { "epoch": 3.9361497839654342, "grad_norm": 0.7422355024747747, "learning_rate": 1.3201268516547382e-06, "loss": 0.4977, "step": 4101 }, { "epoch": 3.9371099375900145, "grad_norm": 1.081956192606628, "learning_rate": 1.317859724171192e-06, "loss": 0.47, "step": 4102 }, { "epoch": 3.9380700912145943, "grad_norm": 1.0748916964510475, "learning_rate": 1.3155942495610574e-06, "loss": 0.4489, "step": 4103 }, { "epoch": 3.9390302448391745, "grad_norm": 1.0783082990356552, "learning_rate": 1.3133304288412768e-06, "loss": 0.5281, "step": 4104 }, { "epoch": 3.9399903984637543, "grad_norm": 1.1452849790373907, "learning_rate": 1.3110682630280603e-06, "loss": 0.4614, "step": 4105 }, { "epoch": 3.940950552088334, "grad_norm": 1.5275412956863934, "learning_rate": 1.3088077531368686e-06, "loss": 0.4981, "step": 4106 }, { "epoch": 3.941910705712914, "grad_norm": 1.122540722178402, "learning_rate": 1.3065489001824194e-06, "loss": 0.4679, "step": 4107 }, { "epoch": 3.942870859337494, "grad_norm": 0.7046774654859396, "learning_rate": 1.3042917051786873e-06, "loss": 0.6447, "step": 4108 }, { "epoch": 3.943831012962074, "grad_norm": 0.8951767715283352, "learning_rate": 1.3020361691389043e-06, "loss": 0.5026, "step": 4109 }, { "epoch": 3.9447911665866537, "grad_norm": 0.8395093487984943, "learning_rate": 1.299782293075556e-06, "loss": 0.4285, "step": 4110 }, { "epoch": 3.945751320211234, "grad_norm": 1.0492925910278683, "learning_rate": 1.297530078000384e-06, "loss": 0.5439, "step": 4111 }, { "epoch": 3.9467114738358138, "grad_norm": 0.9554153972372763, "learning_rate": 1.295279524924382e-06, "loss": 0.485, "step": 4112 }, { "epoch": 3.9476716274603936, "grad_norm": 0.7094820519137514, "learning_rate": 1.2930306348577982e-06, "loss": 0.5709, "step": 4113 }, { "epoch": 3.9486317810849734, "grad_norm": 0.859280440936798, "learning_rate": 1.2907834088101402e-06, "loss": 0.5068, "step": 4114 }, { "epoch": 3.9495919347095536, "grad_norm": 0.6606263850800658, "learning_rate": 1.2885378477901595e-06, "loss": 0.6028, "step": 4115 }, { "epoch": 3.9505520883341334, "grad_norm": 0.9015368433821501, "learning_rate": 1.2862939528058665e-06, "loss": 0.5797, "step": 4116 }, { "epoch": 3.9515122419587136, "grad_norm": 0.8041215537749933, "learning_rate": 1.284051724864518e-06, "loss": 0.5187, "step": 4117 }, { "epoch": 3.9524723955832934, "grad_norm": 0.93058985198674, "learning_rate": 1.2818111649726306e-06, "loss": 0.4613, "step": 4118 }, { "epoch": 3.9534325492078732, "grad_norm": 1.0814574944322364, "learning_rate": 1.2795722741359672e-06, "loss": 0.4258, "step": 4119 }, { "epoch": 3.954392702832453, "grad_norm": 0.8554775553018628, "learning_rate": 1.277335053359539e-06, "loss": 0.5088, "step": 4120 }, { "epoch": 3.9553528564570333, "grad_norm": 0.9508076822048066, "learning_rate": 1.2750995036476133e-06, "loss": 0.588, "step": 4121 }, { "epoch": 3.956313010081613, "grad_norm": 0.852412891674262, "learning_rate": 1.2728656260037025e-06, "loss": 0.4111, "step": 4122 }, { "epoch": 3.957273163706193, "grad_norm": 0.9848792965180964, "learning_rate": 1.2706334214305716e-06, "loss": 0.5017, "step": 4123 }, { "epoch": 3.958233317330773, "grad_norm": 0.5562352672055572, "learning_rate": 1.2684028909302348e-06, "loss": 0.4682, "step": 4124 }, { "epoch": 3.959193470955353, "grad_norm": 0.7826241798327265, "learning_rate": 1.2661740355039499e-06, "loss": 0.5261, "step": 4125 }, { "epoch": 3.9601536245799327, "grad_norm": 0.9622092757967434, "learning_rate": 1.2639468561522284e-06, "loss": 0.5383, "step": 4126 }, { "epoch": 3.9611137782045125, "grad_norm": 1.0120717991242756, "learning_rate": 1.2617213538748268e-06, "loss": 0.5126, "step": 4127 }, { "epoch": 3.9620739318290927, "grad_norm": 0.7268848449086381, "learning_rate": 1.2594975296707495e-06, "loss": 0.5127, "step": 4128 }, { "epoch": 3.9630340854536725, "grad_norm": 0.7758650260208573, "learning_rate": 1.2572753845382484e-06, "loss": 0.5599, "step": 4129 }, { "epoch": 3.9639942390782528, "grad_norm": 1.2247830001139546, "learning_rate": 1.255054919474819e-06, "loss": 0.5018, "step": 4130 }, { "epoch": 3.9649543927028326, "grad_norm": 1.7701762453635013, "learning_rate": 1.252836135477203e-06, "loss": 0.55, "step": 4131 }, { "epoch": 3.9659145463274124, "grad_norm": 1.2006644071689445, "learning_rate": 1.2506190335413943e-06, "loss": 0.5234, "step": 4132 }, { "epoch": 3.966874699951992, "grad_norm": 0.895697532824246, "learning_rate": 1.2484036146626205e-06, "loss": 0.5031, "step": 4133 }, { "epoch": 3.9678348535765724, "grad_norm": 1.02791443398622, "learning_rate": 1.2461898798353639e-06, "loss": 0.4561, "step": 4134 }, { "epoch": 3.968795007201152, "grad_norm": 1.1039106450716663, "learning_rate": 1.2439778300533412e-06, "loss": 0.4346, "step": 4135 }, { "epoch": 3.969755160825732, "grad_norm": 0.7278713489944657, "learning_rate": 1.241767466309522e-06, "loss": 0.4549, "step": 4136 }, { "epoch": 3.9707153144503122, "grad_norm": 0.7961014526727447, "learning_rate": 1.2395587895961158e-06, "loss": 0.5452, "step": 4137 }, { "epoch": 3.971675468074892, "grad_norm": 0.9842629444059643, "learning_rate": 1.2373518009045705e-06, "loss": 0.4689, "step": 4138 }, { "epoch": 3.972635621699472, "grad_norm": 0.7549903610324099, "learning_rate": 1.2351465012255808e-06, "loss": 0.522, "step": 4139 }, { "epoch": 3.9735957753240516, "grad_norm": 0.7206528631856418, "learning_rate": 1.232942891549083e-06, "loss": 0.5129, "step": 4140 }, { "epoch": 3.974555928948632, "grad_norm": 0.6711402159269111, "learning_rate": 1.2307409728642539e-06, "loss": 0.5382, "step": 4141 }, { "epoch": 3.9755160825732117, "grad_norm": 0.7086619999114167, "learning_rate": 1.2285407461595122e-06, "loss": 0.435, "step": 4142 }, { "epoch": 3.976476236197792, "grad_norm": 0.8619286498269527, "learning_rate": 1.2263422124225132e-06, "loss": 0.4451, "step": 4143 }, { "epoch": 3.9774363898223717, "grad_norm": 0.7957636176388156, "learning_rate": 1.2241453726401559e-06, "loss": 0.4936, "step": 4144 }, { "epoch": 3.9783965434469515, "grad_norm": 0.8555345358019707, "learning_rate": 1.2219502277985817e-06, "loss": 0.4189, "step": 4145 }, { "epoch": 3.9793566970715313, "grad_norm": 1.154385960051515, "learning_rate": 1.2197567788831638e-06, "loss": 0.4271, "step": 4146 }, { "epoch": 3.9803168506961115, "grad_norm": 0.8744058237546322, "learning_rate": 1.2175650268785205e-06, "loss": 0.5143, "step": 4147 }, { "epoch": 3.9812770043206913, "grad_norm": 0.8285986056799638, "learning_rate": 1.215374972768502e-06, "loss": 0.5187, "step": 4148 }, { "epoch": 3.982237157945271, "grad_norm": 0.8121854138621586, "learning_rate": 1.213186617536204e-06, "loss": 0.4977, "step": 4149 }, { "epoch": 3.9831973115698514, "grad_norm": 0.9148852135295331, "learning_rate": 1.2109999621639561e-06, "loss": 0.4901, "step": 4150 }, { "epoch": 3.984157465194431, "grad_norm": 0.9679094380427661, "learning_rate": 1.2088150076333222e-06, "loss": 0.5213, "step": 4151 }, { "epoch": 3.985117618819011, "grad_norm": 0.7385145910058086, "learning_rate": 1.2066317549251055e-06, "loss": 0.579, "step": 4152 }, { "epoch": 3.9860777724435907, "grad_norm": 0.7871950476823385, "learning_rate": 1.2044502050193451e-06, "loss": 0.3923, "step": 4153 }, { "epoch": 3.987037926068171, "grad_norm": 0.6268400000410335, "learning_rate": 1.2022703588953155e-06, "loss": 0.5057, "step": 4154 }, { "epoch": 3.9879980796927508, "grad_norm": 0.911223916307204, "learning_rate": 1.2000922175315278e-06, "loss": 0.5164, "step": 4155 }, { "epoch": 3.988958233317331, "grad_norm": 0.9222792521774887, "learning_rate": 1.197915781905723e-06, "loss": 0.4588, "step": 4156 }, { "epoch": 3.989918386941911, "grad_norm": 0.7379850553966277, "learning_rate": 1.1957410529948803e-06, "loss": 0.6044, "step": 4157 }, { "epoch": 3.9908785405664906, "grad_norm": 1.0593049798583, "learning_rate": 1.1935680317752173e-06, "loss": 0.5584, "step": 4158 }, { "epoch": 3.9918386941910704, "grad_norm": 0.995365545114401, "learning_rate": 1.1913967192221742e-06, "loss": 0.4624, "step": 4159 }, { "epoch": 3.9927988478156506, "grad_norm": 0.9543989290067514, "learning_rate": 1.189227116310434e-06, "loss": 0.6056, "step": 4160 }, { "epoch": 3.9937590014402304, "grad_norm": 1.0016748109830997, "learning_rate": 1.1870592240139033e-06, "loss": 0.4294, "step": 4161 }, { "epoch": 3.9947191550648102, "grad_norm": 0.7164015325123054, "learning_rate": 1.18489304330573e-06, "loss": 0.4671, "step": 4162 }, { "epoch": 3.9956793086893905, "grad_norm": 0.8018011662715128, "learning_rate": 1.18272857515829e-06, "loss": 0.5035, "step": 4163 }, { "epoch": 3.9966394623139703, "grad_norm": 0.6186196579761598, "learning_rate": 1.1805658205431868e-06, "loss": 0.3945, "step": 4164 }, { "epoch": 3.99759961593855, "grad_norm": 0.9492012012840304, "learning_rate": 1.1784047804312598e-06, "loss": 0.5269, "step": 4165 }, { "epoch": 3.99855976956313, "grad_norm": 0.7252988615131755, "learning_rate": 1.1762454557925756e-06, "loss": 0.5537, "step": 4166 }, { "epoch": 3.99951992318771, "grad_norm": 0.8165057844562047, "learning_rate": 1.174087847596433e-06, "loss": 0.4595, "step": 4167 }, { "epoch": 4.0, "grad_norm": 2.006203212383796, "learning_rate": 1.1719319568113602e-06, "loss": 0.5165, "step": 4168 }, { "epoch": 4.00096015362458, "grad_norm": 0.9232993754677573, "learning_rate": 1.1697777844051105e-06, "loss": 0.4804, "step": 4169 }, { "epoch": 4.00192030724916, "grad_norm": 1.309062544285563, "learning_rate": 1.1676253313446706e-06, "loss": 0.5118, "step": 4170 }, { "epoch": 4.002880460873739, "grad_norm": 0.8911129430624856, "learning_rate": 1.165474598596254e-06, "loss": 0.5446, "step": 4171 }, { "epoch": 4.00384061449832, "grad_norm": 1.5071906731939637, "learning_rate": 1.1633255871253013e-06, "loss": 0.4819, "step": 4172 }, { "epoch": 4.0048007681229, "grad_norm": 0.8426639925842434, "learning_rate": 1.161178297896482e-06, "loss": 0.459, "step": 4173 }, { "epoch": 4.00576092174748, "grad_norm": 0.9040492897742162, "learning_rate": 1.1590327318736887e-06, "loss": 0.3716, "step": 4174 }, { "epoch": 4.0067210753720595, "grad_norm": 0.999123889304821, "learning_rate": 1.1568888900200432e-06, "loss": 0.5634, "step": 4175 }, { "epoch": 4.007681228996639, "grad_norm": 1.0403815330565804, "learning_rate": 1.1547467732978967e-06, "loss": 0.504, "step": 4176 }, { "epoch": 4.008641382621219, "grad_norm": 0.7413592204780737, "learning_rate": 1.152606382668819e-06, "loss": 0.4162, "step": 4177 }, { "epoch": 4.0096015362458, "grad_norm": 0.7058536671839979, "learning_rate": 1.1504677190936115e-06, "loss": 0.4904, "step": 4178 }, { "epoch": 4.0105616898703795, "grad_norm": 1.6780450367705253, "learning_rate": 1.148330783532292e-06, "loss": 0.3399, "step": 4179 }, { "epoch": 4.011521843494959, "grad_norm": 0.8803205975002225, "learning_rate": 1.1461955769441141e-06, "loss": 0.4519, "step": 4180 }, { "epoch": 4.012481997119539, "grad_norm": 0.7561608043920058, "learning_rate": 1.144062100287548e-06, "loss": 0.4872, "step": 4181 }, { "epoch": 4.013442150744119, "grad_norm": 0.759317518827711, "learning_rate": 1.1419303545202854e-06, "loss": 0.4491, "step": 4182 }, { "epoch": 4.014402304368699, "grad_norm": 0.9359576212224298, "learning_rate": 1.1398003405992468e-06, "loss": 0.3685, "step": 4183 }, { "epoch": 4.0153624579932785, "grad_norm": 0.876582614814508, "learning_rate": 1.1376720594805719e-06, "loss": 0.5576, "step": 4184 }, { "epoch": 4.016322611617859, "grad_norm": 0.8691972137454735, "learning_rate": 1.1355455121196234e-06, "loss": 0.4465, "step": 4185 }, { "epoch": 4.017282765242439, "grad_norm": 0.657038932166332, "learning_rate": 1.1334206994709874e-06, "loss": 0.4549, "step": 4186 }, { "epoch": 4.018242918867019, "grad_norm": 0.8348842838083208, "learning_rate": 1.1312976224884665e-06, "loss": 0.536, "step": 4187 }, { "epoch": 4.019203072491599, "grad_norm": 1.0794217084312434, "learning_rate": 1.1291762821250867e-06, "loss": 0.3896, "step": 4188 }, { "epoch": 4.020163226116178, "grad_norm": 1.2379288638798194, "learning_rate": 1.1270566793331006e-06, "loss": 0.4761, "step": 4189 }, { "epoch": 4.021123379740758, "grad_norm": 1.1411170993648845, "learning_rate": 1.1249388150639702e-06, "loss": 0.4872, "step": 4190 }, { "epoch": 4.022083533365339, "grad_norm": 1.3535637148712907, "learning_rate": 1.1228226902683847e-06, "loss": 0.4712, "step": 4191 }, { "epoch": 4.023043686989919, "grad_norm": 1.203792912775379, "learning_rate": 1.1207083058962465e-06, "loss": 0.4574, "step": 4192 }, { "epoch": 4.0240038406144985, "grad_norm": 0.7066559012248774, "learning_rate": 1.1185956628966832e-06, "loss": 0.5136, "step": 4193 }, { "epoch": 4.024963994239078, "grad_norm": 0.8830135864243964, "learning_rate": 1.1164847622180392e-06, "loss": 0.3721, "step": 4194 }, { "epoch": 4.025924147863658, "grad_norm": 1.3022689821471134, "learning_rate": 1.1143756048078707e-06, "loss": 0.5537, "step": 4195 }, { "epoch": 4.026884301488238, "grad_norm": 0.9035717689946475, "learning_rate": 1.1122681916129586e-06, "loss": 0.5273, "step": 4196 }, { "epoch": 4.027844455112818, "grad_norm": 0.9737690135996574, "learning_rate": 1.1101625235792978e-06, "loss": 0.4521, "step": 4197 }, { "epoch": 4.028804608737398, "grad_norm": 0.6873497194587332, "learning_rate": 1.1080586016521e-06, "loss": 0.4342, "step": 4198 }, { "epoch": 4.029764762361978, "grad_norm": 0.9678677239715283, "learning_rate": 1.1059564267757945e-06, "loss": 0.5662, "step": 4199 }, { "epoch": 4.030724915986558, "grad_norm": 0.821182842498426, "learning_rate": 1.1038559998940228e-06, "loss": 0.459, "step": 4200 }, { "epoch": 4.031685069611138, "grad_norm": 0.9782888809451943, "learning_rate": 1.1017573219496453e-06, "loss": 0.4675, "step": 4201 }, { "epoch": 4.0326452232357175, "grad_norm": 0.8535867101654473, "learning_rate": 1.0996603938847361e-06, "loss": 0.5038, "step": 4202 }, { "epoch": 4.033605376860297, "grad_norm": 1.0568833653513559, "learning_rate": 1.0975652166405836e-06, "loss": 0.4434, "step": 4203 }, { "epoch": 4.034565530484878, "grad_norm": 0.7715758688488206, "learning_rate": 1.0954717911576923e-06, "loss": 0.4449, "step": 4204 }, { "epoch": 4.035525684109458, "grad_norm": 0.7402360059833049, "learning_rate": 1.0933801183757742e-06, "loss": 0.5037, "step": 4205 }, { "epoch": 4.036485837734038, "grad_norm": 1.0106750096942454, "learning_rate": 1.0912901992337626e-06, "loss": 0.3523, "step": 4206 }, { "epoch": 4.037445991358617, "grad_norm": 0.8455281224150393, "learning_rate": 1.0892020346698e-06, "loss": 0.5451, "step": 4207 }, { "epoch": 4.038406144983197, "grad_norm": 1.011750261391204, "learning_rate": 1.087115625621239e-06, "loss": 0.3944, "step": 4208 }, { "epoch": 4.039366298607777, "grad_norm": 0.8623085356964846, "learning_rate": 1.0850309730246472e-06, "loss": 0.437, "step": 4209 }, { "epoch": 4.040326452232357, "grad_norm": 0.7388787357974598, "learning_rate": 1.0829480778158025e-06, "loss": 0.4371, "step": 4210 }, { "epoch": 4.041286605856937, "grad_norm": 0.787699224272905, "learning_rate": 1.0808669409296951e-06, "loss": 0.3974, "step": 4211 }, { "epoch": 4.042246759481517, "grad_norm": 0.875745087481987, "learning_rate": 1.0787875633005256e-06, "loss": 0.3688, "step": 4212 }, { "epoch": 4.043206913106097, "grad_norm": 0.8177165059852899, "learning_rate": 1.0767099458617025e-06, "loss": 0.4584, "step": 4213 }, { "epoch": 4.044167066730677, "grad_norm": 0.6815851799245208, "learning_rate": 1.0746340895458475e-06, "loss": 0.4119, "step": 4214 }, { "epoch": 4.045127220355257, "grad_norm": 0.7710536305468243, "learning_rate": 1.0725599952847894e-06, "loss": 0.556, "step": 4215 }, { "epoch": 4.046087373979836, "grad_norm": 0.7812271348714714, "learning_rate": 1.070487664009567e-06, "loss": 0.496, "step": 4216 }, { "epoch": 4.047047527604417, "grad_norm": 0.96742750064349, "learning_rate": 1.0684170966504298e-06, "loss": 0.4929, "step": 4217 }, { "epoch": 4.048007681228997, "grad_norm": 1.018297425610479, "learning_rate": 1.0663482941368303e-06, "loss": 0.4769, "step": 4218 }, { "epoch": 4.048967834853577, "grad_norm": 0.9837573996217966, "learning_rate": 1.064281257397432e-06, "loss": 0.5019, "step": 4219 }, { "epoch": 4.0499279884781565, "grad_norm": 0.6349464558277421, "learning_rate": 1.062215987360109e-06, "loss": 0.4557, "step": 4220 }, { "epoch": 4.050888142102736, "grad_norm": 1.1816781337235738, "learning_rate": 1.0601524849519357e-06, "loss": 0.3442, "step": 4221 }, { "epoch": 4.051848295727316, "grad_norm": 1.3538521856325505, "learning_rate": 1.0580907510991982e-06, "loss": 0.4541, "step": 4222 }, { "epoch": 4.052808449351896, "grad_norm": 1.1309505990189086, "learning_rate": 1.0560307867273838e-06, "loss": 0.5259, "step": 4223 }, { "epoch": 4.053768602976477, "grad_norm": 0.9534996051900567, "learning_rate": 1.0539725927611915e-06, "loss": 0.4411, "step": 4224 }, { "epoch": 4.054728756601056, "grad_norm": 1.6643310426438258, "learning_rate": 1.0519161701245234e-06, "loss": 0.426, "step": 4225 }, { "epoch": 4.055688910225636, "grad_norm": 1.1199428230627557, "learning_rate": 1.049861519740482e-06, "loss": 0.4233, "step": 4226 }, { "epoch": 4.056649063850216, "grad_norm": 0.7670945534089602, "learning_rate": 1.04780864253138e-06, "loss": 0.505, "step": 4227 }, { "epoch": 4.057609217474796, "grad_norm": 0.9551098363234146, "learning_rate": 1.0457575394187314e-06, "loss": 0.4427, "step": 4228 }, { "epoch": 4.0585693710993755, "grad_norm": 0.7204972950584314, "learning_rate": 1.043708211323255e-06, "loss": 0.5325, "step": 4229 }, { "epoch": 4.059529524723956, "grad_norm": 0.8474175401612934, "learning_rate": 1.0416606591648737e-06, "loss": 0.4637, "step": 4230 }, { "epoch": 4.060489678348536, "grad_norm": 1.160052085298909, "learning_rate": 1.039614883862709e-06, "loss": 0.5098, "step": 4231 }, { "epoch": 4.061449831973116, "grad_norm": 0.7762562754639191, "learning_rate": 1.037570886335087e-06, "loss": 0.4618, "step": 4232 }, { "epoch": 4.062409985597696, "grad_norm": 0.6649732945585918, "learning_rate": 1.0355286674995413e-06, "loss": 0.4115, "step": 4233 }, { "epoch": 4.063370139222275, "grad_norm": 1.2426902773819097, "learning_rate": 1.0334882282727971e-06, "loss": 0.4577, "step": 4234 }, { "epoch": 4.064330292846855, "grad_norm": 0.7186592983937328, "learning_rate": 1.0314495695707887e-06, "loss": 0.5381, "step": 4235 }, { "epoch": 4.065290446471435, "grad_norm": 1.2462910721738365, "learning_rate": 1.029412692308645e-06, "loss": 0.5068, "step": 4236 }, { "epoch": 4.066250600096016, "grad_norm": 0.8297096965431681, "learning_rate": 1.0273775974007017e-06, "loss": 0.4837, "step": 4237 }, { "epoch": 4.0672107537205955, "grad_norm": 1.1908254191846446, "learning_rate": 1.0253442857604911e-06, "loss": 0.5343, "step": 4238 }, { "epoch": 4.068170907345175, "grad_norm": 0.7935440184542987, "learning_rate": 1.0233127583007424e-06, "loss": 0.5347, "step": 4239 }, { "epoch": 4.069131060969755, "grad_norm": 0.8804868963482719, "learning_rate": 1.0212830159333886e-06, "loss": 0.4918, "step": 4240 }, { "epoch": 4.070091214594335, "grad_norm": 0.7804713744375774, "learning_rate": 1.0192550595695584e-06, "loss": 0.491, "step": 4241 }, { "epoch": 4.071051368218915, "grad_norm": 0.7378922957845447, "learning_rate": 1.0172288901195805e-06, "loss": 0.4552, "step": 4242 }, { "epoch": 4.072011521843495, "grad_norm": 0.810537269703588, "learning_rate": 1.0152045084929813e-06, "loss": 0.4816, "step": 4243 }, { "epoch": 4.072971675468075, "grad_norm": 1.2489976183766112, "learning_rate": 1.0131819155984818e-06, "loss": 0.4941, "step": 4244 }, { "epoch": 4.073931829092655, "grad_norm": 0.9218123534207826, "learning_rate": 1.0111611123440041e-06, "loss": 0.5631, "step": 4245 }, { "epoch": 4.074891982717235, "grad_norm": 0.8941219148079584, "learning_rate": 1.0091420996366646e-06, "loss": 0.3759, "step": 4246 }, { "epoch": 4.0758521363418145, "grad_norm": 0.7890333963085507, "learning_rate": 1.0071248783827764e-06, "loss": 0.5941, "step": 4247 }, { "epoch": 4.076812289966394, "grad_norm": 1.4383149769006918, "learning_rate": 1.0051094494878505e-06, "loss": 0.4381, "step": 4248 }, { "epoch": 4.077772443590974, "grad_norm": 0.9127443606592621, "learning_rate": 1.0030958138565883e-06, "loss": 0.4658, "step": 4249 }, { "epoch": 4.078732597215555, "grad_norm": 0.9387852418364859, "learning_rate": 1.001083972392889e-06, "loss": 0.4933, "step": 4250 }, { "epoch": 4.079692750840135, "grad_norm": 0.9962711226005692, "learning_rate": 9.990739259998517e-07, "loss": 0.4482, "step": 4251 }, { "epoch": 4.080652904464714, "grad_norm": 0.8918590473864576, "learning_rate": 9.970656755797598e-07, "loss": 0.4903, "step": 4252 }, { "epoch": 4.081613058089294, "grad_norm": 0.8584939670062984, "learning_rate": 9.950592220340972e-07, "loss": 0.4609, "step": 4253 }, { "epoch": 4.082573211713874, "grad_norm": 1.3259663424855042, "learning_rate": 9.930545662635393e-07, "loss": 0.4451, "step": 4254 }, { "epoch": 4.083533365338454, "grad_norm": 1.1632417306592062, "learning_rate": 9.91051709167955e-07, "loss": 0.4107, "step": 4255 }, { "epoch": 4.0844935189630345, "grad_norm": 0.7551611550458784, "learning_rate": 9.890506516464066e-07, "loss": 0.5007, "step": 4256 }, { "epoch": 4.085453672587614, "grad_norm": 0.8248909199967531, "learning_rate": 9.870513945971444e-07, "loss": 0.5426, "step": 4257 }, { "epoch": 4.086413826212194, "grad_norm": 1.1963998958461926, "learning_rate": 9.85053938917615e-07, "loss": 0.4853, "step": 4258 }, { "epoch": 4.087373979836774, "grad_norm": 0.8207944093770068, "learning_rate": 9.83058285504455e-07, "loss": 0.5262, "step": 4259 }, { "epoch": 4.088334133461354, "grad_norm": 0.767065256823657, "learning_rate": 9.81064435253492e-07, "loss": 0.4484, "step": 4260 }, { "epoch": 4.0892942870859335, "grad_norm": 1.1419590636214552, "learning_rate": 9.790723890597452e-07, "loss": 0.4151, "step": 4261 }, { "epoch": 4.090254440710513, "grad_norm": 1.2655560335198281, "learning_rate": 9.770821478174207e-07, "loss": 0.3541, "step": 4262 }, { "epoch": 4.091214594335094, "grad_norm": 0.7817268026665667, "learning_rate": 9.750937124199162e-07, "loss": 0.4825, "step": 4263 }, { "epoch": 4.092174747959674, "grad_norm": 0.9589389317637074, "learning_rate": 9.731070837598227e-07, "loss": 0.4837, "step": 4264 }, { "epoch": 4.0931349015842535, "grad_norm": 1.0105823072366205, "learning_rate": 9.71122262728913e-07, "loss": 0.4881, "step": 4265 }, { "epoch": 4.094095055208833, "grad_norm": 1.103154389738956, "learning_rate": 9.69139250218154e-07, "loss": 0.4849, "step": 4266 }, { "epoch": 4.095055208833413, "grad_norm": 0.9256765577007436, "learning_rate": 9.67158047117696e-07, "loss": 0.4067, "step": 4267 }, { "epoch": 4.096015362457993, "grad_norm": 0.928073100238676, "learning_rate": 9.651786543168834e-07, "loss": 0.4809, "step": 4268 }, { "epoch": 4.096975516082574, "grad_norm": 1.0548292955861684, "learning_rate": 9.63201072704244e-07, "loss": 0.5011, "step": 4269 }, { "epoch": 4.097935669707153, "grad_norm": 0.8445129017075373, "learning_rate": 9.612253031674906e-07, "loss": 0.5375, "step": 4270 }, { "epoch": 4.098895823331733, "grad_norm": 1.0346308213725393, "learning_rate": 9.592513465935272e-07, "loss": 0.4459, "step": 4271 }, { "epoch": 4.099855976956313, "grad_norm": 0.8251019919848175, "learning_rate": 9.572792038684404e-07, "loss": 0.3945, "step": 4272 }, { "epoch": 4.100816130580893, "grad_norm": 0.686060577985539, "learning_rate": 9.553088758775048e-07, "loss": 0.5028, "step": 4273 }, { "epoch": 4.101776284205473, "grad_norm": 0.6575411045764168, "learning_rate": 9.533403635051813e-07, "loss": 0.4462, "step": 4274 }, { "epoch": 4.102736437830052, "grad_norm": 0.849729125995272, "learning_rate": 9.513736676351104e-07, "loss": 0.5098, "step": 4275 }, { "epoch": 4.103696591454633, "grad_norm": 0.9471291102428688, "learning_rate": 9.494087891501213e-07, "loss": 0.4193, "step": 4276 }, { "epoch": 4.104656745079213, "grad_norm": 0.7796868958567337, "learning_rate": 9.474457289322314e-07, "loss": 0.5198, "step": 4277 }, { "epoch": 4.105616898703793, "grad_norm": 0.7785297596455943, "learning_rate": 9.454844878626329e-07, "loss": 0.4298, "step": 4278 }, { "epoch": 4.1065770523283724, "grad_norm": 0.6673557334895199, "learning_rate": 9.435250668217088e-07, "loss": 0.4636, "step": 4279 }, { "epoch": 4.107537205952952, "grad_norm": 0.9556767855679679, "learning_rate": 9.415674666890178e-07, "loss": 0.4029, "step": 4280 }, { "epoch": 4.108497359577532, "grad_norm": 0.8164471885718875, "learning_rate": 9.396116883433104e-07, "loss": 0.5028, "step": 4281 }, { "epoch": 4.109457513202113, "grad_norm": 1.0255218944873699, "learning_rate": 9.376577326625147e-07, "loss": 0.4349, "step": 4282 }, { "epoch": 4.1104176668266925, "grad_norm": 1.0437147555377209, "learning_rate": 9.35705600523737e-07, "loss": 0.582, "step": 4283 }, { "epoch": 4.111377820451272, "grad_norm": 1.1061768120579432, "learning_rate": 9.337552928032706e-07, "loss": 0.3685, "step": 4284 }, { "epoch": 4.112337974075852, "grad_norm": 0.7534174678734537, "learning_rate": 9.318068103765871e-07, "loss": 0.5692, "step": 4285 }, { "epoch": 4.113298127700432, "grad_norm": 0.76568001787948, "learning_rate": 9.2986015411834e-07, "loss": 0.4779, "step": 4286 }, { "epoch": 4.114258281325012, "grad_norm": 0.7642467774726053, "learning_rate": 9.279153249023637e-07, "loss": 0.5185, "step": 4287 }, { "epoch": 4.1152184349495915, "grad_norm": 0.6904376159531015, "learning_rate": 9.259723236016688e-07, "loss": 0.5102, "step": 4288 }, { "epoch": 4.116178588574172, "grad_norm": 0.7263486023540775, "learning_rate": 9.240311510884487e-07, "loss": 0.5016, "step": 4289 }, { "epoch": 4.117138742198752, "grad_norm": 1.3379132042060935, "learning_rate": 9.220918082340752e-07, "loss": 0.4667, "step": 4290 }, { "epoch": 4.118098895823332, "grad_norm": 0.7298054638671881, "learning_rate": 9.201542959090987e-07, "loss": 0.5069, "step": 4291 }, { "epoch": 4.119059049447912, "grad_norm": 0.7957886478132007, "learning_rate": 9.182186149832484e-07, "loss": 0.4362, "step": 4292 }, { "epoch": 4.120019203072491, "grad_norm": 1.1443551562722623, "learning_rate": 9.162847663254292e-07, "loss": 0.4151, "step": 4293 }, { "epoch": 4.120979356697071, "grad_norm": 0.7715742799774088, "learning_rate": 9.143527508037243e-07, "loss": 0.4366, "step": 4294 }, { "epoch": 4.121939510321652, "grad_norm": 0.9534046820432989, "learning_rate": 9.124225692853995e-07, "loss": 0.3968, "step": 4295 }, { "epoch": 4.122899663946232, "grad_norm": 0.8032210086662163, "learning_rate": 9.104942226368879e-07, "loss": 0.4136, "step": 4296 }, { "epoch": 4.123859817570811, "grad_norm": 0.7600566923324985, "learning_rate": 9.085677117238068e-07, "loss": 0.5399, "step": 4297 }, { "epoch": 4.124819971195391, "grad_norm": 0.7247530315429304, "learning_rate": 9.066430374109419e-07, "loss": 0.3883, "step": 4298 }, { "epoch": 4.125780124819971, "grad_norm": 0.6637632798918436, "learning_rate": 9.047202005622641e-07, "loss": 0.3968, "step": 4299 }, { "epoch": 4.126740278444551, "grad_norm": 0.7388863465373643, "learning_rate": 9.027992020409127e-07, "loss": 0.4531, "step": 4300 }, { "epoch": 4.1277004320691315, "grad_norm": 0.9866891389797265, "learning_rate": 9.008800427092018e-07, "loss": 0.5062, "step": 4301 }, { "epoch": 4.128660585693711, "grad_norm": 0.8004344548077691, "learning_rate": 8.989627234286225e-07, "loss": 0.4645, "step": 4302 }, { "epoch": 4.129620739318291, "grad_norm": 0.7178838133534993, "learning_rate": 8.970472450598394e-07, "loss": 0.4662, "step": 4303 }, { "epoch": 4.130580892942871, "grad_norm": 0.8726704106444644, "learning_rate": 8.951336084626894e-07, "loss": 0.5117, "step": 4304 }, { "epoch": 4.131541046567451, "grad_norm": 0.9446234952960579, "learning_rate": 8.932218144961858e-07, "loss": 0.4367, "step": 4305 }, { "epoch": 4.1325012001920305, "grad_norm": 1.0896945689541557, "learning_rate": 8.913118640185086e-07, "loss": 0.4179, "step": 4306 }, { "epoch": 4.13346135381661, "grad_norm": 0.9702203407853551, "learning_rate": 8.894037578870152e-07, "loss": 0.5006, "step": 4307 }, { "epoch": 4.134421507441191, "grad_norm": 0.7628611243252998, "learning_rate": 8.874974969582379e-07, "loss": 0.4281, "step": 4308 }, { "epoch": 4.135381661065771, "grad_norm": 0.7670203397440188, "learning_rate": 8.855930820878727e-07, "loss": 0.5268, "step": 4309 }, { "epoch": 4.136341814690351, "grad_norm": 0.8494655822526511, "learning_rate": 8.836905141307933e-07, "loss": 0.4756, "step": 4310 }, { "epoch": 4.13730196831493, "grad_norm": 1.257534090874012, "learning_rate": 8.817897939410391e-07, "loss": 0.4011, "step": 4311 }, { "epoch": 4.13826212193951, "grad_norm": 1.073808099350545, "learning_rate": 8.79890922371826e-07, "loss": 0.4692, "step": 4312 }, { "epoch": 4.13922227556409, "grad_norm": 0.7781977614741021, "learning_rate": 8.779939002755378e-07, "loss": 0.51, "step": 4313 }, { "epoch": 4.14018242918867, "grad_norm": 0.8590933254580378, "learning_rate": 8.760987285037248e-07, "loss": 0.4641, "step": 4314 }, { "epoch": 4.14114258281325, "grad_norm": 1.2513012700586943, "learning_rate": 8.742054079071105e-07, "loss": 0.4476, "step": 4315 }, { "epoch": 4.14210273643783, "grad_norm": 0.854846406341309, "learning_rate": 8.723139393355861e-07, "loss": 0.5543, "step": 4316 }, { "epoch": 4.14306289006241, "grad_norm": 0.6787497892983915, "learning_rate": 8.704243236382115e-07, "loss": 0.4223, "step": 4317 }, { "epoch": 4.14402304368699, "grad_norm": 1.2289311806440029, "learning_rate": 8.685365616632175e-07, "loss": 0.5156, "step": 4318 }, { "epoch": 4.14498319731157, "grad_norm": 1.3705252779099841, "learning_rate": 8.666506542579961e-07, "loss": 0.4065, "step": 4319 }, { "epoch": 4.145943350936149, "grad_norm": 0.7799625176226243, "learning_rate": 8.64766602269112e-07, "loss": 0.4687, "step": 4320 }, { "epoch": 4.14690350456073, "grad_norm": 1.1488904906165036, "learning_rate": 8.628844065422992e-07, "loss": 0.4301, "step": 4321 }, { "epoch": 4.14786365818531, "grad_norm": 0.7218834848850836, "learning_rate": 8.610040679224518e-07, "loss": 0.4229, "step": 4322 }, { "epoch": 4.14882381180989, "grad_norm": 0.9549818360899203, "learning_rate": 8.591255872536364e-07, "loss": 0.4124, "step": 4323 }, { "epoch": 4.1497839654344695, "grad_norm": 1.2658571300578119, "learning_rate": 8.572489653790789e-07, "loss": 0.4554, "step": 4324 }, { "epoch": 4.150744119059049, "grad_norm": 0.8725463311352225, "learning_rate": 8.553742031411783e-07, "loss": 0.4417, "step": 4325 }, { "epoch": 4.151704272683629, "grad_norm": 1.5114172776585166, "learning_rate": 8.535013013814952e-07, "loss": 0.5115, "step": 4326 }, { "epoch": 4.15266442630821, "grad_norm": 0.798969466496258, "learning_rate": 8.51630260940754e-07, "loss": 0.5471, "step": 4327 }, { "epoch": 4.1536245799327896, "grad_norm": 0.7271505526868637, "learning_rate": 8.497610826588443e-07, "loss": 0.4346, "step": 4328 }, { "epoch": 4.154584733557369, "grad_norm": 0.8092580496842909, "learning_rate": 8.478937673748211e-07, "loss": 0.5559, "step": 4329 }, { "epoch": 4.155544887181949, "grad_norm": 0.8055906964566754, "learning_rate": 8.460283159269023e-07, "loss": 0.468, "step": 4330 }, { "epoch": 4.156505040806529, "grad_norm": 0.8385596775922433, "learning_rate": 8.441647291524701e-07, "loss": 0.6053, "step": 4331 }, { "epoch": 4.157465194431109, "grad_norm": 0.9902963904701965, "learning_rate": 8.423030078880667e-07, "loss": 0.4254, "step": 4332 }, { "epoch": 4.1584253480556885, "grad_norm": 1.4070908993644713, "learning_rate": 8.404431529693996e-07, "loss": 0.4554, "step": 4333 }, { "epoch": 4.159385501680269, "grad_norm": 1.4417172968276892, "learning_rate": 8.385851652313376e-07, "loss": 0.4892, "step": 4334 }, { "epoch": 4.160345655304849, "grad_norm": 0.8169784594883777, "learning_rate": 8.367290455079125e-07, "loss": 0.4326, "step": 4335 }, { "epoch": 4.161305808929429, "grad_norm": 1.2789568715025366, "learning_rate": 8.348747946323165e-07, "loss": 0.4066, "step": 4336 }, { "epoch": 4.162265962554009, "grad_norm": 0.9996486468039553, "learning_rate": 8.330224134369014e-07, "loss": 0.5158, "step": 4337 }, { "epoch": 4.163226116178588, "grad_norm": 0.9807417828038394, "learning_rate": 8.311719027531817e-07, "loss": 0.5177, "step": 4338 }, { "epoch": 4.164186269803168, "grad_norm": 0.7042764008749778, "learning_rate": 8.293232634118342e-07, "loss": 0.4414, "step": 4339 }, { "epoch": 4.165146423427748, "grad_norm": 0.8946210777862171, "learning_rate": 8.274764962426912e-07, "loss": 0.5495, "step": 4340 }, { "epoch": 4.166106577052329, "grad_norm": 0.8794279064161873, "learning_rate": 8.256316020747479e-07, "loss": 0.4851, "step": 4341 }, { "epoch": 4.1670667306769085, "grad_norm": 1.0446940100869235, "learning_rate": 8.237885817361552e-07, "loss": 0.4504, "step": 4342 }, { "epoch": 4.168026884301488, "grad_norm": 1.4443362856735134, "learning_rate": 8.219474360542273e-07, "loss": 0.4291, "step": 4343 }, { "epoch": 4.168987037926068, "grad_norm": 0.8839421101877749, "learning_rate": 8.201081658554361e-07, "loss": 0.4049, "step": 4344 }, { "epoch": 4.169947191550648, "grad_norm": 0.9986288866243261, "learning_rate": 8.182707719654076e-07, "loss": 0.4337, "step": 4345 }, { "epoch": 4.170907345175228, "grad_norm": 0.9818769631805125, "learning_rate": 8.164352552089288e-07, "loss": 0.4892, "step": 4346 }, { "epoch": 4.171867498799808, "grad_norm": 1.121112598744746, "learning_rate": 8.146016164099446e-07, "loss": 0.4992, "step": 4347 }, { "epoch": 4.172827652424388, "grad_norm": 0.9552138858157704, "learning_rate": 8.127698563915554e-07, "loss": 0.4109, "step": 4348 }, { "epoch": 4.173787806048968, "grad_norm": 0.7664195435406662, "learning_rate": 8.109399759760194e-07, "loss": 0.5697, "step": 4349 }, { "epoch": 4.174747959673548, "grad_norm": 1.120421901474019, "learning_rate": 8.091119759847493e-07, "loss": 0.4711, "step": 4350 }, { "epoch": 4.1757081132981275, "grad_norm": 1.0726943402625635, "learning_rate": 8.07285857238313e-07, "loss": 0.5776, "step": 4351 }, { "epoch": 4.176668266922707, "grad_norm": 1.2074775112216969, "learning_rate": 8.054616205564403e-07, "loss": 0.4246, "step": 4352 }, { "epoch": 4.177628420547288, "grad_norm": 1.0891304678498153, "learning_rate": 8.036392667580084e-07, "loss": 0.4, "step": 4353 }, { "epoch": 4.178588574171868, "grad_norm": 0.8666836733963192, "learning_rate": 8.018187966610535e-07, "loss": 0.4567, "step": 4354 }, { "epoch": 4.179548727796448, "grad_norm": 0.8354639314501893, "learning_rate": 8.000002110827626e-07, "loss": 0.5094, "step": 4355 }, { "epoch": 4.180508881421027, "grad_norm": 0.7681547036199203, "learning_rate": 7.981835108394825e-07, "loss": 0.4149, "step": 4356 }, { "epoch": 4.181469035045607, "grad_norm": 1.091860872943518, "learning_rate": 7.963686967467111e-07, "loss": 0.3389, "step": 4357 }, { "epoch": 4.182429188670187, "grad_norm": 0.9853420419337459, "learning_rate": 7.945557696190959e-07, "loss": 0.403, "step": 4358 }, { "epoch": 4.183389342294767, "grad_norm": 0.8316527700630447, "learning_rate": 7.927447302704422e-07, "loss": 0.5309, "step": 4359 }, { "epoch": 4.1843494959193475, "grad_norm": 0.8515513968965307, "learning_rate": 7.909355795137058e-07, "loss": 0.524, "step": 4360 }, { "epoch": 4.185309649543927, "grad_norm": 0.8148281367449892, "learning_rate": 7.891283181609955e-07, "loss": 0.5437, "step": 4361 }, { "epoch": 4.186269803168507, "grad_norm": 0.6738770199421895, "learning_rate": 7.873229470235722e-07, "loss": 0.4411, "step": 4362 }, { "epoch": 4.187229956793087, "grad_norm": 0.9726783547513081, "learning_rate": 7.855194669118466e-07, "loss": 0.4181, "step": 4363 }, { "epoch": 4.188190110417667, "grad_norm": 0.9774514291922914, "learning_rate": 7.837178786353811e-07, "loss": 0.4549, "step": 4364 }, { "epoch": 4.189150264042246, "grad_norm": 0.915401308853798, "learning_rate": 7.81918183002891e-07, "loss": 0.4763, "step": 4365 }, { "epoch": 4.190110417666826, "grad_norm": 0.7355172223236986, "learning_rate": 7.801203808222396e-07, "loss": 0.6094, "step": 4366 }, { "epoch": 4.191070571291407, "grad_norm": 0.8359333320215213, "learning_rate": 7.783244729004425e-07, "loss": 0.4593, "step": 4367 }, { "epoch": 4.192030724915987, "grad_norm": 0.7231050837148997, "learning_rate": 7.765304600436596e-07, "loss": 0.4974, "step": 4368 }, { "epoch": 4.1929908785405665, "grad_norm": 0.9820068050778495, "learning_rate": 7.747383430572081e-07, "loss": 0.4222, "step": 4369 }, { "epoch": 4.193951032165146, "grad_norm": 0.7380098355777244, "learning_rate": 7.729481227455498e-07, "loss": 0.511, "step": 4370 }, { "epoch": 4.194911185789726, "grad_norm": 0.8019234441761591, "learning_rate": 7.711597999122939e-07, "loss": 0.4473, "step": 4371 }, { "epoch": 4.195871339414306, "grad_norm": 1.1531247218664065, "learning_rate": 7.693733753601995e-07, "loss": 0.4343, "step": 4372 }, { "epoch": 4.196831493038887, "grad_norm": 0.788840518649079, "learning_rate": 7.675888498911743e-07, "loss": 0.4444, "step": 4373 }, { "epoch": 4.197791646663466, "grad_norm": 0.9665236312347353, "learning_rate": 7.658062243062725e-07, "loss": 0.5776, "step": 4374 }, { "epoch": 4.198751800288046, "grad_norm": 0.8581300927574299, "learning_rate": 7.640254994056967e-07, "loss": 0.5341, "step": 4375 }, { "epoch": 4.199711953912626, "grad_norm": 0.9113890500446716, "learning_rate": 7.622466759887931e-07, "loss": 0.4129, "step": 4376 }, { "epoch": 4.200672107537206, "grad_norm": 0.85871988432874, "learning_rate": 7.604697548540574e-07, "loss": 0.4424, "step": 4377 }, { "epoch": 4.201632261161786, "grad_norm": 1.0839432959041035, "learning_rate": 7.58694736799131e-07, "loss": 0.3776, "step": 4378 }, { "epoch": 4.202592414786366, "grad_norm": 0.6997663110690233, "learning_rate": 7.569216226207999e-07, "loss": 0.5284, "step": 4379 }, { "epoch": 4.203552568410946, "grad_norm": 2.026715118284863, "learning_rate": 7.551504131149972e-07, "loss": 0.4699, "step": 4380 }, { "epoch": 4.204512722035526, "grad_norm": 0.8915474552160914, "learning_rate": 7.533811090767984e-07, "loss": 0.5145, "step": 4381 }, { "epoch": 4.205472875660106, "grad_norm": 0.8021039419141872, "learning_rate": 7.516137113004246e-07, "loss": 0.4321, "step": 4382 }, { "epoch": 4.206433029284685, "grad_norm": 1.2936919873587827, "learning_rate": 7.498482205792457e-07, "loss": 0.442, "step": 4383 }, { "epoch": 4.207393182909265, "grad_norm": 0.897384059856078, "learning_rate": 7.480846377057677e-07, "loss": 0.4285, "step": 4384 }, { "epoch": 4.208353336533845, "grad_norm": 0.7969320068756583, "learning_rate": 7.463229634716462e-07, "loss": 0.4773, "step": 4385 }, { "epoch": 4.209313490158426, "grad_norm": 0.9344987608394323, "learning_rate": 7.445631986676755e-07, "loss": 0.4798, "step": 4386 }, { "epoch": 4.2102736437830055, "grad_norm": 0.7865019718075021, "learning_rate": 7.428053440837973e-07, "loss": 0.4901, "step": 4387 }, { "epoch": 4.211233797407585, "grad_norm": 0.8069471233694514, "learning_rate": 7.410494005090946e-07, "loss": 0.5171, "step": 4388 }, { "epoch": 4.212193951032165, "grad_norm": 0.6740959542626899, "learning_rate": 7.392953687317894e-07, "loss": 0.4609, "step": 4389 }, { "epoch": 4.213154104656745, "grad_norm": 0.9249536082733577, "learning_rate": 7.375432495392487e-07, "loss": 0.4951, "step": 4390 }, { "epoch": 4.214114258281325, "grad_norm": 0.8288011717499868, "learning_rate": 7.3579304371798e-07, "loss": 0.4298, "step": 4391 }, { "epoch": 4.215074411905905, "grad_norm": 0.7596175183963028, "learning_rate": 7.34044752053632e-07, "loss": 0.4414, "step": 4392 }, { "epoch": 4.216034565530485, "grad_norm": 0.9756967895656372, "learning_rate": 7.322983753309959e-07, "loss": 0.4323, "step": 4393 }, { "epoch": 4.216994719155065, "grad_norm": 0.897518136916691, "learning_rate": 7.305539143339985e-07, "loss": 0.4836, "step": 4394 }, { "epoch": 4.217954872779645, "grad_norm": 1.1130830512637204, "learning_rate": 7.2881136984571e-07, "loss": 0.4379, "step": 4395 }, { "epoch": 4.2189150264042246, "grad_norm": 1.0408037231435858, "learning_rate": 7.270707426483431e-07, "loss": 0.4217, "step": 4396 }, { "epoch": 4.219875180028804, "grad_norm": 1.304814585990243, "learning_rate": 7.253320335232439e-07, "loss": 0.3997, "step": 4397 }, { "epoch": 4.220835333653384, "grad_norm": 1.10001994851479, "learning_rate": 7.235952432509024e-07, "loss": 0.3635, "step": 4398 }, { "epoch": 4.221795487277965, "grad_norm": 0.88321535999021, "learning_rate": 7.218603726109413e-07, "loss": 0.46, "step": 4399 }, { "epoch": 4.222755640902545, "grad_norm": 0.9169162163386833, "learning_rate": 7.201274223821298e-07, "loss": 0.5309, "step": 4400 }, { "epoch": 4.223715794527124, "grad_norm": 1.0261912735726835, "learning_rate": 7.183963933423699e-07, "loss": 0.4696, "step": 4401 }, { "epoch": 4.224675948151704, "grad_norm": 0.7794879827105895, "learning_rate": 7.166672862687002e-07, "loss": 0.4255, "step": 4402 }, { "epoch": 4.225636101776284, "grad_norm": 0.6306891464658564, "learning_rate": 7.149401019372993e-07, "loss": 0.4012, "step": 4403 }, { "epoch": 4.226596255400864, "grad_norm": 0.9713853993082007, "learning_rate": 7.132148411234818e-07, "loss": 0.4376, "step": 4404 }, { "epoch": 4.2275564090254445, "grad_norm": 0.9111649958590405, "learning_rate": 7.11491504601699e-07, "loss": 0.4021, "step": 4405 }, { "epoch": 4.228516562650024, "grad_norm": 0.7631561225268061, "learning_rate": 7.097700931455397e-07, "loss": 0.449, "step": 4406 }, { "epoch": 4.229476716274604, "grad_norm": 0.7441285696807136, "learning_rate": 7.080506075277243e-07, "loss": 0.4374, "step": 4407 }, { "epoch": 4.230436869899184, "grad_norm": 0.7887187514538087, "learning_rate": 7.063330485201126e-07, "loss": 0.4786, "step": 4408 }, { "epoch": 4.231397023523764, "grad_norm": 0.881970014680061, "learning_rate": 7.046174168936987e-07, "loss": 0.5447, "step": 4409 }, { "epoch": 4.2323571771483435, "grad_norm": 1.2652148042526734, "learning_rate": 7.029037134186112e-07, "loss": 0.3841, "step": 4410 }, { "epoch": 4.233317330772923, "grad_norm": 0.6870851440950798, "learning_rate": 7.011919388641148e-07, "loss": 0.3694, "step": 4411 }, { "epoch": 4.234277484397504, "grad_norm": 0.9386475548652757, "learning_rate": 6.994820939986025e-07, "loss": 0.4259, "step": 4412 }, { "epoch": 4.235237638022084, "grad_norm": 0.898820692885359, "learning_rate": 6.977741795896098e-07, "loss": 0.4926, "step": 4413 }, { "epoch": 4.2361977916466635, "grad_norm": 0.9036547893283904, "learning_rate": 6.960681964038008e-07, "loss": 0.4194, "step": 4414 }, { "epoch": 4.237157945271243, "grad_norm": 0.7651747455316642, "learning_rate": 6.943641452069705e-07, "loss": 0.4624, "step": 4415 }, { "epoch": 4.238118098895823, "grad_norm": 1.0360224615232205, "learning_rate": 6.92662026764051e-07, "loss": 0.4641, "step": 4416 }, { "epoch": 4.239078252520403, "grad_norm": 0.9961097787227992, "learning_rate": 6.909618418391045e-07, "loss": 0.4208, "step": 4417 }, { "epoch": 4.240038406144984, "grad_norm": 1.2035065586425844, "learning_rate": 6.892635911953261e-07, "loss": 0.4578, "step": 4418 }, { "epoch": 4.240998559769563, "grad_norm": 1.3561042614209045, "learning_rate": 6.875672755950441e-07, "loss": 0.4631, "step": 4419 }, { "epoch": 4.241958713394143, "grad_norm": 0.8433848460705656, "learning_rate": 6.858728957997129e-07, "loss": 0.4257, "step": 4420 }, { "epoch": 4.242918867018723, "grad_norm": 0.9656233518699683, "learning_rate": 6.841804525699236e-07, "loss": 0.3909, "step": 4421 }, { "epoch": 4.243879020643303, "grad_norm": 0.7676250566703666, "learning_rate": 6.824899466653961e-07, "loss": 0.4699, "step": 4422 }, { "epoch": 4.244839174267883, "grad_norm": 0.9193430513387799, "learning_rate": 6.808013788449791e-07, "loss": 0.4316, "step": 4423 }, { "epoch": 4.245799327892462, "grad_norm": 0.801509254284077, "learning_rate": 6.79114749866655e-07, "loss": 0.5371, "step": 4424 }, { "epoch": 4.246759481517043, "grad_norm": 0.7993506891986629, "learning_rate": 6.774300604875312e-07, "loss": 0.4468, "step": 4425 }, { "epoch": 4.247719635141623, "grad_norm": 1.2517274034081474, "learning_rate": 6.757473114638458e-07, "loss": 0.516, "step": 4426 }, { "epoch": 4.248679788766203, "grad_norm": 1.2394973558327544, "learning_rate": 6.740665035509714e-07, "loss": 0.4725, "step": 4427 }, { "epoch": 4.2496399423907825, "grad_norm": 0.7858397030318564, "learning_rate": 6.723876375034e-07, "loss": 0.4253, "step": 4428 }, { "epoch": 4.250600096015362, "grad_norm": 0.7603313867258878, "learning_rate": 6.707107140747604e-07, "loss": 0.4465, "step": 4429 }, { "epoch": 4.251560249639942, "grad_norm": 1.0186829098469345, "learning_rate": 6.690357340178011e-07, "loss": 0.4116, "step": 4430 }, { "epoch": 4.252520403264523, "grad_norm": 0.9174477620611604, "learning_rate": 6.673626980844067e-07, "loss": 0.5231, "step": 4431 }, { "epoch": 4.2534805568891025, "grad_norm": 1.2170042466078643, "learning_rate": 6.656916070255842e-07, "loss": 0.3437, "step": 4432 }, { "epoch": 4.254440710513682, "grad_norm": 0.7776467015152414, "learning_rate": 6.640224615914681e-07, "loss": 0.4656, "step": 4433 }, { "epoch": 4.255400864138262, "grad_norm": 0.7638418402366861, "learning_rate": 6.62355262531319e-07, "loss": 0.4523, "step": 4434 }, { "epoch": 4.256361017762842, "grad_norm": 0.7518292290477447, "learning_rate": 6.606900105935265e-07, "loss": 0.5249, "step": 4435 }, { "epoch": 4.257321171387422, "grad_norm": 0.7421585657722738, "learning_rate": 6.590267065256029e-07, "loss": 0.4865, "step": 4436 }, { "epoch": 4.2582813250120015, "grad_norm": 0.8643532347842229, "learning_rate": 6.573653510741901e-07, "loss": 0.5028, "step": 4437 }, { "epoch": 4.259241478636582, "grad_norm": 1.1835511450028287, "learning_rate": 6.557059449850494e-07, "loss": 0.5222, "step": 4438 }, { "epoch": 4.260201632261162, "grad_norm": 0.9940022768391624, "learning_rate": 6.540484890030724e-07, "loss": 0.5259, "step": 4439 }, { "epoch": 4.261161785885742, "grad_norm": 0.9594774581906959, "learning_rate": 6.523929838722726e-07, "loss": 0.3337, "step": 4440 }, { "epoch": 4.262121939510322, "grad_norm": 0.8288160661755188, "learning_rate": 6.507394303357894e-07, "loss": 0.5182, "step": 4441 }, { "epoch": 4.263082093134901, "grad_norm": 1.0420238359428533, "learning_rate": 6.49087829135886e-07, "loss": 0.4382, "step": 4442 }, { "epoch": 4.264042246759481, "grad_norm": 0.7633425845134069, "learning_rate": 6.474381810139446e-07, "loss": 0.4891, "step": 4443 }, { "epoch": 4.265002400384062, "grad_norm": 1.3088508532791028, "learning_rate": 6.457904867104792e-07, "loss": 0.4458, "step": 4444 }, { "epoch": 4.265962554008642, "grad_norm": 0.974607506051096, "learning_rate": 6.441447469651213e-07, "loss": 0.4449, "step": 4445 }, { "epoch": 4.2669227076332215, "grad_norm": 1.2185585365259715, "learning_rate": 6.42500962516624e-07, "loss": 0.572, "step": 4446 }, { "epoch": 4.267882861257801, "grad_norm": 1.1670471944023437, "learning_rate": 6.408591341028652e-07, "loss": 0.4772, "step": 4447 }, { "epoch": 4.268843014882381, "grad_norm": 0.6780441531053685, "learning_rate": 6.392192624608451e-07, "loss": 0.3815, "step": 4448 }, { "epoch": 4.269803168506961, "grad_norm": 0.8732953706476884, "learning_rate": 6.375813483266835e-07, "loss": 0.529, "step": 4449 }, { "epoch": 4.270763322131541, "grad_norm": 0.8743100494079687, "learning_rate": 6.35945392435624e-07, "loss": 0.4544, "step": 4450 }, { "epoch": 4.271723475756121, "grad_norm": 1.063839774970999, "learning_rate": 6.343113955220276e-07, "loss": 0.531, "step": 4451 }, { "epoch": 4.272683629380701, "grad_norm": 1.0383385680026316, "learning_rate": 6.326793583193797e-07, "loss": 0.4859, "step": 4452 }, { "epoch": 4.273643783005281, "grad_norm": 1.3677241025689286, "learning_rate": 6.310492815602831e-07, "loss": 0.5654, "step": 4453 }, { "epoch": 4.274603936629861, "grad_norm": 0.8934589522200643, "learning_rate": 6.294211659764626e-07, "loss": 0.4687, "step": 4454 }, { "epoch": 4.2755640902544405, "grad_norm": 0.8373693813738303, "learning_rate": 6.277950122987631e-07, "loss": 0.5137, "step": 4455 }, { "epoch": 4.27652424387902, "grad_norm": 0.7790777009275738, "learning_rate": 6.261708212571438e-07, "loss": 0.47, "step": 4456 }, { "epoch": 4.277484397503601, "grad_norm": 0.7823685439260714, "learning_rate": 6.245485935806905e-07, "loss": 0.4901, "step": 4457 }, { "epoch": 4.278444551128181, "grad_norm": 0.9588814483038433, "learning_rate": 6.229283299976041e-07, "loss": 0.4209, "step": 4458 }, { "epoch": 4.279404704752761, "grad_norm": 0.7973395699193075, "learning_rate": 6.213100312352005e-07, "loss": 0.5489, "step": 4459 }, { "epoch": 4.28036485837734, "grad_norm": 0.7975980526054264, "learning_rate": 6.196936980199181e-07, "loss": 0.4859, "step": 4460 }, { "epoch": 4.28132501200192, "grad_norm": 1.1475121724426156, "learning_rate": 6.18079331077312e-07, "loss": 0.4834, "step": 4461 }, { "epoch": 4.2822851656265, "grad_norm": 0.7378053232604399, "learning_rate": 6.164669311320543e-07, "loss": 0.4323, "step": 4462 }, { "epoch": 4.28324531925108, "grad_norm": 0.818724350897374, "learning_rate": 6.148564989079347e-07, "loss": 0.4957, "step": 4463 }, { "epoch": 4.2842054728756604, "grad_norm": 0.902056346847682, "learning_rate": 6.13248035127857e-07, "loss": 0.4846, "step": 4464 }, { "epoch": 4.28516562650024, "grad_norm": 0.9905045558291343, "learning_rate": 6.116415405138443e-07, "loss": 0.4227, "step": 4465 }, { "epoch": 4.28612578012482, "grad_norm": 0.7118259088829861, "learning_rate": 6.100370157870356e-07, "loss": 0.5244, "step": 4466 }, { "epoch": 4.2870859337494, "grad_norm": 1.1256910282759405, "learning_rate": 6.08434461667684e-07, "loss": 0.4777, "step": 4467 }, { "epoch": 4.28804608737398, "grad_norm": 0.7731113156472914, "learning_rate": 6.06833878875161e-07, "loss": 0.5308, "step": 4468 }, { "epoch": 4.289006240998559, "grad_norm": 0.9921171701077229, "learning_rate": 6.052352681279478e-07, "loss": 0.4864, "step": 4469 }, { "epoch": 4.28996639462314, "grad_norm": 1.144932093716501, "learning_rate": 6.036386301436448e-07, "loss": 0.5198, "step": 4470 }, { "epoch": 4.29092654824772, "grad_norm": 0.7099882338824444, "learning_rate": 6.020439656389676e-07, "loss": 0.401, "step": 4471 }, { "epoch": 4.2918867018723, "grad_norm": 1.1877589144700407, "learning_rate": 6.004512753297421e-07, "loss": 0.3288, "step": 4472 }, { "epoch": 4.2928468554968795, "grad_norm": 1.1382268845336376, "learning_rate": 5.988605599309116e-07, "loss": 0.439, "step": 4473 }, { "epoch": 4.293807009121459, "grad_norm": 0.8401337674574306, "learning_rate": 5.97271820156528e-07, "loss": 0.4143, "step": 4474 }, { "epoch": 4.294767162746039, "grad_norm": 0.7569676206674237, "learning_rate": 5.95685056719763e-07, "loss": 0.4315, "step": 4475 }, { "epoch": 4.295727316370619, "grad_norm": 0.6669947353772528, "learning_rate": 5.941002703328974e-07, "loss": 0.447, "step": 4476 }, { "epoch": 4.2966874699952, "grad_norm": 0.833791359848185, "learning_rate": 5.925174617073226e-07, "loss": 0.5273, "step": 4477 }, { "epoch": 4.297647623619779, "grad_norm": 0.8048018642909697, "learning_rate": 5.909366315535464e-07, "loss": 0.5064, "step": 4478 }, { "epoch": 4.298607777244359, "grad_norm": 0.85886096588653, "learning_rate": 5.893577805811856e-07, "loss": 0.4549, "step": 4479 }, { "epoch": 4.299567930868939, "grad_norm": 0.7287541615009318, "learning_rate": 5.877809094989706e-07, "loss": 0.4176, "step": 4480 }, { "epoch": 4.300528084493519, "grad_norm": 0.9359898684150765, "learning_rate": 5.862060190147417e-07, "loss": 0.5035, "step": 4481 }, { "epoch": 4.3014882381180986, "grad_norm": 1.0259276964154722, "learning_rate": 5.84633109835449e-07, "loss": 0.3835, "step": 4482 }, { "epoch": 4.302448391742679, "grad_norm": 1.4304028525800667, "learning_rate": 5.830621826671562e-07, "loss": 0.393, "step": 4483 }, { "epoch": 4.303408545367259, "grad_norm": 1.1766818044881837, "learning_rate": 5.814932382150346e-07, "loss": 0.462, "step": 4484 }, { "epoch": 4.304368698991839, "grad_norm": 0.9084847675670823, "learning_rate": 5.799262771833675e-07, "loss": 0.5161, "step": 4485 }, { "epoch": 4.305328852616419, "grad_norm": 0.7752387622184189, "learning_rate": 5.783613002755478e-07, "loss": 0.4862, "step": 4486 }, { "epoch": 4.306289006240998, "grad_norm": 0.8089736887202884, "learning_rate": 5.767983081940731e-07, "loss": 0.4295, "step": 4487 }, { "epoch": 4.307249159865578, "grad_norm": 0.9814416451678882, "learning_rate": 5.752373016405577e-07, "loss": 0.4622, "step": 4488 }, { "epoch": 4.308209313490158, "grad_norm": 0.7456740685599017, "learning_rate": 5.736782813157204e-07, "loss": 0.4564, "step": 4489 }, { "epoch": 4.309169467114739, "grad_norm": 1.5403388497454777, "learning_rate": 5.721212479193871e-07, "loss": 0.4321, "step": 4490 }, { "epoch": 4.3101296207393185, "grad_norm": 0.9600202139765531, "learning_rate": 5.705662021504943e-07, "loss": 0.4611, "step": 4491 }, { "epoch": 4.311089774363898, "grad_norm": 0.6517143038861583, "learning_rate": 5.690131447070851e-07, "loss": 0.5444, "step": 4492 }, { "epoch": 4.312049927988478, "grad_norm": 0.9404292407576655, "learning_rate": 5.674620762863109e-07, "loss": 0.5099, "step": 4493 }, { "epoch": 4.313010081613058, "grad_norm": 0.8739921506921365, "learning_rate": 5.659129975844302e-07, "loss": 0.5233, "step": 4494 }, { "epoch": 4.313970235237638, "grad_norm": 0.7672376701492597, "learning_rate": 5.643659092968062e-07, "loss": 0.5172, "step": 4495 }, { "epoch": 4.314930388862218, "grad_norm": 1.0618195052709622, "learning_rate": 5.628208121179107e-07, "loss": 0.4563, "step": 4496 }, { "epoch": 4.315890542486798, "grad_norm": 0.70352741348272, "learning_rate": 5.612777067413228e-07, "loss": 0.4853, "step": 4497 }, { "epoch": 4.316850696111378, "grad_norm": 1.0649833953418502, "learning_rate": 5.597365938597243e-07, "loss": 0.4359, "step": 4498 }, { "epoch": 4.317810849735958, "grad_norm": 1.0060138065505244, "learning_rate": 5.581974741649066e-07, "loss": 0.4978, "step": 4499 }, { "epoch": 4.3187710033605375, "grad_norm": 0.8688887367034764, "learning_rate": 5.566603483477607e-07, "loss": 0.4977, "step": 4500 }, { "epoch": 4.319731156985117, "grad_norm": 0.8776054376888042, "learning_rate": 5.551252170982874e-07, "loss": 0.5017, "step": 4501 }, { "epoch": 4.320691310609698, "grad_norm": 0.7095576090177874, "learning_rate": 5.535920811055928e-07, "loss": 0.4016, "step": 4502 }, { "epoch": 4.321651464234278, "grad_norm": 0.8052842641367718, "learning_rate": 5.520609410578826e-07, "loss": 0.5357, "step": 4503 }, { "epoch": 4.322611617858858, "grad_norm": 0.7025043311738873, "learning_rate": 5.505317976424712e-07, "loss": 0.4702, "step": 4504 }, { "epoch": 4.323571771483437, "grad_norm": 0.9231003327755292, "learning_rate": 5.49004651545772e-07, "loss": 0.4139, "step": 4505 }, { "epoch": 4.324531925108017, "grad_norm": 0.8983252223497511, "learning_rate": 5.474795034533064e-07, "loss": 0.4215, "step": 4506 }, { "epoch": 4.325492078732597, "grad_norm": 0.8680074650483032, "learning_rate": 5.459563540496987e-07, "loss": 0.3675, "step": 4507 }, { "epoch": 4.326452232357177, "grad_norm": 0.8452083942496762, "learning_rate": 5.444352040186712e-07, "loss": 0.4785, "step": 4508 }, { "epoch": 4.3274123859817575, "grad_norm": 0.813210787908203, "learning_rate": 5.429160540430533e-07, "loss": 0.4392, "step": 4509 }, { "epoch": 4.328372539606337, "grad_norm": 0.6711196013151076, "learning_rate": 5.413989048047752e-07, "loss": 0.4737, "step": 4510 }, { "epoch": 4.329332693230917, "grad_norm": 0.7185833167220337, "learning_rate": 5.398837569848686e-07, "loss": 0.4409, "step": 4511 }, { "epoch": 4.330292846855497, "grad_norm": 1.4068161719592658, "learning_rate": 5.38370611263469e-07, "loss": 0.3947, "step": 4512 }, { "epoch": 4.331253000480077, "grad_norm": 0.7972007740207678, "learning_rate": 5.368594683198087e-07, "loss": 0.4883, "step": 4513 }, { "epoch": 4.3322131541046565, "grad_norm": 1.0194361412969433, "learning_rate": 5.353503288322243e-07, "loss": 0.4679, "step": 4514 }, { "epoch": 4.333173307729236, "grad_norm": 0.8488437794972901, "learning_rate": 5.338431934781535e-07, "loss": 0.5266, "step": 4515 }, { "epoch": 4.334133461353817, "grad_norm": 0.6735259807180475, "learning_rate": 5.323380629341324e-07, "loss": 0.5108, "step": 4516 }, { "epoch": 4.335093614978397, "grad_norm": 1.3874258129745916, "learning_rate": 5.308349378757988e-07, "loss": 0.5384, "step": 4517 }, { "epoch": 4.3360537686029765, "grad_norm": 0.7347476819362757, "learning_rate": 5.293338189778874e-07, "loss": 0.451, "step": 4518 }, { "epoch": 4.337013922227556, "grad_norm": 0.7737274139570938, "learning_rate": 5.278347069142364e-07, "loss": 0.4771, "step": 4519 }, { "epoch": 4.337974075852136, "grad_norm": 0.8551900027614615, "learning_rate": 5.263376023577821e-07, "loss": 0.4654, "step": 4520 }, { "epoch": 4.338934229476716, "grad_norm": 0.8756844966925763, "learning_rate": 5.248425059805557e-07, "loss": 0.4123, "step": 4521 }, { "epoch": 4.339894383101297, "grad_norm": 0.5716344249192349, "learning_rate": 5.23349418453692e-07, "loss": 0.4204, "step": 4522 }, { "epoch": 4.340854536725876, "grad_norm": 0.9929429914718405, "learning_rate": 5.218583404474214e-07, "loss": 0.3509, "step": 4523 }, { "epoch": 4.341814690350456, "grad_norm": 0.7447980410131674, "learning_rate": 5.203692726310738e-07, "loss": 0.4447, "step": 4524 }, { "epoch": 4.342774843975036, "grad_norm": 0.8182229220066706, "learning_rate": 5.188822156730756e-07, "loss": 0.4638, "step": 4525 }, { "epoch": 4.343734997599616, "grad_norm": 0.7794769046876837, "learning_rate": 5.173971702409491e-07, "loss": 0.4785, "step": 4526 }, { "epoch": 4.344695151224196, "grad_norm": 1.1785540165281798, "learning_rate": 5.159141370013165e-07, "loss": 0.5034, "step": 4527 }, { "epoch": 4.345655304848776, "grad_norm": 0.7049126361370234, "learning_rate": 5.144331166198963e-07, "loss": 0.5039, "step": 4528 }, { "epoch": 4.346615458473356, "grad_norm": 0.8114932017007331, "learning_rate": 5.129541097615009e-07, "loss": 0.3828, "step": 4529 }, { "epoch": 4.347575612097936, "grad_norm": 1.7392914008145774, "learning_rate": 5.114771170900434e-07, "loss": 0.4934, "step": 4530 }, { "epoch": 4.348535765722516, "grad_norm": 0.8916641339131138, "learning_rate": 5.100021392685267e-07, "loss": 0.4855, "step": 4531 }, { "epoch": 4.3494959193470955, "grad_norm": 0.8528892138641294, "learning_rate": 5.08529176959055e-07, "loss": 0.5339, "step": 4532 }, { "epoch": 4.350456072971675, "grad_norm": 0.7206504333692724, "learning_rate": 5.070582308228256e-07, "loss": 0.503, "step": 4533 }, { "epoch": 4.351416226596255, "grad_norm": 0.9725538263948202, "learning_rate": 5.05589301520128e-07, "loss": 0.4458, "step": 4534 }, { "epoch": 4.352376380220836, "grad_norm": 0.8920944243383255, "learning_rate": 5.041223897103509e-07, "loss": 0.573, "step": 4535 }, { "epoch": 4.3533365338454155, "grad_norm": 0.855515374486098, "learning_rate": 5.026574960519747e-07, "loss": 0.5278, "step": 4536 }, { "epoch": 4.354296687469995, "grad_norm": 0.5771256039586815, "learning_rate": 5.011946212025747e-07, "loss": 0.4343, "step": 4537 }, { "epoch": 4.355256841094575, "grad_norm": 0.9279473939613587, "learning_rate": 4.997337658188207e-07, "loss": 0.5752, "step": 4538 }, { "epoch": 4.356216994719155, "grad_norm": 0.8789720444153342, "learning_rate": 4.982749305564727e-07, "loss": 0.4637, "step": 4539 }, { "epoch": 4.357177148343735, "grad_norm": 1.061043842823861, "learning_rate": 4.968181160703877e-07, "loss": 0.4529, "step": 4540 }, { "epoch": 4.3581373019683145, "grad_norm": 0.9296144686532484, "learning_rate": 4.953633230145144e-07, "loss": 0.5078, "step": 4541 }, { "epoch": 4.359097455592895, "grad_norm": 0.8862191091197755, "learning_rate": 4.93910552041893e-07, "loss": 0.3591, "step": 4542 }, { "epoch": 4.360057609217475, "grad_norm": 0.7132039707525786, "learning_rate": 4.924598038046591e-07, "loss": 0.4039, "step": 4543 }, { "epoch": 4.361017762842055, "grad_norm": 0.6945126568562365, "learning_rate": 4.910110789540346e-07, "loss": 0.5357, "step": 4544 }, { "epoch": 4.361977916466635, "grad_norm": 0.7684704255331473, "learning_rate": 4.895643781403375e-07, "loss": 0.4107, "step": 4545 }, { "epoch": 4.362938070091214, "grad_norm": 0.8028073604728271, "learning_rate": 4.881197020129797e-07, "loss": 0.5632, "step": 4546 }, { "epoch": 4.363898223715794, "grad_norm": 1.042660405380427, "learning_rate": 4.866770512204566e-07, "loss": 0.4092, "step": 4547 }, { "epoch": 4.364858377340375, "grad_norm": 0.6748406319312957, "learning_rate": 4.852364264103621e-07, "loss": 0.4675, "step": 4548 }, { "epoch": 4.365818530964955, "grad_norm": 1.425172926590198, "learning_rate": 4.837978282293731e-07, "loss": 0.3949, "step": 4549 }, { "epoch": 4.366778684589534, "grad_norm": 0.8745004855226827, "learning_rate": 4.823612573232644e-07, "loss": 0.4711, "step": 4550 }, { "epoch": 4.367738838214114, "grad_norm": 0.9699895803415595, "learning_rate": 4.809267143368979e-07, "loss": 0.4566, "step": 4551 }, { "epoch": 4.368698991838694, "grad_norm": 0.7982870432555138, "learning_rate": 4.794941999142222e-07, "loss": 0.4002, "step": 4552 }, { "epoch": 4.369659145463274, "grad_norm": 0.9156853261055181, "learning_rate": 4.780637146982781e-07, "loss": 0.4899, "step": 4553 }, { "epoch": 4.3706192990878545, "grad_norm": 1.1041218151471073, "learning_rate": 4.7663525933119625e-07, "loss": 0.6033, "step": 4554 }, { "epoch": 4.371579452712434, "grad_norm": 0.7798686501226011, "learning_rate": 4.7520883445419387e-07, "loss": 0.4515, "step": 4555 }, { "epoch": 4.372539606337014, "grad_norm": 0.792631516002396, "learning_rate": 4.737844407075798e-07, "loss": 0.4844, "step": 4556 }, { "epoch": 4.373499759961594, "grad_norm": 0.770240557503258, "learning_rate": 4.723620787307465e-07, "loss": 0.4665, "step": 4557 }, { "epoch": 4.374459913586174, "grad_norm": 0.6996474495764845, "learning_rate": 4.7094174916217817e-07, "loss": 0.458, "step": 4558 }, { "epoch": 4.3754200672107535, "grad_norm": 0.9581904836869652, "learning_rate": 4.69523452639446e-07, "loss": 0.5089, "step": 4559 }, { "epoch": 4.376380220835333, "grad_norm": 1.1618737110760582, "learning_rate": 4.6810718979920734e-07, "loss": 0.4168, "step": 4560 }, { "epoch": 4.377340374459914, "grad_norm": 1.2642035387336457, "learning_rate": 4.666929612772092e-07, "loss": 0.4351, "step": 4561 }, { "epoch": 4.378300528084494, "grad_norm": 0.9448193447862767, "learning_rate": 4.6528076770828045e-07, "loss": 0.4407, "step": 4562 }, { "epoch": 4.379260681709074, "grad_norm": 0.6793147729783259, "learning_rate": 4.638706097263429e-07, "loss": 0.4783, "step": 4563 }, { "epoch": 4.380220835333653, "grad_norm": 0.826303073221589, "learning_rate": 4.6246248796440074e-07, "loss": 0.4769, "step": 4564 }, { "epoch": 4.381180988958233, "grad_norm": 0.8203244276044834, "learning_rate": 4.6105640305454337e-07, "loss": 0.3942, "step": 4565 }, { "epoch": 4.382141142582813, "grad_norm": 0.822813946503897, "learning_rate": 4.5965235562794754e-07, "loss": 0.5581, "step": 4566 }, { "epoch": 4.383101296207393, "grad_norm": 1.2193716259780936, "learning_rate": 4.582503463148763e-07, "loss": 0.3948, "step": 4567 }, { "epoch": 4.384061449831973, "grad_norm": 0.8075314231477937, "learning_rate": 4.5685037574467574e-07, "loss": 0.479, "step": 4568 }, { "epoch": 4.385021603456553, "grad_norm": 0.6722646899307166, "learning_rate": 4.5545244454577863e-07, "loss": 0.3828, "step": 4569 }, { "epoch": 4.385981757081133, "grad_norm": 1.2020543969748223, "learning_rate": 4.540565533457003e-07, "loss": 0.4017, "step": 4570 }, { "epoch": 4.386941910705713, "grad_norm": 0.7625626926874675, "learning_rate": 4.5266270277104116e-07, "loss": 0.4651, "step": 4571 }, { "epoch": 4.387902064330293, "grad_norm": 1.0964645858840152, "learning_rate": 4.512708934474863e-07, "loss": 0.4804, "step": 4572 }, { "epoch": 4.388862217954872, "grad_norm": 0.8746159711557085, "learning_rate": 4.498811259998048e-07, "loss": 0.4033, "step": 4573 }, { "epoch": 4.389822371579453, "grad_norm": 1.0852343613609734, "learning_rate": 4.4849340105184823e-07, "loss": 0.5588, "step": 4574 }, { "epoch": 4.390782525204033, "grad_norm": 0.7259218875960756, "learning_rate": 4.4710771922654936e-07, "loss": 0.4881, "step": 4575 }, { "epoch": 4.391742678828613, "grad_norm": 1.3002785402480983, "learning_rate": 4.4572408114592845e-07, "loss": 0.4132, "step": 4576 }, { "epoch": 4.3927028324531925, "grad_norm": 0.6706165884019654, "learning_rate": 4.4434248743108587e-07, "loss": 0.3691, "step": 4577 }, { "epoch": 4.393662986077772, "grad_norm": 1.1523017214744773, "learning_rate": 4.4296293870220276e-07, "loss": 0.4982, "step": 4578 }, { "epoch": 4.394623139702352, "grad_norm": 1.074151259971911, "learning_rate": 4.415854355785443e-07, "loss": 0.5064, "step": 4579 }, { "epoch": 4.395583293326933, "grad_norm": 0.719490510295043, "learning_rate": 4.4020997867845695e-07, "loss": 0.448, "step": 4580 }, { "epoch": 4.396543446951513, "grad_norm": 0.9135281157111083, "learning_rate": 4.388365686193685e-07, "loss": 0.5501, "step": 4581 }, { "epoch": 4.397503600576092, "grad_norm": 0.7020104396398932, "learning_rate": 4.3746520601778974e-07, "loss": 0.5059, "step": 4582 }, { "epoch": 4.398463754200672, "grad_norm": 0.8404882351192802, "learning_rate": 4.3609589148930766e-07, "loss": 0.4925, "step": 4583 }, { "epoch": 4.399423907825252, "grad_norm": 0.7182870473748246, "learning_rate": 4.3472862564859496e-07, "loss": 0.4594, "step": 4584 }, { "epoch": 4.400384061449832, "grad_norm": 0.890697048899442, "learning_rate": 4.3336340910940186e-07, "loss": 0.501, "step": 4585 }, { "epoch": 4.4013442150744115, "grad_norm": 1.0060460979119374, "learning_rate": 4.320002424845604e-07, "loss": 0.3967, "step": 4586 }, { "epoch": 4.402304368698992, "grad_norm": 1.099295452312327, "learning_rate": 4.3063912638598106e-07, "loss": 0.5107, "step": 4587 }, { "epoch": 4.403264522323572, "grad_norm": 1.185775960102807, "learning_rate": 4.292800614246534e-07, "loss": 0.4791, "step": 4588 }, { "epoch": 4.404224675948152, "grad_norm": 1.227713847454258, "learning_rate": 4.2792304821064833e-07, "loss": 0.4505, "step": 4589 }, { "epoch": 4.405184829572732, "grad_norm": 0.7600400369043754, "learning_rate": 4.265680873531136e-07, "loss": 0.4548, "step": 4590 }, { "epoch": 4.406144983197311, "grad_norm": 0.7775181642756245, "learning_rate": 4.2521517946027715e-07, "loss": 0.4733, "step": 4591 }, { "epoch": 4.407105136821891, "grad_norm": 1.405334667391349, "learning_rate": 4.2386432513944587e-07, "loss": 0.5366, "step": 4592 }, { "epoch": 4.408065290446471, "grad_norm": 0.6844624221275151, "learning_rate": 4.225155249970014e-07, "loss": 0.487, "step": 4593 }, { "epoch": 4.409025444071052, "grad_norm": 1.1440564137939904, "learning_rate": 4.2116877963840786e-07, "loss": 0.5243, "step": 4594 }, { "epoch": 4.4099855976956315, "grad_norm": 0.7835355939371433, "learning_rate": 4.198240896682049e-07, "loss": 0.438, "step": 4595 }, { "epoch": 4.410945751320211, "grad_norm": 0.8165540499210122, "learning_rate": 4.18481455690008e-07, "loss": 0.52, "step": 4596 }, { "epoch": 4.411905904944791, "grad_norm": 0.9341752975341933, "learning_rate": 4.1714087830651226e-07, "loss": 0.6019, "step": 4597 }, { "epoch": 4.412866058569371, "grad_norm": 0.706822763536395, "learning_rate": 4.1580235811948753e-07, "loss": 0.4535, "step": 4598 }, { "epoch": 4.413826212193951, "grad_norm": 0.8642593073145319, "learning_rate": 4.14465895729782e-07, "loss": 0.4508, "step": 4599 }, { "epoch": 4.414786365818531, "grad_norm": 0.9013572778827121, "learning_rate": 4.131314917373208e-07, "loss": 0.4987, "step": 4600 }, { "epoch": 4.415746519443111, "grad_norm": 1.0609069872512749, "learning_rate": 4.1179914674110035e-07, "loss": 0.5225, "step": 4601 }, { "epoch": 4.416706673067691, "grad_norm": 0.6466332660235511, "learning_rate": 4.1046886133919825e-07, "loss": 0.4197, "step": 4602 }, { "epoch": 4.417666826692271, "grad_norm": 0.711166997836075, "learning_rate": 4.0914063612876473e-07, "loss": 0.5569, "step": 4603 }, { "epoch": 4.4186269803168505, "grad_norm": 0.8008278428663917, "learning_rate": 4.0781447170602617e-07, "loss": 0.5547, "step": 4604 }, { "epoch": 4.41958713394143, "grad_norm": 0.737131761609093, "learning_rate": 4.064903686662841e-07, "loss": 0.5416, "step": 4605 }, { "epoch": 4.420547287566011, "grad_norm": 0.7397846271590577, "learning_rate": 4.0516832760391256e-07, "loss": 0.5572, "step": 4606 }, { "epoch": 4.421507441190591, "grad_norm": 0.8059011322514746, "learning_rate": 4.0384834911236294e-07, "loss": 0.5251, "step": 4607 }, { "epoch": 4.422467594815171, "grad_norm": 1.4482902285092738, "learning_rate": 4.025304337841607e-07, "loss": 0.5104, "step": 4608 }, { "epoch": 4.42342774843975, "grad_norm": 0.9304876406832102, "learning_rate": 4.012145822109015e-07, "loss": 0.4373, "step": 4609 }, { "epoch": 4.42438790206433, "grad_norm": 1.1267446869302769, "learning_rate": 3.9990079498325786e-07, "loss": 0.5098, "step": 4610 }, { "epoch": 4.42534805568891, "grad_norm": 0.6463484680833234, "learning_rate": 3.985890726909747e-07, "loss": 0.4499, "step": 4611 }, { "epoch": 4.42630820931349, "grad_norm": 0.7663039452581064, "learning_rate": 3.97279415922871e-07, "loss": 0.5336, "step": 4612 }, { "epoch": 4.4272683629380705, "grad_norm": 0.7927553742470165, "learning_rate": 3.9597182526683763e-07, "loss": 0.4855, "step": 4613 }, { "epoch": 4.42822851656265, "grad_norm": 0.9722405800441338, "learning_rate": 3.946663013098373e-07, "loss": 0.5369, "step": 4614 }, { "epoch": 4.42918867018723, "grad_norm": 0.7766856380577424, "learning_rate": 3.9336284463790566e-07, "loss": 0.4355, "step": 4615 }, { "epoch": 4.43014882381181, "grad_norm": 0.8826128180764007, "learning_rate": 3.9206145583615076e-07, "loss": 0.4471, "step": 4616 }, { "epoch": 4.43110897743639, "grad_norm": 0.7782283739915025, "learning_rate": 3.90762135488752e-07, "loss": 0.5139, "step": 4617 }, { "epoch": 4.4320691310609694, "grad_norm": 0.8231326848275679, "learning_rate": 3.894648841789617e-07, "loss": 0.3735, "step": 4618 }, { "epoch": 4.433029284685549, "grad_norm": 0.8422076080060997, "learning_rate": 3.881697024890996e-07, "loss": 0.4331, "step": 4619 }, { "epoch": 4.43398943831013, "grad_norm": 0.9407844994810795, "learning_rate": 3.8687659100056063e-07, "loss": 0.4001, "step": 4620 }, { "epoch": 4.43494959193471, "grad_norm": 1.1708230022427275, "learning_rate": 3.8558555029380937e-07, "loss": 0.4983, "step": 4621 }, { "epoch": 4.4359097455592895, "grad_norm": 0.7683682243282864, "learning_rate": 3.8429658094837896e-07, "loss": 0.4005, "step": 4622 }, { "epoch": 4.436869899183869, "grad_norm": 0.726558035345366, "learning_rate": 3.830096835428737e-07, "loss": 0.4931, "step": 4623 }, { "epoch": 4.437830052808449, "grad_norm": 0.9793077677537522, "learning_rate": 3.8172485865496876e-07, "loss": 0.4731, "step": 4624 }, { "epoch": 4.438790206433029, "grad_norm": 0.8982035846265182, "learning_rate": 3.8044210686140784e-07, "loss": 0.5051, "step": 4625 }, { "epoch": 4.43975036005761, "grad_norm": 0.7666997755379185, "learning_rate": 3.791614287380063e-07, "loss": 0.5541, "step": 4626 }, { "epoch": 4.440710513682189, "grad_norm": 0.7969330032814169, "learning_rate": 3.778828248596439e-07, "loss": 0.4628, "step": 4627 }, { "epoch": 4.441670667306769, "grad_norm": 0.7316377244708072, "learning_rate": 3.766062958002742e-07, "loss": 0.3509, "step": 4628 }, { "epoch": 4.442630820931349, "grad_norm": 0.9479727687662354, "learning_rate": 3.7533184213291665e-07, "loss": 0.5214, "step": 4629 }, { "epoch": 4.443590974555929, "grad_norm": 0.9782195947037668, "learning_rate": 3.7405946442966023e-07, "loss": 0.4629, "step": 4630 }, { "epoch": 4.444551128180509, "grad_norm": 0.7656472033620342, "learning_rate": 3.727891632616626e-07, "loss": 0.3512, "step": 4631 }, { "epoch": 4.445511281805089, "grad_norm": 0.8474863330347192, "learning_rate": 3.7152093919914644e-07, "loss": 0.4739, "step": 4632 }, { "epoch": 4.446471435429669, "grad_norm": 0.8128135570652125, "learning_rate": 3.702547928114053e-07, "loss": 0.4169, "step": 4633 }, { "epoch": 4.447431589054249, "grad_norm": 1.1766740305278616, "learning_rate": 3.689907246667984e-07, "loss": 0.5698, "step": 4634 }, { "epoch": 4.448391742678829, "grad_norm": 0.7560359087432834, "learning_rate": 3.677287353327519e-07, "loss": 0.5017, "step": 4635 }, { "epoch": 4.449351896303408, "grad_norm": 0.993937453927248, "learning_rate": 3.6646882537576144e-07, "loss": 0.4572, "step": 4636 }, { "epoch": 4.450312049927988, "grad_norm": 1.602243174258801, "learning_rate": 3.652109953613836e-07, "loss": 0.522, "step": 4637 }, { "epoch": 4.451272203552568, "grad_norm": 0.9282952296146263, "learning_rate": 3.639552458542478e-07, "loss": 0.4835, "step": 4638 }, { "epoch": 4.452232357177149, "grad_norm": 0.8036820068875816, "learning_rate": 3.627015774180459e-07, "loss": 0.5626, "step": 4639 }, { "epoch": 4.4531925108017285, "grad_norm": 0.8070379090597591, "learning_rate": 3.614499906155355e-07, "loss": 0.4642, "step": 4640 }, { "epoch": 4.454152664426308, "grad_norm": 0.7742197078635148, "learning_rate": 3.602004860085406e-07, "loss": 0.3882, "step": 4641 }, { "epoch": 4.455112818050888, "grad_norm": 0.790531724974495, "learning_rate": 3.589530641579514e-07, "loss": 0.5274, "step": 4642 }, { "epoch": 4.456072971675468, "grad_norm": 0.7008645278476398, "learning_rate": 3.577077256237216e-07, "loss": 0.473, "step": 4643 }, { "epoch": 4.457033125300048, "grad_norm": 0.7501334601588905, "learning_rate": 3.564644709648707e-07, "loss": 0.4982, "step": 4644 }, { "epoch": 4.4579932789246275, "grad_norm": 0.7394238950286565, "learning_rate": 3.5522330073948177e-07, "loss": 0.5435, "step": 4645 }, { "epoch": 4.458953432549208, "grad_norm": 0.7530290711298909, "learning_rate": 3.53984215504703e-07, "loss": 0.3679, "step": 4646 }, { "epoch": 4.459913586173788, "grad_norm": 1.0850165966903176, "learning_rate": 3.5274721581674664e-07, "loss": 0.5922, "step": 4647 }, { "epoch": 4.460873739798368, "grad_norm": 1.086481873028536, "learning_rate": 3.5151230223088915e-07, "loss": 0.456, "step": 4648 }, { "epoch": 4.461833893422948, "grad_norm": 0.6746488259036413, "learning_rate": 3.5027947530147045e-07, "loss": 0.4463, "step": 4649 }, { "epoch": 4.462794047047527, "grad_norm": 0.9932577176785224, "learning_rate": 3.4904873558189055e-07, "loss": 0.4853, "step": 4650 }, { "epoch": 4.463754200672107, "grad_norm": 1.2459424553644305, "learning_rate": 3.4782008362461817e-07, "loss": 0.527, "step": 4651 }, { "epoch": 4.464714354296688, "grad_norm": 1.0193025734643002, "learning_rate": 3.4659351998118204e-07, "loss": 0.5789, "step": 4652 }, { "epoch": 4.465674507921268, "grad_norm": 0.9620334143337059, "learning_rate": 3.4536904520217184e-07, "loss": 0.3974, "step": 4653 }, { "epoch": 4.466634661545847, "grad_norm": 1.0876195132401631, "learning_rate": 3.441466598372423e-07, "loss": 0.4483, "step": 4654 }, { "epoch": 4.467594815170427, "grad_norm": 0.6822693235624294, "learning_rate": 3.429263644351083e-07, "loss": 0.4656, "step": 4655 }, { "epoch": 4.468554968795007, "grad_norm": 1.1717100921861974, "learning_rate": 3.417081595435484e-07, "loss": 0.478, "step": 4656 }, { "epoch": 4.469515122419587, "grad_norm": 0.8365310143479946, "learning_rate": 3.4049204570940185e-07, "loss": 0.4446, "step": 4657 }, { "epoch": 4.4704752760441675, "grad_norm": 0.9338376985130449, "learning_rate": 3.392780234785681e-07, "loss": 0.5152, "step": 4658 }, { "epoch": 4.471435429668747, "grad_norm": 1.027030542016933, "learning_rate": 3.3806609339600894e-07, "loss": 0.4841, "step": 4659 }, { "epoch": 4.472395583293327, "grad_norm": 1.1629840266201996, "learning_rate": 3.36856256005747e-07, "loss": 0.3672, "step": 4660 }, { "epoch": 4.473355736917907, "grad_norm": 0.9542771662641921, "learning_rate": 3.356485118508657e-07, "loss": 0.5038, "step": 4661 }, { "epoch": 4.474315890542487, "grad_norm": 0.8579100390019719, "learning_rate": 3.3444286147350955e-07, "loss": 0.3716, "step": 4662 }, { "epoch": 4.4752760441670665, "grad_norm": 0.8516303373222301, "learning_rate": 3.332393054148786e-07, "loss": 0.4348, "step": 4663 }, { "epoch": 4.476236197791646, "grad_norm": 0.7641599284662147, "learning_rate": 3.3203784421524056e-07, "loss": 0.5682, "step": 4664 }, { "epoch": 4.477196351416227, "grad_norm": 1.1666139279779757, "learning_rate": 3.3083847841391514e-07, "loss": 0.4765, "step": 4665 }, { "epoch": 4.478156505040807, "grad_norm": 0.9045556267724257, "learning_rate": 3.296412085492867e-07, "loss": 0.5064, "step": 4666 }, { "epoch": 4.4791166586653866, "grad_norm": 0.9624014251815759, "learning_rate": 3.2844603515879647e-07, "loss": 0.3683, "step": 4667 }, { "epoch": 4.480076812289966, "grad_norm": 0.6691840258966838, "learning_rate": 3.272529587789447e-07, "loss": 0.4511, "step": 4668 }, { "epoch": 4.481036965914546, "grad_norm": 1.0618070341581336, "learning_rate": 3.2606197994529076e-07, "loss": 0.433, "step": 4669 }, { "epoch": 4.481997119539126, "grad_norm": 1.3116271387333063, "learning_rate": 3.2487309919245313e-07, "loss": 0.5262, "step": 4670 }, { "epoch": 4.482957273163706, "grad_norm": 0.6858280944409239, "learning_rate": 3.236863170541066e-07, "loss": 0.4521, "step": 4671 }, { "epoch": 4.483917426788286, "grad_norm": 0.8985126591823042, "learning_rate": 3.22501634062985e-07, "loss": 0.4221, "step": 4672 }, { "epoch": 4.484877580412866, "grad_norm": 0.9452010124647932, "learning_rate": 3.213190507508801e-07, "loss": 0.4969, "step": 4673 }, { "epoch": 4.485837734037446, "grad_norm": 0.8364068279288167, "learning_rate": 3.2013856764864126e-07, "loss": 0.6446, "step": 4674 }, { "epoch": 4.486797887662026, "grad_norm": 0.69818785428534, "learning_rate": 3.1896018528617456e-07, "loss": 0.4736, "step": 4675 }, { "epoch": 4.487758041286606, "grad_norm": 1.06413955970531, "learning_rate": 3.17783904192443e-07, "loss": 0.4665, "step": 4676 }, { "epoch": 4.488718194911185, "grad_norm": 1.1629842690923666, "learning_rate": 3.166097248954669e-07, "loss": 0.4272, "step": 4677 }, { "epoch": 4.489678348535766, "grad_norm": 0.797164719117261, "learning_rate": 3.15437647922322e-07, "loss": 0.5003, "step": 4678 }, { "epoch": 4.490638502160346, "grad_norm": 0.7971719501397799, "learning_rate": 3.1426767379914234e-07, "loss": 0.4689, "step": 4679 }, { "epoch": 4.491598655784926, "grad_norm": 0.7978622259739498, "learning_rate": 3.1309980305111674e-07, "loss": 0.373, "step": 4680 }, { "epoch": 4.4925588094095055, "grad_norm": 0.6702093671923518, "learning_rate": 3.119340362024875e-07, "loss": 0.4144, "step": 4681 }, { "epoch": 4.493518963034085, "grad_norm": 1.2180917884818254, "learning_rate": 3.1077037377655827e-07, "loss": 0.5053, "step": 4682 }, { "epoch": 4.494479116658665, "grad_norm": 0.9246921432117737, "learning_rate": 3.096088162956834e-07, "loss": 0.4077, "step": 4683 }, { "epoch": 4.495439270283246, "grad_norm": 0.8963345231313784, "learning_rate": 3.084493642812725e-07, "loss": 0.4683, "step": 4684 }, { "epoch": 4.4963994239078255, "grad_norm": 0.7430583369702585, "learning_rate": 3.072920182537925e-07, "loss": 0.5287, "step": 4685 }, { "epoch": 4.497359577532405, "grad_norm": 0.749108411771166, "learning_rate": 3.0613677873276293e-07, "loss": 0.4917, "step": 4686 }, { "epoch": 4.498319731156985, "grad_norm": 1.1620295248971038, "learning_rate": 3.049836462367589e-07, "loss": 0.4776, "step": 4687 }, { "epoch": 4.499279884781565, "grad_norm": 1.185271312724475, "learning_rate": 3.038326212834103e-07, "loss": 0.3686, "step": 4688 }, { "epoch": 4.500240038406145, "grad_norm": 1.4311371540539202, "learning_rate": 3.0268370438939767e-07, "loss": 0.5411, "step": 4689 }, { "epoch": 4.5012001920307245, "grad_norm": 1.2354132908408686, "learning_rate": 3.015368960704584e-07, "loss": 0.3977, "step": 4690 }, { "epoch": 4.502160345655305, "grad_norm": 0.7519181220685499, "learning_rate": 3.003921968413831e-07, "loss": 0.4506, "step": 4691 }, { "epoch": 4.503120499279885, "grad_norm": 1.3011219816752164, "learning_rate": 2.9924960721601393e-07, "loss": 0.3566, "step": 4692 }, { "epoch": 4.504080652904465, "grad_norm": 0.8014521755847029, "learning_rate": 2.98109127707249e-07, "loss": 0.4279, "step": 4693 }, { "epoch": 4.505040806529045, "grad_norm": 1.2241492950637805, "learning_rate": 2.9697075882703486e-07, "loss": 0.4296, "step": 4694 }, { "epoch": 4.506000960153624, "grad_norm": 1.2040497067778926, "learning_rate": 2.95834501086375e-07, "loss": 0.5073, "step": 4695 }, { "epoch": 4.506961113778204, "grad_norm": 0.7394943645156205, "learning_rate": 2.9470035499532425e-07, "loss": 0.4918, "step": 4696 }, { "epoch": 4.507921267402784, "grad_norm": 1.0708371025533796, "learning_rate": 2.9356832106298584e-07, "loss": 0.4208, "step": 4697 }, { "epoch": 4.508881421027365, "grad_norm": 0.8298761750846005, "learning_rate": 2.9243839979751943e-07, "loss": 0.4862, "step": 4698 }, { "epoch": 4.5098415746519445, "grad_norm": 0.806477304011966, "learning_rate": 2.913105917061343e-07, "loss": 0.5362, "step": 4699 }, { "epoch": 4.510801728276524, "grad_norm": 0.7289639114947551, "learning_rate": 2.9018489729509103e-07, "loss": 0.3918, "step": 4700 }, { "epoch": 4.511761881901104, "grad_norm": 0.9842071010276117, "learning_rate": 2.8906131706970375e-07, "loss": 0.4517, "step": 4701 }, { "epoch": 4.512722035525684, "grad_norm": 0.7549824157329305, "learning_rate": 2.879398515343329e-07, "loss": 0.3774, "step": 4702 }, { "epoch": 4.5136821891502645, "grad_norm": 0.7715248544533407, "learning_rate": 2.8682050119239357e-07, "loss": 0.5129, "step": 4703 }, { "epoch": 4.514642342774844, "grad_norm": 0.8788190108282804, "learning_rate": 2.857032665463499e-07, "loss": 0.5154, "step": 4704 }, { "epoch": 4.515602496399424, "grad_norm": 1.4391367918137075, "learning_rate": 2.8458814809771694e-07, "loss": 0.43, "step": 4705 }, { "epoch": 4.516562650024004, "grad_norm": 0.9654489348306599, "learning_rate": 2.834751463470603e-07, "loss": 0.4502, "step": 4706 }, { "epoch": 4.517522803648584, "grad_norm": 1.1064350431787375, "learning_rate": 2.823642617939926e-07, "loss": 0.4792, "step": 4707 }, { "epoch": 4.5184829572731635, "grad_norm": 1.1516128022520271, "learning_rate": 2.8125549493717984e-07, "loss": 0.4358, "step": 4708 }, { "epoch": 4.519443110897743, "grad_norm": 0.8552526125767584, "learning_rate": 2.8014884627433404e-07, "loss": 0.426, "step": 4709 }, { "epoch": 4.520403264522324, "grad_norm": 0.768569660131042, "learning_rate": 2.7904431630221937e-07, "loss": 0.4757, "step": 4710 }, { "epoch": 4.521363418146904, "grad_norm": 1.2551523153725843, "learning_rate": 2.7794190551664604e-07, "loss": 0.4818, "step": 4711 }, { "epoch": 4.522323571771484, "grad_norm": 1.2520699984852053, "learning_rate": 2.7684161441247584e-07, "loss": 0.4059, "step": 4712 }, { "epoch": 4.523283725396063, "grad_norm": 0.7916241507927148, "learning_rate": 2.75743443483617e-07, "loss": 0.5246, "step": 4713 }, { "epoch": 4.524243879020643, "grad_norm": 1.270504429851397, "learning_rate": 2.746473932230281e-07, "loss": 0.4561, "step": 4714 }, { "epoch": 4.525204032645223, "grad_norm": 0.7980045652151333, "learning_rate": 2.7355346412271223e-07, "loss": 0.5642, "step": 4715 }, { "epoch": 4.526164186269803, "grad_norm": 0.8087464041999383, "learning_rate": 2.7246165667372314e-07, "loss": 0.4479, "step": 4716 }, { "epoch": 4.5271243398943835, "grad_norm": 0.8316284837395602, "learning_rate": 2.713719713661617e-07, "loss": 0.4407, "step": 4717 }, { "epoch": 4.528084493518963, "grad_norm": 0.8071429132710493, "learning_rate": 2.702844086891765e-07, "loss": 0.4334, "step": 4718 }, { "epoch": 4.529044647143543, "grad_norm": 1.0106878492971298, "learning_rate": 2.6919896913096264e-07, "loss": 0.5093, "step": 4719 }, { "epoch": 4.530004800768123, "grad_norm": 1.393267896860484, "learning_rate": 2.6811565317876154e-07, "loss": 0.4594, "step": 4720 }, { "epoch": 4.530964954392703, "grad_norm": 0.9098294338146282, "learning_rate": 2.67034461318863e-07, "loss": 0.3684, "step": 4721 }, { "epoch": 4.531925108017282, "grad_norm": 1.1457415891384533, "learning_rate": 2.659553940366016e-07, "loss": 0.374, "step": 4722 }, { "epoch": 4.532885261641862, "grad_norm": 0.6946732200319012, "learning_rate": 2.6487845181635994e-07, "loss": 0.4641, "step": 4723 }, { "epoch": 4.533845415266443, "grad_norm": 0.8351422638020232, "learning_rate": 2.638036351415668e-07, "loss": 0.3352, "step": 4724 }, { "epoch": 4.534805568891023, "grad_norm": 0.8621252254379258, "learning_rate": 2.627309444946929e-07, "loss": 0.4869, "step": 4725 }, { "epoch": 4.5357657225156025, "grad_norm": 0.9243602715439708, "learning_rate": 2.6166038035725984e-07, "loss": 0.4318, "step": 4726 }, { "epoch": 4.536725876140182, "grad_norm": 0.91634171896971, "learning_rate": 2.605919432098331e-07, "loss": 0.4835, "step": 4727 }, { "epoch": 4.537686029764762, "grad_norm": 1.033872382271473, "learning_rate": 2.5952563353202064e-07, "loss": 0.4762, "step": 4728 }, { "epoch": 4.538646183389343, "grad_norm": 0.7538323009996811, "learning_rate": 2.584614518024786e-07, "loss": 0.4602, "step": 4729 }, { "epoch": 4.539606337013923, "grad_norm": 1.0350316941016902, "learning_rate": 2.573993984989065e-07, "loss": 0.4169, "step": 4730 }, { "epoch": 4.540566490638502, "grad_norm": 1.1394829679606986, "learning_rate": 2.5633947409804907e-07, "loss": 0.4411, "step": 4731 }, { "epoch": 4.541526644263082, "grad_norm": 0.7166114486954768, "learning_rate": 2.552816790756951e-07, "loss": 0.4462, "step": 4732 }, { "epoch": 4.542486797887662, "grad_norm": 1.2729766995429028, "learning_rate": 2.542260139066771e-07, "loss": 0.5247, "step": 4733 }, { "epoch": 4.543446951512242, "grad_norm": 0.9388225633436925, "learning_rate": 2.5317247906487153e-07, "loss": 0.5099, "step": 4734 }, { "epoch": 4.544407105136822, "grad_norm": 1.1682604052430245, "learning_rate": 2.521210750231995e-07, "loss": 0.3991, "step": 4735 }, { "epoch": 4.545367258761402, "grad_norm": 1.6809885200509191, "learning_rate": 2.510718022536246e-07, "loss": 0.5078, "step": 4736 }, { "epoch": 4.546327412385982, "grad_norm": 0.7434751888551592, "learning_rate": 2.5002466122715505e-07, "loss": 0.5102, "step": 4737 }, { "epoch": 4.547287566010562, "grad_norm": 0.8164511219441575, "learning_rate": 2.489796524138394e-07, "loss": 0.5309, "step": 4738 }, { "epoch": 4.548247719635142, "grad_norm": 0.9484028915314612, "learning_rate": 2.479367762827739e-07, "loss": 0.4387, "step": 4739 }, { "epoch": 4.549207873259721, "grad_norm": 0.8402252367025563, "learning_rate": 2.46896033302092e-07, "loss": 0.4516, "step": 4740 }, { "epoch": 4.550168026884301, "grad_norm": 0.6579043985962874, "learning_rate": 2.4585742393897274e-07, "loss": 0.4639, "step": 4741 }, { "epoch": 4.551128180508881, "grad_norm": 0.9131734694548054, "learning_rate": 2.44820948659637e-07, "loss": 0.466, "step": 4742 }, { "epoch": 4.552088334133462, "grad_norm": 1.4068012089655582, "learning_rate": 2.4378660792934715e-07, "loss": 0.5564, "step": 4743 }, { "epoch": 4.5530484877580415, "grad_norm": 0.9884110496388859, "learning_rate": 2.4275440221240786e-07, "loss": 0.3999, "step": 4744 }, { "epoch": 4.554008641382621, "grad_norm": 0.777329431546367, "learning_rate": 2.4172433197216625e-07, "loss": 0.5557, "step": 4745 }, { "epoch": 4.554968795007201, "grad_norm": 0.8265935623306391, "learning_rate": 2.4069639767100805e-07, "loss": 0.4594, "step": 4746 }, { "epoch": 4.555928948631781, "grad_norm": 0.7661443829904606, "learning_rate": 2.39670599770363e-07, "loss": 0.4422, "step": 4747 }, { "epoch": 4.556889102256361, "grad_norm": 0.9517233477650007, "learning_rate": 2.3864693873070054e-07, "loss": 0.4955, "step": 4748 }, { "epoch": 4.5578492558809405, "grad_norm": 1.3197219872122448, "learning_rate": 2.376254150115309e-07, "loss": 0.4741, "step": 4749 }, { "epoch": 4.558809409505521, "grad_norm": 0.8312271786514669, "learning_rate": 2.3660602907140672e-07, "loss": 0.4136, "step": 4750 }, { "epoch": 4.559769563130101, "grad_norm": 1.0453894693884935, "learning_rate": 2.3558878136791697e-07, "loss": 0.4112, "step": 4751 }, { "epoch": 4.560729716754681, "grad_norm": 0.7453365845367181, "learning_rate": 2.3457367235769468e-07, "loss": 0.5904, "step": 4752 }, { "epoch": 4.5616898703792605, "grad_norm": 1.1269632471461644, "learning_rate": 2.335607024964115e-07, "loss": 0.3913, "step": 4753 }, { "epoch": 4.56265002400384, "grad_norm": 1.0593364794655684, "learning_rate": 2.3254987223877868e-07, "loss": 0.3911, "step": 4754 }, { "epoch": 4.563610177628421, "grad_norm": 0.9835490800399789, "learning_rate": 2.315411820385477e-07, "loss": 0.4318, "step": 4755 }, { "epoch": 4.564570331253001, "grad_norm": 0.862050744005447, "learning_rate": 2.3053463234850637e-07, "loss": 0.4432, "step": 4756 }, { "epoch": 4.565530484877581, "grad_norm": 0.6530005906793727, "learning_rate": 2.295302236204866e-07, "loss": 0.49, "step": 4757 }, { "epoch": 4.56649063850216, "grad_norm": 1.0111380743716984, "learning_rate": 2.2852795630535661e-07, "loss": 0.3992, "step": 4758 }, { "epoch": 4.56745079212674, "grad_norm": 0.8856854104938091, "learning_rate": 2.2752783085302265e-07, "loss": 0.4359, "step": 4759 }, { "epoch": 4.56841094575132, "grad_norm": 0.9795166768831115, "learning_rate": 2.2652984771243002e-07, "loss": 0.5555, "step": 4760 }, { "epoch": 4.5693710993759, "grad_norm": 0.8864711760449256, "learning_rate": 2.2553400733156317e-07, "loss": 0.4162, "step": 4761 }, { "epoch": 4.5703312530004805, "grad_norm": 0.7452907065419622, "learning_rate": 2.2454031015744448e-07, "loss": 0.4973, "step": 4762 }, { "epoch": 4.57129140662506, "grad_norm": 1.417750038999514, "learning_rate": 2.2354875663613384e-07, "loss": 0.3524, "step": 4763 }, { "epoch": 4.57225156024964, "grad_norm": 0.9723447187607988, "learning_rate": 2.2255934721272908e-07, "loss": 0.5049, "step": 4764 }, { "epoch": 4.57321171387422, "grad_norm": 0.7009013499874682, "learning_rate": 2.2157208233136495e-07, "loss": 0.5614, "step": 4765 }, { "epoch": 4.5741718674988, "grad_norm": 0.8675953804799497, "learning_rate": 2.2058696243521417e-07, "loss": 0.5068, "step": 4766 }, { "epoch": 4.5751320211233795, "grad_norm": 0.627997662238186, "learning_rate": 2.196039879664874e-07, "loss": 0.5051, "step": 4767 }, { "epoch": 4.576092174747959, "grad_norm": 1.0935809198800226, "learning_rate": 2.186231593664312e-07, "loss": 0.5422, "step": 4768 }, { "epoch": 4.57705232837254, "grad_norm": 0.7779700531211013, "learning_rate": 2.1764447707532722e-07, "loss": 0.4924, "step": 4769 }, { "epoch": 4.57801248199712, "grad_norm": 1.1538294156998674, "learning_rate": 2.1666794153249792e-07, "loss": 0.5158, "step": 4770 }, { "epoch": 4.5789726356216995, "grad_norm": 0.8539535476552191, "learning_rate": 2.1569355317629937e-07, "loss": 0.4811, "step": 4771 }, { "epoch": 4.579932789246279, "grad_norm": 0.6968607635177251, "learning_rate": 2.1472131244412275e-07, "loss": 0.5038, "step": 4772 }, { "epoch": 4.580892942870859, "grad_norm": 1.0610245456351224, "learning_rate": 2.1375121977239676e-07, "loss": 0.5388, "step": 4773 }, { "epoch": 4.581853096495439, "grad_norm": 0.8742698605720799, "learning_rate": 2.127832755965864e-07, "loss": 0.4754, "step": 4774 }, { "epoch": 4.582813250120019, "grad_norm": 0.8559013009494933, "learning_rate": 2.118174803511913e-07, "loss": 0.4272, "step": 4775 }, { "epoch": 4.583773403744599, "grad_norm": 1.3452518658388009, "learning_rate": 2.108538344697464e-07, "loss": 0.5388, "step": 4776 }, { "epoch": 4.584733557369179, "grad_norm": 0.9649151793437266, "learning_rate": 2.0989233838482225e-07, "loss": 0.5226, "step": 4777 }, { "epoch": 4.585693710993759, "grad_norm": 0.7573996183402368, "learning_rate": 2.0893299252802367e-07, "loss": 0.5153, "step": 4778 }, { "epoch": 4.586653864618339, "grad_norm": 0.8689743808309157, "learning_rate": 2.0797579732999063e-07, "loss": 0.4723, "step": 4779 }, { "epoch": 4.587614018242919, "grad_norm": 1.0640685524128548, "learning_rate": 2.0702075322039884e-07, "loss": 0.4627, "step": 4780 }, { "epoch": 4.588574171867499, "grad_norm": 1.3929760071716122, "learning_rate": 2.0606786062795702e-07, "loss": 0.4566, "step": 4781 }, { "epoch": 4.589534325492079, "grad_norm": 1.0156275280030036, "learning_rate": 2.051171199804075e-07, "loss": 0.4254, "step": 4782 }, { "epoch": 4.590494479116659, "grad_norm": 0.7047959090420411, "learning_rate": 2.0416853170452945e-07, "loss": 0.4327, "step": 4783 }, { "epoch": 4.591454632741239, "grad_norm": 0.8658683800583026, "learning_rate": 2.032220962261322e-07, "loss": 0.4385, "step": 4784 }, { "epoch": 4.5924147863658185, "grad_norm": 0.951464739882769, "learning_rate": 2.022778139700615e-07, "loss": 0.4799, "step": 4785 }, { "epoch": 4.593374939990398, "grad_norm": 0.8162494267137287, "learning_rate": 2.0133568536019498e-07, "loss": 0.456, "step": 4786 }, { "epoch": 4.594335093614978, "grad_norm": 0.7471996937348034, "learning_rate": 2.0039571081944431e-07, "loss": 0.4652, "step": 4787 }, { "epoch": 4.595295247239559, "grad_norm": 0.7221690708302403, "learning_rate": 1.994578907697542e-07, "loss": 0.4658, "step": 4788 }, { "epoch": 4.5962554008641385, "grad_norm": 0.801898606661738, "learning_rate": 1.985222256321029e-07, "loss": 0.4509, "step": 4789 }, { "epoch": 4.597215554488718, "grad_norm": 0.816245329495932, "learning_rate": 1.9758871582649887e-07, "loss": 0.474, "step": 4790 }, { "epoch": 4.598175708113298, "grad_norm": 0.8064224691419776, "learning_rate": 1.9665736177198524e-07, "loss": 0.4455, "step": 4791 }, { "epoch": 4.599135861737878, "grad_norm": 0.7187701936022547, "learning_rate": 1.9572816388663706e-07, "loss": 0.5113, "step": 4792 }, { "epoch": 4.600096015362458, "grad_norm": 0.7918434667672134, "learning_rate": 1.9480112258756123e-07, "loss": 0.4803, "step": 4793 }, { "epoch": 4.6010561689870375, "grad_norm": 1.0181110635095563, "learning_rate": 1.9387623829089764e-07, "loss": 0.4705, "step": 4794 }, { "epoch": 4.602016322611618, "grad_norm": 1.2414394967531963, "learning_rate": 1.929535114118153e-07, "loss": 0.448, "step": 4795 }, { "epoch": 4.602976476236198, "grad_norm": 0.670875061462943, "learning_rate": 1.920329423645173e-07, "loss": 0.4704, "step": 4796 }, { "epoch": 4.603936629860778, "grad_norm": 1.4307654990160052, "learning_rate": 1.9111453156223704e-07, "loss": 0.4596, "step": 4797 }, { "epoch": 4.604896783485358, "grad_norm": 0.7316131519846413, "learning_rate": 1.901982794172391e-07, "loss": 0.4137, "step": 4798 }, { "epoch": 4.605856937109937, "grad_norm": 0.7911247270488225, "learning_rate": 1.8928418634082068e-07, "loss": 0.4606, "step": 4799 }, { "epoch": 4.606817090734517, "grad_norm": 0.9403559393725737, "learning_rate": 1.8837225274330572e-07, "loss": 0.3765, "step": 4800 }, { "epoch": 4.607777244359097, "grad_norm": 0.7818230989255862, "learning_rate": 1.87462479034054e-07, "loss": 0.3585, "step": 4801 }, { "epoch": 4.608737397983678, "grad_norm": 1.025377473821061, "learning_rate": 1.8655486562145275e-07, "loss": 0.4328, "step": 4802 }, { "epoch": 4.6096975516082574, "grad_norm": 0.8702840379870368, "learning_rate": 1.8564941291291883e-07, "loss": 0.4556, "step": 4803 }, { "epoch": 4.610657705232837, "grad_norm": 1.6587854042263128, "learning_rate": 1.8474612131490098e-07, "loss": 0.3195, "step": 4804 }, { "epoch": 4.611617858857417, "grad_norm": 0.9779633144045597, "learning_rate": 1.8384499123287713e-07, "loss": 0.5386, "step": 4805 }, { "epoch": 4.612578012481997, "grad_norm": 1.0699800432033595, "learning_rate": 1.8294602307135478e-07, "loss": 0.422, "step": 4806 }, { "epoch": 4.6135381661065775, "grad_norm": 0.8314449051426623, "learning_rate": 1.820492172338717e-07, "loss": 0.3809, "step": 4807 }, { "epoch": 4.614498319731157, "grad_norm": 1.0099640557617928, "learning_rate": 1.8115457412299363e-07, "loss": 0.4401, "step": 4808 }, { "epoch": 4.615458473355737, "grad_norm": 0.7971272183279201, "learning_rate": 1.802620941403166e-07, "loss": 0.4303, "step": 4809 }, { "epoch": 4.616418626980317, "grad_norm": 1.1211005707738178, "learning_rate": 1.7937177768646508e-07, "loss": 0.3734, "step": 4810 }, { "epoch": 4.617378780604897, "grad_norm": 1.1681605442886276, "learning_rate": 1.7848362516109274e-07, "loss": 0.4895, "step": 4811 }, { "epoch": 4.6183389342294765, "grad_norm": 1.154169482726129, "learning_rate": 1.775976369628829e-07, "loss": 0.4629, "step": 4812 }, { "epoch": 4.619299087854056, "grad_norm": 1.0177327264470901, "learning_rate": 1.7671381348954408e-07, "loss": 0.4809, "step": 4813 }, { "epoch": 4.620259241478637, "grad_norm": 1.1119722502596274, "learning_rate": 1.7583215513781725e-07, "loss": 0.5124, "step": 4814 }, { "epoch": 4.621219395103217, "grad_norm": 0.8817465365942875, "learning_rate": 1.749526623034681e-07, "loss": 0.5339, "step": 4815 }, { "epoch": 4.622179548727797, "grad_norm": 0.8993937859103783, "learning_rate": 1.7407533538129185e-07, "loss": 0.4784, "step": 4816 }, { "epoch": 4.623139702352376, "grad_norm": 0.9834057179015229, "learning_rate": 1.732001747651113e-07, "loss": 0.4486, "step": 4817 }, { "epoch": 4.624099855976956, "grad_norm": 1.2699068553881105, "learning_rate": 1.7232718084777666e-07, "loss": 0.4448, "step": 4818 }, { "epoch": 4.625060009601536, "grad_norm": 0.9273301122526332, "learning_rate": 1.7145635402116556e-07, "loss": 0.4315, "step": 4819 }, { "epoch": 4.626020163226116, "grad_norm": 0.7084125451866525, "learning_rate": 1.7058769467618364e-07, "loss": 0.5531, "step": 4820 }, { "epoch": 4.626980316850696, "grad_norm": 1.2439849160749008, "learning_rate": 1.697212032027612e-07, "loss": 0.4293, "step": 4821 }, { "epoch": 4.627940470475276, "grad_norm": 0.7688444094592808, "learning_rate": 1.688568799898582e-07, "loss": 0.5057, "step": 4822 }, { "epoch": 4.628900624099856, "grad_norm": 0.9098249578224078, "learning_rate": 1.6799472542545935e-07, "loss": 0.4924, "step": 4823 }, { "epoch": 4.629860777724436, "grad_norm": 0.8246723901757721, "learning_rate": 1.671347398965767e-07, "loss": 0.4292, "step": 4824 }, { "epoch": 4.630820931349016, "grad_norm": 0.8441050016379568, "learning_rate": 1.6627692378924975e-07, "loss": 0.5338, "step": 4825 }, { "epoch": 4.631781084973595, "grad_norm": 0.6692190880549667, "learning_rate": 1.6542127748854108e-07, "loss": 0.4345, "step": 4826 }, { "epoch": 4.632741238598175, "grad_norm": 0.899537019302587, "learning_rate": 1.6456780137854346e-07, "loss": 0.4217, "step": 4827 }, { "epoch": 4.633701392222756, "grad_norm": 0.8339174751135915, "learning_rate": 1.6371649584237093e-07, "loss": 0.4785, "step": 4828 }, { "epoch": 4.634661545847336, "grad_norm": 0.6120569462401307, "learning_rate": 1.6286736126216674e-07, "loss": 0.4477, "step": 4829 }, { "epoch": 4.6356216994719155, "grad_norm": 0.8104792812715427, "learning_rate": 1.6202039801909763e-07, "loss": 0.4281, "step": 4830 }, { "epoch": 4.636581853096495, "grad_norm": 1.0492401688172464, "learning_rate": 1.6117560649335674e-07, "loss": 0.4193, "step": 4831 }, { "epoch": 4.637542006721075, "grad_norm": 0.9421768578527981, "learning_rate": 1.6033298706416234e-07, "loss": 0.5416, "step": 4832 }, { "epoch": 4.638502160345656, "grad_norm": 1.2825478165877162, "learning_rate": 1.594925401097569e-07, "loss": 0.3725, "step": 4833 }, { "epoch": 4.639462313970236, "grad_norm": 0.8022527078984462, "learning_rate": 1.5865426600740807e-07, "loss": 0.4724, "step": 4834 }, { "epoch": 4.640422467594815, "grad_norm": 0.8090665959660093, "learning_rate": 1.5781816513340764e-07, "loss": 0.5481, "step": 4835 }, { "epoch": 4.641382621219395, "grad_norm": 0.7242875037607582, "learning_rate": 1.5698423786307314e-07, "loss": 0.4545, "step": 4836 }, { "epoch": 4.642342774843975, "grad_norm": 0.963493605252525, "learning_rate": 1.5615248457074516e-07, "loss": 0.5453, "step": 4837 }, { "epoch": 4.643302928468555, "grad_norm": 0.9008133697578701, "learning_rate": 1.5532290562978948e-07, "loss": 0.4022, "step": 4838 }, { "epoch": 4.6442630820931345, "grad_norm": 0.8377983733257375, "learning_rate": 1.5449550141259427e-07, "loss": 0.535, "step": 4839 }, { "epoch": 4.645223235717715, "grad_norm": 1.0638661218559817, "learning_rate": 1.5367027229057307e-07, "loss": 0.4231, "step": 4840 }, { "epoch": 4.646183389342295, "grad_norm": 0.89576212329215, "learning_rate": 1.5284721863416175e-07, "loss": 0.411, "step": 4841 }, { "epoch": 4.647143542966875, "grad_norm": 0.7412906028183777, "learning_rate": 1.5202634081282142e-07, "loss": 0.4785, "step": 4842 }, { "epoch": 4.648103696591455, "grad_norm": 0.9010992284443584, "learning_rate": 1.512076391950351e-07, "loss": 0.5237, "step": 4843 }, { "epoch": 4.649063850216034, "grad_norm": 0.9731586396770459, "learning_rate": 1.5039111414830721e-07, "loss": 0.5225, "step": 4844 }, { "epoch": 4.650024003840614, "grad_norm": 1.20433307265466, "learning_rate": 1.4957676603917005e-07, "loss": 0.4248, "step": 4845 }, { "epoch": 4.650984157465194, "grad_norm": 0.9454831299128075, "learning_rate": 1.4876459523317455e-07, "loss": 0.4331, "step": 4846 }, { "epoch": 4.651944311089775, "grad_norm": 0.6791484979123632, "learning_rate": 1.479546020948952e-07, "loss": 0.4837, "step": 4847 }, { "epoch": 4.6529044647143545, "grad_norm": 0.7268790237338438, "learning_rate": 1.4714678698792894e-07, "loss": 0.4869, "step": 4848 }, { "epoch": 4.653864618338934, "grad_norm": 1.1793053332024215, "learning_rate": 1.4634115027489627e-07, "loss": 0.43, "step": 4849 }, { "epoch": 4.654824771963514, "grad_norm": 1.0999726595379253, "learning_rate": 1.4553769231743842e-07, "loss": 0.4793, "step": 4850 }, { "epoch": 4.655784925588094, "grad_norm": 1.006445670928147, "learning_rate": 1.447364134762197e-07, "loss": 0.4723, "step": 4851 }, { "epoch": 4.656745079212674, "grad_norm": 1.2509823227053158, "learning_rate": 1.439373141109246e-07, "loss": 0.4944, "step": 4852 }, { "epoch": 4.6577052328372535, "grad_norm": 0.8321091046167617, "learning_rate": 1.4314039458026063e-07, "loss": 0.554, "step": 4853 }, { "epoch": 4.658665386461834, "grad_norm": 0.7592827853327345, "learning_rate": 1.4234565524195664e-07, "loss": 0.3538, "step": 4854 }, { "epoch": 4.659625540086414, "grad_norm": 0.8314149965890658, "learning_rate": 1.4155309645276283e-07, "loss": 0.5529, "step": 4855 }, { "epoch": 4.660585693710994, "grad_norm": 0.9064900376176223, "learning_rate": 1.407627185684507e-07, "loss": 0.4979, "step": 4856 }, { "epoch": 4.6615458473355735, "grad_norm": 0.8423366872169219, "learning_rate": 1.3997452194381034e-07, "loss": 0.4849, "step": 4857 }, { "epoch": 4.662506000960153, "grad_norm": 1.090179042027191, "learning_rate": 1.3918850693265763e-07, "loss": 0.3808, "step": 4858 }, { "epoch": 4.663466154584734, "grad_norm": 1.2139587000743401, "learning_rate": 1.3840467388782418e-07, "loss": 0.5113, "step": 4859 }, { "epoch": 4.664426308209314, "grad_norm": 1.0179056591934565, "learning_rate": 1.3762302316116527e-07, "loss": 0.3456, "step": 4860 }, { "epoch": 4.665386461833894, "grad_norm": 0.9385392500390293, "learning_rate": 1.3684355510355518e-07, "loss": 0.5034, "step": 4861 }, { "epoch": 4.666346615458473, "grad_norm": 1.0949546261049083, "learning_rate": 1.3606627006488905e-07, "loss": 0.4846, "step": 4862 }, { "epoch": 4.667306769083053, "grad_norm": 0.7887910194535214, "learning_rate": 1.3529116839408162e-07, "loss": 0.43, "step": 4863 }, { "epoch": 4.668266922707633, "grad_norm": 0.7581049973195865, "learning_rate": 1.3451825043906852e-07, "loss": 0.4495, "step": 4864 }, { "epoch": 4.669227076332213, "grad_norm": 0.8115233285612591, "learning_rate": 1.337475165468033e-07, "loss": 0.4476, "step": 4865 }, { "epoch": 4.6701872299567935, "grad_norm": 0.8932787009018343, "learning_rate": 1.3297896706326031e-07, "loss": 0.532, "step": 4866 }, { "epoch": 4.671147383581373, "grad_norm": 0.7759193460028937, "learning_rate": 1.3221260233343357e-07, "loss": 0.4909, "step": 4867 }, { "epoch": 4.672107537205953, "grad_norm": 0.8220709800287594, "learning_rate": 1.3144842270133573e-07, "loss": 0.4464, "step": 4868 }, { "epoch": 4.673067690830533, "grad_norm": 0.8297348583203794, "learning_rate": 1.3068642850999957e-07, "loss": 0.5163, "step": 4869 }, { "epoch": 4.674027844455113, "grad_norm": 0.7548278493955829, "learning_rate": 1.2992662010147428e-07, "loss": 0.5131, "step": 4870 }, { "epoch": 4.6749879980796925, "grad_norm": 1.2164716796085013, "learning_rate": 1.2916899781683257e-07, "loss": 0.4782, "step": 4871 }, { "epoch": 4.675948151704272, "grad_norm": 0.8144337545067792, "learning_rate": 1.2841356199616072e-07, "loss": 0.4012, "step": 4872 }, { "epoch": 4.676908305328853, "grad_norm": 1.0180538103456755, "learning_rate": 1.276603129785664e-07, "loss": 0.5042, "step": 4873 }, { "epoch": 4.677868458953433, "grad_norm": 0.7599765291210797, "learning_rate": 1.2690925110217523e-07, "loss": 0.3727, "step": 4874 }, { "epoch": 4.6788286125780125, "grad_norm": 0.9814081922728302, "learning_rate": 1.2616037670413084e-07, "loss": 0.5141, "step": 4875 }, { "epoch": 4.679788766202592, "grad_norm": 1.0512830186217923, "learning_rate": 1.254136901205949e-07, "loss": 0.548, "step": 4876 }, { "epoch": 4.680748919827172, "grad_norm": 1.046537019444709, "learning_rate": 1.2466919168674763e-07, "loss": 0.5059, "step": 4877 }, { "epoch": 4.681709073451752, "grad_norm": 0.8335674644115226, "learning_rate": 1.2392688173678503e-07, "loss": 0.4879, "step": 4878 }, { "epoch": 4.682669227076333, "grad_norm": 0.7256321358525952, "learning_rate": 1.2318676060392333e-07, "loss": 0.5871, "step": 4879 }, { "epoch": 4.683629380700912, "grad_norm": 0.8098733760424985, "learning_rate": 1.2244882862039398e-07, "loss": 0.4498, "step": 4880 }, { "epoch": 4.684589534325492, "grad_norm": 0.975398594613558, "learning_rate": 1.217130861174476e-07, "loss": 0.4567, "step": 4881 }, { "epoch": 4.685549687950072, "grad_norm": 1.1508505671833837, "learning_rate": 1.209795334253516e-07, "loss": 0.4565, "step": 4882 }, { "epoch": 4.686509841574652, "grad_norm": 0.8521625945407955, "learning_rate": 1.2024817087338826e-07, "loss": 0.5399, "step": 4883 }, { "epoch": 4.687469995199232, "grad_norm": 0.7495190177596657, "learning_rate": 1.1951899878985983e-07, "loss": 0.4602, "step": 4884 }, { "epoch": 4.688430148823812, "grad_norm": 1.1361208287020443, "learning_rate": 1.1879201750208291e-07, "loss": 0.4002, "step": 4885 }, { "epoch": 4.689390302448392, "grad_norm": 0.7636389697218643, "learning_rate": 1.1806722733639253e-07, "loss": 0.5023, "step": 4886 }, { "epoch": 4.690350456072972, "grad_norm": 0.8195393899029699, "learning_rate": 1.17344628618139e-07, "loss": 0.4905, "step": 4887 }, { "epoch": 4.691310609697552, "grad_norm": 1.2016732057541688, "learning_rate": 1.1662422167168841e-07, "loss": 0.4401, "step": 4888 }, { "epoch": 4.692270763322131, "grad_norm": 1.0877526380171023, "learning_rate": 1.159060068204243e-07, "loss": 0.483, "step": 4889 }, { "epoch": 4.693230916946711, "grad_norm": 0.5717126680612578, "learning_rate": 1.1518998438674656e-07, "loss": 0.397, "step": 4890 }, { "epoch": 4.694191070571291, "grad_norm": 1.1617459555308653, "learning_rate": 1.1447615469206863e-07, "loss": 0.432, "step": 4891 }, { "epoch": 4.695151224195872, "grad_norm": 1.3865550484465146, "learning_rate": 1.1376451805682143e-07, "loss": 0.4427, "step": 4892 }, { "epoch": 4.6961113778204515, "grad_norm": 0.7470770090734316, "learning_rate": 1.1305507480045108e-07, "loss": 0.4626, "step": 4893 }, { "epoch": 4.697071531445031, "grad_norm": 0.8550251815977783, "learning_rate": 1.1234782524141896e-07, "loss": 0.4134, "step": 4894 }, { "epoch": 4.698031685069611, "grad_norm": 0.6526075850206136, "learning_rate": 1.1164276969720278e-07, "loss": 0.4015, "step": 4895 }, { "epoch": 4.698991838694191, "grad_norm": 0.7095442782244794, "learning_rate": 1.109399084842927e-07, "loss": 0.3861, "step": 4896 }, { "epoch": 4.699951992318771, "grad_norm": 0.7876407454395872, "learning_rate": 1.1023924191819691e-07, "loss": 0.4731, "step": 4897 }, { "epoch": 4.7009121459433505, "grad_norm": 0.8449836354241523, "learning_rate": 1.0954077031343658e-07, "loss": 0.3474, "step": 4898 }, { "epoch": 4.701872299567931, "grad_norm": 0.8331767705152524, "learning_rate": 1.0884449398354868e-07, "loss": 0.4685, "step": 4899 }, { "epoch": 4.702832453192511, "grad_norm": 0.7040274001003756, "learning_rate": 1.0815041324108378e-07, "loss": 0.4695, "step": 4900 }, { "epoch": 4.703792606817091, "grad_norm": 0.9883042896346539, "learning_rate": 1.0745852839760706e-07, "loss": 0.4571, "step": 4901 }, { "epoch": 4.704752760441671, "grad_norm": 0.753928293364501, "learning_rate": 1.0676883976369956e-07, "loss": 0.4154, "step": 4902 }, { "epoch": 4.70571291406625, "grad_norm": 0.9874622194023811, "learning_rate": 1.0608134764895306e-07, "loss": 0.4619, "step": 4903 }, { "epoch": 4.70667306769083, "grad_norm": 0.9492837912644471, "learning_rate": 1.053960523619768e-07, "loss": 0.5318, "step": 4904 }, { "epoch": 4.707633221315411, "grad_norm": 1.018686079057083, "learning_rate": 1.0471295421039251e-07, "loss": 0.4202, "step": 4905 }, { "epoch": 4.708593374939991, "grad_norm": 0.6522744185019053, "learning_rate": 1.040320535008349e-07, "loss": 0.4488, "step": 4906 }, { "epoch": 4.70955352856457, "grad_norm": 0.763241653868046, "learning_rate": 1.0335335053895391e-07, "loss": 0.482, "step": 4907 }, { "epoch": 4.71051368218915, "grad_norm": 0.8727960121470273, "learning_rate": 1.0267684562941194e-07, "loss": 0.4615, "step": 4908 }, { "epoch": 4.71147383581373, "grad_norm": 0.7931149899216856, "learning_rate": 1.0200253907588387e-07, "loss": 0.5043, "step": 4909 }, { "epoch": 4.71243398943831, "grad_norm": 0.8344362035186265, "learning_rate": 1.0133043118105923e-07, "loss": 0.4737, "step": 4910 }, { "epoch": 4.7133941430628905, "grad_norm": 0.9529856292957075, "learning_rate": 1.0066052224664058e-07, "loss": 0.5199, "step": 4911 }, { "epoch": 4.71435429668747, "grad_norm": 0.9485902711000596, "learning_rate": 9.999281257334181e-08, "loss": 0.4725, "step": 4912 }, { "epoch": 4.71531445031205, "grad_norm": 0.9096895179745177, "learning_rate": 9.932730246089262e-08, "loss": 0.4303, "step": 4913 }, { "epoch": 4.71627460393663, "grad_norm": 1.0979497076948672, "learning_rate": 9.86639922080307e-08, "loss": 0.5197, "step": 4914 }, { "epoch": 4.71723475756121, "grad_norm": 1.1200194773528456, "learning_rate": 9.800288211251119e-08, "loss": 0.4455, "step": 4915 }, { "epoch": 4.7181949111857895, "grad_norm": 0.7687186621584152, "learning_rate": 9.734397247109784e-08, "loss": 0.4995, "step": 4916 }, { "epoch": 4.719155064810369, "grad_norm": 0.8991970929925109, "learning_rate": 9.668726357956903e-08, "loss": 0.4332, "step": 4917 }, { "epoch": 4.72011521843495, "grad_norm": 1.2073743980947125, "learning_rate": 9.603275573271397e-08, "loss": 0.4724, "step": 4918 }, { "epoch": 4.72107537205953, "grad_norm": 0.7438596053863795, "learning_rate": 9.538044922433432e-08, "loss": 0.4337, "step": 4919 }, { "epoch": 4.72203552568411, "grad_norm": 0.6395774007466823, "learning_rate": 9.47303443472436e-08, "loss": 0.4591, "step": 4920 }, { "epoch": 4.722995679308689, "grad_norm": 1.440249732921376, "learning_rate": 9.408244139326728e-08, "loss": 0.4346, "step": 4921 }, { "epoch": 4.723955832933269, "grad_norm": 1.2055193304195897, "learning_rate": 9.34367406532416e-08, "loss": 0.5024, "step": 4922 }, { "epoch": 4.724915986557849, "grad_norm": 0.6827219336006264, "learning_rate": 9.279324241701415e-08, "loss": 0.5337, "step": 4923 }, { "epoch": 4.725876140182429, "grad_norm": 0.96065917193701, "learning_rate": 9.215194697344554e-08, "loss": 0.4652, "step": 4924 }, { "epoch": 4.726836293807009, "grad_norm": 0.8704346688122111, "learning_rate": 9.151285461040604e-08, "loss": 0.445, "step": 4925 }, { "epoch": 4.727796447431589, "grad_norm": 1.3956580921768234, "learning_rate": 9.08759656147773e-08, "loss": 0.4424, "step": 4926 }, { "epoch": 4.728756601056169, "grad_norm": 1.1026847843787726, "learning_rate": 9.024128027245171e-08, "loss": 0.4485, "step": 4927 }, { "epoch": 4.729716754680749, "grad_norm": 0.7413470468371381, "learning_rate": 8.960879886833307e-08, "loss": 0.4256, "step": 4928 }, { "epoch": 4.730676908305329, "grad_norm": 1.0728969069215688, "learning_rate": 8.897852168633537e-08, "loss": 0.3935, "step": 4929 }, { "epoch": 4.731637061929908, "grad_norm": 0.971436916390455, "learning_rate": 8.835044900938339e-08, "loss": 0.3743, "step": 4930 }, { "epoch": 4.732597215554489, "grad_norm": 0.7668247989226016, "learning_rate": 8.772458111941273e-08, "loss": 0.4554, "step": 4931 }, { "epoch": 4.733557369179069, "grad_norm": 1.0026370581101243, "learning_rate": 8.7100918297367e-08, "loss": 0.4761, "step": 4932 }, { "epoch": 4.734517522803649, "grad_norm": 1.2871470762233315, "learning_rate": 8.647946082320446e-08, "loss": 0.4906, "step": 4933 }, { "epoch": 4.7354776764282285, "grad_norm": 0.9164257718866419, "learning_rate": 8.586020897588864e-08, "loss": 0.5273, "step": 4934 }, { "epoch": 4.736437830052808, "grad_norm": 1.0342485122084313, "learning_rate": 8.524316303339609e-08, "loss": 0.5496, "step": 4935 }, { "epoch": 4.737397983677388, "grad_norm": 0.8055252819736493, "learning_rate": 8.462832327271131e-08, "loss": 0.4136, "step": 4936 }, { "epoch": 4.738358137301969, "grad_norm": 0.6260904702042229, "learning_rate": 8.401568996983022e-08, "loss": 0.4529, "step": 4937 }, { "epoch": 4.7393182909265485, "grad_norm": 1.0122419265479339, "learning_rate": 8.340526339975674e-08, "loss": 0.3929, "step": 4938 }, { "epoch": 4.740278444551128, "grad_norm": 0.7243647483213268, "learning_rate": 8.279704383650555e-08, "loss": 0.4646, "step": 4939 }, { "epoch": 4.741238598175708, "grad_norm": 0.7739978886129899, "learning_rate": 8.219103155309938e-08, "loss": 0.5041, "step": 4940 }, { "epoch": 4.742198751800288, "grad_norm": 0.8954837999687718, "learning_rate": 8.158722682157005e-08, "loss": 0.4303, "step": 4941 }, { "epoch": 4.743158905424868, "grad_norm": 0.6969343750344611, "learning_rate": 8.098562991296022e-08, "loss": 0.409, "step": 4942 }, { "epoch": 4.7441190590494475, "grad_norm": 1.2047577015961035, "learning_rate": 8.03862410973194e-08, "loss": 0.421, "step": 4943 }, { "epoch": 4.745079212674028, "grad_norm": 1.0602052661356354, "learning_rate": 7.978906064370739e-08, "loss": 0.3825, "step": 4944 }, { "epoch": 4.746039366298608, "grad_norm": 1.099892159105913, "learning_rate": 7.919408882019142e-08, "loss": 0.3828, "step": 4945 }, { "epoch": 4.746999519923188, "grad_norm": 0.942143137831873, "learning_rate": 7.86013258938495e-08, "loss": 0.4305, "step": 4946 }, { "epoch": 4.747959673547768, "grad_norm": 0.8233466522691802, "learning_rate": 7.801077213076491e-08, "loss": 0.4423, "step": 4947 }, { "epoch": 4.748919827172347, "grad_norm": 0.8903522742447062, "learning_rate": 7.742242779603115e-08, "loss": 0.4355, "step": 4948 }, { "epoch": 4.749879980796927, "grad_norm": 1.2230470470137618, "learning_rate": 7.683629315375086e-08, "loss": 0.4907, "step": 4949 }, { "epoch": 4.750840134421507, "grad_norm": 0.9063053684032998, "learning_rate": 7.625236846703243e-08, "loss": 0.4888, "step": 4950 }, { "epoch": 4.751800288046088, "grad_norm": 0.8288330487375294, "learning_rate": 7.5670653997994e-08, "loss": 0.5051, "step": 4951 }, { "epoch": 4.7527604416706675, "grad_norm": 0.8945450368295222, "learning_rate": 7.509115000776112e-08, "loss": 0.5505, "step": 4952 }, { "epoch": 4.753720595295247, "grad_norm": 0.6440593755754892, "learning_rate": 7.451385675646627e-08, "loss": 0.4216, "step": 4953 }, { "epoch": 4.754680748919827, "grad_norm": 0.780503327586162, "learning_rate": 7.393877450325105e-08, "loss": 0.5454, "step": 4954 }, { "epoch": 4.755640902544407, "grad_norm": 0.8542908015835958, "learning_rate": 7.336590350626282e-08, "loss": 0.4908, "step": 4955 }, { "epoch": 4.756601056168987, "grad_norm": 0.8662112716900428, "learning_rate": 7.279524402265758e-08, "loss": 0.5418, "step": 4956 }, { "epoch": 4.757561209793567, "grad_norm": 0.7961324892545846, "learning_rate": 7.222679630859875e-08, "loss": 0.4151, "step": 4957 }, { "epoch": 4.758521363418147, "grad_norm": 1.0207487602706176, "learning_rate": 7.166056061925609e-08, "loss": 0.4724, "step": 4958 }, { "epoch": 4.759481517042727, "grad_norm": 0.8175293666538429, "learning_rate": 7.109653720880682e-08, "loss": 0.4653, "step": 4959 }, { "epoch": 4.760441670667307, "grad_norm": 0.7194832230993446, "learning_rate": 7.053472633043457e-08, "loss": 0.438, "step": 4960 }, { "epoch": 4.7614018242918865, "grad_norm": 1.4133328519477804, "learning_rate": 6.99751282363309e-08, "loss": 0.4708, "step": 4961 }, { "epoch": 4.762361977916466, "grad_norm": 0.712995913876599, "learning_rate": 6.941774317769268e-08, "loss": 0.477, "step": 4962 }, { "epoch": 4.763322131541047, "grad_norm": 0.8128832363835004, "learning_rate": 6.886257140472475e-08, "loss": 0.4938, "step": 4963 }, { "epoch": 4.764282285165627, "grad_norm": 1.2852150201205494, "learning_rate": 6.830961316663775e-08, "loss": 0.3657, "step": 4964 }, { "epoch": 4.765242438790207, "grad_norm": 0.7183685885760241, "learning_rate": 6.775886871164872e-08, "loss": 0.427, "step": 4965 }, { "epoch": 4.766202592414786, "grad_norm": 1.0682770395265626, "learning_rate": 6.721033828698098e-08, "loss": 0.4065, "step": 4966 }, { "epoch": 4.767162746039366, "grad_norm": 0.8615966520645691, "learning_rate": 6.666402213886313e-08, "loss": 0.4982, "step": 4967 }, { "epoch": 4.768122899663946, "grad_norm": 0.7945289732589843, "learning_rate": 6.61199205125318e-08, "loss": 0.3699, "step": 4968 }, { "epoch": 4.769083053288526, "grad_norm": 0.9300363254839001, "learning_rate": 6.557803365222714e-08, "loss": 0.3987, "step": 4969 }, { "epoch": 4.7700432069131065, "grad_norm": 0.8302341523969474, "learning_rate": 6.503836180119849e-08, "loss": 0.4736, "step": 4970 }, { "epoch": 4.771003360537686, "grad_norm": 0.6751469388517514, "learning_rate": 6.45009052016965e-08, "loss": 0.5062, "step": 4971 }, { "epoch": 4.771963514162266, "grad_norm": 1.216995934426271, "learning_rate": 6.396566409498039e-08, "loss": 0.4288, "step": 4972 }, { "epoch": 4.772923667786846, "grad_norm": 0.7834064203335489, "learning_rate": 6.343263872131467e-08, "loss": 0.5749, "step": 4973 }, { "epoch": 4.773883821411426, "grad_norm": 0.9073177559408646, "learning_rate": 6.290182931996846e-08, "loss": 0.4522, "step": 4974 }, { "epoch": 4.774843975036005, "grad_norm": 0.7945761850621127, "learning_rate": 6.237323612921675e-08, "loss": 0.4916, "step": 4975 }, { "epoch": 4.775804128660585, "grad_norm": 0.9273178925996909, "learning_rate": 6.184685938633861e-08, "loss": 0.4012, "step": 4976 }, { "epoch": 4.776764282285166, "grad_norm": 0.8526532547218346, "learning_rate": 6.132269932761947e-08, "loss": 0.4585, "step": 4977 }, { "epoch": 4.777724435909746, "grad_norm": 0.6878819032989627, "learning_rate": 6.080075618834891e-08, "loss": 0.4463, "step": 4978 }, { "epoch": 4.7786845895343255, "grad_norm": 0.9149694660708082, "learning_rate": 6.028103020282228e-08, "loss": 0.4939, "step": 4979 }, { "epoch": 4.779644743158905, "grad_norm": 1.0646330953923748, "learning_rate": 5.976352160433796e-08, "loss": 0.4371, "step": 4980 }, { "epoch": 4.780604896783485, "grad_norm": 0.8463300625111078, "learning_rate": 5.9248230625200686e-08, "loss": 0.4554, "step": 4981 }, { "epoch": 4.781565050408066, "grad_norm": 1.0145093270848597, "learning_rate": 5.873515749671932e-08, "loss": 0.4586, "step": 4982 }, { "epoch": 4.782525204032646, "grad_norm": 0.5896430000687316, "learning_rate": 5.822430244920685e-08, "loss": 0.4016, "step": 4983 }, { "epoch": 4.783485357657225, "grad_norm": 1.067340698318332, "learning_rate": 5.771566571197984e-08, "loss": 0.4674, "step": 4984 }, { "epoch": 4.784445511281805, "grad_norm": 0.9808813721457339, "learning_rate": 5.72092475133601e-08, "loss": 0.4877, "step": 4985 }, { "epoch": 4.785405664906385, "grad_norm": 0.8746287649517445, "learning_rate": 5.670504808067412e-08, "loss": 0.4497, "step": 4986 }, { "epoch": 4.786365818530965, "grad_norm": 1.328033589256457, "learning_rate": 5.620306764025085e-08, "loss": 0.5184, "step": 4987 }, { "epoch": 4.787325972155545, "grad_norm": 0.8457112834157048, "learning_rate": 5.57033064174245e-08, "loss": 0.48, "step": 4988 }, { "epoch": 4.788286125780125, "grad_norm": 1.1627671118196217, "learning_rate": 5.5205764636532266e-08, "loss": 0.3387, "step": 4989 }, { "epoch": 4.789246279404705, "grad_norm": 0.9960918332205848, "learning_rate": 5.471044252091551e-08, "loss": 0.5623, "step": 4990 }, { "epoch": 4.790206433029285, "grad_norm": 0.931109941940499, "learning_rate": 5.421734029291858e-08, "loss": 0.4617, "step": 4991 }, { "epoch": 4.791166586653865, "grad_norm": 0.7444235059789088, "learning_rate": 5.372645817389055e-08, "loss": 0.5053, "step": 4992 }, { "epoch": 4.792126740278444, "grad_norm": 1.2532952858141375, "learning_rate": 5.323779638418236e-08, "loss": 0.5238, "step": 4993 }, { "epoch": 4.793086893903024, "grad_norm": 0.7776994766144231, "learning_rate": 5.275135514314911e-08, "loss": 0.4497, "step": 4994 }, { "epoch": 4.794047047527604, "grad_norm": 1.2647076332598313, "learning_rate": 5.226713466915001e-08, "loss": 0.4527, "step": 4995 }, { "epoch": 4.795007201152185, "grad_norm": 0.9198154647395577, "learning_rate": 5.1785135179545643e-08, "loss": 0.3747, "step": 4996 }, { "epoch": 4.7959673547767645, "grad_norm": 0.703861896989188, "learning_rate": 5.130535689070071e-08, "loss": 0.4394, "step": 4997 }, { "epoch": 4.796927508401344, "grad_norm": 0.8778920849151423, "learning_rate": 5.0827800017982376e-08, "loss": 0.3921, "step": 4998 }, { "epoch": 4.797887662025924, "grad_norm": 0.9537724308995854, "learning_rate": 5.035246477576083e-08, "loss": 0.5222, "step": 4999 }, { "epoch": 4.798847815650504, "grad_norm": 0.7141897420500998, "learning_rate": 4.9879351377408734e-08, "loss": 0.4481, "step": 5000 }, { "epoch": 4.799807969275084, "grad_norm": 0.9038259620205664, "learning_rate": 4.940846003530231e-08, "loss": 0.504, "step": 5001 }, { "epoch": 4.8007681228996635, "grad_norm": 0.8273084597854825, "learning_rate": 4.893979096081969e-08, "loss": 0.5391, "step": 5002 }, { "epoch": 4.801728276524244, "grad_norm": 0.7134386809127866, "learning_rate": 4.847334436434037e-08, "loss": 0.4401, "step": 5003 }, { "epoch": 4.802688430148824, "grad_norm": 1.500177434674667, "learning_rate": 4.800912045524797e-08, "loss": 0.5436, "step": 5004 }, { "epoch": 4.803648583773404, "grad_norm": 0.7620735553146109, "learning_rate": 4.7547119441927466e-08, "loss": 0.4199, "step": 5005 }, { "epoch": 4.8046087373979836, "grad_norm": 0.7817727237495927, "learning_rate": 4.7087341531766864e-08, "loss": 0.503, "step": 5006 }, { "epoch": 4.805568891022563, "grad_norm": 0.7371427891230952, "learning_rate": 4.662978693115439e-08, "loss": 0.4259, "step": 5007 }, { "epoch": 4.806529044647144, "grad_norm": 1.0728535498263434, "learning_rate": 4.617445584548186e-08, "loss": 0.503, "step": 5008 }, { "epoch": 4.807489198271724, "grad_norm": 0.9186906313616544, "learning_rate": 4.5721348479143e-08, "loss": 0.4203, "step": 5009 }, { "epoch": 4.808449351896304, "grad_norm": 1.3180225923268514, "learning_rate": 4.527046503553234e-08, "loss": 0.3971, "step": 5010 }, { "epoch": 4.809409505520883, "grad_norm": 0.7699327305477651, "learning_rate": 4.4821805717046864e-08, "loss": 0.4458, "step": 5011 }, { "epoch": 4.810369659145463, "grad_norm": 1.4094145002585667, "learning_rate": 4.437537072508491e-08, "loss": 0.3992, "step": 5012 }, { "epoch": 4.811329812770043, "grad_norm": 1.1420470332687895, "learning_rate": 4.393116026004618e-08, "loss": 0.4055, "step": 5013 }, { "epoch": 4.812289966394623, "grad_norm": 1.7361045132152544, "learning_rate": 4.3489174521332833e-08, "loss": 0.4695, "step": 5014 }, { "epoch": 4.8132501200192035, "grad_norm": 1.1257705802333253, "learning_rate": 4.304941370734616e-08, "loss": 0.403, "step": 5015 }, { "epoch": 4.814210273643783, "grad_norm": 0.8033591178452424, "learning_rate": 4.261187801549105e-08, "loss": 0.5026, "step": 5016 }, { "epoch": 4.815170427268363, "grad_norm": 1.0434338010436313, "learning_rate": 4.217656764217204e-08, "loss": 0.4095, "step": 5017 }, { "epoch": 4.816130580892943, "grad_norm": 1.3588574124231256, "learning_rate": 4.1743482782796166e-08, "loss": 0.4496, "step": 5018 }, { "epoch": 4.817090734517523, "grad_norm": 0.9749891957753111, "learning_rate": 4.1312623631770134e-08, "loss": 0.4315, "step": 5019 }, { "epoch": 4.8180508881421025, "grad_norm": 0.8921856472075009, "learning_rate": 4.08839903825009e-08, "loss": 0.5096, "step": 5020 }, { "epoch": 4.819011041766682, "grad_norm": 0.9092793986976705, "learning_rate": 4.045758322739957e-08, "loss": 0.4727, "step": 5021 }, { "epoch": 4.819971195391263, "grad_norm": 0.8132582425339893, "learning_rate": 4.003340235787412e-08, "loss": 0.5081, "step": 5022 }, { "epoch": 4.820931349015843, "grad_norm": 1.2176344913248394, "learning_rate": 3.961144796433502e-08, "loss": 0.4286, "step": 5023 }, { "epoch": 4.8218915026404225, "grad_norm": 0.7391120856514499, "learning_rate": 3.919172023619355e-08, "loss": 0.4468, "step": 5024 }, { "epoch": 4.822851656265002, "grad_norm": 0.907006660933719, "learning_rate": 3.8774219361860655e-08, "loss": 0.4707, "step": 5025 }, { "epoch": 4.823811809889582, "grad_norm": 0.9773605199709892, "learning_rate": 3.835894552874808e-08, "loss": 0.4994, "step": 5026 }, { "epoch": 4.824771963514162, "grad_norm": 1.0977660091685522, "learning_rate": 3.794589892326839e-08, "loss": 0.4225, "step": 5027 }, { "epoch": 4.825732117138742, "grad_norm": 0.7806541712141686, "learning_rate": 3.75350797308327e-08, "loss": 0.5221, "step": 5028 }, { "epoch": 4.826692270763322, "grad_norm": 0.9939879606367326, "learning_rate": 3.712648813585407e-08, "loss": 0.4529, "step": 5029 }, { "epoch": 4.827652424387902, "grad_norm": 0.8619201435212147, "learning_rate": 3.672012432174465e-08, "loss": 0.4873, "step": 5030 }, { "epoch": 4.828612578012482, "grad_norm": 0.7836767631337458, "learning_rate": 3.631598847091688e-08, "loss": 0.5304, "step": 5031 }, { "epoch": 4.829572731637062, "grad_norm": 0.7654112974306962, "learning_rate": 3.591408076478342e-08, "loss": 0.4794, "step": 5032 }, { "epoch": 4.830532885261642, "grad_norm": 0.8146179836754643, "learning_rate": 3.5514401383755505e-08, "loss": 0.4906, "step": 5033 }, { "epoch": 4.831493038886222, "grad_norm": 0.9863552230355864, "learning_rate": 3.511695050724573e-08, "loss": 0.5116, "step": 5034 }, { "epoch": 4.832453192510802, "grad_norm": 1.1027628157403395, "learning_rate": 3.4721728313664715e-08, "loss": 0.4584, "step": 5035 }, { "epoch": 4.833413346135382, "grad_norm": 1.2507132361259323, "learning_rate": 3.432873498042444e-08, "loss": 0.4047, "step": 5036 }, { "epoch": 4.834373499759962, "grad_norm": 1.0633479299683923, "learning_rate": 3.3937970683934893e-08, "loss": 0.4031, "step": 5037 }, { "epoch": 4.8353336533845415, "grad_norm": 1.0700245385318792, "learning_rate": 3.354943559960577e-08, "loss": 0.4765, "step": 5038 }, { "epoch": 4.836293807009121, "grad_norm": 0.7896934034489211, "learning_rate": 3.3163129901847e-08, "loss": 0.5169, "step": 5039 }, { "epoch": 4.837253960633701, "grad_norm": 0.9685052022657994, "learning_rate": 3.277905376406654e-08, "loss": 0.5015, "step": 5040 }, { "epoch": 4.838214114258282, "grad_norm": 0.7588758726927429, "learning_rate": 3.2397207358672024e-08, "loss": 0.4989, "step": 5041 }, { "epoch": 4.8391742678828615, "grad_norm": 0.7420659483925464, "learning_rate": 3.201759085707079e-08, "loss": 0.4935, "step": 5042 }, { "epoch": 4.840134421507441, "grad_norm": 0.9219864750705471, "learning_rate": 3.164020442966764e-08, "loss": 0.5079, "step": 5043 }, { "epoch": 4.841094575132021, "grad_norm": 1.0341302517030253, "learning_rate": 3.126504824586818e-08, "loss": 0.514, "step": 5044 }, { "epoch": 4.842054728756601, "grad_norm": 0.8930390721450207, "learning_rate": 3.0892122474076025e-08, "loss": 0.4624, "step": 5045 }, { "epoch": 4.843014882381181, "grad_norm": 1.0093298894463756, "learning_rate": 3.0521427281693386e-08, "loss": 0.3957, "step": 5046 }, { "epoch": 4.8439750360057605, "grad_norm": 0.7645366314246456, "learning_rate": 3.015296283512048e-08, "loss": 0.3619, "step": 5047 }, { "epoch": 4.844935189630341, "grad_norm": 0.6796125755640886, "learning_rate": 2.9786729299758343e-08, "loss": 0.4598, "step": 5048 }, { "epoch": 4.845895343254921, "grad_norm": 0.9909166254150091, "learning_rate": 2.942272684000491e-08, "loss": 0.4529, "step": 5049 }, { "epoch": 4.846855496879501, "grad_norm": 0.8893659642801445, "learning_rate": 2.9060955619256702e-08, "loss": 0.3896, "step": 5050 }, { "epoch": 4.847815650504081, "grad_norm": 0.8227575591024602, "learning_rate": 2.8701415799908815e-08, "loss": 0.4899, "step": 5051 }, { "epoch": 4.84877580412866, "grad_norm": 0.7339268781443058, "learning_rate": 2.8344107543356035e-08, "loss": 0.415, "step": 5052 }, { "epoch": 4.84973595775324, "grad_norm": 0.7203616264104085, "learning_rate": 2.7989031009988398e-08, "loss": 0.4556, "step": 5053 }, { "epoch": 4.85069611137782, "grad_norm": 1.28821020732297, "learning_rate": 2.763618635919729e-08, "loss": 0.4012, "step": 5054 }, { "epoch": 4.851656265002401, "grad_norm": 0.8718252818408424, "learning_rate": 2.7285573749370465e-08, "loss": 0.4121, "step": 5055 }, { "epoch": 4.8526164186269805, "grad_norm": 1.1673036117963758, "learning_rate": 2.693719333789424e-08, "loss": 0.4371, "step": 5056 }, { "epoch": 4.85357657225156, "grad_norm": 1.2616066142477538, "learning_rate": 2.6591045281152418e-08, "loss": 0.4536, "step": 5057 }, { "epoch": 4.85453672587614, "grad_norm": 1.2111572938247852, "learning_rate": 2.6247129734528474e-08, "loss": 0.4294, "step": 5058 }, { "epoch": 4.85549687950072, "grad_norm": 0.9908621857228271, "learning_rate": 2.5905446852400594e-08, "loss": 0.341, "step": 5059 }, { "epoch": 4.8564570331253005, "grad_norm": 1.1040753667463135, "learning_rate": 2.5565996788148306e-08, "loss": 0.3958, "step": 5060 }, { "epoch": 4.85741718674988, "grad_norm": 0.932226107181068, "learning_rate": 2.5228779694145832e-08, "loss": 0.4213, "step": 5061 }, { "epoch": 4.85837734037446, "grad_norm": 0.7238169853536627, "learning_rate": 2.4893795721766534e-08, "loss": 0.513, "step": 5062 }, { "epoch": 4.85933749399904, "grad_norm": 1.629384553685994, "learning_rate": 2.4561045021382346e-08, "loss": 0.4476, "step": 5063 }, { "epoch": 4.86029764762362, "grad_norm": 0.7731372941436756, "learning_rate": 2.4230527742359344e-08, "loss": 0.4875, "step": 5064 }, { "epoch": 4.8612578012481995, "grad_norm": 1.2335549626091928, "learning_rate": 2.3902244033065512e-08, "loss": 0.6198, "step": 5065 }, { "epoch": 4.862217954872779, "grad_norm": 0.8533918014496968, "learning_rate": 2.3576194040861866e-08, "loss": 0.4287, "step": 5066 }, { "epoch": 4.86317810849736, "grad_norm": 0.7820208438608377, "learning_rate": 2.3252377912110214e-08, "loss": 0.5585, "step": 5067 }, { "epoch": 4.86413826212194, "grad_norm": 1.1427154284894614, "learning_rate": 2.2930795792167615e-08, "loss": 0.4262, "step": 5068 }, { "epoch": 4.86509841574652, "grad_norm": 0.7767144541654959, "learning_rate": 2.2611447825389154e-08, "loss": 0.6266, "step": 5069 }, { "epoch": 4.866058569371099, "grad_norm": 1.1085525627511914, "learning_rate": 2.2294334155125718e-08, "loss": 0.4888, "step": 5070 }, { "epoch": 4.867018722995679, "grad_norm": 1.05474045892835, "learning_rate": 2.1979454923727327e-08, "loss": 0.4333, "step": 5071 }, { "epoch": 4.867978876620259, "grad_norm": 0.9944282603641644, "learning_rate": 2.1666810272539806e-08, "loss": 0.4382, "step": 5072 }, { "epoch": 4.868939030244839, "grad_norm": 1.04992326479202, "learning_rate": 2.1356400341905337e-08, "loss": 0.552, "step": 5073 }, { "epoch": 4.8698991838694194, "grad_norm": 0.7346988881837826, "learning_rate": 2.1048225271164125e-08, "loss": 0.4651, "step": 5074 }, { "epoch": 4.870859337493999, "grad_norm": 0.8954085670000772, "learning_rate": 2.0742285198652734e-08, "loss": 0.3481, "step": 5075 }, { "epoch": 4.871819491118579, "grad_norm": 0.7561000659317284, "learning_rate": 2.0438580261704087e-08, "loss": 0.4181, "step": 5076 }, { "epoch": 4.872779644743159, "grad_norm": 0.9275813376747075, "learning_rate": 2.0137110596648025e-08, "loss": 0.5413, "step": 5077 }, { "epoch": 4.873739798367739, "grad_norm": 0.9488626020827761, "learning_rate": 1.9837876338811292e-08, "loss": 0.4667, "step": 5078 }, { "epoch": 4.874699951992318, "grad_norm": 0.9264931630469518, "learning_rate": 1.9540877622516995e-08, "loss": 0.5945, "step": 5079 }, { "epoch": 4.875660105616898, "grad_norm": 1.496972071727646, "learning_rate": 1.924611458108461e-08, "loss": 0.4076, "step": 5080 }, { "epoch": 4.876620259241479, "grad_norm": 0.9181303633694966, "learning_rate": 1.8953587346829948e-08, "loss": 0.4408, "step": 5081 }, { "epoch": 4.877580412866059, "grad_norm": 1.0545928041621861, "learning_rate": 1.8663296051065762e-08, "loss": 0.4733, "step": 5082 }, { "epoch": 4.8785405664906385, "grad_norm": 0.8837803153806716, "learning_rate": 1.8375240824100028e-08, "loss": 0.5908, "step": 5083 }, { "epoch": 4.879500720115218, "grad_norm": 1.0876973706554303, "learning_rate": 1.8089421795238205e-08, "loss": 0.3593, "step": 5084 }, { "epoch": 4.880460873739798, "grad_norm": 0.6604402543952191, "learning_rate": 1.7805839092781553e-08, "loss": 0.4476, "step": 5085 }, { "epoch": 4.881421027364379, "grad_norm": 1.0266940678541534, "learning_rate": 1.7524492844026576e-08, "loss": 0.3625, "step": 5086 }, { "epoch": 4.882381180988959, "grad_norm": 0.6798553332839922, "learning_rate": 1.724538317526725e-08, "loss": 0.4988, "step": 5087 }, { "epoch": 4.883341334613538, "grad_norm": 0.9450559239711623, "learning_rate": 1.69685102117928e-08, "loss": 0.5873, "step": 5088 }, { "epoch": 4.884301488238118, "grad_norm": 0.657794475732379, "learning_rate": 1.6693874077888805e-08, "loss": 0.4383, "step": 5089 }, { "epoch": 4.885261641862698, "grad_norm": 0.8129979383909199, "learning_rate": 1.6421474896835544e-08, "loss": 0.3791, "step": 5090 }, { "epoch": 4.886221795487278, "grad_norm": 0.9637728280863701, "learning_rate": 1.615131279091131e-08, "loss": 0.488, "step": 5091 }, { "epoch": 4.8871819491118575, "grad_norm": 0.7562024220115714, "learning_rate": 1.588338788138799e-08, "loss": 0.4641, "step": 5092 }, { "epoch": 4.888142102736438, "grad_norm": 0.7918500443923144, "learning_rate": 1.5617700288534377e-08, "loss": 0.4655, "step": 5093 }, { "epoch": 4.889102256361018, "grad_norm": 0.9544850534592311, "learning_rate": 1.5354250131615622e-08, "loss": 0.4234, "step": 5094 }, { "epoch": 4.890062409985598, "grad_norm": 0.7264681102756085, "learning_rate": 1.5093037528890463e-08, "loss": 0.4689, "step": 5095 }, { "epoch": 4.891022563610178, "grad_norm": 1.1002770956789973, "learning_rate": 1.4834062597615107e-08, "loss": 0.4282, "step": 5096 }, { "epoch": 4.891982717234757, "grad_norm": 0.8233129570912056, "learning_rate": 1.4577325454041003e-08, "loss": 0.4867, "step": 5097 }, { "epoch": 4.892942870859337, "grad_norm": 0.8248870254425921, "learning_rate": 1.4322826213414298e-08, "loss": 0.3289, "step": 5098 }, { "epoch": 4.893903024483917, "grad_norm": 0.9702250376905386, "learning_rate": 1.4070564989976387e-08, "loss": 0.5553, "step": 5099 }, { "epoch": 4.894863178108498, "grad_norm": 0.8413811744306313, "learning_rate": 1.3820541896965579e-08, "loss": 0.4694, "step": 5100 }, { "epoch": 4.8958233317330775, "grad_norm": 0.9298669172874939, "learning_rate": 1.3572757046614315e-08, "loss": 0.4634, "step": 5101 }, { "epoch": 4.896783485357657, "grad_norm": 1.481957768632417, "learning_rate": 1.3327210550150293e-08, "loss": 0.4568, "step": 5102 }, { "epoch": 4.897743638982237, "grad_norm": 0.9185883975467055, "learning_rate": 1.308390251779701e-08, "loss": 0.3786, "step": 5103 }, { "epoch": 4.898703792606817, "grad_norm": 0.950170159275753, "learning_rate": 1.2842833058772653e-08, "loss": 0.4832, "step": 5104 }, { "epoch": 4.899663946231397, "grad_norm": 1.1318729723239085, "learning_rate": 1.260400228129066e-08, "loss": 0.4564, "step": 5105 }, { "epoch": 4.9006240998559765, "grad_norm": 0.8330918192969402, "learning_rate": 1.2367410292560277e-08, "loss": 0.5293, "step": 5106 }, { "epoch": 4.901584253480557, "grad_norm": 0.9055144007336328, "learning_rate": 1.2133057198784326e-08, "loss": 0.4997, "step": 5107 }, { "epoch": 4.902544407105137, "grad_norm": 0.6513580076365965, "learning_rate": 1.1900943105160878e-08, "loss": 0.4061, "step": 5108 }, { "epoch": 4.903504560729717, "grad_norm": 0.9915808369039375, "learning_rate": 1.1671068115884921e-08, "loss": 0.5563, "step": 5109 }, { "epoch": 4.9044647143542965, "grad_norm": 1.0670110138792797, "learning_rate": 1.1443432334144466e-08, "loss": 0.5163, "step": 5110 }, { "epoch": 4.905424867978876, "grad_norm": 0.8777705346104406, "learning_rate": 1.1218035862122223e-08, "loss": 0.4398, "step": 5111 }, { "epoch": 4.906385021603457, "grad_norm": 0.8320563788627845, "learning_rate": 1.09948788009967e-08, "loss": 0.4887, "step": 5112 }, { "epoch": 4.907345175228037, "grad_norm": 0.8999332800498788, "learning_rate": 1.0773961250939991e-08, "loss": 0.458, "step": 5113 }, { "epoch": 4.908305328852617, "grad_norm": 0.9919509999774143, "learning_rate": 1.0555283311121101e-08, "loss": 0.3277, "step": 5114 }, { "epoch": 4.909265482477196, "grad_norm": 1.1770181261800745, "learning_rate": 1.0338845079700954e-08, "loss": 0.4768, "step": 5115 }, { "epoch": 4.910225636101776, "grad_norm": 1.3563543716858768, "learning_rate": 1.0124646653836279e-08, "loss": 0.5518, "step": 5116 }, { "epoch": 4.911185789726356, "grad_norm": 0.9098414488591223, "learning_rate": 9.912688129679603e-09, "loss": 0.5431, "step": 5117 }, { "epoch": 4.912145943350936, "grad_norm": 1.3650384676047957, "learning_rate": 9.702969602375378e-09, "loss": 0.4119, "step": 5118 }, { "epoch": 4.9131060969755165, "grad_norm": 1.041579869229159, "learning_rate": 9.495491166065519e-09, "loss": 0.5021, "step": 5119 }, { "epoch": 4.914066250600096, "grad_norm": 0.696486475222185, "learning_rate": 9.290252913883856e-09, "loss": 0.4164, "step": 5120 }, { "epoch": 4.915026404224676, "grad_norm": 1.0190343633724177, "learning_rate": 9.087254937960032e-09, "loss": 0.4473, "step": 5121 }, { "epoch": 4.915986557849256, "grad_norm": 1.245288802132973, "learning_rate": 8.886497329417265e-09, "loss": 0.3665, "step": 5122 }, { "epoch": 4.916946711473836, "grad_norm": 0.9855226306348528, "learning_rate": 8.687980178374023e-09, "loss": 0.4575, "step": 5123 }, { "epoch": 4.9179068650984155, "grad_norm": 1.018563194735285, "learning_rate": 8.491703573942356e-09, "loss": 0.4142, "step": 5124 }, { "epoch": 4.918867018722995, "grad_norm": 0.7874256401597032, "learning_rate": 8.29766760422901e-09, "loss": 0.4988, "step": 5125 }, { "epoch": 4.919827172347576, "grad_norm": 0.7164379236411765, "learning_rate": 8.105872356333755e-09, "loss": 0.5507, "step": 5126 }, { "epoch": 4.920787325972156, "grad_norm": 1.3962972111121241, "learning_rate": 7.916317916352168e-09, "loss": 0.4475, "step": 5127 }, { "epoch": 4.9217474795967355, "grad_norm": 1.0187625225619636, "learning_rate": 7.729004369373405e-09, "loss": 0.4116, "step": 5128 }, { "epoch": 4.922707633221315, "grad_norm": 0.7527051931815145, "learning_rate": 7.543931799479653e-09, "loss": 0.5078, "step": 5129 }, { "epoch": 4.923667786845895, "grad_norm": 1.1824996400939631, "learning_rate": 7.3611002897489015e-09, "loss": 0.4686, "step": 5130 }, { "epoch": 4.924627940470475, "grad_norm": 1.0103043291721752, "learning_rate": 7.180509922251611e-09, "loss": 0.4629, "step": 5131 }, { "epoch": 4.925588094095055, "grad_norm": 1.2908230975696109, "learning_rate": 7.002160778053491e-09, "loss": 0.544, "step": 5132 }, { "epoch": 4.926548247719635, "grad_norm": 0.7305148500074796, "learning_rate": 6.826052937212724e-09, "loss": 0.4304, "step": 5133 }, { "epoch": 4.927508401344215, "grad_norm": 1.0847828492203662, "learning_rate": 6.652186478783296e-09, "loss": 0.4435, "step": 5134 }, { "epoch": 4.928468554968795, "grad_norm": 0.7851624765877688, "learning_rate": 6.480561480811109e-09, "loss": 0.5453, "step": 5135 }, { "epoch": 4.929428708593375, "grad_norm": 0.8540088687705054, "learning_rate": 6.311178020336761e-09, "loss": 0.4455, "step": 5136 }, { "epoch": 4.930388862217955, "grad_norm": 0.8072295433519338, "learning_rate": 6.14403617339554e-09, "loss": 0.5041, "step": 5137 }, { "epoch": 4.931349015842535, "grad_norm": 1.0536817149215467, "learning_rate": 5.979136015015208e-09, "loss": 0.4557, "step": 5138 }, { "epoch": 4.932309169467115, "grad_norm": 0.9075157739383655, "learning_rate": 5.8164776192171094e-09, "loss": 0.4536, "step": 5139 }, { "epoch": 4.933269323091695, "grad_norm": 1.1476957558902572, "learning_rate": 5.656061059017837e-09, "loss": 0.386, "step": 5140 }, { "epoch": 4.934229476716275, "grad_norm": 0.8206538382527856, "learning_rate": 5.4978864064259005e-09, "loss": 0.4516, "step": 5141 }, { "epoch": 4.9351896303408544, "grad_norm": 1.0442810282050246, "learning_rate": 5.3419537324445044e-09, "loss": 0.5025, "step": 5142 }, { "epoch": 4.936149783965434, "grad_norm": 0.7684469454995162, "learning_rate": 5.18826310706988e-09, "loss": 0.461, "step": 5143 }, { "epoch": 4.937109937590014, "grad_norm": 0.8917712800580302, "learning_rate": 5.0368145992929495e-09, "loss": 0.4728, "step": 5144 }, { "epoch": 4.938070091214595, "grad_norm": 0.9396908320541816, "learning_rate": 4.887608277096001e-09, "loss": 0.3717, "step": 5145 }, { "epoch": 4.9390302448391745, "grad_norm": 0.8771142255825808, "learning_rate": 4.740644207457678e-09, "loss": 0.4955, "step": 5146 }, { "epoch": 4.939990398463754, "grad_norm": 1.0734160359965934, "learning_rate": 4.5959224563474345e-09, "loss": 0.4621, "step": 5147 }, { "epoch": 4.940950552088334, "grad_norm": 0.733287272299971, "learning_rate": 4.4534430887299694e-09, "loss": 0.4438, "step": 5148 }, { "epoch": 4.941910705712914, "grad_norm": 0.849530419663069, "learning_rate": 4.3132061685619005e-09, "loss": 0.4951, "step": 5149 }, { "epoch": 4.942870859337494, "grad_norm": 0.8935156942900468, "learning_rate": 4.1752117587950945e-09, "loss": 0.4779, "step": 5150 }, { "epoch": 4.9438310129620735, "grad_norm": 0.989712816784453, "learning_rate": 4.0394599213733346e-09, "loss": 0.5291, "step": 5151 }, { "epoch": 4.944791166586654, "grad_norm": 0.9149994611771824, "learning_rate": 3.905950717233986e-09, "loss": 0.4334, "step": 5152 }, { "epoch": 4.945751320211234, "grad_norm": 1.1674769408072119, "learning_rate": 3.774684206308554e-09, "loss": 0.5056, "step": 5153 }, { "epoch": 4.946711473835814, "grad_norm": 0.9499596404349934, "learning_rate": 3.6456604475199055e-09, "loss": 0.4389, "step": 5154 }, { "epoch": 4.947671627460394, "grad_norm": 0.8390458259946793, "learning_rate": 3.5188794987867093e-09, "loss": 0.4583, "step": 5155 }, { "epoch": 4.948631781084973, "grad_norm": 0.7286754389302613, "learning_rate": 3.394341417018998e-09, "loss": 0.3918, "step": 5156 }, { "epoch": 4.949591934709553, "grad_norm": 0.7602302684257762, "learning_rate": 3.272046258120942e-09, "loss": 0.4557, "step": 5157 }, { "epoch": 4.950552088334134, "grad_norm": 0.7737036030779597, "learning_rate": 3.151994076988629e-09, "loss": 0.5459, "step": 5158 }, { "epoch": 4.951512241958714, "grad_norm": 0.7180901460575704, "learning_rate": 3.034184927513395e-09, "loss": 0.4948, "step": 5159 }, { "epoch": 4.952472395583293, "grad_norm": 1.1473621729871353, "learning_rate": 2.9186188625779376e-09, "loss": 0.5037, "step": 5160 }, { "epoch": 4.953432549207873, "grad_norm": 0.9458990696901555, "learning_rate": 2.805295934058538e-09, "loss": 0.4697, "step": 5161 }, { "epoch": 4.954392702832453, "grad_norm": 1.1340408256274432, "learning_rate": 2.694216192824506e-09, "loss": 0.4904, "step": 5162 }, { "epoch": 4.955352856457033, "grad_norm": 0.7815188358693571, "learning_rate": 2.5853796887387316e-09, "loss": 0.5352, "step": 5163 }, { "epoch": 4.9563130100816135, "grad_norm": 0.69204181911625, "learning_rate": 2.47878647065658e-09, "loss": 0.4421, "step": 5164 }, { "epoch": 4.957273163706193, "grad_norm": 0.8269983690421522, "learning_rate": 2.374436586426443e-09, "loss": 0.4065, "step": 5165 }, { "epoch": 4.958233317330773, "grad_norm": 0.9422006923896841, "learning_rate": 2.27233008289085e-09, "loss": 0.5472, "step": 5166 }, { "epoch": 4.959193470955353, "grad_norm": 0.9556905610945488, "learning_rate": 2.1724670058825838e-09, "loss": 0.4556, "step": 5167 }, { "epoch": 4.960153624579933, "grad_norm": 0.7857921879178307, "learning_rate": 2.074847400230784e-09, "loss": 0.4524, "step": 5168 }, { "epoch": 4.9611137782045125, "grad_norm": 0.9366024342545007, "learning_rate": 1.9794713097548433e-09, "loss": 0.4778, "step": 5169 }, { "epoch": 4.962073931829092, "grad_norm": 0.7052284577448736, "learning_rate": 1.886338777268293e-09, "loss": 0.45, "step": 5170 }, { "epoch": 4.963034085453673, "grad_norm": 0.7827365414195531, "learning_rate": 1.7954498445771352e-09, "loss": 0.5261, "step": 5171 }, { "epoch": 4.963994239078253, "grad_norm": 0.7965128404438521, "learning_rate": 1.7068045524815114e-09, "loss": 0.3534, "step": 5172 }, { "epoch": 4.964954392702833, "grad_norm": 0.7793822718587337, "learning_rate": 1.6204029407718147e-09, "loss": 0.4082, "step": 5173 }, { "epoch": 4.965914546327412, "grad_norm": 1.5015426916080958, "learning_rate": 1.5362450482336865e-09, "loss": 0.448, "step": 5174 }, { "epoch": 4.966874699951992, "grad_norm": 0.849618293959942, "learning_rate": 1.4543309126446858e-09, "loss": 0.5537, "step": 5175 }, { "epoch": 4.967834853576572, "grad_norm": 0.8243558420108481, "learning_rate": 1.374660570774844e-09, "loss": 0.396, "step": 5176 }, { "epoch": 4.968795007201152, "grad_norm": 0.8910521895348387, "learning_rate": 1.2972340583877751e-09, "loss": 0.4298, "step": 5177 }, { "epoch": 4.969755160825732, "grad_norm": 0.7299870942064338, "learning_rate": 1.2220514102390114e-09, "loss": 0.3863, "step": 5178 }, { "epoch": 4.970715314450312, "grad_norm": 0.7413579782631262, "learning_rate": 1.149112660077667e-09, "loss": 0.4193, "step": 5179 }, { "epoch": 4.971675468074892, "grad_norm": 0.7783002624428172, "learning_rate": 1.0784178406442192e-09, "loss": 0.4703, "step": 5180 }, { "epoch": 4.972635621699472, "grad_norm": 0.9379274466944767, "learning_rate": 1.0099669836743931e-09, "loss": 0.4261, "step": 5181 }, { "epoch": 4.973595775324052, "grad_norm": 0.9579292711922635, "learning_rate": 9.43760119893611e-10, "loss": 0.4366, "step": 5182 }, { "epoch": 4.974555928948631, "grad_norm": 0.8707900483313223, "learning_rate": 8.797972790219877e-10, "loss": 0.4057, "step": 5183 }, { "epoch": 4.975516082573212, "grad_norm": 0.8631414457810108, "learning_rate": 8.180784897715565e-10, "loss": 0.5697, "step": 5184 }, { "epoch": 4.976476236197792, "grad_norm": 1.1065510869169313, "learning_rate": 7.586037798479329e-10, "loss": 0.5134, "step": 5185 }, { "epoch": 4.977436389822372, "grad_norm": 0.7725099943602061, "learning_rate": 7.013731759475395e-10, "loss": 0.4906, "step": 5186 }, { "epoch": 4.9783965434469515, "grad_norm": 0.7789717283826445, "learning_rate": 6.463867037614923e-10, "loss": 0.5757, "step": 5187 }, { "epoch": 4.979356697071531, "grad_norm": 0.7645833236483804, "learning_rate": 5.936443879728249e-10, "loss": 0.5332, "step": 5188 }, { "epoch": 4.980316850696111, "grad_norm": 0.806711439661396, "learning_rate": 5.431462522559327e-10, "loss": 0.4377, "step": 5189 }, { "epoch": 4.981277004320692, "grad_norm": 1.0105591950180912, "learning_rate": 4.948923192793497e-10, "loss": 0.5062, "step": 5190 }, { "epoch": 4.9822371579452716, "grad_norm": 0.7391194758123448, "learning_rate": 4.4888261070408223e-10, "loss": 0.5032, "step": 5191 }, { "epoch": 4.983197311569851, "grad_norm": 1.1037328658299064, "learning_rate": 4.0511714718305396e-10, "loss": 0.3893, "step": 5192 }, { "epoch": 4.984157465194431, "grad_norm": 1.0119345423533812, "learning_rate": 3.6359594836277156e-10, "loss": 0.4207, "step": 5193 }, { "epoch": 4.985117618819011, "grad_norm": 0.8234977501394226, "learning_rate": 3.24319032880549e-10, "loss": 0.478, "step": 5194 }, { "epoch": 4.986077772443591, "grad_norm": 0.7787474709100735, "learning_rate": 2.872864183678381e-10, "loss": 0.4777, "step": 5195 }, { "epoch": 4.9870379260681705, "grad_norm": 1.1017375591818004, "learning_rate": 2.5249812144856335e-10, "loss": 0.547, "step": 5196 }, { "epoch": 4.987998079692751, "grad_norm": 0.9605564681074006, "learning_rate": 2.1995415773856666e-10, "loss": 0.5238, "step": 5197 }, { "epoch": 4.988958233317331, "grad_norm": 1.0687278528307413, "learning_rate": 1.8965454184671773e-10, "loss": 0.4792, "step": 5198 }, { "epoch": 4.989918386941911, "grad_norm": 0.9432528863977742, "learning_rate": 1.615992873732486e-10, "loss": 0.5079, "step": 5199 }, { "epoch": 4.990878540566491, "grad_norm": 1.1771290104573302, "learning_rate": 1.3578840691308438e-10, "loss": 0.3844, "step": 5200 }, { "epoch": 4.99183869419107, "grad_norm": 0.9491703556463479, "learning_rate": 1.1222191205195743e-10, "loss": 0.453, "step": 5201 }, { "epoch": 4.99279884781565, "grad_norm": 0.7639070322183352, "learning_rate": 9.089981336807275e-11, "loss": 0.4219, "step": 5202 }, { "epoch": 4.99375900144023, "grad_norm": 0.8779777149590328, "learning_rate": 7.182212043377323e-11, "loss": 0.5304, "step": 5203 }, { "epoch": 4.994719155064811, "grad_norm": 1.1706647505959993, "learning_rate": 5.4988841812209095e-11, "loss": 0.5506, "step": 5204 }, { "epoch": 4.9956793086893905, "grad_norm": 0.8845981532684919, "learning_rate": 4.039998505900311e-11, "loss": 0.4713, "step": 5205 }, { "epoch": 4.99663946231397, "grad_norm": 1.1937333060548272, "learning_rate": 2.8055556724471132e-11, "loss": 0.3481, "step": 5206 }, { "epoch": 4.99759961593855, "grad_norm": 0.9688319428237797, "learning_rate": 1.7955562348626054e-11, "loss": 0.4988, "step": 5207 }, { "epoch": 4.99855976956313, "grad_norm": 1.320308930695614, "learning_rate": 1.0100006465618706e-11, "loss": 0.4024, "step": 5208 }, { "epoch": 4.99951992318771, "grad_norm": 0.7887433224043555, "learning_rate": 4.4888926020725254e-12, "loss": 0.4998, "step": 5209 }, { "epoch": 5.0, "grad_norm": 0.7887433224043555, "learning_rate": 1.1222232765284447e-12, "loss": 0.3158, "step": 5210 }, { "epoch": 5.0, "step": 5210, "total_flos": 611590839336960.0, "train_loss": 0.5804602591410251, "train_runtime": 128031.3025, "train_samples_per_second": 0.325, "train_steps_per_second": 0.041 } ], "logging_steps": 1, "max_steps": 5210, "num_input_tokens_seen": 0, "num_train_epochs": 5, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 611590839336960.0, "train_batch_size": 1, "trial_name": null, "trial_params": null }