{ "best_metric": null, "best_model_checkpoint": null, "epoch": 2.9943851768669285, "eval_steps": 500, "global_step": 333, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.008983717012914094, "grad_norm": 5.839033078230819, "learning_rate": 6.655117647058823e-06, "loss": 0.851, "step": 1 }, { "epoch": 0.017967434025828188, "grad_norm": 5.796609591178738, "learning_rate": 1.3310235294117646e-05, "loss": 0.8589, "step": 2 }, { "epoch": 0.02695115103874228, "grad_norm": 4.022069478985531, "learning_rate": 1.996535294117647e-05, "loss": 0.8091, "step": 3 }, { "epoch": 0.035934868051656375, "grad_norm": 5.021116871190826, "learning_rate": 2.6620470588235293e-05, "loss": 0.8296, "step": 4 }, { "epoch": 0.044918585064570464, "grad_norm": 6.766543105088873, "learning_rate": 3.3275588235294115e-05, "loss": 0.8406, "step": 5 }, { "epoch": 0.05390230207748456, "grad_norm": 3.7488525352605198, "learning_rate": 3.993070588235294e-05, "loss": 0.7848, "step": 6 }, { "epoch": 0.06288601909039865, "grad_norm": 3.232161490369813, "learning_rate": 4.658582352941176e-05, "loss": 0.7269, "step": 7 }, { "epoch": 0.07186973610331275, "grad_norm": 3.5299607181225734, "learning_rate": 5.3240941176470586e-05, "loss": 0.7032, "step": 8 }, { "epoch": 0.08085345311622684, "grad_norm": 2.4635657277606833, "learning_rate": 5.989605882352941e-05, "loss": 0.686, "step": 9 }, { "epoch": 0.08983717012914093, "grad_norm": 3.5184074662375378, "learning_rate": 6.655117647058823e-05, "loss": 0.6731, "step": 10 }, { "epoch": 0.09882088714205503, "grad_norm": 2.034739373521881, "learning_rate": 7.320629411764706e-05, "loss": 0.6436, "step": 11 }, { "epoch": 0.10780460415496912, "grad_norm": 3.484093426894694, "learning_rate": 7.986141176470588e-05, "loss": 0.6546, "step": 12 }, { "epoch": 0.11678832116788321, "grad_norm": 1.7458659318282235, "learning_rate": 8.65165294117647e-05, "loss": 0.6243, "step": 13 }, { "epoch": 0.1257720381807973, "grad_norm": 2.696049020472868, "learning_rate": 9.317164705882352e-05, "loss": 0.6431, "step": 14 }, { "epoch": 0.13475575519371139, "grad_norm": 2.7403846964581335, "learning_rate": 9.982676470588235e-05, "loss": 0.6337, "step": 15 }, { "epoch": 0.1437394722066255, "grad_norm": 1.9984598625107888, "learning_rate": 0.00010648188235294117, "loss": 0.6219, "step": 16 }, { "epoch": 0.1527231892195396, "grad_norm": 1.8055549976045306, "learning_rate": 0.000113137, "loss": 0.6009, "step": 17 }, { "epoch": 0.16170690623245368, "grad_norm": 1.5503606010043973, "learning_rate": 0.00011979211764705882, "loss": 0.5964, "step": 18 }, { "epoch": 0.17069062324536777, "grad_norm": 2.35861768584361, "learning_rate": 0.00012644723529411765, "loss": 0.6148, "step": 19 }, { "epoch": 0.17967434025828186, "grad_norm": 1.5396308193214678, "learning_rate": 0.00013310235294117646, "loss": 0.6018, "step": 20 }, { "epoch": 0.18865805727119594, "grad_norm": 1.6143825069098845, "learning_rate": 0.0001397574705882353, "loss": 0.5954, "step": 21 }, { "epoch": 0.19764177428411006, "grad_norm": 1.6144901661060405, "learning_rate": 0.0001464125882352941, "loss": 0.6004, "step": 22 }, { "epoch": 0.20662549129702415, "grad_norm": 3.3171185005484856, "learning_rate": 0.00015306770588235295, "loss": 0.6111, "step": 23 }, { "epoch": 0.21560920830993824, "grad_norm": 2.2411018078854936, "learning_rate": 0.00015972282352941176, "loss": 0.6119, "step": 24 }, { "epoch": 0.22459292532285233, "grad_norm": 8.161797640058992, "learning_rate": 0.0001663779411764706, "loss": 0.593, "step": 25 }, { "epoch": 0.23357664233576642, "grad_norm": 846.7701836327712, "learning_rate": 0.0001730330588235294, "loss": 5.673, "step": 26 }, { "epoch": 0.2425603593486805, "grad_norm": 4.629361945191611, "learning_rate": 0.00017968817647058823, "loss": 0.7795, "step": 27 }, { "epoch": 0.2515440763615946, "grad_norm": 85.87349809516256, "learning_rate": 0.00018634329411764704, "loss": 0.7672, "step": 28 }, { "epoch": 0.2605277933745087, "grad_norm": 2.3423299747514212, "learning_rate": 0.00019299841176470588, "loss": 0.6615, "step": 29 }, { "epoch": 0.26951151038742277, "grad_norm": 2.629813380478519, "learning_rate": 0.0001996535294117647, "loss": 0.6385, "step": 30 }, { "epoch": 0.2784952274003369, "grad_norm": 1.2741888860891866, "learning_rate": 0.00020630864705882353, "loss": 0.612, "step": 31 }, { "epoch": 0.287478944413251, "grad_norm": 31.843678922273796, "learning_rate": 0.00021296376470588234, "loss": 0.6641, "step": 32 }, { "epoch": 0.29646266142616506, "grad_norm": 5.150042714870819, "learning_rate": 0.00021961888235294118, "loss": 0.9157, "step": 33 }, { "epoch": 0.3054463784390792, "grad_norm": 610.734122596688, "learning_rate": 0.000226274, "loss": 1.7373, "step": 34 }, { "epoch": 0.31443009545199324, "grad_norm": 46.58948260555198, "learning_rate": 0.00022626775506336304, "loss": 0.8771, "step": 35 }, { "epoch": 0.32341381246490736, "grad_norm": 6.962976785617738, "learning_rate": 0.00022624902094286824, "loss": 0.9851, "step": 36 }, { "epoch": 0.3323975294778215, "grad_norm": 21.673571852019826, "learning_rate": 0.00022621779970668783, "loss": 0.9069, "step": 37 }, { "epoch": 0.34138124649073553, "grad_norm": 2.892556097499312, "learning_rate": 0.00022617409480152153, "loss": 0.8182, "step": 38 }, { "epoch": 0.35036496350364965, "grad_norm": 15.60543009302285, "learning_rate": 0.00022611791105221654, "loss": 0.9486, "step": 39 }, { "epoch": 0.3593486805165637, "grad_norm": 11.538473656686941, "learning_rate": 0.00022604925466123451, "loss": 0.8591, "step": 40 }, { "epoch": 0.36833239752947783, "grad_norm": 2.5218994606008907, "learning_rate": 0.00022596813320796707, "loss": 0.7781, "step": 41 }, { "epoch": 0.3773161145423919, "grad_norm": 4.662799540456364, "learning_rate": 0.00022587455564789883, "loss": 0.8526, "step": 42 }, { "epoch": 0.386299831555306, "grad_norm": 565.0656362499724, "learning_rate": 0.00022576853231161902, "loss": 1.738, "step": 43 }, { "epoch": 0.3952835485682201, "grad_norm": 7.6600755456980325, "learning_rate": 0.00022565007490368076, "loss": 0.8999, "step": 44 }, { "epoch": 0.4042672655811342, "grad_norm": 287.49302909066023, "learning_rate": 0.00022551919650130918, "loss": 7.1473, "step": 45 }, { "epoch": 0.4132509825940483, "grad_norm": 13.25846612995419, "learning_rate": 0.00022537591155295756, "loss": 1.5197, "step": 46 }, { "epoch": 0.42223469960696236, "grad_norm": 6.573021527190113, "learning_rate": 0.00022522023587671233, "loss": 1.0434, "step": 47 }, { "epoch": 0.4312184166198765, "grad_norm": 2.561584699671566, "learning_rate": 0.0002250521866585469, "loss": 0.8122, "step": 48 }, { "epoch": 0.4402021336327906, "grad_norm": 12.292027687573608, "learning_rate": 0.00022487178245042422, "loss": 0.9547, "step": 49 }, { "epoch": 0.44918585064570465, "grad_norm": 14.460331630727685, "learning_rate": 0.0002246790431682489, "loss": 1.026, "step": 50 }, { "epoch": 0.45816956765861877, "grad_norm": 4.017308882255896, "learning_rate": 0.00022447399008966853, "loss": 0.8958, "step": 51 }, { "epoch": 0.46715328467153283, "grad_norm": 1.8162087381759398, "learning_rate": 0.0002242566458517245, "loss": 0.7584, "step": 52 }, { "epoch": 0.47613700168444695, "grad_norm": 2.229007766044214, "learning_rate": 0.0002240270344483534, "loss": 0.7519, "step": 53 }, { "epoch": 0.485120718697361, "grad_norm": 1.275167700368963, "learning_rate": 0.00022378518122773768, "loss": 0.7169, "step": 54 }, { "epoch": 0.4941044357102751, "grad_norm": 1.2052448804014857, "learning_rate": 0.00022353111288950776, "loss": 0.7069, "step": 55 }, { "epoch": 0.5030881527231892, "grad_norm": 2.809114390406924, "learning_rate": 0.00022326485748179416, "loss": 0.6934, "step": 56 }, { "epoch": 0.5120718697361033, "grad_norm": 1.7295819486463446, "learning_rate": 0.00022298644439813125, "loss": 0.6997, "step": 57 }, { "epoch": 0.5210555867490174, "grad_norm": 1.176047538360094, "learning_rate": 0.00022269590437421234, "loss": 0.6587, "step": 58 }, { "epoch": 0.5300393037619315, "grad_norm": 0.9770345550489453, "learning_rate": 0.0002223932694844966, "loss": 0.6444, "step": 59 }, { "epoch": 0.5390230207748455, "grad_norm": 1.496641892403964, "learning_rate": 0.00022207857313866796, "loss": 0.6558, "step": 60 }, { "epoch": 0.5480067377877597, "grad_norm": 1.7254473728830027, "learning_rate": 0.00022175185007794712, "loss": 0.6647, "step": 61 }, { "epoch": 0.5569904548006738, "grad_norm": 1.148244331773896, "learning_rate": 0.000221413136371256, "loss": 0.6244, "step": 62 }, { "epoch": 0.5659741718135879, "grad_norm": 1.082240763317917, "learning_rate": 0.00022106246941123604, "loss": 0.616, "step": 63 }, { "epoch": 0.574957888826502, "grad_norm": 1.0463382313968703, "learning_rate": 0.0002206998879101201, "loss": 0.6097, "step": 64 }, { "epoch": 0.583941605839416, "grad_norm": 1.5314621991483217, "learning_rate": 0.00022032543189545893, "loss": 0.6261, "step": 65 }, { "epoch": 0.5929253228523301, "grad_norm": 0.8440358318436397, "learning_rate": 0.00021993914270570204, "loss": 0.5992, "step": 66 }, { "epoch": 0.6019090398652442, "grad_norm": 1.1721068603539129, "learning_rate": 0.0002195410629856343, "loss": 0.6003, "step": 67 }, { "epoch": 0.6108927568781584, "grad_norm": 1.0225694078862206, "learning_rate": 0.00021913123668166815, "loss": 0.5952, "step": 68 }, { "epoch": 0.6198764738910725, "grad_norm": 1.223954503187751, "learning_rate": 0.00021870970903699184, "loss": 0.5844, "step": 69 }, { "epoch": 0.6288601909039865, "grad_norm": 0.7517434069071857, "learning_rate": 0.00021827652658657518, "loss": 0.5736, "step": 70 }, { "epoch": 0.6378439079169006, "grad_norm": 0.542782981372059, "learning_rate": 0.00021783173715203175, "loss": 0.5677, "step": 71 }, { "epoch": 0.6468276249298147, "grad_norm": 0.8491645074764822, "learning_rate": 0.00021737538983634003, "loss": 0.5567, "step": 72 }, { "epoch": 0.6558113419427288, "grad_norm": 0.9512722570962481, "learning_rate": 0.00021690753501842243, "loss": 0.5652, "step": 73 }, { "epoch": 0.664795058955643, "grad_norm": 0.7420499518682178, "learning_rate": 0.00021642822434758365, "loss": 0.5521, "step": 74 }, { "epoch": 0.673778775968557, "grad_norm": 0.6023655209710193, "learning_rate": 0.0002159375107378088, "loss": 0.5484, "step": 75 }, { "epoch": 0.6827624929814711, "grad_norm": 0.6704929086374853, "learning_rate": 0.00021543544836192198, "loss": 0.5629, "step": 76 }, { "epoch": 0.6917462099943852, "grad_norm": 0.8511116323519582, "learning_rate": 0.00021492209264560583, "loss": 0.5477, "step": 77 }, { "epoch": 0.7007299270072993, "grad_norm": 0.8293067237834507, "learning_rate": 0.00021439750026128255, "loss": 0.5486, "step": 78 }, { "epoch": 0.7097136440202133, "grad_norm": 0.7951812792060676, "learning_rate": 0.00021386172912185786, "loss": 0.5386, "step": 79 }, { "epoch": 0.7186973610331274, "grad_norm": 0.5942026077567073, "learning_rate": 0.0002133148383743272, "loss": 0.5526, "step": 80 }, { "epoch": 0.7276810780460415, "grad_norm": 0.7074314677815655, "learning_rate": 0.00021275688839324653, "loss": 0.5355, "step": 81 }, { "epoch": 0.7366647950589557, "grad_norm": 1.0064759251882633, "learning_rate": 0.00021218794077406699, "loss": 0.5214, "step": 82 }, { "epoch": 0.7456485120718698, "grad_norm": 0.5936960428857333, "learning_rate": 0.00021160805832633507, "loss": 0.5334, "step": 83 }, { "epoch": 0.7546322290847838, "grad_norm": 0.5587606186465214, "learning_rate": 0.00021101730506675872, "loss": 0.5246, "step": 84 }, { "epoch": 0.7636159460976979, "grad_norm": 0.7563054405639342, "learning_rate": 0.00021041574621214016, "loss": 0.5211, "step": 85 }, { "epoch": 0.772599663110612, "grad_norm": 0.4650706352210191, "learning_rate": 0.00020980344817217607, "loss": 0.52, "step": 86 }, { "epoch": 0.7815833801235261, "grad_norm": 0.46292748233999875, "learning_rate": 0.00020918047854212644, "loss": 0.5086, "step": 87 }, { "epoch": 0.7905670971364402, "grad_norm": 0.45822776585768843, "learning_rate": 0.00020854690609535217, "loss": 0.5205, "step": 88 }, { "epoch": 0.7995508141493542, "grad_norm": 0.3680018488886641, "learning_rate": 0.0002079028007757228, "loss": 0.4995, "step": 89 }, { "epoch": 0.8085345311622684, "grad_norm": 0.3651907315130092, "learning_rate": 0.000207248233689895, "loss": 0.5108, "step": 90 }, { "epoch": 0.8175182481751825, "grad_norm": 0.38739099960496165, "learning_rate": 0.00020658327709946274, "loss": 0.5105, "step": 91 }, { "epoch": 0.8265019651880966, "grad_norm": 0.37587537297348633, "learning_rate": 0.00020590800441297976, "loss": 0.5043, "step": 92 }, { "epoch": 0.8354856822010107, "grad_norm": 0.5725177706736702, "learning_rate": 0.00020522249017785566, "loss": 0.5056, "step": 93 }, { "epoch": 0.8444693992139247, "grad_norm": 1.0428309894412633, "learning_rate": 0.00020452681007212612, "loss": 0.5189, "step": 94 }, { "epoch": 0.8534531162268388, "grad_norm": 1.1604159616874143, "learning_rate": 0.0002038210408960984, "loss": 0.5167, "step": 95 }, { "epoch": 0.862436833239753, "grad_norm": 0.5208537644478249, "learning_rate": 0.0002031052605638728, "loss": 0.5145, "step": 96 }, { "epoch": 0.8714205502526671, "grad_norm": 0.83804291321802, "learning_rate": 0.00020237954809474134, "loss": 0.5158, "step": 97 }, { "epoch": 0.8804042672655812, "grad_norm": 0.7064273962256518, "learning_rate": 0.00020164398360446436, "loss": 0.4931, "step": 98 }, { "epoch": 0.8893879842784952, "grad_norm": 0.5057800234712763, "learning_rate": 0.00020089864829642596, "loss": 0.5075, "step": 99 }, { "epoch": 0.8983717012914093, "grad_norm": 0.5940667934373025, "learning_rate": 0.0002001436244526695, "loss": 0.4907, "step": 100 }, { "epoch": 0.9073554183043234, "grad_norm": 0.5019525607511636, "learning_rate": 0.00019937899542481408, "loss": 0.5041, "step": 101 }, { "epoch": 0.9163391353172375, "grad_norm": 0.4491147515500972, "learning_rate": 0.00019860484562485276, "loss": 0.488, "step": 102 }, { "epoch": 0.9253228523301515, "grad_norm": 0.5412361814676229, "learning_rate": 0.00019782126051583386, "loss": 0.5023, "step": 103 }, { "epoch": 0.9343065693430657, "grad_norm": 0.3383672378854014, "learning_rate": 0.00019702832660242615, "loss": 0.489, "step": 104 }, { "epoch": 0.9432902863559798, "grad_norm": 0.36276664892975297, "learning_rate": 0.0001962261314213691, "loss": 0.4996, "step": 105 }, { "epoch": 0.9522740033688939, "grad_norm": 0.3454995598586516, "learning_rate": 0.00019541476353180916, "loss": 0.4856, "step": 106 }, { "epoch": 0.961257720381808, "grad_norm": 0.311134046456636, "learning_rate": 0.00019459431250552317, "loss": 0.4855, "step": 107 }, { "epoch": 0.970241437394722, "grad_norm": 0.3425949299888489, "learning_rate": 0.0001937648689170301, "loss": 0.4762, "step": 108 }, { "epoch": 0.9792251544076361, "grad_norm": 0.3496537132478162, "learning_rate": 0.00019292652433359177, "loss": 0.4834, "step": 109 }, { "epoch": 0.9882088714205502, "grad_norm": 0.4269583108093911, "learning_rate": 0.00019207937130510442, "loss": 0.4946, "step": 110 }, { "epoch": 0.9971925884334644, "grad_norm": 0.4147252598563634, "learning_rate": 0.0001912235033538814, "loss": 0.4846, "step": 111 }, { "epoch": 1.0075800112296462, "grad_norm": 0.8355251290758435, "learning_rate": 0.00019035901496432887, "loss": 0.8602, "step": 112 }, { "epoch": 1.0165637282425604, "grad_norm": 1.6156052899406323, "learning_rate": 0.0001894860015725148, "loss": 0.4935, "step": 113 }, { "epoch": 1.0255474452554745, "grad_norm": 0.46493272524584955, "learning_rate": 0.0001886045595556335, "loss": 0.4655, "step": 114 }, { "epoch": 1.0345311622683886, "grad_norm": 0.979120652286332, "learning_rate": 0.00018771478622136586, "loss": 0.4778, "step": 115 }, { "epoch": 1.0435148792813027, "grad_norm": 0.7656350182275576, "learning_rate": 0.0001868167797971369, "loss": 0.4786, "step": 116 }, { "epoch": 1.0524985962942168, "grad_norm": 0.5200988865180658, "learning_rate": 0.00018591063941927197, "loss": 0.4667, "step": 117 }, { "epoch": 1.0614823133071307, "grad_norm": 0.5373218112082903, "learning_rate": 0.00018499646512205238, "loss": 0.4698, "step": 118 }, { "epoch": 1.0704660303200448, "grad_norm": 0.3981583489798308, "learning_rate": 0.00018407435782667207, "loss": 0.4573, "step": 119 }, { "epoch": 1.079449747332959, "grad_norm": 0.4729814422997812, "learning_rate": 0.0001831444193300964, "loss": 0.4537, "step": 120 }, { "epoch": 1.088433464345873, "grad_norm": 0.40160292629773175, "learning_rate": 0.00018220675229382407, "loss": 0.4607, "step": 121 }, { "epoch": 1.0974171813587872, "grad_norm": 0.34191199159101826, "learning_rate": 0.0001812614602325536, "loss": 0.4491, "step": 122 }, { "epoch": 1.1064008983717013, "grad_norm": 0.38115499724656887, "learning_rate": 0.00018030864750275598, "loss": 0.4649, "step": 123 }, { "epoch": 1.1153846153846154, "grad_norm": 0.2939815268385644, "learning_rate": 0.0001793484192911539, "loss": 0.4405, "step": 124 }, { "epoch": 1.1243683323975295, "grad_norm": 0.38182282790459066, "learning_rate": 0.00017838088160310975, "loss": 0.4673, "step": 125 }, { "epoch": 1.1333520494104437, "grad_norm": 0.26951517499854655, "learning_rate": 0.00017740614125092276, "loss": 0.4366, "step": 126 }, { "epoch": 1.1423357664233578, "grad_norm": 0.32986266756899896, "learning_rate": 0.00017642430584203776, "loss": 0.4506, "step": 127 }, { "epoch": 1.1513194834362717, "grad_norm": 0.29552314624204684, "learning_rate": 0.0001754354837671654, "loss": 0.4409, "step": 128 }, { "epoch": 1.1603032004491858, "grad_norm": 0.2715172818791699, "learning_rate": 0.0001744397841883164, "loss": 0.455, "step": 129 }, { "epoch": 1.1692869174621, "grad_norm": 0.30375726886135684, "learning_rate": 0.00017343731702675036, "loss": 0.4453, "step": 130 }, { "epoch": 1.178270634475014, "grad_norm": 0.20896293317807582, "learning_rate": 0.00017242819295084123, "loss": 0.4447, "step": 131 }, { "epoch": 1.1872543514879281, "grad_norm": 0.26828433742607094, "learning_rate": 0.0001714125233638595, "loss": 0.444, "step": 132 }, { "epoch": 1.1962380685008422, "grad_norm": 0.21271544817391416, "learning_rate": 0.00017039042039167416, "loss": 0.4434, "step": 133 }, { "epoch": 1.2052217855137564, "grad_norm": 0.22231269773708437, "learning_rate": 0.00016936199687037412, "loss": 0.4392, "step": 134 }, { "epoch": 1.2142055025266705, "grad_norm": 0.21858395274334466, "learning_rate": 0.00016832736633381174, "loss": 0.44, "step": 135 }, { "epoch": 1.2231892195395846, "grad_norm": 0.2359281892515482, "learning_rate": 0.00016728664300106905, "loss": 0.4395, "step": 136 }, { "epoch": 1.2321729365524985, "grad_norm": 0.21457493342859105, "learning_rate": 0.0001662399417638485, "loss": 0.442, "step": 137 }, { "epoch": 1.2411566535654126, "grad_norm": 0.17832258654038546, "learning_rate": 0.00016518737817378936, "loss": 0.4379, "step": 138 }, { "epoch": 1.2501403705783267, "grad_norm": 0.20100339820701033, "learning_rate": 0.0001641290684297112, "loss": 0.4371, "step": 139 }, { "epoch": 1.2591240875912408, "grad_norm": 0.23303786372370514, "learning_rate": 0.00016306512936478605, "loss": 0.4353, "step": 140 }, { "epoch": 1.268107804604155, "grad_norm": 0.25687763850096407, "learning_rate": 0.00016199567843364066, "loss": 0.4379, "step": 141 }, { "epoch": 1.277091521617069, "grad_norm": 0.23974642432643137, "learning_rate": 0.00016092083369938952, "loss": 0.4364, "step": 142 }, { "epoch": 1.2860752386299832, "grad_norm": 0.20652369486142858, "learning_rate": 0.00015984071382060174, "loss": 0.4385, "step": 143 }, { "epoch": 1.2950589556428973, "grad_norm": 0.20585656880887795, "learning_rate": 0.00015875543803820115, "loss": 0.4425, "step": 144 }, { "epoch": 1.3040426726558114, "grad_norm": 0.22079211853825997, "learning_rate": 0.00015766512616230286, "loss": 0.4251, "step": 145 }, { "epoch": 1.3130263896687255, "grad_norm": 0.19534999766663985, "learning_rate": 0.00015656989855898653, "loss": 0.4374, "step": 146 }, { "epoch": 1.3220101066816397, "grad_norm": 0.19699056774469947, "learning_rate": 0.00015546987613700858, "loss": 0.4385, "step": 147 }, { "epoch": 1.3309938236945535, "grad_norm": 0.25025460629960233, "learning_rate": 0.00015436518033445427, "loss": 0.4414, "step": 148 }, { "epoch": 1.3399775407074677, "grad_norm": 0.22783106244726478, "learning_rate": 0.00015325593310533135, "loss": 0.4292, "step": 149 }, { "epoch": 1.3489612577203818, "grad_norm": 0.24800972153045137, "learning_rate": 0.00015214225690610695, "loss": 0.4367, "step": 150 }, { "epoch": 1.357944974733296, "grad_norm": 0.2755781135491422, "learning_rate": 0.00015102427468218884, "loss": 0.4343, "step": 151 }, { "epoch": 1.36692869174621, "grad_norm": 0.3302080679814042, "learning_rate": 0.00014990210985435274, "loss": 0.4225, "step": 152 }, { "epoch": 1.3759124087591241, "grad_norm": 0.3884646633170279, "learning_rate": 0.00014877588630511702, "loss": 0.4291, "step": 153 }, { "epoch": 1.3848961257720382, "grad_norm": 0.4487151204051441, "learning_rate": 0.0001476457283650669, "loss": 0.4314, "step": 154 }, { "epoch": 1.3938798427849521, "grad_norm": 0.4142847149164689, "learning_rate": 0.00014651176079912844, "loss": 0.4248, "step": 155 }, { "epoch": 1.4028635597978663, "grad_norm": 0.3852901369815094, "learning_rate": 0.00014537410879279536, "loss": 0.4233, "step": 156 }, { "epoch": 1.4118472768107804, "grad_norm": 0.364518339635796, "learning_rate": 0.00014423289793830883, "loss": 0.4362, "step": 157 }, { "epoch": 1.4208309938236945, "grad_norm": 0.35417387758842084, "learning_rate": 0.00014308825422079265, "loss": 0.4252, "step": 158 }, { "epoch": 1.4298147108366086, "grad_norm": 0.3557663368174752, "learning_rate": 0.00014194030400434496, "loss": 0.4395, "step": 159 }, { "epoch": 1.4387984278495227, "grad_norm": 0.34124968887701046, "learning_rate": 0.00014078917401808824, "loss": 0.4355, "step": 160 }, { "epoch": 1.4477821448624368, "grad_norm": 0.253304289924485, "learning_rate": 0.0001396349913421788, "loss": 0.4355, "step": 161 }, { "epoch": 1.456765861875351, "grad_norm": 0.2979898627616292, "learning_rate": 0.00013847788339377788, "loss": 0.4411, "step": 162 }, { "epoch": 1.465749578888265, "grad_norm": 0.2927807074804572, "learning_rate": 0.0001373179779129849, "loss": 0.434, "step": 163 }, { "epoch": 1.4747332959011792, "grad_norm": 0.23083501058351574, "learning_rate": 0.00013615540294873585, "loss": 0.4296, "step": 164 }, { "epoch": 1.4837170129140933, "grad_norm": 0.25440061192162694, "learning_rate": 0.00013499028684466692, "loss": 0.4316, "step": 165 }, { "epoch": 1.4927007299270074, "grad_norm": 0.3151406421517585, "learning_rate": 0.00013382275822494612, "loss": 0.4268, "step": 166 }, { "epoch": 1.5016844469399215, "grad_norm": 0.28632589462844177, "learning_rate": 0.00013265294598007347, "loss": 0.4338, "step": 167 }, { "epoch": 1.5106681639528357, "grad_norm": 0.20398795355439606, "learning_rate": 0.00013148097925265212, "loss": 0.4335, "step": 168 }, { "epoch": 1.5196518809657495, "grad_norm": 0.3190354573542016, "learning_rate": 0.00013030698742313152, "loss": 0.4206, "step": 169 }, { "epoch": 1.5286355979786637, "grad_norm": 0.3741573095309751, "learning_rate": 0.00012913110009552428, "loss": 0.4322, "step": 170 }, { "epoch": 1.5376193149915778, "grad_norm": 0.2617731827572085, "learning_rate": 0.0001279534470830984, "loss": 0.4282, "step": 171 }, { "epoch": 1.546603032004492, "grad_norm": 0.31625614240914174, "learning_rate": 0.00012677415839404646, "loss": 0.4315, "step": 172 }, { "epoch": 1.5555867490174058, "grad_norm": 0.3559962927646376, "learning_rate": 0.00012559336421713333, "loss": 0.4255, "step": 173 }, { "epoch": 1.56457046603032, "grad_norm": 0.23926556769719626, "learning_rate": 0.00012441119490732357, "loss": 0.4197, "step": 174 }, { "epoch": 1.573554183043234, "grad_norm": 0.19009337870934856, "learning_rate": 0.00012322778097139102, "loss": 0.4333, "step": 175 }, { "epoch": 1.5825379000561481, "grad_norm": 0.2467283593538461, "learning_rate": 0.00012204325305351117, "loss": 0.4327, "step": 176 }, { "epoch": 1.5915216170690623, "grad_norm": 0.20155153552210148, "learning_rate": 0.00012085774192083878, "loss": 0.424, "step": 177 }, { "epoch": 1.6005053340819764, "grad_norm": 0.2055516806210948, "learning_rate": 0.00011967137844907157, "loss": 0.4198, "step": 178 }, { "epoch": 1.6094890510948905, "grad_norm": 0.21888134195759792, "learning_rate": 0.00011848429360800205, "loss": 0.4279, "step": 179 }, { "epoch": 1.6184727681078046, "grad_norm": 0.18186305811122733, "learning_rate": 0.00011729661844705912, "loss": 0.4249, "step": 180 }, { "epoch": 1.6274564851207187, "grad_norm": 0.13875458201438734, "learning_rate": 0.00011610848408084054, "loss": 0.4275, "step": 181 }, { "epoch": 1.6364402021336328, "grad_norm": 0.1717023475061491, "learning_rate": 0.0001149200216746385, "loss": 0.4225, "step": 182 }, { "epoch": 1.645423919146547, "grad_norm": 0.15952136716834545, "learning_rate": 0.00011373136242995958, "loss": 0.4153, "step": 183 }, { "epoch": 1.654407636159461, "grad_norm": 0.15537668646212147, "learning_rate": 0.00011254263757004044, "loss": 0.4164, "step": 184 }, { "epoch": 1.6633913531723752, "grad_norm": 0.160282534660321, "learning_rate": 0.00011135397832536151, "loss": 0.4253, "step": 185 }, { "epoch": 1.6723750701852893, "grad_norm": 0.16123326071336702, "learning_rate": 0.00011016551591915949, "loss": 0.4239, "step": 186 }, { "epoch": 1.6813587871982034, "grad_norm": 0.148970328863511, "learning_rate": 0.0001089773815529409, "loss": 0.4245, "step": 187 }, { "epoch": 1.6903425042111173, "grad_norm": 0.16045971509449883, "learning_rate": 0.00010778970639199795, "loss": 0.4198, "step": 188 }, { "epoch": 1.6993262212240314, "grad_norm": 0.16138201760304008, "learning_rate": 0.00010660262155092845, "loss": 0.4275, "step": 189 }, { "epoch": 1.7083099382369455, "grad_norm": 0.16720529564166428, "learning_rate": 0.00010541625807916123, "loss": 0.4175, "step": 190 }, { "epoch": 1.7172936552498597, "grad_norm": 0.16645688605969786, "learning_rate": 0.00010423074694648884, "loss": 0.4272, "step": 191 }, { "epoch": 1.7262773722627736, "grad_norm": 0.23842852394974162, "learning_rate": 0.00010304621902860899, "loss": 0.4274, "step": 192 }, { "epoch": 1.7352610892756877, "grad_norm": 0.23233366002775901, "learning_rate": 0.00010186280509267643, "loss": 0.4275, "step": 193 }, { "epoch": 1.7442448062886018, "grad_norm": 0.16088479076799017, "learning_rate": 0.00010068063578286667, "loss": 0.4172, "step": 194 }, { "epoch": 1.753228523301516, "grad_norm": 0.1990526955294185, "learning_rate": 9.949984160595352e-05, "loss": 0.4279, "step": 195 }, { "epoch": 1.76221224031443, "grad_norm": 0.2080695351190495, "learning_rate": 9.832055291690164e-05, "loss": 0.4315, "step": 196 }, { "epoch": 1.7711959573273441, "grad_norm": 0.1890634090210766, "learning_rate": 9.714289990447578e-05, "loss": 0.4182, "step": 197 }, { "epoch": 1.7801796743402583, "grad_norm": 0.20016634340121803, "learning_rate": 9.596701257686851e-05, "loss": 0.4119, "step": 198 }, { "epoch": 1.7891633913531724, "grad_norm": 0.23107873036017035, "learning_rate": 9.479302074734792e-05, "loss": 0.4162, "step": 199 }, { "epoch": 1.7981471083660865, "grad_norm": 0.1830895281774673, "learning_rate": 9.362105401992656e-05, "loss": 0.4177, "step": 200 }, { "epoch": 1.8071308253790006, "grad_norm": 0.16770209709045605, "learning_rate": 9.245124177505392e-05, "loss": 0.421, "step": 201 }, { "epoch": 1.8161145423919147, "grad_norm": 0.2297597515397976, "learning_rate": 9.12837131553331e-05, "loss": 0.434, "step": 202 }, { "epoch": 1.8250982594048288, "grad_norm": 0.16824933783531398, "learning_rate": 9.011859705126419e-05, "loss": 0.418, "step": 203 }, { "epoch": 1.834081976417743, "grad_norm": 0.16377185436749506, "learning_rate": 8.895602208701511e-05, "loss": 0.4113, "step": 204 }, { "epoch": 1.843065693430657, "grad_norm": 0.22604619484876676, "learning_rate": 8.779611660622215e-05, "loss": 0.4239, "step": 205 }, { "epoch": 1.8520494104435712, "grad_norm": 0.2215922382832308, "learning_rate": 8.66390086578212e-05, "loss": 0.4121, "step": 206 }, { "epoch": 1.861033127456485, "grad_norm": 0.20334095034573255, "learning_rate": 8.548482598191181e-05, "loss": 0.4193, "step": 207 }, { "epoch": 1.8700168444693992, "grad_norm": 0.1829539632348045, "learning_rate": 8.433369599565508e-05, "loss": 0.4296, "step": 208 }, { "epoch": 1.8790005614823133, "grad_norm": 0.15716250429615242, "learning_rate": 8.318574577920736e-05, "loss": 0.4164, "step": 209 }, { "epoch": 1.8879842784952274, "grad_norm": 0.23307948623719757, "learning_rate": 8.204110206169114e-05, "loss": 0.4172, "step": 210 }, { "epoch": 1.8969679955081415, "grad_norm": 0.1761132573267027, "learning_rate": 8.08998912072046e-05, "loss": 0.4167, "step": 211 }, { "epoch": 1.9059517125210554, "grad_norm": 0.1690942036116379, "learning_rate": 7.976223920087153e-05, "loss": 0.4247, "step": 212 }, { "epoch": 1.9149354295339696, "grad_norm": 0.17381964913986805, "learning_rate": 7.86282716349331e-05, "loss": 0.4076, "step": 213 }, { "epoch": 1.9239191465468837, "grad_norm": 0.1558189761050885, "learning_rate": 7.749811369488296e-05, "loss": 0.425, "step": 214 }, { "epoch": 1.9329028635597978, "grad_norm": 0.16691148084521792, "learning_rate": 7.637189014564727e-05, "loss": 0.4152, "step": 215 }, { "epoch": 1.941886580572712, "grad_norm": 0.14084306621184936, "learning_rate": 7.524972531781114e-05, "loss": 0.4201, "step": 216 }, { "epoch": 1.950870297585626, "grad_norm": 0.1673195421371356, "learning_rate": 7.413174309389303e-05, "loss": 0.4084, "step": 217 }, { "epoch": 1.9598540145985401, "grad_norm": 0.16765695738584874, "learning_rate": 7.301806689466864e-05, "loss": 0.4275, "step": 218 }, { "epoch": 1.9688377316114543, "grad_norm": 0.16869078852623143, "learning_rate": 7.190881966554573e-05, "loss": 0.418, "step": 219 }, { "epoch": 1.9778214486243684, "grad_norm": 0.16808033042759418, "learning_rate": 7.080412386299138e-05, "loss": 0.4141, "step": 220 }, { "epoch": 1.9868051656372825, "grad_norm": 0.19910528453140844, "learning_rate": 6.970410144101348e-05, "loss": 0.4029, "step": 221 }, { "epoch": 1.9957888826501966, "grad_norm": 0.20801296591215268, "learning_rate": 6.860887383769717e-05, "loss": 0.4266, "step": 222 }, { "epoch": 2.0061763054463784, "grad_norm": 0.3587587352425674, "learning_rate": 6.751856196179887e-05, "loss": 0.7214, "step": 223 }, { "epoch": 2.0151600224592925, "grad_norm": 0.387007013989843, "learning_rate": 6.643328617939827e-05, "loss": 0.3541, "step": 224 }, { "epoch": 2.0241437394722066, "grad_norm": 0.27250154771422735, "learning_rate": 6.53531663006105e-05, "loss": 0.3741, "step": 225 }, { "epoch": 2.0331274564851207, "grad_norm": 0.31369626134171097, "learning_rate": 6.427832156635937e-05, "loss": 0.3623, "step": 226 }, { "epoch": 2.042111173498035, "grad_norm": 0.3899920698764709, "learning_rate": 6.320887063521393e-05, "loss": 0.3666, "step": 227 }, { "epoch": 2.051094890510949, "grad_norm": 0.22822552402245974, "learning_rate": 6.214493157028881e-05, "loss": 0.3588, "step": 228 }, { "epoch": 2.060078607523863, "grad_norm": 0.3578216864292053, "learning_rate": 6.108662182621064e-05, "loss": 0.3734, "step": 229 }, { "epoch": 2.069062324536777, "grad_norm": 0.25875022892169425, "learning_rate": 6.003405823615149e-05, "loss": 0.3595, "step": 230 }, { "epoch": 2.0780460415496913, "grad_norm": 0.2595984420837872, "learning_rate": 5.898735699893096e-05, "loss": 0.3605, "step": 231 }, { "epoch": 2.0870297585626054, "grad_norm": 0.298435693379116, "learning_rate": 5.794663366618828e-05, "loss": 0.3632, "step": 232 }, { "epoch": 2.0960134755755195, "grad_norm": 0.2237606329851956, "learning_rate": 5.691200312962588e-05, "loss": 0.3498, "step": 233 }, { "epoch": 2.1049971925884337, "grad_norm": 0.26971282475418673, "learning_rate": 5.5883579608325826e-05, "loss": 0.3629, "step": 234 }, { "epoch": 2.1139809096013478, "grad_norm": 0.2272539853787037, "learning_rate": 5.486147663614048e-05, "loss": 0.3616, "step": 235 }, { "epoch": 2.1229646266142614, "grad_norm": 0.2772404901897913, "learning_rate": 5.3845807049158815e-05, "loss": 0.3568, "step": 236 }, { "epoch": 2.1319483436271756, "grad_norm": 0.18005249012493324, "learning_rate": 5.2836682973249665e-05, "loss": 0.3655, "step": 237 }, { "epoch": 2.1409320606400897, "grad_norm": 0.25329620107734796, "learning_rate": 5.1834215811683654e-05, "loss": 0.3549, "step": 238 }, { "epoch": 2.149915777653004, "grad_norm": 0.16701885107466424, "learning_rate": 5.0838516232834614e-05, "loss": 0.3566, "step": 239 }, { "epoch": 2.158899494665918, "grad_norm": 0.2641235688031744, "learning_rate": 4.9849694157962234e-05, "loss": 0.3588, "step": 240 }, { "epoch": 2.167883211678832, "grad_norm": 0.15619382389891984, "learning_rate": 4.886785874907724e-05, "loss": 0.3488, "step": 241 }, { "epoch": 2.176866928691746, "grad_norm": 0.2501699730380541, "learning_rate": 4.7893118396890284e-05, "loss": 0.3715, "step": 242 }, { "epoch": 2.1858506457046603, "grad_norm": 0.16211610093557965, "learning_rate": 4.6925580708846104e-05, "loss": 0.3765, "step": 243 }, { "epoch": 2.1948343627175744, "grad_norm": 0.21430975519576806, "learning_rate": 4.596535249724404e-05, "loss": 0.3519, "step": 244 }, { "epoch": 2.2038180797304885, "grad_norm": 0.1482343492785817, "learning_rate": 4.501253976744641e-05, "loss": 0.3561, "step": 245 }, { "epoch": 2.2128017967434026, "grad_norm": 0.18600390573451364, "learning_rate": 4.406724770617595e-05, "loss": 0.3515, "step": 246 }, { "epoch": 2.2217855137563167, "grad_norm": 0.13270293221384985, "learning_rate": 4.3129580669903586e-05, "loss": 0.3612, "step": 247 }, { "epoch": 2.230769230769231, "grad_norm": 0.15896036552877565, "learning_rate": 4.2199642173327955e-05, "loss": 0.3482, "step": 248 }, { "epoch": 2.239752947782145, "grad_norm": 0.13144953626310965, "learning_rate": 4.127753487794768e-05, "loss": 0.3559, "step": 249 }, { "epoch": 2.248736664795059, "grad_norm": 0.14893603771330527, "learning_rate": 4.036336058072806e-05, "loss": 0.356, "step": 250 }, { "epoch": 2.257720381807973, "grad_norm": 0.13694686666323366, "learning_rate": 3.9457220202863104e-05, "loss": 0.3606, "step": 251 }, { "epoch": 2.2667040988208873, "grad_norm": 0.14835982621693597, "learning_rate": 3.855921377863414e-05, "loss": 0.3667, "step": 252 }, { "epoch": 2.2756878158338014, "grad_norm": 0.13903851563190525, "learning_rate": 3.766944044436649e-05, "loss": 0.3548, "step": 253 }, { "epoch": 2.2846715328467155, "grad_norm": 0.13419589581861444, "learning_rate": 3.678799842748521e-05, "loss": 0.3716, "step": 254 }, { "epoch": 2.293655249859629, "grad_norm": 0.14685606142797603, "learning_rate": 3.5914985035671156e-05, "loss": 0.3571, "step": 255 }, { "epoch": 2.3026389668725433, "grad_norm": 0.1304570595276937, "learning_rate": 3.5050496646118584e-05, "loss": 0.3655, "step": 256 }, { "epoch": 2.3116226838854574, "grad_norm": 0.12826171716327225, "learning_rate": 3.4194628694895594e-05, "loss": 0.364, "step": 257 }, { "epoch": 2.3206064008983716, "grad_norm": 0.11075309460103316, "learning_rate": 3.334747566640824e-05, "loss": 0.351, "step": 258 }, { "epoch": 2.3295901179112857, "grad_norm": 0.1312845965125087, "learning_rate": 3.2509131082969915e-05, "loss": 0.3687, "step": 259 }, { "epoch": 2.3385738349242, "grad_norm": 0.11983646002909432, "learning_rate": 3.167968749447683e-05, "loss": 0.3454, "step": 260 }, { "epoch": 2.347557551937114, "grad_norm": 0.11479008167106183, "learning_rate": 3.0859236468190844e-05, "loss": 0.3654, "step": 261 }, { "epoch": 2.356541268950028, "grad_norm": 0.11632197031117864, "learning_rate": 3.00478685786309e-05, "loss": 0.3617, "step": 262 }, { "epoch": 2.365524985962942, "grad_norm": 0.12594569973876463, "learning_rate": 2.9245673397573843e-05, "loss": 0.3582, "step": 263 }, { "epoch": 2.3745087029758563, "grad_norm": 0.124698262560256, "learning_rate": 2.8452739484166123e-05, "loss": 0.3628, "step": 264 }, { "epoch": 2.3834924199887704, "grad_norm": 0.11918130322626423, "learning_rate": 2.7669154375147227e-05, "loss": 0.3617, "step": 265 }, { "epoch": 2.3924761370016845, "grad_norm": 0.11920547285141371, "learning_rate": 2.6895004575185922e-05, "loss": 0.3657, "step": 266 }, { "epoch": 2.4014598540145986, "grad_norm": 0.10937163911147071, "learning_rate": 2.6130375547330496e-05, "loss": 0.3519, "step": 267 }, { "epoch": 2.4104435710275127, "grad_norm": 0.14631208309450952, "learning_rate": 2.5375351703574044e-05, "loss": 0.3631, "step": 268 }, { "epoch": 2.419427288040427, "grad_norm": 0.1265003307705914, "learning_rate": 2.4630016395535618e-05, "loss": 0.3679, "step": 269 }, { "epoch": 2.428411005053341, "grad_norm": 0.11607388860612512, "learning_rate": 2.389445190525866e-05, "loss": 0.3475, "step": 270 }, { "epoch": 2.437394722066255, "grad_norm": 0.13926049695780746, "learning_rate": 2.316873943612722e-05, "loss": 0.3509, "step": 271 }, { "epoch": 2.446378439079169, "grad_norm": 0.12235594276612768, "learning_rate": 2.2452959103901613e-05, "loss": 0.3597, "step": 272 }, { "epoch": 2.4553621560920833, "grad_norm": 0.11752333623641517, "learning_rate": 2.1747189927873877e-05, "loss": 0.3515, "step": 273 }, { "epoch": 2.464345873104997, "grad_norm": 0.11683542380499232, "learning_rate": 2.1051509822144332e-05, "loss": 0.3695, "step": 274 }, { "epoch": 2.473329590117911, "grad_norm": 0.12084518265976886, "learning_rate": 2.0365995587020234e-05, "loss": 0.3592, "step": 275 }, { "epoch": 2.482313307130825, "grad_norm": 0.10536387180642533, "learning_rate": 1.969072290053725e-05, "loss": 0.3634, "step": 276 }, { "epoch": 2.4912970241437393, "grad_norm": 0.12459678439311449, "learning_rate": 1.902576631010499e-05, "loss": 0.3444, "step": 277 }, { "epoch": 2.5002807411566534, "grad_norm": 0.10423725455797198, "learning_rate": 1.8371199224277213e-05, "loss": 0.3446, "step": 278 }, { "epoch": 2.5092644581695676, "grad_norm": 0.09677902356707294, "learning_rate": 1.772709390464784e-05, "loss": 0.3565, "step": 279 }, { "epoch": 2.5182481751824817, "grad_norm": 0.1056308438167975, "learning_rate": 1.7093521457873555e-05, "loss": 0.3667, "step": 280 }, { "epoch": 2.527231892195396, "grad_norm": 0.09014603021736677, "learning_rate": 1.647055182782392e-05, "loss": 0.3486, "step": 281 }, { "epoch": 2.53621560920831, "grad_norm": 0.10287065883652714, "learning_rate": 1.5858253787859857e-05, "loss": 0.3525, "step": 282 }, { "epoch": 2.545199326221224, "grad_norm": 0.09914950751521003, "learning_rate": 1.5256694933241261e-05, "loss": 0.3675, "step": 283 }, { "epoch": 2.554183043234138, "grad_norm": 0.09453069458816908, "learning_rate": 1.466594167366493e-05, "loss": 0.3443, "step": 284 }, { "epoch": 2.5631667602470523, "grad_norm": 0.10390944505822122, "learning_rate": 1.4086059225933016e-05, "loss": 0.3719, "step": 285 }, { "epoch": 2.5721504772599664, "grad_norm": 0.10438780375230042, "learning_rate": 1.3517111606753471e-05, "loss": 0.3605, "step": 286 }, { "epoch": 2.5811341942728805, "grad_norm": 0.09859350856353293, "learning_rate": 1.2959161625672802e-05, "loss": 0.3616, "step": 287 }, { "epoch": 2.5901179112857946, "grad_norm": 0.09622498419041688, "learning_rate": 1.2412270878142156e-05, "loss": 0.3501, "step": 288 }, { "epoch": 2.5991016282987087, "grad_norm": 0.09113584600196091, "learning_rate": 1.1876499738717436e-05, "loss": 0.3503, "step": 289 }, { "epoch": 2.608085345311623, "grad_norm": 0.09741692104467439, "learning_rate": 1.1351907354394194e-05, "loss": 0.3657, "step": 290 }, { "epoch": 2.6170690623245365, "grad_norm": 0.09206757777787074, "learning_rate": 1.0838551638078013e-05, "loss": 0.3603, "step": 291 }, { "epoch": 2.626052779337451, "grad_norm": 0.08188402631450145, "learning_rate": 1.0336489262191212e-05, "loss": 0.3462, "step": 292 }, { "epoch": 2.6350364963503647, "grad_norm": 0.08808466701218914, "learning_rate": 9.845775652416357e-06, "loss": 0.3632, "step": 293 }, { "epoch": 2.6440202133632793, "grad_norm": 0.0943618643568796, "learning_rate": 9.366464981577584e-06, "loss": 0.3594, "step": 294 }, { "epoch": 2.653003930376193, "grad_norm": 0.09593739977292585, "learning_rate": 8.89861016365997e-06, "loss": 0.3515, "step": 295 }, { "epoch": 2.661987647389107, "grad_norm": 0.08248015664645339, "learning_rate": 8.442262847968263e-06, "loss": 0.3608, "step": 296 }, { "epoch": 2.670971364402021, "grad_norm": 0.08411868089198224, "learning_rate": 7.997473413424846e-06, "loss": 0.3592, "step": 297 }, { "epoch": 2.6799550814149353, "grad_norm": 0.09490876979791782, "learning_rate": 7.564290963008149e-06, "loss": 0.358, "step": 298 }, { "epoch": 2.6889387984278494, "grad_norm": 0.08867586845979594, "learning_rate": 7.142763318331872e-06, "loss": 0.351, "step": 299 }, { "epoch": 2.6979225154407636, "grad_norm": 0.08779362567358494, "learning_rate": 6.732937014365695e-06, "loss": 0.3494, "step": 300 }, { "epoch": 2.7069062324536777, "grad_norm": 0.08301900637144422, "learning_rate": 6.3348572942979654e-06, "loss": 0.3587, "step": 301 }, { "epoch": 2.715889949466592, "grad_norm": 0.08891924464679533, "learning_rate": 5.948568104541074e-06, "loss": 0.3651, "step": 302 }, { "epoch": 2.724873666479506, "grad_norm": 0.08730964097529133, "learning_rate": 5.574112089879872e-06, "loss": 0.3523, "step": 303 }, { "epoch": 2.73385738349242, "grad_norm": 0.08450880465007451, "learning_rate": 5.211530588763962e-06, "loss": 0.3482, "step": 304 }, { "epoch": 2.742841100505334, "grad_norm": 0.08342255207562929, "learning_rate": 4.860863628744007e-06, "loss": 0.3564, "step": 305 }, { "epoch": 2.7518248175182483, "grad_norm": 0.08975850015832335, "learning_rate": 4.522149922052897e-06, "loss": 0.3491, "step": 306 }, { "epoch": 2.7608085345311624, "grad_norm": 0.08240771555091773, "learning_rate": 4.195426861332049e-06, "loss": 0.3588, "step": 307 }, { "epoch": 2.7697922515440765, "grad_norm": 0.08798778487887464, "learning_rate": 3.880730515503412e-06, "loss": 0.3427, "step": 308 }, { "epoch": 2.7787759685569906, "grad_norm": 0.08113053094229192, "learning_rate": 3.57809562578763e-06, "loss": 0.3728, "step": 309 }, { "epoch": 2.7877596855699043, "grad_norm": 0.08279617250380948, "learning_rate": 3.2875556018687533e-06, "loss": 0.3418, "step": 310 }, { "epoch": 2.796743402582819, "grad_norm": 0.07729029098691793, "learning_rate": 3.0091425182058514e-06, "loss": 0.3617, "step": 311 }, { "epoch": 2.8057271195957325, "grad_norm": 0.08856029522706829, "learning_rate": 2.742887110492231e-06, "loss": 0.3602, "step": 312 }, { "epoch": 2.814710836608647, "grad_norm": 0.09170437352580232, "learning_rate": 2.4888187722622945e-06, "loss": 0.3532, "step": 313 }, { "epoch": 2.8236945536215607, "grad_norm": 0.08713999383289339, "learning_rate": 2.2469655516466e-06, "loss": 0.3579, "step": 314 }, { "epoch": 2.832678270634475, "grad_norm": 0.08158645722200215, "learning_rate": 2.017354148275491e-06, "loss": 0.3527, "step": 315 }, { "epoch": 2.841661987647389, "grad_norm": 0.0775962718555909, "learning_rate": 1.8000099103314957e-06, "loss": 0.357, "step": 316 }, { "epoch": 2.850645704660303, "grad_norm": 0.07417615515514962, "learning_rate": 1.5949568317510827e-06, "loss": 0.3633, "step": 317 }, { "epoch": 2.859629421673217, "grad_norm": 0.07076814193057895, "learning_rate": 1.402217549575769e-06, "loss": 0.357, "step": 318 }, { "epoch": 2.8686131386861313, "grad_norm": 0.08051085296339765, "learning_rate": 1.2218133414530984e-06, "loss": 0.3528, "step": 319 }, { "epoch": 2.8775968556990454, "grad_norm": 0.07853549988975052, "learning_rate": 1.0537641232876473e-06, "loss": 0.3614, "step": 320 }, { "epoch": 2.8865805727119596, "grad_norm": 0.07073622608834755, "learning_rate": 8.980884470424321e-07, "loss": 0.3551, "step": 321 }, { "epoch": 2.8955642897248737, "grad_norm": 0.07143610759857762, "learning_rate": 7.548034986908066e-07, "loss": 0.3545, "step": 322 }, { "epoch": 2.904548006737788, "grad_norm": 0.0749903587104429, "learning_rate": 6.239250963192269e-07, "loss": 0.3583, "step": 323 }, { "epoch": 2.913531723750702, "grad_norm": 0.07746634488254509, "learning_rate": 5.054676883809827e-07, "loss": 0.3533, "step": 324 }, { "epoch": 2.922515440763616, "grad_norm": 0.07444763137173055, "learning_rate": 3.994443521011485e-07, "loss": 0.3627, "step": 325 }, { "epoch": 2.93149915777653, "grad_norm": 0.06987734521185171, "learning_rate": 3.058667920329281e-07, "loss": 0.348, "step": 326 }, { "epoch": 2.9404828747894443, "grad_norm": 0.07171007945104789, "learning_rate": 2.2474533876546995e-07, "loss": 0.3455, "step": 327 }, { "epoch": 2.9494665918023584, "grad_norm": 0.0727402518849161, "learning_rate": 1.560889477834654e-07, "loss": 0.3579, "step": 328 }, { "epoch": 2.958450308815272, "grad_norm": 0.07322322016212719, "learning_rate": 9.99051984784689e-08, "loss": 0.3528, "step": 329 }, { "epoch": 2.9674340258281866, "grad_norm": 0.07567114973675214, "learning_rate": 5.620029331218986e-08, "loss": 0.3564, "step": 330 }, { "epoch": 2.9764177428411003, "grad_norm": 0.07625451494420618, "learning_rate": 2.4979057131732006e-08, "loss": 0.3594, "step": 331 }, { "epoch": 2.985401459854015, "grad_norm": 0.07133050968488629, "learning_rate": 6.2449366369555176e-09, "loss": 0.3521, "step": 332 }, { "epoch": 2.9943851768669285, "grad_norm": 0.07662216533129715, "learning_rate": 0.0, "loss": 0.3558, "step": 333 }, { "epoch": 2.9943851768669285, "step": 333, "total_flos": 4607052149424128.0, "train_loss": 0.5244213240640657, "train_runtime": 115493.2352, "train_samples_per_second": 2.96, "train_steps_per_second": 0.003 } ], "logging_steps": 1.0, "max_steps": 333, "num_input_tokens_seen": 0, "num_train_epochs": 3, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 4607052149424128.0, "train_batch_size": 1, "trial_name": null, "trial_params": null }