{ "best_metric": null, "best_model_checkpoint": null, "epoch": 1.0, "eval_steps": 500, "global_step": 518, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0019305019305019305, "grad_norm": 9.087823867797852, "learning_rate": 9.980694980694981e-06, "loss": 763.5779, "step": 1 }, { "epoch": 0.003861003861003861, "grad_norm": 8.02014446258545, "learning_rate": 9.961389961389962e-06, "loss": 742.0983, "step": 2 }, { "epoch": 0.005791505791505791, "grad_norm": 6.846795558929443, "learning_rate": 9.942084942084944e-06, "loss": 754.556, "step": 3 }, { "epoch": 0.007722007722007722, "grad_norm": 5.642313003540039, "learning_rate": 9.922779922779924e-06, "loss": 760.8958, "step": 4 }, { "epoch": 0.009652509652509652, "grad_norm": 4.954174995422363, "learning_rate": 9.903474903474905e-06, "loss": 773.4886, "step": 5 }, { "epoch": 0.011583011583011582, "grad_norm": 4.209826946258545, "learning_rate": 9.884169884169885e-06, "loss": 752.0172, "step": 6 }, { "epoch": 0.013513513513513514, "grad_norm": 3.8238916397094727, "learning_rate": 9.864864864864865e-06, "loss": 765.1586, "step": 7 }, { "epoch": 0.015444015444015444, "grad_norm": 3.472975730895996, "learning_rate": 9.845559845559847e-06, "loss": 758.0544, "step": 8 }, { "epoch": 0.017374517374517374, "grad_norm": 3.3365626335144043, "learning_rate": 9.826254826254828e-06, "loss": 760.6899, "step": 9 }, { "epoch": 0.019305019305019305, "grad_norm": 3.205853223800659, "learning_rate": 9.806949806949808e-06, "loss": 763.1088, "step": 10 }, { "epoch": 0.021235521235521235, "grad_norm": 2.9912545680999756, "learning_rate": 9.787644787644789e-06, "loss": 751.3445, "step": 11 }, { "epoch": 0.023166023166023165, "grad_norm": 2.9894907474517822, "learning_rate": 9.768339768339769e-06, "loss": 762.5296, "step": 12 }, { "epoch": 0.025096525096525095, "grad_norm": 2.973778009414673, "learning_rate": 9.749034749034751e-06, "loss": 766.5692, "step": 13 }, { "epoch": 0.02702702702702703, "grad_norm": 2.866206645965576, "learning_rate": 9.729729729729732e-06, "loss": 759.659, "step": 14 }, { "epoch": 0.02895752895752896, "grad_norm": 2.7454729080200195, "learning_rate": 9.71042471042471e-06, "loss": 749.6011, "step": 15 }, { "epoch": 0.03088803088803089, "grad_norm": 2.7870497703552246, "learning_rate": 9.69111969111969e-06, "loss": 757.2128, "step": 16 }, { "epoch": 0.032818532818532815, "grad_norm": 2.727757453918457, "learning_rate": 9.671814671814673e-06, "loss": 752.8752, "step": 17 }, { "epoch": 0.03474903474903475, "grad_norm": 2.761842966079712, "learning_rate": 9.652509652509653e-06, "loss": 763.5405, "step": 18 }, { "epoch": 0.03667953667953668, "grad_norm": 2.64555025100708, "learning_rate": 9.633204633204634e-06, "loss": 752.1604, "step": 19 }, { "epoch": 0.03861003861003861, "grad_norm": 2.604011297225952, "learning_rate": 9.613899613899614e-06, "loss": 745.1989, "step": 20 }, { "epoch": 0.04054054054054054, "grad_norm": 2.5989909172058105, "learning_rate": 9.594594594594594e-06, "loss": 744.575, "step": 21 }, { "epoch": 0.04247104247104247, "grad_norm": 2.646167278289795, "learning_rate": 9.575289575289576e-06, "loss": 753.7839, "step": 22 }, { "epoch": 0.0444015444015444, "grad_norm": 2.6674346923828125, "learning_rate": 9.555984555984557e-06, "loss": 764.2125, "step": 23 }, { "epoch": 0.04633204633204633, "grad_norm": 2.6234052181243896, "learning_rate": 9.536679536679537e-06, "loss": 753.5481, "step": 24 }, { "epoch": 0.04826254826254826, "grad_norm": 2.550206422805786, "learning_rate": 9.517374517374518e-06, "loss": 744.3314, "step": 25 }, { "epoch": 0.05019305019305019, "grad_norm": 2.634002208709717, "learning_rate": 9.498069498069498e-06, "loss": 756.7723, "step": 26 }, { "epoch": 0.052123552123552123, "grad_norm": 2.632336378097534, "learning_rate": 9.47876447876448e-06, "loss": 756.9872, "step": 27 }, { "epoch": 0.05405405405405406, "grad_norm": 2.628897190093994, "learning_rate": 9.45945945945946e-06, "loss": 756.0316, "step": 28 }, { "epoch": 0.055984555984555984, "grad_norm": 2.6032018661499023, "learning_rate": 9.440154440154441e-06, "loss": 754.0662, "step": 29 }, { "epoch": 0.05791505791505792, "grad_norm": 2.5998055934906006, "learning_rate": 9.420849420849421e-06, "loss": 749.3468, "step": 30 }, { "epoch": 0.059845559845559844, "grad_norm": 2.623134136199951, "learning_rate": 9.401544401544402e-06, "loss": 753.1194, "step": 31 }, { "epoch": 0.06177606177606178, "grad_norm": 2.636357069015503, "learning_rate": 9.382239382239384e-06, "loss": 755.1291, "step": 32 }, { "epoch": 0.0637065637065637, "grad_norm": 2.709622621536255, "learning_rate": 9.362934362934364e-06, "loss": 774.9333, "step": 33 }, { "epoch": 0.06563706563706563, "grad_norm": 2.6019175052642822, "learning_rate": 9.343629343629345e-06, "loss": 758.8455, "step": 34 }, { "epoch": 0.06756756756756757, "grad_norm": 2.620014190673828, "learning_rate": 9.324324324324325e-06, "loss": 751.6978, "step": 35 }, { "epoch": 0.0694980694980695, "grad_norm": 2.6363117694854736, "learning_rate": 9.305019305019305e-06, "loss": 755.4987, "step": 36 }, { "epoch": 0.07142857142857142, "grad_norm": 2.697289228439331, "learning_rate": 9.285714285714288e-06, "loss": 769.0797, "step": 37 }, { "epoch": 0.07335907335907337, "grad_norm": 2.684602975845337, "learning_rate": 9.266409266409268e-06, "loss": 771.785, "step": 38 }, { "epoch": 0.07528957528957529, "grad_norm": 2.6869068145751953, "learning_rate": 9.247104247104248e-06, "loss": 771.0751, "step": 39 }, { "epoch": 0.07722007722007722, "grad_norm": 2.65386962890625, "learning_rate": 9.227799227799229e-06, "loss": 766.6141, "step": 40 }, { "epoch": 0.07915057915057915, "grad_norm": 2.5492303371429443, "learning_rate": 9.20849420849421e-06, "loss": 741.012, "step": 41 }, { "epoch": 0.08108108108108109, "grad_norm": 2.659393548965454, "learning_rate": 9.189189189189191e-06, "loss": 761.9696, "step": 42 }, { "epoch": 0.08301158301158301, "grad_norm": 2.6672518253326416, "learning_rate": 9.169884169884172e-06, "loss": 762.8466, "step": 43 }, { "epoch": 0.08494208494208494, "grad_norm": 2.60377836227417, "learning_rate": 9.15057915057915e-06, "loss": 758.5039, "step": 44 }, { "epoch": 0.08687258687258688, "grad_norm": 2.622608184814453, "learning_rate": 9.13127413127413e-06, "loss": 750.5551, "step": 45 }, { "epoch": 0.0888030888030888, "grad_norm": 2.6671438217163086, "learning_rate": 9.111969111969113e-06, "loss": 766.5024, "step": 46 }, { "epoch": 0.09073359073359073, "grad_norm": 2.5831432342529297, "learning_rate": 9.092664092664093e-06, "loss": 747.624, "step": 47 }, { "epoch": 0.09266409266409266, "grad_norm": 2.61625075340271, "learning_rate": 9.073359073359074e-06, "loss": 759.9783, "step": 48 }, { "epoch": 0.0945945945945946, "grad_norm": 2.6503899097442627, "learning_rate": 9.054054054054054e-06, "loss": 766.5831, "step": 49 }, { "epoch": 0.09652509652509653, "grad_norm": 2.713566303253174, "learning_rate": 9.034749034749034e-06, "loss": 764.6612, "step": 50 }, { "epoch": 0.09845559845559845, "grad_norm": 2.6607675552368164, "learning_rate": 9.015444015444017e-06, "loss": 759.2146, "step": 51 }, { "epoch": 0.10038610038610038, "grad_norm": 2.703293800354004, "learning_rate": 8.996138996138997e-06, "loss": 771.3636, "step": 52 }, { "epoch": 0.10231660231660232, "grad_norm": 2.6380341053009033, "learning_rate": 8.976833976833977e-06, "loss": 762.6519, "step": 53 }, { "epoch": 0.10424710424710425, "grad_norm": 2.6804656982421875, "learning_rate": 8.957528957528958e-06, "loss": 764.4567, "step": 54 }, { "epoch": 0.10617760617760617, "grad_norm": 2.6722824573516846, "learning_rate": 8.938223938223938e-06, "loss": 762.7928, "step": 55 }, { "epoch": 0.10810810810810811, "grad_norm": 2.5956473350524902, "learning_rate": 8.91891891891892e-06, "loss": 757.1573, "step": 56 }, { "epoch": 0.11003861003861004, "grad_norm": 2.67158579826355, "learning_rate": 8.8996138996139e-06, "loss": 762.9989, "step": 57 }, { "epoch": 0.11196911196911197, "grad_norm": 2.6719319820404053, "learning_rate": 8.880308880308881e-06, "loss": 765.219, "step": 58 }, { "epoch": 0.1138996138996139, "grad_norm": 2.690995931625366, "learning_rate": 8.861003861003861e-06, "loss": 755.7704, "step": 59 }, { "epoch": 0.11583011583011583, "grad_norm": 2.672006845474243, "learning_rate": 8.841698841698842e-06, "loss": 755.6411, "step": 60 }, { "epoch": 0.11776061776061776, "grad_norm": 2.6625969409942627, "learning_rate": 8.822393822393824e-06, "loss": 755.6705, "step": 61 }, { "epoch": 0.11969111969111969, "grad_norm": 2.6644089221954346, "learning_rate": 8.803088803088804e-06, "loss": 755.7343, "step": 62 }, { "epoch": 0.12162162162162163, "grad_norm": 2.6913063526153564, "learning_rate": 8.783783783783785e-06, "loss": 760.8112, "step": 63 }, { "epoch": 0.12355212355212356, "grad_norm": 2.640939474105835, "learning_rate": 8.764478764478765e-06, "loss": 752.9957, "step": 64 }, { "epoch": 0.12548262548262548, "grad_norm": 2.706070899963379, "learning_rate": 8.745173745173746e-06, "loss": 751.4651, "step": 65 }, { "epoch": 0.1274131274131274, "grad_norm": 2.6751809120178223, "learning_rate": 8.725868725868728e-06, "loss": 760.4512, "step": 66 }, { "epoch": 0.12934362934362933, "grad_norm": 2.731750965118408, "learning_rate": 8.706563706563708e-06, "loss": 766.3732, "step": 67 }, { "epoch": 0.13127413127413126, "grad_norm": 2.688737392425537, "learning_rate": 8.687258687258689e-06, "loss": 749.8923, "step": 68 }, { "epoch": 0.13320463320463322, "grad_norm": 2.67354679107666, "learning_rate": 8.667953667953669e-06, "loss": 751.1083, "step": 69 }, { "epoch": 0.13513513513513514, "grad_norm": 2.6414129734039307, "learning_rate": 8.64864864864865e-06, "loss": 733.0116, "step": 70 }, { "epoch": 0.13706563706563707, "grad_norm": 3.0136189460754395, "learning_rate": 8.629343629343631e-06, "loss": 764.5856, "step": 71 }, { "epoch": 0.138996138996139, "grad_norm": 2.7174079418182373, "learning_rate": 8.61003861003861e-06, "loss": 753.4861, "step": 72 }, { "epoch": 0.14092664092664092, "grad_norm": 2.714317560195923, "learning_rate": 8.59073359073359e-06, "loss": 754.7225, "step": 73 }, { "epoch": 0.14285714285714285, "grad_norm": 2.7697298526763916, "learning_rate": 8.571428571428571e-06, "loss": 765.4945, "step": 74 }, { "epoch": 0.14478764478764478, "grad_norm": 2.7908132076263428, "learning_rate": 8.552123552123553e-06, "loss": 779.0895, "step": 75 }, { "epoch": 0.14671814671814673, "grad_norm": 2.7896623611450195, "learning_rate": 8.532818532818533e-06, "loss": 767.8585, "step": 76 }, { "epoch": 0.14864864864864866, "grad_norm": 2.718611240386963, "learning_rate": 8.513513513513514e-06, "loss": 760.3116, "step": 77 }, { "epoch": 0.15057915057915058, "grad_norm": 2.78110671043396, "learning_rate": 8.494208494208494e-06, "loss": 761.649, "step": 78 }, { "epoch": 0.1525096525096525, "grad_norm": 2.822859048843384, "learning_rate": 8.474903474903475e-06, "loss": 766.2858, "step": 79 }, { "epoch": 0.15444015444015444, "grad_norm": 2.719345808029175, "learning_rate": 8.455598455598457e-06, "loss": 755.3741, "step": 80 }, { "epoch": 0.15637065637065636, "grad_norm": 2.710378408432007, "learning_rate": 8.436293436293437e-06, "loss": 758.8147, "step": 81 }, { "epoch": 0.1583011583011583, "grad_norm": 2.7957656383514404, "learning_rate": 8.416988416988418e-06, "loss": 761.3373, "step": 82 }, { "epoch": 0.16023166023166024, "grad_norm": 2.8298914432525635, "learning_rate": 8.397683397683398e-06, "loss": 769.8754, "step": 83 }, { "epoch": 0.16216216216216217, "grad_norm": 2.762465715408325, "learning_rate": 8.378378378378378e-06, "loss": 769.7825, "step": 84 }, { "epoch": 0.1640926640926641, "grad_norm": 2.773038625717163, "learning_rate": 8.35907335907336e-06, "loss": 749.3358, "step": 85 }, { "epoch": 0.16602316602316602, "grad_norm": 2.7703516483306885, "learning_rate": 8.33976833976834e-06, "loss": 757.9563, "step": 86 }, { "epoch": 0.16795366795366795, "grad_norm": 2.930047035217285, "learning_rate": 8.320463320463321e-06, "loss": 764.3655, "step": 87 }, { "epoch": 0.16988416988416988, "grad_norm": 2.7261102199554443, "learning_rate": 8.301158301158302e-06, "loss": 748.2144, "step": 88 }, { "epoch": 0.1718146718146718, "grad_norm": 2.793009042739868, "learning_rate": 8.281853281853282e-06, "loss": 754.8672, "step": 89 }, { "epoch": 0.17374517374517376, "grad_norm": 2.8894758224487305, "learning_rate": 8.262548262548264e-06, "loss": 764.7795, "step": 90 }, { "epoch": 0.17567567567567569, "grad_norm": 2.8510689735412598, "learning_rate": 8.243243243243245e-06, "loss": 753.4908, "step": 91 }, { "epoch": 0.1776061776061776, "grad_norm": 2.757417917251587, "learning_rate": 8.223938223938225e-06, "loss": 760.0729, "step": 92 }, { "epoch": 0.17953667953667954, "grad_norm": 2.8439671993255615, "learning_rate": 8.204633204633205e-06, "loss": 771.863, "step": 93 }, { "epoch": 0.18146718146718147, "grad_norm": 2.8313705921173096, "learning_rate": 8.185328185328186e-06, "loss": 765.174, "step": 94 }, { "epoch": 0.1833976833976834, "grad_norm": 2.7856626510620117, "learning_rate": 8.166023166023168e-06, "loss": 758.1021, "step": 95 }, { "epoch": 0.18532818532818532, "grad_norm": 2.8734123706817627, "learning_rate": 8.146718146718148e-06, "loss": 768.5622, "step": 96 }, { "epoch": 0.18725868725868725, "grad_norm": 2.833714723587036, "learning_rate": 8.127413127413129e-06, "loss": 753.1698, "step": 97 }, { "epoch": 0.1891891891891892, "grad_norm": 2.8014986515045166, "learning_rate": 8.108108108108109e-06, "loss": 752.7419, "step": 98 }, { "epoch": 0.19111969111969113, "grad_norm": 2.7595949172973633, "learning_rate": 8.08880308880309e-06, "loss": 755.9596, "step": 99 }, { "epoch": 0.19305019305019305, "grad_norm": 2.7833123207092285, "learning_rate": 8.06949806949807e-06, "loss": 765.9607, "step": 100 }, { "epoch": 0.19498069498069498, "grad_norm": 2.777548313140869, "learning_rate": 8.05019305019305e-06, "loss": 766.981, "step": 101 }, { "epoch": 0.1969111969111969, "grad_norm": 2.7974212169647217, "learning_rate": 8.03088803088803e-06, "loss": 753.3157, "step": 102 }, { "epoch": 0.19884169884169883, "grad_norm": 3.194911003112793, "learning_rate": 8.011583011583011e-06, "loss": 773.3227, "step": 103 }, { "epoch": 0.20077220077220076, "grad_norm": 2.7325661182403564, "learning_rate": 7.992277992277993e-06, "loss": 746.1143, "step": 104 }, { "epoch": 0.20270270270270271, "grad_norm": 2.740203619003296, "learning_rate": 7.972972972972974e-06, "loss": 750.4532, "step": 105 }, { "epoch": 0.20463320463320464, "grad_norm": 2.764615297317505, "learning_rate": 7.953667953667954e-06, "loss": 757.8738, "step": 106 }, { "epoch": 0.20656370656370657, "grad_norm": 2.783639430999756, "learning_rate": 7.934362934362934e-06, "loss": 759.7949, "step": 107 }, { "epoch": 0.2084942084942085, "grad_norm": 2.933793783187866, "learning_rate": 7.915057915057915e-06, "loss": 771.2281, "step": 108 }, { "epoch": 0.21042471042471042, "grad_norm": 2.795241117477417, "learning_rate": 7.895752895752897e-06, "loss": 747.9264, "step": 109 }, { "epoch": 0.21235521235521235, "grad_norm": 2.755584239959717, "learning_rate": 7.876447876447877e-06, "loss": 746.8754, "step": 110 }, { "epoch": 0.21428571428571427, "grad_norm": 2.802136182785034, "learning_rate": 7.857142857142858e-06, "loss": 745.1733, "step": 111 }, { "epoch": 0.21621621621621623, "grad_norm": 3.5105533599853516, "learning_rate": 7.837837837837838e-06, "loss": 745.6346, "step": 112 }, { "epoch": 0.21814671814671815, "grad_norm": 2.8462111949920654, "learning_rate": 7.818532818532818e-06, "loss": 757.0023, "step": 113 }, { "epoch": 0.22007722007722008, "grad_norm": 3.079543113708496, "learning_rate": 7.7992277992278e-06, "loss": 757.2842, "step": 114 }, { "epoch": 0.222007722007722, "grad_norm": 2.8306119441986084, "learning_rate": 7.779922779922781e-06, "loss": 752.3796, "step": 115 }, { "epoch": 0.22393822393822393, "grad_norm": 3.1957056522369385, "learning_rate": 7.760617760617761e-06, "loss": 753.2894, "step": 116 }, { "epoch": 0.22586872586872586, "grad_norm": 2.90783953666687, "learning_rate": 7.741312741312742e-06, "loss": 753.8042, "step": 117 }, { "epoch": 0.2277992277992278, "grad_norm": 2.8529882431030273, "learning_rate": 7.722007722007722e-06, "loss": 758.0192, "step": 118 }, { "epoch": 0.22972972972972974, "grad_norm": 2.8292617797851562, "learning_rate": 7.702702702702704e-06, "loss": 756.563, "step": 119 }, { "epoch": 0.23166023166023167, "grad_norm": 2.8797755241394043, "learning_rate": 7.683397683397685e-06, "loss": 761.1346, "step": 120 }, { "epoch": 0.2335907335907336, "grad_norm": 2.7390382289886475, "learning_rate": 7.664092664092665e-06, "loss": 750.3627, "step": 121 }, { "epoch": 0.23552123552123552, "grad_norm": 2.7686545848846436, "learning_rate": 7.644787644787645e-06, "loss": 750.1412, "step": 122 }, { "epoch": 0.23745173745173745, "grad_norm": 2.798008680343628, "learning_rate": 7.625482625482627e-06, "loss": 742.588, "step": 123 }, { "epoch": 0.23938223938223938, "grad_norm": 4.088344097137451, "learning_rate": 7.606177606177607e-06, "loss": 758.9719, "step": 124 }, { "epoch": 0.2413127413127413, "grad_norm": 2.9229252338409424, "learning_rate": 7.5868725868725875e-06, "loss": 756.5762, "step": 125 }, { "epoch": 0.24324324324324326, "grad_norm": 3.1966676712036133, "learning_rate": 7.567567567567569e-06, "loss": 768.2112, "step": 126 }, { "epoch": 0.24517374517374518, "grad_norm": 3.255308151245117, "learning_rate": 7.548262548262549e-06, "loss": 768.8184, "step": 127 }, { "epoch": 0.2471042471042471, "grad_norm": 2.803837537765503, "learning_rate": 7.5289575289575304e-06, "loss": 758.1424, "step": 128 }, { "epoch": 0.24903474903474904, "grad_norm": 2.8075811862945557, "learning_rate": 7.509652509652511e-06, "loss": 748.3246, "step": 129 }, { "epoch": 0.25096525096525096, "grad_norm": 3.1468892097473145, "learning_rate": 7.49034749034749e-06, "loss": 759.9876, "step": 130 }, { "epoch": 0.2528957528957529, "grad_norm": 3.324525833129883, "learning_rate": 7.471042471042471e-06, "loss": 756.2051, "step": 131 }, { "epoch": 0.2548262548262548, "grad_norm": 2.8439035415649414, "learning_rate": 7.451737451737452e-06, "loss": 755.392, "step": 132 }, { "epoch": 0.25675675675675674, "grad_norm": 2.882779598236084, "learning_rate": 7.4324324324324324e-06, "loss": 771.8354, "step": 133 }, { "epoch": 0.25868725868725867, "grad_norm": 3.0841407775878906, "learning_rate": 7.413127413127414e-06, "loss": 754.2871, "step": 134 }, { "epoch": 0.2606177606177606, "grad_norm": 3.2702057361602783, "learning_rate": 7.393822393822394e-06, "loss": 742.1606, "step": 135 }, { "epoch": 0.2625482625482625, "grad_norm": 2.7956244945526123, "learning_rate": 7.3745173745173745e-06, "loss": 750.0869, "step": 136 }, { "epoch": 0.2644787644787645, "grad_norm": 3.004718542098999, "learning_rate": 7.355212355212356e-06, "loss": 781.2869, "step": 137 }, { "epoch": 0.26640926640926643, "grad_norm": 3.8438968658447266, "learning_rate": 7.335907335907336e-06, "loss": 755.309, "step": 138 }, { "epoch": 0.26833976833976836, "grad_norm": 3.0307114124298096, "learning_rate": 7.316602316602317e-06, "loss": 757.5778, "step": 139 }, { "epoch": 0.2702702702702703, "grad_norm": 2.9648211002349854, "learning_rate": 7.297297297297298e-06, "loss": 753.4247, "step": 140 }, { "epoch": 0.2722007722007722, "grad_norm": 2.9691784381866455, "learning_rate": 7.277992277992278e-06, "loss": 757.2734, "step": 141 }, { "epoch": 0.27413127413127414, "grad_norm": 2.8389272689819336, "learning_rate": 7.2586872586872595e-06, "loss": 758.1973, "step": 142 }, { "epoch": 0.27606177606177607, "grad_norm": 2.8457741737365723, "learning_rate": 7.23938223938224e-06, "loss": 759.3056, "step": 143 }, { "epoch": 0.277992277992278, "grad_norm": 2.9947755336761475, "learning_rate": 7.22007722007722e-06, "loss": 756.6718, "step": 144 }, { "epoch": 0.2799227799227799, "grad_norm": 3.285741090774536, "learning_rate": 7.2007722007722015e-06, "loss": 741.2257, "step": 145 }, { "epoch": 0.28185328185328185, "grad_norm": 3.118593692779541, "learning_rate": 7.181467181467182e-06, "loss": 757.1205, "step": 146 }, { "epoch": 0.28378378378378377, "grad_norm": 3.1158804893493652, "learning_rate": 7.162162162162163e-06, "loss": 747.1119, "step": 147 }, { "epoch": 0.2857142857142857, "grad_norm": 2.9939019680023193, "learning_rate": 7.1428571428571436e-06, "loss": 738.4834, "step": 148 }, { "epoch": 0.2876447876447876, "grad_norm": 3.792679786682129, "learning_rate": 7.123552123552124e-06, "loss": 758.1683, "step": 149 }, { "epoch": 0.28957528957528955, "grad_norm": 3.3988595008850098, "learning_rate": 7.104247104247105e-06, "loss": 757.7793, "step": 150 }, { "epoch": 0.2915057915057915, "grad_norm": 3.4164488315582275, "learning_rate": 7.084942084942086e-06, "loss": 753.686, "step": 151 }, { "epoch": 0.29343629343629346, "grad_norm": 4.295523166656494, "learning_rate": 7.065637065637067e-06, "loss": 749.4221, "step": 152 }, { "epoch": 0.2953667953667954, "grad_norm": 3.1985435485839844, "learning_rate": 7.046332046332047e-06, "loss": 766.7484, "step": 153 }, { "epoch": 0.2972972972972973, "grad_norm": 2.811443567276001, "learning_rate": 7.027027027027028e-06, "loss": 744.4705, "step": 154 }, { "epoch": 0.29922779922779924, "grad_norm": 2.8985869884490967, "learning_rate": 7.007722007722009e-06, "loss": 771.2823, "step": 155 }, { "epoch": 0.30115830115830117, "grad_norm": 2.913670301437378, "learning_rate": 6.988416988416989e-06, "loss": 767.1945, "step": 156 }, { "epoch": 0.3030888030888031, "grad_norm": 3.091149091720581, "learning_rate": 6.9691119691119706e-06, "loss": 770.168, "step": 157 }, { "epoch": 0.305019305019305, "grad_norm": 2.89846134185791, "learning_rate": 6.949806949806951e-06, "loss": 746.1279, "step": 158 }, { "epoch": 0.30694980694980695, "grad_norm": 2.981823444366455, "learning_rate": 6.9305019305019305e-06, "loss": 749.2059, "step": 159 }, { "epoch": 0.3088803088803089, "grad_norm": 3.6476705074310303, "learning_rate": 6.911196911196911e-06, "loss": 757.4434, "step": 160 }, { "epoch": 0.3108108108108108, "grad_norm": 2.946174383163452, "learning_rate": 6.891891891891892e-06, "loss": 759.8816, "step": 161 }, { "epoch": 0.3127413127413127, "grad_norm": 3.030103921890259, "learning_rate": 6.872586872586873e-06, "loss": 757.3468, "step": 162 }, { "epoch": 0.31467181467181465, "grad_norm": 3.3172245025634766, "learning_rate": 6.853281853281854e-06, "loss": 753.3054, "step": 163 }, { "epoch": 0.3166023166023166, "grad_norm": 3.128265857696533, "learning_rate": 6.833976833976834e-06, "loss": 764.0436, "step": 164 }, { "epoch": 0.3185328185328185, "grad_norm": 2.8332438468933105, "learning_rate": 6.814671814671815e-06, "loss": 747.4676, "step": 165 }, { "epoch": 0.3204633204633205, "grad_norm": 2.9192118644714355, "learning_rate": 6.795366795366796e-06, "loss": 759.479, "step": 166 }, { "epoch": 0.3223938223938224, "grad_norm": 2.926225423812866, "learning_rate": 6.776061776061776e-06, "loss": 749.5502, "step": 167 }, { "epoch": 0.32432432432432434, "grad_norm": 3.739107608795166, "learning_rate": 6.7567567567567575e-06, "loss": 761.9246, "step": 168 }, { "epoch": 0.32625482625482627, "grad_norm": 3.1689376831054688, "learning_rate": 6.737451737451738e-06, "loss": 759.9313, "step": 169 }, { "epoch": 0.3281853281853282, "grad_norm": 4.102996349334717, "learning_rate": 6.718146718146718e-06, "loss": 755.9019, "step": 170 }, { "epoch": 0.3301158301158301, "grad_norm": 4.32927131652832, "learning_rate": 6.6988416988417e-06, "loss": 748.2408, "step": 171 }, { "epoch": 0.33204633204633205, "grad_norm": 2.9341273307800293, "learning_rate": 6.67953667953668e-06, "loss": 767.1997, "step": 172 }, { "epoch": 0.333976833976834, "grad_norm": 2.9747681617736816, "learning_rate": 6.66023166023166e-06, "loss": 749.2424, "step": 173 }, { "epoch": 0.3359073359073359, "grad_norm": 3.03954815864563, "learning_rate": 6.640926640926642e-06, "loss": 753.1185, "step": 174 }, { "epoch": 0.33783783783783783, "grad_norm": 2.896865129470825, "learning_rate": 6.621621621621622e-06, "loss": 760.0892, "step": 175 }, { "epoch": 0.33976833976833976, "grad_norm": 4.517345905303955, "learning_rate": 6.602316602316603e-06, "loss": 761.2517, "step": 176 }, { "epoch": 0.3416988416988417, "grad_norm": 3.4095852375030518, "learning_rate": 6.583011583011584e-06, "loss": 760.7, "step": 177 }, { "epoch": 0.3436293436293436, "grad_norm": 3.5481455326080322, "learning_rate": 6.563706563706564e-06, "loss": 761.6207, "step": 178 }, { "epoch": 0.34555984555984554, "grad_norm": 4.07427978515625, "learning_rate": 6.544401544401545e-06, "loss": 762.6121, "step": 179 }, { "epoch": 0.3474903474903475, "grad_norm": 3.08604097366333, "learning_rate": 6.525096525096526e-06, "loss": 743.1675, "step": 180 }, { "epoch": 0.34942084942084944, "grad_norm": 4.791824817657471, "learning_rate": 6.505791505791507e-06, "loss": 773.7855, "step": 181 }, { "epoch": 0.35135135135135137, "grad_norm": 5.513062953948975, "learning_rate": 6.486486486486487e-06, "loss": 755.7749, "step": 182 }, { "epoch": 0.3532818532818533, "grad_norm": 4.703281879425049, "learning_rate": 6.467181467181468e-06, "loss": 762.9774, "step": 183 }, { "epoch": 0.3552123552123552, "grad_norm": 4.766940593719482, "learning_rate": 6.447876447876449e-06, "loss": 757.7225, "step": 184 }, { "epoch": 0.35714285714285715, "grad_norm": 2.995068073272705, "learning_rate": 6.4285714285714295e-06, "loss": 757.5352, "step": 185 }, { "epoch": 0.3590733590733591, "grad_norm": 4.172158241271973, "learning_rate": 6.409266409266411e-06, "loss": 758.1395, "step": 186 }, { "epoch": 0.361003861003861, "grad_norm": 3.057084798812866, "learning_rate": 6.389961389961391e-06, "loss": 763.8903, "step": 187 }, { "epoch": 0.36293436293436293, "grad_norm": 3.4208884239196777, "learning_rate": 6.370656370656371e-06, "loss": 755.9663, "step": 188 }, { "epoch": 0.36486486486486486, "grad_norm": 3.5881717205047607, "learning_rate": 6.351351351351351e-06, "loss": 764.6107, "step": 189 }, { "epoch": 0.3667953667953668, "grad_norm": 4.713850498199463, "learning_rate": 6.332046332046332e-06, "loss": 748.8102, "step": 190 }, { "epoch": 0.3687258687258687, "grad_norm": 3.244720935821533, "learning_rate": 6.312741312741313e-06, "loss": 764.7598, "step": 191 }, { "epoch": 0.37065637065637064, "grad_norm": 4.447683334350586, "learning_rate": 6.293436293436294e-06, "loss": 752.6754, "step": 192 }, { "epoch": 0.37258687258687256, "grad_norm": 3.9772603511810303, "learning_rate": 6.274131274131274e-06, "loss": 756.0471, "step": 193 }, { "epoch": 0.3745173745173745, "grad_norm": 4.803740978240967, "learning_rate": 6.254826254826255e-06, "loss": 769.636, "step": 194 }, { "epoch": 0.3764478764478765, "grad_norm": 4.907277584075928, "learning_rate": 6.235521235521236e-06, "loss": 765.9659, "step": 195 }, { "epoch": 0.3783783783783784, "grad_norm": 3.6414902210235596, "learning_rate": 6.2162162162162164e-06, "loss": 760.1011, "step": 196 }, { "epoch": 0.3803088803088803, "grad_norm": 3.6037609577178955, "learning_rate": 6.196911196911197e-06, "loss": 750.7067, "step": 197 }, { "epoch": 0.38223938223938225, "grad_norm": 2.8772242069244385, "learning_rate": 6.177606177606178e-06, "loss": 756.6414, "step": 198 }, { "epoch": 0.3841698841698842, "grad_norm": 5.2583184242248535, "learning_rate": 6.1583011583011585e-06, "loss": 751.6912, "step": 199 }, { "epoch": 0.3861003861003861, "grad_norm": 2.8050296306610107, "learning_rate": 6.13899613899614e-06, "loss": 743.5164, "step": 200 }, { "epoch": 0.38803088803088803, "grad_norm": 3.5677103996276855, "learning_rate": 6.11969111969112e-06, "loss": 764.8164, "step": 201 }, { "epoch": 0.38996138996138996, "grad_norm": 3.1550533771514893, "learning_rate": 6.1003861003861005e-06, "loss": 762.9319, "step": 202 }, { "epoch": 0.3918918918918919, "grad_norm": 2.991600751876831, "learning_rate": 6.081081081081082e-06, "loss": 770.1368, "step": 203 }, { "epoch": 0.3938223938223938, "grad_norm": 2.9870667457580566, "learning_rate": 6.061776061776062e-06, "loss": 751.1348, "step": 204 }, { "epoch": 0.39575289575289574, "grad_norm": 2.980593204498291, "learning_rate": 6.0424710424710434e-06, "loss": 769.3608, "step": 205 }, { "epoch": 0.39768339768339767, "grad_norm": 3.3218741416931152, "learning_rate": 6.023166023166024e-06, "loss": 763.5135, "step": 206 }, { "epoch": 0.3996138996138996, "grad_norm": 3.1396753787994385, "learning_rate": 6.003861003861004e-06, "loss": 757.1105, "step": 207 }, { "epoch": 0.4015444015444015, "grad_norm": 3.2232918739318848, "learning_rate": 5.9845559845559855e-06, "loss": 743.5112, "step": 208 }, { "epoch": 0.4034749034749035, "grad_norm": 3.3327605724334717, "learning_rate": 5.965250965250966e-06, "loss": 766.4379, "step": 209 }, { "epoch": 0.40540540540540543, "grad_norm": 3.4697072505950928, "learning_rate": 5.945945945945947e-06, "loss": 767.7292, "step": 210 }, { "epoch": 0.40733590733590735, "grad_norm": 4.161364555358887, "learning_rate": 5.9266409266409275e-06, "loss": 746.888, "step": 211 }, { "epoch": 0.4092664092664093, "grad_norm": 3.0299108028411865, "learning_rate": 5.907335907335908e-06, "loss": 773.0205, "step": 212 }, { "epoch": 0.4111969111969112, "grad_norm": 3.702420949935913, "learning_rate": 5.888030888030889e-06, "loss": 762.8275, "step": 213 }, { "epoch": 0.41312741312741313, "grad_norm": 4.044224262237549, "learning_rate": 5.86872586872587e-06, "loss": 762.0264, "step": 214 }, { "epoch": 0.41505791505791506, "grad_norm": 3.09975266456604, "learning_rate": 5.84942084942085e-06, "loss": 738.4187, "step": 215 }, { "epoch": 0.416988416988417, "grad_norm": 3.779367208480835, "learning_rate": 5.83011583011583e-06, "loss": 764.1588, "step": 216 }, { "epoch": 0.4189189189189189, "grad_norm": 3.9344935417175293, "learning_rate": 5.810810810810811e-06, "loss": 751.987, "step": 217 }, { "epoch": 0.42084942084942084, "grad_norm": 3.073119878768921, "learning_rate": 5.791505791505791e-06, "loss": 763.5579, "step": 218 }, { "epoch": 0.42277992277992277, "grad_norm": 5.221838474273682, "learning_rate": 5.7722007722007725e-06, "loss": 736.679, "step": 219 }, { "epoch": 0.4247104247104247, "grad_norm": 3.210583209991455, "learning_rate": 5.752895752895753e-06, "loss": 743.9155, "step": 220 }, { "epoch": 0.4266409266409266, "grad_norm": 3.032778739929199, "learning_rate": 5.733590733590734e-06, "loss": 756.9208, "step": 221 }, { "epoch": 0.42857142857142855, "grad_norm": 3.052091598510742, "learning_rate": 5.7142857142857145e-06, "loss": 753.4189, "step": 222 }, { "epoch": 0.4305019305019305, "grad_norm": 2.934615135192871, "learning_rate": 5.694980694980695e-06, "loss": 744.2302, "step": 223 }, { "epoch": 0.43243243243243246, "grad_norm": 3.453640937805176, "learning_rate": 5.675675675675676e-06, "loss": 753.5289, "step": 224 }, { "epoch": 0.4343629343629344, "grad_norm": 4.493204116821289, "learning_rate": 5.6563706563706566e-06, "loss": 773.2119, "step": 225 }, { "epoch": 0.4362934362934363, "grad_norm": 3.1990344524383545, "learning_rate": 5.637065637065637e-06, "loss": 771.2296, "step": 226 }, { "epoch": 0.43822393822393824, "grad_norm": 4.990328788757324, "learning_rate": 5.617760617760618e-06, "loss": 766.2954, "step": 227 }, { "epoch": 0.44015444015444016, "grad_norm": 6.218008518218994, "learning_rate": 5.598455598455599e-06, "loss": 754.2188, "step": 228 }, { "epoch": 0.4420849420849421, "grad_norm": 4.668118476867676, "learning_rate": 5.57915057915058e-06, "loss": 764.908, "step": 229 }, { "epoch": 0.444015444015444, "grad_norm": 3.144949436187744, "learning_rate": 5.55984555984556e-06, "loss": 769.907, "step": 230 }, { "epoch": 0.44594594594594594, "grad_norm": 3.761617422103882, "learning_rate": 5.540540540540541e-06, "loss": 762.5698, "step": 231 }, { "epoch": 0.44787644787644787, "grad_norm": 3.3352720737457275, "learning_rate": 5.521235521235522e-06, "loss": 760.8379, "step": 232 }, { "epoch": 0.4498069498069498, "grad_norm": 3.646754741668701, "learning_rate": 5.501930501930502e-06, "loss": 762.5335, "step": 233 }, { "epoch": 0.4517374517374517, "grad_norm": 5.615658283233643, "learning_rate": 5.4826254826254836e-06, "loss": 763.7914, "step": 234 }, { "epoch": 0.45366795366795365, "grad_norm": 3.1293540000915527, "learning_rate": 5.463320463320464e-06, "loss": 765.2655, "step": 235 }, { "epoch": 0.4555984555984556, "grad_norm": 3.9679007530212402, "learning_rate": 5.444015444015444e-06, "loss": 776.4496, "step": 236 }, { "epoch": 0.4575289575289575, "grad_norm": 4.110316753387451, "learning_rate": 5.424710424710426e-06, "loss": 769.2201, "step": 237 }, { "epoch": 0.4594594594594595, "grad_norm": 3.5175106525421143, "learning_rate": 5.405405405405406e-06, "loss": 759.53, "step": 238 }, { "epoch": 0.4613899613899614, "grad_norm": 3.206360101699829, "learning_rate": 5.386100386100387e-06, "loss": 753.7003, "step": 239 }, { "epoch": 0.46332046332046334, "grad_norm": 2.994739294052124, "learning_rate": 5.366795366795368e-06, "loss": 750.8613, "step": 240 }, { "epoch": 0.46525096525096526, "grad_norm": 2.966420888900757, "learning_rate": 5.347490347490348e-06, "loss": 755.6821, "step": 241 }, { "epoch": 0.4671814671814672, "grad_norm": 5.035081386566162, "learning_rate": 5.328185328185329e-06, "loss": 753.4547, "step": 242 }, { "epoch": 0.4691119691119691, "grad_norm": 3.399747371673584, "learning_rate": 5.30888030888031e-06, "loss": 752.4637, "step": 243 }, { "epoch": 0.47104247104247104, "grad_norm": 2.9818832874298096, "learning_rate": 5.28957528957529e-06, "loss": 761.8242, "step": 244 }, { "epoch": 0.47297297297297297, "grad_norm": 3.317462921142578, "learning_rate": 5.2702702702702705e-06, "loss": 741.8611, "step": 245 }, { "epoch": 0.4749034749034749, "grad_norm": 3.9936470985412598, "learning_rate": 5.250965250965251e-06, "loss": 755.0592, "step": 246 }, { "epoch": 0.4768339768339768, "grad_norm": 3.263794422149658, "learning_rate": 5.231660231660231e-06, "loss": 748.0864, "step": 247 }, { "epoch": 0.47876447876447875, "grad_norm": 3.1874217987060547, "learning_rate": 5.212355212355213e-06, "loss": 751.8619, "step": 248 }, { "epoch": 0.4806949806949807, "grad_norm": 4.754833698272705, "learning_rate": 5.193050193050193e-06, "loss": 763.7399, "step": 249 }, { "epoch": 0.4826254826254826, "grad_norm": 3.21230149269104, "learning_rate": 5.173745173745173e-06, "loss": 757.1412, "step": 250 }, { "epoch": 0.48455598455598453, "grad_norm": 3.2287521362304688, "learning_rate": 5.154440154440155e-06, "loss": 749.6309, "step": 251 }, { "epoch": 0.4864864864864865, "grad_norm": 3.137535572052002, "learning_rate": 5.135135135135135e-06, "loss": 759.4156, "step": 252 }, { "epoch": 0.48841698841698844, "grad_norm": 3.6146438121795654, "learning_rate": 5.115830115830116e-06, "loss": 741.8262, "step": 253 }, { "epoch": 0.49034749034749037, "grad_norm": 2.9531517028808594, "learning_rate": 5.096525096525097e-06, "loss": 763.7745, "step": 254 }, { "epoch": 0.4922779922779923, "grad_norm": 3.0513949394226074, "learning_rate": 5.077220077220077e-06, "loss": 755.271, "step": 255 }, { "epoch": 0.4942084942084942, "grad_norm": 3.2384207248687744, "learning_rate": 5.057915057915058e-06, "loss": 765.3377, "step": 256 }, { "epoch": 0.49613899613899615, "grad_norm": 3.8821310997009277, "learning_rate": 5.038610038610039e-06, "loss": 763.2088, "step": 257 }, { "epoch": 0.4980694980694981, "grad_norm": 2.926225423812866, "learning_rate": 5.01930501930502e-06, "loss": 758.5596, "step": 258 }, { "epoch": 0.5, "grad_norm": 2.9751245975494385, "learning_rate": 5e-06, "loss": 757.9719, "step": 259 }, { "epoch": 0.5019305019305019, "grad_norm": 4.772404193878174, "learning_rate": 4.980694980694981e-06, "loss": 757.3872, "step": 260 }, { "epoch": 0.5038610038610039, "grad_norm": 3.096656322479248, "learning_rate": 4.961389961389962e-06, "loss": 760.4723, "step": 261 }, { "epoch": 0.5057915057915058, "grad_norm": 3.504852056503296, "learning_rate": 4.9420849420849425e-06, "loss": 747.486, "step": 262 }, { "epoch": 0.5077220077220077, "grad_norm": 3.33747935295105, "learning_rate": 4.922779922779924e-06, "loss": 758.9821, "step": 263 }, { "epoch": 0.5096525096525096, "grad_norm": 3.4812850952148438, "learning_rate": 4.903474903474904e-06, "loss": 765.7698, "step": 264 }, { "epoch": 0.5115830115830116, "grad_norm": 3.174971580505371, "learning_rate": 4.8841698841698845e-06, "loss": 756.7213, "step": 265 }, { "epoch": 0.5135135135135135, "grad_norm": 3.601083755493164, "learning_rate": 4.864864864864866e-06, "loss": 762.7404, "step": 266 }, { "epoch": 0.5154440154440154, "grad_norm": 3.3433072566986084, "learning_rate": 4.845559845559845e-06, "loss": 747.8812, "step": 267 }, { "epoch": 0.5173745173745173, "grad_norm": 3.7494962215423584, "learning_rate": 4.8262548262548266e-06, "loss": 756.0776, "step": 268 }, { "epoch": 0.5193050193050193, "grad_norm": 4.609169960021973, "learning_rate": 4.806949806949807e-06, "loss": 750.7801, "step": 269 }, { "epoch": 0.5212355212355212, "grad_norm": 5.737427711486816, "learning_rate": 4.787644787644788e-06, "loss": 751.1343, "step": 270 }, { "epoch": 0.5231660231660231, "grad_norm": 3.3391857147216797, "learning_rate": 4.768339768339769e-06, "loss": 760.2618, "step": 271 }, { "epoch": 0.525096525096525, "grad_norm": 3.1758346557617188, "learning_rate": 4.749034749034749e-06, "loss": 756.4235, "step": 272 }, { "epoch": 0.527027027027027, "grad_norm": 4.515329360961914, "learning_rate": 4.72972972972973e-06, "loss": 770.3212, "step": 273 }, { "epoch": 0.528957528957529, "grad_norm": 3.2932803630828857, "learning_rate": 4.710424710424711e-06, "loss": 751.3934, "step": 274 }, { "epoch": 0.5308880308880309, "grad_norm": 3.047992467880249, "learning_rate": 4.691119691119692e-06, "loss": 749.6434, "step": 275 }, { "epoch": 0.5328185328185329, "grad_norm": 3.2720017433166504, "learning_rate": 4.671814671814672e-06, "loss": 771.3862, "step": 276 }, { "epoch": 0.5347490347490348, "grad_norm": 3.6635942459106445, "learning_rate": 4.652509652509653e-06, "loss": 758.3706, "step": 277 }, { "epoch": 0.5366795366795367, "grad_norm": 3.4579808712005615, "learning_rate": 4.633204633204634e-06, "loss": 747.9475, "step": 278 }, { "epoch": 0.5386100386100386, "grad_norm": 3.525111675262451, "learning_rate": 4.613899613899614e-06, "loss": 756.5895, "step": 279 }, { "epoch": 0.5405405405405406, "grad_norm": 3.336205005645752, "learning_rate": 4.594594594594596e-06, "loss": 772.9059, "step": 280 }, { "epoch": 0.5424710424710425, "grad_norm": 3.758382558822632, "learning_rate": 4.575289575289575e-06, "loss": 769.8763, "step": 281 }, { "epoch": 0.5444015444015444, "grad_norm": 3.054797649383545, "learning_rate": 4.5559845559845564e-06, "loss": 768.0063, "step": 282 }, { "epoch": 0.5463320463320464, "grad_norm": 4.434202671051025, "learning_rate": 4.536679536679537e-06, "loss": 760.1692, "step": 283 }, { "epoch": 0.5482625482625483, "grad_norm": 4.3827667236328125, "learning_rate": 4.517374517374517e-06, "loss": 744.6565, "step": 284 }, { "epoch": 0.5501930501930502, "grad_norm": 6.237070560455322, "learning_rate": 4.4980694980694985e-06, "loss": 767.5054, "step": 285 }, { "epoch": 0.5521235521235521, "grad_norm": 3.162343740463257, "learning_rate": 4.478764478764479e-06, "loss": 762.6232, "step": 286 }, { "epoch": 0.5540540540540541, "grad_norm": 3.9073193073272705, "learning_rate": 4.45945945945946e-06, "loss": 778.0848, "step": 287 }, { "epoch": 0.555984555984556, "grad_norm": 6.03090763092041, "learning_rate": 4.4401544401544405e-06, "loss": 759.6763, "step": 288 }, { "epoch": 0.5579150579150579, "grad_norm": 4.3667311668396, "learning_rate": 4.420849420849421e-06, "loss": 762.0499, "step": 289 }, { "epoch": 0.5598455598455598, "grad_norm": 3.5320377349853516, "learning_rate": 4.401544401544402e-06, "loss": 769.951, "step": 290 }, { "epoch": 0.5617760617760618, "grad_norm": 3.7424259185791016, "learning_rate": 4.382239382239383e-06, "loss": 767.4916, "step": 291 }, { "epoch": 0.5637065637065637, "grad_norm": 3.3271844387054443, "learning_rate": 4.362934362934364e-06, "loss": 761.4987, "step": 292 }, { "epoch": 0.5656370656370656, "grad_norm": 2.957545757293701, "learning_rate": 4.343629343629344e-06, "loss": 762.5405, "step": 293 }, { "epoch": 0.5675675675675675, "grad_norm": 3.087904453277588, "learning_rate": 4.324324324324325e-06, "loss": 752.0869, "step": 294 }, { "epoch": 0.5694980694980695, "grad_norm": 3.4047775268554688, "learning_rate": 4.305019305019305e-06, "loss": 764.2693, "step": 295 }, { "epoch": 0.5714285714285714, "grad_norm": 3.244701385498047, "learning_rate": 4.2857142857142855e-06, "loss": 754.4578, "step": 296 }, { "epoch": 0.5733590733590733, "grad_norm": 3.3166253566741943, "learning_rate": 4.266409266409267e-06, "loss": 773.7292, "step": 297 }, { "epoch": 0.5752895752895753, "grad_norm": 3.0471668243408203, "learning_rate": 4.247104247104247e-06, "loss": 778.6743, "step": 298 }, { "epoch": 0.5772200772200772, "grad_norm": 3.6165595054626465, "learning_rate": 4.227799227799228e-06, "loss": 755.6964, "step": 299 }, { "epoch": 0.5791505791505791, "grad_norm": 3.426254987716675, "learning_rate": 4.208494208494209e-06, "loss": 768.4984, "step": 300 }, { "epoch": 0.581081081081081, "grad_norm": 4.2183518409729, "learning_rate": 4.189189189189189e-06, "loss": 756.4444, "step": 301 }, { "epoch": 0.583011583011583, "grad_norm": 4.252375602722168, "learning_rate": 4.16988416988417e-06, "loss": 760.7186, "step": 302 }, { "epoch": 0.584942084942085, "grad_norm": 3.898979425430298, "learning_rate": 4.150579150579151e-06, "loss": 755.8693, "step": 303 }, { "epoch": 0.5868725868725869, "grad_norm": 4.586599349975586, "learning_rate": 4.131274131274132e-06, "loss": 767.7788, "step": 304 }, { "epoch": 0.5888030888030888, "grad_norm": 3.4545576572418213, "learning_rate": 4.1119691119691125e-06, "loss": 755.6498, "step": 305 }, { "epoch": 0.5907335907335908, "grad_norm": 3.056624174118042, "learning_rate": 4.092664092664093e-06, "loss": 779.4406, "step": 306 }, { "epoch": 0.5926640926640927, "grad_norm": 3.1570804119110107, "learning_rate": 4.073359073359074e-06, "loss": 756.1407, "step": 307 }, { "epoch": 0.5945945945945946, "grad_norm": 5.404672622680664, "learning_rate": 4.0540540540540545e-06, "loss": 770.681, "step": 308 }, { "epoch": 0.5965250965250966, "grad_norm": 3.0374257564544678, "learning_rate": 4.034749034749035e-06, "loss": 759.7466, "step": 309 }, { "epoch": 0.5984555984555985, "grad_norm": 3.2685699462890625, "learning_rate": 4.015444015444015e-06, "loss": 762.7603, "step": 310 }, { "epoch": 0.6003861003861004, "grad_norm": 7.29334831237793, "learning_rate": 3.996138996138997e-06, "loss": 742.7489, "step": 311 }, { "epoch": 0.6023166023166023, "grad_norm": 3.852389335632324, "learning_rate": 3.976833976833977e-06, "loss": 767.3596, "step": 312 }, { "epoch": 0.6042471042471043, "grad_norm": 3.200193166732788, "learning_rate": 3.957528957528957e-06, "loss": 766.8765, "step": 313 }, { "epoch": 0.6061776061776062, "grad_norm": 3.5071451663970947, "learning_rate": 3.938223938223939e-06, "loss": 776.7089, "step": 314 }, { "epoch": 0.6081081081081081, "grad_norm": 4.16587495803833, "learning_rate": 3.918918918918919e-06, "loss": 741.2016, "step": 315 }, { "epoch": 0.61003861003861, "grad_norm": 5.703463554382324, "learning_rate": 3.8996138996139e-06, "loss": 756.0547, "step": 316 }, { "epoch": 0.611969111969112, "grad_norm": 5.430912017822266, "learning_rate": 3.880308880308881e-06, "loss": 756.722, "step": 317 }, { "epoch": 0.6138996138996139, "grad_norm": 8.650154113769531, "learning_rate": 3.861003861003861e-06, "loss": 745.7667, "step": 318 }, { "epoch": 0.6158301158301158, "grad_norm": 13.80824089050293, "learning_rate": 3.841698841698842e-06, "loss": 755.2433, "step": 319 }, { "epoch": 0.6177606177606177, "grad_norm": 7.08932638168335, "learning_rate": 3.822393822393823e-06, "loss": 753.7036, "step": 320 }, { "epoch": 0.6196911196911197, "grad_norm": 3.6433534622192383, "learning_rate": 3.8030888030888036e-06, "loss": 773.3873, "step": 321 }, { "epoch": 0.6216216216216216, "grad_norm": 3.2302098274230957, "learning_rate": 3.7837837837837844e-06, "loss": 758.14, "step": 322 }, { "epoch": 0.6235521235521235, "grad_norm": 3.731865644454956, "learning_rate": 3.7644787644787652e-06, "loss": 766.5978, "step": 323 }, { "epoch": 0.6254826254826255, "grad_norm": 3.244114875793457, "learning_rate": 3.745173745173745e-06, "loss": 748.5593, "step": 324 }, { "epoch": 0.6274131274131274, "grad_norm": 6.71386194229126, "learning_rate": 3.725868725868726e-06, "loss": 750.8797, "step": 325 }, { "epoch": 0.6293436293436293, "grad_norm": 6.857434272766113, "learning_rate": 3.706563706563707e-06, "loss": 751.0732, "step": 326 }, { "epoch": 0.6312741312741312, "grad_norm": 3.1057608127593994, "learning_rate": 3.6872586872586872e-06, "loss": 751.9279, "step": 327 }, { "epoch": 0.6332046332046332, "grad_norm": 4.462062835693359, "learning_rate": 3.667953667953668e-06, "loss": 761.2999, "step": 328 }, { "epoch": 0.6351351351351351, "grad_norm": 5.394342422485352, "learning_rate": 3.648648648648649e-06, "loss": 765.5782, "step": 329 }, { "epoch": 0.637065637065637, "grad_norm": 4.586513996124268, "learning_rate": 3.6293436293436297e-06, "loss": 759.7944, "step": 330 }, { "epoch": 0.638996138996139, "grad_norm": 6.204144477844238, "learning_rate": 3.61003861003861e-06, "loss": 749.0181, "step": 331 }, { "epoch": 0.640926640926641, "grad_norm": 3.065887451171875, "learning_rate": 3.590733590733591e-06, "loss": 764.4531, "step": 332 }, { "epoch": 0.6428571428571429, "grad_norm": 3.5210766792297363, "learning_rate": 3.5714285714285718e-06, "loss": 757.1459, "step": 333 }, { "epoch": 0.6447876447876448, "grad_norm": 4.145437240600586, "learning_rate": 3.5521235521235526e-06, "loss": 772.7172, "step": 334 }, { "epoch": 0.6467181467181468, "grad_norm": 3.4393627643585205, "learning_rate": 3.5328185328185334e-06, "loss": 748.3174, "step": 335 }, { "epoch": 0.6486486486486487, "grad_norm": 6.470264911651611, "learning_rate": 3.513513513513514e-06, "loss": 761.2096, "step": 336 }, { "epoch": 0.6505791505791506, "grad_norm": 3.4703409671783447, "learning_rate": 3.4942084942084947e-06, "loss": 752.6991, "step": 337 }, { "epoch": 0.6525096525096525, "grad_norm": 9.870447158813477, "learning_rate": 3.4749034749034755e-06, "loss": 771.034, "step": 338 }, { "epoch": 0.6544401544401545, "grad_norm": 9.164502143859863, "learning_rate": 3.4555984555984555e-06, "loss": 759.8434, "step": 339 }, { "epoch": 0.6563706563706564, "grad_norm": 8.678630828857422, "learning_rate": 3.4362934362934363e-06, "loss": 760.7397, "step": 340 }, { "epoch": 0.6583011583011583, "grad_norm": 4.10188102722168, "learning_rate": 3.416988416988417e-06, "loss": 752.9614, "step": 341 }, { "epoch": 0.6602316602316602, "grad_norm": 5.000187873840332, "learning_rate": 3.397683397683398e-06, "loss": 741.5472, "step": 342 }, { "epoch": 0.6621621621621622, "grad_norm": 5.491066932678223, "learning_rate": 3.3783783783783788e-06, "loss": 752.3487, "step": 343 }, { "epoch": 0.6640926640926641, "grad_norm": 4.743298530578613, "learning_rate": 3.359073359073359e-06, "loss": 740.1313, "step": 344 }, { "epoch": 0.666023166023166, "grad_norm": 3.8254621028900146, "learning_rate": 3.33976833976834e-06, "loss": 751.6027, "step": 345 }, { "epoch": 0.667953667953668, "grad_norm": 3.4094951152801514, "learning_rate": 3.320463320463321e-06, "loss": 747.6785, "step": 346 }, { "epoch": 0.6698841698841699, "grad_norm": 4.3971757888793945, "learning_rate": 3.3011583011583016e-06, "loss": 765.18, "step": 347 }, { "epoch": 0.6718146718146718, "grad_norm": 5.712260723114014, "learning_rate": 3.281853281853282e-06, "loss": 754.3199, "step": 348 }, { "epoch": 0.6737451737451737, "grad_norm": 2.944014310836792, "learning_rate": 3.262548262548263e-06, "loss": 759.125, "step": 349 }, { "epoch": 0.6756756756756757, "grad_norm": 5.154186725616455, "learning_rate": 3.2432432432432437e-06, "loss": 773.4568, "step": 350 }, { "epoch": 0.6776061776061776, "grad_norm": 4.158810138702393, "learning_rate": 3.2239382239382245e-06, "loss": 773.0182, "step": 351 }, { "epoch": 0.6795366795366795, "grad_norm": 2.9692142009735107, "learning_rate": 3.2046332046332054e-06, "loss": 759.3475, "step": 352 }, { "epoch": 0.6814671814671814, "grad_norm": 5.837861061096191, "learning_rate": 3.1853281853281853e-06, "loss": 758.8844, "step": 353 }, { "epoch": 0.6833976833976834, "grad_norm": 4.23380708694458, "learning_rate": 3.166023166023166e-06, "loss": 762.8468, "step": 354 }, { "epoch": 0.6853281853281853, "grad_norm": 5.026414394378662, "learning_rate": 3.146718146718147e-06, "loss": 738.6548, "step": 355 }, { "epoch": 0.6872586872586872, "grad_norm": 4.041588306427002, "learning_rate": 3.1274131274131274e-06, "loss": 766.1196, "step": 356 }, { "epoch": 0.6891891891891891, "grad_norm": 4.39390230178833, "learning_rate": 3.1081081081081082e-06, "loss": 749.96, "step": 357 }, { "epoch": 0.6911196911196911, "grad_norm": 4.870336055755615, "learning_rate": 3.088803088803089e-06, "loss": 750.3604, "step": 358 }, { "epoch": 0.693050193050193, "grad_norm": 6.5269036293029785, "learning_rate": 3.06949806949807e-06, "loss": 765.9295, "step": 359 }, { "epoch": 0.694980694980695, "grad_norm": 3.1647534370422363, "learning_rate": 3.0501930501930503e-06, "loss": 747.9913, "step": 360 }, { "epoch": 0.696911196911197, "grad_norm": 3.435527801513672, "learning_rate": 3.030888030888031e-06, "loss": 770.7856, "step": 361 }, { "epoch": 0.6988416988416989, "grad_norm": 5.05522346496582, "learning_rate": 3.011583011583012e-06, "loss": 773.0812, "step": 362 }, { "epoch": 0.7007722007722008, "grad_norm": 4.161330223083496, "learning_rate": 2.9922779922779927e-06, "loss": 771.5272, "step": 363 }, { "epoch": 0.7027027027027027, "grad_norm": 4.4345526695251465, "learning_rate": 2.9729729729729736e-06, "loss": 754.7959, "step": 364 }, { "epoch": 0.7046332046332047, "grad_norm": 2.9777281284332275, "learning_rate": 2.953667953667954e-06, "loss": 749.3711, "step": 365 }, { "epoch": 0.7065637065637066, "grad_norm": 2.855987310409546, "learning_rate": 2.934362934362935e-06, "loss": 749.5988, "step": 366 }, { "epoch": 0.7084942084942085, "grad_norm": 3.307720184326172, "learning_rate": 2.915057915057915e-06, "loss": 761.0093, "step": 367 }, { "epoch": 0.7104247104247104, "grad_norm": 3.787365198135376, "learning_rate": 2.8957528957528956e-06, "loss": 751.6588, "step": 368 }, { "epoch": 0.7123552123552124, "grad_norm": 3.029486656188965, "learning_rate": 2.8764478764478764e-06, "loss": 755.2015, "step": 369 }, { "epoch": 0.7142857142857143, "grad_norm": 3.0125091075897217, "learning_rate": 2.8571428571428573e-06, "loss": 766.543, "step": 370 }, { "epoch": 0.7162162162162162, "grad_norm": 4.560346603393555, "learning_rate": 2.837837837837838e-06, "loss": 768.441, "step": 371 }, { "epoch": 0.7181467181467182, "grad_norm": 3.190322160720825, "learning_rate": 2.8185328185328185e-06, "loss": 753.3053, "step": 372 }, { "epoch": 0.7200772200772201, "grad_norm": 3.4040372371673584, "learning_rate": 2.7992277992277993e-06, "loss": 769.1228, "step": 373 }, { "epoch": 0.722007722007722, "grad_norm": 5.434439182281494, "learning_rate": 2.77992277992278e-06, "loss": 754.5753, "step": 374 }, { "epoch": 0.7239382239382239, "grad_norm": 3.3807380199432373, "learning_rate": 2.760617760617761e-06, "loss": 768.5261, "step": 375 }, { "epoch": 0.7258687258687259, "grad_norm": 3.2653305530548096, "learning_rate": 2.7413127413127418e-06, "loss": 768.1964, "step": 376 }, { "epoch": 0.7277992277992278, "grad_norm": 3.041628360748291, "learning_rate": 2.722007722007722e-06, "loss": 771.9323, "step": 377 }, { "epoch": 0.7297297297297297, "grad_norm": 3.092245101928711, "learning_rate": 2.702702702702703e-06, "loss": 753.942, "step": 378 }, { "epoch": 0.7316602316602316, "grad_norm": 4.467896938323975, "learning_rate": 2.683397683397684e-06, "loss": 749.8477, "step": 379 }, { "epoch": 0.7335907335907336, "grad_norm": 2.989520788192749, "learning_rate": 2.6640926640926647e-06, "loss": 759.7245, "step": 380 }, { "epoch": 0.7355212355212355, "grad_norm": 2.9288792610168457, "learning_rate": 2.644787644787645e-06, "loss": 761.0784, "step": 381 }, { "epoch": 0.7374517374517374, "grad_norm": 3.411576986312866, "learning_rate": 2.6254826254826255e-06, "loss": 769.0735, "step": 382 }, { "epoch": 0.7393822393822393, "grad_norm": 7.500954627990723, "learning_rate": 2.6061776061776063e-06, "loss": 752.552, "step": 383 }, { "epoch": 0.7413127413127413, "grad_norm": 4.436896324157715, "learning_rate": 2.5868725868725867e-06, "loss": 772.7829, "step": 384 }, { "epoch": 0.7432432432432432, "grad_norm": 3.8016252517700195, "learning_rate": 2.5675675675675675e-06, "loss": 756.5428, "step": 385 }, { "epoch": 0.7451737451737451, "grad_norm": 3.1768035888671875, "learning_rate": 2.5482625482625484e-06, "loss": 761.3637, "step": 386 }, { "epoch": 0.747104247104247, "grad_norm": 3.24792218208313, "learning_rate": 2.528957528957529e-06, "loss": 764.2406, "step": 387 }, { "epoch": 0.749034749034749, "grad_norm": 3.6789443492889404, "learning_rate": 2.50965250965251e-06, "loss": 764.6515, "step": 388 }, { "epoch": 0.750965250965251, "grad_norm": 2.998337507247925, "learning_rate": 2.4903474903474904e-06, "loss": 745.9545, "step": 389 }, { "epoch": 0.752895752895753, "grad_norm": 5.5688796043396, "learning_rate": 2.4710424710424712e-06, "loss": 757.1937, "step": 390 }, { "epoch": 0.7548262548262549, "grad_norm": 3.491425037384033, "learning_rate": 2.451737451737452e-06, "loss": 747.8929, "step": 391 }, { "epoch": 0.7567567567567568, "grad_norm": 5.370115280151367, "learning_rate": 2.432432432432433e-06, "loss": 760.0563, "step": 392 }, { "epoch": 0.7586872586872587, "grad_norm": 3.907552480697632, "learning_rate": 2.4131274131274133e-06, "loss": 752.7516, "step": 393 }, { "epoch": 0.7606177606177607, "grad_norm": 3.052731513977051, "learning_rate": 2.393822393822394e-06, "loss": 752.1655, "step": 394 }, { "epoch": 0.7625482625482626, "grad_norm": 5.308294296264648, "learning_rate": 2.3745173745173745e-06, "loss": 767.8257, "step": 395 }, { "epoch": 0.7644787644787645, "grad_norm": 3.72119140625, "learning_rate": 2.3552123552123553e-06, "loss": 764.8139, "step": 396 }, { "epoch": 0.7664092664092664, "grad_norm": 3.117110252380371, "learning_rate": 2.335907335907336e-06, "loss": 766.5566, "step": 397 }, { "epoch": 0.7683397683397684, "grad_norm": 3.8100202083587646, "learning_rate": 2.316602316602317e-06, "loss": 767.1085, "step": 398 }, { "epoch": 0.7702702702702703, "grad_norm": 3.1219418048858643, "learning_rate": 2.297297297297298e-06, "loss": 750.5438, "step": 399 }, { "epoch": 0.7722007722007722, "grad_norm": 3.212282180786133, "learning_rate": 2.2779922779922782e-06, "loss": 748.5366, "step": 400 }, { "epoch": 0.7741312741312741, "grad_norm": 3.0505459308624268, "learning_rate": 2.2586872586872586e-06, "loss": 743.7197, "step": 401 }, { "epoch": 0.7760617760617761, "grad_norm": 3.131042718887329, "learning_rate": 2.2393822393822394e-06, "loss": 756.7755, "step": 402 }, { "epoch": 0.777992277992278, "grad_norm": 4.418597221374512, "learning_rate": 2.2200772200772203e-06, "loss": 754.0249, "step": 403 }, { "epoch": 0.7799227799227799, "grad_norm": 4.271742820739746, "learning_rate": 2.200772200772201e-06, "loss": 768.3597, "step": 404 }, { "epoch": 0.7818532818532818, "grad_norm": 3.788788080215454, "learning_rate": 2.181467181467182e-06, "loss": 758.9531, "step": 405 }, { "epoch": 0.7837837837837838, "grad_norm": 3.5197062492370605, "learning_rate": 2.1621621621621623e-06, "loss": 754.1716, "step": 406 }, { "epoch": 0.7857142857142857, "grad_norm": 3.414795398712158, "learning_rate": 2.1428571428571427e-06, "loss": 751.0524, "step": 407 }, { "epoch": 0.7876447876447876, "grad_norm": 3.1258320808410645, "learning_rate": 2.1235521235521236e-06, "loss": 761.6645, "step": 408 }, { "epoch": 0.7895752895752896, "grad_norm": 6.547064781188965, "learning_rate": 2.1042471042471044e-06, "loss": 765.9225, "step": 409 }, { "epoch": 0.7915057915057915, "grad_norm": 4.772284030914307, "learning_rate": 2.084942084942085e-06, "loss": 764.9897, "step": 410 }, { "epoch": 0.7934362934362934, "grad_norm": 4.287495136260986, "learning_rate": 2.065637065637066e-06, "loss": 756.7947, "step": 411 }, { "epoch": 0.7953667953667953, "grad_norm": 5.286579132080078, "learning_rate": 2.0463320463320464e-06, "loss": 756.2988, "step": 412 }, { "epoch": 0.7972972972972973, "grad_norm": 3.132582902908325, "learning_rate": 2.0270270270270273e-06, "loss": 760.6235, "step": 413 }, { "epoch": 0.7992277992277992, "grad_norm": 4.088491439819336, "learning_rate": 2.0077220077220077e-06, "loss": 751.6579, "step": 414 }, { "epoch": 0.8011583011583011, "grad_norm": 3.853825807571411, "learning_rate": 1.9884169884169885e-06, "loss": 751.7198, "step": 415 }, { "epoch": 0.803088803088803, "grad_norm": 3.8805510997772217, "learning_rate": 1.9691119691119693e-06, "loss": 754.175, "step": 416 }, { "epoch": 0.805019305019305, "grad_norm": 3.465831756591797, "learning_rate": 1.94980694980695e-06, "loss": 763.5966, "step": 417 }, { "epoch": 0.806949806949807, "grad_norm": 3.4799764156341553, "learning_rate": 1.9305019305019305e-06, "loss": 757.6487, "step": 418 }, { "epoch": 0.8088803088803089, "grad_norm": 4.306357383728027, "learning_rate": 1.9111969111969114e-06, "loss": 770.5115, "step": 419 }, { "epoch": 0.8108108108108109, "grad_norm": 5.999144554138184, "learning_rate": 1.8918918918918922e-06, "loss": 750.0549, "step": 420 }, { "epoch": 0.8127413127413128, "grad_norm": 3.353970527648926, "learning_rate": 1.8725868725868726e-06, "loss": 770.1407, "step": 421 }, { "epoch": 0.8146718146718147, "grad_norm": 3.1123435497283936, "learning_rate": 1.8532818532818534e-06, "loss": 755.7714, "step": 422 }, { "epoch": 0.8166023166023166, "grad_norm": 3.5919506549835205, "learning_rate": 1.833976833976834e-06, "loss": 768.7866, "step": 423 }, { "epoch": 0.8185328185328186, "grad_norm": 5.815296649932861, "learning_rate": 1.8146718146718149e-06, "loss": 769.049, "step": 424 }, { "epoch": 0.8204633204633205, "grad_norm": 6.508057117462158, "learning_rate": 1.7953667953667955e-06, "loss": 744.9335, "step": 425 }, { "epoch": 0.8223938223938224, "grad_norm": 3.8051586151123047, "learning_rate": 1.7760617760617763e-06, "loss": 755.9141, "step": 426 }, { "epoch": 0.8243243243243243, "grad_norm": 3.628269910812378, "learning_rate": 1.756756756756757e-06, "loss": 758.3513, "step": 427 }, { "epoch": 0.8262548262548263, "grad_norm": 5.430158615112305, "learning_rate": 1.7374517374517377e-06, "loss": 754.9781, "step": 428 }, { "epoch": 0.8281853281853282, "grad_norm": 3.8903748989105225, "learning_rate": 1.7181467181467181e-06, "loss": 762.0396, "step": 429 }, { "epoch": 0.8301158301158301, "grad_norm": 3.2768187522888184, "learning_rate": 1.698841698841699e-06, "loss": 748.2692, "step": 430 }, { "epoch": 0.832046332046332, "grad_norm": 4.247284889221191, "learning_rate": 1.6795366795366796e-06, "loss": 759.6968, "step": 431 }, { "epoch": 0.833976833976834, "grad_norm": 3.205296039581299, "learning_rate": 1.6602316602316604e-06, "loss": 745.4493, "step": 432 }, { "epoch": 0.8359073359073359, "grad_norm": 3.3943800926208496, "learning_rate": 1.640926640926641e-06, "loss": 759.4459, "step": 433 }, { "epoch": 0.8378378378378378, "grad_norm": 3.8933515548706055, "learning_rate": 1.6216216216216219e-06, "loss": 768.5825, "step": 434 }, { "epoch": 0.8397683397683398, "grad_norm": 3.123276710510254, "learning_rate": 1.6023166023166027e-06, "loss": 749.004, "step": 435 }, { "epoch": 0.8416988416988417, "grad_norm": 3.028196334838867, "learning_rate": 1.583011583011583e-06, "loss": 754.6147, "step": 436 }, { "epoch": 0.8436293436293436, "grad_norm": 5.627590179443359, "learning_rate": 1.5637065637065637e-06, "loss": 764.9931, "step": 437 }, { "epoch": 0.8455598455598455, "grad_norm": 4.444796562194824, "learning_rate": 1.5444015444015445e-06, "loss": 756.2324, "step": 438 }, { "epoch": 0.8474903474903475, "grad_norm": 4.994812965393066, "learning_rate": 1.5250965250965251e-06, "loss": 764.1788, "step": 439 }, { "epoch": 0.8494208494208494, "grad_norm": 3.8100221157073975, "learning_rate": 1.505791505791506e-06, "loss": 748.0216, "step": 440 }, { "epoch": 0.8513513513513513, "grad_norm": 3.284438371658325, "learning_rate": 1.4864864864864868e-06, "loss": 744.6481, "step": 441 }, { "epoch": 0.8532818532818532, "grad_norm": 3.754772901535034, "learning_rate": 1.4671814671814674e-06, "loss": 759.6836, "step": 442 }, { "epoch": 0.8552123552123552, "grad_norm": 3.4503631591796875, "learning_rate": 1.4478764478764478e-06, "loss": 757.1558, "step": 443 }, { "epoch": 0.8571428571428571, "grad_norm": 3.9186127185821533, "learning_rate": 1.4285714285714286e-06, "loss": 749.7752, "step": 444 }, { "epoch": 0.859073359073359, "grad_norm": 5.067336082458496, "learning_rate": 1.4092664092664092e-06, "loss": 768.12, "step": 445 }, { "epoch": 0.861003861003861, "grad_norm": 3.5622620582580566, "learning_rate": 1.38996138996139e-06, "loss": 749.7698, "step": 446 }, { "epoch": 0.862934362934363, "grad_norm": 3.2070393562316895, "learning_rate": 1.3706563706563709e-06, "loss": 765.8137, "step": 447 }, { "epoch": 0.8648648648648649, "grad_norm": 3.920966386795044, "learning_rate": 1.3513513513513515e-06, "loss": 763.2885, "step": 448 }, { "epoch": 0.8667953667953668, "grad_norm": 4.828410625457764, "learning_rate": 1.3320463320463323e-06, "loss": 744.1356, "step": 449 }, { "epoch": 0.8687258687258688, "grad_norm": 2.9684271812438965, "learning_rate": 1.3127413127413127e-06, "loss": 748.5911, "step": 450 }, { "epoch": 0.8706563706563707, "grad_norm": 3.4309277534484863, "learning_rate": 1.2934362934362933e-06, "loss": 754.4804, "step": 451 }, { "epoch": 0.8725868725868726, "grad_norm": 3.2176759243011475, "learning_rate": 1.2741312741312742e-06, "loss": 763.7363, "step": 452 }, { "epoch": 0.8745173745173745, "grad_norm": 4.412746429443359, "learning_rate": 1.254826254826255e-06, "loss": 762.2845, "step": 453 }, { "epoch": 0.8764478764478765, "grad_norm": 3.3320093154907227, "learning_rate": 1.2355212355212356e-06, "loss": 737.4437, "step": 454 }, { "epoch": 0.8783783783783784, "grad_norm": 3.1412718296051025, "learning_rate": 1.2162162162162164e-06, "loss": 752.1414, "step": 455 }, { "epoch": 0.8803088803088803, "grad_norm": 2.9298524856567383, "learning_rate": 1.196911196911197e-06, "loss": 749.2881, "step": 456 }, { "epoch": 0.8822393822393823, "grad_norm": 3.4683337211608887, "learning_rate": 1.1776061776061777e-06, "loss": 750.4893, "step": 457 }, { "epoch": 0.8841698841698842, "grad_norm": 8.293436050415039, "learning_rate": 1.1583011583011585e-06, "loss": 759.465, "step": 458 }, { "epoch": 0.8861003861003861, "grad_norm": 3.0078611373901367, "learning_rate": 1.1389961389961391e-06, "loss": 765.3209, "step": 459 }, { "epoch": 0.888030888030888, "grad_norm": 3.5911474227905273, "learning_rate": 1.1196911196911197e-06, "loss": 748.1154, "step": 460 }, { "epoch": 0.88996138996139, "grad_norm": 3.626723527908325, "learning_rate": 1.1003861003861005e-06, "loss": 763.6529, "step": 461 }, { "epoch": 0.8918918918918919, "grad_norm": 3.322850465774536, "learning_rate": 1.0810810810810812e-06, "loss": 757.5947, "step": 462 }, { "epoch": 0.8938223938223938, "grad_norm": 3.866856575012207, "learning_rate": 1.0617760617760618e-06, "loss": 761.439, "step": 463 }, { "epoch": 0.8957528957528957, "grad_norm": 3.293731927871704, "learning_rate": 1.0424710424710426e-06, "loss": 763.1173, "step": 464 }, { "epoch": 0.8976833976833977, "grad_norm": 3.4967000484466553, "learning_rate": 1.0231660231660232e-06, "loss": 771.8659, "step": 465 }, { "epoch": 0.8996138996138996, "grad_norm": 2.9513790607452393, "learning_rate": 1.0038610038610038e-06, "loss": 759.2766, "step": 466 }, { "epoch": 0.9015444015444015, "grad_norm": 3.051725387573242, "learning_rate": 9.845559845559847e-07, "loss": 761.7484, "step": 467 }, { "epoch": 0.9034749034749034, "grad_norm": 3.222289562225342, "learning_rate": 9.652509652509653e-07, "loss": 753.144, "step": 468 }, { "epoch": 0.9054054054054054, "grad_norm": 4.781485557556152, "learning_rate": 9.459459459459461e-07, "loss": 766.5778, "step": 469 }, { "epoch": 0.9073359073359073, "grad_norm": 9.611944198608398, "learning_rate": 9.266409266409267e-07, "loss": 757.17, "step": 470 }, { "epoch": 0.9092664092664092, "grad_norm": 5.166945457458496, "learning_rate": 9.073359073359074e-07, "loss": 758.192, "step": 471 }, { "epoch": 0.9111969111969112, "grad_norm": 3.2668616771698, "learning_rate": 8.880308880308882e-07, "loss": 772.9589, "step": 472 }, { "epoch": 0.9131274131274131, "grad_norm": 3.808668851852417, "learning_rate": 8.687258687258689e-07, "loss": 743.6304, "step": 473 }, { "epoch": 0.915057915057915, "grad_norm": 3.6966071128845215, "learning_rate": 8.494208494208495e-07, "loss": 739.0352, "step": 474 }, { "epoch": 0.916988416988417, "grad_norm": 5.23868465423584, "learning_rate": 8.301158301158302e-07, "loss": 758.2206, "step": 475 }, { "epoch": 0.918918918918919, "grad_norm": 5.4658732414245605, "learning_rate": 8.108108108108109e-07, "loss": 745.1991, "step": 476 }, { "epoch": 0.9208494208494209, "grad_norm": 4.291003227233887, "learning_rate": 7.915057915057915e-07, "loss": 765.403, "step": 477 }, { "epoch": 0.9227799227799228, "grad_norm": 3.528947353363037, "learning_rate": 7.722007722007723e-07, "loss": 764.4042, "step": 478 }, { "epoch": 0.9247104247104247, "grad_norm": 3.179927349090576, "learning_rate": 7.52895752895753e-07, "loss": 757.1093, "step": 479 }, { "epoch": 0.9266409266409267, "grad_norm": 4.567697048187256, "learning_rate": 7.335907335907337e-07, "loss": 759.746, "step": 480 }, { "epoch": 0.9285714285714286, "grad_norm": 3.929335355758667, "learning_rate": 7.142857142857143e-07, "loss": 758.0447, "step": 481 }, { "epoch": 0.9305019305019305, "grad_norm": 5.094730377197266, "learning_rate": 6.94980694980695e-07, "loss": 757.1946, "step": 482 }, { "epoch": 0.9324324324324325, "grad_norm": 3.7125465869903564, "learning_rate": 6.756756756756758e-07, "loss": 770.7656, "step": 483 }, { "epoch": 0.9343629343629344, "grad_norm": 3.3825149536132812, "learning_rate": 6.563706563706564e-07, "loss": 747.5059, "step": 484 }, { "epoch": 0.9362934362934363, "grad_norm": 3.3807835578918457, "learning_rate": 6.370656370656371e-07, "loss": 767.5899, "step": 485 }, { "epoch": 0.9382239382239382, "grad_norm": 2.997955322265625, "learning_rate": 6.177606177606178e-07, "loss": 758.338, "step": 486 }, { "epoch": 0.9401544401544402, "grad_norm": 3.478146553039551, "learning_rate": 5.984555984555985e-07, "loss": 758.3116, "step": 487 }, { "epoch": 0.9420849420849421, "grad_norm": 3.0303502082824707, "learning_rate": 5.791505791505792e-07, "loss": 752.7023, "step": 488 }, { "epoch": 0.944015444015444, "grad_norm": 4.182223796844482, "learning_rate": 5.598455598455599e-07, "loss": 767.7703, "step": 489 }, { "epoch": 0.9459459459459459, "grad_norm": 3.0645620822906494, "learning_rate": 5.405405405405406e-07, "loss": 756.7206, "step": 490 }, { "epoch": 0.9478764478764479, "grad_norm": 3.5093557834625244, "learning_rate": 5.212355212355213e-07, "loss": 759.0178, "step": 491 }, { "epoch": 0.9498069498069498, "grad_norm": 5.231660842895508, "learning_rate": 5.019305019305019e-07, "loss": 758.1909, "step": 492 }, { "epoch": 0.9517374517374517, "grad_norm": 4.316704273223877, "learning_rate": 4.826254826254826e-07, "loss": 756.6906, "step": 493 }, { "epoch": 0.9536679536679536, "grad_norm": 3.842200994491577, "learning_rate": 4.6332046332046336e-07, "loss": 757.8163, "step": 494 }, { "epoch": 0.9555984555984556, "grad_norm": 3.351325511932373, "learning_rate": 4.440154440154441e-07, "loss": 762.0304, "step": 495 }, { "epoch": 0.9575289575289575, "grad_norm": 3.0159342288970947, "learning_rate": 4.2471042471042474e-07, "loss": 761.4176, "step": 496 }, { "epoch": 0.9594594594594594, "grad_norm": 3.074551582336426, "learning_rate": 4.0540540540540546e-07, "loss": 751.5521, "step": 497 }, { "epoch": 0.9613899613899614, "grad_norm": 3.896904945373535, "learning_rate": 3.8610038610038613e-07, "loss": 763.3478, "step": 498 }, { "epoch": 0.9633204633204633, "grad_norm": 3.919020175933838, "learning_rate": 3.6679536679536685e-07, "loss": 768.0314, "step": 499 }, { "epoch": 0.9652509652509652, "grad_norm": 2.949843645095825, "learning_rate": 3.474903474903475e-07, "loss": 755.5469, "step": 500 }, { "epoch": 0.9671814671814671, "grad_norm": 3.6389143466949463, "learning_rate": 3.281853281853282e-07, "loss": 744.184, "step": 501 }, { "epoch": 0.9691119691119691, "grad_norm": 4.229357719421387, "learning_rate": 3.088803088803089e-07, "loss": 761.7263, "step": 502 }, { "epoch": 0.971042471042471, "grad_norm": 3.1734883785247803, "learning_rate": 2.895752895752896e-07, "loss": 758.8358, "step": 503 }, { "epoch": 0.972972972972973, "grad_norm": 3.089228868484497, "learning_rate": 2.702702702702703e-07, "loss": 760.397, "step": 504 }, { "epoch": 0.974903474903475, "grad_norm": 3.926948070526123, "learning_rate": 2.5096525096525096e-07, "loss": 755.6573, "step": 505 }, { "epoch": 0.9768339768339769, "grad_norm": 2.9857828617095947, "learning_rate": 2.3166023166023168e-07, "loss": 767.6666, "step": 506 }, { "epoch": 0.9787644787644788, "grad_norm": 4.68814754486084, "learning_rate": 2.1235521235521237e-07, "loss": 767.6501, "step": 507 }, { "epoch": 0.9806949806949807, "grad_norm": 3.457016706466675, "learning_rate": 1.9305019305019306e-07, "loss": 760.5441, "step": 508 }, { "epoch": 0.9826254826254827, "grad_norm": 4.365758419036865, "learning_rate": 1.7374517374517376e-07, "loss": 760.8904, "step": 509 }, { "epoch": 0.9845559845559846, "grad_norm": 3.6141843795776367, "learning_rate": 1.5444015444015445e-07, "loss": 754.2842, "step": 510 }, { "epoch": 0.9864864864864865, "grad_norm": 4.281587600708008, "learning_rate": 1.3513513513513515e-07, "loss": 754.181, "step": 511 }, { "epoch": 0.9884169884169884, "grad_norm": 3.3735556602478027, "learning_rate": 1.1583011583011584e-07, "loss": 755.9177, "step": 512 }, { "epoch": 0.9903474903474904, "grad_norm": 3.126232624053955, "learning_rate": 9.652509652509653e-08, "loss": 750.3485, "step": 513 }, { "epoch": 0.9922779922779923, "grad_norm": 5.9563727378845215, "learning_rate": 7.722007722007723e-08, "loss": 766.379, "step": 514 }, { "epoch": 0.9942084942084942, "grad_norm": 3.0199837684631348, "learning_rate": 5.791505791505792e-08, "loss": 753.6818, "step": 515 }, { "epoch": 0.9961389961389961, "grad_norm": 3.0909945964813232, "learning_rate": 3.861003861003861e-08, "loss": 749.71, "step": 516 }, { "epoch": 0.9980694980694981, "grad_norm": 3.416806221008301, "learning_rate": 1.9305019305019306e-08, "loss": 772.7269, "step": 517 }, { "epoch": 1.0, "grad_norm": 5.65108585357666, "learning_rate": 0.0, "loss": 744.4813, "step": 518 }, { "epoch": 1.0, "step": 518, "total_flos": 1.1089078720895386e+18, "train_loss": 758.2892525573495, "train_runtime": 2215.2409, "train_samples_per_second": 239.051, "train_steps_per_second": 0.234 } ], "logging_steps": 1, "max_steps": 518, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 500, "total_flos": 1.1089078720895386e+18, "train_batch_size": 1024, "trial_name": null, "trial_params": null }