diff --git "a/trainer_state.json" "b/trainer_state.json" new file mode 100644--- /dev/null +++ "b/trainer_state.json" @@ -0,0 +1,5524 @@ +{ + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 2.0, + "eval_steps": 500, + "global_step": 782, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.0025575447570332483, + "grad_norm": 21.953563303736964, + "learning_rate": 9.999959651660741e-06, + "loss": 1.3458, + "step": 1 + }, + { + "epoch": 0.005115089514066497, + "grad_norm": 11.652464724143648, + "learning_rate": 9.999838607294157e-06, + "loss": 1.293, + "step": 2 + }, + { + "epoch": 0.0076726342710997444, + "grad_norm": 14.373069724258055, + "learning_rate": 9.999636868853824e-06, + "loss": 0.9527, + "step": 3 + }, + { + "epoch": 0.010230179028132993, + "grad_norm": 9.875205922264723, + "learning_rate": 9.999354439595668e-06, + "loss": 0.9394, + "step": 4 + }, + { + "epoch": 0.01278772378516624, + "grad_norm": 15.116196776322745, + "learning_rate": 9.998991324077906e-06, + "loss": 0.8439, + "step": 5 + }, + { + "epoch": 0.015345268542199489, + "grad_norm": 11.057152894995923, + "learning_rate": 9.998547528160987e-06, + "loss": 0.8405, + "step": 6 + }, + { + "epoch": 0.017902813299232736, + "grad_norm": 9.365688071554938, + "learning_rate": 9.998023059007477e-06, + "loss": 0.8734, + "step": 7 + }, + { + "epoch": 0.020460358056265986, + "grad_norm": 8.392619519229692, + "learning_rate": 9.997417925081963e-06, + "loss": 0.7449, + "step": 8 + }, + { + "epoch": 0.023017902813299233, + "grad_norm": 9.340778471253389, + "learning_rate": 9.996732136150902e-06, + "loss": 0.7876, + "step": 9 + }, + { + "epoch": 0.02557544757033248, + "grad_norm": 8.20102806123342, + "learning_rate": 9.995965703282472e-06, + "loss": 0.7954, + "step": 10 + }, + { + "epoch": 0.028132992327365727, + "grad_norm": 8.76971572309029, + "learning_rate": 9.995118638846394e-06, + "loss": 0.8333, + "step": 11 + }, + { + "epoch": 0.030690537084398978, + "grad_norm": 7.557620850007077, + "learning_rate": 9.99419095651372e-06, + "loss": 0.8554, + "step": 12 + }, + { + "epoch": 0.03324808184143223, + "grad_norm": 7.743714182477662, + "learning_rate": 9.993182671256633e-06, + "loss": 0.8637, + "step": 13 + }, + { + "epoch": 0.03580562659846547, + "grad_norm": 7.32084774886768, + "learning_rate": 9.992093799348182e-06, + "loss": 0.8621, + "step": 14 + }, + { + "epoch": 0.03836317135549872, + "grad_norm": 7.6800585762630655, + "learning_rate": 9.990924358362037e-06, + "loss": 0.7075, + "step": 15 + }, + { + "epoch": 0.04092071611253197, + "grad_norm": 7.053712401961599, + "learning_rate": 9.9896743671722e-06, + "loss": 0.8078, + "step": 16 + }, + { + "epoch": 0.043478260869565216, + "grad_norm": 7.957894711517596, + "learning_rate": 9.988343845952697e-06, + "loss": 0.8534, + "step": 17 + }, + { + "epoch": 0.04603580562659847, + "grad_norm": 7.77708191899406, + "learning_rate": 9.986932816177258e-06, + "loss": 0.7898, + "step": 18 + }, + { + "epoch": 0.04859335038363171, + "grad_norm": 8.86563729131372, + "learning_rate": 9.985441300618966e-06, + "loss": 0.7867, + "step": 19 + }, + { + "epoch": 0.05115089514066496, + "grad_norm": 7.9582904003704655, + "learning_rate": 9.98386932334989e-06, + "loss": 0.7112, + "step": 20 + }, + { + "epoch": 0.05370843989769821, + "grad_norm": 7.46439913319433, + "learning_rate": 9.982216909740703e-06, + "loss": 0.7207, + "step": 21 + }, + { + "epoch": 0.056265984654731455, + "grad_norm": 7.502417267747896, + "learning_rate": 9.980484086460258e-06, + "loss": 0.7635, + "step": 22 + }, + { + "epoch": 0.058823529411764705, + "grad_norm": 8.077410576519545, + "learning_rate": 9.978670881475173e-06, + "loss": 0.7568, + "step": 23 + }, + { + "epoch": 0.061381074168797956, + "grad_norm": 8.530645418517887, + "learning_rate": 9.976777324049374e-06, + "loss": 0.8006, + "step": 24 + }, + { + "epoch": 0.0639386189258312, + "grad_norm": 7.884014258298183, + "learning_rate": 9.974803444743617e-06, + "loss": 0.741, + "step": 25 + }, + { + "epoch": 0.06649616368286446, + "grad_norm": 6.917821196530823, + "learning_rate": 9.972749275415005e-06, + "loss": 0.6668, + "step": 26 + }, + { + "epoch": 0.06905370843989769, + "grad_norm": 6.951997744847425, + "learning_rate": 9.970614849216465e-06, + "loss": 0.7619, + "step": 27 + }, + { + "epoch": 0.07161125319693094, + "grad_norm": 6.541689929654185, + "learning_rate": 9.96840020059622e-06, + "loss": 0.7667, + "step": 28 + }, + { + "epoch": 0.0741687979539642, + "grad_norm": 7.022380585505507, + "learning_rate": 9.966105365297226e-06, + "loss": 0.7176, + "step": 29 + }, + { + "epoch": 0.07672634271099744, + "grad_norm": 6.912790106638039, + "learning_rate": 9.963730380356599e-06, + "loss": 0.7199, + "step": 30 + }, + { + "epoch": 0.0792838874680307, + "grad_norm": 6.0140784283728275, + "learning_rate": 9.96127528410502e-06, + "loss": 0.7356, + "step": 31 + }, + { + "epoch": 0.08184143222506395, + "grad_norm": 6.50099127837641, + "learning_rate": 9.958740116166113e-06, + "loss": 0.6741, + "step": 32 + }, + { + "epoch": 0.08439897698209718, + "grad_norm": 6.196159967289502, + "learning_rate": 9.9561249174558e-06, + "loss": 0.6453, + "step": 33 + }, + { + "epoch": 0.08695652173913043, + "grad_norm": 7.115083432855918, + "learning_rate": 9.953429730181653e-06, + "loss": 0.6921, + "step": 34 + }, + { + "epoch": 0.08951406649616368, + "grad_norm": 8.373221612447116, + "learning_rate": 9.950654597842209e-06, + "loss": 0.6904, + "step": 35 + }, + { + "epoch": 0.09207161125319693, + "grad_norm": 7.708793853244428, + "learning_rate": 9.947799565226253e-06, + "loss": 0.7295, + "step": 36 + }, + { + "epoch": 0.09462915601023018, + "grad_norm": 8.81215615740888, + "learning_rate": 9.944864678412118e-06, + "loss": 0.6806, + "step": 37 + }, + { + "epoch": 0.09718670076726342, + "grad_norm": 7.456995220385941, + "learning_rate": 9.94184998476693e-06, + "loss": 0.6659, + "step": 38 + }, + { + "epoch": 0.09974424552429667, + "grad_norm": 7.636846126721769, + "learning_rate": 9.938755532945838e-06, + "loss": 0.7073, + "step": 39 + }, + { + "epoch": 0.10230179028132992, + "grad_norm": 8.018319876671674, + "learning_rate": 9.93558137289124e-06, + "loss": 0.8063, + "step": 40 + }, + { + "epoch": 0.10485933503836317, + "grad_norm": 6.429546682427428, + "learning_rate": 9.932327555831972e-06, + "loss": 0.5498, + "step": 41 + }, + { + "epoch": 0.10741687979539642, + "grad_norm": 6.106558471191083, + "learning_rate": 9.928994134282477e-06, + "loss": 0.6125, + "step": 42 + }, + { + "epoch": 0.10997442455242967, + "grad_norm": 7.220242425194578, + "learning_rate": 9.925581162041967e-06, + "loss": 0.6553, + "step": 43 + }, + { + "epoch": 0.11253196930946291, + "grad_norm": 7.406518370609504, + "learning_rate": 9.922088694193546e-06, + "loss": 0.7235, + "step": 44 + }, + { + "epoch": 0.11508951406649616, + "grad_norm": 6.232863127816167, + "learning_rate": 9.918516787103322e-06, + "loss": 0.6302, + "step": 45 + }, + { + "epoch": 0.11764705882352941, + "grad_norm": 6.960585716404723, + "learning_rate": 9.91486549841951e-06, + "loss": 0.6631, + "step": 46 + }, + { + "epoch": 0.12020460358056266, + "grad_norm": 7.014565196822424, + "learning_rate": 9.911134887071477e-06, + "loss": 0.6013, + "step": 47 + }, + { + "epoch": 0.12276214833759591, + "grad_norm": 6.747724971317866, + "learning_rate": 9.907325013268816e-06, + "loss": 0.7311, + "step": 48 + }, + { + "epoch": 0.12531969309462915, + "grad_norm": 6.343001355283702, + "learning_rate": 9.903435938500356e-06, + "loss": 0.5662, + "step": 49 + }, + { + "epoch": 0.1278772378516624, + "grad_norm": 6.68715844293194, + "learning_rate": 9.899467725533181e-06, + "loss": 0.6746, + "step": 50 + }, + { + "epoch": 0.13043478260869565, + "grad_norm": 6.670242029824649, + "learning_rate": 9.895420438411616e-06, + "loss": 0.5636, + "step": 51 + }, + { + "epoch": 0.1329923273657289, + "grad_norm": 7.6565238394007, + "learning_rate": 9.89129414245618e-06, + "loss": 0.6959, + "step": 52 + }, + { + "epoch": 0.13554987212276215, + "grad_norm": 6.322260761537018, + "learning_rate": 9.887088904262557e-06, + "loss": 0.6475, + "step": 53 + }, + { + "epoch": 0.13810741687979539, + "grad_norm": 7.424958009645482, + "learning_rate": 9.882804791700488e-06, + "loss": 0.7381, + "step": 54 + }, + { + "epoch": 0.14066496163682865, + "grad_norm": 6.484080185567617, + "learning_rate": 9.878441873912712e-06, + "loss": 0.5978, + "step": 55 + }, + { + "epoch": 0.1432225063938619, + "grad_norm": 7.904610864584768, + "learning_rate": 9.87400022131382e-06, + "loss": 0.7156, + "step": 56 + }, + { + "epoch": 0.14578005115089515, + "grad_norm": 6.091768577876145, + "learning_rate": 9.869479905589136e-06, + "loss": 0.5674, + "step": 57 + }, + { + "epoch": 0.1483375959079284, + "grad_norm": 6.338206394229405, + "learning_rate": 9.864880999693551e-06, + "loss": 0.5511, + "step": 58 + }, + { + "epoch": 0.15089514066496162, + "grad_norm": 7.393436860670507, + "learning_rate": 9.860203577850353e-06, + "loss": 0.6305, + "step": 59 + }, + { + "epoch": 0.1534526854219949, + "grad_norm": 6.520098900315275, + "learning_rate": 9.855447715550024e-06, + "loss": 0.6191, + "step": 60 + }, + { + "epoch": 0.15601023017902813, + "grad_norm": 6.415362809418058, + "learning_rate": 9.850613489549018e-06, + "loss": 0.6296, + "step": 61 + }, + { + "epoch": 0.1585677749360614, + "grad_norm": 7.9580764103093635, + "learning_rate": 9.845700977868536e-06, + "loss": 0.6384, + "step": 62 + }, + { + "epoch": 0.16112531969309463, + "grad_norm": 7.1082899489573315, + "learning_rate": 9.840710259793251e-06, + "loss": 0.6021, + "step": 63 + }, + { + "epoch": 0.1636828644501279, + "grad_norm": 6.386504712691233, + "learning_rate": 9.835641415870038e-06, + "loss": 0.6873, + "step": 64 + }, + { + "epoch": 0.16624040920716113, + "grad_norm": 7.4797007816006555, + "learning_rate": 9.830494527906671e-06, + "loss": 0.5648, + "step": 65 + }, + { + "epoch": 0.16879795396419436, + "grad_norm": 6.748679766187366, + "learning_rate": 9.825269678970502e-06, + "loss": 0.566, + "step": 66 + }, + { + "epoch": 0.17135549872122763, + "grad_norm": 6.56170099270124, + "learning_rate": 9.819966953387122e-06, + "loss": 0.6398, + "step": 67 + }, + { + "epoch": 0.17391304347826086, + "grad_norm": 5.941005918686002, + "learning_rate": 9.814586436738998e-06, + "loss": 0.6658, + "step": 68 + }, + { + "epoch": 0.17647058823529413, + "grad_norm": 5.942719440981736, + "learning_rate": 9.809128215864096e-06, + "loss": 0.5971, + "step": 69 + }, + { + "epoch": 0.17902813299232737, + "grad_norm": 6.4968806279365925, + "learning_rate": 9.803592378854476e-06, + "loss": 0.6047, + "step": 70 + }, + { + "epoch": 0.1815856777493606, + "grad_norm": 5.577555703482892, + "learning_rate": 9.797979015054868e-06, + "loss": 0.5534, + "step": 71 + }, + { + "epoch": 0.18414322250639387, + "grad_norm": 6.262423254220968, + "learning_rate": 9.792288215061237e-06, + "loss": 0.5755, + "step": 72 + }, + { + "epoch": 0.1867007672634271, + "grad_norm": 7.17658874763877, + "learning_rate": 9.786520070719313e-06, + "loss": 0.5511, + "step": 73 + }, + { + "epoch": 0.18925831202046037, + "grad_norm": 5.532267867303149, + "learning_rate": 9.780674675123113e-06, + "loss": 0.4965, + "step": 74 + }, + { + "epoch": 0.1918158567774936, + "grad_norm": 7.040268722404183, + "learning_rate": 9.77475212261344e-06, + "loss": 0.6644, + "step": 75 + }, + { + "epoch": 0.19437340153452684, + "grad_norm": 6.774497918030386, + "learning_rate": 9.768752508776358e-06, + "loss": 0.491, + "step": 76 + }, + { + "epoch": 0.1969309462915601, + "grad_norm": 5.776934238941505, + "learning_rate": 9.762675930441647e-06, + "loss": 0.4861, + "step": 77 + }, + { + "epoch": 0.19948849104859334, + "grad_norm": 6.47046232974907, + "learning_rate": 9.756522485681247e-06, + "loss": 0.6273, + "step": 78 + }, + { + "epoch": 0.2020460358056266, + "grad_norm": 6.712438744552012, + "learning_rate": 9.750292273807666e-06, + "loss": 0.7368, + "step": 79 + }, + { + "epoch": 0.20460358056265984, + "grad_norm": 8.615585919768426, + "learning_rate": 9.743985395372387e-06, + "loss": 0.6335, + "step": 80 + }, + { + "epoch": 0.2071611253196931, + "grad_norm": 8.071799145378323, + "learning_rate": 9.737601952164238e-06, + "loss": 0.6612, + "step": 81 + }, + { + "epoch": 0.20971867007672634, + "grad_norm": 6.823667160150434, + "learning_rate": 9.73114204720775e-06, + "loss": 0.5456, + "step": 82 + }, + { + "epoch": 0.21227621483375958, + "grad_norm": 5.888063194216176, + "learning_rate": 9.724605784761501e-06, + "loss": 0.5959, + "step": 83 + }, + { + "epoch": 0.21483375959079284, + "grad_norm": 7.210449720464657, + "learning_rate": 9.717993270316421e-06, + "loss": 0.5919, + "step": 84 + }, + { + "epoch": 0.21739130434782608, + "grad_norm": 7.06901541191221, + "learning_rate": 9.711304610594104e-06, + "loss": 0.5882, + "step": 85 + }, + { + "epoch": 0.21994884910485935, + "grad_norm": 6.374122555190542, + "learning_rate": 9.704539913545073e-06, + "loss": 0.6133, + "step": 86 + }, + { + "epoch": 0.22250639386189258, + "grad_norm": 6.630925756972144, + "learning_rate": 9.697699288347043e-06, + "loss": 0.5664, + "step": 87 + }, + { + "epoch": 0.22506393861892582, + "grad_norm": 5.966658672241774, + "learning_rate": 9.690782845403164e-06, + "loss": 0.6244, + "step": 88 + }, + { + "epoch": 0.22762148337595908, + "grad_norm": 5.5770246123900264, + "learning_rate": 9.683790696340229e-06, + "loss": 0.5334, + "step": 89 + }, + { + "epoch": 0.23017902813299232, + "grad_norm": 6.522499319244262, + "learning_rate": 9.676722954006878e-06, + "loss": 0.6739, + "step": 90 + }, + { + "epoch": 0.23273657289002558, + "grad_norm": 6.4235805218363025, + "learning_rate": 9.669579732471779e-06, + "loss": 0.6595, + "step": 91 + }, + { + "epoch": 0.23529411764705882, + "grad_norm": 6.3431696131235284, + "learning_rate": 9.66236114702178e-06, + "loss": 0.6023, + "step": 92 + }, + { + "epoch": 0.23785166240409208, + "grad_norm": 6.429883886786425, + "learning_rate": 9.655067314160058e-06, + "loss": 0.5986, + "step": 93 + }, + { + "epoch": 0.24040920716112532, + "grad_norm": 6.701368521252926, + "learning_rate": 9.647698351604227e-06, + "loss": 0.6569, + "step": 94 + }, + { + "epoch": 0.24296675191815856, + "grad_norm": 6.626338500498468, + "learning_rate": 9.640254378284447e-06, + "loss": 0.5552, + "step": 95 + }, + { + "epoch": 0.24552429667519182, + "grad_norm": 6.872036104522696, + "learning_rate": 9.632735514341508e-06, + "loss": 0.5384, + "step": 96 + }, + { + "epoch": 0.24808184143222506, + "grad_norm": 7.976561364452108, + "learning_rate": 9.625141881124874e-06, + "loss": 0.6225, + "step": 97 + }, + { + "epoch": 0.2506393861892583, + "grad_norm": 6.979721775572128, + "learning_rate": 9.617473601190743e-06, + "loss": 0.5937, + "step": 98 + }, + { + "epoch": 0.2531969309462916, + "grad_norm": 7.519173375318072, + "learning_rate": 9.609730798300056e-06, + "loss": 0.5673, + "step": 99 + }, + { + "epoch": 0.2557544757033248, + "grad_norm": 6.501376080036557, + "learning_rate": 9.601913597416513e-06, + "loss": 0.6167, + "step": 100 + }, + { + "epoch": 0.25831202046035806, + "grad_norm": 7.232025532728033, + "learning_rate": 9.594022124704541e-06, + "loss": 0.6528, + "step": 101 + }, + { + "epoch": 0.2608695652173913, + "grad_norm": 5.821804513422366, + "learning_rate": 9.586056507527266e-06, + "loss": 0.5839, + "step": 102 + }, + { + "epoch": 0.26342710997442453, + "grad_norm": 7.447246210872055, + "learning_rate": 9.578016874444459e-06, + "loss": 0.5425, + "step": 103 + }, + { + "epoch": 0.2659846547314578, + "grad_norm": 5.023721202880345, + "learning_rate": 9.569903355210457e-06, + "loss": 0.4649, + "step": 104 + }, + { + "epoch": 0.26854219948849106, + "grad_norm": 5.3855220410063165, + "learning_rate": 9.561716080772072e-06, + "loss": 0.5362, + "step": 105 + }, + { + "epoch": 0.2710997442455243, + "grad_norm": 5.755012965635693, + "learning_rate": 9.55345518326647e-06, + "loss": 0.637, + "step": 106 + }, + { + "epoch": 0.27365728900255754, + "grad_norm": 6.317278618265475, + "learning_rate": 9.545120796019056e-06, + "loss": 0.6073, + "step": 107 + }, + { + "epoch": 0.27621483375959077, + "grad_norm": 6.556750652969424, + "learning_rate": 9.5367130535413e-06, + "loss": 0.6106, + "step": 108 + }, + { + "epoch": 0.27877237851662406, + "grad_norm": 7.234204791297877, + "learning_rate": 9.528232091528578e-06, + "loss": 0.5537, + "step": 109 + }, + { + "epoch": 0.2813299232736573, + "grad_norm": 6.44863975829238, + "learning_rate": 9.519678046857987e-06, + "loss": 0.6654, + "step": 110 + }, + { + "epoch": 0.28388746803069054, + "grad_norm": 5.882964477582141, + "learning_rate": 9.511051057586125e-06, + "loss": 0.5723, + "step": 111 + }, + { + "epoch": 0.2864450127877238, + "grad_norm": 5.682881601819309, + "learning_rate": 9.502351262946865e-06, + "loss": 0.5325, + "step": 112 + }, + { + "epoch": 0.289002557544757, + "grad_norm": 5.907234833224297, + "learning_rate": 9.493578803349117e-06, + "loss": 0.6238, + "step": 113 + }, + { + "epoch": 0.2915601023017903, + "grad_norm": 6.539157912120955, + "learning_rate": 9.48473382037455e-06, + "loss": 0.6228, + "step": 114 + }, + { + "epoch": 0.29411764705882354, + "grad_norm": 6.263387218751874, + "learning_rate": 9.475816456775313e-06, + "loss": 0.5954, + "step": 115 + }, + { + "epoch": 0.2966751918158568, + "grad_norm": 5.93110985413684, + "learning_rate": 9.466826856471728e-06, + "loss": 0.6008, + "step": 116 + }, + { + "epoch": 0.29923273657289, + "grad_norm": 6.326093043591936, + "learning_rate": 9.457765164549979e-06, + "loss": 0.5834, + "step": 117 + }, + { + "epoch": 0.30179028132992325, + "grad_norm": 8.333855109469555, + "learning_rate": 9.448631527259749e-06, + "loss": 0.7357, + "step": 118 + }, + { + "epoch": 0.30434782608695654, + "grad_norm": 7.352912958641833, + "learning_rate": 9.439426092011877e-06, + "loss": 0.5621, + "step": 119 + }, + { + "epoch": 0.3069053708439898, + "grad_norm": 6.812130929920315, + "learning_rate": 9.430149007375974e-06, + "loss": 0.6281, + "step": 120 + }, + { + "epoch": 0.309462915601023, + "grad_norm": 6.244862089177942, + "learning_rate": 9.42080042307802e-06, + "loss": 0.6083, + "step": 121 + }, + { + "epoch": 0.31202046035805625, + "grad_norm": 5.91824427054237, + "learning_rate": 9.411380489997962e-06, + "loss": 0.5141, + "step": 122 + }, + { + "epoch": 0.3145780051150895, + "grad_norm": 5.9138584985365075, + "learning_rate": 9.401889360167256e-06, + "loss": 0.5525, + "step": 123 + }, + { + "epoch": 0.3171355498721228, + "grad_norm": 6.264682424037789, + "learning_rate": 9.392327186766434e-06, + "loss": 0.5049, + "step": 124 + }, + { + "epoch": 0.319693094629156, + "grad_norm": 5.838854049045799, + "learning_rate": 9.382694124122624e-06, + "loss": 0.5835, + "step": 125 + }, + { + "epoch": 0.32225063938618925, + "grad_norm": 6.39551091760669, + "learning_rate": 9.372990327707057e-06, + "loss": 0.5132, + "step": 126 + }, + { + "epoch": 0.3248081841432225, + "grad_norm": 6.716488574379217, + "learning_rate": 9.36321595413256e-06, + "loss": 0.5372, + "step": 127 + }, + { + "epoch": 0.3273657289002558, + "grad_norm": 5.742568960781894, + "learning_rate": 9.353371161151032e-06, + "loss": 0.6203, + "step": 128 + }, + { + "epoch": 0.329923273657289, + "grad_norm": 6.772868273583815, + "learning_rate": 9.34345610765089e-06, + "loss": 0.4926, + "step": 129 + }, + { + "epoch": 0.33248081841432225, + "grad_norm": 7.185477633140159, + "learning_rate": 9.333470953654513e-06, + "loss": 0.6842, + "step": 130 + }, + { + "epoch": 0.3350383631713555, + "grad_norm": 6.296716433927121, + "learning_rate": 9.32341586031565e-06, + "loss": 0.5163, + "step": 131 + }, + { + "epoch": 0.3375959079283887, + "grad_norm": 8.081152679238958, + "learning_rate": 9.31329098991683e-06, + "loss": 0.7902, + "step": 132 + }, + { + "epoch": 0.340153452685422, + "grad_norm": 6.861874747525335, + "learning_rate": 9.303096505866734e-06, + "loss": 0.5645, + "step": 133 + }, + { + "epoch": 0.34271099744245526, + "grad_norm": 6.171268132782227, + "learning_rate": 9.292832572697566e-06, + "loss": 0.6114, + "step": 134 + }, + { + "epoch": 0.3452685421994885, + "grad_norm": 6.552747924757749, + "learning_rate": 9.282499356062385e-06, + "loss": 0.5954, + "step": 135 + }, + { + "epoch": 0.34782608695652173, + "grad_norm": 6.847724657541522, + "learning_rate": 9.272097022732444e-06, + "loss": 0.6485, + "step": 136 + }, + { + "epoch": 0.35038363171355497, + "grad_norm": 6.507209809740027, + "learning_rate": 9.261625740594494e-06, + "loss": 0.6159, + "step": 137 + }, + { + "epoch": 0.35294117647058826, + "grad_norm": 7.252605300424873, + "learning_rate": 9.251085678648072e-06, + "loss": 0.5576, + "step": 138 + }, + { + "epoch": 0.3554987212276215, + "grad_norm": 7.119115054572448, + "learning_rate": 9.240477007002777e-06, + "loss": 0.7135, + "step": 139 + }, + { + "epoch": 0.35805626598465473, + "grad_norm": 7.349847252060697, + "learning_rate": 9.22979989687552e-06, + "loss": 0.6444, + "step": 140 + }, + { + "epoch": 0.36061381074168797, + "grad_norm": 5.765325529890724, + "learning_rate": 9.219054520587766e-06, + "loss": 0.4233, + "step": 141 + }, + { + "epoch": 0.3631713554987212, + "grad_norm": 5.112678313504443, + "learning_rate": 9.208241051562753e-06, + "loss": 0.5447, + "step": 142 + }, + { + "epoch": 0.3657289002557545, + "grad_norm": 7.144208807574378, + "learning_rate": 9.197359664322684e-06, + "loss": 0.5891, + "step": 143 + }, + { + "epoch": 0.36828644501278773, + "grad_norm": 6.278104080681577, + "learning_rate": 9.186410534485924e-06, + "loss": 0.5701, + "step": 144 + }, + { + "epoch": 0.37084398976982097, + "grad_norm": 6.887145963626663, + "learning_rate": 9.175393838764153e-06, + "loss": 0.5502, + "step": 145 + }, + { + "epoch": 0.3734015345268542, + "grad_norm": 6.023815922626531, + "learning_rate": 9.164309754959523e-06, + "loss": 0.5286, + "step": 146 + }, + { + "epoch": 0.37595907928388744, + "grad_norm": 5.517318940161725, + "learning_rate": 9.153158461961782e-06, + "loss": 0.4433, + "step": 147 + }, + { + "epoch": 0.37851662404092073, + "grad_norm": 6.019318968473694, + "learning_rate": 9.14194013974539e-06, + "loss": 0.5065, + "step": 148 + }, + { + "epoch": 0.38107416879795397, + "grad_norm": 5.6444749821227385, + "learning_rate": 9.130654969366619e-06, + "loss": 0.501, + "step": 149 + }, + { + "epoch": 0.3836317135549872, + "grad_norm": 6.559732586020477, + "learning_rate": 9.11930313296062e-06, + "loss": 0.6101, + "step": 150 + }, + { + "epoch": 0.38618925831202044, + "grad_norm": 6.2279396393587705, + "learning_rate": 9.107884813738492e-06, + "loss": 0.5938, + "step": 151 + }, + { + "epoch": 0.3887468030690537, + "grad_norm": 7.319750378815782, + "learning_rate": 9.096400195984322e-06, + "loss": 0.4252, + "step": 152 + }, + { + "epoch": 0.391304347826087, + "grad_norm": 7.088600256578154, + "learning_rate": 9.08484946505221e-06, + "loss": 0.5793, + "step": 153 + }, + { + "epoch": 0.3938618925831202, + "grad_norm": 6.573409174775933, + "learning_rate": 9.073232807363283e-06, + "loss": 0.5026, + "step": 154 + }, + { + "epoch": 0.39641943734015345, + "grad_norm": 7.980317521312881, + "learning_rate": 9.061550410402677e-06, + "loss": 0.6736, + "step": 155 + }, + { + "epoch": 0.3989769820971867, + "grad_norm": 6.87810163307716, + "learning_rate": 9.049802462716521e-06, + "loss": 0.493, + "step": 156 + }, + { + "epoch": 0.40153452685422, + "grad_norm": 7.407535434490462, + "learning_rate": 9.037989153908882e-06, + "loss": 0.5762, + "step": 157 + }, + { + "epoch": 0.4040920716112532, + "grad_norm": 6.310545687817972, + "learning_rate": 9.026110674638722e-06, + "loss": 0.5802, + "step": 158 + }, + { + "epoch": 0.40664961636828645, + "grad_norm": 6.538278944739297, + "learning_rate": 9.0141672166168e-06, + "loss": 0.4665, + "step": 159 + }, + { + "epoch": 0.4092071611253197, + "grad_norm": 6.25525186329276, + "learning_rate": 9.002158972602599e-06, + "loss": 0.65, + "step": 160 + }, + { + "epoch": 0.4117647058823529, + "grad_norm": 7.134439997933502, + "learning_rate": 8.990086136401199e-06, + "loss": 0.6436, + "step": 161 + }, + { + "epoch": 0.4143222506393862, + "grad_norm": 6.907730334313879, + "learning_rate": 8.977948902860154e-06, + "loss": 0.6688, + "step": 162 + }, + { + "epoch": 0.41687979539641945, + "grad_norm": 5.60949076779962, + "learning_rate": 8.965747467866355e-06, + "loss": 0.4263, + "step": 163 + }, + { + "epoch": 0.4194373401534527, + "grad_norm": 7.153866044006984, + "learning_rate": 8.953482028342853e-06, + "loss": 0.707, + "step": 164 + }, + { + "epoch": 0.4219948849104859, + "grad_norm": 5.749873195369449, + "learning_rate": 8.9411527822457e-06, + "loss": 0.5522, + "step": 165 + }, + { + "epoch": 0.42455242966751916, + "grad_norm": 6.450113679037296, + "learning_rate": 8.92875992856073e-06, + "loss": 0.491, + "step": 166 + }, + { + "epoch": 0.42710997442455245, + "grad_norm": 7.10664387732353, + "learning_rate": 8.916303667300373e-06, + "loss": 0.5526, + "step": 167 + }, + { + "epoch": 0.4296675191815857, + "grad_norm": 5.908129226489756, + "learning_rate": 8.903784199500412e-06, + "loss": 0.4589, + "step": 168 + }, + { + "epoch": 0.4322250639386189, + "grad_norm": 6.036965674381112, + "learning_rate": 8.89120172721674e-06, + "loss": 0.5393, + "step": 169 + }, + { + "epoch": 0.43478260869565216, + "grad_norm": 6.557422143271899, + "learning_rate": 8.8785564535221e-06, + "loss": 0.4947, + "step": 170 + }, + { + "epoch": 0.4373401534526854, + "grad_norm": 7.091654604179938, + "learning_rate": 8.86584858250281e-06, + "loss": 0.5744, + "step": 171 + }, + { + "epoch": 0.4398976982097187, + "grad_norm": 6.839942726126082, + "learning_rate": 8.853078319255466e-06, + "loss": 0.5621, + "step": 172 + }, + { + "epoch": 0.4424552429667519, + "grad_norm": 6.705305452145543, + "learning_rate": 8.840245869883635e-06, + "loss": 0.6277, + "step": 173 + }, + { + "epoch": 0.44501278772378516, + "grad_norm": 6.410914353629197, + "learning_rate": 8.827351441494525e-06, + "loss": 0.5795, + "step": 174 + }, + { + "epoch": 0.4475703324808184, + "grad_norm": 6.178508928681474, + "learning_rate": 8.814395242195642e-06, + "loss": 0.5039, + "step": 175 + }, + { + "epoch": 0.45012787723785164, + "grad_norm": 5.283454576226612, + "learning_rate": 8.80137748109144e-06, + "loss": 0.4565, + "step": 176 + }, + { + "epoch": 0.45268542199488493, + "grad_norm": 5.78397157032685, + "learning_rate": 8.78829836827993e-06, + "loss": 0.5435, + "step": 177 + }, + { + "epoch": 0.45524296675191817, + "grad_norm": 5.014711664858047, + "learning_rate": 8.77515811484931e-06, + "loss": 0.4744, + "step": 178 + }, + { + "epoch": 0.4578005115089514, + "grad_norm": 5.673473816390766, + "learning_rate": 8.761956932874539e-06, + "loss": 0.4794, + "step": 179 + }, + { + "epoch": 0.46035805626598464, + "grad_norm": 6.548843702355434, + "learning_rate": 8.748695035413925e-06, + "loss": 0.5124, + "step": 180 + }, + { + "epoch": 0.4629156010230179, + "grad_norm": 5.294743068866496, + "learning_rate": 8.735372636505681e-06, + "loss": 0.4964, + "step": 181 + }, + { + "epoch": 0.46547314578005117, + "grad_norm": 5.307479046129796, + "learning_rate": 8.72198995116448e-06, + "loss": 0.4848, + "step": 182 + }, + { + "epoch": 0.4680306905370844, + "grad_norm": 5.862290813329295, + "learning_rate": 8.708547195377968e-06, + "loss": 0.6168, + "step": 183 + }, + { + "epoch": 0.47058823529411764, + "grad_norm": 6.660472406940894, + "learning_rate": 8.695044586103297e-06, + "loss": 0.6317, + "step": 184 + }, + { + "epoch": 0.4731457800511509, + "grad_norm": 6.239173715990654, + "learning_rate": 8.68148234126361e-06, + "loss": 0.5712, + "step": 185 + }, + { + "epoch": 0.47570332480818417, + "grad_norm": 6.855169313192307, + "learning_rate": 8.667860679744529e-06, + "loss": 0.4569, + "step": 186 + }, + { + "epoch": 0.4782608695652174, + "grad_norm": 6.098286103768463, + "learning_rate": 8.65417982139062e-06, + "loss": 0.5377, + "step": 187 + }, + { + "epoch": 0.48081841432225064, + "grad_norm": 6.687927077404218, + "learning_rate": 8.640439987001855e-06, + "loss": 0.5466, + "step": 188 + }, + { + "epoch": 0.4833759590792839, + "grad_norm": 5.06986215439764, + "learning_rate": 8.626641398330027e-06, + "loss": 0.4597, + "step": 189 + }, + { + "epoch": 0.4859335038363171, + "grad_norm": 6.615207228778788, + "learning_rate": 8.612784278075195e-06, + "loss": 0.6486, + "step": 190 + }, + { + "epoch": 0.4884910485933504, + "grad_norm": 8.206032710195597, + "learning_rate": 8.598868849882074e-06, + "loss": 0.53, + "step": 191 + }, + { + "epoch": 0.49104859335038364, + "grad_norm": 6.512201203410748, + "learning_rate": 8.58489533833643e-06, + "loss": 0.5075, + "step": 192 + }, + { + "epoch": 0.4936061381074169, + "grad_norm": 6.623708653660542, + "learning_rate": 8.570863968961456e-06, + "loss": 0.4697, + "step": 193 + }, + { + "epoch": 0.4961636828644501, + "grad_norm": 4.93093459028815, + "learning_rate": 8.556774968214134e-06, + "loss": 0.5169, + "step": 194 + }, + { + "epoch": 0.49872122762148335, + "grad_norm": 5.787452319450779, + "learning_rate": 8.542628563481577e-06, + "loss": 0.5482, + "step": 195 + }, + { + "epoch": 0.5012787723785166, + "grad_norm": 5.88850708880366, + "learning_rate": 8.52842498307736e-06, + "loss": 0.6134, + "step": 196 + }, + { + "epoch": 0.5038363171355499, + "grad_norm": 6.696557687225988, + "learning_rate": 8.514164456237835e-06, + "loss": 0.6447, + "step": 197 + }, + { + "epoch": 0.5063938618925832, + "grad_norm": 7.195386414110228, + "learning_rate": 8.499847213118431e-06, + "loss": 0.5117, + "step": 198 + }, + { + "epoch": 0.5089514066496164, + "grad_norm": 5.825208905742397, + "learning_rate": 8.485473484789944e-06, + "loss": 0.5152, + "step": 199 + }, + { + "epoch": 0.5115089514066496, + "grad_norm": 6.118462869888847, + "learning_rate": 8.471043503234796e-06, + "loss": 0.6536, + "step": 200 + }, + { + "epoch": 0.5140664961636828, + "grad_norm": 7.551332041886624, + "learning_rate": 8.45655750134331e-06, + "loss": 0.6084, + "step": 201 + }, + { + "epoch": 0.5166240409207161, + "grad_norm": 5.225118027592022, + "learning_rate": 8.442015712909926e-06, + "loss": 0.4555, + "step": 202 + }, + { + "epoch": 0.5191815856777494, + "grad_norm": 5.3592959184851265, + "learning_rate": 8.427418372629456e-06, + "loss": 0.4821, + "step": 203 + }, + { + "epoch": 0.5217391304347826, + "grad_norm": 6.212770160868918, + "learning_rate": 8.412765716093273e-06, + "loss": 0.5149, + "step": 204 + }, + { + "epoch": 0.5242966751918159, + "grad_norm": 5.519259201547804, + "learning_rate": 8.398057979785515e-06, + "loss": 0.4876, + "step": 205 + }, + { + "epoch": 0.5268542199488491, + "grad_norm": 6.847844140759948, + "learning_rate": 8.383295401079284e-06, + "loss": 0.5245, + "step": 206 + }, + { + "epoch": 0.5294117647058824, + "grad_norm": 6.473882379231715, + "learning_rate": 8.368478218232787e-06, + "loss": 0.5319, + "step": 207 + }, + { + "epoch": 0.5319693094629157, + "grad_norm": 4.995240865011453, + "learning_rate": 8.353606670385514e-06, + "loss": 0.4201, + "step": 208 + }, + { + "epoch": 0.5345268542199488, + "grad_norm": 5.272197527549254, + "learning_rate": 8.338680997554372e-06, + "loss": 0.4832, + "step": 209 + }, + { + "epoch": 0.5370843989769821, + "grad_norm": 5.585738561949535, + "learning_rate": 8.3237014406298e-06, + "loss": 0.4929, + "step": 210 + }, + { + "epoch": 0.5396419437340153, + "grad_norm": 6.239332915949274, + "learning_rate": 8.308668241371897e-06, + "loss": 0.4171, + "step": 211 + }, + { + "epoch": 0.5421994884910486, + "grad_norm": 5.322513595323884, + "learning_rate": 8.293581642406517e-06, + "loss": 0.4073, + "step": 212 + }, + { + "epoch": 0.5447570332480819, + "grad_norm": 7.46575629890418, + "learning_rate": 8.278441887221338e-06, + "loss": 0.6626, + "step": 213 + }, + { + "epoch": 0.5473145780051151, + "grad_norm": 6.052661000824651, + "learning_rate": 8.263249220161957e-06, + "loss": 0.5068, + "step": 214 + }, + { + "epoch": 0.5498721227621484, + "grad_norm": 6.708569886061961, + "learning_rate": 8.248003886427927e-06, + "loss": 0.4966, + "step": 215 + }, + { + "epoch": 0.5524296675191815, + "grad_norm": 5.68545952396897, + "learning_rate": 8.232706132068806e-06, + "loss": 0.3861, + "step": 216 + }, + { + "epoch": 0.5549872122762148, + "grad_norm": 5.421452942064916, + "learning_rate": 8.217356203980187e-06, + "loss": 0.3885, + "step": 217 + }, + { + "epoch": 0.5575447570332481, + "grad_norm": 6.028892220556533, + "learning_rate": 8.201954349899712e-06, + "loss": 0.5848, + "step": 218 + }, + { + "epoch": 0.5601023017902813, + "grad_norm": 5.85865954700368, + "learning_rate": 8.186500818403076e-06, + "loss": 0.4014, + "step": 219 + }, + { + "epoch": 0.5626598465473146, + "grad_norm": 7.224336653003557, + "learning_rate": 8.17099585890001e-06, + "loss": 0.6191, + "step": 220 + }, + { + "epoch": 0.5652173913043478, + "grad_norm": 6.30772616724702, + "learning_rate": 8.155439721630265e-06, + "loss": 0.4756, + "step": 221 + }, + { + "epoch": 0.5677749360613811, + "grad_norm": 6.236289350658551, + "learning_rate": 8.139832657659557e-06, + "loss": 0.4964, + "step": 222 + }, + { + "epoch": 0.5703324808184144, + "grad_norm": 6.209791638515158, + "learning_rate": 8.124174918875532e-06, + "loss": 0.5958, + "step": 223 + }, + { + "epoch": 0.5728900255754475, + "grad_norm": 6.377139101398714, + "learning_rate": 8.108466757983695e-06, + "loss": 0.3906, + "step": 224 + }, + { + "epoch": 0.5754475703324808, + "grad_norm": 5.907378705805276, + "learning_rate": 8.092708428503324e-06, + "loss": 0.5376, + "step": 225 + }, + { + "epoch": 0.578005115089514, + "grad_norm": 6.743429798340147, + "learning_rate": 8.076900184763394e-06, + "loss": 0.4802, + "step": 226 + }, + { + "epoch": 0.5805626598465473, + "grad_norm": 5.376964104341389, + "learning_rate": 8.061042281898453e-06, + "loss": 0.4509, + "step": 227 + }, + { + "epoch": 0.5831202046035806, + "grad_norm": 6.511105613676549, + "learning_rate": 8.04513497584452e-06, + "loss": 0.4214, + "step": 228 + }, + { + "epoch": 0.5856777493606138, + "grad_norm": 6.852114152108356, + "learning_rate": 8.02917852333495e-06, + "loss": 0.6038, + "step": 229 + }, + { + "epoch": 0.5882352941176471, + "grad_norm": 5.572255565814968, + "learning_rate": 8.013173181896283e-06, + "loss": 0.5224, + "step": 230 + }, + { + "epoch": 0.5907928388746803, + "grad_norm": 7.0760399259778435, + "learning_rate": 7.9971192098441e-06, + "loss": 0.4502, + "step": 231 + }, + { + "epoch": 0.5933503836317136, + "grad_norm": 5.207560218712082, + "learning_rate": 7.981016866278843e-06, + "loss": 0.4027, + "step": 232 + }, + { + "epoch": 0.5959079283887468, + "grad_norm": 5.796635616276233, + "learning_rate": 7.964866411081645e-06, + "loss": 0.5675, + "step": 233 + }, + { + "epoch": 0.59846547314578, + "grad_norm": 5.708164631421739, + "learning_rate": 7.94866810491012e-06, + "loss": 0.4437, + "step": 234 + }, + { + "epoch": 0.6010230179028133, + "grad_norm": 6.436431203161013, + "learning_rate": 7.93242220919417e-06, + "loss": 0.5474, + "step": 235 + }, + { + "epoch": 0.6035805626598465, + "grad_norm": 5.574144310350591, + "learning_rate": 7.916128986131761e-06, + "loss": 0.5439, + "step": 236 + }, + { + "epoch": 0.6061381074168798, + "grad_norm": 4.8240456797819835, + "learning_rate": 7.899788698684687e-06, + "loss": 0.4686, + "step": 237 + }, + { + "epoch": 0.6086956521739131, + "grad_norm": 6.5914849870729055, + "learning_rate": 7.883401610574338e-06, + "loss": 0.5512, + "step": 238 + }, + { + "epoch": 0.6112531969309463, + "grad_norm": 4.739332602957458, + "learning_rate": 7.866967986277423e-06, + "loss": 0.4204, + "step": 239 + }, + { + "epoch": 0.6138107416879796, + "grad_norm": 6.116120073202256, + "learning_rate": 7.850488091021726e-06, + "loss": 0.5596, + "step": 240 + }, + { + "epoch": 0.6163682864450127, + "grad_norm": 6.299579832647148, + "learning_rate": 7.833962190781809e-06, + "loss": 0.5729, + "step": 241 + }, + { + "epoch": 0.618925831202046, + "grad_norm": 5.77832842987742, + "learning_rate": 7.817390552274721e-06, + "loss": 0.4062, + "step": 242 + }, + { + "epoch": 0.6214833759590793, + "grad_norm": 5.6928424134185365, + "learning_rate": 7.800773442955703e-06, + "loss": 0.562, + "step": 243 + }, + { + "epoch": 0.6240409207161125, + "grad_norm": 5.754032663780959, + "learning_rate": 7.784111131013858e-06, + "loss": 0.4763, + "step": 244 + }, + { + "epoch": 0.6265984654731458, + "grad_norm": 6.1254790208347, + "learning_rate": 7.767403885367832e-06, + "loss": 0.4931, + "step": 245 + }, + { + "epoch": 0.629156010230179, + "grad_norm": 6.088136454995643, + "learning_rate": 7.750651975661471e-06, + "loss": 0.5366, + "step": 246 + }, + { + "epoch": 0.6317135549872123, + "grad_norm": 6.525801581028963, + "learning_rate": 7.733855672259472e-06, + "loss": 0.5869, + "step": 247 + }, + { + "epoch": 0.6342710997442456, + "grad_norm": 5.142226783902718, + "learning_rate": 7.717015246243012e-06, + "loss": 0.4107, + "step": 248 + }, + { + "epoch": 0.6368286445012787, + "grad_norm": 5.884475685733821, + "learning_rate": 7.700130969405377e-06, + "loss": 0.5575, + "step": 249 + }, + { + "epoch": 0.639386189258312, + "grad_norm": 5.430956350007929, + "learning_rate": 7.683203114247587e-06, + "loss": 0.4316, + "step": 250 + }, + { + "epoch": 0.6419437340153452, + "grad_norm": 5.852470007876826, + "learning_rate": 7.66623195397397e-06, + "loss": 0.5228, + "step": 251 + }, + { + "epoch": 0.6445012787723785, + "grad_norm": 6.316931409524609, + "learning_rate": 7.649217762487786e-06, + "loss": 0.6069, + "step": 252 + }, + { + "epoch": 0.6470588235294118, + "grad_norm": 5.981156306158716, + "learning_rate": 7.63216081438678e-06, + "loss": 0.4525, + "step": 253 + }, + { + "epoch": 0.649616368286445, + "grad_norm": 6.193124213697377, + "learning_rate": 7.615061384958764e-06, + "loss": 0.6367, + "step": 254 + }, + { + "epoch": 0.6521739130434783, + "grad_norm": 6.145394633019291, + "learning_rate": 7.597919750177168e-06, + "loss": 0.5622, + "step": 255 + }, + { + "epoch": 0.6547314578005116, + "grad_norm": 6.076895232152138, + "learning_rate": 7.580736186696593e-06, + "loss": 0.5016, + "step": 256 + }, + { + "epoch": 0.6572890025575447, + "grad_norm": 5.562852949209647, + "learning_rate": 7.563510971848339e-06, + "loss": 0.5739, + "step": 257 + }, + { + "epoch": 0.659846547314578, + "grad_norm": 6.358261854476947, + "learning_rate": 7.546244383635929e-06, + "loss": 0.5783, + "step": 258 + }, + { + "epoch": 0.6624040920716112, + "grad_norm": 5.61873313563532, + "learning_rate": 7.528936700730627e-06, + "loss": 0.5671, + "step": 259 + }, + { + "epoch": 0.6649616368286445, + "grad_norm": 5.007471092297137, + "learning_rate": 7.5115882024669375e-06, + "loss": 0.4238, + "step": 260 + }, + { + "epoch": 0.6675191815856778, + "grad_norm": 5.02718344671977, + "learning_rate": 7.494199168838099e-06, + "loss": 0.431, + "step": 261 + }, + { + "epoch": 0.670076726342711, + "grad_norm": 5.822025275525143, + "learning_rate": 7.476769880491561e-06, + "loss": 0.555, + "step": 262 + }, + { + "epoch": 0.6726342710997443, + "grad_norm": 5.852672525450696, + "learning_rate": 7.459300618724462e-06, + "loss": 0.4537, + "step": 263 + }, + { + "epoch": 0.6751918158567775, + "grad_norm": 6.36830749484907, + "learning_rate": 7.44179166547908e-06, + "loss": 0.5466, + "step": 264 + }, + { + "epoch": 0.6777493606138107, + "grad_norm": 5.187641704740303, + "learning_rate": 7.42424330333829e-06, + "loss": 0.4966, + "step": 265 + }, + { + "epoch": 0.680306905370844, + "grad_norm": 5.862839321803861, + "learning_rate": 7.406655815520998e-06, + "loss": 0.4902, + "step": 266 + }, + { + "epoch": 0.6828644501278772, + "grad_norm": 6.529161660718858, + "learning_rate": 7.389029485877577e-06, + "loss": 0.493, + "step": 267 + }, + { + "epoch": 0.6854219948849105, + "grad_norm": 5.732050686572585, + "learning_rate": 7.371364598885276e-06, + "loss": 0.4744, + "step": 268 + }, + { + "epoch": 0.6879795396419437, + "grad_norm": 5.533174363200175, + "learning_rate": 7.353661439643638e-06, + "loss": 0.3833, + "step": 269 + }, + { + "epoch": 0.690537084398977, + "grad_norm": 5.768168615527615, + "learning_rate": 7.335920293869891e-06, + "loss": 0.423, + "step": 270 + }, + { + "epoch": 0.6930946291560103, + "grad_norm": 5.852266644103708, + "learning_rate": 7.318141447894344e-06, + "loss": 0.3371, + "step": 271 + }, + { + "epoch": 0.6956521739130435, + "grad_norm": 6.038563114564619, + "learning_rate": 7.300325188655762e-06, + "loss": 0.4891, + "step": 272 + }, + { + "epoch": 0.6982097186700768, + "grad_norm": 6.280803826327464, + "learning_rate": 7.28247180369673e-06, + "loss": 0.5385, + "step": 273 + }, + { + "epoch": 0.7007672634271099, + "grad_norm": 7.292365659382516, + "learning_rate": 7.264581581159024e-06, + "loss": 0.6148, + "step": 274 + }, + { + "epoch": 0.7033248081841432, + "grad_norm": 6.763240999324924, + "learning_rate": 7.246654809778951e-06, + "loss": 0.5272, + "step": 275 + }, + { + "epoch": 0.7058823529411765, + "grad_norm": 6.444401975777849, + "learning_rate": 7.2286917788826926e-06, + "loss": 0.4879, + "step": 276 + }, + { + "epoch": 0.7084398976982097, + "grad_norm": 5.037923525497081, + "learning_rate": 7.210692778381634e-06, + "loss": 0.5377, + "step": 277 + }, + { + "epoch": 0.710997442455243, + "grad_norm": 6.327806611970394, + "learning_rate": 7.192658098767686e-06, + "loss": 0.4654, + "step": 278 + }, + { + "epoch": 0.7135549872122762, + "grad_norm": 5.832786135763086, + "learning_rate": 7.174588031108598e-06, + "loss": 0.5921, + "step": 279 + }, + { + "epoch": 0.7161125319693095, + "grad_norm": 6.032098832742715, + "learning_rate": 7.1564828670432595e-06, + "loss": 0.5032, + "step": 280 + }, + { + "epoch": 0.7186700767263428, + "grad_norm": 7.528807903355475, + "learning_rate": 7.138342898776989e-06, + "loss": 0.5143, + "step": 281 + }, + { + "epoch": 0.7212276214833759, + "grad_norm": 5.662236290695636, + "learning_rate": 7.120168419076825e-06, + "loss": 0.5752, + "step": 282 + }, + { + "epoch": 0.7237851662404092, + "grad_norm": 5.922836277812778, + "learning_rate": 7.101959721266798e-06, + "loss": 0.5907, + "step": 283 + }, + { + "epoch": 0.7263427109974424, + "grad_norm": 6.258012070363337, + "learning_rate": 7.083717099223192e-06, + "loss": 0.5447, + "step": 284 + }, + { + "epoch": 0.7289002557544757, + "grad_norm": 5.261480296532744, + "learning_rate": 7.0654408473698084e-06, + "loss": 0.4521, + "step": 285 + }, + { + "epoch": 0.731457800511509, + "grad_norm": 5.918110722172615, + "learning_rate": 7.047131260673214e-06, + "loss": 0.4637, + "step": 286 + }, + { + "epoch": 0.7340153452685422, + "grad_norm": 5.741282290810403, + "learning_rate": 7.0287886346379755e-06, + "loss": 0.4131, + "step": 287 + }, + { + "epoch": 0.7365728900255755, + "grad_norm": 5.943919434881143, + "learning_rate": 7.010413265301888e-06, + "loss": 0.4712, + "step": 288 + }, + { + "epoch": 0.7391304347826086, + "grad_norm": 6.1059644383499885, + "learning_rate": 6.9920054492312086e-06, + "loss": 0.6022, + "step": 289 + }, + { + "epoch": 0.7416879795396419, + "grad_norm": 6.884474367848085, + "learning_rate": 6.97356548351586e-06, + "loss": 0.5212, + "step": 290 + }, + { + "epoch": 0.7442455242966752, + "grad_norm": 5.758493578440039, + "learning_rate": 6.9550936657646386e-06, + "loss": 0.507, + "step": 291 + }, + { + "epoch": 0.7468030690537084, + "grad_norm": 5.5122177192122415, + "learning_rate": 6.936590294100414e-06, + "loss": 0.4096, + "step": 292 + }, + { + "epoch": 0.7493606138107417, + "grad_norm": 6.529147733060143, + "learning_rate": 6.918055667155311e-06, + "loss": 0.4668, + "step": 293 + }, + { + "epoch": 0.7519181585677749, + "grad_norm": 5.580434227838566, + "learning_rate": 6.899490084065897e-06, + "loss": 0.4825, + "step": 294 + }, + { + "epoch": 0.7544757033248082, + "grad_norm": 6.141771398723171, + "learning_rate": 6.8808938444683505e-06, + "loss": 0.5189, + "step": 295 + }, + { + "epoch": 0.7570332480818415, + "grad_norm": 6.258445455734687, + "learning_rate": 6.862267248493624e-06, + "loss": 0.4217, + "step": 296 + }, + { + "epoch": 0.7595907928388747, + "grad_norm": 5.577447249480196, + "learning_rate": 6.843610596762606e-06, + "loss": 0.4574, + "step": 297 + }, + { + "epoch": 0.7621483375959079, + "grad_norm": 6.508949986966596, + "learning_rate": 6.824924190381257e-06, + "loss": 0.4512, + "step": 298 + }, + { + "epoch": 0.7647058823529411, + "grad_norm": 6.487835617711066, + "learning_rate": 6.806208330935766e-06, + "loss": 0.4817, + "step": 299 + }, + { + "epoch": 0.7672634271099744, + "grad_norm": 6.733508290939032, + "learning_rate": 6.7874633204876705e-06, + "loss": 0.4648, + "step": 300 + }, + { + "epoch": 0.7698209718670077, + "grad_norm": 6.6391089444926195, + "learning_rate": 6.768689461568987e-06, + "loss": 0.4959, + "step": 301 + }, + { + "epoch": 0.7723785166240409, + "grad_norm": 5.426573235588597, + "learning_rate": 6.7498870571773275e-06, + "loss": 0.4101, + "step": 302 + }, + { + "epoch": 0.7749360613810742, + "grad_norm": 5.27403325499086, + "learning_rate": 6.731056410771008e-06, + "loss": 0.4183, + "step": 303 + }, + { + "epoch": 0.7774936061381074, + "grad_norm": 5.602097858442588, + "learning_rate": 6.712197826264154e-06, + "loss": 0.4712, + "step": 304 + }, + { + "epoch": 0.7800511508951407, + "grad_norm": 5.602023911663575, + "learning_rate": 6.69331160802179e-06, + "loss": 0.376, + "step": 305 + }, + { + "epoch": 0.782608695652174, + "grad_norm": 6.057331404811353, + "learning_rate": 6.674398060854931e-06, + "loss": 0.3333, + "step": 306 + }, + { + "epoch": 0.7851662404092071, + "grad_norm": 5.453910778706793, + "learning_rate": 6.655457490015667e-06, + "loss": 0.5251, + "step": 307 + }, + { + "epoch": 0.7877237851662404, + "grad_norm": 6.209463244054028, + "learning_rate": 6.636490201192229e-06, + "loss": 0.5256, + "step": 308 + }, + { + "epoch": 0.7902813299232737, + "grad_norm": 5.6125006489249145, + "learning_rate": 6.617496500504056e-06, + "loss": 0.35, + "step": 309 + }, + { + "epoch": 0.7928388746803069, + "grad_norm": 6.414498410153366, + "learning_rate": 6.5984766944968636e-06, + "loss": 0.5181, + "step": 310 + }, + { + "epoch": 0.7953964194373402, + "grad_norm": 6.101747662704522, + "learning_rate": 6.579431090137681e-06, + "loss": 0.4106, + "step": 311 + }, + { + "epoch": 0.7979539641943734, + "grad_norm": 6.462777333488606, + "learning_rate": 6.560359994809916e-06, + "loss": 0.6125, + "step": 312 + }, + { + "epoch": 0.8005115089514067, + "grad_norm": 5.9920053305051875, + "learning_rate": 6.541263716308375e-06, + "loss": 0.4968, + "step": 313 + }, + { + "epoch": 0.80306905370844, + "grad_norm": 6.671005371719509, + "learning_rate": 6.522142562834307e-06, + "loss": 0.5637, + "step": 314 + }, + { + "epoch": 0.8056265984654731, + "grad_norm": 5.361336122168199, + "learning_rate": 6.502996842990431e-06, + "loss": 0.4208, + "step": 315 + }, + { + "epoch": 0.8081841432225064, + "grad_norm": 5.670064103939166, + "learning_rate": 6.483826865775941e-06, + "loss": 0.5278, + "step": 316 + }, + { + "epoch": 0.8107416879795396, + "grad_norm": 5.5103998057715105, + "learning_rate": 6.46463294058154e-06, + "loss": 0.4007, + "step": 317 + }, + { + "epoch": 0.8132992327365729, + "grad_norm": 5.659095784663181, + "learning_rate": 6.445415377184427e-06, + "loss": 0.4742, + "step": 318 + }, + { + "epoch": 0.8158567774936062, + "grad_norm": 6.30132561670194, + "learning_rate": 6.426174485743309e-06, + "loss": 0.4078, + "step": 319 + }, + { + "epoch": 0.8184143222506394, + "grad_norm": 5.643268096385628, + "learning_rate": 6.4069105767933944e-06, + "loss": 0.46, + "step": 320 + }, + { + "epoch": 0.8209718670076727, + "grad_norm": 7.788725418859061, + "learning_rate": 6.387623961241375e-06, + "loss": 0.6119, + "step": 321 + }, + { + "epoch": 0.8235294117647058, + "grad_norm": 5.927896020719375, + "learning_rate": 6.368314950360416e-06, + "loss": 0.5225, + "step": 322 + }, + { + "epoch": 0.8260869565217391, + "grad_norm": 5.296955151964955, + "learning_rate": 6.348983855785122e-06, + "loss": 0.3126, + "step": 323 + }, + { + "epoch": 0.8286445012787724, + "grad_norm": 5.095733276738074, + "learning_rate": 6.3296309895065215e-06, + "loss": 0.3639, + "step": 324 + }, + { + "epoch": 0.8312020460358056, + "grad_norm": 6.080988913298908, + "learning_rate": 6.310256663867019e-06, + "loss": 0.5063, + "step": 325 + }, + { + "epoch": 0.8337595907928389, + "grad_norm": 7.036245894709906, + "learning_rate": 6.290861191555359e-06, + "loss": 0.4578, + "step": 326 + }, + { + "epoch": 0.8363171355498721, + "grad_norm": 5.580633409599807, + "learning_rate": 6.271444885601583e-06, + "loss": 0.4639, + "step": 327 + }, + { + "epoch": 0.8388746803069054, + "grad_norm": 6.523213064272758, + "learning_rate": 6.252008059371968e-06, + "loss": 0.4699, + "step": 328 + }, + { + "epoch": 0.8414322250639387, + "grad_norm": 4.85798591447732, + "learning_rate": 6.2325510265639785e-06, + "loss": 0.2973, + "step": 329 + }, + { + "epoch": 0.8439897698209718, + "grad_norm": 6.805564012992218, + "learning_rate": 6.213074101201202e-06, + "loss": 0.4894, + "step": 330 + }, + { + "epoch": 0.8465473145780051, + "grad_norm": 6.052583253032932, + "learning_rate": 6.193577597628268e-06, + "loss": 0.5193, + "step": 331 + }, + { + "epoch": 0.8491048593350383, + "grad_norm": 6.230894069829904, + "learning_rate": 6.174061830505801e-06, + "loss": 0.5028, + "step": 332 + }, + { + "epoch": 0.8516624040920716, + "grad_norm": 7.201079594756455, + "learning_rate": 6.154527114805312e-06, + "loss": 0.618, + "step": 333 + }, + { + "epoch": 0.8542199488491049, + "grad_norm": 6.600993850416883, + "learning_rate": 6.1349737658041385e-06, + "loss": 0.5133, + "step": 334 + }, + { + "epoch": 0.8567774936061381, + "grad_norm": 5.724822321191247, + "learning_rate": 6.115402099080345e-06, + "loss": 0.4838, + "step": 335 + }, + { + "epoch": 0.8593350383631714, + "grad_norm": 5.613108225355487, + "learning_rate": 6.095812430507627e-06, + "loss": 0.3442, + "step": 336 + }, + { + "epoch": 0.8618925831202046, + "grad_norm": 5.965477408637899, + "learning_rate": 6.076205076250227e-06, + "loss": 0.5109, + "step": 337 + }, + { + "epoch": 0.8644501278772379, + "grad_norm": 5.9857901067552, + "learning_rate": 6.056580352757813e-06, + "loss": 0.4511, + "step": 338 + }, + { + "epoch": 0.8670076726342711, + "grad_norm": 6.174131332105638, + "learning_rate": 6.036938576760388e-06, + "loss": 0.4419, + "step": 339 + }, + { + "epoch": 0.8695652173913043, + "grad_norm": 5.075847963553367, + "learning_rate": 6.0172800652631706e-06, + "loss": 0.3777, + "step": 340 + }, + { + "epoch": 0.8721227621483376, + "grad_norm": 7.190594951132575, + "learning_rate": 5.997605135541472e-06, + "loss": 0.5106, + "step": 341 + }, + { + "epoch": 0.8746803069053708, + "grad_norm": 6.747196680979683, + "learning_rate": 5.977914105135594e-06, + "loss": 0.4762, + "step": 342 + }, + { + "epoch": 0.8772378516624041, + "grad_norm": 5.714370912906624, + "learning_rate": 5.9582072918456805e-06, + "loss": 0.362, + "step": 343 + }, + { + "epoch": 0.8797953964194374, + "grad_norm": 5.236876408531075, + "learning_rate": 5.938485013726612e-06, + "loss": 0.3947, + "step": 344 + }, + { + "epoch": 0.8823529411764706, + "grad_norm": 5.116217278468624, + "learning_rate": 5.918747589082853e-06, + "loss": 0.4747, + "step": 345 + }, + { + "epoch": 0.8849104859335039, + "grad_norm": 4.749747087957306, + "learning_rate": 5.898995336463326e-06, + "loss": 0.4274, + "step": 346 + }, + { + "epoch": 0.887468030690537, + "grad_norm": 5.230418676152823, + "learning_rate": 5.879228574656269e-06, + "loss": 0.3441, + "step": 347 + }, + { + "epoch": 0.8900255754475703, + "grad_norm": 6.303242756857959, + "learning_rate": 5.859447622684084e-06, + "loss": 0.5131, + "step": 348 + }, + { + "epoch": 0.8925831202046036, + "grad_norm": 5.4435933577635645, + "learning_rate": 5.839652799798197e-06, + "loss": 0.4243, + "step": 349 + }, + { + "epoch": 0.8951406649616368, + "grad_norm": 6.921408500128556, + "learning_rate": 5.819844425473899e-06, + "loss": 0.5549, + "step": 350 + }, + { + "epoch": 0.8976982097186701, + "grad_norm": 6.635487149449039, + "learning_rate": 5.800022819405194e-06, + "loss": 0.5061, + "step": 351 + }, + { + "epoch": 0.9002557544757033, + "grad_norm": 6.3209563326259515, + "learning_rate": 5.780188301499636e-06, + "loss": 0.5999, + "step": 352 + }, + { + "epoch": 0.9028132992327366, + "grad_norm": 6.077422261762329, + "learning_rate": 5.760341191873167e-06, + "loss": 0.5111, + "step": 353 + }, + { + "epoch": 0.9053708439897699, + "grad_norm": 5.867259188193717, + "learning_rate": 5.740481810844952e-06, + "loss": 0.4771, + "step": 354 + }, + { + "epoch": 0.907928388746803, + "grad_norm": 4.777893746653604, + "learning_rate": 5.720610478932211e-06, + "loss": 0.3242, + "step": 355 + }, + { + "epoch": 0.9104859335038363, + "grad_norm": 5.764419825204445, + "learning_rate": 5.700727516845038e-06, + "loss": 0.3306, + "step": 356 + }, + { + "epoch": 0.9130434782608695, + "grad_norm": 5.510318712985209, + "learning_rate": 5.680833245481234e-06, + "loss": 0.4642, + "step": 357 + }, + { + "epoch": 0.9156010230179028, + "grad_norm": 6.53182819796998, + "learning_rate": 5.660927985921122e-06, + "loss": 0.51, + "step": 358 + }, + { + "epoch": 0.9181585677749361, + "grad_norm": 6.086318246788371, + "learning_rate": 5.641012059422369e-06, + "loss": 0.5472, + "step": 359 + }, + { + "epoch": 0.9207161125319693, + "grad_norm": 5.544550425971534, + "learning_rate": 5.621085787414799e-06, + "loss": 0.4603, + "step": 360 + }, + { + "epoch": 0.9232736572890026, + "grad_norm": 6.1043161575819616, + "learning_rate": 5.601149491495206e-06, + "loss": 0.485, + "step": 361 + }, + { + "epoch": 0.9258312020460358, + "grad_norm": 5.827986489165051, + "learning_rate": 5.581203493422161e-06, + "loss": 0.5864, + "step": 362 + }, + { + "epoch": 0.928388746803069, + "grad_norm": 5.172049016763376, + "learning_rate": 5.561248115110822e-06, + "loss": 0.4517, + "step": 363 + }, + { + "epoch": 0.9309462915601023, + "grad_norm": 6.640832021653832, + "learning_rate": 5.541283678627742e-06, + "loss": 0.3703, + "step": 364 + }, + { + "epoch": 0.9335038363171355, + "grad_norm": 5.361432485515485, + "learning_rate": 5.521310506185661e-06, + "loss": 0.4262, + "step": 365 + }, + { + "epoch": 0.9360613810741688, + "grad_norm": 5.808037599792696, + "learning_rate": 5.501328920138314e-06, + "loss": 0.544, + "step": 366 + }, + { + "epoch": 0.9386189258312021, + "grad_norm": 6.0541662716251095, + "learning_rate": 5.481339242975227e-06, + "loss": 0.4024, + "step": 367 + }, + { + "epoch": 0.9411764705882353, + "grad_norm": 6.232047697753629, + "learning_rate": 5.46134179731651e-06, + "loss": 0.4862, + "step": 368 + }, + { + "epoch": 0.9437340153452686, + "grad_norm": 6.000932213910604, + "learning_rate": 5.441336905907653e-06, + "loss": 0.4635, + "step": 369 + }, + { + "epoch": 0.9462915601023018, + "grad_norm": 7.611546241496742, + "learning_rate": 5.421324891614312e-06, + "loss": 0.4135, + "step": 370 + }, + { + "epoch": 0.948849104859335, + "grad_norm": 5.362330490202002, + "learning_rate": 5.4013060774171055e-06, + "loss": 0.4506, + "step": 371 + }, + { + "epoch": 0.9514066496163683, + "grad_norm": 5.346120483245015, + "learning_rate": 5.3812807864063946e-06, + "loss": 0.4576, + "step": 372 + }, + { + "epoch": 0.9539641943734015, + "grad_norm": 6.100965821026688, + "learning_rate": 5.361249341777075e-06, + "loss": 0.5165, + "step": 373 + }, + { + "epoch": 0.9565217391304348, + "grad_norm": 4.676613084116823, + "learning_rate": 5.341212066823356e-06, + "loss": 0.4383, + "step": 374 + }, + { + "epoch": 0.959079283887468, + "grad_norm": 6.564693244827939, + "learning_rate": 5.321169284933543e-06, + "loss": 0.5044, + "step": 375 + }, + { + "epoch": 0.9616368286445013, + "grad_norm": 4.752286756978166, + "learning_rate": 5.3011213195848245e-06, + "loss": 0.5422, + "step": 376 + }, + { + "epoch": 0.9641943734015346, + "grad_norm": 6.298216680263071, + "learning_rate": 5.281068494338039e-06, + "loss": 0.3751, + "step": 377 + }, + { + "epoch": 0.9667519181585678, + "grad_norm": 5.041453892793781, + "learning_rate": 5.26101113283247e-06, + "loss": 0.3732, + "step": 378 + }, + { + "epoch": 0.969309462915601, + "grad_norm": 6.072922237394072, + "learning_rate": 5.240949558780605e-06, + "loss": 0.4873, + "step": 379 + }, + { + "epoch": 0.9718670076726342, + "grad_norm": 5.712375997105542, + "learning_rate": 5.220884095962924e-06, + "loss": 0.4877, + "step": 380 + }, + { + "epoch": 0.9744245524296675, + "grad_norm": 5.5897510763688585, + "learning_rate": 5.200815068222666e-06, + "loss": 0.386, + "step": 381 + }, + { + "epoch": 0.9769820971867008, + "grad_norm": 6.160592176666778, + "learning_rate": 5.1807427994606065e-06, + "loss": 0.369, + "step": 382 + }, + { + "epoch": 0.979539641943734, + "grad_norm": 5.482848722330923, + "learning_rate": 5.1606676136298305e-06, + "loss": 0.4618, + "step": 383 + }, + { + "epoch": 0.9820971867007673, + "grad_norm": 5.96411837712334, + "learning_rate": 5.140589834730503e-06, + "loss": 0.4286, + "step": 384 + }, + { + "epoch": 0.9846547314578005, + "grad_norm": 6.021924443213883, + "learning_rate": 5.120509786804635e-06, + "loss": 0.4545, + "step": 385 + }, + { + "epoch": 0.9872122762148338, + "grad_norm": 4.860619712046072, + "learning_rate": 5.100427793930862e-06, + "loss": 0.4847, + "step": 386 + }, + { + "epoch": 0.989769820971867, + "grad_norm": 6.16249371180202, + "learning_rate": 5.08034418021921e-06, + "loss": 0.4119, + "step": 387 + }, + { + "epoch": 0.9923273657289002, + "grad_norm": 5.190813395638479, + "learning_rate": 5.06025926980586e-06, + "loss": 0.3609, + "step": 388 + }, + { + "epoch": 0.9948849104859335, + "grad_norm": 5.3092155597782025, + "learning_rate": 5.040173386847926e-06, + "loss": 0.3655, + "step": 389 + }, + { + "epoch": 0.9974424552429667, + "grad_norm": 5.7321767766000935, + "learning_rate": 5.0200868555182155e-06, + "loss": 0.5465, + "step": 390 + }, + { + "epoch": 1.0, + "grad_norm": 5.719257228559791, + "learning_rate": 5e-06, + "loss": 0.4347, + "step": 391 + }, + { + "epoch": 1.0025575447570332, + "grad_norm": 4.2391486408050145, + "learning_rate": 4.979913144481785e-06, + "loss": 0.2179, + "step": 392 + }, + { + "epoch": 1.0051150895140666, + "grad_norm": 4.604106412801713, + "learning_rate": 4.959826613152074e-06, + "loss": 0.2675, + "step": 393 + }, + { + "epoch": 1.0076726342710998, + "grad_norm": 4.113700302207306, + "learning_rate": 4.939740730194141e-06, + "loss": 0.2586, + "step": 394 + }, + { + "epoch": 1.010230179028133, + "grad_norm": 4.894518417900452, + "learning_rate": 4.919655819780792e-06, + "loss": 0.278, + "step": 395 + }, + { + "epoch": 1.0127877237851663, + "grad_norm": 4.3903147870940815, + "learning_rate": 4.899572206069138e-06, + "loss": 0.2175, + "step": 396 + }, + { + "epoch": 1.0153452685421995, + "grad_norm": 4.395292367085451, + "learning_rate": 4.879490213195366e-06, + "loss": 0.2597, + "step": 397 + }, + { + "epoch": 1.0179028132992327, + "grad_norm": 4.399892144315872, + "learning_rate": 4.8594101652694996e-06, + "loss": 0.2806, + "step": 398 + }, + { + "epoch": 1.020460358056266, + "grad_norm": 5.345621237626395, + "learning_rate": 4.839332386370171e-06, + "loss": 0.2571, + "step": 399 + }, + { + "epoch": 1.0230179028132993, + "grad_norm": 4.533667138931473, + "learning_rate": 4.819257200539394e-06, + "loss": 0.2646, + "step": 400 + }, + { + "epoch": 1.0255754475703325, + "grad_norm": 5.037019368108468, + "learning_rate": 4.799184931777337e-06, + "loss": 0.1862, + "step": 401 + }, + { + "epoch": 1.0281329923273657, + "grad_norm": 5.628048303078724, + "learning_rate": 4.779115904037079e-06, + "loss": 0.239, + "step": 402 + }, + { + "epoch": 1.030690537084399, + "grad_norm": 6.657437275552853, + "learning_rate": 4.759050441219395e-06, + "loss": 0.2032, + "step": 403 + }, + { + "epoch": 1.0332480818414322, + "grad_norm": 7.066364647253674, + "learning_rate": 4.738988867167531e-06, + "loss": 0.2686, + "step": 404 + }, + { + "epoch": 1.0358056265984654, + "grad_norm": 6.242504066160711, + "learning_rate": 4.718931505661961e-06, + "loss": 0.2039, + "step": 405 + }, + { + "epoch": 1.0383631713554988, + "grad_norm": 7.053347520596888, + "learning_rate": 4.698878680415176e-06, + "loss": 0.2677, + "step": 406 + }, + { + "epoch": 1.040920716112532, + "grad_norm": 6.980982091790442, + "learning_rate": 4.678830715066458e-06, + "loss": 0.2986, + "step": 407 + }, + { + "epoch": 1.0434782608695652, + "grad_norm": 7.1587957747958715, + "learning_rate": 4.6587879331766465e-06, + "loss": 0.2742, + "step": 408 + }, + { + "epoch": 1.0460358056265984, + "grad_norm": 6.506842088570043, + "learning_rate": 4.638750658222927e-06, + "loss": 0.327, + "step": 409 + }, + { + "epoch": 1.0485933503836318, + "grad_norm": 5.343374587345889, + "learning_rate": 4.618719213593605e-06, + "loss": 0.2399, + "step": 410 + }, + { + "epoch": 1.051150895140665, + "grad_norm": 5.952604432226821, + "learning_rate": 4.598693922582896e-06, + "loss": 0.3362, + "step": 411 + }, + { + "epoch": 1.0537084398976981, + "grad_norm": 5.292017124731555, + "learning_rate": 4.5786751083856895e-06, + "loss": 0.2229, + "step": 412 + }, + { + "epoch": 1.0562659846547315, + "grad_norm": 5.7674633387114, + "learning_rate": 4.558663094092348e-06, + "loss": 0.2716, + "step": 413 + }, + { + "epoch": 1.0588235294117647, + "grad_norm": 4.654675739900228, + "learning_rate": 4.53865820268349e-06, + "loss": 0.2439, + "step": 414 + }, + { + "epoch": 1.061381074168798, + "grad_norm": 6.404992169914481, + "learning_rate": 4.518660757024774e-06, + "loss": 0.2713, + "step": 415 + }, + { + "epoch": 1.0639386189258313, + "grad_norm": 4.940754639892674, + "learning_rate": 4.498671079861686e-06, + "loss": 0.2225, + "step": 416 + }, + { + "epoch": 1.0664961636828645, + "grad_norm": 4.76891987704093, + "learning_rate": 4.478689493814341e-06, + "loss": 0.2676, + "step": 417 + }, + { + "epoch": 1.0690537084398977, + "grad_norm": 5.023886841031966, + "learning_rate": 4.4587163213722595e-06, + "loss": 0.2594, + "step": 418 + }, + { + "epoch": 1.0716112531969308, + "grad_norm": 4.184649476108494, + "learning_rate": 4.438751884889179e-06, + "loss": 0.1791, + "step": 419 + }, + { + "epoch": 1.0741687979539642, + "grad_norm": 5.029459043202698, + "learning_rate": 4.41879650657784e-06, + "loss": 0.2588, + "step": 420 + }, + { + "epoch": 1.0767263427109974, + "grad_norm": 4.795405784853634, + "learning_rate": 4.398850508504795e-06, + "loss": 0.2273, + "step": 421 + }, + { + "epoch": 1.0792838874680306, + "grad_norm": 4.137020254719259, + "learning_rate": 4.3789142125852015e-06, + "loss": 0.2682, + "step": 422 + }, + { + "epoch": 1.081841432225064, + "grad_norm": 5.222698810349911, + "learning_rate": 4.358987940577631e-06, + "loss": 0.2186, + "step": 423 + }, + { + "epoch": 1.0843989769820972, + "grad_norm": 5.743293194926402, + "learning_rate": 4.339072014078879e-06, + "loss": 0.1999, + "step": 424 + }, + { + "epoch": 1.0869565217391304, + "grad_norm": 5.052413496789277, + "learning_rate": 4.319166754518768e-06, + "loss": 0.2047, + "step": 425 + }, + { + "epoch": 1.0895140664961638, + "grad_norm": 6.345187477594701, + "learning_rate": 4.299272483154963e-06, + "loss": 0.314, + "step": 426 + }, + { + "epoch": 1.092071611253197, + "grad_norm": 5.440791188116355, + "learning_rate": 4.27938952106779e-06, + "loss": 0.1874, + "step": 427 + }, + { + "epoch": 1.0946291560102301, + "grad_norm": 4.59572793131487, + "learning_rate": 4.259518189155049e-06, + "loss": 0.2259, + "step": 428 + }, + { + "epoch": 1.0971867007672633, + "grad_norm": 7.1276467094843925, + "learning_rate": 4.2396588081268355e-06, + "loss": 0.2556, + "step": 429 + }, + { + "epoch": 1.0997442455242967, + "grad_norm": 6.0356348861027564, + "learning_rate": 4.219811698500365e-06, + "loss": 0.3379, + "step": 430 + }, + { + "epoch": 1.10230179028133, + "grad_norm": 5.364328953584096, + "learning_rate": 4.199977180594807e-06, + "loss": 0.1789, + "step": 431 + }, + { + "epoch": 1.104859335038363, + "grad_norm": 6.395865991308588, + "learning_rate": 4.1801555745261025e-06, + "loss": 0.3364, + "step": 432 + }, + { + "epoch": 1.1074168797953965, + "grad_norm": 5.307623822753343, + "learning_rate": 4.160347200201804e-06, + "loss": 0.3123, + "step": 433 + }, + { + "epoch": 1.1099744245524297, + "grad_norm": 5.8960456968886055, + "learning_rate": 4.140552377315918e-06, + "loss": 0.3115, + "step": 434 + }, + { + "epoch": 1.1125319693094629, + "grad_norm": 5.5625443910431915, + "learning_rate": 4.120771425343733e-06, + "loss": 0.2276, + "step": 435 + }, + { + "epoch": 1.1150895140664963, + "grad_norm": 5.922218273593377, + "learning_rate": 4.101004663536675e-06, + "loss": 0.2569, + "step": 436 + }, + { + "epoch": 1.1176470588235294, + "grad_norm": 5.263570887474491, + "learning_rate": 4.081252410917148e-06, + "loss": 0.2444, + "step": 437 + }, + { + "epoch": 1.1202046035805626, + "grad_norm": 6.181574590354267, + "learning_rate": 4.061514986273391e-06, + "loss": 0.251, + "step": 438 + }, + { + "epoch": 1.1227621483375958, + "grad_norm": 6.1580046071843935, + "learning_rate": 4.041792708154321e-06, + "loss": 0.2596, + "step": 439 + }, + { + "epoch": 1.1253196930946292, + "grad_norm": 5.458696870991704, + "learning_rate": 4.022085894864408e-06, + "loss": 0.2237, + "step": 440 + }, + { + "epoch": 1.1278772378516624, + "grad_norm": 4.70048293838936, + "learning_rate": 4.0023948644585294e-06, + "loss": 0.2476, + "step": 441 + }, + { + "epoch": 1.1304347826086956, + "grad_norm": 4.466773072234314, + "learning_rate": 3.982719934736832e-06, + "loss": 0.2, + "step": 442 + }, + { + "epoch": 1.132992327365729, + "grad_norm": 4.27013775097264, + "learning_rate": 3.963061423239612e-06, + "loss": 0.1978, + "step": 443 + }, + { + "epoch": 1.1355498721227621, + "grad_norm": 5.689677983167152, + "learning_rate": 3.943419647242189e-06, + "loss": 0.3034, + "step": 444 + }, + { + "epoch": 1.1381074168797953, + "grad_norm": 4.25975811296343, + "learning_rate": 3.923794923749775e-06, + "loss": 0.2187, + "step": 445 + }, + { + "epoch": 1.1406649616368287, + "grad_norm": 4.876551747060212, + "learning_rate": 3.904187569492373e-06, + "loss": 0.257, + "step": 446 + }, + { + "epoch": 1.143222506393862, + "grad_norm": 7.4635507464539765, + "learning_rate": 3.884597900919656e-06, + "loss": 0.2295, + "step": 447 + }, + { + "epoch": 1.145780051150895, + "grad_norm": 5.715365590596455, + "learning_rate": 3.865026234195863e-06, + "loss": 0.2771, + "step": 448 + }, + { + "epoch": 1.1483375959079285, + "grad_norm": 5.275576356708176, + "learning_rate": 3.8454728851946885e-06, + "loss": 0.1754, + "step": 449 + }, + { + "epoch": 1.1508951406649617, + "grad_norm": 5.062656319020689, + "learning_rate": 3.8259381694942e-06, + "loss": 0.2382, + "step": 450 + }, + { + "epoch": 1.1534526854219949, + "grad_norm": 5.798164920877668, + "learning_rate": 3.806422402371733e-06, + "loss": 0.288, + "step": 451 + }, + { + "epoch": 1.156010230179028, + "grad_norm": 5.272021205396943, + "learning_rate": 3.786925898798801e-06, + "loss": 0.2654, + "step": 452 + }, + { + "epoch": 1.1585677749360614, + "grad_norm": 4.739102069258541, + "learning_rate": 3.767448973436021e-06, + "loss": 0.1996, + "step": 453 + }, + { + "epoch": 1.1611253196930946, + "grad_norm": 5.933924764643274, + "learning_rate": 3.7479919406280334e-06, + "loss": 0.3014, + "step": 454 + }, + { + "epoch": 1.1636828644501278, + "grad_norm": 5.230108631129741, + "learning_rate": 3.728555114398419e-06, + "loss": 0.2306, + "step": 455 + }, + { + "epoch": 1.1662404092071612, + "grad_norm": 5.1311748950507905, + "learning_rate": 3.709138808444641e-06, + "loss": 0.2426, + "step": 456 + }, + { + "epoch": 1.1687979539641944, + "grad_norm": 4.8518342992156365, + "learning_rate": 3.689743336132982e-06, + "loss": 0.148, + "step": 457 + }, + { + "epoch": 1.1713554987212276, + "grad_norm": 4.141242492696908, + "learning_rate": 3.6703690104934806e-06, + "loss": 0.1969, + "step": 458 + }, + { + "epoch": 1.1739130434782608, + "grad_norm": 5.575504964227216, + "learning_rate": 3.6510161442148783e-06, + "loss": 0.2847, + "step": 459 + }, + { + "epoch": 1.1764705882352942, + "grad_norm": 5.205832366429269, + "learning_rate": 3.6316850496395863e-06, + "loss": 0.3003, + "step": 460 + }, + { + "epoch": 1.1790281329923273, + "grad_norm": 5.640735065432005, + "learning_rate": 3.6123760387586265e-06, + "loss": 0.2454, + "step": 461 + }, + { + "epoch": 1.1815856777493605, + "grad_norm": 5.129282050106555, + "learning_rate": 3.5930894232066072e-06, + "loss": 0.1335, + "step": 462 + }, + { + "epoch": 1.184143222506394, + "grad_norm": 5.640092025988574, + "learning_rate": 3.5738255142566912e-06, + "loss": 0.2181, + "step": 463 + }, + { + "epoch": 1.186700767263427, + "grad_norm": 5.386833673551597, + "learning_rate": 3.5545846228155743e-06, + "loss": 0.2176, + "step": 464 + }, + { + "epoch": 1.1892583120204603, + "grad_norm": 5.704105112383357, + "learning_rate": 3.5353670594184623e-06, + "loss": 0.2497, + "step": 465 + }, + { + "epoch": 1.1918158567774937, + "grad_norm": 5.93251011789455, + "learning_rate": 3.516173134224059e-06, + "loss": 0.3096, + "step": 466 + }, + { + "epoch": 1.1943734015345269, + "grad_norm": 5.135070353719214, + "learning_rate": 3.4970031570095707e-06, + "loss": 0.2201, + "step": 467 + }, + { + "epoch": 1.19693094629156, + "grad_norm": 5.865866790423223, + "learning_rate": 3.477857437165694e-06, + "loss": 0.2694, + "step": 468 + }, + { + "epoch": 1.1994884910485935, + "grad_norm": 6.268703077398429, + "learning_rate": 3.458736283691626e-06, + "loss": 0.2555, + "step": 469 + }, + { + "epoch": 1.2020460358056266, + "grad_norm": 6.588443786412173, + "learning_rate": 3.4396400051900846e-06, + "loss": 0.2546, + "step": 470 + }, + { + "epoch": 1.2046035805626598, + "grad_norm": 5.327149282563148, + "learning_rate": 3.4205689098623195e-06, + "loss": 0.1783, + "step": 471 + }, + { + "epoch": 1.207161125319693, + "grad_norm": 4.958795590477734, + "learning_rate": 3.401523305503139e-06, + "loss": 0.1755, + "step": 472 + }, + { + "epoch": 1.2097186700767264, + "grad_norm": 5.8814512868036735, + "learning_rate": 3.3825034994959445e-06, + "loss": 0.2751, + "step": 473 + }, + { + "epoch": 1.2122762148337596, + "grad_norm": 6.346180998183863, + "learning_rate": 3.3635097988077724e-06, + "loss": 0.2825, + "step": 474 + }, + { + "epoch": 1.2148337595907928, + "grad_norm": 4.824141112150008, + "learning_rate": 3.3445425099843343e-06, + "loss": 0.193, + "step": 475 + }, + { + "epoch": 1.2173913043478262, + "grad_norm": 5.212763506921242, + "learning_rate": 3.3256019391450696e-06, + "loss": 0.2539, + "step": 476 + }, + { + "epoch": 1.2199488491048593, + "grad_norm": 5.731843763443905, + "learning_rate": 3.3066883919782116e-06, + "loss": 0.2647, + "step": 477 + }, + { + "epoch": 1.2225063938618925, + "grad_norm": 5.069916401820976, + "learning_rate": 3.287802173735848e-06, + "loss": 0.2466, + "step": 478 + }, + { + "epoch": 1.2250639386189257, + "grad_norm": 5.351892499599417, + "learning_rate": 3.268943589228992e-06, + "loss": 0.2369, + "step": 479 + }, + { + "epoch": 1.227621483375959, + "grad_norm": 5.226527301587174, + "learning_rate": 3.250112942822673e-06, + "loss": 0.2063, + "step": 480 + }, + { + "epoch": 1.2301790281329923, + "grad_norm": 5.599367983952576, + "learning_rate": 3.231310538431015e-06, + "loss": 0.3108, + "step": 481 + }, + { + "epoch": 1.2327365728900257, + "grad_norm": 4.949232012109662, + "learning_rate": 3.212536679512332e-06, + "loss": 0.2778, + "step": 482 + }, + { + "epoch": 1.2352941176470589, + "grad_norm": 4.692407945943108, + "learning_rate": 3.1937916690642356e-06, + "loss": 0.2972, + "step": 483 + }, + { + "epoch": 1.237851662404092, + "grad_norm": 4.946087662063809, + "learning_rate": 3.1750758096187446e-06, + "loss": 0.2155, + "step": 484 + }, + { + "epoch": 1.2404092071611252, + "grad_norm": 4.8219004232287, + "learning_rate": 3.1563894032373977e-06, + "loss": 0.22, + "step": 485 + }, + { + "epoch": 1.2429667519181586, + "grad_norm": 5.621400027096903, + "learning_rate": 3.137732751506376e-06, + "loss": 0.2286, + "step": 486 + }, + { + "epoch": 1.2455242966751918, + "grad_norm": 5.479705903908969, + "learning_rate": 3.1191061555316503e-06, + "loss": 0.2534, + "step": 487 + }, + { + "epoch": 1.248081841432225, + "grad_norm": 5.794553272942662, + "learning_rate": 3.1005099159341044e-06, + "loss": 0.2618, + "step": 488 + }, + { + "epoch": 1.2506393861892584, + "grad_norm": 6.017425638082386, + "learning_rate": 3.08194433284469e-06, + "loss": 0.2843, + "step": 489 + }, + { + "epoch": 1.2531969309462916, + "grad_norm": 4.5326844807976805, + "learning_rate": 3.0634097058995877e-06, + "loss": 0.1928, + "step": 490 + }, + { + "epoch": 1.2557544757033248, + "grad_norm": 6.4463104104564115, + "learning_rate": 3.0449063342353635e-06, + "loss": 0.2478, + "step": 491 + }, + { + "epoch": 1.258312020460358, + "grad_norm": 4.5887061422293565, + "learning_rate": 3.0264345164841426e-06, + "loss": 0.204, + "step": 492 + }, + { + "epoch": 1.2608695652173914, + "grad_norm": 5.508160093203401, + "learning_rate": 3.007994550768793e-06, + "loss": 0.2159, + "step": 493 + }, + { + "epoch": 1.2634271099744245, + "grad_norm": 4.737111003577314, + "learning_rate": 2.989586734698113e-06, + "loss": 0.2093, + "step": 494 + }, + { + "epoch": 1.265984654731458, + "grad_norm": 4.947584431847608, + "learning_rate": 2.971211365362028e-06, + "loss": 0.1984, + "step": 495 + }, + { + "epoch": 1.2685421994884911, + "grad_norm": 4.800447464091687, + "learning_rate": 2.9528687393267865e-06, + "loss": 0.2396, + "step": 496 + }, + { + "epoch": 1.2710997442455243, + "grad_norm": 5.424287981614177, + "learning_rate": 2.934559152630192e-06, + "loss": 0.2752, + "step": 497 + }, + { + "epoch": 1.2736572890025575, + "grad_norm": 4.885472933623559, + "learning_rate": 2.9162829007768103e-06, + "loss": 0.2778, + "step": 498 + }, + { + "epoch": 1.2762148337595907, + "grad_norm": 5.051334757075423, + "learning_rate": 2.898040278733203e-06, + "loss": 0.2728, + "step": 499 + }, + { + "epoch": 1.278772378516624, + "grad_norm": 5.103121044913984, + "learning_rate": 2.879831580923176e-06, + "loss": 0.2276, + "step": 500 + }, + { + "epoch": 1.278772378516624, + "eval_loss": 0.4383206367492676, + "eval_runtime": 0.9493, + "eval_samples_per_second": 33.71, + "eval_steps_per_second": 8.428, + "step": 500 + }, + { + "epoch": 1.2813299232736572, + "grad_norm": 6.302964407184529, + "learning_rate": 2.8616571012230134e-06, + "loss": 0.324, + "step": 501 + }, + { + "epoch": 1.2838874680306906, + "grad_norm": 4.570682847412226, + "learning_rate": 2.843517132956742e-06, + "loss": 0.2638, + "step": 502 + }, + { + "epoch": 1.2864450127877238, + "grad_norm": 5.1557886110649065, + "learning_rate": 2.8254119688914017e-06, + "loss": 0.2901, + "step": 503 + }, + { + "epoch": 1.289002557544757, + "grad_norm": 4.59205694943921, + "learning_rate": 2.8073419012323154e-06, + "loss": 0.246, + "step": 504 + }, + { + "epoch": 1.2915601023017902, + "grad_norm": 5.220862830802117, + "learning_rate": 2.789307221618369e-06, + "loss": 0.2665, + "step": 505 + }, + { + "epoch": 1.2941176470588236, + "grad_norm": 5.6445669568957895, + "learning_rate": 2.771308221117309e-06, + "loss": 0.2565, + "step": 506 + }, + { + "epoch": 1.2966751918158568, + "grad_norm": 5.578642462958587, + "learning_rate": 2.7533451902210512e-06, + "loss": 0.2805, + "step": 507 + }, + { + "epoch": 1.29923273657289, + "grad_norm": 5.9960066898200814, + "learning_rate": 2.7354184188409773e-06, + "loss": 0.3228, + "step": 508 + }, + { + "epoch": 1.3017902813299234, + "grad_norm": 4.257760920803555, + "learning_rate": 2.71752819630327e-06, + "loss": 0.1833, + "step": 509 + }, + { + "epoch": 1.3043478260869565, + "grad_norm": 6.1683068408323605, + "learning_rate": 2.6996748113442397e-06, + "loss": 0.185, + "step": 510 + }, + { + "epoch": 1.3069053708439897, + "grad_norm": 5.042423224754131, + "learning_rate": 2.6818585521056573e-06, + "loss": 0.241, + "step": 511 + }, + { + "epoch": 1.309462915601023, + "grad_norm": 4.727233078424547, + "learning_rate": 2.66407970613011e-06, + "loss": 0.1559, + "step": 512 + }, + { + "epoch": 1.3120204603580563, + "grad_norm": 5.0241427635079505, + "learning_rate": 2.646338560356363e-06, + "loss": 0.1877, + "step": 513 + }, + { + "epoch": 1.3145780051150895, + "grad_norm": 6.126061319891909, + "learning_rate": 2.6286354011147252e-06, + "loss": 0.2004, + "step": 514 + }, + { + "epoch": 1.317135549872123, + "grad_norm": 5.24912663925928, + "learning_rate": 2.6109705141224255e-06, + "loss": 0.2364, + "step": 515 + }, + { + "epoch": 1.319693094629156, + "grad_norm": 6.034124083029447, + "learning_rate": 2.593344184479003e-06, + "loss": 0.2788, + "step": 516 + }, + { + "epoch": 1.3222506393861893, + "grad_norm": 6.608163697606231, + "learning_rate": 2.575756696661713e-06, + "loss": 0.2416, + "step": 517 + }, + { + "epoch": 1.3248081841432224, + "grad_norm": 4.758894846786754, + "learning_rate": 2.5582083345209217e-06, + "loss": 0.2124, + "step": 518 + }, + { + "epoch": 1.3273657289002558, + "grad_norm": 5.719504905312417, + "learning_rate": 2.540699381275539e-06, + "loss": 0.297, + "step": 519 + }, + { + "epoch": 1.329923273657289, + "grad_norm": 5.86724061218003, + "learning_rate": 2.5232301195084395e-06, + "loss": 0.3234, + "step": 520 + }, + { + "epoch": 1.3324808184143222, + "grad_norm": 5.7644650842566945, + "learning_rate": 2.5058008311619035e-06, + "loss": 0.2615, + "step": 521 + }, + { + "epoch": 1.3350383631713556, + "grad_norm": 5.745064713338971, + "learning_rate": 2.488411797533064e-06, + "loss": 0.2118, + "step": 522 + }, + { + "epoch": 1.3375959079283888, + "grad_norm": 4.576869292927031, + "learning_rate": 2.4710632992693737e-06, + "loss": 0.2278, + "step": 523 + }, + { + "epoch": 1.340153452685422, + "grad_norm": 5.037272698300459, + "learning_rate": 2.4537556163640726e-06, + "loss": 0.236, + "step": 524 + }, + { + "epoch": 1.3427109974424551, + "grad_norm": 5.421366472711992, + "learning_rate": 2.4364890281516633e-06, + "loss": 0.2577, + "step": 525 + }, + { + "epoch": 1.3452685421994885, + "grad_norm": 4.866078798724115, + "learning_rate": 2.4192638133034074e-06, + "loss": 0.2318, + "step": 526 + }, + { + "epoch": 1.3478260869565217, + "grad_norm": 4.7964708165794265, + "learning_rate": 2.4020802498228333e-06, + "loss": 0.2054, + "step": 527 + }, + { + "epoch": 1.350383631713555, + "grad_norm": 4.928572310449122, + "learning_rate": 2.384938615041238e-06, + "loss": 0.2681, + "step": 528 + }, + { + "epoch": 1.3529411764705883, + "grad_norm": 4.737745386088544, + "learning_rate": 2.3678391856132203e-06, + "loss": 0.23, + "step": 529 + }, + { + "epoch": 1.3554987212276215, + "grad_norm": 4.546792180996154, + "learning_rate": 2.350782237512215e-06, + "loss": 0.2158, + "step": 530 + }, + { + "epoch": 1.3580562659846547, + "grad_norm": 5.14316611376246, + "learning_rate": 2.3337680460260314e-06, + "loss": 0.196, + "step": 531 + }, + { + "epoch": 1.3606138107416879, + "grad_norm": 5.834646238531157, + "learning_rate": 2.316796885752415e-06, + "loss": 0.2587, + "step": 532 + }, + { + "epoch": 1.3631713554987213, + "grad_norm": 5.780876751105411, + "learning_rate": 2.299869030594622e-06, + "loss": 0.1942, + "step": 533 + }, + { + "epoch": 1.3657289002557544, + "grad_norm": 4.548997507254711, + "learning_rate": 2.2829847537569904e-06, + "loss": 0.176, + "step": 534 + }, + { + "epoch": 1.3682864450127878, + "grad_norm": 5.731302369481431, + "learning_rate": 2.266144327740531e-06, + "loss": 0.3019, + "step": 535 + }, + { + "epoch": 1.370843989769821, + "grad_norm": 5.073679478403773, + "learning_rate": 2.2493480243385298e-06, + "loss": 0.2065, + "step": 536 + }, + { + "epoch": 1.3734015345268542, + "grad_norm": 5.850299158046723, + "learning_rate": 2.2325961146321683e-06, + "loss": 0.2714, + "step": 537 + }, + { + "epoch": 1.3759590792838874, + "grad_norm": 5.7360121373777035, + "learning_rate": 2.2158888689861434e-06, + "loss": 0.2249, + "step": 538 + }, + { + "epoch": 1.3785166240409208, + "grad_norm": 4.828788116522103, + "learning_rate": 2.1992265570442974e-06, + "loss": 0.2304, + "step": 539 + }, + { + "epoch": 1.381074168797954, + "grad_norm": 5.160394779683274, + "learning_rate": 2.182609447725279e-06, + "loss": 0.2259, + "step": 540 + }, + { + "epoch": 1.3836317135549872, + "grad_norm": 5.290912122436337, + "learning_rate": 2.1660378092181935e-06, + "loss": 0.2104, + "step": 541 + }, + { + "epoch": 1.3861892583120206, + "grad_norm": 5.983793408233051, + "learning_rate": 2.149511908978275e-06, + "loss": 0.2372, + "step": 542 + }, + { + "epoch": 1.3887468030690537, + "grad_norm": 5.146327357469625, + "learning_rate": 2.1330320137225773e-06, + "loss": 0.2594, + "step": 543 + }, + { + "epoch": 1.391304347826087, + "grad_norm": 6.1200616170554465, + "learning_rate": 2.1165983894256647e-06, + "loss": 0.2858, + "step": 544 + }, + { + "epoch": 1.39386189258312, + "grad_norm": 5.18913029029246, + "learning_rate": 2.100211301315315e-06, + "loss": 0.1924, + "step": 545 + }, + { + "epoch": 1.3964194373401535, + "grad_norm": 5.568160243169466, + "learning_rate": 2.0838710138682412e-06, + "loss": 0.193, + "step": 546 + }, + { + "epoch": 1.3989769820971867, + "grad_norm": 4.998024067501082, + "learning_rate": 2.0675777908058307e-06, + "loss": 0.2753, + "step": 547 + }, + { + "epoch": 1.40153452685422, + "grad_norm": 5.644184048080268, + "learning_rate": 2.051331895089882e-06, + "loss": 0.2865, + "step": 548 + }, + { + "epoch": 1.4040920716112533, + "grad_norm": 5.40264585527916, + "learning_rate": 2.035133588918356e-06, + "loss": 0.183, + "step": 549 + }, + { + "epoch": 1.4066496163682864, + "grad_norm": 4.477185903868405, + "learning_rate": 2.0189831337211573e-06, + "loss": 0.1957, + "step": 550 + }, + { + "epoch": 1.4092071611253196, + "grad_norm": 5.199329102721851, + "learning_rate": 2.0028807901559027e-06, + "loss": 0.2067, + "step": 551 + }, + { + "epoch": 1.4117647058823528, + "grad_norm": 4.040859440248337, + "learning_rate": 1.9868268181037186e-06, + "loss": 0.188, + "step": 552 + }, + { + "epoch": 1.4143222506393862, + "grad_norm": 4.554963712348274, + "learning_rate": 1.970821476665051e-06, + "loss": 0.1872, + "step": 553 + }, + { + "epoch": 1.4168797953964194, + "grad_norm": 5.430125296226688, + "learning_rate": 1.9548650241554812e-06, + "loss": 0.2154, + "step": 554 + }, + { + "epoch": 1.4194373401534528, + "grad_norm": 5.328931412586536, + "learning_rate": 1.9389577181015496e-06, + "loss": 0.1755, + "step": 555 + }, + { + "epoch": 1.421994884910486, + "grad_norm": 4.249854605239616, + "learning_rate": 1.923099815236608e-06, + "loss": 0.1593, + "step": 556 + }, + { + "epoch": 1.4245524296675192, + "grad_norm": 3.907795738528054, + "learning_rate": 1.9072915714966761e-06, + "loss": 0.1251, + "step": 557 + }, + { + "epoch": 1.4271099744245523, + "grad_norm": 5.529382370073762, + "learning_rate": 1.8915332420163074e-06, + "loss": 0.1817, + "step": 558 + }, + { + "epoch": 1.4296675191815857, + "grad_norm": 5.035709320601147, + "learning_rate": 1.8758250811244682e-06, + "loss": 0.2079, + "step": 559 + }, + { + "epoch": 1.432225063938619, + "grad_norm": 4.581843899498873, + "learning_rate": 1.8601673423404449e-06, + "loss": 0.2037, + "step": 560 + }, + { + "epoch": 1.434782608695652, + "grad_norm": 6.5976247360019045, + "learning_rate": 1.8445602783697375e-06, + "loss": 0.2877, + "step": 561 + }, + { + "epoch": 1.4373401534526855, + "grad_norm": 4.959100730308061, + "learning_rate": 1.8290041410999893e-06, + "loss": 0.1957, + "step": 562 + }, + { + "epoch": 1.4398976982097187, + "grad_norm": 5.169599476739708, + "learning_rate": 1.8134991815969238e-06, + "loss": 0.2358, + "step": 563 + }, + { + "epoch": 1.4424552429667519, + "grad_norm": 5.894324327311631, + "learning_rate": 1.798045650100289e-06, + "loss": 0.224, + "step": 564 + }, + { + "epoch": 1.445012787723785, + "grad_norm": 5.802424880479901, + "learning_rate": 1.782643796019814e-06, + "loss": 0.2429, + "step": 565 + }, + { + "epoch": 1.4475703324808185, + "grad_norm": 4.922016237101965, + "learning_rate": 1.7672938679311957e-06, + "loss": 0.2266, + "step": 566 + }, + { + "epoch": 1.4501278772378516, + "grad_norm": 5.755816715889569, + "learning_rate": 1.7519961135720737e-06, + "loss": 0.2376, + "step": 567 + }, + { + "epoch": 1.452685421994885, + "grad_norm": 5.220912348487513, + "learning_rate": 1.736750779838044e-06, + "loss": 0.2216, + "step": 568 + }, + { + "epoch": 1.4552429667519182, + "grad_norm": 5.210521733834325, + "learning_rate": 1.7215581127786624e-06, + "loss": 0.2615, + "step": 569 + }, + { + "epoch": 1.4578005115089514, + "grad_norm": 5.082754952247903, + "learning_rate": 1.7064183575934856e-06, + "loss": 0.2395, + "step": 570 + }, + { + "epoch": 1.4603580562659846, + "grad_norm": 4.668118713570096, + "learning_rate": 1.6913317586281048e-06, + "loss": 0.1761, + "step": 571 + }, + { + "epoch": 1.4629156010230178, + "grad_norm": 5.582283863667842, + "learning_rate": 1.676298559370202e-06, + "loss": 0.2342, + "step": 572 + }, + { + "epoch": 1.4654731457800512, + "grad_norm": 5.88971284160264, + "learning_rate": 1.6613190024456293e-06, + "loss": 0.3086, + "step": 573 + }, + { + "epoch": 1.4680306905370843, + "grad_norm": 5.053648614417428, + "learning_rate": 1.6463933296144863e-06, + "loss": 0.2169, + "step": 574 + }, + { + "epoch": 1.4705882352941178, + "grad_norm": 5.3893999941494055, + "learning_rate": 1.6315217817672142e-06, + "loss": 0.2483, + "step": 575 + }, + { + "epoch": 1.473145780051151, + "grad_norm": 6.003001211927369, + "learning_rate": 1.6167045989207185e-06, + "loss": 0.2488, + "step": 576 + }, + { + "epoch": 1.4757033248081841, + "grad_norm": 5.8557134317975965, + "learning_rate": 1.6019420202144853e-06, + "loss": 0.2349, + "step": 577 + }, + { + "epoch": 1.4782608695652173, + "grad_norm": 5.263808302802599, + "learning_rate": 1.5872342839067305e-06, + "loss": 0.1857, + "step": 578 + }, + { + "epoch": 1.4808184143222507, + "grad_norm": 4.433242818261617, + "learning_rate": 1.5725816273705453e-06, + "loss": 0.1888, + "step": 579 + }, + { + "epoch": 1.4833759590792839, + "grad_norm": 4.993905414062928, + "learning_rate": 1.5579842870900746e-06, + "loss": 0.2154, + "step": 580 + }, + { + "epoch": 1.485933503836317, + "grad_norm": 4.381712629412114, + "learning_rate": 1.5434424986566938e-06, + "loss": 0.222, + "step": 581 + }, + { + "epoch": 1.4884910485933505, + "grad_norm": 4.174610676856918, + "learning_rate": 1.5289564967652033e-06, + "loss": 0.1991, + "step": 582 + }, + { + "epoch": 1.4910485933503836, + "grad_norm": 4.819672292661032, + "learning_rate": 1.5145265152100574e-06, + "loss": 0.2425, + "step": 583 + }, + { + "epoch": 1.4936061381074168, + "grad_norm": 5.297609781727035, + "learning_rate": 1.5001527868815702e-06, + "loss": 0.3006, + "step": 584 + }, + { + "epoch": 1.49616368286445, + "grad_norm": 5.79687615069299, + "learning_rate": 1.4858355437621663e-06, + "loss": 0.27, + "step": 585 + }, + { + "epoch": 1.4987212276214834, + "grad_norm": 4.762605653912011, + "learning_rate": 1.4715750169226417e-06, + "loss": 0.2548, + "step": 586 + }, + { + "epoch": 1.5012787723785166, + "grad_norm": 5.7198048630752725, + "learning_rate": 1.457371436518424e-06, + "loss": 0.2594, + "step": 587 + }, + { + "epoch": 1.50383631713555, + "grad_norm": 4.802403477803473, + "learning_rate": 1.4432250317858675e-06, + "loss": 0.19, + "step": 588 + }, + { + "epoch": 1.5063938618925832, + "grad_norm": 6.635420914435053, + "learning_rate": 1.4291360310385455e-06, + "loss": 0.2405, + "step": 589 + }, + { + "epoch": 1.5089514066496164, + "grad_norm": 4.539889740139929, + "learning_rate": 1.4151046616635727e-06, + "loss": 0.1389, + "step": 590 + }, + { + "epoch": 1.5115089514066495, + "grad_norm": 5.38811895384659, + "learning_rate": 1.4011311501179287e-06, + "loss": 0.2662, + "step": 591 + }, + { + "epoch": 1.5140664961636827, + "grad_norm": 5.109717895825474, + "learning_rate": 1.3872157219248045e-06, + "loss": 0.2043, + "step": 592 + }, + { + "epoch": 1.5166240409207161, + "grad_norm": 5.0959884186276625, + "learning_rate": 1.373358601669973e-06, + "loss": 0.1719, + "step": 593 + }, + { + "epoch": 1.5191815856777495, + "grad_norm": 5.469942867350689, + "learning_rate": 1.3595600129981469e-06, + "loss": 0.1644, + "step": 594 + }, + { + "epoch": 1.5217391304347827, + "grad_norm": 3.8395720981267885, + "learning_rate": 1.3458201786093795e-06, + "loss": 0.1706, + "step": 595 + }, + { + "epoch": 1.5242966751918159, + "grad_norm": 5.518424328264906, + "learning_rate": 1.3321393202554739e-06, + "loss": 0.23, + "step": 596 + }, + { + "epoch": 1.526854219948849, + "grad_norm": 6.325750802451315, + "learning_rate": 1.3185176587363919e-06, + "loss": 0.3093, + "step": 597 + }, + { + "epoch": 1.5294117647058822, + "grad_norm": 5.306860678214338, + "learning_rate": 1.3049554138967052e-06, + "loss": 0.1669, + "step": 598 + }, + { + "epoch": 1.5319693094629157, + "grad_norm": 5.531828559483773, + "learning_rate": 1.2914528046220332e-06, + "loss": 0.2521, + "step": 599 + }, + { + "epoch": 1.5345268542199488, + "grad_norm": 4.928627869404515, + "learning_rate": 1.278010048835523e-06, + "loss": 0.1606, + "step": 600 + }, + { + "epoch": 1.5370843989769822, + "grad_norm": 4.736176509824429, + "learning_rate": 1.2646273634943195e-06, + "loss": 0.2382, + "step": 601 + }, + { + "epoch": 1.5396419437340154, + "grad_norm": 5.273813096512506, + "learning_rate": 1.2513049645860759e-06, + "loss": 0.1809, + "step": 602 + }, + { + "epoch": 1.5421994884910486, + "grad_norm": 5.824236440729651, + "learning_rate": 1.2380430671254618e-06, + "loss": 0.2194, + "step": 603 + }, + { + "epoch": 1.5447570332480818, + "grad_norm": 5.0892118395960395, + "learning_rate": 1.224841885150691e-06, + "loss": 0.2326, + "step": 604 + }, + { + "epoch": 1.547314578005115, + "grad_norm": 4.776070837603081, + "learning_rate": 1.2117016317200702e-06, + "loss": 0.202, + "step": 605 + }, + { + "epoch": 1.5498721227621484, + "grad_norm": 4.827682184305226, + "learning_rate": 1.1986225189085627e-06, + "loss": 0.2047, + "step": 606 + }, + { + "epoch": 1.5524296675191815, + "grad_norm": 6.354947289868109, + "learning_rate": 1.185604757804359e-06, + "loss": 0.2489, + "step": 607 + }, + { + "epoch": 1.554987212276215, + "grad_norm": 6.389932140184679, + "learning_rate": 1.172648558505477e-06, + "loss": 0.2039, + "step": 608 + }, + { + "epoch": 1.5575447570332481, + "grad_norm": 5.4707664838796815, + "learning_rate": 1.1597541301163655e-06, + "loss": 0.2242, + "step": 609 + }, + { + "epoch": 1.5601023017902813, + "grad_norm": 5.852672190013367, + "learning_rate": 1.1469216807445348e-06, + "loss": 0.1804, + "step": 610 + }, + { + "epoch": 1.5626598465473145, + "grad_norm": 6.351885487364482, + "learning_rate": 1.1341514174971907e-06, + "loss": 0.2128, + "step": 611 + }, + { + "epoch": 1.5652173913043477, + "grad_norm": 5.1770067311650605, + "learning_rate": 1.1214435464779006e-06, + "loss": 0.2993, + "step": 612 + }, + { + "epoch": 1.567774936061381, + "grad_norm": 5.447984746676171, + "learning_rate": 1.1087982727832613e-06, + "loss": 0.307, + "step": 613 + }, + { + "epoch": 1.5703324808184145, + "grad_norm": 5.182727825447981, + "learning_rate": 1.0962158004995893e-06, + "loss": 0.2687, + "step": 614 + }, + { + "epoch": 1.5728900255754477, + "grad_norm": 4.64809564738005, + "learning_rate": 1.083696332699628e-06, + "loss": 0.1572, + "step": 615 + }, + { + "epoch": 1.5754475703324808, + "grad_norm": 4.7325713746238876, + "learning_rate": 1.0712400714392723e-06, + "loss": 0.18, + "step": 616 + }, + { + "epoch": 1.578005115089514, + "grad_norm": 5.269623806320352, + "learning_rate": 1.058847217754303e-06, + "loss": 0.2037, + "step": 617 + }, + { + "epoch": 1.5805626598465472, + "grad_norm": 5.304757901803767, + "learning_rate": 1.0465179716571467e-06, + "loss": 0.1971, + "step": 618 + }, + { + "epoch": 1.5831202046035806, + "grad_norm": 6.728413029180757, + "learning_rate": 1.034252532133646e-06, + "loss": 0.2494, + "step": 619 + }, + { + "epoch": 1.5856777493606138, + "grad_norm": 3.4223539307887454, + "learning_rate": 1.0220510971398473e-06, + "loss": 0.1613, + "step": 620 + }, + { + "epoch": 1.5882352941176472, + "grad_norm": 5.305411837383943, + "learning_rate": 1.0099138635988026e-06, + "loss": 0.2349, + "step": 621 + }, + { + "epoch": 1.5907928388746804, + "grad_norm": 4.856914484373969, + "learning_rate": 9.978410273974015e-07, + "loss": 0.1723, + "step": 622 + }, + { + "epoch": 1.5933503836317136, + "grad_norm": 5.358684756870008, + "learning_rate": 9.858327833832004e-07, + "loss": 0.2593, + "step": 623 + }, + { + "epoch": 1.5959079283887467, + "grad_norm": 4.542731858430077, + "learning_rate": 9.738893253612808e-07, + "loss": 0.1939, + "step": 624 + }, + { + "epoch": 1.59846547314578, + "grad_norm": 4.5148886714925345, + "learning_rate": 9.620108460911181e-07, + "loss": 0.1901, + "step": 625 + }, + { + "epoch": 1.6010230179028133, + "grad_norm": 5.107745783747687, + "learning_rate": 9.50197537283481e-07, + "loss": 0.2062, + "step": 626 + }, + { + "epoch": 1.6035805626598465, + "grad_norm": 6.591147405424452, + "learning_rate": 9.384495895973227e-07, + "loss": 0.2293, + "step": 627 + }, + { + "epoch": 1.60613810741688, + "grad_norm": 5.607830445145212, + "learning_rate": 9.267671926367166e-07, + "loss": 0.2449, + "step": 628 + }, + { + "epoch": 1.608695652173913, + "grad_norm": 5.457362074950791, + "learning_rate": 9.151505349477901e-07, + "loss": 0.2638, + "step": 629 + }, + { + "epoch": 1.6112531969309463, + "grad_norm": 4.821051497667833, + "learning_rate": 9.035998040156801e-07, + "loss": 0.1902, + "step": 630 + }, + { + "epoch": 1.6138107416879794, + "grad_norm": 5.171122943753494, + "learning_rate": 8.921151862615091e-07, + "loss": 0.1422, + "step": 631 + }, + { + "epoch": 1.6163682864450126, + "grad_norm": 6.259146518225325, + "learning_rate": 8.806968670393801e-07, + "loss": 0.315, + "step": 632 + }, + { + "epoch": 1.618925831202046, + "grad_norm": 5.397528592846895, + "learning_rate": 8.693450306333818e-07, + "loss": 0.21, + "step": 633 + }, + { + "epoch": 1.6214833759590794, + "grad_norm": 5.180249332098257, + "learning_rate": 8.580598602546109e-07, + "loss": 0.2556, + "step": 634 + }, + { + "epoch": 1.6240409207161126, + "grad_norm": 4.562634677645485, + "learning_rate": 8.4684153803822e-07, + "loss": 0.2216, + "step": 635 + }, + { + "epoch": 1.6265984654731458, + "grad_norm": 5.390722528674729, + "learning_rate": 8.356902450404792e-07, + "loss": 0.233, + "step": 636 + }, + { + "epoch": 1.629156010230179, + "grad_norm": 4.775191692612564, + "learning_rate": 8.246061612358475e-07, + "loss": 0.2287, + "step": 637 + }, + { + "epoch": 1.6317135549872122, + "grad_norm": 4.852828279434425, + "learning_rate": 8.135894655140758e-07, + "loss": 0.2191, + "step": 638 + }, + { + "epoch": 1.6342710997442456, + "grad_norm": 5.704310246599364, + "learning_rate": 8.026403356773161e-07, + "loss": 0.2047, + "step": 639 + }, + { + "epoch": 1.6368286445012787, + "grad_norm": 4.793807479675851, + "learning_rate": 7.91758948437249e-07, + "loss": 0.1618, + "step": 640 + }, + { + "epoch": 1.6393861892583121, + "grad_norm": 4.754502329271982, + "learning_rate": 7.809454794122346e-07, + "loss": 0.2781, + "step": 641 + }, + { + "epoch": 1.6419437340153453, + "grad_norm": 7.018986504016533, + "learning_rate": 7.702001031244816e-07, + "loss": 0.2729, + "step": 642 + }, + { + "epoch": 1.6445012787723785, + "grad_norm": 4.41916704046766, + "learning_rate": 7.595229929972253e-07, + "loss": 0.2092, + "step": 643 + }, + { + "epoch": 1.6470588235294117, + "grad_norm": 5.336932050022912, + "learning_rate": 7.489143213519301e-07, + "loss": 0.2267, + "step": 644 + }, + { + "epoch": 1.6496163682864449, + "grad_norm": 4.654218849496965, + "learning_rate": 7.383742594055077e-07, + "loss": 0.2136, + "step": 645 + }, + { + "epoch": 1.6521739130434783, + "grad_norm": 5.1623224835718515, + "learning_rate": 7.279029772675572e-07, + "loss": 0.221, + "step": 646 + }, + { + "epoch": 1.6547314578005117, + "grad_norm": 5.686897560893601, + "learning_rate": 7.17500643937617e-07, + "loss": 0.243, + "step": 647 + }, + { + "epoch": 1.6572890025575449, + "grad_norm": 4.738802336926236, + "learning_rate": 7.071674273024353e-07, + "loss": 0.2059, + "step": 648 + }, + { + "epoch": 1.659846547314578, + "grad_norm": 6.305944976355385, + "learning_rate": 6.969034941332664e-07, + "loss": 0.2147, + "step": 649 + }, + { + "epoch": 1.6624040920716112, + "grad_norm": 5.190380929928794, + "learning_rate": 6.86709010083172e-07, + "loss": 0.1909, + "step": 650 + }, + { + "epoch": 1.6649616368286444, + "grad_norm": 5.659990773064172, + "learning_rate": 6.765841396843514e-07, + "loss": 0.1913, + "step": 651 + }, + { + "epoch": 1.6675191815856778, + "grad_norm": 4.841853616494702, + "learning_rate": 6.665290463454882e-07, + "loss": 0.201, + "step": 652 + }, + { + "epoch": 1.670076726342711, + "grad_norm": 6.0589020041416575, + "learning_rate": 6.565438923491102e-07, + "loss": 0.2908, + "step": 653 + }, + { + "epoch": 1.6726342710997444, + "grad_norm": 4.716079545501844, + "learning_rate": 6.466288388489689e-07, + "loss": 0.2169, + "step": 654 + }, + { + "epoch": 1.6751918158567776, + "grad_norm": 4.268196527173102, + "learning_rate": 6.367840458674401e-07, + "loss": 0.202, + "step": 655 + }, + { + "epoch": 1.6777493606138107, + "grad_norm": 5.8944147568821625, + "learning_rate": 6.270096722929442e-07, + "loss": 0.2586, + "step": 656 + }, + { + "epoch": 1.680306905370844, + "grad_norm": 5.564876452450793, + "learning_rate": 6.173058758773775e-07, + "loss": 0.2391, + "step": 657 + }, + { + "epoch": 1.682864450127877, + "grad_norm": 5.003498689191179, + "learning_rate": 6.076728132335669e-07, + "loss": 0.1756, + "step": 658 + }, + { + "epoch": 1.6854219948849105, + "grad_norm": 4.993122576064924, + "learning_rate": 5.981106398327463e-07, + "loss": 0.1984, + "step": 659 + }, + { + "epoch": 1.6879795396419437, + "grad_norm": 5.118160492207756, + "learning_rate": 5.886195100020408e-07, + "loss": 0.2879, + "step": 660 + }, + { + "epoch": 1.690537084398977, + "grad_norm": 5.396969385626712, + "learning_rate": 5.7919957692198e-07, + "loss": 0.1767, + "step": 661 + }, + { + "epoch": 1.6930946291560103, + "grad_norm": 4.709841482974692, + "learning_rate": 5.698509926240275e-07, + "loss": 0.2265, + "step": 662 + }, + { + "epoch": 1.6956521739130435, + "grad_norm": 5.74378925311425, + "learning_rate": 5.60573907988124e-07, + "loss": 0.2585, + "step": 663 + }, + { + "epoch": 1.6982097186700766, + "grad_norm": 5.340843111211026, + "learning_rate": 5.513684727402529e-07, + "loss": 0.1713, + "step": 664 + }, + { + "epoch": 1.7007672634271098, + "grad_norm": 5.095292121289202, + "learning_rate": 5.422348354500217e-07, + "loss": 0.1969, + "step": 665 + }, + { + "epoch": 1.7033248081841432, + "grad_norm": 6.38491199555891, + "learning_rate": 5.331731435282705e-07, + "loss": 0.1764, + "step": 666 + }, + { + "epoch": 1.7058823529411766, + "grad_norm": 6.33773326776292, + "learning_rate": 5.241835432246888e-07, + "loss": 0.2176, + "step": 667 + }, + { + "epoch": 1.7084398976982098, + "grad_norm": 4.747290198021267, + "learning_rate": 5.152661796254505e-07, + "loss": 0.2194, + "step": 668 + }, + { + "epoch": 1.710997442455243, + "grad_norm": 5.180305167036278, + "learning_rate": 5.064211966508837e-07, + "loss": 0.1838, + "step": 669 + }, + { + "epoch": 1.7135549872122762, + "grad_norm": 5.158569886057061, + "learning_rate": 4.976487370531352e-07, + "loss": 0.1714, + "step": 670 + }, + { + "epoch": 1.7161125319693094, + "grad_norm": 6.7699503709640405, + "learning_rate": 4.88948942413876e-07, + "loss": 0.2794, + "step": 671 + }, + { + "epoch": 1.7186700767263428, + "grad_norm": 4.330269645663795, + "learning_rate": 4.803219531420128e-07, + "loss": 0.1567, + "step": 672 + }, + { + "epoch": 1.721227621483376, + "grad_norm": 6.071163454166977, + "learning_rate": 4.717679084714222e-07, + "loss": 0.2268, + "step": 673 + }, + { + "epoch": 1.7237851662404093, + "grad_norm": 4.455492690019572, + "learning_rate": 4.6328694645870254e-07, + "loss": 0.2092, + "step": 674 + }, + { + "epoch": 1.7263427109974425, + "grad_norm": 5.5297942207932165, + "learning_rate": 4.5487920398094465e-07, + "loss": 0.2037, + "step": 675 + }, + { + "epoch": 1.7289002557544757, + "grad_norm": 4.8818412705137115, + "learning_rate": 4.46544816733529e-07, + "loss": 0.2475, + "step": 676 + }, + { + "epoch": 1.7314578005115089, + "grad_norm": 5.627569483357689, + "learning_rate": 4.382839192279303e-07, + "loss": 0.2585, + "step": 677 + }, + { + "epoch": 1.734015345268542, + "grad_norm": 5.1192641037653965, + "learning_rate": 4.3009664478954384e-07, + "loss": 0.2449, + "step": 678 + }, + { + "epoch": 1.7365728900255755, + "grad_norm": 5.286870696275029, + "learning_rate": 4.219831255555423e-07, + "loss": 0.2004, + "step": 679 + }, + { + "epoch": 1.7391304347826086, + "grad_norm": 5.5327277136832675, + "learning_rate": 4.139434924727359e-07, + "loss": 0.1921, + "step": 680 + }, + { + "epoch": 1.741687979539642, + "grad_norm": 4.552608688952828, + "learning_rate": 4.059778752954607e-07, + "loss": 0.1432, + "step": 681 + }, + { + "epoch": 1.7442455242966752, + "grad_norm": 5.216152442361206, + "learning_rate": 3.9808640258348686e-07, + "loss": 0.1754, + "step": 682 + }, + { + "epoch": 1.7468030690537084, + "grad_norm": 5.756818223369158, + "learning_rate": 3.9026920169994374e-07, + "loss": 0.21, + "step": 683 + }, + { + "epoch": 1.7493606138107416, + "grad_norm": 5.29491810051506, + "learning_rate": 3.825263988092587e-07, + "loss": 0.2228, + "step": 684 + }, + { + "epoch": 1.7519181585677748, + "grad_norm": 5.164127220528273, + "learning_rate": 3.7485811887512714e-07, + "loss": 0.2144, + "step": 685 + }, + { + "epoch": 1.7544757033248082, + "grad_norm": 5.314203439099788, + "learning_rate": 3.672644856584928e-07, + "loss": 0.2911, + "step": 686 + }, + { + "epoch": 1.7570332480818416, + "grad_norm": 4.972659541709094, + "learning_rate": 3.597456217155526e-07, + "loss": 0.1519, + "step": 687 + }, + { + "epoch": 1.7595907928388748, + "grad_norm": 5.199722512215739, + "learning_rate": 3.523016483957742e-07, + "loss": 0.2739, + "step": 688 + }, + { + "epoch": 1.762148337595908, + "grad_norm": 5.21999031799005, + "learning_rate": 3.4493268583994434e-07, + "loss": 0.201, + "step": 689 + }, + { + "epoch": 1.7647058823529411, + "grad_norm": 4.364821712930802, + "learning_rate": 3.3763885297822153e-07, + "loss": 0.1881, + "step": 690 + }, + { + "epoch": 1.7672634271099743, + "grad_norm": 5.202401016558692, + "learning_rate": 3.3042026752822254e-07, + "loss": 0.2125, + "step": 691 + }, + { + "epoch": 1.7698209718670077, + "grad_norm": 5.684342025889591, + "learning_rate": 3.2327704599312283e-07, + "loss": 0.2636, + "step": 692 + }, + { + "epoch": 1.772378516624041, + "grad_norm": 5.59338748722469, + "learning_rate": 3.16209303659773e-07, + "loss": 0.2489, + "step": 693 + }, + { + "epoch": 1.7749360613810743, + "grad_norm": 4.23116707170702, + "learning_rate": 3.0921715459683753e-07, + "loss": 0.153, + "step": 694 + }, + { + "epoch": 1.7774936061381075, + "grad_norm": 5.558061912478462, + "learning_rate": 3.0230071165295804e-07, + "loss": 0.2291, + "step": 695 + }, + { + "epoch": 1.7800511508951407, + "grad_norm": 5.197951332859319, + "learning_rate": 2.95460086454929e-07, + "loss": 0.2267, + "step": 696 + }, + { + "epoch": 1.7826086956521738, + "grad_norm": 5.668569655261014, + "learning_rate": 2.88695389405898e-07, + "loss": 0.2191, + "step": 697 + }, + { + "epoch": 1.785166240409207, + "grad_norm": 4.00218617616878, + "learning_rate": 2.820067296835799e-07, + "loss": 0.1415, + "step": 698 + }, + { + "epoch": 1.7877237851662404, + "grad_norm": 5.580094964915473, + "learning_rate": 2.753942152385014e-07, + "loss": 0.2282, + "step": 699 + }, + { + "epoch": 1.7902813299232738, + "grad_norm": 5.157143860228644, + "learning_rate": 2.688579527922514e-07, + "loss": 0.228, + "step": 700 + }, + { + "epoch": 1.792838874680307, + "grad_norm": 6.375062738221611, + "learning_rate": 2.6239804783576294e-07, + "loss": 0.227, + "step": 701 + }, + { + "epoch": 1.7953964194373402, + "grad_norm": 5.7437796904243115, + "learning_rate": 2.560146046276135e-07, + "loss": 0.2823, + "step": 702 + }, + { + "epoch": 1.7979539641943734, + "grad_norm": 5.624690347371044, + "learning_rate": 2.4970772619233475e-07, + "loss": 0.2042, + "step": 703 + }, + { + "epoch": 1.8005115089514065, + "grad_norm": 6.498262506148002, + "learning_rate": 2.4347751431875453e-07, + "loss": 0.2598, + "step": 704 + }, + { + "epoch": 1.80306905370844, + "grad_norm": 6.186981848018963, + "learning_rate": 2.373240695583534e-07, + "loss": 0.2505, + "step": 705 + }, + { + "epoch": 1.8056265984654731, + "grad_norm": 4.425992548953008, + "learning_rate": 2.3124749122364286e-07, + "loss": 0.2145, + "step": 706 + }, + { + "epoch": 1.8081841432225065, + "grad_norm": 4.690718811893992, + "learning_rate": 2.2524787738656073e-07, + "loss": 0.2106, + "step": 707 + }, + { + "epoch": 1.8107416879795397, + "grad_norm": 6.156573679807675, + "learning_rate": 2.1932532487688784e-07, + "loss": 0.222, + "step": 708 + }, + { + "epoch": 1.813299232736573, + "grad_norm": 5.586951769770138, + "learning_rate": 2.1347992928068884e-07, + "loss": 0.2469, + "step": 709 + }, + { + "epoch": 1.815856777493606, + "grad_norm": 5.817151417919622, + "learning_rate": 2.0771178493876387e-07, + "loss": 0.2514, + "step": 710 + }, + { + "epoch": 1.8184143222506393, + "grad_norm": 5.920190459702125, + "learning_rate": 2.0202098494513157e-07, + "loss": 0.2034, + "step": 711 + }, + { + "epoch": 1.8209718670076727, + "grad_norm": 5.8230410453933015, + "learning_rate": 1.964076211455246e-07, + "loss": 0.2364, + "step": 712 + }, + { + "epoch": 1.8235294117647058, + "grad_norm": 5.7392724174161405, + "learning_rate": 1.908717841359048e-07, + "loss": 0.2135, + "step": 713 + }, + { + "epoch": 1.8260869565217392, + "grad_norm": 5.775722433349359, + "learning_rate": 1.8541356326100436e-07, + "loss": 0.2462, + "step": 714 + }, + { + "epoch": 1.8286445012787724, + "grad_norm": 5.591670023546011, + "learning_rate": 1.800330466128808e-07, + "loss": 0.228, + "step": 715 + }, + { + "epoch": 1.8312020460358056, + "grad_norm": 4.711360255327199, + "learning_rate": 1.7473032102949983e-07, + "loss": 0.2075, + "step": 716 + }, + { + "epoch": 1.8337595907928388, + "grad_norm": 5.929377308741077, + "learning_rate": 1.695054720933309e-07, + "loss": 0.257, + "step": 717 + }, + { + "epoch": 1.836317135549872, + "grad_norm": 5.279542039345601, + "learning_rate": 1.6435858412996275e-07, + "loss": 0.1529, + "step": 718 + }, + { + "epoch": 1.8388746803069054, + "grad_norm": 5.1397676431266595, + "learning_rate": 1.5928974020674947e-07, + "loss": 0.1491, + "step": 719 + }, + { + "epoch": 1.8414322250639388, + "grad_norm": 5.512326623165319, + "learning_rate": 1.542990221314644e-07, + "loss": 0.1774, + "step": 720 + }, + { + "epoch": 1.843989769820972, + "grad_norm": 5.66627970312488, + "learning_rate": 1.4938651045098174e-07, + "loss": 0.2278, + "step": 721 + }, + { + "epoch": 1.8465473145780051, + "grad_norm": 4.424031839884274, + "learning_rate": 1.445522844499775e-07, + "loss": 0.1713, + "step": 722 + }, + { + "epoch": 1.8491048593350383, + "grad_norm": 6.625716850892438, + "learning_rate": 1.3979642214964728e-07, + "loss": 0.254, + "step": 723 + }, + { + "epoch": 1.8516624040920715, + "grad_norm": 5.181446281414914, + "learning_rate": 1.3511900030644954e-07, + "loss": 0.1753, + "step": 724 + }, + { + "epoch": 1.854219948849105, + "grad_norm": 6.694348870428919, + "learning_rate": 1.3052009441086533e-07, + "loss": 0.2535, + "step": 725 + }, + { + "epoch": 1.856777493606138, + "grad_norm": 6.09347774518519, + "learning_rate": 1.2599977868618052e-07, + "loss": 0.2617, + "step": 726 + }, + { + "epoch": 1.8593350383631715, + "grad_norm": 6.126918278593983, + "learning_rate": 1.215581260872889e-07, + "loss": 0.1734, + "step": 727 + }, + { + "epoch": 1.8618925831202047, + "grad_norm": 4.9394061667025175, + "learning_rate": 1.1719520829951203e-07, + "loss": 0.1727, + "step": 728 + }, + { + "epoch": 1.8644501278772379, + "grad_norm": 5.679604325861919, + "learning_rate": 1.1291109573744574e-07, + "loss": 0.2195, + "step": 729 + }, + { + "epoch": 1.867007672634271, + "grad_norm": 5.5914554870990525, + "learning_rate": 1.087058575438199e-07, + "loss": 0.2528, + "step": 730 + }, + { + "epoch": 1.8695652173913042, + "grad_norm": 5.877663743949046, + "learning_rate": 1.0457956158838545e-07, + "loss": 0.2491, + "step": 731 + }, + { + "epoch": 1.8721227621483376, + "grad_norm": 3.8851223721617663, + "learning_rate": 1.0053227446681912e-07, + "loss": 0.1317, + "step": 732 + }, + { + "epoch": 1.8746803069053708, + "grad_norm": 5.412598503950858, + "learning_rate": 9.656406149964548e-08, + "loss": 0.2314, + "step": 733 + }, + { + "epoch": 1.8772378516624042, + "grad_norm": 4.937142526302332, + "learning_rate": 9.267498673118547e-08, + "loss": 0.1964, + "step": 734 + }, + { + "epoch": 1.8797953964194374, + "grad_norm": 5.6093925489434735, + "learning_rate": 8.886511292852395e-08, + "loss": 0.2169, + "step": 735 + }, + { + "epoch": 1.8823529411764706, + "grad_norm": 5.544589584646933, + "learning_rate": 8.513450158049109e-08, + "loss": 0.2262, + "step": 736 + }, + { + "epoch": 1.8849104859335037, + "grad_norm": 4.522441136690872, + "learning_rate": 8.148321289667749e-08, + "loss": 0.1836, + "step": 737 + }, + { + "epoch": 1.887468030690537, + "grad_norm": 4.735747110712485, + "learning_rate": 7.791130580645623e-08, + "loss": 0.2037, + "step": 738 + }, + { + "epoch": 1.8900255754475703, + "grad_norm": 5.424105506041224, + "learning_rate": 7.441883795803462e-08, + "loss": 0.181, + "step": 739 + }, + { + "epoch": 1.8925831202046037, + "grad_norm": 5.57291250707421, + "learning_rate": 7.100586571752444e-08, + "loss": 0.2009, + "step": 740 + }, + { + "epoch": 1.895140664961637, + "grad_norm": 4.723852414529119, + "learning_rate": 6.767244416802988e-08, + "loss": 0.2135, + "step": 741 + }, + { + "epoch": 1.89769820971867, + "grad_norm": 4.625168893146648, + "learning_rate": 6.441862710876102e-08, + "loss": 0.182, + "step": 742 + }, + { + "epoch": 1.9002557544757033, + "grad_norm": 5.742248230826967, + "learning_rate": 6.124446705416343e-08, + "loss": 0.2256, + "step": 743 + }, + { + "epoch": 1.9028132992327365, + "grad_norm": 5.446320051746138, + "learning_rate": 5.815001523307162e-08, + "loss": 0.1844, + "step": 744 + }, + { + "epoch": 1.9053708439897699, + "grad_norm": 5.776856594363444, + "learning_rate": 5.513532158788193e-08, + "loss": 0.2793, + "step": 745 + }, + { + "epoch": 1.907928388746803, + "grad_norm": 6.11898381729892, + "learning_rate": 5.220043477374759e-08, + "loss": 0.228, + "step": 746 + }, + { + "epoch": 1.9104859335038364, + "grad_norm": 5.485214100078597, + "learning_rate": 4.934540215779271e-08, + "loss": 0.2315, + "step": 747 + }, + { + "epoch": 1.9130434782608696, + "grad_norm": 5.1671494114407155, + "learning_rate": 4.657026981834623e-08, + "loss": 0.2191, + "step": 748 + }, + { + "epoch": 1.9156010230179028, + "grad_norm": 5.544282249484376, + "learning_rate": 4.3875082544201364e-08, + "loss": 0.2446, + "step": 749 + }, + { + "epoch": 1.918158567774936, + "grad_norm": 4.828585643362734, + "learning_rate": 4.125988383388957e-08, + "loss": 0.2293, + "step": 750 + }, + { + "epoch": 1.9207161125319692, + "grad_norm": 6.32383213055493, + "learning_rate": 3.87247158949805e-08, + "loss": 0.2332, + "step": 751 + }, + { + "epoch": 1.9232736572890026, + "grad_norm": 5.560416321895529, + "learning_rate": 3.626961964340203e-08, + "loss": 0.2383, + "step": 752 + }, + { + "epoch": 1.9258312020460358, + "grad_norm": 5.865352141528642, + "learning_rate": 3.389463470277576e-08, + "loss": 0.2104, + "step": 753 + }, + { + "epoch": 1.9283887468030692, + "grad_norm": 5.341153765939635, + "learning_rate": 3.159979940378088e-08, + "loss": 0.2128, + "step": 754 + }, + { + "epoch": 1.9309462915601023, + "grad_norm": 5.291758252142984, + "learning_rate": 2.938515078353521e-08, + "loss": 0.1886, + "step": 755 + }, + { + "epoch": 1.9335038363171355, + "grad_norm": 5.163447364959534, + "learning_rate": 2.725072458499567e-08, + "loss": 0.2996, + "step": 756 + }, + { + "epoch": 1.9360613810741687, + "grad_norm": 6.676816256652347, + "learning_rate": 2.519655525638376e-08, + "loss": 0.2902, + "step": 757 + }, + { + "epoch": 1.938618925831202, + "grad_norm": 6.29139443707462, + "learning_rate": 2.3222675950627106e-08, + "loss": 0.1572, + "step": 758 + }, + { + "epoch": 1.9411764705882353, + "grad_norm": 4.037989400305309, + "learning_rate": 2.1329118524827662e-08, + "loss": 0.1417, + "step": 759 + }, + { + "epoch": 1.9437340153452687, + "grad_norm": 4.5423783674836695, + "learning_rate": 1.9515913539743247e-08, + "loss": 0.233, + "step": 760 + }, + { + "epoch": 1.9462915601023019, + "grad_norm": 3.859294316006792, + "learning_rate": 1.7783090259297918e-08, + "loss": 0.1918, + "step": 761 + }, + { + "epoch": 1.948849104859335, + "grad_norm": 5.265098117553532, + "learning_rate": 1.613067665010959e-08, + "loss": 0.1965, + "step": 762 + }, + { + "epoch": 1.9514066496163682, + "grad_norm": 4.559779542352908, + "learning_rate": 1.4558699381034825e-08, + "loss": 0.1888, + "step": 763 + }, + { + "epoch": 1.9539641943734014, + "grad_norm": 6.603926656306259, + "learning_rate": 1.3067183822742525e-08, + "loss": 0.2729, + "step": 764 + }, + { + "epoch": 1.9565217391304348, + "grad_norm": 5.966454788112439, + "learning_rate": 1.1656154047303691e-08, + "loss": 0.1994, + "step": 765 + }, + { + "epoch": 1.959079283887468, + "grad_norm": 5.135081772282031, + "learning_rate": 1.0325632827801745e-08, + "loss": 0.1875, + "step": 766 + }, + { + "epoch": 1.9616368286445014, + "grad_norm": 4.6463930984383275, + "learning_rate": 9.075641637964483e-09, + "loss": 0.213, + "step": 767 + }, + { + "epoch": 1.9641943734015346, + "grad_norm": 5.19696586284704, + "learning_rate": 7.906200651819907e-09, + "loss": 0.188, + "step": 768 + }, + { + "epoch": 1.9667519181585678, + "grad_norm": 4.1586996620802585, + "learning_rate": 6.817328743368712e-09, + "loss": 0.1314, + "step": 769 + }, + { + "epoch": 1.969309462915601, + "grad_norm": 6.428796482081383, + "learning_rate": 5.809043486279531e-09, + "loss": 0.265, + "step": 770 + }, + { + "epoch": 1.9718670076726341, + "grad_norm": 6.071118019652648, + "learning_rate": 4.881361153606934e-09, + "loss": 0.1938, + "step": 771 + }, + { + "epoch": 1.9744245524296675, + "grad_norm": 6.004256284137117, + "learning_rate": 4.034296717527752e-09, + "loss": 0.258, + "step": 772 + }, + { + "epoch": 1.976982097186701, + "grad_norm": 4.424170682035935, + "learning_rate": 3.2678638490996064e-09, + "loss": 0.1944, + "step": 773 + }, + { + "epoch": 1.979539641943734, + "grad_norm": 4.5443816733282025, + "learning_rate": 2.5820749180388573e-09, + "loss": 0.225, + "step": 774 + }, + { + "epoch": 1.9820971867007673, + "grad_norm": 6.813889752181503, + "learning_rate": 1.976940992523546e-09, + "loss": 0.242, + "step": 775 + }, + { + "epoch": 1.9846547314578005, + "grad_norm": 6.943362317683093, + "learning_rate": 1.4524718390140913e-09, + "loss": 0.3044, + "step": 776 + }, + { + "epoch": 1.9872122762148337, + "grad_norm": 5.440758095117894, + "learning_rate": 1.0086759220934162e-09, + "loss": 0.2411, + "step": 777 + }, + { + "epoch": 1.989769820971867, + "grad_norm": 3.8628110410030914, + "learning_rate": 6.455604043331676e-10, + "loss": 0.1659, + "step": 778 + }, + { + "epoch": 1.9923273657289002, + "grad_norm": 5.170928482570995, + "learning_rate": 3.631311461765874e-10, + "loss": 0.1393, + "step": 779 + }, + { + "epoch": 1.9948849104859336, + "grad_norm": 5.769015016748028, + "learning_rate": 1.6139270584358823e-10, + "loss": 0.178, + "step": 780 + }, + { + "epoch": 1.9974424552429668, + "grad_norm": 5.070011448256778, + "learning_rate": 4.034833925969928e-11, + "loss": 0.2705, + "step": 781 + }, + { + "epoch": 2.0, + "grad_norm": 6.222117398419603, + "learning_rate": 0.0, + "loss": 0.1879, + "step": 782 + }, + { + "epoch": 2.0, + "step": 782, + "total_flos": 850857689088.0, + "train_loss": 0.39088617330012115, + "train_runtime": 416.9902, + "train_samples_per_second": 15.003, + "train_steps_per_second": 1.875 + } + ], + "logging_steps": 1, + "max_steps": 782, + "num_input_tokens_seen": 0, + "num_train_epochs": 2, + "save_steps": 70000, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": true + }, + "attributes": {} + } + }, + "total_flos": 850857689088.0, + "train_batch_size": 2, + "trial_name": null, + "trial_params": null +}