{ "best_metric": null, "best_model_checkpoint": null, "epoch": 4.990689013035381, "eval_steps": 500, "global_step": 335, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0148975791433892, "grad_norm": 5.950184932951318, "learning_rate": 2.3529411764705885e-06, "loss": 0.7975, "step": 1 }, { "epoch": 0.0297951582867784, "grad_norm": 5.889999633747208, "learning_rate": 4.705882352941177e-06, "loss": 0.7918, "step": 2 }, { "epoch": 0.0446927374301676, "grad_norm": 5.479080731430823, "learning_rate": 7.058823529411766e-06, "loss": 0.7842, "step": 3 }, { "epoch": 0.0595903165735568, "grad_norm": 4.072102472920601, "learning_rate": 9.411764705882354e-06, "loss": 0.7305, "step": 4 }, { "epoch": 0.074487895716946, "grad_norm": 2.0993469862660445, "learning_rate": 1.1764705882352942e-05, "loss": 0.6919, "step": 5 }, { "epoch": 0.0893854748603352, "grad_norm": 5.471012079326577, "learning_rate": 1.4117647058823532e-05, "loss": 0.7321, "step": 6 }, { "epoch": 0.1042830540037244, "grad_norm": 7.497778887994976, "learning_rate": 1.647058823529412e-05, "loss": 0.7089, "step": 7 }, { "epoch": 0.1191806331471136, "grad_norm": 7.744362603097892, "learning_rate": 1.8823529411764708e-05, "loss": 0.7259, "step": 8 }, { "epoch": 0.1340782122905028, "grad_norm": 4.336088431882978, "learning_rate": 2.1176470588235296e-05, "loss": 0.6722, "step": 9 }, { "epoch": 0.148975791433892, "grad_norm": 3.1425590071036407, "learning_rate": 2.3529411764705884e-05, "loss": 0.6323, "step": 10 }, { "epoch": 0.16387337057728119, "grad_norm": 2.293080261987849, "learning_rate": 2.5882352941176475e-05, "loss": 0.6058, "step": 11 }, { "epoch": 0.1787709497206704, "grad_norm": 1.5511769350436764, "learning_rate": 2.8235294117647063e-05, "loss": 0.5911, "step": 12 }, { "epoch": 0.19366852886405958, "grad_norm": 1.43403410737388, "learning_rate": 3.0588235294117644e-05, "loss": 0.5715, "step": 13 }, { "epoch": 0.2085661080074488, "grad_norm": 1.0282624052041764, "learning_rate": 3.294117647058824e-05, "loss": 0.5482, "step": 14 }, { "epoch": 0.22346368715083798, "grad_norm": 22.401057880346002, "learning_rate": 3.529411764705883e-05, "loss": 0.6398, "step": 15 }, { "epoch": 0.2383612662942272, "grad_norm": 4.036215605056668, "learning_rate": 3.7647058823529415e-05, "loss": 0.5932, "step": 16 }, { "epoch": 0.2532588454376164, "grad_norm": 2.1089398812874904, "learning_rate": 4e-05, "loss": 0.5449, "step": 17 }, { "epoch": 0.2681564245810056, "grad_norm": 1.8917019656777936, "learning_rate": 4.235294117647059e-05, "loss": 0.545, "step": 18 }, { "epoch": 0.28305400372439476, "grad_norm": 1.9916573775756592, "learning_rate": 4.470588235294118e-05, "loss": 0.5414, "step": 19 }, { "epoch": 0.297951582867784, "grad_norm": 1.8920562076710838, "learning_rate": 4.705882352941177e-05, "loss": 0.5407, "step": 20 }, { "epoch": 0.3128491620111732, "grad_norm": 1.0243631917296254, "learning_rate": 4.941176470588236e-05, "loss": 0.5197, "step": 21 }, { "epoch": 0.32774674115456237, "grad_norm": 1.8353601120734953, "learning_rate": 5.176470588235295e-05, "loss": 0.5242, "step": 22 }, { "epoch": 0.3426443202979516, "grad_norm": 1.0556187539851671, "learning_rate": 5.411764705882354e-05, "loss": 0.5115, "step": 23 }, { "epoch": 0.3575418994413408, "grad_norm": 1.6827786939863305, "learning_rate": 5.6470588235294126e-05, "loss": 0.5057, "step": 24 }, { "epoch": 0.37243947858473, "grad_norm": 1.3016042554320595, "learning_rate": 5.8823529411764714e-05, "loss": 0.5114, "step": 25 }, { "epoch": 0.38733705772811916, "grad_norm": 1.2977216399034888, "learning_rate": 6.117647058823529e-05, "loss": 0.4991, "step": 26 }, { "epoch": 0.4022346368715084, "grad_norm": 1.223225396686375, "learning_rate": 6.352941176470589e-05, "loss": 0.4941, "step": 27 }, { "epoch": 0.4171322160148976, "grad_norm": 0.9631867501185041, "learning_rate": 6.588235294117648e-05, "loss": 0.4906, "step": 28 }, { "epoch": 0.43202979515828677, "grad_norm": 1.2189198129233703, "learning_rate": 6.823529411764707e-05, "loss": 0.4913, "step": 29 }, { "epoch": 0.44692737430167595, "grad_norm": 0.7374132716188752, "learning_rate": 7.058823529411765e-05, "loss": 0.4785, "step": 30 }, { "epoch": 0.4618249534450652, "grad_norm": 4.329279729767168, "learning_rate": 7.294117647058824e-05, "loss": 0.4855, "step": 31 }, { "epoch": 0.4767225325884544, "grad_norm": 1.9935350979555255, "learning_rate": 7.529411764705883e-05, "loss": 0.5068, "step": 32 }, { "epoch": 0.49162011173184356, "grad_norm": 1.0099023827824438, "learning_rate": 7.764705882352942e-05, "loss": 0.4878, "step": 33 }, { "epoch": 0.5065176908752328, "grad_norm": 1.453165427026921, "learning_rate": 8e-05, "loss": 0.4887, "step": 34 }, { "epoch": 0.521415270018622, "grad_norm": 1.0681365388503938, "learning_rate": 7.999782132320701e-05, "loss": 0.4792, "step": 35 }, { "epoch": 0.5363128491620112, "grad_norm": 1.2991278121605818, "learning_rate": 7.999128553015966e-05, "loss": 0.4746, "step": 36 }, { "epoch": 0.5512104283054003, "grad_norm": 0.9705929073068399, "learning_rate": 7.998039333282696e-05, "loss": 0.47, "step": 37 }, { "epoch": 0.5661080074487895, "grad_norm": 1.288262936234187, "learning_rate": 7.996514591773782e-05, "loss": 0.4758, "step": 38 }, { "epoch": 0.5810055865921788, "grad_norm": 0.6816553781708893, "learning_rate": 7.99455449458517e-05, "loss": 0.4627, "step": 39 }, { "epoch": 0.595903165735568, "grad_norm": 0.9231036421804909, "learning_rate": 7.992159255237773e-05, "loss": 0.4691, "step": 40 }, { "epoch": 0.6108007448789572, "grad_norm": 1.1402552499674083, "learning_rate": 7.989329134654207e-05, "loss": 0.4638, "step": 41 }, { "epoch": 0.6256983240223464, "grad_norm": 0.7253677973155603, "learning_rate": 7.986064441130378e-05, "loss": 0.4593, "step": 42 }, { "epoch": 0.6405959031657356, "grad_norm": 1.1918043936517428, "learning_rate": 7.982365530301885e-05, "loss": 0.4624, "step": 43 }, { "epoch": 0.6554934823091247, "grad_norm": 1.0659452922693904, "learning_rate": 7.978232805105288e-05, "loss": 0.4514, "step": 44 }, { "epoch": 0.6703910614525139, "grad_norm": 0.9974435758912656, "learning_rate": 7.97366671573421e-05, "loss": 0.4495, "step": 45 }, { "epoch": 0.6852886405959032, "grad_norm": 0.741244345164511, "learning_rate": 7.968667759590298e-05, "loss": 0.4498, "step": 46 }, { "epoch": 0.7001862197392924, "grad_norm": 0.700212224489924, "learning_rate": 7.96323648122904e-05, "loss": 0.4496, "step": 47 }, { "epoch": 0.7150837988826816, "grad_norm": 0.6861544024192645, "learning_rate": 7.957373472300442e-05, "loss": 0.4425, "step": 48 }, { "epoch": 0.7299813780260708, "grad_norm": 0.9097284558328156, "learning_rate": 7.951079371484578e-05, "loss": 0.4435, "step": 49 }, { "epoch": 0.74487895716946, "grad_norm": 0.8086828171768753, "learning_rate": 7.944354864422016e-05, "loss": 0.4427, "step": 50 }, { "epoch": 0.7597765363128491, "grad_norm": 0.7336340975598995, "learning_rate": 7.937200683639133e-05, "loss": 0.4418, "step": 51 }, { "epoch": 0.7746741154562383, "grad_norm": 0.709266378863673, "learning_rate": 7.929617608468308e-05, "loss": 0.4413, "step": 52 }, { "epoch": 0.7895716945996276, "grad_norm": 0.6997957286653358, "learning_rate": 7.921606464963037e-05, "loss": 0.4343, "step": 53 }, { "epoch": 0.8044692737430168, "grad_norm": 0.6312526710744601, "learning_rate": 7.913168125807943e-05, "loss": 0.4357, "step": 54 }, { "epoch": 0.819366852886406, "grad_norm": 0.8682993581299321, "learning_rate": 7.90430351022371e-05, "loss": 0.4256, "step": 55 }, { "epoch": 0.8342644320297952, "grad_norm": 1.0211685354189273, "learning_rate": 7.895013583866949e-05, "loss": 0.4383, "step": 56 }, { "epoch": 0.8491620111731844, "grad_norm": 0.5942703334288042, "learning_rate": 7.88529935872501e-05, "loss": 0.4266, "step": 57 }, { "epoch": 0.8640595903165735, "grad_norm": 0.4905469818523975, "learning_rate": 7.875161893005736e-05, "loss": 0.4275, "step": 58 }, { "epoch": 0.8789571694599627, "grad_norm": 0.713524523381961, "learning_rate": 7.864602291022193e-05, "loss": 0.425, "step": 59 }, { "epoch": 0.8938547486033519, "grad_norm": 0.7215883306108987, "learning_rate": 7.85362170307237e-05, "loss": 0.4296, "step": 60 }, { "epoch": 0.9087523277467412, "grad_norm": 0.5476669418552049, "learning_rate": 7.842221325313873e-05, "loss": 0.4264, "step": 61 }, { "epoch": 0.9236499068901304, "grad_norm": 0.5026608565405364, "learning_rate": 7.830402399633624e-05, "loss": 0.4279, "step": 62 }, { "epoch": 0.9385474860335196, "grad_norm": 0.43642895836746987, "learning_rate": 7.818166213512581e-05, "loss": 0.4231, "step": 63 }, { "epoch": 0.9534450651769087, "grad_norm": 0.5516886415086357, "learning_rate": 7.805514099885479e-05, "loss": 0.4286, "step": 64 }, { "epoch": 0.9683426443202979, "grad_norm": 0.6202502097645083, "learning_rate": 7.792447436995634e-05, "loss": 0.4238, "step": 65 }, { "epoch": 0.9832402234636871, "grad_norm": 0.8483095514593049, "learning_rate": 7.778967648244807e-05, "loss": 0.4251, "step": 66 }, { "epoch": 0.9981378026070763, "grad_norm": 0.9010008832838037, "learning_rate": 7.765076202038145e-05, "loss": 0.4227, "step": 67 }, { "epoch": 1.0130353817504656, "grad_norm": 1.1210899234881495, "learning_rate": 7.750774611624222e-05, "loss": 0.7678, "step": 68 }, { "epoch": 1.0279329608938548, "grad_norm": 1.4294187108260912, "learning_rate": 7.736064434930193e-05, "loss": 0.4298, "step": 69 }, { "epoch": 1.042830540037244, "grad_norm": 0.5162696432741285, "learning_rate": 7.720947274392087e-05, "loss": 0.4051, "step": 70 }, { "epoch": 1.0577281191806331, "grad_norm": 1.2710462444765749, "learning_rate": 7.705424776780249e-05, "loss": 0.4149, "step": 71 }, { "epoch": 1.0726256983240223, "grad_norm": 0.7995934416533542, "learning_rate": 7.689498633019941e-05, "loss": 0.4007, "step": 72 }, { "epoch": 1.0875232774674115, "grad_norm": 0.880390966869298, "learning_rate": 7.673170578007157e-05, "loss": 0.4116, "step": 73 }, { "epoch": 1.1024208566108007, "grad_norm": 0.7628454204466086, "learning_rate": 7.656442390419622e-05, "loss": 0.4008, "step": 74 }, { "epoch": 1.1173184357541899, "grad_norm": 0.5370866632430045, "learning_rate": 7.63931589252304e-05, "loss": 0.3935, "step": 75 }, { "epoch": 1.132216014897579, "grad_norm": 0.6503753033899171, "learning_rate": 7.621792949972588e-05, "loss": 0.3983, "step": 76 }, { "epoch": 1.1471135940409685, "grad_norm": 0.483185187236749, "learning_rate": 7.603875471609677e-05, "loss": 0.4092, "step": 77 }, { "epoch": 1.1620111731843576, "grad_norm": 0.5469246081737635, "learning_rate": 7.585565409254025e-05, "loss": 0.3988, "step": 78 }, { "epoch": 1.1769087523277468, "grad_norm": 0.3809737733256841, "learning_rate": 7.566864757491027e-05, "loss": 0.4049, "step": 79 }, { "epoch": 1.191806331471136, "grad_norm": 0.4978345458094043, "learning_rate": 7.547775553454485e-05, "loss": 0.3991, "step": 80 }, { "epoch": 1.2067039106145252, "grad_norm": 0.3419277603644534, "learning_rate": 7.528299876604689e-05, "loss": 0.4015, "step": 81 }, { "epoch": 1.2216014897579144, "grad_norm": 0.3978591421270835, "learning_rate": 7.508439848501899e-05, "loss": 0.3948, "step": 82 }, { "epoch": 1.2364990689013036, "grad_norm": 0.3215672869140669, "learning_rate": 7.488197632575232e-05, "loss": 0.3946, "step": 83 }, { "epoch": 1.2513966480446927, "grad_norm": 0.2805092237683431, "learning_rate": 7.467575433886989e-05, "loss": 0.3925, "step": 84 }, { "epoch": 1.266294227188082, "grad_norm": 0.29668081700935095, "learning_rate": 7.44657549889246e-05, "loss": 0.3944, "step": 85 }, { "epoch": 1.2811918063314711, "grad_norm": 0.32093150433950507, "learning_rate": 7.425200115195193e-05, "loss": 0.395, "step": 86 }, { "epoch": 1.2960893854748603, "grad_norm": 0.3114671727422004, "learning_rate": 7.403451611297808e-05, "loss": 0.3896, "step": 87 }, { "epoch": 1.3109869646182495, "grad_norm": 0.2621069849515514, "learning_rate": 7.381332356348343e-05, "loss": 0.3972, "step": 88 }, { "epoch": 1.3258845437616387, "grad_norm": 0.35041652635951187, "learning_rate": 7.358844759882168e-05, "loss": 0.3947, "step": 89 }, { "epoch": 1.3407821229050279, "grad_norm": 0.35165133417817274, "learning_rate": 7.335991271559512e-05, "loss": 0.3924, "step": 90 }, { "epoch": 1.355679702048417, "grad_norm": 0.3622625923423102, "learning_rate": 7.312774380898608e-05, "loss": 0.3802, "step": 91 }, { "epoch": 1.3705772811918062, "grad_norm": 0.34568371596915276, "learning_rate": 7.289196617004499e-05, "loss": 0.3921, "step": 92 }, { "epoch": 1.3854748603351954, "grad_norm": 0.36106236048997714, "learning_rate": 7.265260548293535e-05, "loss": 0.3852, "step": 93 }, { "epoch": 1.4003724394785848, "grad_norm": 0.3591694116875136, "learning_rate": 7.24096878221359e-05, "loss": 0.3864, "step": 94 }, { "epoch": 1.415270018621974, "grad_norm": 0.4611887631357433, "learning_rate": 7.216323964960013e-05, "loss": 0.3917, "step": 95 }, { "epoch": 1.4301675977653632, "grad_norm": 0.5900177988285087, "learning_rate": 7.191328781187374e-05, "loss": 0.3889, "step": 96 }, { "epoch": 1.4450651769087524, "grad_norm": 0.6476785738427469, "learning_rate": 7.165985953717017e-05, "loss": 0.3976, "step": 97 }, { "epoch": 1.4599627560521415, "grad_norm": 0.5959898548929913, "learning_rate": 7.140298243240444e-05, "loss": 0.3991, "step": 98 }, { "epoch": 1.4748603351955307, "grad_norm": 0.45154508108061203, "learning_rate": 7.114268448018589e-05, "loss": 0.395, "step": 99 }, { "epoch": 1.48975791433892, "grad_norm": 0.41405552449106015, "learning_rate": 7.087899403576992e-05, "loss": 0.3904, "step": 100 }, { "epoch": 1.504655493482309, "grad_norm": 0.550596726311639, "learning_rate": 7.06119398239691e-05, "loss": 0.3809, "step": 101 }, { "epoch": 1.5195530726256983, "grad_norm": 0.6087570876386633, "learning_rate": 7.034155093602413e-05, "loss": 0.3962, "step": 102 }, { "epoch": 1.5344506517690877, "grad_norm": 0.533129391668596, "learning_rate": 7.006785682643479e-05, "loss": 0.3976, "step": 103 }, { "epoch": 1.5493482309124769, "grad_norm": 0.43458513735835697, "learning_rate": 6.979088730975128e-05, "loss": 0.3917, "step": 104 }, { "epoch": 1.564245810055866, "grad_norm": 0.3518341619336429, "learning_rate": 6.951067255732655e-05, "loss": 0.3858, "step": 105 }, { "epoch": 1.5791433891992552, "grad_norm": 0.40015247854184544, "learning_rate": 6.92272430940295e-05, "loss": 0.3864, "step": 106 }, { "epoch": 1.5940409683426444, "grad_norm": 0.3910716089734902, "learning_rate": 6.894062979491987e-05, "loss": 0.3883, "step": 107 }, { "epoch": 1.6089385474860336, "grad_norm": 0.35066296337841696, "learning_rate": 6.865086388188476e-05, "loss": 0.386, "step": 108 }, { "epoch": 1.6238361266294228, "grad_norm": 0.3072713840073755, "learning_rate": 6.835797692023774e-05, "loss": 0.39, "step": 109 }, { "epoch": 1.638733705772812, "grad_norm": 0.3550143349129471, "learning_rate": 6.806200081528008e-05, "loss": 0.3872, "step": 110 }, { "epoch": 1.6536312849162011, "grad_norm": 0.37053547013887106, "learning_rate": 6.776296780882537e-05, "loss": 0.3813, "step": 111 }, { "epoch": 1.6685288640595903, "grad_norm": 0.24071852433891697, "learning_rate": 6.746091047568716e-05, "loss": 0.3812, "step": 112 }, { "epoch": 1.6834264432029795, "grad_norm": 0.3042807960699796, "learning_rate": 6.715586172013054e-05, "loss": 0.3835, "step": 113 }, { "epoch": 1.6983240223463687, "grad_norm": 0.3538152445413312, "learning_rate": 6.684785477228777e-05, "loss": 0.3813, "step": 114 }, { "epoch": 1.7132216014897579, "grad_norm": 0.24086260757195502, "learning_rate": 6.653692318453831e-05, "loss": 0.3852, "step": 115 }, { "epoch": 1.728119180633147, "grad_norm": 0.3025847544952239, "learning_rate": 6.622310082785384e-05, "loss": 0.3913, "step": 116 }, { "epoch": 1.7430167597765363, "grad_norm": 0.31641711838551906, "learning_rate": 6.590642188810869e-05, "loss": 0.3827, "step": 117 }, { "epoch": 1.7579143389199254, "grad_norm": 0.2688168797114581, "learning_rate": 6.558692086235565e-05, "loss": 0.3744, "step": 118 }, { "epoch": 1.7728119180633146, "grad_norm": 0.2984297462672916, "learning_rate": 6.526463255506828e-05, "loss": 0.3811, "step": 119 }, { "epoch": 1.7877094972067038, "grad_norm": 0.30746963837396896, "learning_rate": 6.493959207434934e-05, "loss": 0.3758, "step": 120 }, { "epoch": 1.802607076350093, "grad_norm": 0.3230084067714464, "learning_rate": 6.461183482810646e-05, "loss": 0.3853, "step": 121 }, { "epoch": 1.8175046554934822, "grad_norm": 0.4000243849303593, "learning_rate": 6.42813965201949e-05, "loss": 0.3882, "step": 122 }, { "epoch": 1.8324022346368714, "grad_norm": 0.4288088964712755, "learning_rate": 6.394831314652835e-05, "loss": 0.3858, "step": 123 }, { "epoch": 1.8472998137802608, "grad_norm": 0.5085831972309272, "learning_rate": 6.361262099115761e-05, "loss": 0.3856, "step": 124 }, { "epoch": 1.86219739292365, "grad_norm": 0.6105698024621092, "learning_rate": 6.327435662231812e-05, "loss": 0.3849, "step": 125 }, { "epoch": 1.8770949720670391, "grad_norm": 0.5824359200910605, "learning_rate": 6.293355688844637e-05, "loss": 0.3866, "step": 126 }, { "epoch": 1.8919925512104283, "grad_norm": 0.44326416200929236, "learning_rate": 6.259025891416594e-05, "loss": 0.3737, "step": 127 }, { "epoch": 1.9068901303538175, "grad_norm": 0.3430006009849836, "learning_rate": 6.224450009624332e-05, "loss": 0.3741, "step": 128 }, { "epoch": 1.9217877094972067, "grad_norm": 0.3048045239289814, "learning_rate": 6.18963180995141e-05, "loss": 0.3767, "step": 129 }, { "epoch": 1.9366852886405959, "grad_norm": 0.3664037873157842, "learning_rate": 6.154575085278012e-05, "loss": 0.3857, "step": 130 }, { "epoch": 1.9515828677839853, "grad_norm": 0.3595451501590318, "learning_rate": 6.119283654467761e-05, "loss": 0.3806, "step": 131 }, { "epoch": 1.9664804469273744, "grad_norm": 0.3076574382248462, "learning_rate": 6.083761361951722e-05, "loss": 0.3802, "step": 132 }, { "epoch": 1.9813780260707636, "grad_norm": 0.25896414972948856, "learning_rate": 6.048012077309612e-05, "loss": 0.3773, "step": 133 }, { "epoch": 1.9962756052141528, "grad_norm": 0.2438114619698701, "learning_rate": 6.01203969484827e-05, "loss": 0.3832, "step": 134 }, { "epoch": 2.011173184357542, "grad_norm": 0.4880986651628094, "learning_rate": 5.975848133177442e-05, "loss": 0.7025, "step": 135 }, { "epoch": 2.026070763500931, "grad_norm": 0.7683631863694373, "learning_rate": 5.939441334782901e-05, "loss": 0.3581, "step": 136 }, { "epoch": 2.0409683426443204, "grad_norm": 1.0088180311972472, "learning_rate": 5.9028232655969866e-05, "loss": 0.3662, "step": 137 }, { "epoch": 2.0558659217877095, "grad_norm": 0.8036028379230191, "learning_rate": 5.865997914566577e-05, "loss": 0.3606, "step": 138 }, { "epoch": 2.0707635009310987, "grad_norm": 0.6488117999773151, "learning_rate": 5.8289692932185546e-05, "loss": 0.3619, "step": 139 }, { "epoch": 2.085661080074488, "grad_norm": 0.6603945376014646, "learning_rate": 5.791741435222821e-05, "loss": 0.3613, "step": 140 }, { "epoch": 2.100558659217877, "grad_norm": 0.6137565370101866, "learning_rate": 5.7543183959528886e-05, "loss": 0.3523, "step": 141 }, { "epoch": 2.1154562383612663, "grad_norm": 0.6393236118969376, "learning_rate": 5.716704252044116e-05, "loss": 0.3543, "step": 142 }, { "epoch": 2.1303538175046555, "grad_norm": 0.37602818408299177, "learning_rate": 5.678903100949625e-05, "loss": 0.3564, "step": 143 }, { "epoch": 2.1452513966480447, "grad_norm": 0.4873243740997193, "learning_rate": 5.640919060493948e-05, "loss": 0.3547, "step": 144 }, { "epoch": 2.160148975791434, "grad_norm": 0.399733374633036, "learning_rate": 5.602756268424457e-05, "loss": 0.3521, "step": 145 }, { "epoch": 2.175046554934823, "grad_norm": 0.38098879738078906, "learning_rate": 5.564418881960624e-05, "loss": 0.3525, "step": 146 }, { "epoch": 2.189944134078212, "grad_norm": 0.3112584196059825, "learning_rate": 5.5259110773411596e-05, "loss": 0.3578, "step": 147 }, { "epoch": 2.2048417132216014, "grad_norm": 0.33974576612810525, "learning_rate": 5.487237049369075e-05, "loss": 0.3485, "step": 148 }, { "epoch": 2.2197392923649906, "grad_norm": 0.31599699558663985, "learning_rate": 5.448401010954733e-05, "loss": 0.3482, "step": 149 }, { "epoch": 2.2346368715083798, "grad_norm": 0.23588411232355225, "learning_rate": 5.4094071926569146e-05, "loss": 0.3465, "step": 150 }, { "epoch": 2.249534450651769, "grad_norm": 0.259141065976224, "learning_rate": 5.370259842221972e-05, "loss": 0.3574, "step": 151 }, { "epoch": 2.264432029795158, "grad_norm": 0.2948627119507419, "learning_rate": 5.330963224121096e-05, "loss": 0.356, "step": 152 }, { "epoch": 2.2793296089385473, "grad_norm": 0.2398464878022089, "learning_rate": 5.291521619085785e-05, "loss": 0.3536, "step": 153 }, { "epoch": 2.294227188081937, "grad_norm": 0.2554013267182894, "learning_rate": 5.251939323641516e-05, "loss": 0.3494, "step": 154 }, { "epoch": 2.3091247672253257, "grad_norm": 0.2781655892263444, "learning_rate": 5.212220649639715e-05, "loss": 0.3511, "step": 155 }, { "epoch": 2.3240223463687153, "grad_norm": 0.22065610655196213, "learning_rate": 5.172369923788046e-05, "loss": 0.3589, "step": 156 }, { "epoch": 2.338919925512104, "grad_norm": 0.2632172433601213, "learning_rate": 5.132391487179088e-05, "loss": 0.3537, "step": 157 }, { "epoch": 2.3538175046554937, "grad_norm": 0.18801127961637412, "learning_rate": 5.092289694817446e-05, "loss": 0.3483, "step": 158 }, { "epoch": 2.368715083798883, "grad_norm": 0.2368052254172045, "learning_rate": 5.052068915145336e-05, "loss": 0.3571, "step": 159 }, { "epoch": 2.383612662942272, "grad_norm": 0.23655539210427956, "learning_rate": 5.011733529566723e-05, "loss": 0.3446, "step": 160 }, { "epoch": 2.398510242085661, "grad_norm": 0.2138112386109768, "learning_rate": 4.971287931970033e-05, "loss": 0.3482, "step": 161 }, { "epoch": 2.4134078212290504, "grad_norm": 0.23467262592443583, "learning_rate": 4.9307365282495075e-05, "loss": 0.3473, "step": 162 }, { "epoch": 2.4283054003724396, "grad_norm": 0.1804084648863834, "learning_rate": 4.890083735825258e-05, "loss": 0.3531, "step": 163 }, { "epoch": 2.4432029795158288, "grad_norm": 0.22285709273589918, "learning_rate": 4.849333983162056e-05, "loss": 0.3516, "step": 164 }, { "epoch": 2.458100558659218, "grad_norm": 0.14828204605953096, "learning_rate": 4.808491709286921e-05, "loss": 0.356, "step": 165 }, { "epoch": 2.472998137802607, "grad_norm": 0.1797938362516586, "learning_rate": 4.76756136330557e-05, "loss": 0.3491, "step": 166 }, { "epoch": 2.4878957169459963, "grad_norm": 0.15466564418943565, "learning_rate": 4.726547403917746e-05, "loss": 0.3435, "step": 167 }, { "epoch": 2.5027932960893855, "grad_norm": 0.1866313977972178, "learning_rate": 4.685454298931527e-05, "loss": 0.3506, "step": 168 }, { "epoch": 2.5176908752327747, "grad_norm": 0.15144628774380284, "learning_rate": 4.6442865247766203e-05, "loss": 0.3547, "step": 169 }, { "epoch": 2.532588454376164, "grad_norm": 0.17455463171543423, "learning_rate": 4.603048566016735e-05, "loss": 0.3486, "step": 170 }, { "epoch": 2.547486033519553, "grad_norm": 0.17359950511666877, "learning_rate": 4.5617449148610584e-05, "loss": 0.3483, "step": 171 }, { "epoch": 2.5623836126629422, "grad_norm": 0.16169017429437474, "learning_rate": 4.520380070674902e-05, "loss": 0.3477, "step": 172 }, { "epoch": 2.5772811918063314, "grad_norm": 0.17061911155312887, "learning_rate": 4.478958539489569e-05, "loss": 0.3571, "step": 173 }, { "epoch": 2.5921787709497206, "grad_norm": 0.16695037154209072, "learning_rate": 4.437484833511499e-05, "loss": 0.3522, "step": 174 }, { "epoch": 2.60707635009311, "grad_norm": 0.15955082753766311, "learning_rate": 4.395963470630723e-05, "loss": 0.3413, "step": 175 }, { "epoch": 2.621973929236499, "grad_norm": 0.1573403347105783, "learning_rate": 4.3543989739287326e-05, "loss": 0.3477, "step": 176 }, { "epoch": 2.636871508379888, "grad_norm": 0.14729071549065167, "learning_rate": 4.312795871185742e-05, "loss": 0.3483, "step": 177 }, { "epoch": 2.6517690875232773, "grad_norm": 0.1625891228202278, "learning_rate": 4.2711586943874774e-05, "loss": 0.3452, "step": 178 }, { "epoch": 2.6666666666666665, "grad_norm": 0.15062427849577625, "learning_rate": 4.2294919792314794e-05, "loss": 0.3488, "step": 179 }, { "epoch": 2.6815642458100557, "grad_norm": 0.14387393723407899, "learning_rate": 4.1878002646330144e-05, "loss": 0.3524, "step": 180 }, { "epoch": 2.6964618249534453, "grad_norm": 0.14777699560280422, "learning_rate": 4.1460880922306367e-05, "loss": 0.3458, "step": 181 }, { "epoch": 2.711359404096834, "grad_norm": 0.1590574368277688, "learning_rate": 4.1043600058914436e-05, "loss": 0.3479, "step": 182 }, { "epoch": 2.7262569832402237, "grad_norm": 0.1285624178683975, "learning_rate": 4.0626205512161034e-05, "loss": 0.3503, "step": 183 }, { "epoch": 2.7411545623836124, "grad_norm": 0.14369454205895854, "learning_rate": 4.020874275043679e-05, "loss": 0.346, "step": 184 }, { "epoch": 2.756052141527002, "grad_norm": 0.12193167492676163, "learning_rate": 3.979125724956324e-05, "loss": 0.3506, "step": 185 }, { "epoch": 2.770949720670391, "grad_norm": 0.1326021753158276, "learning_rate": 3.937379448783898e-05, "loss": 0.3502, "step": 186 }, { "epoch": 2.7858472998137804, "grad_norm": 0.12810706794198765, "learning_rate": 3.895639994108558e-05, "loss": 0.3541, "step": 187 }, { "epoch": 2.8007448789571696, "grad_norm": 0.12656208043359207, "learning_rate": 3.853911907769365e-05, "loss": 0.3497, "step": 188 }, { "epoch": 2.815642458100559, "grad_norm": 0.11404167595412021, "learning_rate": 3.812199735366986e-05, "loss": 0.3473, "step": 189 }, { "epoch": 2.830540037243948, "grad_norm": 0.13878407632584827, "learning_rate": 3.770508020768522e-05, "loss": 0.3489, "step": 190 }, { "epoch": 2.845437616387337, "grad_norm": 0.12311309579828897, "learning_rate": 3.728841305612524e-05, "loss": 0.3464, "step": 191 }, { "epoch": 2.8603351955307263, "grad_norm": 0.13332912284616857, "learning_rate": 3.687204128814259e-05, "loss": 0.3428, "step": 192 }, { "epoch": 2.8752327746741155, "grad_norm": 0.1384361081760211, "learning_rate": 3.645601026071269e-05, "loss": 0.3505, "step": 193 }, { "epoch": 2.8901303538175047, "grad_norm": 0.13736333082695462, "learning_rate": 3.604036529369277e-05, "loss": 0.347, "step": 194 }, { "epoch": 2.905027932960894, "grad_norm": 0.11583463546593212, "learning_rate": 3.5625151664885036e-05, "loss": 0.3461, "step": 195 }, { "epoch": 2.919925512104283, "grad_norm": 0.12889980341532484, "learning_rate": 3.5210414605104314e-05, "loss": 0.3428, "step": 196 }, { "epoch": 2.9348230912476723, "grad_norm": 0.11316852370701413, "learning_rate": 3.4796199293250987e-05, "loss": 0.3479, "step": 197 }, { "epoch": 2.9497206703910615, "grad_norm": 0.11454972761265435, "learning_rate": 3.438255085138943e-05, "loss": 0.3464, "step": 198 }, { "epoch": 2.9646182495344506, "grad_norm": 0.13041453651138907, "learning_rate": 3.396951433983266e-05, "loss": 0.3458, "step": 199 }, { "epoch": 2.97951582867784, "grad_norm": 0.12148037255351198, "learning_rate": 3.355713475223382e-05, "loss": 0.3463, "step": 200 }, { "epoch": 2.994413407821229, "grad_norm": 0.1255719642211306, "learning_rate": 3.314545701068475e-05, "loss": 0.3491, "step": 201 }, { "epoch": 3.009310986964618, "grad_norm": 0.3022359318173022, "learning_rate": 3.2734525960822545e-05, "loss": 0.6352, "step": 202 }, { "epoch": 3.0242085661080074, "grad_norm": 0.24062134731666432, "learning_rate": 3.232438636694431e-05, "loss": 0.3224, "step": 203 }, { "epoch": 3.0391061452513966, "grad_norm": 0.18412193702307067, "learning_rate": 3.191508290713079e-05, "loss": 0.3206, "step": 204 }, { "epoch": 3.0540037243947857, "grad_norm": 0.20860348558828723, "learning_rate": 3.150666016837947e-05, "loss": 0.3264, "step": 205 }, { "epoch": 3.068901303538175, "grad_norm": 0.20845238316157091, "learning_rate": 3.109916264174743e-05, "loss": 0.327, "step": 206 }, { "epoch": 3.083798882681564, "grad_norm": 0.20010239080813352, "learning_rate": 3.069263471750493e-05, "loss": 0.3223, "step": 207 }, { "epoch": 3.0986964618249533, "grad_norm": 0.18760823228422985, "learning_rate": 3.0287120680299677e-05, "loss": 0.3238, "step": 208 }, { "epoch": 3.1135940409683425, "grad_norm": 0.21319162553294677, "learning_rate": 2.988266470433277e-05, "loss": 0.3249, "step": 209 }, { "epoch": 3.1284916201117317, "grad_norm": 0.17329474811121787, "learning_rate": 2.9479310848546644e-05, "loss": 0.3243, "step": 210 }, { "epoch": 3.143389199255121, "grad_norm": 0.20404750548823786, "learning_rate": 2.9077103051825567e-05, "loss": 0.3203, "step": 211 }, { "epoch": 3.1582867783985105, "grad_norm": 0.17940098271660299, "learning_rate": 2.8676085128209133e-05, "loss": 0.3237, "step": 212 }, { "epoch": 3.1731843575418996, "grad_norm": 0.16237217554172287, "learning_rate": 2.8276300762119553e-05, "loss": 0.3196, "step": 213 }, { "epoch": 3.188081936685289, "grad_norm": 0.15801986044567953, "learning_rate": 2.787779350360286e-05, "loss": 0.3196, "step": 214 }, { "epoch": 3.202979515828678, "grad_norm": 0.1485056929542617, "learning_rate": 2.748060676358484e-05, "loss": 0.3192, "step": 215 }, { "epoch": 3.217877094972067, "grad_norm": 0.15223329100287147, "learning_rate": 2.7084783809142164e-05, "loss": 0.3192, "step": 216 }, { "epoch": 3.2327746741154564, "grad_norm": 0.13836533584306443, "learning_rate": 2.6690367758789046e-05, "loss": 0.328, "step": 217 }, { "epoch": 3.2476722532588456, "grad_norm": 0.1351886559378493, "learning_rate": 2.6297401577780295e-05, "loss": 0.322, "step": 218 }, { "epoch": 3.2625698324022347, "grad_norm": 0.12857547848905795, "learning_rate": 2.5905928073430854e-05, "loss": 0.319, "step": 219 }, { "epoch": 3.277467411545624, "grad_norm": 0.1274395295470198, "learning_rate": 2.5515989890452674e-05, "loss": 0.3237, "step": 220 }, { "epoch": 3.292364990689013, "grad_norm": 0.11899376687857942, "learning_rate": 2.5127629506309264e-05, "loss": 0.3205, "step": 221 }, { "epoch": 3.3072625698324023, "grad_norm": 0.1304655308172865, "learning_rate": 2.474088922658842e-05, "loss": 0.3188, "step": 222 }, { "epoch": 3.3221601489757915, "grad_norm": 0.11611090203712811, "learning_rate": 2.4355811180393767e-05, "loss": 0.3195, "step": 223 }, { "epoch": 3.3370577281191807, "grad_norm": 0.11604416368224107, "learning_rate": 2.397243731575543e-05, "loss": 0.3201, "step": 224 }, { "epoch": 3.35195530726257, "grad_norm": 0.11134670769841407, "learning_rate": 2.359080939506052e-05, "loss": 0.32, "step": 225 }, { "epoch": 3.366852886405959, "grad_norm": 0.10870133266703798, "learning_rate": 2.3210968990503755e-05, "loss": 0.3208, "step": 226 }, { "epoch": 3.381750465549348, "grad_norm": 0.1110502941686079, "learning_rate": 2.2832957479558866e-05, "loss": 0.317, "step": 227 }, { "epoch": 3.3966480446927374, "grad_norm": 0.10438158313371027, "learning_rate": 2.245681604047114e-05, "loss": 0.3179, "step": 228 }, { "epoch": 3.4115456238361266, "grad_norm": 0.10282665897445756, "learning_rate": 2.2082585647771807e-05, "loss": 0.3228, "step": 229 }, { "epoch": 3.4264432029795158, "grad_norm": 0.10672800838711503, "learning_rate": 2.171030706781446e-05, "loss": 0.3222, "step": 230 }, { "epoch": 3.441340782122905, "grad_norm": 0.09786842543478856, "learning_rate": 2.1340020854334246e-05, "loss": 0.3181, "step": 231 }, { "epoch": 3.456238361266294, "grad_norm": 0.10404230481342372, "learning_rate": 2.0971767344030144e-05, "loss": 0.3225, "step": 232 }, { "epoch": 3.4711359404096833, "grad_norm": 0.10712648472255477, "learning_rate": 2.0605586652170998e-05, "loss": 0.321, "step": 233 }, { "epoch": 3.4860335195530725, "grad_norm": 0.1044874161797147, "learning_rate": 2.0241518668225595e-05, "loss": 0.3252, "step": 234 }, { "epoch": 3.5009310986964617, "grad_norm": 0.10451587153618398, "learning_rate": 1.98796030515173e-05, "loss": 0.3173, "step": 235 }, { "epoch": 3.515828677839851, "grad_norm": 0.09831156843059215, "learning_rate": 1.9519879226903903e-05, "loss": 0.3188, "step": 236 }, { "epoch": 3.5307262569832405, "grad_norm": 0.09438833971964583, "learning_rate": 1.9162386380482795e-05, "loss": 0.3174, "step": 237 }, { "epoch": 3.5456238361266292, "grad_norm": 0.10433270782878293, "learning_rate": 1.88071634553224e-05, "loss": 0.3219, "step": 238 }, { "epoch": 3.560521415270019, "grad_norm": 0.09292286071725575, "learning_rate": 1.845424914721988e-05, "loss": 0.3252, "step": 239 }, { "epoch": 3.5754189944134076, "grad_norm": 0.09413410282284697, "learning_rate": 1.81036819004859e-05, "loss": 0.3163, "step": 240 }, { "epoch": 3.5903165735567972, "grad_norm": 0.10158989698055416, "learning_rate": 1.7755499903756704e-05, "loss": 0.3259, "step": 241 }, { "epoch": 3.605214152700186, "grad_norm": 0.09464740249041373, "learning_rate": 1.7409741085834066e-05, "loss": 0.3191, "step": 242 }, { "epoch": 3.6201117318435756, "grad_norm": 0.10726107957817572, "learning_rate": 1.7066443111553627e-05, "loss": 0.3211, "step": 243 }, { "epoch": 3.635009310986965, "grad_norm": 0.08917870235234628, "learning_rate": 1.6725643377681893e-05, "loss": 0.3225, "step": 244 }, { "epoch": 3.649906890130354, "grad_norm": 0.10677910900640168, "learning_rate": 1.638737900884239e-05, "loss": 0.3214, "step": 245 }, { "epoch": 3.664804469273743, "grad_norm": 0.09774898119089985, "learning_rate": 1.6051686853471667e-05, "loss": 0.3252, "step": 246 }, { "epoch": 3.6797020484171323, "grad_norm": 0.08773810545171451, "learning_rate": 1.5718603479805113e-05, "loss": 0.3235, "step": 247 }, { "epoch": 3.6945996275605215, "grad_norm": 0.10111349625699705, "learning_rate": 1.538816517189356e-05, "loss": 0.3228, "step": 248 }, { "epoch": 3.7094972067039107, "grad_norm": 0.09592458935558433, "learning_rate": 1.5060407925650662e-05, "loss": 0.3234, "step": 249 }, { "epoch": 3.7243947858473, "grad_norm": 0.08715032888076826, "learning_rate": 1.4735367444931722e-05, "loss": 0.3192, "step": 250 }, { "epoch": 3.739292364990689, "grad_norm": 0.09053851330147095, "learning_rate": 1.4413079137644358e-05, "loss": 0.3162, "step": 251 }, { "epoch": 3.7541899441340782, "grad_norm": 0.10377676484499614, "learning_rate": 1.4093578111891333e-05, "loss": 0.3189, "step": 252 }, { "epoch": 3.7690875232774674, "grad_norm": 0.08949080189520761, "learning_rate": 1.377689917214617e-05, "loss": 0.3258, "step": 253 }, { "epoch": 3.7839851024208566, "grad_norm": 0.09258472675428549, "learning_rate": 1.3463076815461703e-05, "loss": 0.3195, "step": 254 }, { "epoch": 3.798882681564246, "grad_norm": 0.0951200280500826, "learning_rate": 1.3152145227712221e-05, "loss": 0.314, "step": 255 }, { "epoch": 3.813780260707635, "grad_norm": 0.09226678982549726, "learning_rate": 1.284413827986946e-05, "loss": 0.3255, "step": 256 }, { "epoch": 3.828677839851024, "grad_norm": 0.08915216421489361, "learning_rate": 1.2539089524312855e-05, "loss": 0.3261, "step": 257 }, { "epoch": 3.8435754189944134, "grad_norm": 0.0925569123285621, "learning_rate": 1.2237032191174642e-05, "loss": 0.3202, "step": 258 }, { "epoch": 3.8584729981378025, "grad_norm": 0.08684065982720417, "learning_rate": 1.1937999184719926e-05, "loss": 0.3212, "step": 259 }, { "epoch": 3.8733705772811917, "grad_norm": 0.08765050211323365, "learning_rate": 1.1642023079762281e-05, "loss": 0.319, "step": 260 }, { "epoch": 3.888268156424581, "grad_norm": 0.08724442004118642, "learning_rate": 1.1349136118115242e-05, "loss": 0.3154, "step": 261 }, { "epoch": 3.90316573556797, "grad_norm": 0.08586178309454286, "learning_rate": 1.1059370205080157e-05, "loss": 0.3226, "step": 262 }, { "epoch": 3.9180633147113593, "grad_norm": 0.08501873030376662, "learning_rate": 1.07727569059705e-05, "loss": 0.3229, "step": 263 }, { "epoch": 3.9329608938547485, "grad_norm": 0.09206134247645327, "learning_rate": 1.0489327442673459e-05, "loss": 0.3227, "step": 264 }, { "epoch": 3.9478584729981376, "grad_norm": 0.08410101453326498, "learning_rate": 1.0209112690248726e-05, "loss": 0.3177, "step": 265 }, { "epoch": 3.9627560521415273, "grad_norm": 0.08821111302237976, "learning_rate": 9.932143173565225e-06, "loss": 0.3227, "step": 266 }, { "epoch": 3.977653631284916, "grad_norm": 0.08616646529797505, "learning_rate": 9.658449063975875e-06, "loss": 0.3216, "step": 267 }, { "epoch": 3.9925512104283056, "grad_norm": 0.0885838583519875, "learning_rate": 9.388060176030907e-06, "loss": 0.3226, "step": 268 }, { "epoch": 4.007448789571694, "grad_norm": 0.22436110155064673, "learning_rate": 9.12100596423009e-06, "loss": 0.577, "step": 269 }, { "epoch": 4.022346368715084, "grad_norm": 0.11607630428391699, "learning_rate": 8.857315519814111e-06, "loss": 0.306, "step": 270 }, { "epoch": 4.037243947858473, "grad_norm": 0.11774451222490284, "learning_rate": 8.597017567595562e-06, "loss": 0.3064, "step": 271 }, { "epoch": 4.052141527001862, "grad_norm": 0.09907950082186576, "learning_rate": 8.34014046282984e-06, "loss": 0.302, "step": 272 }, { "epoch": 4.067039106145251, "grad_norm": 0.10570787971335635, "learning_rate": 8.086712188126263e-06, "loss": 0.3021, "step": 273 }, { "epoch": 4.081936685288641, "grad_norm": 0.10877019752281963, "learning_rate": 7.836760350399881e-06, "loss": 0.2992, "step": 274 }, { "epoch": 4.0968342644320295, "grad_norm": 0.1173773952444213, "learning_rate": 7.5903121778641096e-06, "loss": 0.3077, "step": 275 }, { "epoch": 4.111731843575419, "grad_norm": 0.10898666220477518, "learning_rate": 7.347394517064663e-06, "loss": 0.3068, "step": 276 }, { "epoch": 4.126629422718808, "grad_norm": 0.09803810087456757, "learning_rate": 7.108033829955028e-06, "loss": 0.3077, "step": 277 }, { "epoch": 4.1415270018621975, "grad_norm": 0.09446749801813671, "learning_rate": 6.87225619101394e-06, "loss": 0.3061, "step": 278 }, { "epoch": 4.156424581005586, "grad_norm": 0.0951410747387855, "learning_rate": 6.640087284404888e-06, "loss": 0.3054, "step": 279 }, { "epoch": 4.171322160148976, "grad_norm": 0.10491597814630026, "learning_rate": 6.411552401178327e-06, "loss": 0.3098, "step": 280 }, { "epoch": 4.186219739292365, "grad_norm": 0.09593380058395089, "learning_rate": 6.186676436516581e-06, "loss": 0.3008, "step": 281 }, { "epoch": 4.201117318435754, "grad_norm": 0.09551371288826128, "learning_rate": 5.965483887021934e-06, "loss": 0.3087, "step": 282 }, { "epoch": 4.216014897579143, "grad_norm": 0.0895153508605015, "learning_rate": 5.747998848048091e-06, "loss": 0.3091, "step": 283 }, { "epoch": 4.230912476722533, "grad_norm": 0.09197880309307534, "learning_rate": 5.534245011075414e-06, "loss": 0.298, "step": 284 }, { "epoch": 4.245810055865922, "grad_norm": 0.08575532851271551, "learning_rate": 5.3242456611301095e-06, "loss": 0.3, "step": 285 }, { "epoch": 4.260707635009311, "grad_norm": 0.08276481700085347, "learning_rate": 5.118023674247692e-06, "loss": 0.2984, "step": 286 }, { "epoch": 4.275605214152701, "grad_norm": 0.08713205903583655, "learning_rate": 4.915601514981024e-06, "loss": 0.3078, "step": 287 }, { "epoch": 4.290502793296089, "grad_norm": 0.08363731970724225, "learning_rate": 4.717001233953116e-06, "loss": 0.2987, "step": 288 }, { "epoch": 4.305400372439479, "grad_norm": 0.08529607146039025, "learning_rate": 4.522244465455154e-06, "loss": 0.3038, "step": 289 }, { "epoch": 4.320297951582868, "grad_norm": 0.08215569802598276, "learning_rate": 4.33135242508973e-06, "loss": 0.3043, "step": 290 }, { "epoch": 4.335195530726257, "grad_norm": 0.08149200923150887, "learning_rate": 4.144345907459753e-06, "loss": 0.3013, "step": 291 }, { "epoch": 4.350093109869646, "grad_norm": 0.08086624632566386, "learning_rate": 3.961245283903239e-06, "loss": 0.308, "step": 292 }, { "epoch": 4.364990689013036, "grad_norm": 0.07896402407099593, "learning_rate": 3.7820705002741353e-06, "loss": 0.3052, "step": 293 }, { "epoch": 4.379888268156424, "grad_norm": 0.08231952485165063, "learning_rate": 3.6068410747696112e-06, "loss": 0.3002, "step": 294 }, { "epoch": 4.394785847299814, "grad_norm": 0.07789166320140198, "learning_rate": 3.435576095803792e-06, "loss": 0.3021, "step": 295 }, { "epoch": 4.409683426443203, "grad_norm": 0.07892696916845467, "learning_rate": 3.268294219928434e-06, "loss": 0.3046, "step": 296 }, { "epoch": 4.424581005586592, "grad_norm": 0.07646811095500806, "learning_rate": 3.1050136698005963e-06, "loss": 0.3025, "step": 297 }, { "epoch": 4.439478584729981, "grad_norm": 0.07957607014767461, "learning_rate": 2.9457522321975253e-06, "loss": 0.3079, "step": 298 }, { "epoch": 4.454376163873371, "grad_norm": 0.08390752993938617, "learning_rate": 2.790527256079134e-06, "loss": 0.3066, "step": 299 }, { "epoch": 4.4692737430167595, "grad_norm": 0.07541814294925331, "learning_rate": 2.6393556506980834e-06, "loss": 0.3071, "step": 300 }, { "epoch": 4.484171322160149, "grad_norm": 0.0714738276840467, "learning_rate": 2.4922538837577916e-06, "loss": 0.3078, "step": 301 }, { "epoch": 4.499068901303538, "grad_norm": 0.0729695749304457, "learning_rate": 2.349237979618555e-06, "loss": 0.3112, "step": 302 }, { "epoch": 4.5139664804469275, "grad_norm": 0.10108195097380965, "learning_rate": 2.2103235175519355e-06, "loss": 0.3052, "step": 303 }, { "epoch": 4.528864059590316, "grad_norm": 0.07138906578544997, "learning_rate": 2.0755256300436687e-06, "loss": 0.3014, "step": 304 }, { "epoch": 4.543761638733706, "grad_norm": 0.07356056598380585, "learning_rate": 1.944859001145223e-06, "loss": 0.3065, "step": 305 }, { "epoch": 4.558659217877095, "grad_norm": 0.07516469607423958, "learning_rate": 1.8183378648741979e-06, "loss": 0.3049, "step": 306 }, { "epoch": 4.573556797020484, "grad_norm": 0.07371825538928153, "learning_rate": 1.6959760036637662e-06, "loss": 0.304, "step": 307 }, { "epoch": 4.588454376163874, "grad_norm": 0.07502909475594224, "learning_rate": 1.5777867468612874e-06, "loss": 0.3011, "step": 308 }, { "epoch": 4.603351955307263, "grad_norm": 0.07352157905066296, "learning_rate": 1.4637829692763128e-06, "loss": 0.3029, "step": 309 }, { "epoch": 4.618249534450651, "grad_norm": 0.07644054766848495, "learning_rate": 1.353977089778078e-06, "loss": 0.3046, "step": 310 }, { "epoch": 4.633147113594041, "grad_norm": 0.07171131988386027, "learning_rate": 1.2483810699426458e-06, "loss": 0.3068, "step": 311 }, { "epoch": 4.648044692737431, "grad_norm": 0.07057850167398541, "learning_rate": 1.1470064127499091e-06, "loss": 0.3064, "step": 312 }, { "epoch": 4.662942271880819, "grad_norm": 0.07626745016001804, "learning_rate": 1.0498641613305182e-06, "loss": 0.3007, "step": 313 }, { "epoch": 4.677839851024208, "grad_norm": 0.0693829644961785, "learning_rate": 9.569648977629176e-07, "loss": 0.3049, "step": 314 }, { "epoch": 4.692737430167598, "grad_norm": 0.0674761214613311, "learning_rate": 8.683187419205797e-07, "loss": 0.3013, "step": 315 }, { "epoch": 4.707635009310987, "grad_norm": 0.06823222699165288, "learning_rate": 7.839353503696379e-07, "loss": 0.3071, "step": 316 }, { "epoch": 4.722532588454376, "grad_norm": 0.06859479836699696, "learning_rate": 7.038239153169324e-07, "loss": 0.301, "step": 317 }, { "epoch": 4.737430167597766, "grad_norm": 0.07031736775350605, "learning_rate": 6.279931636086912e-07, "loss": 0.3086, "step": 318 }, { "epoch": 4.752327746741154, "grad_norm": 0.06915744071480374, "learning_rate": 5.564513557798501e-07, "loss": 0.3048, "step": 319 }, { "epoch": 4.767225325884544, "grad_norm": 0.06815891873681693, "learning_rate": 4.892062851542356e-07, "loss": 0.309, "step": 320 }, { "epoch": 4.782122905027933, "grad_norm": 0.06935050248754245, "learning_rate": 4.2626527699558996e-07, "loss": 0.3018, "step": 321 }, { "epoch": 4.797020484171322, "grad_norm": 0.06784753003557113, "learning_rate": 3.6763518770960517e-07, "loss": 0.3051, "step": 322 }, { "epoch": 4.811918063314711, "grad_norm": 0.06808652827602746, "learning_rate": 3.133224040970273e-07, "loss": 0.3071, "step": 323 }, { "epoch": 4.826815642458101, "grad_norm": 0.07121784851648261, "learning_rate": 2.6333284265790627e-07, "loss": 0.3047, "step": 324 }, { "epoch": 4.8417132216014895, "grad_norm": 0.06700344502725675, "learning_rate": 2.1767194894712462e-07, "loss": 0.305, "step": 325 }, { "epoch": 4.856610800744879, "grad_norm": 0.06927419764551407, "learning_rate": 1.763446969811522e-07, "loss": 0.306, "step": 326 }, { "epoch": 4.871508379888268, "grad_norm": 0.06947638763948212, "learning_rate": 1.3935558869622168e-07, "loss": 0.3042, "step": 327 }, { "epoch": 4.8864059590316575, "grad_norm": 0.06866693792434822, "learning_rate": 1.0670865345793425e-07, "loss": 0.3094, "step": 328 }, { "epoch": 4.901303538175046, "grad_norm": 0.06859589228192586, "learning_rate": 7.840744762229069e-08, "loss": 0.3054, "step": 329 }, { "epoch": 4.916201117318436, "grad_norm": 0.06927380611470749, "learning_rate": 5.445505414831242e-08, "loss": 0.3054, "step": 330 }, { "epoch": 4.931098696461825, "grad_norm": 0.06680927663742126, "learning_rate": 3.485408226218567e-08, "loss": 0.2998, "step": 331 }, { "epoch": 4.945996275605214, "grad_norm": 0.06773777244031369, "learning_rate": 1.960666717304438e-08, "loss": 0.3058, "step": 332 }, { "epoch": 4.960893854748603, "grad_norm": 0.06864556914107234, "learning_rate": 8.714469840351848e-09, "loss": 0.2984, "step": 333 }, { "epoch": 4.975791433891993, "grad_norm": 0.0678647360547641, "learning_rate": 2.178676792996548e-09, "loss": 0.3051, "step": 334 }, { "epoch": 4.990689013035381, "grad_norm": 0.06662103871858636, "learning_rate": 0.0, "loss": 0.2973, "step": 335 }, { "epoch": 4.990689013035381, "step": 335, "total_flos": 8.620546826447094e+18, "train_loss": 0.380745083004681, "train_runtime": 51129.1279, "train_samples_per_second": 3.36, "train_steps_per_second": 0.007 } ], "logging_steps": 1.0, "max_steps": 335, "num_input_tokens_seen": 0, "num_train_epochs": 5, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 8.620546826447094e+18, "train_batch_size": 1, "trial_name": null, "trial_params": null }