{ "best_global_step": null, "best_metric": null, "best_model_checkpoint": null, "epoch": 5.0, "eval_steps": 500, "global_step": 135, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.037037037037037035, "grad_norm": 3.889946460723877, "learning_rate": 0.0, "loss": 1.0255, "step": 1 }, { "epoch": 0.07407407407407407, "grad_norm": 3.7213633060455322, "learning_rate": 3.846153846153847e-06, "loss": 0.99, "step": 2 }, { "epoch": 0.1111111111111111, "grad_norm": 2.358224868774414, "learning_rate": 7.692307692307694e-06, "loss": 1.0113, "step": 3 }, { "epoch": 0.14814814814814814, "grad_norm": 3.7503502368927, "learning_rate": 1.153846153846154e-05, "loss": 0.952, "step": 4 }, { "epoch": 0.18518518518518517, "grad_norm": 6.118247032165527, "learning_rate": 1.5384615384615387e-05, "loss": 1.0118, "step": 5 }, { "epoch": 0.2222222222222222, "grad_norm": 11.75301456451416, "learning_rate": 1.923076923076923e-05, "loss": 1.0014, "step": 6 }, { "epoch": 0.25925925925925924, "grad_norm": 2.9081552028656006, "learning_rate": 2.307692307692308e-05, "loss": 1.0193, "step": 7 }, { "epoch": 0.2962962962962963, "grad_norm": 4.145256996154785, "learning_rate": 2.6923076923076923e-05, "loss": 0.9929, "step": 8 }, { "epoch": 0.3333333333333333, "grad_norm": 2.380268096923828, "learning_rate": 3.0769230769230774e-05, "loss": 0.968, "step": 9 }, { "epoch": 0.37037037037037035, "grad_norm": 2.615178108215332, "learning_rate": 3.461538461538462e-05, "loss": 0.9375, "step": 10 }, { "epoch": 0.4074074074074074, "grad_norm": 2.8492236137390137, "learning_rate": 3.846153846153846e-05, "loss": 0.8548, "step": 11 }, { "epoch": 0.4444444444444444, "grad_norm": 3.7682080268859863, "learning_rate": 4.230769230769231e-05, "loss": 0.9339, "step": 12 }, { "epoch": 0.48148148148148145, "grad_norm": 2.53898286819458, "learning_rate": 4.615384615384616e-05, "loss": 0.8744, "step": 13 }, { "epoch": 0.5185185185185185, "grad_norm": 3.0268638134002686, "learning_rate": 5e-05, "loss": 0.8486, "step": 14 }, { "epoch": 0.5555555555555556, "grad_norm": 1.9204188585281372, "learning_rate": 4.9992540519072045e-05, "loss": 0.8193, "step": 15 }, { "epoch": 0.5925925925925926, "grad_norm": 2.533561944961548, "learning_rate": 4.9970167022408685e-05, "loss": 0.8744, "step": 16 }, { "epoch": 0.6296296296296297, "grad_norm": 1.509847640991211, "learning_rate": 4.993289434509185e-05, "loss": 0.849, "step": 17 }, { "epoch": 0.6666666666666666, "grad_norm": 1.8710334300994873, "learning_rate": 4.988074720132825e-05, "loss": 0.7792, "step": 18 }, { "epoch": 0.7037037037037037, "grad_norm": 1.9412447214126587, "learning_rate": 4.9813760168062285e-05, "loss": 0.8514, "step": 19 }, { "epoch": 0.7407407407407407, "grad_norm": 1.3795047998428345, "learning_rate": 4.9731977662049233e-05, "loss": 0.8437, "step": 20 }, { "epoch": 0.7777777777777778, "grad_norm": 1.3450686931610107, "learning_rate": 4.9635453910404125e-05, "loss": 0.7967, "step": 21 }, { "epoch": 0.8148148148148148, "grad_norm": 1.3413690328598022, "learning_rate": 4.9524252914645555e-05, "loss": 0.8021, "step": 22 }, { "epoch": 0.8518518518518519, "grad_norm": 1.650895357131958, "learning_rate": 4.939844840825861e-05, "loss": 0.798, "step": 23 }, { "epoch": 0.8888888888888888, "grad_norm": 1.3016217947006226, "learning_rate": 4.9258123807804715e-05, "loss": 0.8069, "step": 24 }, { "epoch": 0.9259259259259259, "grad_norm": 1.2827719449996948, "learning_rate": 4.910337215761104e-05, "loss": 0.7925, "step": 25 }, { "epoch": 0.9629629629629629, "grad_norm": 1.6255881786346436, "learning_rate": 4.8934296068076105e-05, "loss": 0.823, "step": 26 }, { "epoch": 1.0, "grad_norm": 1.3472081422805786, "learning_rate": 4.875100764763238e-05, "loss": 0.8121, "step": 27 }, { "epoch": 1.037037037037037, "grad_norm": 1.3420181274414062, "learning_rate": 4.855362842841111e-05, "loss": 0.7562, "step": 28 }, { "epoch": 1.074074074074074, "grad_norm": 1.0205856561660767, "learning_rate": 4.834228928565864e-05, "loss": 0.7176, "step": 29 }, { "epoch": 1.1111111111111112, "grad_norm": 1.5821263790130615, "learning_rate": 4.811713035095761e-05, "loss": 0.7531, "step": 30 }, { "epoch": 1.1481481481481481, "grad_norm": 1.0714203119277954, "learning_rate": 4.7878300919310606e-05, "loss": 0.7189, "step": 31 }, { "epoch": 1.1851851851851851, "grad_norm": 1.2463059425354004, "learning_rate": 4.7625959350147905e-05, "loss": 0.7524, "step": 32 }, { "epoch": 1.2222222222222223, "grad_norm": 1.087406039237976, "learning_rate": 4.7360272962324815e-05, "loss": 0.7422, "step": 33 }, { "epoch": 1.2592592592592593, "grad_norm": 1.4159847497940063, "learning_rate": 4.70814179231785e-05, "loss": 0.7596, "step": 34 }, { "epoch": 1.2962962962962963, "grad_norm": 1.084489345550537, "learning_rate": 4.678957913171748e-05, "loss": 0.7427, "step": 35 }, { "epoch": 1.3333333333333333, "grad_norm": 1.2119760513305664, "learning_rate": 4.648495009602168e-05, "loss": 0.7429, "step": 36 }, { "epoch": 1.3703703703703702, "grad_norm": 1.1701514720916748, "learning_rate": 4.616773280493393e-05, "loss": 0.7292, "step": 37 }, { "epoch": 1.4074074074074074, "grad_norm": 1.2118418216705322, "learning_rate": 4.5838137594128254e-05, "loss": 0.6576, "step": 38 }, { "epoch": 1.4444444444444444, "grad_norm": 1.8038779497146606, "learning_rate": 4.549638300664369e-05, "loss": 0.7537, "step": 39 }, { "epoch": 1.4814814814814814, "grad_norm": 1.0828356742858887, "learning_rate": 4.5142695647975993e-05, "loss": 0.6865, "step": 40 }, { "epoch": 1.5185185185185186, "grad_norm": 1.5504570007324219, "learning_rate": 4.47773100358235e-05, "loss": 0.6754, "step": 41 }, { "epoch": 1.5555555555555556, "grad_norm": 1.0614900588989258, "learning_rate": 4.44004684445867e-05, "loss": 0.6384, "step": 42 }, { "epoch": 1.5925925925925926, "grad_norm": 1.7621930837631226, "learning_rate": 4.401242074472448e-05, "loss": 0.7096, "step": 43 }, { "epoch": 1.6296296296296298, "grad_norm": 1.2761889696121216, "learning_rate": 4.361342423707385e-05, "loss": 0.6737, "step": 44 }, { "epoch": 1.6666666666666665, "grad_norm": 1.1437257528305054, "learning_rate": 4.320374348224273e-05, "loss": 0.6176, "step": 45 }, { "epoch": 1.7037037037037037, "grad_norm": 1.2675424814224243, "learning_rate": 4.2783650125189096e-05, "loss": 0.6927, "step": 46 }, { "epoch": 1.7407407407407407, "grad_norm": 1.4404767751693726, "learning_rate": 4.2353422715102714e-05, "loss": 0.674, "step": 47 }, { "epoch": 1.7777777777777777, "grad_norm": 1.0773519277572632, "learning_rate": 4.191334652070895e-05, "loss": 0.6334, "step": 48 }, { "epoch": 1.8148148148148149, "grad_norm": 1.1088582277297974, "learning_rate": 4.146371334111702e-05, "loss": 0.6316, "step": 49 }, { "epoch": 1.8518518518518519, "grad_norm": 1.1710089445114136, "learning_rate": 4.1004821312338285e-05, "loss": 0.6552, "step": 50 }, { "epoch": 1.8888888888888888, "grad_norm": 1.1748957633972168, "learning_rate": 4.053697470960268e-05, "loss": 0.647, "step": 51 }, { "epoch": 1.925925925925926, "grad_norm": 0.9752252697944641, "learning_rate": 4.006048374560445e-05, "loss": 0.6347, "step": 52 }, { "epoch": 1.9629629629629628, "grad_norm": 1.0895061492919922, "learning_rate": 3.9575664364811015e-05, "loss": 0.6857, "step": 53 }, { "epoch": 2.0, "grad_norm": 1.7490938901901245, "learning_rate": 3.90828380339712e-05, "loss": 0.6664, "step": 54 }, { "epoch": 2.037037037037037, "grad_norm": 1.1272878646850586, "learning_rate": 3.858233152896195e-05, "loss": 0.633, "step": 55 }, { "epoch": 2.074074074074074, "grad_norm": 1.1926578283309937, "learning_rate": 3.8074476718114706e-05, "loss": 0.5644, "step": 56 }, { "epoch": 2.111111111111111, "grad_norm": 1.3418350219726562, "learning_rate": 3.7559610342165064e-05, "loss": 0.6357, "step": 57 }, { "epoch": 2.148148148148148, "grad_norm": 1.3300039768218994, "learning_rate": 3.7038073790971875e-05, "loss": 0.5759, "step": 58 }, { "epoch": 2.185185185185185, "grad_norm": 1.1557310819625854, "learning_rate": 3.65102128771535e-05, "loss": 0.6032, "step": 59 }, { "epoch": 2.2222222222222223, "grad_norm": 1.384039044380188, "learning_rate": 3.597637760679167e-05, "loss": 0.5991, "step": 60 }, { "epoch": 2.259259259259259, "grad_norm": 1.12711763381958, "learning_rate": 3.543692194735464e-05, "loss": 0.6423, "step": 61 }, { "epoch": 2.2962962962962963, "grad_norm": 1.4164304733276367, "learning_rate": 3.4892203592993786e-05, "loss": 0.6083, "step": 62 }, { "epoch": 2.3333333333333335, "grad_norm": 1.0840896368026733, "learning_rate": 3.434258372736915e-05, "loss": 0.619, "step": 63 }, { "epoch": 2.3703703703703702, "grad_norm": 1.3002837896347046, "learning_rate": 3.3788426784161216e-05, "loss": 0.6012, "step": 64 }, { "epoch": 2.4074074074074074, "grad_norm": 1.4222713708877563, "learning_rate": 3.323010020542765e-05, "loss": 0.5493, "step": 65 }, { "epoch": 2.4444444444444446, "grad_norm": 1.0047050714492798, "learning_rate": 3.2667974197965405e-05, "loss": 0.6409, "step": 66 }, { "epoch": 2.4814814814814814, "grad_norm": 1.1294559240341187, "learning_rate": 3.210242148783952e-05, "loss": 0.5421, "step": 67 }, { "epoch": 2.5185185185185186, "grad_norm": 1.3252534866333008, "learning_rate": 3.1533817073241556e-05, "loss": 0.5746, "step": 68 }, { "epoch": 2.5555555555555554, "grad_norm": 1.2919150590896606, "learning_rate": 3.096253797584139e-05, "loss": 0.5149, "step": 69 }, { "epoch": 2.5925925925925926, "grad_norm": 1.1059527397155762, "learning_rate": 3.03889629907974e-05, "loss": 0.6009, "step": 70 }, { "epoch": 2.6296296296296298, "grad_norm": 1.0076607465744019, "learning_rate": 2.981347243559061e-05, "loss": 0.551, "step": 71 }, { "epoch": 2.6666666666666665, "grad_norm": 1.179908037185669, "learning_rate": 2.923644789784955e-05, "loss": 0.4946, "step": 72 }, { "epoch": 2.7037037037037037, "grad_norm": 1.197548508644104, "learning_rate": 2.86582719823328e-05, "loss": 0.5694, "step": 73 }, { "epoch": 2.7407407407407405, "grad_norm": 1.1806666851043701, "learning_rate": 2.807932805723725e-05, "loss": 0.5638, "step": 74 }, { "epoch": 2.7777777777777777, "grad_norm": 1.3463941812515259, "learning_rate": 2.7500000000000004e-05, "loss": 0.5082, "step": 75 }, { "epoch": 2.814814814814815, "grad_norm": 1.1747078895568848, "learning_rate": 2.692067194276276e-05, "loss": 0.5078, "step": 76 }, { "epoch": 2.851851851851852, "grad_norm": 1.1876875162124634, "learning_rate": 2.6341728017667205e-05, "loss": 0.5297, "step": 77 }, { "epoch": 2.888888888888889, "grad_norm": 1.0407406091690063, "learning_rate": 2.5763552102150456e-05, "loss": 0.5319, "step": 78 }, { "epoch": 2.925925925925926, "grad_norm": 1.3862707614898682, "learning_rate": 2.518652756440939e-05, "loss": 0.4949, "step": 79 }, { "epoch": 2.962962962962963, "grad_norm": 1.0679614543914795, "learning_rate": 2.4611037009202603e-05, "loss": 0.5533, "step": 80 }, { "epoch": 3.0, "grad_norm": 1.4187452793121338, "learning_rate": 2.4037462024158607e-05, "loss": 0.574, "step": 81 }, { "epoch": 3.037037037037037, "grad_norm": 1.0800102949142456, "learning_rate": 2.3466182926758456e-05, "loss": 0.5141, "step": 82 }, { "epoch": 3.074074074074074, "grad_norm": 1.321203351020813, "learning_rate": 2.289757851216049e-05, "loss": 0.4526, "step": 83 }, { "epoch": 3.111111111111111, "grad_norm": 1.0968624353408813, "learning_rate": 2.2332025802034607e-05, "loss": 0.523, "step": 84 }, { "epoch": 3.148148148148148, "grad_norm": 1.173933744430542, "learning_rate": 2.176989979457236e-05, "loss": 0.4723, "step": 85 }, { "epoch": 3.185185185185185, "grad_norm": 1.1096935272216797, "learning_rate": 2.1211573215838792e-05, "loss": 0.4802, "step": 86 }, { "epoch": 3.2222222222222223, "grad_norm": 1.2036195993423462, "learning_rate": 2.0657416272630853e-05, "loss": 0.4906, "step": 87 }, { "epoch": 3.259259259259259, "grad_norm": 1.3066295385360718, "learning_rate": 2.010779640700622e-05, "loss": 0.5162, "step": 88 }, { "epoch": 3.2962962962962963, "grad_norm": 1.0549880266189575, "learning_rate": 1.9563078052645367e-05, "loss": 0.5089, "step": 89 }, { "epoch": 3.3333333333333335, "grad_norm": 1.0702770948410034, "learning_rate": 1.9023622393208336e-05, "loss": 0.4899, "step": 90 }, { "epoch": 3.3703703703703702, "grad_norm": 1.1546988487243652, "learning_rate": 1.84897871228465e-05, "loss": 0.4927, "step": 91 }, { "epoch": 3.4074074074074074, "grad_norm": 1.0201380252838135, "learning_rate": 1.796192620902814e-05, "loss": 0.4419, "step": 92 }, { "epoch": 3.4444444444444446, "grad_norm": 1.1780834197998047, "learning_rate": 1.744038965783493e-05, "loss": 0.505, "step": 93 }, { "epoch": 3.4814814814814814, "grad_norm": 1.4081705808639526, "learning_rate": 1.692552328188531e-05, "loss": 0.4298, "step": 94 }, { "epoch": 3.5185185185185186, "grad_norm": 1.1309189796447754, "learning_rate": 1.6417668471038057e-05, "loss": 0.4648, "step": 95 }, { "epoch": 3.5555555555555554, "grad_norm": 1.2165582180023193, "learning_rate": 1.5917161966028815e-05, "loss": 0.4143, "step": 96 }, { "epoch": 3.5925925925925926, "grad_norm": 1.1092609167099, "learning_rate": 1.542433563518899e-05, "loss": 0.4837, "step": 97 }, { "epoch": 3.6296296296296298, "grad_norm": 1.0920662879943848, "learning_rate": 1.4939516254395546e-05, "loss": 0.4345, "step": 98 }, { "epoch": 3.6666666666666665, "grad_norm": 1.0206769704818726, "learning_rate": 1.446302529039732e-05, "loss": 0.3851, "step": 99 }, { "epoch": 3.7037037037037037, "grad_norm": 1.0061428546905518, "learning_rate": 1.399517868766172e-05, "loss": 0.4538, "step": 100 }, { "epoch": 3.7407407407407405, "grad_norm": 1.2543085813522339, "learning_rate": 1.3536286658882989e-05, "loss": 0.4501, "step": 101 }, { "epoch": 3.7777777777777777, "grad_norm": 1.2222250699996948, "learning_rate": 1.3086653479291062e-05, "loss": 0.4127, "step": 102 }, { "epoch": 3.814814814814815, "grad_norm": 1.0591493844985962, "learning_rate": 1.2646577284897284e-05, "loss": 0.4031, "step": 103 }, { "epoch": 3.851851851851852, "grad_norm": 1.2473130226135254, "learning_rate": 1.2216349874810906e-05, "loss": 0.4178, "step": 104 }, { "epoch": 3.888888888888889, "grad_norm": 1.3380539417266846, "learning_rate": 1.1796256517757267e-05, "loss": 0.4176, "step": 105 }, { "epoch": 3.925925925925926, "grad_norm": 1.4147883653640747, "learning_rate": 1.1386575762926155e-05, "loss": 0.4069, "step": 106 }, { "epoch": 3.962962962962963, "grad_norm": 1.1942764520645142, "learning_rate": 1.0987579255275524e-05, "loss": 0.4365, "step": 107 }, { "epoch": 4.0, "grad_norm": 1.0244306325912476, "learning_rate": 1.0599531555413309e-05, "loss": 0.4665, "step": 108 }, { "epoch": 4.037037037037037, "grad_norm": 1.2454806566238403, "learning_rate": 1.0222689964176502e-05, "loss": 0.4011, "step": 109 }, { "epoch": 4.074074074074074, "grad_norm": 0.9134472608566284, "learning_rate": 9.857304352024019e-06, "loss": 0.3615, "step": 110 }, { "epoch": 4.111111111111111, "grad_norm": 0.9509828090667725, "learning_rate": 9.503616993356315e-06, "loss": 0.4154, "step": 111 }, { "epoch": 4.148148148148148, "grad_norm": 0.8373136520385742, "learning_rate": 9.161862405871748e-06, "loss": 0.3683, "step": 112 }, { "epoch": 4.185185185185185, "grad_norm": 0.9871260523796082, "learning_rate": 8.832267195066075e-06, "loss": 0.3695, "step": 113 }, { "epoch": 4.222222222222222, "grad_norm": 1.120859980583191, "learning_rate": 8.515049903978325e-06, "loss": 0.3897, "step": 114 }, { "epoch": 4.2592592592592595, "grad_norm": 1.2524614334106445, "learning_rate": 8.210420868282522e-06, "loss": 0.4148, "step": 115 }, { "epoch": 4.296296296296296, "grad_norm": 1.1435761451721191, "learning_rate": 7.918582076821507e-06, "loss": 0.4006, "step": 116 }, { "epoch": 4.333333333333333, "grad_norm": 1.2172073125839233, "learning_rate": 7.639727037675181e-06, "loss": 0.3915, "step": 117 }, { "epoch": 4.37037037037037, "grad_norm": 1.1755033731460571, "learning_rate": 7.374040649852105e-06, "loss": 0.4021, "step": 118 }, { "epoch": 4.407407407407407, "grad_norm": 1.001423716545105, "learning_rate": 7.121699080689394e-06, "loss": 0.3457, "step": 119 }, { "epoch": 4.444444444444445, "grad_norm": 1.1704602241516113, "learning_rate": 6.882869649042397e-06, "loss": 0.4105, "step": 120 }, { "epoch": 4.481481481481482, "grad_norm": 1.072298288345337, "learning_rate": 6.657710714341364e-06, "loss": 0.3461, "step": 121 }, { "epoch": 4.518518518518518, "grad_norm": 0.9239315986633301, "learning_rate": 6.446371571588896e-06, "loss": 0.3633, "step": 122 }, { "epoch": 4.555555555555555, "grad_norm": 0.8906875252723694, "learning_rate": 6.248992352367622e-06, "loss": 0.3273, "step": 123 }, { "epoch": 4.592592592592593, "grad_norm": 0.9496861696243286, "learning_rate": 6.065703931923894e-06, "loss": 0.391, "step": 124 }, { "epoch": 4.62962962962963, "grad_norm": 0.9626538753509521, "learning_rate": 5.896627842388961e-06, "loss": 0.3506, "step": 125 }, { "epoch": 4.666666666666667, "grad_norm": 0.8809230327606201, "learning_rate": 5.741876192195292e-06, "loss": 0.2992, "step": 126 }, { "epoch": 4.703703703703704, "grad_norm": 1.044419288635254, "learning_rate": 5.601551591741394e-06, "loss": 0.3594, "step": 127 }, { "epoch": 4.7407407407407405, "grad_norm": 1.0764929056167603, "learning_rate": 5.47574708535445e-06, "loss": 0.3684, "step": 128 }, { "epoch": 4.777777777777778, "grad_norm": 1.0782520771026611, "learning_rate": 5.364546089595883e-06, "loss": 0.3323, "step": 129 }, { "epoch": 4.814814814814815, "grad_norm": 0.9487962126731873, "learning_rate": 5.268022337950767e-06, "loss": 0.3182, "step": 130 }, { "epoch": 4.851851851851852, "grad_norm": 1.0227851867675781, "learning_rate": 5.186239831937717e-06, "loss": 0.3416, "step": 131 }, { "epoch": 4.888888888888889, "grad_norm": 1.0411008596420288, "learning_rate": 5.119252798671747e-06, "loss": 0.3446, "step": 132 }, { "epoch": 4.925925925925926, "grad_norm": 0.8823951482772827, "learning_rate": 5.0671056549081495e-06, "loss": 0.3355, "step": 133 }, { "epoch": 4.962962962962963, "grad_norm": 0.8825088739395142, "learning_rate": 5.029832977591314e-06, "loss": 0.3553, "step": 134 }, { "epoch": 5.0, "grad_norm": 1.035265326499939, "learning_rate": 5.007459480927957e-06, "loss": 0.3773, "step": 135 } ], "logging_steps": 1, "max_steps": 135, "num_input_tokens_seen": 0, "num_train_epochs": 5, "save_steps": 27, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 2.2332169934340096e+17, "train_batch_size": 2, "trial_name": null, "trial_params": null }