| { | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 3.9741935483870967, | |
| "eval_steps": 500, | |
| "global_step": 308, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.025806451612903226, | |
| "grad_norm": 3.5142855644226074, | |
| "learning_rate": 0.00019870129870129872, | |
| "loss": 8.9214, | |
| "step": 2 | |
| }, | |
| { | |
| "epoch": 0.05161290322580645, | |
| "grad_norm": 3.772717237472534, | |
| "learning_rate": 0.00019740259740259742, | |
| "loss": 8.3405, | |
| "step": 4 | |
| }, | |
| { | |
| "epoch": 0.07741935483870968, | |
| "grad_norm": 6.7029924392700195, | |
| "learning_rate": 0.00019610389610389613, | |
| "loss": 8.1379, | |
| "step": 6 | |
| }, | |
| { | |
| "epoch": 0.1032258064516129, | |
| "grad_norm": 3.7924954891204834, | |
| "learning_rate": 0.0001948051948051948, | |
| "loss": 5.913, | |
| "step": 8 | |
| }, | |
| { | |
| "epoch": 0.12903225806451613, | |
| "grad_norm": 4.214364051818848, | |
| "learning_rate": 0.00019350649350649354, | |
| "loss": 7.0624, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.15483870967741936, | |
| "grad_norm": 3.0830295085906982, | |
| "learning_rate": 0.00019220779220779222, | |
| "loss": 6.6031, | |
| "step": 12 | |
| }, | |
| { | |
| "epoch": 0.18064516129032257, | |
| "grad_norm": 2.293950080871582, | |
| "learning_rate": 0.00019090909090909092, | |
| "loss": 6.1543, | |
| "step": 14 | |
| }, | |
| { | |
| "epoch": 0.2064516129032258, | |
| "grad_norm": 2.8121275901794434, | |
| "learning_rate": 0.00018961038961038963, | |
| "loss": 5.6952, | |
| "step": 16 | |
| }, | |
| { | |
| "epoch": 0.23225806451612904, | |
| "grad_norm": 3.2042553424835205, | |
| "learning_rate": 0.00018831168831168833, | |
| "loss": 5.6916, | |
| "step": 18 | |
| }, | |
| { | |
| "epoch": 0.25806451612903225, | |
| "grad_norm": 3.4788341522216797, | |
| "learning_rate": 0.000187012987012987, | |
| "loss": 5.575, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.2838709677419355, | |
| "grad_norm": 1.1461117267608643, | |
| "learning_rate": 0.00018571428571428572, | |
| "loss": 5.7922, | |
| "step": 22 | |
| }, | |
| { | |
| "epoch": 0.3096774193548387, | |
| "grad_norm": 1.7745267152786255, | |
| "learning_rate": 0.00018441558441558442, | |
| "loss": 5.5945, | |
| "step": 24 | |
| }, | |
| { | |
| "epoch": 0.33548387096774196, | |
| "grad_norm": 1.0893975496292114, | |
| "learning_rate": 0.00018311688311688313, | |
| "loss": 6.1593, | |
| "step": 26 | |
| }, | |
| { | |
| "epoch": 0.36129032258064514, | |
| "grad_norm": 2.169384241104126, | |
| "learning_rate": 0.00018181818181818183, | |
| "loss": 5.8243, | |
| "step": 28 | |
| }, | |
| { | |
| "epoch": 0.3870967741935484, | |
| "grad_norm": 1.8578031063079834, | |
| "learning_rate": 0.00018051948051948054, | |
| "loss": 5.5041, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.4129032258064516, | |
| "grad_norm": 1.448111653327942, | |
| "learning_rate": 0.00017922077922077922, | |
| "loss": 5.3471, | |
| "step": 32 | |
| }, | |
| { | |
| "epoch": 0.43870967741935485, | |
| "grad_norm": 1.828591227531433, | |
| "learning_rate": 0.00017792207792207792, | |
| "loss": 5.0863, | |
| "step": 34 | |
| }, | |
| { | |
| "epoch": 0.4645161290322581, | |
| "grad_norm": 4.83806848526001, | |
| "learning_rate": 0.00017662337662337663, | |
| "loss": 6.5823, | |
| "step": 36 | |
| }, | |
| { | |
| "epoch": 0.49032258064516127, | |
| "grad_norm": 2.1270275115966797, | |
| "learning_rate": 0.00017532467532467534, | |
| "loss": 5.3268, | |
| "step": 38 | |
| }, | |
| { | |
| "epoch": 0.5161290322580645, | |
| "grad_norm": 1.42393958568573, | |
| "learning_rate": 0.00017402597402597401, | |
| "loss": 5.4203, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.5419354838709678, | |
| "grad_norm": 2.768852472305298, | |
| "learning_rate": 0.00017272727272727275, | |
| "loss": 5.3403, | |
| "step": 42 | |
| }, | |
| { | |
| "epoch": 0.567741935483871, | |
| "grad_norm": 0.7816085815429688, | |
| "learning_rate": 0.00017142857142857143, | |
| "loss": 5.6736, | |
| "step": 44 | |
| }, | |
| { | |
| "epoch": 0.5935483870967742, | |
| "grad_norm": 0.8368009924888611, | |
| "learning_rate": 0.00017012987012987013, | |
| "loss": 5.4177, | |
| "step": 46 | |
| }, | |
| { | |
| "epoch": 0.6193548387096774, | |
| "grad_norm": 1.7966578006744385, | |
| "learning_rate": 0.00016883116883116884, | |
| "loss": 5.8706, | |
| "step": 48 | |
| }, | |
| { | |
| "epoch": 0.6451612903225806, | |
| "grad_norm": 0.694948673248291, | |
| "learning_rate": 0.00016753246753246754, | |
| "loss": 5.2186, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.6709677419354839, | |
| "grad_norm": 2.0851364135742188, | |
| "learning_rate": 0.00016623376623376625, | |
| "loss": 5.452, | |
| "step": 52 | |
| }, | |
| { | |
| "epoch": 0.6967741935483871, | |
| "grad_norm": 1.705665946006775, | |
| "learning_rate": 0.00016493506493506495, | |
| "loss": 5.5373, | |
| "step": 54 | |
| }, | |
| { | |
| "epoch": 0.7225806451612903, | |
| "grad_norm": 1.5386658906936646, | |
| "learning_rate": 0.00016363636363636366, | |
| "loss": 5.8638, | |
| "step": 56 | |
| }, | |
| { | |
| "epoch": 0.7483870967741936, | |
| "grad_norm": 1.3625737428665161, | |
| "learning_rate": 0.00016233766233766234, | |
| "loss": 5.2913, | |
| "step": 58 | |
| }, | |
| { | |
| "epoch": 0.7741935483870968, | |
| "grad_norm": 1.357426643371582, | |
| "learning_rate": 0.00016103896103896104, | |
| "loss": 5.7317, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "grad_norm": 1.1882686614990234, | |
| "learning_rate": 0.00015974025974025975, | |
| "loss": 5.9749, | |
| "step": 62 | |
| }, | |
| { | |
| "epoch": 0.8258064516129032, | |
| "grad_norm": 1.0777816772460938, | |
| "learning_rate": 0.00015844155844155845, | |
| "loss": 5.4791, | |
| "step": 64 | |
| }, | |
| { | |
| "epoch": 0.8516129032258064, | |
| "grad_norm": 0.7732660174369812, | |
| "learning_rate": 0.00015714285714285716, | |
| "loss": 5.9732, | |
| "step": 66 | |
| }, | |
| { | |
| "epoch": 0.8774193548387097, | |
| "grad_norm": 1.0730479955673218, | |
| "learning_rate": 0.00015584415584415587, | |
| "loss": 5.4338, | |
| "step": 68 | |
| }, | |
| { | |
| "epoch": 0.9032258064516129, | |
| "grad_norm": 1.4834214448928833, | |
| "learning_rate": 0.00015454545454545454, | |
| "loss": 5.1802, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 0.9290322580645162, | |
| "grad_norm": 2.2870066165924072, | |
| "learning_rate": 0.00015324675324675325, | |
| "loss": 6.3573, | |
| "step": 72 | |
| }, | |
| { | |
| "epoch": 0.9548387096774194, | |
| "grad_norm": 0.935196578502655, | |
| "learning_rate": 0.00015194805194805196, | |
| "loss": 5.0633, | |
| "step": 74 | |
| }, | |
| { | |
| "epoch": 0.9806451612903225, | |
| "grad_norm": 0.994687020778656, | |
| "learning_rate": 0.00015064935064935066, | |
| "loss": 5.8048, | |
| "step": 76 | |
| }, | |
| { | |
| "epoch": 1.0064516129032257, | |
| "grad_norm": 2.274411201477051, | |
| "learning_rate": 0.00014935064935064934, | |
| "loss": 4.6835, | |
| "step": 78 | |
| }, | |
| { | |
| "epoch": 1.032258064516129, | |
| "grad_norm": 1.2067323923110962, | |
| "learning_rate": 0.00014805194805194807, | |
| "loss": 5.4713, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 1.0580645161290323, | |
| "grad_norm": 0.9473636150360107, | |
| "learning_rate": 0.00014675324675324675, | |
| "loss": 5.8508, | |
| "step": 82 | |
| }, | |
| { | |
| "epoch": 1.0838709677419356, | |
| "grad_norm": 5.7191996574401855, | |
| "learning_rate": 0.00014545454545454546, | |
| "loss": 4.9011, | |
| "step": 84 | |
| }, | |
| { | |
| "epoch": 1.1096774193548387, | |
| "grad_norm": 2.201740026473999, | |
| "learning_rate": 0.00014415584415584416, | |
| "loss": 5.4717, | |
| "step": 86 | |
| }, | |
| { | |
| "epoch": 1.135483870967742, | |
| "grad_norm": 1.0441229343414307, | |
| "learning_rate": 0.00014285714285714287, | |
| "loss": 5.5206, | |
| "step": 88 | |
| }, | |
| { | |
| "epoch": 1.1612903225806452, | |
| "grad_norm": 1.0501593351364136, | |
| "learning_rate": 0.00014155844155844155, | |
| "loss": 5.2199, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 1.1870967741935483, | |
| "grad_norm": 1.1907446384429932, | |
| "learning_rate": 0.00014025974025974028, | |
| "loss": 4.8319, | |
| "step": 92 | |
| }, | |
| { | |
| "epoch": 1.2129032258064516, | |
| "grad_norm": 4.133155345916748, | |
| "learning_rate": 0.00013896103896103896, | |
| "loss": 5.1049, | |
| "step": 94 | |
| }, | |
| { | |
| "epoch": 1.238709677419355, | |
| "grad_norm": 1.399916172027588, | |
| "learning_rate": 0.00013766233766233766, | |
| "loss": 5.5704, | |
| "step": 96 | |
| }, | |
| { | |
| "epoch": 1.2645161290322582, | |
| "grad_norm": 1.0567469596862793, | |
| "learning_rate": 0.00013636363636363637, | |
| "loss": 5.3263, | |
| "step": 98 | |
| }, | |
| { | |
| "epoch": 1.2903225806451613, | |
| "grad_norm": 1.2196253538131714, | |
| "learning_rate": 0.00013506493506493507, | |
| "loss": 5.2236, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 1.3161290322580645, | |
| "grad_norm": 2.465505838394165, | |
| "learning_rate": 0.00013376623376623375, | |
| "loss": 5.0547, | |
| "step": 102 | |
| }, | |
| { | |
| "epoch": 1.3419354838709676, | |
| "grad_norm": 0.9022129774093628, | |
| "learning_rate": 0.00013246753246753249, | |
| "loss": 5.8387, | |
| "step": 104 | |
| }, | |
| { | |
| "epoch": 1.367741935483871, | |
| "grad_norm": 1.6524704694747925, | |
| "learning_rate": 0.0001311688311688312, | |
| "loss": 5.4351, | |
| "step": 106 | |
| }, | |
| { | |
| "epoch": 1.3935483870967742, | |
| "grad_norm": 1.2133516073226929, | |
| "learning_rate": 0.00012987012987012987, | |
| "loss": 5.4708, | |
| "step": 108 | |
| }, | |
| { | |
| "epoch": 1.4193548387096775, | |
| "grad_norm": 1.019903302192688, | |
| "learning_rate": 0.00012857142857142858, | |
| "loss": 5.8655, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 1.4451612903225808, | |
| "grad_norm": 1.1822901964187622, | |
| "learning_rate": 0.00012727272727272728, | |
| "loss": 5.3089, | |
| "step": 112 | |
| }, | |
| { | |
| "epoch": 1.4709677419354839, | |
| "grad_norm": 0.8840041160583496, | |
| "learning_rate": 0.000125974025974026, | |
| "loss": 5.0657, | |
| "step": 114 | |
| }, | |
| { | |
| "epoch": 1.4967741935483871, | |
| "grad_norm": 0.6745492219924927, | |
| "learning_rate": 0.00012467532467532467, | |
| "loss": 4.8004, | |
| "step": 116 | |
| }, | |
| { | |
| "epoch": 1.5225806451612902, | |
| "grad_norm": 1.4586549997329712, | |
| "learning_rate": 0.0001233766233766234, | |
| "loss": 5.0572, | |
| "step": 118 | |
| }, | |
| { | |
| "epoch": 1.5483870967741935, | |
| "grad_norm": 1.5153924226760864, | |
| "learning_rate": 0.00012207792207792208, | |
| "loss": 4.4382, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 1.5741935483870968, | |
| "grad_norm": 0.9981886148452759, | |
| "learning_rate": 0.0001207792207792208, | |
| "loss": 4.8087, | |
| "step": 122 | |
| }, | |
| { | |
| "epoch": 1.6, | |
| "grad_norm": 0.7403278350830078, | |
| "learning_rate": 0.00011948051948051949, | |
| "loss": 5.2025, | |
| "step": 124 | |
| }, | |
| { | |
| "epoch": 1.6258064516129034, | |
| "grad_norm": 1.1075459718704224, | |
| "learning_rate": 0.0001181818181818182, | |
| "loss": 4.946, | |
| "step": 126 | |
| }, | |
| { | |
| "epoch": 1.6516129032258065, | |
| "grad_norm": 0.6300385594367981, | |
| "learning_rate": 0.00011688311688311689, | |
| "loss": 5.6817, | |
| "step": 128 | |
| }, | |
| { | |
| "epoch": 1.6774193548387095, | |
| "grad_norm": 0.8701953291893005, | |
| "learning_rate": 0.00011558441558441559, | |
| "loss": 5.6561, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 1.7032258064516128, | |
| "grad_norm": 1.133817434310913, | |
| "learning_rate": 0.00011428571428571428, | |
| "loss": 5.6528, | |
| "step": 132 | |
| }, | |
| { | |
| "epoch": 1.729032258064516, | |
| "grad_norm": 1.6048352718353271, | |
| "learning_rate": 0.000112987012987013, | |
| "loss": 4.8602, | |
| "step": 134 | |
| }, | |
| { | |
| "epoch": 1.7548387096774194, | |
| "grad_norm": 1.2783055305480957, | |
| "learning_rate": 0.00011168831168831168, | |
| "loss": 4.1662, | |
| "step": 136 | |
| }, | |
| { | |
| "epoch": 1.7806451612903227, | |
| "grad_norm": 1.6137133836746216, | |
| "learning_rate": 0.0001103896103896104, | |
| "loss": 5.2295, | |
| "step": 138 | |
| }, | |
| { | |
| "epoch": 1.8064516129032258, | |
| "grad_norm": 0.6374461054801941, | |
| "learning_rate": 0.00010909090909090909, | |
| "loss": 5.3992, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 1.832258064516129, | |
| "grad_norm": 0.8323061466217041, | |
| "learning_rate": 0.0001077922077922078, | |
| "loss": 5.1712, | |
| "step": 142 | |
| }, | |
| { | |
| "epoch": 1.8580645161290321, | |
| "grad_norm": 0.6804484724998474, | |
| "learning_rate": 0.00010649350649350649, | |
| "loss": 5.3369, | |
| "step": 144 | |
| }, | |
| { | |
| "epoch": 1.8838709677419354, | |
| "grad_norm": 0.9051455855369568, | |
| "learning_rate": 0.0001051948051948052, | |
| "loss": 5.3726, | |
| "step": 146 | |
| }, | |
| { | |
| "epoch": 1.9096774193548387, | |
| "grad_norm": 0.5890415906906128, | |
| "learning_rate": 0.00010389610389610389, | |
| "loss": 5.4903, | |
| "step": 148 | |
| }, | |
| { | |
| "epoch": 1.935483870967742, | |
| "grad_norm": 0.5433252453804016, | |
| "learning_rate": 0.00010259740259740261, | |
| "loss": 5.4437, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 1.9612903225806453, | |
| "grad_norm": 1.353121042251587, | |
| "learning_rate": 0.0001012987012987013, | |
| "loss": 4.1137, | |
| "step": 152 | |
| }, | |
| { | |
| "epoch": 1.9870967741935484, | |
| "grad_norm": 1.0199609994888306, | |
| "learning_rate": 0.0001, | |
| "loss": 5.2558, | |
| "step": 154 | |
| }, | |
| { | |
| "epoch": 2.0129032258064514, | |
| "grad_norm": 0.6021209359169006, | |
| "learning_rate": 9.870129870129871e-05, | |
| "loss": 5.5338, | |
| "step": 156 | |
| }, | |
| { | |
| "epoch": 2.0387096774193547, | |
| "grad_norm": 0.9102515578269958, | |
| "learning_rate": 9.74025974025974e-05, | |
| "loss": 4.5988, | |
| "step": 158 | |
| }, | |
| { | |
| "epoch": 2.064516129032258, | |
| "grad_norm": 0.7101506590843201, | |
| "learning_rate": 9.610389610389611e-05, | |
| "loss": 5.2622, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 2.0903225806451613, | |
| "grad_norm": 0.631308913230896, | |
| "learning_rate": 9.480519480519481e-05, | |
| "loss": 5.4718, | |
| "step": 162 | |
| }, | |
| { | |
| "epoch": 2.1161290322580646, | |
| "grad_norm": 0.7462102770805359, | |
| "learning_rate": 9.35064935064935e-05, | |
| "loss": 5.0117, | |
| "step": 164 | |
| }, | |
| { | |
| "epoch": 2.141935483870968, | |
| "grad_norm": 0.6367154717445374, | |
| "learning_rate": 9.220779220779221e-05, | |
| "loss": 4.9968, | |
| "step": 166 | |
| }, | |
| { | |
| "epoch": 2.167741935483871, | |
| "grad_norm": 0.6959227323532104, | |
| "learning_rate": 9.090909090909092e-05, | |
| "loss": 5.6022, | |
| "step": 168 | |
| }, | |
| { | |
| "epoch": 2.193548387096774, | |
| "grad_norm": 0.6119377613067627, | |
| "learning_rate": 8.961038961038961e-05, | |
| "loss": 4.9405, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 2.2193548387096773, | |
| "grad_norm": 0.6411863565444946, | |
| "learning_rate": 8.831168831168831e-05, | |
| "loss": 5.1413, | |
| "step": 172 | |
| }, | |
| { | |
| "epoch": 2.2451612903225806, | |
| "grad_norm": 1.615324854850769, | |
| "learning_rate": 8.701298701298701e-05, | |
| "loss": 4.7308, | |
| "step": 174 | |
| }, | |
| { | |
| "epoch": 2.270967741935484, | |
| "grad_norm": 0.5708146691322327, | |
| "learning_rate": 8.571428571428571e-05, | |
| "loss": 5.4497, | |
| "step": 176 | |
| }, | |
| { | |
| "epoch": 2.296774193548387, | |
| "grad_norm": 0.8313891291618347, | |
| "learning_rate": 8.441558441558442e-05, | |
| "loss": 5.3687, | |
| "step": 178 | |
| }, | |
| { | |
| "epoch": 2.3225806451612905, | |
| "grad_norm": 1.1731419563293457, | |
| "learning_rate": 8.311688311688312e-05, | |
| "loss": 5.3314, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 2.3483870967741938, | |
| "grad_norm": 0.7381497025489807, | |
| "learning_rate": 8.181818181818183e-05, | |
| "loss": 5.1178, | |
| "step": 182 | |
| }, | |
| { | |
| "epoch": 2.3741935483870966, | |
| "grad_norm": 0.6883618831634521, | |
| "learning_rate": 8.051948051948052e-05, | |
| "loss": 5.9266, | |
| "step": 184 | |
| }, | |
| { | |
| "epoch": 2.4, | |
| "grad_norm": 0.735080361366272, | |
| "learning_rate": 7.922077922077923e-05, | |
| "loss": 4.7453, | |
| "step": 186 | |
| }, | |
| { | |
| "epoch": 2.425806451612903, | |
| "grad_norm": 0.6184589862823486, | |
| "learning_rate": 7.792207792207793e-05, | |
| "loss": 5.0924, | |
| "step": 188 | |
| }, | |
| { | |
| "epoch": 2.4516129032258065, | |
| "grad_norm": 0.6305899024009705, | |
| "learning_rate": 7.662337662337662e-05, | |
| "loss": 5.1167, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 2.47741935483871, | |
| "grad_norm": 0.815334677696228, | |
| "learning_rate": 7.532467532467533e-05, | |
| "loss": 5.4719, | |
| "step": 192 | |
| }, | |
| { | |
| "epoch": 2.5032258064516126, | |
| "grad_norm": 0.6805923581123352, | |
| "learning_rate": 7.402597402597404e-05, | |
| "loss": 4.4392, | |
| "step": 194 | |
| }, | |
| { | |
| "epoch": 2.5290322580645164, | |
| "grad_norm": 0.6633741855621338, | |
| "learning_rate": 7.272727272727273e-05, | |
| "loss": 5.1926, | |
| "step": 196 | |
| }, | |
| { | |
| "epoch": 2.554838709677419, | |
| "grad_norm": 0.6042450666427612, | |
| "learning_rate": 7.142857142857143e-05, | |
| "loss": 4.7079, | |
| "step": 198 | |
| }, | |
| { | |
| "epoch": 2.5806451612903225, | |
| "grad_norm": 0.5892207622528076, | |
| "learning_rate": 7.012987012987014e-05, | |
| "loss": 4.9938, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 2.606451612903226, | |
| "grad_norm": 0.6753908395767212, | |
| "learning_rate": 6.883116883116883e-05, | |
| "loss": 5.0097, | |
| "step": 202 | |
| }, | |
| { | |
| "epoch": 2.632258064516129, | |
| "grad_norm": 0.5345973968505859, | |
| "learning_rate": 6.753246753246754e-05, | |
| "loss": 5.1074, | |
| "step": 204 | |
| }, | |
| { | |
| "epoch": 2.6580645161290324, | |
| "grad_norm": 0.4924313724040985, | |
| "learning_rate": 6.623376623376624e-05, | |
| "loss": 5.0817, | |
| "step": 206 | |
| }, | |
| { | |
| "epoch": 2.6838709677419352, | |
| "grad_norm": 0.5989976525306702, | |
| "learning_rate": 6.493506493506494e-05, | |
| "loss": 4.2039, | |
| "step": 208 | |
| }, | |
| { | |
| "epoch": 2.709677419354839, | |
| "grad_norm": 1.0730255842208862, | |
| "learning_rate": 6.363636363636364e-05, | |
| "loss": 4.2224, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 2.735483870967742, | |
| "grad_norm": 0.991038978099823, | |
| "learning_rate": 6.233766233766233e-05, | |
| "loss": 4.8869, | |
| "step": 212 | |
| }, | |
| { | |
| "epoch": 2.761290322580645, | |
| "grad_norm": 0.5952357053756714, | |
| "learning_rate": 6.103896103896104e-05, | |
| "loss": 5.1753, | |
| "step": 214 | |
| }, | |
| { | |
| "epoch": 2.7870967741935484, | |
| "grad_norm": 0.6798732876777649, | |
| "learning_rate": 5.9740259740259744e-05, | |
| "loss": 4.3396, | |
| "step": 216 | |
| }, | |
| { | |
| "epoch": 2.8129032258064517, | |
| "grad_norm": 0.5361295342445374, | |
| "learning_rate": 5.844155844155844e-05, | |
| "loss": 5.2857, | |
| "step": 218 | |
| }, | |
| { | |
| "epoch": 2.838709677419355, | |
| "grad_norm": 0.6326772570610046, | |
| "learning_rate": 5.714285714285714e-05, | |
| "loss": 4.7385, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 2.864516129032258, | |
| "grad_norm": 0.639589786529541, | |
| "learning_rate": 5.584415584415584e-05, | |
| "loss": 4.3803, | |
| "step": 222 | |
| }, | |
| { | |
| "epoch": 2.8903225806451616, | |
| "grad_norm": 0.7248474955558777, | |
| "learning_rate": 5.4545454545454546e-05, | |
| "loss": 4.6668, | |
| "step": 224 | |
| }, | |
| { | |
| "epoch": 2.9161290322580644, | |
| "grad_norm": 0.7551538348197937, | |
| "learning_rate": 5.3246753246753245e-05, | |
| "loss": 5.016, | |
| "step": 226 | |
| }, | |
| { | |
| "epoch": 2.9419354838709677, | |
| "grad_norm": 0.4990728497505188, | |
| "learning_rate": 5.1948051948051944e-05, | |
| "loss": 5.0843, | |
| "step": 228 | |
| }, | |
| { | |
| "epoch": 2.967741935483871, | |
| "grad_norm": 0.5729503035545349, | |
| "learning_rate": 5.064935064935065e-05, | |
| "loss": 4.573, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 2.9935483870967743, | |
| "grad_norm": 1.4705737829208374, | |
| "learning_rate": 4.9350649350649355e-05, | |
| "loss": 4.9812, | |
| "step": 232 | |
| }, | |
| { | |
| "epoch": 3.0193548387096776, | |
| "grad_norm": 0.7053755521774292, | |
| "learning_rate": 4.8051948051948054e-05, | |
| "loss": 4.7651, | |
| "step": 234 | |
| }, | |
| { | |
| "epoch": 3.0451612903225804, | |
| "grad_norm": 0.6120907664299011, | |
| "learning_rate": 4.675324675324675e-05, | |
| "loss": 5.2057, | |
| "step": 236 | |
| }, | |
| { | |
| "epoch": 3.0709677419354837, | |
| "grad_norm": 0.6173492074012756, | |
| "learning_rate": 4.545454545454546e-05, | |
| "loss": 5.0342, | |
| "step": 238 | |
| }, | |
| { | |
| "epoch": 3.096774193548387, | |
| "grad_norm": 0.5435605049133301, | |
| "learning_rate": 4.415584415584416e-05, | |
| "loss": 4.9538, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 3.1225806451612903, | |
| "grad_norm": 1.4004778861999512, | |
| "learning_rate": 4.2857142857142856e-05, | |
| "loss": 4.3856, | |
| "step": 242 | |
| }, | |
| { | |
| "epoch": 3.1483870967741936, | |
| "grad_norm": 1.1300957202911377, | |
| "learning_rate": 4.155844155844156e-05, | |
| "loss": 5.0085, | |
| "step": 244 | |
| }, | |
| { | |
| "epoch": 3.174193548387097, | |
| "grad_norm": 0.6826758980751038, | |
| "learning_rate": 4.025974025974026e-05, | |
| "loss": 5.3525, | |
| "step": 246 | |
| }, | |
| { | |
| "epoch": 3.2, | |
| "grad_norm": 0.6162336468696594, | |
| "learning_rate": 3.8961038961038966e-05, | |
| "loss": 5.1639, | |
| "step": 248 | |
| }, | |
| { | |
| "epoch": 3.225806451612903, | |
| "grad_norm": 0.7551366686820984, | |
| "learning_rate": 3.7662337662337665e-05, | |
| "loss": 5.1913, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 3.2516129032258063, | |
| "grad_norm": 0.5899360775947571, | |
| "learning_rate": 3.6363636363636364e-05, | |
| "loss": 5.022, | |
| "step": 252 | |
| }, | |
| { | |
| "epoch": 3.2774193548387096, | |
| "grad_norm": 0.6666110157966614, | |
| "learning_rate": 3.506493506493507e-05, | |
| "loss": 4.697, | |
| "step": 254 | |
| }, | |
| { | |
| "epoch": 3.303225806451613, | |
| "grad_norm": 0.5686184167861938, | |
| "learning_rate": 3.376623376623377e-05, | |
| "loss": 4.8758, | |
| "step": 256 | |
| }, | |
| { | |
| "epoch": 3.329032258064516, | |
| "grad_norm": 0.7564727663993835, | |
| "learning_rate": 3.246753246753247e-05, | |
| "loss": 4.9468, | |
| "step": 258 | |
| }, | |
| { | |
| "epoch": 3.3548387096774195, | |
| "grad_norm": 0.5773691534996033, | |
| "learning_rate": 3.1168831168831166e-05, | |
| "loss": 5.1122, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 3.3806451612903228, | |
| "grad_norm": 0.642393171787262, | |
| "learning_rate": 2.9870129870129872e-05, | |
| "loss": 5.233, | |
| "step": 262 | |
| }, | |
| { | |
| "epoch": 3.4064516129032256, | |
| "grad_norm": 0.6513245701789856, | |
| "learning_rate": 2.857142857142857e-05, | |
| "loss": 4.2779, | |
| "step": 264 | |
| }, | |
| { | |
| "epoch": 3.432258064516129, | |
| "grad_norm": 0.8267136812210083, | |
| "learning_rate": 2.7272727272727273e-05, | |
| "loss": 5.2998, | |
| "step": 266 | |
| }, | |
| { | |
| "epoch": 3.458064516129032, | |
| "grad_norm": 0.5494163632392883, | |
| "learning_rate": 2.5974025974025972e-05, | |
| "loss": 5.4667, | |
| "step": 268 | |
| }, | |
| { | |
| "epoch": 3.4838709677419355, | |
| "grad_norm": 0.7057967782020569, | |
| "learning_rate": 2.4675324675324678e-05, | |
| "loss": 5.1893, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 3.509677419354839, | |
| "grad_norm": 0.5400364398956299, | |
| "learning_rate": 2.3376623376623376e-05, | |
| "loss": 4.9972, | |
| "step": 272 | |
| }, | |
| { | |
| "epoch": 3.535483870967742, | |
| "grad_norm": 0.4732670485973358, | |
| "learning_rate": 2.207792207792208e-05, | |
| "loss": 4.896, | |
| "step": 274 | |
| }, | |
| { | |
| "epoch": 3.5612903225806454, | |
| "grad_norm": 0.5432953834533691, | |
| "learning_rate": 2.077922077922078e-05, | |
| "loss": 5.0452, | |
| "step": 276 | |
| }, | |
| { | |
| "epoch": 3.587096774193548, | |
| "grad_norm": 0.5461270213127136, | |
| "learning_rate": 1.9480519480519483e-05, | |
| "loss": 4.7124, | |
| "step": 278 | |
| }, | |
| { | |
| "epoch": 3.6129032258064515, | |
| "grad_norm": 0.6231604218482971, | |
| "learning_rate": 1.8181818181818182e-05, | |
| "loss": 4.3152, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 3.638709677419355, | |
| "grad_norm": 0.42820078134536743, | |
| "learning_rate": 1.6883116883116884e-05, | |
| "loss": 4.9029, | |
| "step": 282 | |
| }, | |
| { | |
| "epoch": 3.664516129032258, | |
| "grad_norm": 0.7605751752853394, | |
| "learning_rate": 1.5584415584415583e-05, | |
| "loss": 4.8563, | |
| "step": 284 | |
| }, | |
| { | |
| "epoch": 3.6903225806451614, | |
| "grad_norm": 0.8210684061050415, | |
| "learning_rate": 1.4285714285714285e-05, | |
| "loss": 3.7461, | |
| "step": 286 | |
| }, | |
| { | |
| "epoch": 3.7161290322580647, | |
| "grad_norm": 0.6016200184822083, | |
| "learning_rate": 1.2987012987012986e-05, | |
| "loss": 4.7194, | |
| "step": 288 | |
| }, | |
| { | |
| "epoch": 3.741935483870968, | |
| "grad_norm": 0.7171183228492737, | |
| "learning_rate": 1.1688311688311688e-05, | |
| "loss": 4.5957, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 3.767741935483871, | |
| "grad_norm": 0.6558433175086975, | |
| "learning_rate": 1.038961038961039e-05, | |
| "loss": 4.5994, | |
| "step": 292 | |
| }, | |
| { | |
| "epoch": 3.793548387096774, | |
| "grad_norm": 0.5295835733413696, | |
| "learning_rate": 9.090909090909091e-06, | |
| "loss": 5.5671, | |
| "step": 294 | |
| }, | |
| { | |
| "epoch": 3.8193548387096774, | |
| "grad_norm": 0.6055201292037964, | |
| "learning_rate": 7.792207792207792e-06, | |
| "loss": 5.0694, | |
| "step": 296 | |
| }, | |
| { | |
| "epoch": 3.8451612903225807, | |
| "grad_norm": 0.6041186451911926, | |
| "learning_rate": 6.493506493506493e-06, | |
| "loss": 5.0007, | |
| "step": 298 | |
| }, | |
| { | |
| "epoch": 3.870967741935484, | |
| "grad_norm": 0.7281818985939026, | |
| "learning_rate": 5.194805194805195e-06, | |
| "loss": 4.3153, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 3.896774193548387, | |
| "grad_norm": 0.4925851821899414, | |
| "learning_rate": 3.896103896103896e-06, | |
| "loss": 5.5631, | |
| "step": 302 | |
| }, | |
| { | |
| "epoch": 3.9225806451612906, | |
| "grad_norm": 0.6916934847831726, | |
| "learning_rate": 2.5974025974025976e-06, | |
| "loss": 5.3932, | |
| "step": 304 | |
| }, | |
| { | |
| "epoch": 3.9483870967741934, | |
| "grad_norm": 0.6561426520347595, | |
| "learning_rate": 1.2987012987012988e-06, | |
| "loss": 5.3373, | |
| "step": 306 | |
| }, | |
| { | |
| "epoch": 3.9741935483870967, | |
| "grad_norm": 0.6102042198181152, | |
| "learning_rate": 0.0, | |
| "loss": 5.1064, | |
| "step": 308 | |
| }, | |
| { | |
| "epoch": 3.9741935483870967, | |
| "step": 308, | |
| "total_flos": 723765039226044.0, | |
| "train_loss": 5.270187915145577, | |
| "train_runtime": 746.2778, | |
| "train_samples_per_second": 3.318, | |
| "train_steps_per_second": 0.413 | |
| } | |
| ], | |
| "logging_steps": 2, | |
| "max_steps": 308, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 4, | |
| "save_steps": 500, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": true | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 723765039226044.0, | |
| "train_batch_size": 2, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |