{ "best_global_step": null, "best_metric": null, "best_model_checkpoint": null, "epoch": 3.0, "eval_steps": 500, "global_step": 441, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.034013605442176874, "grad_norm": 2.7028610706329346, "learning_rate": 4.998414279387743e-05, "loss": 6.0793, "num_input_tokens_seen": 3168, "step": 5 }, { "epoch": 0.06802721088435375, "grad_norm": 2.967193603515625, "learning_rate": 4.9936591291588586e-05, "loss": 5.4017, "num_input_tokens_seen": 6544, "step": 10 }, { "epoch": 0.10204081632653061, "grad_norm": 1.1446479558944702, "learning_rate": 4.985740581585134e-05, "loss": 4.3102, "num_input_tokens_seen": 10464, "step": 15 }, { "epoch": 0.1360544217687075, "grad_norm": 1.7880395650863647, "learning_rate": 4.9746686819498546e-05, "loss": 4.232, "num_input_tokens_seen": 13296, "step": 20 }, { "epoch": 0.17006802721088435, "grad_norm": 2.237711191177368, "learning_rate": 4.960457475804594e-05, "loss": 3.9065, "num_input_tokens_seen": 16320, "step": 25 }, { "epoch": 0.20408163265306123, "grad_norm": 1.7026973962783813, "learning_rate": 4.94312499115136e-05, "loss": 3.8348, "num_input_tokens_seen": 19504, "step": 30 }, { "epoch": 0.23809523809523808, "grad_norm": 1.4949020147323608, "learning_rate": 4.922693215572695e-05, "loss": 3.5761, "num_input_tokens_seen": 22704, "step": 35 }, { "epoch": 0.272108843537415, "grad_norm": 2.234846830368042, "learning_rate": 4.899188068338743e-05, "loss": 3.7531, "num_input_tokens_seen": 25728, "step": 40 }, { "epoch": 0.30612244897959184, "grad_norm": 1.7678470611572266, "learning_rate": 4.8726393675266716e-05, "loss": 3.5748, "num_input_tokens_seen": 29024, "step": 45 }, { "epoch": 0.3401360544217687, "grad_norm": 1.8059829473495483, "learning_rate": 4.84308079219417e-05, "loss": 3.772, "num_input_tokens_seen": 32464, "step": 50 }, { "epoch": 0.3741496598639456, "grad_norm": 2.300635576248169, "learning_rate": 4.810549839654973e-05, "loss": 3.807, "num_input_tokens_seen": 35424, "step": 55 }, { "epoch": 0.40816326530612246, "grad_norm": 2.1998870372772217, "learning_rate": 4.7750877779106666e-05, "loss": 3.5115, "num_input_tokens_seen": 38208, "step": 60 }, { "epoch": 0.4421768707482993, "grad_norm": 1.6165343523025513, "learning_rate": 4.736739593299058e-05, "loss": 3.7613, "num_input_tokens_seen": 41280, "step": 65 }, { "epoch": 0.47619047619047616, "grad_norm": 2.3959059715270996, "learning_rate": 4.6955539334255716e-05, "loss": 3.6662, "num_input_tokens_seen": 44144, "step": 70 }, { "epoch": 0.5102040816326531, "grad_norm": 1.5641539096832275, "learning_rate": 4.651583045450041e-05, "loss": 3.627, "num_input_tokens_seen": 47120, "step": 75 }, { "epoch": 0.54421768707483, "grad_norm": 1.9732143878936768, "learning_rate": 4.604882709807187e-05, "loss": 3.6502, "num_input_tokens_seen": 50272, "step": 80 }, { "epoch": 0.5782312925170068, "grad_norm": 2.6501049995422363, "learning_rate": 4.5555121694448735e-05, "loss": 3.7163, "num_input_tokens_seen": 53056, "step": 85 }, { "epoch": 0.6122448979591837, "grad_norm": 1.9555385112762451, "learning_rate": 4.503534054669892e-05, "loss": 3.6891, "num_input_tokens_seen": 55968, "step": 90 }, { "epoch": 0.6462585034013606, "grad_norm": 1.728500485420227, "learning_rate": 4.44901430369663e-05, "loss": 3.6063, "num_input_tokens_seen": 58944, "step": 95 }, { "epoch": 0.6802721088435374, "grad_norm": 1.6137477159500122, "learning_rate": 4.392022078999405e-05, "loss": 3.5177, "num_input_tokens_seen": 61872, "step": 100 }, { "epoch": 0.7142857142857143, "grad_norm": 1.7164677381515503, "learning_rate": 4.332629679574566e-05, "loss": 3.376, "num_input_tokens_seen": 64704, "step": 105 }, { "epoch": 0.7482993197278912, "grad_norm": 1.568602442741394, "learning_rate": 4.270912449223699e-05, "loss": 3.4565, "num_input_tokens_seen": 67776, "step": 110 }, { "epoch": 0.782312925170068, "grad_norm": 2.5001213550567627, "learning_rate": 4.206948680974242e-05, "loss": 3.6683, "num_input_tokens_seen": 70784, "step": 115 }, { "epoch": 0.8163265306122449, "grad_norm": 2.0534141063690186, "learning_rate": 4.140819517758795e-05, "loss": 3.5794, "num_input_tokens_seen": 74336, "step": 120 }, { "epoch": 0.8503401360544217, "grad_norm": 2.1421139240264893, "learning_rate": 4.072608849479106e-05, "loss": 3.6307, "num_input_tokens_seen": 77232, "step": 125 }, { "epoch": 0.8843537414965986, "grad_norm": 1.939579963684082, "learning_rate": 4.002403206585307e-05, "loss": 3.5575, "num_input_tokens_seen": 80352, "step": 130 }, { "epoch": 0.9183673469387755, "grad_norm": 2.4437503814697266, "learning_rate": 3.9302916503054246e-05, "loss": 3.5487, "num_input_tokens_seen": 84016, "step": 135 }, { "epoch": 0.9523809523809523, "grad_norm": 2.416694164276123, "learning_rate": 3.856365659664399e-05, "loss": 3.6761, "num_input_tokens_seen": 87120, "step": 140 }, { "epoch": 0.9863945578231292, "grad_norm": 2.3587541580200195, "learning_rate": 3.780719015435943e-05, "loss": 3.1846, "num_input_tokens_seen": 90544, "step": 145 }, { "epoch": 1.0204081632653061, "grad_norm": 2.34908390045166, "learning_rate": 3.703447681174458e-05, "loss": 3.2783, "num_input_tokens_seen": 94712, "step": 150 }, { "epoch": 1.054421768707483, "grad_norm": 2.1918578147888184, "learning_rate": 3.624649681477923e-05, "loss": 3.494, "num_input_tokens_seen": 97912, "step": 155 }, { "epoch": 1.08843537414966, "grad_norm": 2.6867949962615967, "learning_rate": 3.544424977636198e-05, "loss": 3.2551, "num_input_tokens_seen": 100696, "step": 160 }, { "epoch": 1.1224489795918366, "grad_norm": 3.030991315841675, "learning_rate": 3.4628753408224765e-05, "loss": 3.4821, "num_input_tokens_seen": 104552, "step": 165 }, { "epoch": 1.1564625850340136, "grad_norm": 2.9311370849609375, "learning_rate": 3.3801042229887756e-05, "loss": 3.4083, "num_input_tokens_seen": 107800, "step": 170 }, { "epoch": 1.1904761904761905, "grad_norm": 2.682126760482788, "learning_rate": 3.2962166256292113e-05, "loss": 3.4593, "num_input_tokens_seen": 110840, "step": 175 }, { "epoch": 1.2244897959183674, "grad_norm": 2.8780510425567627, "learning_rate": 3.211318966577581e-05, "loss": 3.5256, "num_input_tokens_seen": 113640, "step": 180 }, { "epoch": 1.2585034013605443, "grad_norm": 2.4468441009521484, "learning_rate": 3.1255189450081977e-05, "loss": 2.9337, "num_input_tokens_seen": 117752, "step": 185 }, { "epoch": 1.2925170068027212, "grad_norm": 2.369248628616333, "learning_rate": 3.0389254048112493e-05, "loss": 3.2217, "num_input_tokens_seen": 121512, "step": 190 }, { "epoch": 1.3265306122448979, "grad_norm": 1.5555424690246582, "learning_rate": 2.9516481965159975e-05, "loss": 3.4546, "num_input_tokens_seen": 124840, "step": 195 }, { "epoch": 1.3605442176870748, "grad_norm": 2.8310885429382324, "learning_rate": 2.863798037936983e-05, "loss": 3.559, "num_input_tokens_seen": 127960, "step": 200 }, { "epoch": 1.3945578231292517, "grad_norm": 2.4837112426757812, "learning_rate": 2.775486373720003e-05, "loss": 3.2782, "num_input_tokens_seen": 130968, "step": 205 }, { "epoch": 1.4285714285714286, "grad_norm": 2.7569165229797363, "learning_rate": 2.686825233966061e-05, "loss": 3.4279, "num_input_tokens_seen": 133736, "step": 210 }, { "epoch": 1.4625850340136055, "grad_norm": 2.901097297668457, "learning_rate": 2.597927092112607e-05, "loss": 3.3297, "num_input_tokens_seen": 136552, "step": 215 }, { "epoch": 1.4965986394557822, "grad_norm": 2.221858263015747, "learning_rate": 2.5089047222523838e-05, "loss": 3.2559, "num_input_tokens_seen": 139832, "step": 220 }, { "epoch": 1.5306122448979593, "grad_norm": 2.864755153656006, "learning_rate": 2.419871056070862e-05, "loss": 3.1332, "num_input_tokens_seen": 142840, "step": 225 }, { "epoch": 1.564625850340136, "grad_norm": 2.6603639125823975, "learning_rate": 2.3309390395837633e-05, "loss": 3.4912, "num_input_tokens_seen": 145704, "step": 230 }, { "epoch": 1.598639455782313, "grad_norm": 1.888742446899414, "learning_rate": 2.2422214898563916e-05, "loss": 3.4511, "num_input_tokens_seen": 148632, "step": 235 }, { "epoch": 1.6326530612244898, "grad_norm": 2.04915714263916, "learning_rate": 2.1538309518865646e-05, "loss": 3.4876, "num_input_tokens_seen": 151608, "step": 240 }, { "epoch": 1.6666666666666665, "grad_norm": 2.1959643363952637, "learning_rate": 2.0658795558326743e-05, "loss": 3.474, "num_input_tokens_seen": 154936, "step": 245 }, { "epoch": 1.7006802721088436, "grad_norm": 2.2664616107940674, "learning_rate": 1.9784788747679982e-05, "loss": 3.2019, "num_input_tokens_seen": 157912, "step": 250 }, { "epoch": 1.7346938775510203, "grad_norm": 2.486283540725708, "learning_rate": 1.8917397831417286e-05, "loss": 3.2353, "num_input_tokens_seen": 160536, "step": 255 }, { "epoch": 1.7687074829931972, "grad_norm": 2.9121577739715576, "learning_rate": 1.80577231612625e-05, "loss": 3.3723, "num_input_tokens_seen": 163976, "step": 260 }, { "epoch": 1.8027210884353742, "grad_norm": 3.1636898517608643, "learning_rate": 1.720685530029105e-05, "loss": 3.3386, "num_input_tokens_seen": 167144, "step": 265 }, { "epoch": 1.836734693877551, "grad_norm": 3.5455474853515625, "learning_rate": 1.6365873639467315e-05, "loss": 3.4695, "num_input_tokens_seen": 169752, "step": 270 }, { "epoch": 1.870748299319728, "grad_norm": 2.9391326904296875, "learning_rate": 1.553584502835456e-05, "loss": 3.0469, "num_input_tokens_seen": 173416, "step": 275 }, { "epoch": 1.9047619047619047, "grad_norm": 2.9568731784820557, "learning_rate": 1.4717822421734718e-05, "loss": 3.3184, "num_input_tokens_seen": 176392, "step": 280 }, { "epoch": 1.9387755102040818, "grad_norm": 3.0478389263153076, "learning_rate": 1.3912843543854664e-05, "loss": 3.3918, "num_input_tokens_seen": 179272, "step": 285 }, { "epoch": 1.9727891156462585, "grad_norm": 2.8810529708862305, "learning_rate": 1.3121929571993685e-05, "loss": 3.5476, "num_input_tokens_seen": 182616, "step": 290 }, { "epoch": 2.006802721088435, "grad_norm": 2.8993630409240723, "learning_rate": 1.2346083841021928e-05, "loss": 3.3684, "num_input_tokens_seen": 185424, "step": 295 }, { "epoch": 2.0408163265306123, "grad_norm": 1.4063067436218262, "learning_rate": 1.158629057059343e-05, "loss": 3.2909, "num_input_tokens_seen": 189040, "step": 300 }, { "epoch": 2.074829931972789, "grad_norm": 4.071662425994873, "learning_rate": 1.084351361658816e-05, "loss": 3.2623, "num_input_tokens_seen": 191856, "step": 305 }, { "epoch": 2.108843537414966, "grad_norm": 3.5538320541381836, "learning_rate": 1.0118695248387141e-05, "loss": 3.3487, "num_input_tokens_seen": 194720, "step": 310 }, { "epoch": 2.142857142857143, "grad_norm": 2.685102939605713, "learning_rate": 9.412754953531663e-06, "loss": 3.2409, "num_input_tokens_seen": 197808, "step": 315 }, { "epoch": 2.17687074829932, "grad_norm": 3.332947015762329, "learning_rate": 8.726588271282978e-06, "loss": 3.2031, "num_input_tokens_seen": 201040, "step": 320 }, { "epoch": 2.2108843537414966, "grad_norm": 3.082350492477417, "learning_rate": 8.061065656562269e-06, "loss": 2.9128, "num_input_tokens_seen": 204944, "step": 325 }, { "epoch": 2.2448979591836733, "grad_norm": 4.072040557861328, "learning_rate": 7.417031375712008e-06, "loss": 3.4445, "num_input_tokens_seen": 207712, "step": 330 }, { "epoch": 2.2789115646258504, "grad_norm": 3.686593770980835, "learning_rate": 6.795302435479523e-06, "loss": 3.0001, "num_input_tokens_seen": 210576, "step": 335 }, { "epoch": 2.312925170068027, "grad_norm": 2.2774229049682617, "learning_rate": 6.196667546581405e-06, "loss": 3.3204, "num_input_tokens_seen": 213712, "step": 340 }, { "epoch": 2.3469387755102042, "grad_norm": 3.1405270099639893, "learning_rate": 5.621886123163708e-06, "loss": 3.1907, "num_input_tokens_seen": 216464, "step": 345 }, { "epoch": 2.380952380952381, "grad_norm": 2.792036771774292, "learning_rate": 5.071687319426946e-06, "loss": 3.298, "num_input_tokens_seen": 219616, "step": 350 }, { "epoch": 2.4149659863945576, "grad_norm": 3.187742233276367, "learning_rate": 4.5467691046382435e-06, "loss": 3.2584, "num_input_tokens_seen": 222848, "step": 355 }, { "epoch": 2.4489795918367347, "grad_norm": 3.170606851577759, "learning_rate": 4.047797377703985e-06, "loss": 3.3844, "num_input_tokens_seen": 225568, "step": 360 }, { "epoch": 2.4829931972789114, "grad_norm": 3.8285155296325684, "learning_rate": 3.5754051224260176e-06, "loss": 3.3962, "num_input_tokens_seen": 228832, "step": 365 }, { "epoch": 2.5170068027210886, "grad_norm": 3.4845287799835205, "learning_rate": 3.130191604513352e-06, "loss": 3.1933, "num_input_tokens_seen": 231968, "step": 370 }, { "epoch": 2.5510204081632653, "grad_norm": 3.385293960571289, "learning_rate": 2.7127216113677635e-06, "loss": 3.1452, "num_input_tokens_seen": 235840, "step": 375 }, { "epoch": 2.5850340136054424, "grad_norm": 4.231153964996338, "learning_rate": 2.323524735607749e-06, "loss": 3.3513, "num_input_tokens_seen": 238768, "step": 380 }, { "epoch": 2.619047619047619, "grad_norm": 2.916943311691284, "learning_rate": 1.9630947032398067e-06, "loss": 2.9965, "num_input_tokens_seen": 242240, "step": 385 }, { "epoch": 2.6530612244897958, "grad_norm": 2.958503484725952, "learning_rate": 1.6318887473292243e-06, "loss": 3.1061, "num_input_tokens_seen": 245248, "step": 390 }, { "epoch": 2.687074829931973, "grad_norm": 2.5356791019439697, "learning_rate": 1.3303270279649477e-06, "loss": 3.0728, "num_input_tokens_seen": 248416, "step": 395 }, { "epoch": 2.7210884353741496, "grad_norm": 3.929884672164917, "learning_rate": 1.0587920992543853e-06, "loss": 3.1985, "num_input_tokens_seen": 251712, "step": 400 }, { "epoch": 2.7551020408163263, "grad_norm": 3.698723077774048, "learning_rate": 8.176284240242638e-07, "loss": 3.3142, "num_input_tokens_seen": 254496, "step": 405 }, { "epoch": 2.7891156462585034, "grad_norm": 3.604297637939453, "learning_rate": 6.071419368431779e-07, "loss": 3.4238, "num_input_tokens_seen": 257760, "step": 410 }, { "epoch": 2.8231292517006805, "grad_norm": 4.10955810546875, "learning_rate": 4.2759965592018283e-07, "loss": 3.3226, "num_input_tokens_seen": 260784, "step": 415 }, { "epoch": 2.857142857142857, "grad_norm": 3.1334779262542725, "learning_rate": 2.7922934437178695e-07, "loss": 3.2627, "num_input_tokens_seen": 263792, "step": 420 }, { "epoch": 2.891156462585034, "grad_norm": 3.3782756328582764, "learning_rate": 1.6221922128700217e-07, "loss": 3.2338, "num_input_tokens_seen": 266784, "step": 425 }, { "epoch": 2.925170068027211, "grad_norm": 2.9588587284088135, "learning_rate": 7.671772295704815e-08, "loss": 3.1991, "num_input_tokens_seen": 269936, "step": 430 }, { "epoch": 2.9591836734693877, "grad_norm": 3.1428213119506836, "learning_rate": 2.2833314572542895e-08, "loss": 3.3015, "num_input_tokens_seen": 272704, "step": 435 }, { "epoch": 2.9931972789115644, "grad_norm": 4.047586917877197, "learning_rate": 6.343526271379574e-10, "loss": 3.4811, "num_input_tokens_seen": 276208, "step": 440 }, { "epoch": 3.0, "num_input_tokens_seen": 276856, "step": 441, "total_flos": 2191990201638912.0, "train_loss": 3.4709230188339477, "train_runtime": 2415.5729, "train_samples_per_second": 2.92, "train_steps_per_second": 0.183 } ], "logging_steps": 5, "max_steps": 441, "num_input_tokens_seen": 276856, "num_train_epochs": 3, "save_steps": 100, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 2191990201638912.0, "train_batch_size": 2, "trial_name": null, "trial_params": null }