| { | |
| "best_global_step": null, | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 1.3386243386243386, | |
| "eval_steps": 500, | |
| "global_step": 80, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.016931216931216932, | |
| "grad_norm": 2322.343994140625, | |
| "learning_rate": 0.0, | |
| "loss": 1.2269, | |
| "step": 1 | |
| }, | |
| { | |
| "epoch": 0.033862433862433865, | |
| "grad_norm": 12.033977508544922, | |
| "learning_rate": 3.3333333333333333e-06, | |
| "loss": 1.2059, | |
| "step": 2 | |
| }, | |
| { | |
| "epoch": 0.050793650793650794, | |
| "grad_norm": 10.809624671936035, | |
| "learning_rate": 6.666666666666667e-06, | |
| "loss": 1.2198, | |
| "step": 3 | |
| }, | |
| { | |
| "epoch": 0.06772486772486773, | |
| "grad_norm": 5.77828311920166, | |
| "learning_rate": 1e-05, | |
| "loss": 1.0599, | |
| "step": 4 | |
| }, | |
| { | |
| "epoch": 0.08465608465608465, | |
| "grad_norm": 5.281270503997803, | |
| "learning_rate": 1.3333333333333333e-05, | |
| "loss": 0.9754, | |
| "step": 5 | |
| }, | |
| { | |
| "epoch": 0.10158730158730159, | |
| "grad_norm": 5.149123191833496, | |
| "learning_rate": 1.6666666666666667e-05, | |
| "loss": 0.9476, | |
| "step": 6 | |
| }, | |
| { | |
| "epoch": 0.11851851851851852, | |
| "grad_norm": 6.636379718780518, | |
| "learning_rate": 2e-05, | |
| "loss": 1.0585, | |
| "step": 7 | |
| }, | |
| { | |
| "epoch": 0.13544973544973546, | |
| "grad_norm": 3.8481030464172363, | |
| "learning_rate": 1.9998370105646414e-05, | |
| "loss": 0.9413, | |
| "step": 8 | |
| }, | |
| { | |
| "epoch": 0.1523809523809524, | |
| "grad_norm": 3.835693120956421, | |
| "learning_rate": 1.999348095389677e-05, | |
| "loss": 0.8891, | |
| "step": 9 | |
| }, | |
| { | |
| "epoch": 0.1693121693121693, | |
| "grad_norm": 2.8596138954162598, | |
| "learning_rate": 1.998533413851124e-05, | |
| "loss": 0.8795, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.18624338624338624, | |
| "grad_norm": 2.8256287574768066, | |
| "learning_rate": 1.9973932315179502e-05, | |
| "loss": 0.8446, | |
| "step": 11 | |
| }, | |
| { | |
| "epoch": 0.20317460317460317, | |
| "grad_norm": 5.52194356918335, | |
| "learning_rate": 1.9959279200655044e-05, | |
| "loss": 0.8027, | |
| "step": 12 | |
| }, | |
| { | |
| "epoch": 0.2201058201058201, | |
| "grad_norm": 2.0485758781433105, | |
| "learning_rate": 1.9941379571543597e-05, | |
| "loss": 0.7859, | |
| "step": 13 | |
| }, | |
| { | |
| "epoch": 0.23703703703703705, | |
| "grad_norm": 2.2072129249572754, | |
| "learning_rate": 1.9920239262746045e-05, | |
| "loss": 0.7766, | |
| "step": 14 | |
| }, | |
| { | |
| "epoch": 0.25396825396825395, | |
| "grad_norm": 2.154557943344116, | |
| "learning_rate": 1.9895865165556375e-05, | |
| "loss": 0.7366, | |
| "step": 15 | |
| }, | |
| { | |
| "epoch": 0.2708994708994709, | |
| "grad_norm": 1.9130762815475464, | |
| "learning_rate": 1.9868265225415263e-05, | |
| "loss": 0.7432, | |
| "step": 16 | |
| }, | |
| { | |
| "epoch": 0.2878306878306878, | |
| "grad_norm": 1.6504809856414795, | |
| "learning_rate": 1.9837448439320027e-05, | |
| "loss": 0.6905, | |
| "step": 17 | |
| }, | |
| { | |
| "epoch": 0.3047619047619048, | |
| "grad_norm": 1.7797411680221558, | |
| "learning_rate": 1.9803424852891803e-05, | |
| "loss": 0.6799, | |
| "step": 18 | |
| }, | |
| { | |
| "epoch": 0.3216931216931217, | |
| "grad_norm": 1.7259820699691772, | |
| "learning_rate": 1.976620555710087e-05, | |
| "loss": 0.6706, | |
| "step": 19 | |
| }, | |
| { | |
| "epoch": 0.3386243386243386, | |
| "grad_norm": 1.4618961811065674, | |
| "learning_rate": 1.9725802684651235e-05, | |
| "loss": 0.6687, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.35555555555555557, | |
| "grad_norm": 1.5048692226409912, | |
| "learning_rate": 1.9682229406025635e-05, | |
| "loss": 0.6762, | |
| "step": 21 | |
| }, | |
| { | |
| "epoch": 0.3724867724867725, | |
| "grad_norm": 1.4504609107971191, | |
| "learning_rate": 1.963549992519223e-05, | |
| "loss": 0.6631, | |
| "step": 22 | |
| }, | |
| { | |
| "epoch": 0.38941798941798944, | |
| "grad_norm": 1.4441351890563965, | |
| "learning_rate": 1.9585629474974413e-05, | |
| "loss": 0.6394, | |
| "step": 23 | |
| }, | |
| { | |
| "epoch": 0.40634920634920635, | |
| "grad_norm": 1.2873355150222778, | |
| "learning_rate": 1.953263431208523e-05, | |
| "loss": 0.6279, | |
| "step": 24 | |
| }, | |
| { | |
| "epoch": 0.42328042328042326, | |
| "grad_norm": 1.4676573276519775, | |
| "learning_rate": 1.9476531711828027e-05, | |
| "loss": 0.6238, | |
| "step": 25 | |
| }, | |
| { | |
| "epoch": 0.4402116402116402, | |
| "grad_norm": 1.3487250804901123, | |
| "learning_rate": 1.9417339962465084e-05, | |
| "loss": 0.6253, | |
| "step": 26 | |
| }, | |
| { | |
| "epoch": 0.45714285714285713, | |
| "grad_norm": 23.1207275390625, | |
| "learning_rate": 1.935507835925601e-05, | |
| "loss": 0.6166, | |
| "step": 27 | |
| }, | |
| { | |
| "epoch": 0.4740740740740741, | |
| "grad_norm": 1.408158540725708, | |
| "learning_rate": 1.9289767198167918e-05, | |
| "loss": 0.5644, | |
| "step": 28 | |
| }, | |
| { | |
| "epoch": 0.491005291005291, | |
| "grad_norm": 1.199977159500122, | |
| "learning_rate": 1.9221427769259333e-05, | |
| "loss": 0.5865, | |
| "step": 29 | |
| }, | |
| { | |
| "epoch": 0.5079365079365079, | |
| "grad_norm": 1.1895148754119873, | |
| "learning_rate": 1.9150082349740123e-05, | |
| "loss": 0.5619, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.5248677248677248, | |
| "grad_norm": 1.1904869079589844, | |
| "learning_rate": 1.9075754196709574e-05, | |
| "loss": 0.5642, | |
| "step": 31 | |
| }, | |
| { | |
| "epoch": 0.5417989417989418, | |
| "grad_norm": 1.4911994934082031, | |
| "learning_rate": 1.899846753957507e-05, | |
| "loss": 0.5684, | |
| "step": 32 | |
| }, | |
| { | |
| "epoch": 0.5587301587301587, | |
| "grad_norm": 1.2807954549789429, | |
| "learning_rate": 1.8918247572153822e-05, | |
| "loss": 0.532, | |
| "step": 33 | |
| }, | |
| { | |
| "epoch": 0.5756613756613757, | |
| "grad_norm": 1.1882894039154053, | |
| "learning_rate": 1.883512044446023e-05, | |
| "loss": 0.5527, | |
| "step": 34 | |
| }, | |
| { | |
| "epoch": 0.5925925925925926, | |
| "grad_norm": 1.3212125301361084, | |
| "learning_rate": 1.8749113254181498e-05, | |
| "loss": 0.5621, | |
| "step": 35 | |
| }, | |
| { | |
| "epoch": 0.6095238095238096, | |
| "grad_norm": 1.4091153144836426, | |
| "learning_rate": 1.866025403784439e-05, | |
| "loss": 0.5205, | |
| "step": 36 | |
| }, | |
| { | |
| "epoch": 0.6264550264550265, | |
| "grad_norm": 1.1614974737167358, | |
| "learning_rate": 1.8568571761675893e-05, | |
| "loss": 0.5181, | |
| "step": 37 | |
| }, | |
| { | |
| "epoch": 0.6433862433862434, | |
| "grad_norm": 1.0888781547546387, | |
| "learning_rate": 1.8474096312160866e-05, | |
| "loss": 0.5359, | |
| "step": 38 | |
| }, | |
| { | |
| "epoch": 0.6603174603174603, | |
| "grad_norm": 1.303357720375061, | |
| "learning_rate": 1.837685848629965e-05, | |
| "loss": 0.5256, | |
| "step": 39 | |
| }, | |
| { | |
| "epoch": 0.6772486772486772, | |
| "grad_norm": 1.1001622676849365, | |
| "learning_rate": 1.827688998156891e-05, | |
| "loss": 0.5277, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.6941798941798942, | |
| "grad_norm": 16.82403564453125, | |
| "learning_rate": 1.817422338558892e-05, | |
| "loss": 0.5301, | |
| "step": 41 | |
| }, | |
| { | |
| "epoch": 0.7111111111111111, | |
| "grad_norm": 1.2562721967697144, | |
| "learning_rate": 1.8068892165500704e-05, | |
| "loss": 0.5019, | |
| "step": 42 | |
| }, | |
| { | |
| "epoch": 0.728042328042328, | |
| "grad_norm": 1.143244981765747, | |
| "learning_rate": 1.796093065705644e-05, | |
| "loss": 0.5098, | |
| "step": 43 | |
| }, | |
| { | |
| "epoch": 0.744973544973545, | |
| "grad_norm": 1.2346585988998413, | |
| "learning_rate": 1.7850374053426725e-05, | |
| "loss": 0.5079, | |
| "step": 44 | |
| }, | |
| { | |
| "epoch": 0.7619047619047619, | |
| "grad_norm": 1.292527675628662, | |
| "learning_rate": 1.7737258393728363e-05, | |
| "loss": 0.4987, | |
| "step": 45 | |
| }, | |
| { | |
| "epoch": 0.7788359788359789, | |
| "grad_norm": 1.0964988470077515, | |
| "learning_rate": 1.7621620551276366e-05, | |
| "loss": 0.52, | |
| "step": 46 | |
| }, | |
| { | |
| "epoch": 0.7957671957671958, | |
| "grad_norm": 1.3093878030776978, | |
| "learning_rate": 1.7503498221564026e-05, | |
| "loss": 0.4911, | |
| "step": 47 | |
| }, | |
| { | |
| "epoch": 0.8126984126984127, | |
| "grad_norm": 1.2538305521011353, | |
| "learning_rate": 1.7382929909974988e-05, | |
| "loss": 0.4805, | |
| "step": 48 | |
| }, | |
| { | |
| "epoch": 0.8296296296296296, | |
| "grad_norm": 1.5297569036483765, | |
| "learning_rate": 1.725995491923131e-05, | |
| "loss": 0.4834, | |
| "step": 49 | |
| }, | |
| { | |
| "epoch": 0.8465608465608465, | |
| "grad_norm": 1.126209020614624, | |
| "learning_rate": 1.7134613336581602e-05, | |
| "loss": 0.474, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.8634920634920635, | |
| "grad_norm": 1.2999428510665894, | |
| "learning_rate": 1.7006946020733426e-05, | |
| "loss": 0.4662, | |
| "step": 51 | |
| }, | |
| { | |
| "epoch": 0.8804232804232804, | |
| "grad_norm": 3.900702714920044, | |
| "learning_rate": 1.6876994588534234e-05, | |
| "loss": 0.4728, | |
| "step": 52 | |
| }, | |
| { | |
| "epoch": 0.8973544973544973, | |
| "grad_norm": 1.1961653232574463, | |
| "learning_rate": 1.6744801401405138e-05, | |
| "loss": 0.4638, | |
| "step": 53 | |
| }, | |
| { | |
| "epoch": 0.9142857142857143, | |
| "grad_norm": 1.1492871046066284, | |
| "learning_rate": 1.6610409551532006e-05, | |
| "loss": 0.4608, | |
| "step": 54 | |
| }, | |
| { | |
| "epoch": 0.9312169312169312, | |
| "grad_norm": 1.1923617124557495, | |
| "learning_rate": 1.647386284781828e-05, | |
| "loss": 0.4742, | |
| "step": 55 | |
| }, | |
| { | |
| "epoch": 0.9481481481481482, | |
| "grad_norm": 1.0757859945297241, | |
| "learning_rate": 1.6335205801604242e-05, | |
| "loss": 0.4757, | |
| "step": 56 | |
| }, | |
| { | |
| "epoch": 0.9650793650793651, | |
| "grad_norm": 1.0793986320495605, | |
| "learning_rate": 1.6194483612157232e-05, | |
| "loss": 0.4543, | |
| "step": 57 | |
| }, | |
| { | |
| "epoch": 0.982010582010582, | |
| "grad_norm": 1.0309922695159912, | |
| "learning_rate": 1.6051742151937655e-05, | |
| "loss": 0.4637, | |
| "step": 58 | |
| }, | |
| { | |
| "epoch": 0.9989417989417989, | |
| "grad_norm": 1.0178518295288086, | |
| "learning_rate": 1.590702795164551e-05, | |
| "loss": 0.4845, | |
| "step": 59 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "grad_norm": 1.0178518295288086, | |
| "learning_rate": 1.57603881850524e-05, | |
| "loss": 0.4151, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 1.016931216931217, | |
| "grad_norm": 6.670537948608398, | |
| "learning_rate": 1.5611870653623826e-05, | |
| "loss": 0.4583, | |
| "step": 61 | |
| }, | |
| { | |
| "epoch": 1.0338624338624338, | |
| "grad_norm": 1.1750141382217407, | |
| "learning_rate": 1.546152377093697e-05, | |
| "loss": 0.4378, | |
| "step": 62 | |
| }, | |
| { | |
| "epoch": 1.0507936507936508, | |
| "grad_norm": 1.258583903312683, | |
| "learning_rate": 1.530939654689887e-05, | |
| "loss": 0.4698, | |
| "step": 63 | |
| }, | |
| { | |
| "epoch": 1.0677248677248676, | |
| "grad_norm": 1.1863641738891602, | |
| "learning_rate": 1.515553857177022e-05, | |
| "loss": 0.4493, | |
| "step": 64 | |
| }, | |
| { | |
| "epoch": 1.0846560846560847, | |
| "grad_norm": 1.2087653875350952, | |
| "learning_rate": 1.5000000000000002e-05, | |
| "loss": 0.4495, | |
| "step": 65 | |
| }, | |
| { | |
| "epoch": 1.1015873015873017, | |
| "grad_norm": 1.1390602588653564, | |
| "learning_rate": 1.4842831533876196e-05, | |
| "loss": 0.4628, | |
| "step": 66 | |
| }, | |
| { | |
| "epoch": 1.1185185185185185, | |
| "grad_norm": 1.0523955821990967, | |
| "learning_rate": 1.4684084406997903e-05, | |
| "loss": 0.4444, | |
| "step": 67 | |
| }, | |
| { | |
| "epoch": 1.1354497354497355, | |
| "grad_norm": 1.3320904970169067, | |
| "learning_rate": 1.4523810367574271e-05, | |
| "loss": 0.4487, | |
| "step": 68 | |
| }, | |
| { | |
| "epoch": 1.1523809523809523, | |
| "grad_norm": 2.979928970336914, | |
| "learning_rate": 1.4362061661555675e-05, | |
| "loss": 0.4568, | |
| "step": 69 | |
| }, | |
| { | |
| "epoch": 1.1693121693121693, | |
| "grad_norm": 1.2313185930252075, | |
| "learning_rate": 1.4198891015602648e-05, | |
| "loss": 0.4359, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 1.1862433862433863, | |
| "grad_norm": 1.1036418676376343, | |
| "learning_rate": 1.4034351619898088e-05, | |
| "loss": 0.4698, | |
| "step": 71 | |
| }, | |
| { | |
| "epoch": 1.2031746031746031, | |
| "grad_norm": 1.1762608289718628, | |
| "learning_rate": 1.3868497110808394e-05, | |
| "loss": 0.4313, | |
| "step": 72 | |
| }, | |
| { | |
| "epoch": 1.2201058201058201, | |
| "grad_norm": 1.033484697341919, | |
| "learning_rate": 1.3701381553399147e-05, | |
| "loss": 0.45, | |
| "step": 73 | |
| }, | |
| { | |
| "epoch": 1.237037037037037, | |
| "grad_norm": 2.2186121940612793, | |
| "learning_rate": 1.3533059423811026e-05, | |
| "loss": 0.4345, | |
| "step": 74 | |
| }, | |
| { | |
| "epoch": 1.253968253968254, | |
| "grad_norm": 1.0926917791366577, | |
| "learning_rate": 1.3363585591501751e-05, | |
| "loss": 0.4317, | |
| "step": 75 | |
| }, | |
| { | |
| "epoch": 1.270899470899471, | |
| "grad_norm": 1.094223976135254, | |
| "learning_rate": 1.31930153013598e-05, | |
| "loss": 0.4274, | |
| "step": 76 | |
| }, | |
| { | |
| "epoch": 1.2878306878306878, | |
| "grad_norm": 1.0197666883468628, | |
| "learning_rate": 1.3021404155695728e-05, | |
| "loss": 0.4465, | |
| "step": 77 | |
| }, | |
| { | |
| "epoch": 1.3047619047619048, | |
| "grad_norm": 1.0113344192504883, | |
| "learning_rate": 1.2848808096117003e-05, | |
| "loss": 0.4256, | |
| "step": 78 | |
| }, | |
| { | |
| "epoch": 1.3216931216931216, | |
| "grad_norm": 0.9916861653327942, | |
| "learning_rate": 1.2675283385292212e-05, | |
| "loss": 0.4291, | |
| "step": 79 | |
| }, | |
| { | |
| "epoch": 1.3386243386243386, | |
| "grad_norm": 1.0909327268600464, | |
| "learning_rate": 1.250088658861063e-05, | |
| "loss": 0.4387, | |
| "step": 80 | |
| } | |
| ], | |
| "logging_steps": 1.0, | |
| "max_steps": 180, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 3, | |
| "save_steps": 5, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": false | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 5.636674904683184e+18, | |
| "train_batch_size": 2, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |