{ "best_metric": null, "best_model_checkpoint": null, "epoch": 0.5050505050505051, "eval_steps": 500, "global_step": 5000, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.005050505050505051, "grad_norm": 0.00640869140625, "learning_rate": 0.00019800000000000002, "loss": 1.2756, "step": 50 }, { "epoch": 0.010101010101010102, "grad_norm": 0.0008697509765625, "learning_rate": 0.000196, "loss": 0.9178, "step": 100 }, { "epoch": 0.015151515151515152, "grad_norm": 0.0002803802490234375, "learning_rate": 0.000194, "loss": 1.2995, "step": 150 }, { "epoch": 0.020202020202020204, "grad_norm": 63.75, "learning_rate": 0.000192, "loss": 1.0395, "step": 200 }, { "epoch": 0.025252525252525252, "grad_norm": 0.00970458984375, "learning_rate": 0.00019, "loss": 0.9509, "step": 250 }, { "epoch": 0.030303030303030304, "grad_norm": 0.00909423828125, "learning_rate": 0.000188, "loss": 0.7246, "step": 300 }, { "epoch": 0.03535353535353535, "grad_norm": 3.8125, "learning_rate": 0.00018600000000000002, "loss": 0.9811, "step": 350 }, { "epoch": 0.04040404040404041, "grad_norm": 0.00823974609375, "learning_rate": 0.00018400000000000003, "loss": 0.734, "step": 400 }, { "epoch": 0.045454545454545456, "grad_norm": 0.0203857421875, "learning_rate": 0.000182, "loss": 0.9583, "step": 450 }, { "epoch": 0.050505050505050504, "grad_norm": 84.0, "learning_rate": 0.00018, "loss": 0.7122, "step": 500 }, { "epoch": 0.05555555555555555, "grad_norm": 0.005859375, "learning_rate": 0.00017800000000000002, "loss": 0.6662, "step": 550 }, { "epoch": 0.06060606060606061, "grad_norm": 7.963180541992188e-05, "learning_rate": 0.00017600000000000002, "loss": 0.7423, "step": 600 }, { "epoch": 0.06565656565656566, "grad_norm": 4.5625, "learning_rate": 0.000174, "loss": 1.175, "step": 650 }, { "epoch": 0.0707070707070707, "grad_norm": 0.0014801025390625, "learning_rate": 0.000172, "loss": 0.6812, "step": 700 }, { "epoch": 0.07575757575757576, "grad_norm": 0.0634765625, "learning_rate": 0.00017, "loss": 0.5917, "step": 750 }, { "epoch": 0.08080808080808081, "grad_norm": 7.375, "learning_rate": 0.000168, "loss": 0.725, "step": 800 }, { "epoch": 0.08585858585858586, "grad_norm": 35.0, "learning_rate": 0.000166, "loss": 0.7658, "step": 850 }, { "epoch": 0.09090909090909091, "grad_norm": 6.625, "learning_rate": 0.000164, "loss": 0.7727, "step": 900 }, { "epoch": 0.09595959595959595, "grad_norm": 0.03759765625, "learning_rate": 0.000162, "loss": 0.5597, "step": 950 }, { "epoch": 0.10101010101010101, "grad_norm": 0.005096435546875, "learning_rate": 0.00016, "loss": 0.5977, "step": 1000 }, { "epoch": 0.10606060606060606, "grad_norm": 0.0037078857421875, "learning_rate": 0.00015800000000000002, "loss": 1.0553, "step": 1050 }, { "epoch": 0.1111111111111111, "grad_norm": 11.875, "learning_rate": 0.00015600000000000002, "loss": 0.7937, "step": 1100 }, { "epoch": 0.11616161616161616, "grad_norm": 0.00087738037109375, "learning_rate": 0.000154, "loss": 0.5609, "step": 1150 }, { "epoch": 0.12121212121212122, "grad_norm": 0.0003604888916015625, "learning_rate": 0.000152, "loss": 0.43, "step": 1200 }, { "epoch": 0.12626262626262627, "grad_norm": 0.00592041015625, "learning_rate": 0.00015000000000000001, "loss": 0.5442, "step": 1250 }, { "epoch": 0.13131313131313133, "grad_norm": 4.875, "learning_rate": 0.000148, "loss": 0.7502, "step": 1300 }, { "epoch": 0.13636363636363635, "grad_norm": 5.0, "learning_rate": 0.000146, "loss": 0.9968, "step": 1350 }, { "epoch": 0.1414141414141414, "grad_norm": 0.016357421875, "learning_rate": 0.000144, "loss": 0.5973, "step": 1400 }, { "epoch": 0.14646464646464646, "grad_norm": 0.92578125, "learning_rate": 0.000142, "loss": 1.0503, "step": 1450 }, { "epoch": 0.15151515151515152, "grad_norm": 0.003082275390625, "learning_rate": 0.00014, "loss": 0.526, "step": 1500 }, { "epoch": 0.15656565656565657, "grad_norm": 0.017333984375, "learning_rate": 0.000138, "loss": 0.6893, "step": 1550 }, { "epoch": 0.16161616161616163, "grad_norm": 5.125, "learning_rate": 0.00013600000000000003, "loss": 0.6892, "step": 1600 }, { "epoch": 0.16666666666666666, "grad_norm": 0.00186920166015625, "learning_rate": 0.000134, "loss": 0.5635, "step": 1650 }, { "epoch": 0.1717171717171717, "grad_norm": 4.125, "learning_rate": 0.000132, "loss": 0.7575, "step": 1700 }, { "epoch": 0.17676767676767677, "grad_norm": 27.375, "learning_rate": 0.00013000000000000002, "loss": 0.6113, "step": 1750 }, { "epoch": 0.18181818181818182, "grad_norm": 0.0019989013671875, "learning_rate": 0.00012800000000000002, "loss": 0.5127, "step": 1800 }, { "epoch": 0.18686868686868688, "grad_norm": 0.000514984130859375, "learning_rate": 0.000126, "loss": 0.6001, "step": 1850 }, { "epoch": 0.1919191919191919, "grad_norm": 0.00531005859375, "learning_rate": 0.000124, "loss": 0.6199, "step": 1900 }, { "epoch": 0.19696969696969696, "grad_norm": 0.486328125, "learning_rate": 0.000122, "loss": 0.8243, "step": 1950 }, { "epoch": 0.20202020202020202, "grad_norm": 0.00141143798828125, "learning_rate": 0.00012, "loss": 0.5386, "step": 2000 }, { "epoch": 0.20707070707070707, "grad_norm": 7.65625, "learning_rate": 0.000118, "loss": 0.4405, "step": 2050 }, { "epoch": 0.21212121212121213, "grad_norm": 9.625, "learning_rate": 0.000116, "loss": 0.7058, "step": 2100 }, { "epoch": 0.21717171717171718, "grad_norm": 11.9375, "learning_rate": 0.00011399999999999999, "loss": 0.6288, "step": 2150 }, { "epoch": 0.2222222222222222, "grad_norm": 0.4375, "learning_rate": 0.00011200000000000001, "loss": 0.5395, "step": 2200 }, { "epoch": 0.22727272727272727, "grad_norm": 2.671875, "learning_rate": 0.00011000000000000002, "loss": 0.8054, "step": 2250 }, { "epoch": 0.23232323232323232, "grad_norm": 0.0034332275390625, "learning_rate": 0.00010800000000000001, "loss": 0.5827, "step": 2300 }, { "epoch": 0.23737373737373738, "grad_norm": 0.0113525390625, "learning_rate": 0.00010600000000000002, "loss": 0.512, "step": 2350 }, { "epoch": 0.24242424242424243, "grad_norm": 2.6875, "learning_rate": 0.00010400000000000001, "loss": 0.6449, "step": 2400 }, { "epoch": 0.2474747474747475, "grad_norm": 8.875, "learning_rate": 0.00010200000000000001, "loss": 0.5955, "step": 2450 }, { "epoch": 0.25252525252525254, "grad_norm": 0.0003261566162109375, "learning_rate": 0.0001, "loss": 0.4586, "step": 2500 }, { "epoch": 0.25757575757575757, "grad_norm": 0.003814697265625, "learning_rate": 9.8e-05, "loss": 0.6287, "step": 2550 }, { "epoch": 0.26262626262626265, "grad_norm": 0.000823974609375, "learning_rate": 9.6e-05, "loss": 0.8459, "step": 2600 }, { "epoch": 0.2676767676767677, "grad_norm": 0.0208740234375, "learning_rate": 9.4e-05, "loss": 0.4845, "step": 2650 }, { "epoch": 0.2727272727272727, "grad_norm": 0.0003643035888671875, "learning_rate": 9.200000000000001e-05, "loss": 0.8213, "step": 2700 }, { "epoch": 0.2777777777777778, "grad_norm": 0.006927490234375, "learning_rate": 9e-05, "loss": 0.5009, "step": 2750 }, { "epoch": 0.2828282828282828, "grad_norm": 0.0078125, "learning_rate": 8.800000000000001e-05, "loss": 0.5776, "step": 2800 }, { "epoch": 0.2878787878787879, "grad_norm": 7.4375, "learning_rate": 8.6e-05, "loss": 0.5102, "step": 2850 }, { "epoch": 0.29292929292929293, "grad_norm": 10.1875, "learning_rate": 8.4e-05, "loss": 0.5194, "step": 2900 }, { "epoch": 0.29797979797979796, "grad_norm": 0.00469970703125, "learning_rate": 8.2e-05, "loss": 0.9075, "step": 2950 }, { "epoch": 0.30303030303030304, "grad_norm": 7.4375, "learning_rate": 8e-05, "loss": 0.6495, "step": 3000 }, { "epoch": 0.30808080808080807, "grad_norm": 33.25, "learning_rate": 7.800000000000001e-05, "loss": 0.3913, "step": 3050 }, { "epoch": 0.31313131313131315, "grad_norm": 0.00080108642578125, "learning_rate": 7.6e-05, "loss": 0.4946, "step": 3100 }, { "epoch": 0.3181818181818182, "grad_norm": 4.0625, "learning_rate": 7.4e-05, "loss": 0.5223, "step": 3150 }, { "epoch": 0.32323232323232326, "grad_norm": 0.0032806396484375, "learning_rate": 7.2e-05, "loss": 0.5745, "step": 3200 }, { "epoch": 0.3282828282828283, "grad_norm": 0.0004863739013671875, "learning_rate": 7e-05, "loss": 0.5042, "step": 3250 }, { "epoch": 0.3333333333333333, "grad_norm": 0.15625, "learning_rate": 6.800000000000001e-05, "loss": 0.5351, "step": 3300 }, { "epoch": 0.3383838383838384, "grad_norm": 12.125, "learning_rate": 6.6e-05, "loss": 0.3649, "step": 3350 }, { "epoch": 0.3434343434343434, "grad_norm": 0.02880859375, "learning_rate": 6.400000000000001e-05, "loss": 0.5516, "step": 3400 }, { "epoch": 0.3484848484848485, "grad_norm": 0.005828857421875, "learning_rate": 6.2e-05, "loss": 0.5899, "step": 3450 }, { "epoch": 0.35353535353535354, "grad_norm": 6.59375, "learning_rate": 6e-05, "loss": 0.4999, "step": 3500 }, { "epoch": 0.35858585858585856, "grad_norm": 0.0026397705078125, "learning_rate": 5.8e-05, "loss": 0.6304, "step": 3550 }, { "epoch": 0.36363636363636365, "grad_norm": 0.0013427734375, "learning_rate": 5.6000000000000006e-05, "loss": 0.7675, "step": 3600 }, { "epoch": 0.3686868686868687, "grad_norm": 14.375, "learning_rate": 5.4000000000000005e-05, "loss": 0.4504, "step": 3650 }, { "epoch": 0.37373737373737376, "grad_norm": 0.0235595703125, "learning_rate": 5.2000000000000004e-05, "loss": 0.4757, "step": 3700 }, { "epoch": 0.3787878787878788, "grad_norm": 0.0018157958984375, "learning_rate": 5e-05, "loss": 0.4315, "step": 3750 }, { "epoch": 0.3838383838383838, "grad_norm": 0.00012302398681640625, "learning_rate": 4.8e-05, "loss": 0.5263, "step": 3800 }, { "epoch": 0.3888888888888889, "grad_norm": 0.88671875, "learning_rate": 4.600000000000001e-05, "loss": 0.3134, "step": 3850 }, { "epoch": 0.3939393939393939, "grad_norm": 0.0245361328125, "learning_rate": 4.4000000000000006e-05, "loss": 0.3785, "step": 3900 }, { "epoch": 0.398989898989899, "grad_norm": 5.78125, "learning_rate": 4.2e-05, "loss": 0.4354, "step": 3950 }, { "epoch": 0.40404040404040403, "grad_norm": 0.000598907470703125, "learning_rate": 4e-05, "loss": 0.7522, "step": 4000 }, { "epoch": 0.4090909090909091, "grad_norm": 6.65625, "learning_rate": 3.8e-05, "loss": 0.5139, "step": 4050 }, { "epoch": 0.41414141414141414, "grad_norm": 4.9375, "learning_rate": 3.6e-05, "loss": 0.3349, "step": 4100 }, { "epoch": 0.41919191919191917, "grad_norm": 34.75, "learning_rate": 3.4000000000000007e-05, "loss": 0.85, "step": 4150 }, { "epoch": 0.42424242424242425, "grad_norm": 0.01953125, "learning_rate": 3.2000000000000005e-05, "loss": 0.4747, "step": 4200 }, { "epoch": 0.4292929292929293, "grad_norm": 0.000492095947265625, "learning_rate": 3e-05, "loss": 0.4613, "step": 4250 }, { "epoch": 0.43434343434343436, "grad_norm": 0.0079345703125, "learning_rate": 2.8000000000000003e-05, "loss": 0.66, "step": 4300 }, { "epoch": 0.4393939393939394, "grad_norm": 0.0013885498046875, "learning_rate": 2.6000000000000002e-05, "loss": 0.7939, "step": 4350 }, { "epoch": 0.4444444444444444, "grad_norm": 14.75, "learning_rate": 2.4e-05, "loss": 0.7121, "step": 4400 }, { "epoch": 0.4494949494949495, "grad_norm": 6.03125, "learning_rate": 2.2000000000000003e-05, "loss": 0.37, "step": 4450 }, { "epoch": 0.45454545454545453, "grad_norm": 1.6015625, "learning_rate": 2e-05, "loss": 0.5789, "step": 4500 }, { "epoch": 0.4595959595959596, "grad_norm": 0.0054931640625, "learning_rate": 1.8e-05, "loss": 0.3011, "step": 4550 }, { "epoch": 0.46464646464646464, "grad_norm": 0.00122833251953125, "learning_rate": 1.6000000000000003e-05, "loss": 0.4864, "step": 4600 }, { "epoch": 0.4696969696969697, "grad_norm": 9.9375, "learning_rate": 1.4000000000000001e-05, "loss": 0.9442, "step": 4650 }, { "epoch": 0.47474747474747475, "grad_norm": 0.040283203125, "learning_rate": 1.2e-05, "loss": 0.5764, "step": 4700 }, { "epoch": 0.4797979797979798, "grad_norm": 17.125, "learning_rate": 1e-05, "loss": 0.9053, "step": 4750 }, { "epoch": 0.48484848484848486, "grad_norm": 5.40625, "learning_rate": 8.000000000000001e-06, "loss": 0.455, "step": 4800 }, { "epoch": 0.4898989898989899, "grad_norm": 0.000843048095703125, "learning_rate": 6e-06, "loss": 0.4908, "step": 4850 }, { "epoch": 0.494949494949495, "grad_norm": 0.00081634521484375, "learning_rate": 4.000000000000001e-06, "loss": 0.4581, "step": 4900 }, { "epoch": 0.5, "grad_norm": 17.125, "learning_rate": 2.0000000000000003e-06, "loss": 0.5418, "step": 4950 }, { "epoch": 0.5050505050505051, "grad_norm": 9.625, "learning_rate": 0.0, "loss": 0.7143, "step": 5000 } ], "logging_steps": 50, "max_steps": 5000, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 1000, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 2.084143267725312e+16, "train_batch_size": 1, "trial_name": null, "trial_params": null }