| { | |
| "best_metric": 1.6195639371871948, | |
| "best_model_checkpoint": "./output/checkpoints/2024-06-11_15-20-56/checkpoint-10", | |
| "epoch": 3.0, | |
| "eval_steps": 1, | |
| "global_step": 111, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.02702702702702703, | |
| "grad_norm": 3.045227289199829, | |
| "learning_rate": 3.3333333333333335e-05, | |
| "loss": 5.5922, | |
| "step": 1 | |
| }, | |
| { | |
| "epoch": 0.02702702702702703, | |
| "eval_loss": 5.6431684494018555, | |
| "eval_runtime": 11.0821, | |
| "eval_samples_per_second": 11.189, | |
| "eval_steps_per_second": 0.722, | |
| "step": 1 | |
| }, | |
| { | |
| "epoch": 0.05405405405405406, | |
| "grad_norm": 3.084507942199707, | |
| "learning_rate": 6.666666666666667e-05, | |
| "loss": 5.6812, | |
| "step": 2 | |
| }, | |
| { | |
| "epoch": 0.05405405405405406, | |
| "eval_loss": 5.540425777435303, | |
| "eval_runtime": 11.083, | |
| "eval_samples_per_second": 11.188, | |
| "eval_steps_per_second": 0.722, | |
| "step": 2 | |
| }, | |
| { | |
| "epoch": 0.08108108108108109, | |
| "grad_norm": 3.1143131256103516, | |
| "learning_rate": 0.0001, | |
| "loss": 5.5805, | |
| "step": 3 | |
| }, | |
| { | |
| "epoch": 0.08108108108108109, | |
| "eval_loss": 5.062227725982666, | |
| "eval_runtime": 11.1285, | |
| "eval_samples_per_second": 11.143, | |
| "eval_steps_per_second": 0.719, | |
| "step": 3 | |
| }, | |
| { | |
| "epoch": 0.10810810810810811, | |
| "grad_norm": 3.3044817447662354, | |
| "learning_rate": 0.00013333333333333334, | |
| "loss": 5.0744, | |
| "step": 4 | |
| }, | |
| { | |
| "epoch": 0.10810810810810811, | |
| "eval_loss": 4.176051139831543, | |
| "eval_runtime": 11.2317, | |
| "eval_samples_per_second": 11.04, | |
| "eval_steps_per_second": 0.712, | |
| "step": 4 | |
| }, | |
| { | |
| "epoch": 0.13513513513513514, | |
| "grad_norm": 3.3503236770629883, | |
| "learning_rate": 0.0001666666666666667, | |
| "loss": 4.1348, | |
| "step": 5 | |
| }, | |
| { | |
| "epoch": 0.13513513513513514, | |
| "eval_loss": 3.137871265411377, | |
| "eval_runtime": 11.1384, | |
| "eval_samples_per_second": 11.133, | |
| "eval_steps_per_second": 0.718, | |
| "step": 5 | |
| }, | |
| { | |
| "epoch": 0.16216216216216217, | |
| "grad_norm": 2.9682626724243164, | |
| "learning_rate": 0.0002, | |
| "loss": 3.0425, | |
| "step": 6 | |
| }, | |
| { | |
| "epoch": 0.16216216216216217, | |
| "eval_loss": 2.459784746170044, | |
| "eval_runtime": 11.1086, | |
| "eval_samples_per_second": 11.163, | |
| "eval_steps_per_second": 0.72, | |
| "step": 6 | |
| }, | |
| { | |
| "epoch": 0.1891891891891892, | |
| "grad_norm": 1.8949077129364014, | |
| "learning_rate": 0.00023333333333333336, | |
| "loss": 2.3291, | |
| "step": 7 | |
| }, | |
| { | |
| "epoch": 0.1891891891891892, | |
| "eval_loss": 2.064051866531372, | |
| "eval_runtime": 11.1293, | |
| "eval_samples_per_second": 11.142, | |
| "eval_steps_per_second": 0.719, | |
| "step": 7 | |
| }, | |
| { | |
| "epoch": 0.21621621621621623, | |
| "grad_norm": 1.2975975275039673, | |
| "learning_rate": 0.0002666666666666667, | |
| "loss": 1.9481, | |
| "step": 8 | |
| }, | |
| { | |
| "epoch": 0.21621621621621623, | |
| "eval_loss": 1.826060175895691, | |
| "eval_runtime": 11.1457, | |
| "eval_samples_per_second": 11.125, | |
| "eval_steps_per_second": 0.718, | |
| "step": 8 | |
| }, | |
| { | |
| "epoch": 0.24324324324324326, | |
| "grad_norm": 0.9844012260437012, | |
| "learning_rate": 0.00030000000000000003, | |
| "loss": 1.6863, | |
| "step": 9 | |
| }, | |
| { | |
| "epoch": 0.24324324324324326, | |
| "eval_loss": 1.7158275842666626, | |
| "eval_runtime": 11.1384, | |
| "eval_samples_per_second": 11.133, | |
| "eval_steps_per_second": 0.718, | |
| "step": 9 | |
| }, | |
| { | |
| "epoch": 0.2702702702702703, | |
| "grad_norm": 0.9899163246154785, | |
| "learning_rate": 0.0003333333333333334, | |
| "loss": 1.6293, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.2702702702702703, | |
| "eval_loss": 1.6195639371871948, | |
| "eval_runtime": 11.1633, | |
| "eval_samples_per_second": 11.108, | |
| "eval_steps_per_second": 0.717, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.2972972972972973, | |
| "grad_norm": 0.2760697901248932, | |
| "learning_rate": 0.00036666666666666667, | |
| "loss": 1.5182, | |
| "step": 11 | |
| }, | |
| { | |
| "epoch": 0.2972972972972973, | |
| "eval_loss": 1.5875523090362549, | |
| "eval_runtime": 11.1331, | |
| "eval_samples_per_second": 11.138, | |
| "eval_steps_per_second": 0.719, | |
| "step": 11 | |
| }, | |
| { | |
| "epoch": 0.32432432432432434, | |
| "grad_norm": 0.4246121644973755, | |
| "learning_rate": 0.0004, | |
| "loss": 1.432, | |
| "step": 12 | |
| }, | |
| { | |
| "epoch": 0.32432432432432434, | |
| "eval_loss": 1.574428677558899, | |
| "eval_runtime": 11.1913, | |
| "eval_samples_per_second": 11.08, | |
| "eval_steps_per_second": 0.715, | |
| "step": 12 | |
| }, | |
| { | |
| "epoch": 0.35135135135135137, | |
| "grad_norm": 0.4968096911907196, | |
| "learning_rate": 0.000395959595959596, | |
| "loss": 1.3155, | |
| "step": 13 | |
| }, | |
| { | |
| "epoch": 0.35135135135135137, | |
| "eval_loss": 1.6574310064315796, | |
| "eval_runtime": 11.1636, | |
| "eval_samples_per_second": 11.108, | |
| "eval_steps_per_second": 0.717, | |
| "step": 13 | |
| }, | |
| { | |
| "epoch": 0.3783783783783784, | |
| "grad_norm": 0.2221263200044632, | |
| "learning_rate": 0.0003919191919191919, | |
| "loss": 1.2013, | |
| "step": 14 | |
| }, | |
| { | |
| "epoch": 0.3783783783783784, | |
| "eval_loss": 1.7277523279190063, | |
| "eval_runtime": 11.2226, | |
| "eval_samples_per_second": 11.049, | |
| "eval_steps_per_second": 0.713, | |
| "step": 14 | |
| }, | |
| { | |
| "epoch": 0.40540540540540543, | |
| "grad_norm": 0.24387121200561523, | |
| "learning_rate": 0.0003878787878787879, | |
| "loss": 1.1303, | |
| "step": 15 | |
| }, | |
| { | |
| "epoch": 0.40540540540540543, | |
| "eval_loss": 1.7132700681686401, | |
| "eval_runtime": 11.1622, | |
| "eval_samples_per_second": 11.109, | |
| "eval_steps_per_second": 0.717, | |
| "step": 15 | |
| }, | |
| { | |
| "epoch": 0.43243243243243246, | |
| "grad_norm": 0.2101132869720459, | |
| "learning_rate": 0.00038383838383838383, | |
| "loss": 1.1181, | |
| "step": 16 | |
| }, | |
| { | |
| "epoch": 0.43243243243243246, | |
| "eval_loss": 1.6620469093322754, | |
| "eval_runtime": 11.1581, | |
| "eval_samples_per_second": 11.113, | |
| "eval_steps_per_second": 0.717, | |
| "step": 16 | |
| }, | |
| { | |
| "epoch": 0.4594594594594595, | |
| "grad_norm": 0.1956334114074707, | |
| "learning_rate": 0.0003797979797979798, | |
| "loss": 1.0897, | |
| "step": 17 | |
| }, | |
| { | |
| "epoch": 0.4594594594594595, | |
| "eval_loss": 1.6307740211486816, | |
| "eval_runtime": 11.1407, | |
| "eval_samples_per_second": 11.13, | |
| "eval_steps_per_second": 0.718, | |
| "step": 17 | |
| }, | |
| { | |
| "epoch": 0.4864864864864865, | |
| "grad_norm": 0.1778702437877655, | |
| "learning_rate": 0.0003757575757575758, | |
| "loss": 1.0865, | |
| "step": 18 | |
| }, | |
| { | |
| "epoch": 0.4864864864864865, | |
| "eval_loss": 1.6229671239852905, | |
| "eval_runtime": 11.2453, | |
| "eval_samples_per_second": 11.027, | |
| "eval_steps_per_second": 0.711, | |
| "step": 18 | |
| }, | |
| { | |
| "epoch": 0.5135135135135135, | |
| "grad_norm": 0.15137992799282074, | |
| "learning_rate": 0.0003717171717171717, | |
| "loss": 1.0591, | |
| "step": 19 | |
| }, | |
| { | |
| "epoch": 0.5135135135135135, | |
| "eval_loss": 1.6271958351135254, | |
| "eval_runtime": 11.1705, | |
| "eval_samples_per_second": 11.101, | |
| "eval_steps_per_second": 0.716, | |
| "step": 19 | |
| }, | |
| { | |
| "epoch": 0.5405405405405406, | |
| "grad_norm": 0.11392053216695786, | |
| "learning_rate": 0.0003676767676767677, | |
| "loss": 1.0392, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.5405405405405406, | |
| "eval_loss": 1.6411267518997192, | |
| "eval_runtime": 11.1409, | |
| "eval_samples_per_second": 11.13, | |
| "eval_steps_per_second": 0.718, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.5675675675675675, | |
| "grad_norm": 0.0866568386554718, | |
| "learning_rate": 0.00036363636363636367, | |
| "loss": 1.0073, | |
| "step": 21 | |
| }, | |
| { | |
| "epoch": 0.5675675675675675, | |
| "eval_loss": 1.6590908765792847, | |
| "eval_runtime": 11.1242, | |
| "eval_samples_per_second": 11.147, | |
| "eval_steps_per_second": 0.719, | |
| "step": 21 | |
| }, | |
| { | |
| "epoch": 0.5945945945945946, | |
| "grad_norm": 0.10577922314405441, | |
| "learning_rate": 0.0003595959595959596, | |
| "loss": 0.9898, | |
| "step": 22 | |
| }, | |
| { | |
| "epoch": 0.5945945945945946, | |
| "eval_loss": 1.6635982990264893, | |
| "eval_runtime": 11.2221, | |
| "eval_samples_per_second": 11.05, | |
| "eval_steps_per_second": 0.713, | |
| "step": 22 | |
| }, | |
| { | |
| "epoch": 0.6216216216216216, | |
| "grad_norm": 0.1029527559876442, | |
| "learning_rate": 0.00035555555555555557, | |
| "loss": 1.0017, | |
| "step": 23 | |
| }, | |
| { | |
| "epoch": 0.6216216216216216, | |
| "eval_loss": 1.6551401615142822, | |
| "eval_runtime": 11.1896, | |
| "eval_samples_per_second": 11.082, | |
| "eval_steps_per_second": 0.715, | |
| "step": 23 | |
| }, | |
| { | |
| "epoch": 0.6486486486486487, | |
| "grad_norm": 0.09740535914897919, | |
| "learning_rate": 0.00035151515151515155, | |
| "loss": 1.0052, | |
| "step": 24 | |
| }, | |
| { | |
| "epoch": 0.6486486486486487, | |
| "eval_loss": 1.642539620399475, | |
| "eval_runtime": 11.1624, | |
| "eval_samples_per_second": 11.109, | |
| "eval_steps_per_second": 0.717, | |
| "step": 24 | |
| }, | |
| { | |
| "epoch": 0.6756756756756757, | |
| "grad_norm": 0.09230296313762665, | |
| "learning_rate": 0.00034747474747474753, | |
| "loss": 0.9873, | |
| "step": 25 | |
| }, | |
| { | |
| "epoch": 0.6756756756756757, | |
| "eval_loss": 1.6282507181167603, | |
| "eval_runtime": 11.0959, | |
| "eval_samples_per_second": 11.175, | |
| "eval_steps_per_second": 0.721, | |
| "step": 25 | |
| }, | |
| { | |
| "epoch": 0.7027027027027027, | |
| "grad_norm": 0.08403297513723373, | |
| "learning_rate": 0.00034343434343434346, | |
| "loss": 0.9743, | |
| "step": 26 | |
| }, | |
| { | |
| "epoch": 0.7027027027027027, | |
| "eval_loss": 1.6223111152648926, | |
| "eval_runtime": 11.143, | |
| "eval_samples_per_second": 11.128, | |
| "eval_steps_per_second": 0.718, | |
| "step": 26 | |
| }, | |
| { | |
| "epoch": 0.7297297297297297, | |
| "grad_norm": 0.0808207094669342, | |
| "learning_rate": 0.00033939393939393943, | |
| "loss": 0.9752, | |
| "step": 27 | |
| }, | |
| { | |
| "epoch": 0.7297297297297297, | |
| "eval_loss": 1.6211432218551636, | |
| "eval_runtime": 11.2057, | |
| "eval_samples_per_second": 11.066, | |
| "eval_steps_per_second": 0.714, | |
| "step": 27 | |
| }, | |
| { | |
| "epoch": 0.7567567567567568, | |
| "grad_norm": 0.08751657605171204, | |
| "learning_rate": 0.00033535353535353536, | |
| "loss": 0.984, | |
| "step": 28 | |
| }, | |
| { | |
| "epoch": 0.7567567567567568, | |
| "eval_loss": 1.6278128623962402, | |
| "eval_runtime": 11.2228, | |
| "eval_samples_per_second": 11.049, | |
| "eval_steps_per_second": 0.713, | |
| "step": 28 | |
| }, | |
| { | |
| "epoch": 0.7837837837837838, | |
| "grad_norm": 0.08614286035299301, | |
| "learning_rate": 0.00033131313131313134, | |
| "loss": 0.9609, | |
| "step": 29 | |
| }, | |
| { | |
| "epoch": 0.7837837837837838, | |
| "eval_loss": 1.640842318534851, | |
| "eval_runtime": 11.1451, | |
| "eval_samples_per_second": 11.126, | |
| "eval_steps_per_second": 0.718, | |
| "step": 29 | |
| }, | |
| { | |
| "epoch": 0.8108108108108109, | |
| "grad_norm": 0.09667662531137466, | |
| "learning_rate": 0.0003272727272727273, | |
| "loss": 0.9947, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.8108108108108109, | |
| "eval_loss": 1.6455894708633423, | |
| "eval_runtime": 11.163, | |
| "eval_samples_per_second": 11.108, | |
| "eval_steps_per_second": 0.717, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.8378378378378378, | |
| "grad_norm": 0.07943412661552429, | |
| "learning_rate": 0.00032323232323232324, | |
| "loss": 0.9682, | |
| "step": 31 | |
| }, | |
| { | |
| "epoch": 0.8378378378378378, | |
| "eval_loss": 1.6466931104660034, | |
| "eval_runtime": 11.1798, | |
| "eval_samples_per_second": 11.091, | |
| "eval_steps_per_second": 0.716, | |
| "step": 31 | |
| }, | |
| { | |
| "epoch": 0.8648648648648649, | |
| "grad_norm": 0.08821859955787659, | |
| "learning_rate": 0.0003191919191919192, | |
| "loss": 0.9828, | |
| "step": 32 | |
| }, | |
| { | |
| "epoch": 0.8648648648648649, | |
| "eval_loss": 1.6367429494857788, | |
| "eval_runtime": 11.1174, | |
| "eval_samples_per_second": 11.154, | |
| "eval_steps_per_second": 0.72, | |
| "step": 32 | |
| }, | |
| { | |
| "epoch": 0.8918918918918919, | |
| "grad_norm": 0.07129044830799103, | |
| "learning_rate": 0.00031515151515151515, | |
| "loss": 0.9315, | |
| "step": 33 | |
| }, | |
| { | |
| "epoch": 0.8918918918918919, | |
| "eval_loss": 1.632204294204712, | |
| "eval_runtime": 11.1509, | |
| "eval_samples_per_second": 11.12, | |
| "eval_steps_per_second": 0.717, | |
| "step": 33 | |
| }, | |
| { | |
| "epoch": 0.918918918918919, | |
| "grad_norm": 0.06505230814218521, | |
| "learning_rate": 0.0003111111111111111, | |
| "loss": 0.9541, | |
| "step": 34 | |
| }, | |
| { | |
| "epoch": 0.918918918918919, | |
| "eval_loss": 1.6281696557998657, | |
| "eval_runtime": 11.2284, | |
| "eval_samples_per_second": 11.043, | |
| "eval_steps_per_second": 0.712, | |
| "step": 34 | |
| }, | |
| { | |
| "epoch": 0.9459459459459459, | |
| "grad_norm": 0.0691852867603302, | |
| "learning_rate": 0.00030707070707070705, | |
| "loss": 0.9753, | |
| "step": 35 | |
| }, | |
| { | |
| "epoch": 0.9459459459459459, | |
| "eval_loss": 1.6272518634796143, | |
| "eval_runtime": 11.1646, | |
| "eval_samples_per_second": 11.106, | |
| "eval_steps_per_second": 0.717, | |
| "step": 35 | |
| }, | |
| { | |
| "epoch": 0.972972972972973, | |
| "grad_norm": 0.07006494700908661, | |
| "learning_rate": 0.00030303030303030303, | |
| "loss": 0.9399, | |
| "step": 36 | |
| }, | |
| { | |
| "epoch": 0.972972972972973, | |
| "eval_loss": 1.6321260929107666, | |
| "eval_runtime": 11.2226, | |
| "eval_samples_per_second": 11.049, | |
| "eval_steps_per_second": 0.713, | |
| "step": 36 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "grad_norm": 0.08096865564584732, | |
| "learning_rate": 0.000298989898989899, | |
| "loss": 0.9456, | |
| "step": 37 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "eval_loss": 1.6329066753387451, | |
| "eval_runtime": 11.1546, | |
| "eval_samples_per_second": 11.116, | |
| "eval_steps_per_second": 0.717, | |
| "step": 37 | |
| }, | |
| { | |
| "epoch": 1.027027027027027, | |
| "grad_norm": 0.06808632612228394, | |
| "learning_rate": 0.00029494949494949493, | |
| "loss": 0.9535, | |
| "step": 38 | |
| }, | |
| { | |
| "epoch": 1.027027027027027, | |
| "eval_loss": 1.6349523067474365, | |
| "eval_runtime": 11.1118, | |
| "eval_samples_per_second": 11.159, | |
| "eval_steps_per_second": 0.72, | |
| "step": 38 | |
| }, | |
| { | |
| "epoch": 1.054054054054054, | |
| "grad_norm": 0.06536230444908142, | |
| "learning_rate": 0.0002909090909090909, | |
| "loss": 0.913, | |
| "step": 39 | |
| }, | |
| { | |
| "epoch": 1.054054054054054, | |
| "eval_loss": 1.6410040855407715, | |
| "eval_runtime": 11.252, | |
| "eval_samples_per_second": 11.02, | |
| "eval_steps_per_second": 0.711, | |
| "step": 39 | |
| }, | |
| { | |
| "epoch": 1.0810810810810811, | |
| "grad_norm": 0.06929512321949005, | |
| "learning_rate": 0.00028686868686868684, | |
| "loss": 0.8924, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 1.0810810810810811, | |
| "eval_loss": 1.6510140895843506, | |
| "eval_runtime": 11.1385, | |
| "eval_samples_per_second": 11.133, | |
| "eval_steps_per_second": 0.718, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 1.1081081081081081, | |
| "grad_norm": 0.06684686243534088, | |
| "learning_rate": 0.0002828282828282828, | |
| "loss": 0.901, | |
| "step": 41 | |
| }, | |
| { | |
| "epoch": 1.1081081081081081, | |
| "eval_loss": 1.6621599197387695, | |
| "eval_runtime": 11.1957, | |
| "eval_samples_per_second": 11.076, | |
| "eval_steps_per_second": 0.715, | |
| "step": 41 | |
| }, | |
| { | |
| "epoch": 1.135135135135135, | |
| "grad_norm": 0.07049503922462463, | |
| "learning_rate": 0.0002787878787878788, | |
| "loss": 0.9256, | |
| "step": 42 | |
| }, | |
| { | |
| "epoch": 1.135135135135135, | |
| "eval_loss": 1.6734611988067627, | |
| "eval_runtime": 11.1581, | |
| "eval_samples_per_second": 11.113, | |
| "eval_steps_per_second": 0.717, | |
| "step": 42 | |
| }, | |
| { | |
| "epoch": 1.1621621621621623, | |
| "grad_norm": 0.07425494492053986, | |
| "learning_rate": 0.0002747474747474748, | |
| "loss": 0.9002, | |
| "step": 43 | |
| }, | |
| { | |
| "epoch": 1.1621621621621623, | |
| "eval_loss": 1.6794347763061523, | |
| "eval_runtime": 11.0906, | |
| "eval_samples_per_second": 11.181, | |
| "eval_steps_per_second": 0.721, | |
| "step": 43 | |
| }, | |
| { | |
| "epoch": 1.1891891891891893, | |
| "grad_norm": 0.07607400417327881, | |
| "learning_rate": 0.00027070707070707075, | |
| "loss": 0.8958, | |
| "step": 44 | |
| }, | |
| { | |
| "epoch": 1.1891891891891893, | |
| "eval_loss": 1.6791983842849731, | |
| "eval_runtime": 11.9226, | |
| "eval_samples_per_second": 10.4, | |
| "eval_steps_per_second": 0.671, | |
| "step": 44 | |
| }, | |
| { | |
| "epoch": 1.2162162162162162, | |
| "grad_norm": 0.08622392266988754, | |
| "learning_rate": 0.0002666666666666667, | |
| "loss": 0.9143, | |
| "step": 45 | |
| }, | |
| { | |
| "epoch": 1.2162162162162162, | |
| "eval_loss": 1.671402096748352, | |
| "eval_runtime": 11.1349, | |
| "eval_samples_per_second": 11.136, | |
| "eval_steps_per_second": 0.718, | |
| "step": 45 | |
| }, | |
| { | |
| "epoch": 1.2432432432432432, | |
| "grad_norm": 0.07751356065273285, | |
| "learning_rate": 0.00026262626262626266, | |
| "loss": 0.9092, | |
| "step": 46 | |
| }, | |
| { | |
| "epoch": 1.2432432432432432, | |
| "eval_loss": 1.6581931114196777, | |
| "eval_runtime": 11.1166, | |
| "eval_samples_per_second": 11.154, | |
| "eval_steps_per_second": 0.72, | |
| "step": 46 | |
| }, | |
| { | |
| "epoch": 1.2702702702702702, | |
| "grad_norm": 0.07954052835702896, | |
| "learning_rate": 0.00025858585858585864, | |
| "loss": 0.9116, | |
| "step": 47 | |
| }, | |
| { | |
| "epoch": 1.2702702702702702, | |
| "eval_loss": 1.6473214626312256, | |
| "eval_runtime": 11.1774, | |
| "eval_samples_per_second": 11.094, | |
| "eval_steps_per_second": 0.716, | |
| "step": 47 | |
| }, | |
| { | |
| "epoch": 1.2972972972972974, | |
| "grad_norm": 0.08470755070447922, | |
| "learning_rate": 0.00025454545454545456, | |
| "loss": 0.9039, | |
| "step": 48 | |
| }, | |
| { | |
| "epoch": 1.2972972972972974, | |
| "eval_loss": 1.6490856409072876, | |
| "eval_runtime": 11.1926, | |
| "eval_samples_per_second": 11.079, | |
| "eval_steps_per_second": 0.715, | |
| "step": 48 | |
| }, | |
| { | |
| "epoch": 1.3243243243243243, | |
| "grad_norm": 0.08538611233234406, | |
| "learning_rate": 0.00025050505050505054, | |
| "loss": 0.8945, | |
| "step": 49 | |
| }, | |
| { | |
| "epoch": 1.3243243243243243, | |
| "eval_loss": 1.6652594804763794, | |
| "eval_runtime": 11.189, | |
| "eval_samples_per_second": 11.082, | |
| "eval_steps_per_second": 0.715, | |
| "step": 49 | |
| }, | |
| { | |
| "epoch": 1.3513513513513513, | |
| "grad_norm": 0.08962593227624893, | |
| "learning_rate": 0.00024646464646464647, | |
| "loss": 0.8878, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 1.3513513513513513, | |
| "eval_loss": 1.6737427711486816, | |
| "eval_runtime": 11.194, | |
| "eval_samples_per_second": 11.077, | |
| "eval_steps_per_second": 0.715, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 1.3783783783783785, | |
| "grad_norm": 0.09344533085823059, | |
| "learning_rate": 0.00024242424242424245, | |
| "loss": 0.8921, | |
| "step": 51 | |
| }, | |
| { | |
| "epoch": 1.3783783783783785, | |
| "eval_loss": 1.6767627000808716, | |
| "eval_runtime": 11.9154, | |
| "eval_samples_per_second": 10.407, | |
| "eval_steps_per_second": 0.671, | |
| "step": 51 | |
| }, | |
| { | |
| "epoch": 1.4054054054054055, | |
| "grad_norm": 0.08334413915872574, | |
| "learning_rate": 0.00023838383838383837, | |
| "loss": 0.8879, | |
| "step": 52 | |
| }, | |
| { | |
| "epoch": 1.4054054054054055, | |
| "eval_loss": 1.6724966764450073, | |
| "eval_runtime": 11.0787, | |
| "eval_samples_per_second": 11.193, | |
| "eval_steps_per_second": 0.722, | |
| "step": 52 | |
| }, | |
| { | |
| "epoch": 1.4324324324324325, | |
| "grad_norm": 0.0849594995379448, | |
| "learning_rate": 0.00023434343434343435, | |
| "loss": 0.9021, | |
| "step": 53 | |
| }, | |
| { | |
| "epoch": 1.4324324324324325, | |
| "eval_loss": 1.6656012535095215, | |
| "eval_runtime": 11.1933, | |
| "eval_samples_per_second": 11.078, | |
| "eval_steps_per_second": 0.715, | |
| "step": 53 | |
| }, | |
| { | |
| "epoch": 1.4594594594594594, | |
| "grad_norm": 0.09131031483411789, | |
| "learning_rate": 0.00023030303030303033, | |
| "loss": 0.8975, | |
| "step": 54 | |
| }, | |
| { | |
| "epoch": 1.4594594594594594, | |
| "eval_loss": 1.6548563241958618, | |
| "eval_runtime": 11.2025, | |
| "eval_samples_per_second": 11.069, | |
| "eval_steps_per_second": 0.714, | |
| "step": 54 | |
| }, | |
| { | |
| "epoch": 1.4864864864864864, | |
| "grad_norm": 0.0867348462343216, | |
| "learning_rate": 0.00022626262626262625, | |
| "loss": 0.8746, | |
| "step": 55 | |
| }, | |
| { | |
| "epoch": 1.4864864864864864, | |
| "eval_loss": 1.6583256721496582, | |
| "eval_runtime": 11.1258, | |
| "eval_samples_per_second": 11.145, | |
| "eval_steps_per_second": 0.719, | |
| "step": 55 | |
| }, | |
| { | |
| "epoch": 1.5135135135135136, | |
| "grad_norm": 0.0924694687128067, | |
| "learning_rate": 0.00022222222222222223, | |
| "loss": 0.8818, | |
| "step": 56 | |
| }, | |
| { | |
| "epoch": 1.5135135135135136, | |
| "eval_loss": 1.6754804849624634, | |
| "eval_runtime": 11.1671, | |
| "eval_samples_per_second": 11.104, | |
| "eval_steps_per_second": 0.716, | |
| "step": 56 | |
| }, | |
| { | |
| "epoch": 1.5405405405405406, | |
| "grad_norm": 0.09178721904754639, | |
| "learning_rate": 0.00021818181818181818, | |
| "loss": 0.9022, | |
| "step": 57 | |
| }, | |
| { | |
| "epoch": 1.5405405405405406, | |
| "eval_loss": 1.690301775932312, | |
| "eval_runtime": 11.0856, | |
| "eval_samples_per_second": 11.186, | |
| "eval_steps_per_second": 0.722, | |
| "step": 57 | |
| }, | |
| { | |
| "epoch": 1.5675675675675675, | |
| "grad_norm": 0.08697674423456192, | |
| "learning_rate": 0.00021414141414141416, | |
| "loss": 0.8596, | |
| "step": 58 | |
| }, | |
| { | |
| "epoch": 1.5675675675675675, | |
| "eval_loss": 1.7041698694229126, | |
| "eval_runtime": 11.2652, | |
| "eval_samples_per_second": 11.007, | |
| "eval_steps_per_second": 0.71, | |
| "step": 58 | |
| }, | |
| { | |
| "epoch": 1.5945945945945947, | |
| "grad_norm": 0.08802448958158493, | |
| "learning_rate": 0.00021010101010101014, | |
| "loss": 0.8572, | |
| "step": 59 | |
| }, | |
| { | |
| "epoch": 1.5945945945945947, | |
| "eval_loss": 1.7124710083007812, | |
| "eval_runtime": 11.152, | |
| "eval_samples_per_second": 11.119, | |
| "eval_steps_per_second": 0.717, | |
| "step": 59 | |
| }, | |
| { | |
| "epoch": 1.6216216216216215, | |
| "grad_norm": 0.10100586712360382, | |
| "learning_rate": 0.00020606060606060607, | |
| "loss": 0.8883, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 1.6216216216216215, | |
| "eval_loss": 1.7054810523986816, | |
| "eval_runtime": 11.1538, | |
| "eval_samples_per_second": 11.117, | |
| "eval_steps_per_second": 0.717, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 1.6486486486486487, | |
| "grad_norm": 0.11259682476520538, | |
| "learning_rate": 0.00020202020202020205, | |
| "loss": 0.8968, | |
| "step": 61 | |
| }, | |
| { | |
| "epoch": 1.6486486486486487, | |
| "eval_loss": 1.683681845664978, | |
| "eval_runtime": 11.1309, | |
| "eval_samples_per_second": 11.14, | |
| "eval_steps_per_second": 0.719, | |
| "step": 61 | |
| }, | |
| { | |
| "epoch": 1.6756756756756757, | |
| "grad_norm": 0.0911870002746582, | |
| "learning_rate": 0.000197979797979798, | |
| "loss": 0.8287, | |
| "step": 62 | |
| }, | |
| { | |
| "epoch": 1.6756756756756757, | |
| "eval_loss": 1.6670082807540894, | |
| "eval_runtime": 11.1469, | |
| "eval_samples_per_second": 11.124, | |
| "eval_steps_per_second": 0.718, | |
| "step": 62 | |
| }, | |
| { | |
| "epoch": 1.7027027027027026, | |
| "grad_norm": 0.08852899819612503, | |
| "learning_rate": 0.00019393939393939395, | |
| "loss": 0.8699, | |
| "step": 63 | |
| }, | |
| { | |
| "epoch": 1.7027027027027026, | |
| "eval_loss": 1.6602628231048584, | |
| "eval_runtime": 11.2189, | |
| "eval_samples_per_second": 11.053, | |
| "eval_steps_per_second": 0.713, | |
| "step": 63 | |
| }, | |
| { | |
| "epoch": 1.7297297297297298, | |
| "grad_norm": 0.0944983959197998, | |
| "learning_rate": 0.0001898989898989899, | |
| "loss": 0.8812, | |
| "step": 64 | |
| }, | |
| { | |
| "epoch": 1.7297297297297298, | |
| "eval_loss": 1.664315104484558, | |
| "eval_runtime": 11.1484, | |
| "eval_samples_per_second": 11.123, | |
| "eval_steps_per_second": 0.718, | |
| "step": 64 | |
| }, | |
| { | |
| "epoch": 1.7567567567567568, | |
| "grad_norm": 0.09581159055233002, | |
| "learning_rate": 0.00018585858585858586, | |
| "loss": 0.8671, | |
| "step": 65 | |
| }, | |
| { | |
| "epoch": 1.7567567567567568, | |
| "eval_loss": 1.6748771667480469, | |
| "eval_runtime": 11.2306, | |
| "eval_samples_per_second": 11.041, | |
| "eval_steps_per_second": 0.712, | |
| "step": 65 | |
| }, | |
| { | |
| "epoch": 1.7837837837837838, | |
| "grad_norm": 0.09136416763067245, | |
| "learning_rate": 0.00018181818181818183, | |
| "loss": 0.8857, | |
| "step": 66 | |
| }, | |
| { | |
| "epoch": 1.7837837837837838, | |
| "eval_loss": 1.6853251457214355, | |
| "eval_runtime": 11.1546, | |
| "eval_samples_per_second": 11.116, | |
| "eval_steps_per_second": 0.717, | |
| "step": 66 | |
| }, | |
| { | |
| "epoch": 1.810810810810811, | |
| "grad_norm": 0.09602421522140503, | |
| "learning_rate": 0.00017777777777777779, | |
| "loss": 0.8578, | |
| "step": 67 | |
| }, | |
| { | |
| "epoch": 1.810810810810811, | |
| "eval_loss": 1.6973600387573242, | |
| "eval_runtime": 11.2294, | |
| "eval_samples_per_second": 11.042, | |
| "eval_steps_per_second": 0.712, | |
| "step": 67 | |
| }, | |
| { | |
| "epoch": 1.8378378378378377, | |
| "grad_norm": 0.10458902269601822, | |
| "learning_rate": 0.00017373737373737377, | |
| "loss": 0.8767, | |
| "step": 68 | |
| }, | |
| { | |
| "epoch": 1.8378378378378377, | |
| "eval_loss": 1.7042148113250732, | |
| "eval_runtime": 11.1244, | |
| "eval_samples_per_second": 11.147, | |
| "eval_steps_per_second": 0.719, | |
| "step": 68 | |
| }, | |
| { | |
| "epoch": 1.864864864864865, | |
| "grad_norm": 0.12431007623672485, | |
| "learning_rate": 0.00016969696969696972, | |
| "loss": 0.8914, | |
| "step": 69 | |
| }, | |
| { | |
| "epoch": 1.864864864864865, | |
| "eval_loss": 1.702877402305603, | |
| "eval_runtime": 11.144, | |
| "eval_samples_per_second": 11.127, | |
| "eval_steps_per_second": 0.718, | |
| "step": 69 | |
| }, | |
| { | |
| "epoch": 1.8918918918918919, | |
| "grad_norm": 0.11106838285923004, | |
| "learning_rate": 0.00016565656565656567, | |
| "loss": 0.8843, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 1.8918918918918919, | |
| "eval_loss": 1.691541314125061, | |
| "eval_runtime": 11.1751, | |
| "eval_samples_per_second": 11.096, | |
| "eval_steps_per_second": 0.716, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 1.9189189189189189, | |
| "grad_norm": 0.0916941687464714, | |
| "learning_rate": 0.00016161616161616162, | |
| "loss": 0.8576, | |
| "step": 71 | |
| }, | |
| { | |
| "epoch": 1.9189189189189189, | |
| "eval_loss": 1.68495774269104, | |
| "eval_runtime": 11.208, | |
| "eval_samples_per_second": 11.063, | |
| "eval_steps_per_second": 0.714, | |
| "step": 71 | |
| }, | |
| { | |
| "epoch": 1.945945945945946, | |
| "grad_norm": 0.10394521802663803, | |
| "learning_rate": 0.00015757575757575757, | |
| "loss": 0.853, | |
| "step": 72 | |
| }, | |
| { | |
| "epoch": 1.945945945945946, | |
| "eval_loss": 1.6894781589508057, | |
| "eval_runtime": 11.1546, | |
| "eval_samples_per_second": 11.117, | |
| "eval_steps_per_second": 0.717, | |
| "step": 72 | |
| }, | |
| { | |
| "epoch": 1.972972972972973, | |
| "grad_norm": 0.09884631633758545, | |
| "learning_rate": 0.00015353535353535353, | |
| "loss": 0.8346, | |
| "step": 73 | |
| }, | |
| { | |
| "epoch": 1.972972972972973, | |
| "eval_loss": 1.7004313468933105, | |
| "eval_runtime": 11.107, | |
| "eval_samples_per_second": 11.164, | |
| "eval_steps_per_second": 0.72, | |
| "step": 73 | |
| }, | |
| { | |
| "epoch": 2.0, | |
| "grad_norm": 0.11277178674936295, | |
| "learning_rate": 0.0001494949494949495, | |
| "loss": 0.8524, | |
| "step": 74 | |
| }, | |
| { | |
| "epoch": 2.0, | |
| "eval_loss": 1.704146385192871, | |
| "eval_runtime": 11.1746, | |
| "eval_samples_per_second": 11.097, | |
| "eval_steps_per_second": 0.716, | |
| "step": 74 | |
| }, | |
| { | |
| "epoch": 2.027027027027027, | |
| "grad_norm": 0.0982976108789444, | |
| "learning_rate": 0.00014545454545454546, | |
| "loss": 0.8267, | |
| "step": 75 | |
| }, | |
| { | |
| "epoch": 2.027027027027027, | |
| "eval_loss": 1.7128726243972778, | |
| "eval_runtime": 11.1189, | |
| "eval_samples_per_second": 11.152, | |
| "eval_steps_per_second": 0.719, | |
| "step": 75 | |
| }, | |
| { | |
| "epoch": 2.054054054054054, | |
| "grad_norm": 0.09641794860363007, | |
| "learning_rate": 0.0001414141414141414, | |
| "loss": 0.8157, | |
| "step": 76 | |
| }, | |
| { | |
| "epoch": 2.054054054054054, | |
| "eval_loss": 1.7283172607421875, | |
| "eval_runtime": 11.1515, | |
| "eval_samples_per_second": 11.12, | |
| "eval_steps_per_second": 0.717, | |
| "step": 76 | |
| }, | |
| { | |
| "epoch": 2.081081081081081, | |
| "grad_norm": 0.10205162316560745, | |
| "learning_rate": 0.0001373737373737374, | |
| "loss": 0.7977, | |
| "step": 77 | |
| }, | |
| { | |
| "epoch": 2.081081081081081, | |
| "eval_loss": 1.7365342378616333, | |
| "eval_runtime": 11.0683, | |
| "eval_samples_per_second": 11.203, | |
| "eval_steps_per_second": 0.723, | |
| "step": 77 | |
| }, | |
| { | |
| "epoch": 2.108108108108108, | |
| "grad_norm": 0.1005796566605568, | |
| "learning_rate": 0.00013333333333333334, | |
| "loss": 0.815, | |
| "step": 78 | |
| }, | |
| { | |
| "epoch": 2.108108108108108, | |
| "eval_loss": 1.7438631057739258, | |
| "eval_runtime": 11.1586, | |
| "eval_samples_per_second": 11.113, | |
| "eval_steps_per_second": 0.717, | |
| "step": 78 | |
| }, | |
| { | |
| "epoch": 2.135135135135135, | |
| "grad_norm": 0.111351877450943, | |
| "learning_rate": 0.00012929292929292932, | |
| "loss": 0.7987, | |
| "step": 79 | |
| }, | |
| { | |
| "epoch": 2.135135135135135, | |
| "eval_loss": 1.7426629066467285, | |
| "eval_runtime": 11.1248, | |
| "eval_samples_per_second": 11.146, | |
| "eval_steps_per_second": 0.719, | |
| "step": 79 | |
| }, | |
| { | |
| "epoch": 2.1621621621621623, | |
| "grad_norm": 0.10649854689836502, | |
| "learning_rate": 0.00012525252525252527, | |
| "loss": 0.8072, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 2.1621621621621623, | |
| "eval_loss": 1.7457417249679565, | |
| "eval_runtime": 11.1907, | |
| "eval_samples_per_second": 11.081, | |
| "eval_steps_per_second": 0.715, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 2.189189189189189, | |
| "grad_norm": 0.09818455576896667, | |
| "learning_rate": 0.00012121212121212122, | |
| "loss": 0.8024, | |
| "step": 81 | |
| }, | |
| { | |
| "epoch": 2.189189189189189, | |
| "eval_loss": 1.7508054971694946, | |
| "eval_runtime": 11.1556, | |
| "eval_samples_per_second": 11.115, | |
| "eval_steps_per_second": 0.717, | |
| "step": 81 | |
| }, | |
| { | |
| "epoch": 2.2162162162162162, | |
| "grad_norm": 0.1122928112745285, | |
| "learning_rate": 0.00011717171717171717, | |
| "loss": 0.8208, | |
| "step": 82 | |
| }, | |
| { | |
| "epoch": 2.2162162162162162, | |
| "eval_loss": 1.7618576288223267, | |
| "eval_runtime": 11.2048, | |
| "eval_samples_per_second": 11.067, | |
| "eval_steps_per_second": 0.714, | |
| "step": 82 | |
| }, | |
| { | |
| "epoch": 2.2432432432432434, | |
| "grad_norm": 0.10680953413248062, | |
| "learning_rate": 0.00011313131313131313, | |
| "loss": 0.7858, | |
| "step": 83 | |
| }, | |
| { | |
| "epoch": 2.2432432432432434, | |
| "eval_loss": 1.7693935632705688, | |
| "eval_runtime": 11.1465, | |
| "eval_samples_per_second": 11.125, | |
| "eval_steps_per_second": 0.718, | |
| "step": 83 | |
| }, | |
| { | |
| "epoch": 2.27027027027027, | |
| "grad_norm": 0.11239504814147949, | |
| "learning_rate": 0.00010909090909090909, | |
| "loss": 0.8082, | |
| "step": 84 | |
| }, | |
| { | |
| "epoch": 2.27027027027027, | |
| "eval_loss": 1.780279517173767, | |
| "eval_runtime": 11.1781, | |
| "eval_samples_per_second": 11.093, | |
| "eval_steps_per_second": 0.716, | |
| "step": 84 | |
| }, | |
| { | |
| "epoch": 2.2972972972972974, | |
| "grad_norm": 0.12527017295360565, | |
| "learning_rate": 0.00010505050505050507, | |
| "loss": 0.7943, | |
| "step": 85 | |
| }, | |
| { | |
| "epoch": 2.2972972972972974, | |
| "eval_loss": 1.7797414064407349, | |
| "eval_runtime": 11.1811, | |
| "eval_samples_per_second": 11.09, | |
| "eval_steps_per_second": 0.715, | |
| "step": 85 | |
| }, | |
| { | |
| "epoch": 2.3243243243243246, | |
| "grad_norm": 0.13715100288391113, | |
| "learning_rate": 0.00010101010101010102, | |
| "loss": 0.8338, | |
| "step": 86 | |
| }, | |
| { | |
| "epoch": 2.3243243243243246, | |
| "eval_loss": 1.7673975229263306, | |
| "eval_runtime": 11.217, | |
| "eval_samples_per_second": 11.055, | |
| "eval_steps_per_second": 0.713, | |
| "step": 86 | |
| }, | |
| { | |
| "epoch": 2.3513513513513513, | |
| "grad_norm": 0.11464574187994003, | |
| "learning_rate": 9.696969696969698e-05, | |
| "loss": 0.8038, | |
| "step": 87 | |
| }, | |
| { | |
| "epoch": 2.3513513513513513, | |
| "eval_loss": 1.75497305393219, | |
| "eval_runtime": 11.1384, | |
| "eval_samples_per_second": 11.133, | |
| "eval_steps_per_second": 0.718, | |
| "step": 87 | |
| }, | |
| { | |
| "epoch": 2.3783783783783785, | |
| "grad_norm": 0.11703913658857346, | |
| "learning_rate": 9.292929292929293e-05, | |
| "loss": 0.812, | |
| "step": 88 | |
| }, | |
| { | |
| "epoch": 2.3783783783783785, | |
| "eval_loss": 1.7427462339401245, | |
| "eval_runtime": 11.1796, | |
| "eval_samples_per_second": 11.092, | |
| "eval_steps_per_second": 0.716, | |
| "step": 88 | |
| }, | |
| { | |
| "epoch": 2.4054054054054053, | |
| "grad_norm": 0.1178952306509018, | |
| "learning_rate": 8.888888888888889e-05, | |
| "loss": 0.8117, | |
| "step": 89 | |
| }, | |
| { | |
| "epoch": 2.4054054054054053, | |
| "eval_loss": 1.7358033657073975, | |
| "eval_runtime": 11.1223, | |
| "eval_samples_per_second": 11.149, | |
| "eval_steps_per_second": 0.719, | |
| "step": 89 | |
| }, | |
| { | |
| "epoch": 2.4324324324324325, | |
| "grad_norm": 0.1348622441291809, | |
| "learning_rate": 8.484848484848486e-05, | |
| "loss": 0.8172, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 2.4324324324324325, | |
| "eval_loss": 1.7397310733795166, | |
| "eval_runtime": 11.1841, | |
| "eval_samples_per_second": 11.087, | |
| "eval_steps_per_second": 0.715, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 2.4594594594594597, | |
| "grad_norm": 0.12706562876701355, | |
| "learning_rate": 8.080808080808081e-05, | |
| "loss": 0.808, | |
| "step": 91 | |
| }, | |
| { | |
| "epoch": 2.4594594594594597, | |
| "eval_loss": 1.751862645149231, | |
| "eval_runtime": 11.0963, | |
| "eval_samples_per_second": 11.175, | |
| "eval_steps_per_second": 0.721, | |
| "step": 91 | |
| }, | |
| { | |
| "epoch": 2.4864864864864864, | |
| "grad_norm": 0.11583199352025986, | |
| "learning_rate": 7.676767676767676e-05, | |
| "loss": 0.8108, | |
| "step": 92 | |
| }, | |
| { | |
| "epoch": 2.4864864864864864, | |
| "eval_loss": 1.764027714729309, | |
| "eval_runtime": 11.1926, | |
| "eval_samples_per_second": 11.079, | |
| "eval_steps_per_second": 0.715, | |
| "step": 92 | |
| }, | |
| { | |
| "epoch": 2.5135135135135136, | |
| "grad_norm": 0.12227907031774521, | |
| "learning_rate": 7.272727272727273e-05, | |
| "loss": 0.8086, | |
| "step": 93 | |
| }, | |
| { | |
| "epoch": 2.5135135135135136, | |
| "eval_loss": 1.7744636535644531, | |
| "eval_runtime": 11.1394, | |
| "eval_samples_per_second": 11.132, | |
| "eval_steps_per_second": 0.718, | |
| "step": 93 | |
| }, | |
| { | |
| "epoch": 2.5405405405405403, | |
| "grad_norm": 0.1270534247159958, | |
| "learning_rate": 6.86868686868687e-05, | |
| "loss": 0.8079, | |
| "step": 94 | |
| }, | |
| { | |
| "epoch": 2.5405405405405403, | |
| "eval_loss": 1.7796293497085571, | |
| "eval_runtime": 11.1924, | |
| "eval_samples_per_second": 11.079, | |
| "eval_steps_per_second": 0.715, | |
| "step": 94 | |
| }, | |
| { | |
| "epoch": 2.5675675675675675, | |
| "grad_norm": 0.1266915649175644, | |
| "learning_rate": 6.464646464646466e-05, | |
| "loss": 0.7876, | |
| "step": 95 | |
| }, | |
| { | |
| "epoch": 2.5675675675675675, | |
| "eval_loss": 1.7799245119094849, | |
| "eval_runtime": 11.1921, | |
| "eval_samples_per_second": 11.079, | |
| "eval_steps_per_second": 0.715, | |
| "step": 95 | |
| }, | |
| { | |
| "epoch": 2.5945945945945947, | |
| "grad_norm": 0.13456638157367706, | |
| "learning_rate": 6.060606060606061e-05, | |
| "loss": 0.7756, | |
| "step": 96 | |
| }, | |
| { | |
| "epoch": 2.5945945945945947, | |
| "eval_loss": 1.7748299837112427, | |
| "eval_runtime": 11.1592, | |
| "eval_samples_per_second": 11.112, | |
| "eval_steps_per_second": 0.717, | |
| "step": 96 | |
| }, | |
| { | |
| "epoch": 2.6216216216216215, | |
| "grad_norm": 0.12409545481204987, | |
| "learning_rate": 5.6565656565656563e-05, | |
| "loss": 0.795, | |
| "step": 97 | |
| }, | |
| { | |
| "epoch": 2.6216216216216215, | |
| "eval_loss": 1.7676475048065186, | |
| "eval_runtime": 11.1442, | |
| "eval_samples_per_second": 11.127, | |
| "eval_steps_per_second": 0.718, | |
| "step": 97 | |
| }, | |
| { | |
| "epoch": 2.6486486486486487, | |
| "grad_norm": 0.11815664917230606, | |
| "learning_rate": 5.2525252525252536e-05, | |
| "loss": 0.7706, | |
| "step": 98 | |
| }, | |
| { | |
| "epoch": 2.6486486486486487, | |
| "eval_loss": 1.7634869813919067, | |
| "eval_runtime": 11.2593, | |
| "eval_samples_per_second": 11.013, | |
| "eval_steps_per_second": 0.711, | |
| "step": 98 | |
| }, | |
| { | |
| "epoch": 2.6756756756756754, | |
| "grad_norm": 0.11971355229616165, | |
| "learning_rate": 4.848484848484849e-05, | |
| "loss": 0.792, | |
| "step": 99 | |
| }, | |
| { | |
| "epoch": 2.6756756756756754, | |
| "eval_loss": 1.7610678672790527, | |
| "eval_runtime": 11.1535, | |
| "eval_samples_per_second": 11.118, | |
| "eval_steps_per_second": 0.717, | |
| "step": 99 | |
| }, | |
| { | |
| "epoch": 2.7027027027027026, | |
| "grad_norm": 0.12585307657718658, | |
| "learning_rate": 4.4444444444444447e-05, | |
| "loss": 0.7942, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 2.7027027027027026, | |
| "eval_loss": 1.7610435485839844, | |
| "eval_runtime": 11.0721, | |
| "eval_samples_per_second": 11.199, | |
| "eval_steps_per_second": 0.723, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 2.72972972972973, | |
| "grad_norm": 0.11809434741735458, | |
| "learning_rate": 4.0404040404040405e-05, | |
| "loss": 0.7697, | |
| "step": 101 | |
| }, | |
| { | |
| "epoch": 2.72972972972973, | |
| "eval_loss": 1.7632513046264648, | |
| "eval_runtime": 11.1597, | |
| "eval_samples_per_second": 11.111, | |
| "eval_steps_per_second": 0.717, | |
| "step": 101 | |
| }, | |
| { | |
| "epoch": 2.756756756756757, | |
| "grad_norm": 0.12419259548187256, | |
| "learning_rate": 3.6363636363636364e-05, | |
| "loss": 0.7789, | |
| "step": 102 | |
| }, | |
| { | |
| "epoch": 2.756756756756757, | |
| "eval_loss": 1.7658323049545288, | |
| "eval_runtime": 11.1861, | |
| "eval_samples_per_second": 11.085, | |
| "eval_steps_per_second": 0.715, | |
| "step": 102 | |
| }, | |
| { | |
| "epoch": 2.7837837837837838, | |
| "grad_norm": 0.12964174151420593, | |
| "learning_rate": 3.232323232323233e-05, | |
| "loss": 0.8047, | |
| "step": 103 | |
| }, | |
| { | |
| "epoch": 2.7837837837837838, | |
| "eval_loss": 1.7687405347824097, | |
| "eval_runtime": 11.0868, | |
| "eval_samples_per_second": 11.184, | |
| "eval_steps_per_second": 0.722, | |
| "step": 103 | |
| }, | |
| { | |
| "epoch": 2.810810810810811, | |
| "grad_norm": 0.1257963329553604, | |
| "learning_rate": 2.8282828282828282e-05, | |
| "loss": 0.7653, | |
| "step": 104 | |
| }, | |
| { | |
| "epoch": 2.810810810810811, | |
| "eval_loss": 1.7723119258880615, | |
| "eval_runtime": 11.1173, | |
| "eval_samples_per_second": 11.154, | |
| "eval_steps_per_second": 0.72, | |
| "step": 104 | |
| }, | |
| { | |
| "epoch": 2.8378378378378377, | |
| "grad_norm": 0.12039487063884735, | |
| "learning_rate": 2.4242424242424244e-05, | |
| "loss": 0.7753, | |
| "step": 105 | |
| }, | |
| { | |
| "epoch": 2.8378378378378377, | |
| "eval_loss": 1.7749425172805786, | |
| "eval_runtime": 11.157, | |
| "eval_samples_per_second": 11.114, | |
| "eval_steps_per_second": 0.717, | |
| "step": 105 | |
| }, | |
| { | |
| "epoch": 2.864864864864865, | |
| "grad_norm": 0.13161887228488922, | |
| "learning_rate": 2.0202020202020203e-05, | |
| "loss": 0.7631, | |
| "step": 106 | |
| }, | |
| { | |
| "epoch": 2.864864864864865, | |
| "eval_loss": 1.7760539054870605, | |
| "eval_runtime": 11.1933, | |
| "eval_samples_per_second": 11.078, | |
| "eval_steps_per_second": 0.715, | |
| "step": 106 | |
| }, | |
| { | |
| "epoch": 2.891891891891892, | |
| "grad_norm": 0.12347672879695892, | |
| "learning_rate": 1.6161616161616165e-05, | |
| "loss": 0.7821, | |
| "step": 107 | |
| }, | |
| { | |
| "epoch": 2.891891891891892, | |
| "eval_loss": 1.7769430875778198, | |
| "eval_runtime": 11.169, | |
| "eval_samples_per_second": 11.102, | |
| "eval_steps_per_second": 0.716, | |
| "step": 107 | |
| }, | |
| { | |
| "epoch": 2.918918918918919, | |
| "grad_norm": 0.12539447844028473, | |
| "learning_rate": 1.2121212121212122e-05, | |
| "loss": 0.7682, | |
| "step": 108 | |
| }, | |
| { | |
| "epoch": 2.918918918918919, | |
| "eval_loss": 1.77769136428833, | |
| "eval_runtime": 11.2241, | |
| "eval_samples_per_second": 11.048, | |
| "eval_steps_per_second": 0.713, | |
| "step": 108 | |
| }, | |
| { | |
| "epoch": 2.945945945945946, | |
| "grad_norm": 0.12666372954845428, | |
| "learning_rate": 8.080808080808082e-06, | |
| "loss": 0.7893, | |
| "step": 109 | |
| }, | |
| { | |
| "epoch": 2.945945945945946, | |
| "eval_loss": 1.7779371738433838, | |
| "eval_runtime": 11.1071, | |
| "eval_samples_per_second": 11.164, | |
| "eval_steps_per_second": 0.72, | |
| "step": 109 | |
| }, | |
| { | |
| "epoch": 2.972972972972973, | |
| "grad_norm": 0.11765623837709427, | |
| "learning_rate": 4.040404040404041e-06, | |
| "loss": 0.7729, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 2.972972972972973, | |
| "eval_loss": 1.7782450914382935, | |
| "eval_runtime": 11.121, | |
| "eval_samples_per_second": 11.15, | |
| "eval_steps_per_second": 0.719, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 3.0, | |
| "grad_norm": 0.14280778169631958, | |
| "learning_rate": 0.0, | |
| "loss": 0.7783, | |
| "step": 111 | |
| }, | |
| { | |
| "epoch": 3.0, | |
| "eval_loss": 1.7782894372940063, | |
| "eval_runtime": 11.0882, | |
| "eval_samples_per_second": 11.183, | |
| "eval_steps_per_second": 0.721, | |
| "step": 111 | |
| }, | |
| { | |
| "epoch": 3.0, | |
| "step": 111, | |
| "total_flos": 4.092563650668134e+16, | |
| "train_loss": 1.148636352371525, | |
| "train_runtime": 2293.4445, | |
| "train_samples_per_second": 1.54, | |
| "train_steps_per_second": 0.048 | |
| } | |
| ], | |
| "logging_steps": 1, | |
| "max_steps": 111, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 3, | |
| "save_steps": 10, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": false | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 4.092563650668134e+16, | |
| "train_batch_size": 32, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |