{ "best_metric": null, "best_model_checkpoint": null, "epoch": 2.98989898989899, "eval_steps": 20, "global_step": 111, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.05387205387205387, "grad_norm": 30.91269874572754, "learning_rate": 0.0001, "loss": 9.4247, "step": 2 }, { "epoch": 0.10774410774410774, "grad_norm": 17.899776458740234, "learning_rate": 9.818181818181818e-05, "loss": 8.2334, "step": 4 }, { "epoch": 0.16161616161616163, "grad_norm": 11.289294242858887, "learning_rate": 9.636363636363637e-05, "loss": 6.7049, "step": 6 }, { "epoch": 0.21548821548821548, "grad_norm": 13.52355670928955, "learning_rate": 9.454545454545455e-05, "loss": 6.0078, "step": 8 }, { "epoch": 0.26936026936026936, "grad_norm": 6.83944845199585, "learning_rate": 9.272727272727273e-05, "loss": 5.4206, "step": 10 }, { "epoch": 0.32323232323232326, "grad_norm": 5.749248504638672, "learning_rate": 9.090909090909092e-05, "loss": 5.0018, "step": 12 }, { "epoch": 0.3771043771043771, "grad_norm": 6.119119644165039, "learning_rate": 8.90909090909091e-05, "loss": 4.7585, "step": 14 }, { "epoch": 0.43097643097643096, "grad_norm": 5.070394992828369, "learning_rate": 8.727272727272727e-05, "loss": 4.4387, "step": 16 }, { "epoch": 0.48484848484848486, "grad_norm": 5.378049850463867, "learning_rate": 8.545454545454545e-05, "loss": 4.1869, "step": 18 }, { "epoch": 0.5387205387205387, "grad_norm": 6.204185962677002, "learning_rate": 8.363636363636364e-05, "loss": 4.0268, "step": 20 }, { "epoch": 0.5387205387205387, "eval_loss": 3.4558703899383545, "eval_runtime": 8.1602, "eval_samples_per_second": 145.708, "eval_steps_per_second": 18.259, "step": 20 }, { "epoch": 0.5925925925925926, "grad_norm": 4.224447727203369, "learning_rate": 8.181818181818183e-05, "loss": 3.8953, "step": 22 }, { "epoch": 0.6464646464646465, "grad_norm": 3.6069717407226562, "learning_rate": 8e-05, "loss": 3.8231, "step": 24 }, { "epoch": 0.7003367003367004, "grad_norm": 2.707709550857544, "learning_rate": 7.818181818181818e-05, "loss": 3.7116, "step": 26 }, { "epoch": 0.7542087542087542, "grad_norm": 3.31360125541687, "learning_rate": 7.636363636363637e-05, "loss": 3.5999, "step": 28 }, { "epoch": 0.8080808080808081, "grad_norm": 2.9818968772888184, "learning_rate": 7.454545454545455e-05, "loss": 3.5958, "step": 30 }, { "epoch": 0.8619528619528619, "grad_norm": 2.68215274810791, "learning_rate": 7.272727272727273e-05, "loss": 3.5518, "step": 32 }, { "epoch": 0.9158249158249159, "grad_norm": 1.9956828355789185, "learning_rate": 7.090909090909092e-05, "loss": 3.4949, "step": 34 }, { "epoch": 0.9696969696969697, "grad_norm": 1.9735620021820068, "learning_rate": 6.90909090909091e-05, "loss": 3.4185, "step": 36 }, { "epoch": 1.0235690235690236, "grad_norm": 1.8688853979110718, "learning_rate": 6.727272727272727e-05, "loss": 3.4284, "step": 38 }, { "epoch": 1.0774410774410774, "grad_norm": 1.9668887853622437, "learning_rate": 6.545454545454546e-05, "loss": 3.3888, "step": 40 }, { "epoch": 1.0774410774410774, "eval_loss": 3.2203786373138428, "eval_runtime": 8.1727, "eval_samples_per_second": 145.484, "eval_steps_per_second": 18.231, "step": 40 }, { "epoch": 1.1313131313131313, "grad_norm": 1.8594480752944946, "learning_rate": 6.363636363636364e-05, "loss": 3.3016, "step": 42 }, { "epoch": 1.1851851851851851, "grad_norm": 2.0851259231567383, "learning_rate": 6.181818181818182e-05, "loss": 3.2937, "step": 44 }, { "epoch": 1.239057239057239, "grad_norm": 2.022775411605835, "learning_rate": 6e-05, "loss": 3.3097, "step": 46 }, { "epoch": 1.2929292929292928, "grad_norm": 1.7249592542648315, "learning_rate": 5.818181818181818e-05, "loss": 3.2856, "step": 48 }, { "epoch": 1.3468013468013469, "grad_norm": 1.894808292388916, "learning_rate": 5.636363636363636e-05, "loss": 3.2579, "step": 50 }, { "epoch": 1.4006734006734007, "grad_norm": 2.0816519260406494, "learning_rate": 5.4545454545454546e-05, "loss": 3.2534, "step": 52 }, { "epoch": 1.4545454545454546, "grad_norm": 1.8632744550704956, "learning_rate": 5.272727272727272e-05, "loss": 3.2358, "step": 54 }, { "epoch": 1.5084175084175084, "grad_norm": 1.8851394653320312, "learning_rate": 5.090909090909091e-05, "loss": 3.2256, "step": 56 }, { "epoch": 1.5622895622895623, "grad_norm": 1.7529875040054321, "learning_rate": 4.909090909090909e-05, "loss": 3.2148, "step": 58 }, { "epoch": 1.6161616161616161, "grad_norm": 1.8536018133163452, "learning_rate": 4.7272727272727275e-05, "loss": 3.2608, "step": 60 }, { "epoch": 1.6161616161616161, "eval_loss": 3.1798043251037598, "eval_runtime": 8.1729, "eval_samples_per_second": 145.48, "eval_steps_per_second": 18.231, "step": 60 }, { "epoch": 1.67003367003367, "grad_norm": 1.9621059894561768, "learning_rate": 4.545454545454546e-05, "loss": 3.2202, "step": 62 }, { "epoch": 1.723905723905724, "grad_norm": 1.9148567914962769, "learning_rate": 4.3636363636363636e-05, "loss": 3.2282, "step": 64 }, { "epoch": 1.7777777777777777, "grad_norm": 1.7804056406021118, "learning_rate": 4.181818181818182e-05, "loss": 3.2654, "step": 66 }, { "epoch": 1.8316498316498318, "grad_norm": 1.8891396522521973, "learning_rate": 4e-05, "loss": 3.1924, "step": 68 }, { "epoch": 1.8855218855218854, "grad_norm": 1.8429640531539917, "learning_rate": 3.818181818181819e-05, "loss": 3.1756, "step": 70 }, { "epoch": 1.9393939393939394, "grad_norm": 1.795774221420288, "learning_rate": 3.6363636363636364e-05, "loss": 3.2097, "step": 72 }, { "epoch": 1.9932659932659933, "grad_norm": 1.8442139625549316, "learning_rate": 3.454545454545455e-05, "loss": 3.2271, "step": 74 }, { "epoch": 2.047138047138047, "grad_norm": 1.6865243911743164, "learning_rate": 3.272727272727273e-05, "loss": 3.1261, "step": 76 }, { "epoch": 2.101010101010101, "grad_norm": 1.6551978588104248, "learning_rate": 3.090909090909091e-05, "loss": 3.101, "step": 78 }, { "epoch": 2.154882154882155, "grad_norm": 1.7607771158218384, "learning_rate": 2.909090909090909e-05, "loss": 3.0763, "step": 80 }, { "epoch": 2.154882154882155, "eval_loss": 3.162412405014038, "eval_runtime": 8.1804, "eval_samples_per_second": 145.348, "eval_steps_per_second": 18.214, "step": 80 }, { "epoch": 2.208754208754209, "grad_norm": 1.720595359802246, "learning_rate": 2.7272727272727273e-05, "loss": 3.1138, "step": 82 }, { "epoch": 2.2626262626262625, "grad_norm": 1.7866605520248413, "learning_rate": 2.5454545454545454e-05, "loss": 3.1239, "step": 84 }, { "epoch": 2.3164983164983166, "grad_norm": 1.7226550579071045, "learning_rate": 2.3636363636363637e-05, "loss": 3.0789, "step": 86 }, { "epoch": 2.3703703703703702, "grad_norm": 1.8532516956329346, "learning_rate": 2.1818181818181818e-05, "loss": 3.1286, "step": 88 }, { "epoch": 2.4242424242424243, "grad_norm": 1.7198731899261475, "learning_rate": 2e-05, "loss": 3.1123, "step": 90 }, { "epoch": 2.478114478114478, "grad_norm": 1.8436646461486816, "learning_rate": 1.8181818181818182e-05, "loss": 3.0712, "step": 92 }, { "epoch": 2.531986531986532, "grad_norm": 1.811488151550293, "learning_rate": 1.6363636363636366e-05, "loss": 3.0727, "step": 94 }, { "epoch": 2.5858585858585856, "grad_norm": 1.7093764543533325, "learning_rate": 1.4545454545454545e-05, "loss": 3.0966, "step": 96 }, { "epoch": 2.6397306397306397, "grad_norm": 1.8103761672973633, "learning_rate": 1.2727272727272727e-05, "loss": 3.084, "step": 98 }, { "epoch": 2.6936026936026938, "grad_norm": 1.8916672468185425, "learning_rate": 1.0909090909090909e-05, "loss": 3.0912, "step": 100 }, { "epoch": 2.6936026936026938, "eval_loss": 3.153109550476074, "eval_runtime": 8.1741, "eval_samples_per_second": 145.459, "eval_steps_per_second": 18.228, "step": 100 }, { "epoch": 2.7474747474747474, "grad_norm": 1.7815495729446411, "learning_rate": 9.090909090909091e-06, "loss": 3.0743, "step": 102 }, { "epoch": 2.8013468013468015, "grad_norm": 1.8348023891448975, "learning_rate": 7.272727272727272e-06, "loss": 3.092, "step": 104 }, { "epoch": 2.855218855218855, "grad_norm": 1.807790756225586, "learning_rate": 5.4545454545454545e-06, "loss": 3.0913, "step": 106 }, { "epoch": 2.909090909090909, "grad_norm": 1.7084001302719116, "learning_rate": 3.636363636363636e-06, "loss": 3.07, "step": 108 }, { "epoch": 2.962962962962963, "grad_norm": 1.719460129737854, "learning_rate": 1.818181818181818e-06, "loss": 3.0812, "step": 110 }, { "epoch": 2.98989898989899, "step": 111, "total_flos": 1.389220151427072e+16, "train_loss": 3.7213772073522344, "train_runtime": 747.8744, "train_samples_per_second": 38.144, "train_steps_per_second": 0.148 }, { "epoch": 2.98989898989899, "eval_loss": 3.1522939205169678, "eval_runtime": 8.1573, "eval_samples_per_second": 145.759, "eval_steps_per_second": 18.266, "step": 111 }, { "epoch": 2.98989898989899, "eval_loss": 3.1559650897979736, "eval_runtime": 8.1292, "eval_samples_per_second": 146.262, "eval_steps_per_second": 18.329, "step": 111 } ], "logging_steps": 2, "max_steps": 111, "num_input_tokens_seen": 0, "num_train_epochs": 3, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 1.389220151427072e+16, "train_batch_size": 32, "trial_name": null, "trial_params": null }