| { | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 12.421052631578947, | |
| "eval_steps": 500, | |
| "global_step": 56, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.21052631578947367, | |
| "grad_norm": 19.218000411987305, | |
| "learning_rate": 2e-05, | |
| "loss": 3.886, | |
| "step": 1 | |
| }, | |
| { | |
| "epoch": 0.42105263157894735, | |
| "grad_norm": 19.777868270874023, | |
| "learning_rate": 4e-05, | |
| "loss": 4.4615, | |
| "step": 2 | |
| }, | |
| { | |
| "epoch": 0.631578947368421, | |
| "grad_norm": 17.180343627929688, | |
| "learning_rate": 6e-05, | |
| "loss": 3.7595, | |
| "step": 3 | |
| }, | |
| { | |
| "epoch": 0.8421052631578947, | |
| "grad_norm": 17.781330108642578, | |
| "learning_rate": 8e-05, | |
| "loss": 3.3302, | |
| "step": 4 | |
| }, | |
| { | |
| "epoch": 1.1052631578947367, | |
| "grad_norm": 20.543649673461914, | |
| "learning_rate": 0.0001, | |
| "loss": 4.6154, | |
| "step": 5 | |
| }, | |
| { | |
| "epoch": 1.3157894736842106, | |
| "grad_norm": 11.29702091217041, | |
| "learning_rate": 0.00012, | |
| "loss": 2.5755, | |
| "step": 6 | |
| }, | |
| { | |
| "epoch": 1.526315789473684, | |
| "grad_norm": 8.194554328918457, | |
| "learning_rate": 0.00014, | |
| "loss": 1.6726, | |
| "step": 7 | |
| }, | |
| { | |
| "epoch": 1.736842105263158, | |
| "grad_norm": 6.040785312652588, | |
| "learning_rate": 0.00016, | |
| "loss": 1.2606, | |
| "step": 8 | |
| }, | |
| { | |
| "epoch": 1.9473684210526314, | |
| "grad_norm": 12.296614646911621, | |
| "learning_rate": 0.00018, | |
| "loss": 1.9246, | |
| "step": 9 | |
| }, | |
| { | |
| "epoch": 2.2105263157894735, | |
| "grad_norm": 6.707492828369141, | |
| "learning_rate": 0.0002, | |
| "loss": 1.0017, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 2.4210526315789473, | |
| "grad_norm": 4.79848575592041, | |
| "learning_rate": 0.00019976687691905393, | |
| "loss": 0.6978, | |
| "step": 11 | |
| }, | |
| { | |
| "epoch": 2.6315789473684212, | |
| "grad_norm": 3.3441014289855957, | |
| "learning_rate": 0.00019906859460363307, | |
| "loss": 0.5996, | |
| "step": 12 | |
| }, | |
| { | |
| "epoch": 2.8421052631578947, | |
| "grad_norm": 3.113868474960327, | |
| "learning_rate": 0.00019790840876823232, | |
| "loss": 0.5782, | |
| "step": 13 | |
| }, | |
| { | |
| "epoch": 3.1052631578947367, | |
| "grad_norm": 7.5327229499816895, | |
| "learning_rate": 0.00019629172873477995, | |
| "loss": 0.8014, | |
| "step": 14 | |
| }, | |
| { | |
| "epoch": 3.3157894736842106, | |
| "grad_norm": 2.6903626918792725, | |
| "learning_rate": 0.00019422609221188207, | |
| "loss": 0.3285, | |
| "step": 15 | |
| }, | |
| { | |
| "epoch": 3.526315789473684, | |
| "grad_norm": 3.672346591949463, | |
| "learning_rate": 0.00019172113015054532, | |
| "loss": 0.3814, | |
| "step": 16 | |
| }, | |
| { | |
| "epoch": 3.736842105263158, | |
| "grad_norm": 1.4505695104599, | |
| "learning_rate": 0.0001887885218402375, | |
| "loss": 0.2401, | |
| "step": 17 | |
| }, | |
| { | |
| "epoch": 3.9473684210526314, | |
| "grad_norm": 2.6290504932403564, | |
| "learning_rate": 0.00018544194045464886, | |
| "loss": 0.3272, | |
| "step": 18 | |
| }, | |
| { | |
| "epoch": 4.2105263157894735, | |
| "grad_norm": 1.3445160388946533, | |
| "learning_rate": 0.0001816969893010442, | |
| "loss": 0.1871, | |
| "step": 19 | |
| }, | |
| { | |
| "epoch": 4.421052631578947, | |
| "grad_norm": 1.851083755493164, | |
| "learning_rate": 0.000177571129070442, | |
| "loss": 0.1781, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 4.631578947368421, | |
| "grad_norm": 1.0952624082565308, | |
| "learning_rate": 0.00017308359642781242, | |
| "loss": 0.1966, | |
| "step": 21 | |
| }, | |
| { | |
| "epoch": 4.842105263157895, | |
| "grad_norm": 2.4012107849121094, | |
| "learning_rate": 0.00016825531432186543, | |
| "loss": 0.1738, | |
| "step": 22 | |
| }, | |
| { | |
| "epoch": 5.105263157894737, | |
| "grad_norm": 1.6361955404281616, | |
| "learning_rate": 0.00016310879443260528, | |
| "loss": 0.2131, | |
| "step": 23 | |
| }, | |
| { | |
| "epoch": 5.315789473684211, | |
| "grad_norm": 2.245840072631836, | |
| "learning_rate": 0.00015766803221148673, | |
| "loss": 0.1646, | |
| "step": 24 | |
| }, | |
| { | |
| "epoch": 5.526315789473684, | |
| "grad_norm": 1.684137225151062, | |
| "learning_rate": 0.00015195839500354335, | |
| "loss": 0.184, | |
| "step": 25 | |
| }, | |
| { | |
| "epoch": 5.7368421052631575, | |
| "grad_norm": 0.9733522534370422, | |
| "learning_rate": 0.00014600650377311522, | |
| "loss": 0.1465, | |
| "step": 26 | |
| }, | |
| { | |
| "epoch": 5.947368421052632, | |
| "grad_norm": 2.4283876419067383, | |
| "learning_rate": 0.00013984010898462416, | |
| "loss": 0.2177, | |
| "step": 27 | |
| }, | |
| { | |
| "epoch": 6.2105263157894735, | |
| "grad_norm": 1.4879282712936401, | |
| "learning_rate": 0.00013348796121709862, | |
| "loss": 0.1299, | |
| "step": 28 | |
| }, | |
| { | |
| "epoch": 6.421052631578947, | |
| "grad_norm": 1.0862969160079956, | |
| "learning_rate": 0.00012697967711570242, | |
| "loss": 0.1115, | |
| "step": 29 | |
| }, | |
| { | |
| "epoch": 6.631578947368421, | |
| "grad_norm": 0.7216919660568237, | |
| "learning_rate": 0.0001203456013052634, | |
| "loss": 0.0971, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 6.842105263157895, | |
| "grad_norm": 1.4882885217666626, | |
| "learning_rate": 0.00011361666490962468, | |
| "loss": 0.1453, | |
| "step": 31 | |
| }, | |
| { | |
| "epoch": 7.105263157894737, | |
| "grad_norm": 1.220837950706482, | |
| "learning_rate": 0.0001068242413364671, | |
| "loss": 0.1634, | |
| "step": 32 | |
| }, | |
| { | |
| "epoch": 7.315789473684211, | |
| "grad_norm": 0.6117944121360779, | |
| "learning_rate": 0.0001, | |
| "loss": 0.1026, | |
| "step": 33 | |
| }, | |
| { | |
| "epoch": 7.526315789473684, | |
| "grad_norm": 0.9110010266304016, | |
| "learning_rate": 9.317575866353292e-05, | |
| "loss": 0.103, | |
| "step": 34 | |
| }, | |
| { | |
| "epoch": 7.7368421052631575, | |
| "grad_norm": 0.9700250625610352, | |
| "learning_rate": 8.638333509037536e-05, | |
| "loss": 0.1205, | |
| "step": 35 | |
| }, | |
| { | |
| "epoch": 7.947368421052632, | |
| "grad_norm": 0.9124184250831604, | |
| "learning_rate": 7.965439869473664e-05, | |
| "loss": 0.1095, | |
| "step": 36 | |
| }, | |
| { | |
| "epoch": 8.210526315789474, | |
| "grad_norm": 0.49950140714645386, | |
| "learning_rate": 7.302032288429756e-05, | |
| "loss": 0.0796, | |
| "step": 37 | |
| }, | |
| { | |
| "epoch": 8.421052631578947, | |
| "grad_norm": 0.8700776100158691, | |
| "learning_rate": 6.651203878290139e-05, | |
| "loss": 0.0774, | |
| "step": 38 | |
| }, | |
| { | |
| "epoch": 8.631578947368421, | |
| "grad_norm": 0.7889358401298523, | |
| "learning_rate": 6.015989101537586e-05, | |
| "loss": 0.0835, | |
| "step": 39 | |
| }, | |
| { | |
| "epoch": 8.842105263157894, | |
| "grad_norm": 0.5236338973045349, | |
| "learning_rate": 5.399349622688479e-05, | |
| "loss": 0.0581, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 9.105263157894736, | |
| "grad_norm": 0.9376251697540283, | |
| "learning_rate": 4.804160499645667e-05, | |
| "loss": 0.0936, | |
| "step": 41 | |
| }, | |
| { | |
| "epoch": 9.31578947368421, | |
| "grad_norm": 0.7259882092475891, | |
| "learning_rate": 4.2331967788513295e-05, | |
| "loss": 0.0712, | |
| "step": 42 | |
| }, | |
| { | |
| "epoch": 9.526315789473685, | |
| "grad_norm": 0.5035570859909058, | |
| "learning_rate": 3.689120556739475e-05, | |
| "loss": 0.0776, | |
| "step": 43 | |
| }, | |
| { | |
| "epoch": 9.736842105263158, | |
| "grad_norm": 0.4928078055381775, | |
| "learning_rate": 3.174468567813461e-05, | |
| "loss": 0.0483, | |
| "step": 44 | |
| }, | |
| { | |
| "epoch": 9.947368421052632, | |
| "grad_norm": 1.1208339929580688, | |
| "learning_rate": 2.691640357218759e-05, | |
| "loss": 0.1032, | |
| "step": 45 | |
| }, | |
| { | |
| "epoch": 10.210526315789474, | |
| "grad_norm": 0.5036032795906067, | |
| "learning_rate": 2.242887092955801e-05, | |
| "loss": 0.0745, | |
| "step": 46 | |
| }, | |
| { | |
| "epoch": 10.421052631578947, | |
| "grad_norm": 0.4454878866672516, | |
| "learning_rate": 1.8303010698955804e-05, | |
| "loss": 0.0526, | |
| "step": 47 | |
| }, | |
| { | |
| "epoch": 10.631578947368421, | |
| "grad_norm": 0.48385077714920044, | |
| "learning_rate": 1.4558059545351143e-05, | |
| "loss": 0.0609, | |
| "step": 48 | |
| }, | |
| { | |
| "epoch": 10.842105263157894, | |
| "grad_norm": 0.8407765030860901, | |
| "learning_rate": 1.1211478159762478e-05, | |
| "loss": 0.0638, | |
| "step": 49 | |
| }, | |
| { | |
| "epoch": 11.105263157894736, | |
| "grad_norm": 0.4083283841609955, | |
| "learning_rate": 8.278869849454718e-06, | |
| "loss": 0.065, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 11.31578947368421, | |
| "grad_norm": 0.4340791702270508, | |
| "learning_rate": 5.77390778811796e-06, | |
| "loss": 0.0545, | |
| "step": 51 | |
| }, | |
| { | |
| "epoch": 11.526315789473685, | |
| "grad_norm": 0.5537849068641663, | |
| "learning_rate": 3.7082712652200867e-06, | |
| "loss": 0.0511, | |
| "step": 52 | |
| }, | |
| { | |
| "epoch": 11.736842105263158, | |
| "grad_norm": 0.36356380581855774, | |
| "learning_rate": 2.091591231767709e-06, | |
| "loss": 0.0553, | |
| "step": 53 | |
| }, | |
| { | |
| "epoch": 11.947368421052632, | |
| "grad_norm": 1.3945286273956299, | |
| "learning_rate": 9.314053963669245e-07, | |
| "loss": 0.1255, | |
| "step": 54 | |
| }, | |
| { | |
| "epoch": 12.210526315789474, | |
| "grad_norm": 0.5195576548576355, | |
| "learning_rate": 2.3312308094607382e-07, | |
| "loss": 0.0559, | |
| "step": 55 | |
| }, | |
| { | |
| "epoch": 12.421052631578947, | |
| "grad_norm": 0.5838670134544373, | |
| "learning_rate": 0.0, | |
| "loss": 0.0631, | |
| "step": 56 | |
| } | |
| ], | |
| "logging_steps": 1, | |
| "max_steps": 56, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 14, | |
| "save_steps": 4, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": true | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 50350673362944.0, | |
| "train_batch_size": 1, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |