| { |
| "best_global_step": null, |
| "best_metric": null, |
| "best_model_checkpoint": null, |
| "epoch": 0.003688845852353945, |
| "eval_steps": 10, |
| "global_step": 60, |
| "is_hyper_param_search": false, |
| "is_local_process_zero": true, |
| "is_world_process_zero": true, |
| "log_history": [ |
| { |
| "epoch": 6.148076420589908e-05, |
| "grad_norm": 0.003479025326669216, |
| "learning_rate": 0.0, |
| "loss": 0.0, |
| "step": 1 |
| }, |
| { |
| "epoch": 0.00012296152841179817, |
| "grad_norm": 0.005437533836811781, |
| "learning_rate": 4e-05, |
| "loss": 0.0001, |
| "step": 2 |
| }, |
| { |
| "epoch": 0.00018444229261769724, |
| "grad_norm": 0.00029893239843659103, |
| "learning_rate": 8e-05, |
| "loss": 0.0, |
| "step": 3 |
| }, |
| { |
| "epoch": 0.00024592305682359633, |
| "grad_norm": 0.00016204842540901154, |
| "learning_rate": 0.00012, |
| "loss": 0.0, |
| "step": 4 |
| }, |
| { |
| "epoch": 0.0003074038210294954, |
| "grad_norm": 0.00012208402040414512, |
| "learning_rate": 0.00016, |
| "loss": 0.0, |
| "step": 5 |
| }, |
| { |
| "epoch": 0.0003688845852353945, |
| "grad_norm": 0.0022303895093500614, |
| "learning_rate": 0.0002, |
| "loss": 0.0, |
| "step": 6 |
| }, |
| { |
| "epoch": 0.00043036534944129355, |
| "grad_norm": 0.0020995561499148607, |
| "learning_rate": 0.00019636363636363636, |
| "loss": 0.0, |
| "step": 7 |
| }, |
| { |
| "epoch": 0.0004918461136471927, |
| "grad_norm": 0.008035325445234776, |
| "learning_rate": 0.00019272727272727274, |
| "loss": 0.0001, |
| "step": 8 |
| }, |
| { |
| "epoch": 0.0005533268778530917, |
| "grad_norm": 0.05308253690600395, |
| "learning_rate": 0.0001890909090909091, |
| "loss": 0.0026, |
| "step": 9 |
| }, |
| { |
| "epoch": 0.0006148076420589908, |
| "grad_norm": 0.005033898167312145, |
| "learning_rate": 0.00018545454545454545, |
| "loss": 0.0002, |
| "step": 10 |
| }, |
| { |
| "epoch": 0.0006762884062648898, |
| "grad_norm": 0.000410953420214355, |
| "learning_rate": 0.00018181818181818183, |
| "loss": 0.0, |
| "step": 11 |
| }, |
| { |
| "epoch": 0.000737769170470789, |
| "grad_norm": 0.044024836272001266, |
| "learning_rate": 0.0001781818181818182, |
| "loss": 0.0004, |
| "step": 12 |
| }, |
| { |
| "epoch": 0.0007992499346766881, |
| "grad_norm": 0.01820911094546318, |
| "learning_rate": 0.00017454545454545454, |
| "loss": 0.0001, |
| "step": 13 |
| }, |
| { |
| "epoch": 0.0008607306988825871, |
| "grad_norm": 0.012782749719917774, |
| "learning_rate": 0.0001709090909090909, |
| "loss": 0.0002, |
| "step": 14 |
| }, |
| { |
| "epoch": 0.0009222114630884862, |
| "grad_norm": 0.0022124142851680517, |
| "learning_rate": 0.00016727272727272728, |
| "loss": 0.0, |
| "step": 15 |
| }, |
| { |
| "epoch": 0.0009836922272943853, |
| "grad_norm": 0.007108100224286318, |
| "learning_rate": 0.00016363636363636366, |
| "loss": 0.0001, |
| "step": 16 |
| }, |
| { |
| "epoch": 0.0010451729915002842, |
| "grad_norm": 0.000606991583481431, |
| "learning_rate": 0.00016, |
| "loss": 0.0, |
| "step": 17 |
| }, |
| { |
| "epoch": 0.0011066537557061834, |
| "grad_norm": 0.005945051088929176, |
| "learning_rate": 0.00015636363636363637, |
| "loss": 0.0002, |
| "step": 18 |
| }, |
| { |
| "epoch": 0.0011681345199120825, |
| "grad_norm": 0.05237742140889168, |
| "learning_rate": 0.00015272727272727275, |
| "loss": 0.0007, |
| "step": 19 |
| }, |
| { |
| "epoch": 0.0012296152841179816, |
| "grad_norm": 0.01273553166538477, |
| "learning_rate": 0.0001490909090909091, |
| "loss": 0.0002, |
| "step": 20 |
| }, |
| { |
| "epoch": 0.0012910960483238807, |
| "grad_norm": 0.0036445350851863623, |
| "learning_rate": 0.00014545454545454546, |
| "loss": 0.0001, |
| "step": 21 |
| }, |
| { |
| "epoch": 0.0013525768125297797, |
| "grad_norm": 0.08508655428886414, |
| "learning_rate": 0.00014181818181818184, |
| "loss": 0.0083, |
| "step": 22 |
| }, |
| { |
| "epoch": 0.0014140575767356788, |
| "grad_norm": 0.043661970645189285, |
| "learning_rate": 0.0001381818181818182, |
| "loss": 0.004, |
| "step": 23 |
| }, |
| { |
| "epoch": 0.001475538340941578, |
| "grad_norm": 0.000659774465020746, |
| "learning_rate": 0.00013454545454545455, |
| "loss": 0.0, |
| "step": 24 |
| }, |
| { |
| "epoch": 0.001537019105147477, |
| "grad_norm": 0.0005294107249937952, |
| "learning_rate": 0.00013090909090909093, |
| "loss": 0.0, |
| "step": 25 |
| }, |
| { |
| "epoch": 0.0015984998693533761, |
| "grad_norm": 0.00016782127204351127, |
| "learning_rate": 0.00012727272727272728, |
| "loss": 0.0, |
| "step": 26 |
| }, |
| { |
| "epoch": 0.001659980633559275, |
| "grad_norm": 0.041384413838386536, |
| "learning_rate": 0.00012363636363636364, |
| "loss": 0.0008, |
| "step": 27 |
| }, |
| { |
| "epoch": 0.0017214613977651742, |
| "grad_norm": 0.009205852635204792, |
| "learning_rate": 0.00012, |
| "loss": 0.0001, |
| "step": 28 |
| }, |
| { |
| "epoch": 0.0017829421619710733, |
| "grad_norm": 0.00016076133761089295, |
| "learning_rate": 0.00011636363636363636, |
| "loss": 0.0, |
| "step": 29 |
| }, |
| { |
| "epoch": 0.0018444229261769724, |
| "grad_norm": 0.0002819328219629824, |
| "learning_rate": 0.00011272727272727272, |
| "loss": 0.0, |
| "step": 30 |
| }, |
| { |
| "epoch": 0.0019059036903828716, |
| "grad_norm": 0.00022894897847436368, |
| "learning_rate": 0.00010909090909090909, |
| "loss": 0.0, |
| "step": 31 |
| }, |
| { |
| "epoch": 0.0019673844545887707, |
| "grad_norm": 0.014002679847180843, |
| "learning_rate": 0.00010545454545454545, |
| "loss": 0.0022, |
| "step": 32 |
| }, |
| { |
| "epoch": 0.00202886521879467, |
| "grad_norm": 0.00016014272114261985, |
| "learning_rate": 0.00010181818181818181, |
| "loss": 0.0, |
| "step": 33 |
| }, |
| { |
| "epoch": 0.0020903459830005685, |
| "grad_norm": 0.20411507785320282, |
| "learning_rate": 9.818181818181818e-05, |
| "loss": 0.0012, |
| "step": 34 |
| }, |
| { |
| "epoch": 0.0021518267472064676, |
| "grad_norm": 0.013666906394064426, |
| "learning_rate": 9.454545454545455e-05, |
| "loss": 0.0001, |
| "step": 35 |
| }, |
| { |
| "epoch": 0.0022133075114123667, |
| "grad_norm": 0.005049742292612791, |
| "learning_rate": 9.090909090909092e-05, |
| "loss": 0.0001, |
| "step": 36 |
| }, |
| { |
| "epoch": 0.002274788275618266, |
| "grad_norm": 0.0020620787981897593, |
| "learning_rate": 8.727272727272727e-05, |
| "loss": 0.0, |
| "step": 37 |
| }, |
| { |
| "epoch": 0.002336269039824165, |
| "grad_norm": 0.023437755182385445, |
| "learning_rate": 8.363636363636364e-05, |
| "loss": 0.0009, |
| "step": 38 |
| }, |
| { |
| "epoch": 0.002397749804030064, |
| "grad_norm": 0.0012406132882460952, |
| "learning_rate": 8e-05, |
| "loss": 0.0001, |
| "step": 39 |
| }, |
| { |
| "epoch": 0.0024592305682359632, |
| "grad_norm": 0.0004515462787821889, |
| "learning_rate": 7.636363636363637e-05, |
| "loss": 0.0, |
| "step": 40 |
| }, |
| { |
| "epoch": 0.0025207113324418624, |
| "grad_norm": 0.0053300075232982635, |
| "learning_rate": 7.272727272727273e-05, |
| "loss": 0.0, |
| "step": 41 |
| }, |
| { |
| "epoch": 0.0025821920966477615, |
| "grad_norm": 0.012812405824661255, |
| "learning_rate": 6.90909090909091e-05, |
| "loss": 0.0005, |
| "step": 42 |
| }, |
| { |
| "epoch": 0.0026436728608536606, |
| "grad_norm": 0.0002906050067394972, |
| "learning_rate": 6.545454545454546e-05, |
| "loss": 0.0, |
| "step": 43 |
| }, |
| { |
| "epoch": 0.0027051536250595593, |
| "grad_norm": 0.0020773536525666714, |
| "learning_rate": 6.181818181818182e-05, |
| "loss": 0.0001, |
| "step": 44 |
| }, |
| { |
| "epoch": 0.0027666343892654584, |
| "grad_norm": 0.0004460048221517354, |
| "learning_rate": 5.818181818181818e-05, |
| "loss": 0.0, |
| "step": 45 |
| }, |
| { |
| "epoch": 0.0028281151534713576, |
| "grad_norm": 0.004530001897364855, |
| "learning_rate": 5.4545454545454546e-05, |
| "loss": 0.0001, |
| "step": 46 |
| }, |
| { |
| "epoch": 0.0028895959176772567, |
| "grad_norm": 0.0008538271649740636, |
| "learning_rate": 5.090909090909091e-05, |
| "loss": 0.0, |
| "step": 47 |
| }, |
| { |
| "epoch": 0.002951076681883156, |
| "grad_norm": 0.01173599623143673, |
| "learning_rate": 4.7272727272727275e-05, |
| "loss": 0.0006, |
| "step": 48 |
| }, |
| { |
| "epoch": 0.003012557446089055, |
| "grad_norm": 0.10636617243289948, |
| "learning_rate": 4.3636363636363636e-05, |
| "loss": 0.009, |
| "step": 49 |
| }, |
| { |
| "epoch": 0.003074038210294954, |
| "grad_norm": 0.019629845395684242, |
| "learning_rate": 4e-05, |
| "loss": 0.0004, |
| "step": 50 |
| }, |
| { |
| "epoch": 0.003135518974500853, |
| "grad_norm": 0.025700179859995842, |
| "learning_rate": 3.6363636363636364e-05, |
| "loss": 0.0001, |
| "step": 51 |
| }, |
| { |
| "epoch": 0.0031969997387067523, |
| "grad_norm": 0.004359859973192215, |
| "learning_rate": 3.272727272727273e-05, |
| "loss": 0.0, |
| "step": 52 |
| }, |
| { |
| "epoch": 0.003258480502912651, |
| "grad_norm": 0.0004983420949429274, |
| "learning_rate": 2.909090909090909e-05, |
| "loss": 0.0, |
| "step": 53 |
| }, |
| { |
| "epoch": 0.00331996126711855, |
| "grad_norm": 0.0008296012529172003, |
| "learning_rate": 2.5454545454545454e-05, |
| "loss": 0.0, |
| "step": 54 |
| }, |
| { |
| "epoch": 0.0033814420313244492, |
| "grad_norm": 0.0012659059138968587, |
| "learning_rate": 2.1818181818181818e-05, |
| "loss": 0.0, |
| "step": 55 |
| }, |
| { |
| "epoch": 0.0034429227955303484, |
| "grad_norm": 0.0003076361317653209, |
| "learning_rate": 1.8181818181818182e-05, |
| "loss": 0.0, |
| "step": 56 |
| }, |
| { |
| "epoch": 0.0035044035597362475, |
| "grad_norm": 0.0007758835563436151, |
| "learning_rate": 1.4545454545454545e-05, |
| "loss": 0.0, |
| "step": 57 |
| }, |
| { |
| "epoch": 0.0035658843239421466, |
| "grad_norm": 0.00053775793639943, |
| "learning_rate": 1.0909090909090909e-05, |
| "loss": 0.0, |
| "step": 58 |
| }, |
| { |
| "epoch": 0.0036273650881480457, |
| "grad_norm": 0.0034126462414860725, |
| "learning_rate": 7.272727272727272e-06, |
| "loss": 0.0002, |
| "step": 59 |
| }, |
| { |
| "epoch": 0.003688845852353945, |
| "grad_norm": 0.0009407071629539132, |
| "learning_rate": 3.636363636363636e-06, |
| "loss": 0.0, |
| "step": 60 |
| } |
| ], |
| "logging_steps": 1, |
| "max_steps": 60, |
| "num_input_tokens_seen": 0, |
| "num_train_epochs": 1, |
| "save_steps": 500, |
| "stateful_callbacks": { |
| "TrainerControl": { |
| "args": { |
| "should_epoch_stop": false, |
| "should_evaluate": false, |
| "should_log": false, |
| "should_save": true, |
| "should_training_stop": true |
| }, |
| "attributes": {} |
| } |
| }, |
| "total_flos": 1.12138907332608e+16, |
| "train_batch_size": 2, |
| "trial_name": null, |
| "trial_params": null |
| } |
|
|