| { |
| "best_global_step": null, |
| "best_metric": null, |
| "best_model_checkpoint": null, |
| "epoch": 15.0, |
| "eval_steps": 500, |
| "global_step": 375, |
| "is_hyper_param_search": false, |
| "is_local_process_zero": true, |
| "is_world_process_zero": true, |
| "log_history": [ |
| { |
| "epoch": 0.2, |
| "grad_norm": 1.3508174419403076, |
| "learning_rate": 0.0002, |
| "loss": 2.5227, |
| "step": 5 |
| }, |
| { |
| "epoch": 0.4, |
| "grad_norm": 1.8602172136306763, |
| "learning_rate": 0.0002, |
| "loss": 2.6438, |
| "step": 10 |
| }, |
| { |
| "epoch": 0.6, |
| "grad_norm": 1.7185380458831787, |
| "learning_rate": 0.0002, |
| "loss": 0.7979, |
| "step": 15 |
| }, |
| { |
| "epoch": 0.8, |
| "grad_norm": 1.46760892868042, |
| "learning_rate": 0.0002, |
| "loss": 0.4091, |
| "step": 20 |
| }, |
| { |
| "epoch": 1.0, |
| "grad_norm": 0.5913628935813904, |
| "learning_rate": 0.0002, |
| "loss": 0.274, |
| "step": 25 |
| }, |
| { |
| "epoch": 1.2, |
| "grad_norm": 0.46462497115135193, |
| "learning_rate": 0.0002, |
| "loss": 0.1921, |
| "step": 30 |
| }, |
| { |
| "epoch": 1.4, |
| "grad_norm": 0.47151607275009155, |
| "learning_rate": 0.0002, |
| "loss": 0.2399, |
| "step": 35 |
| }, |
| { |
| "epoch": 1.6, |
| "grad_norm": 0.45026347041130066, |
| "learning_rate": 0.0002, |
| "loss": 0.5266, |
| "step": 40 |
| }, |
| { |
| "epoch": 1.8, |
| "grad_norm": 0.8638098835945129, |
| "learning_rate": 0.0002, |
| "loss": 0.1596, |
| "step": 45 |
| }, |
| { |
| "epoch": 2.0, |
| "grad_norm": 0.41460809111595154, |
| "learning_rate": 0.0002, |
| "loss": 0.291, |
| "step": 50 |
| }, |
| { |
| "epoch": 2.2, |
| "grad_norm": 0.30362772941589355, |
| "learning_rate": 0.0002, |
| "loss": 0.0818, |
| "step": 55 |
| }, |
| { |
| "epoch": 2.4, |
| "grad_norm": 0.46699783205986023, |
| "learning_rate": 0.0002, |
| "loss": 0.1198, |
| "step": 60 |
| }, |
| { |
| "epoch": 2.6, |
| "grad_norm": 0.3688370883464813, |
| "learning_rate": 0.0002, |
| "loss": 0.0892, |
| "step": 65 |
| }, |
| { |
| "epoch": 2.8, |
| "grad_norm": 0.6380140781402588, |
| "learning_rate": 0.0002, |
| "loss": 0.1068, |
| "step": 70 |
| }, |
| { |
| "epoch": 3.0, |
| "grad_norm": 0.31224119663238525, |
| "learning_rate": 0.0002, |
| "loss": 0.1673, |
| "step": 75 |
| }, |
| { |
| "epoch": 3.2, |
| "grad_norm": 0.455768883228302, |
| "learning_rate": 0.0002, |
| "loss": 0.0405, |
| "step": 80 |
| }, |
| { |
| "epoch": 3.4, |
| "grad_norm": 0.29344627261161804, |
| "learning_rate": 0.0002, |
| "loss": 0.1494, |
| "step": 85 |
| }, |
| { |
| "epoch": 3.6, |
| "grad_norm": 0.4940664768218994, |
| "learning_rate": 0.0002, |
| "loss": 0.0483, |
| "step": 90 |
| }, |
| { |
| "epoch": 3.8, |
| "grad_norm": 0.2362147718667984, |
| "learning_rate": 0.0002, |
| "loss": 0.0396, |
| "step": 95 |
| }, |
| { |
| "epoch": 4.0, |
| "grad_norm": 0.2028852105140686, |
| "learning_rate": 0.0002, |
| "loss": 0.0475, |
| "step": 100 |
| }, |
| { |
| "epoch": 4.2, |
| "grad_norm": 0.12950176000595093, |
| "learning_rate": 0.0002, |
| "loss": 0.0272, |
| "step": 105 |
| }, |
| { |
| "epoch": 4.4, |
| "grad_norm": 0.3273729383945465, |
| "learning_rate": 0.0002, |
| "loss": 0.0302, |
| "step": 110 |
| }, |
| { |
| "epoch": 4.6, |
| "grad_norm": 0.2938006818294525, |
| "learning_rate": 0.0002, |
| "loss": 0.0487, |
| "step": 115 |
| }, |
| { |
| "epoch": 4.8, |
| "grad_norm": 0.6171326041221619, |
| "learning_rate": 0.0002, |
| "loss": 0.1138, |
| "step": 120 |
| }, |
| { |
| "epoch": 5.0, |
| "grad_norm": 0.2811235785484314, |
| "learning_rate": 0.0002, |
| "loss": 0.0313, |
| "step": 125 |
| }, |
| { |
| "epoch": 5.2, |
| "grad_norm": 0.5635420680046082, |
| "learning_rate": 0.0002, |
| "loss": 0.0242, |
| "step": 130 |
| }, |
| { |
| "epoch": 5.4, |
| "grad_norm": 0.5491178035736084, |
| "learning_rate": 0.0002, |
| "loss": 0.0486, |
| "step": 135 |
| }, |
| { |
| "epoch": 5.6, |
| "grad_norm": 0.1552320271730423, |
| "learning_rate": 0.0002, |
| "loss": 0.0332, |
| "step": 140 |
| }, |
| { |
| "epoch": 5.8, |
| "grad_norm": 0.4321906864643097, |
| "learning_rate": 0.0002, |
| "loss": 0.0241, |
| "step": 145 |
| }, |
| { |
| "epoch": 6.0, |
| "grad_norm": 0.3396056592464447, |
| "learning_rate": 0.0002, |
| "loss": 0.034, |
| "step": 150 |
| }, |
| { |
| "epoch": 6.2, |
| "grad_norm": 0.09246141463518143, |
| "learning_rate": 0.0002, |
| "loss": 0.02, |
| "step": 155 |
| }, |
| { |
| "epoch": 6.4, |
| "grad_norm": 0.1537373661994934, |
| "learning_rate": 0.0002, |
| "loss": 0.0258, |
| "step": 160 |
| }, |
| { |
| "epoch": 6.6, |
| "grad_norm": 0.08097207546234131, |
| "learning_rate": 0.0002, |
| "loss": 0.032, |
| "step": 165 |
| }, |
| { |
| "epoch": 6.8, |
| "grad_norm": 0.10840635746717453, |
| "learning_rate": 0.0002, |
| "loss": 0.0343, |
| "step": 170 |
| }, |
| { |
| "epoch": 7.0, |
| "grad_norm": 0.07061374932527542, |
| "learning_rate": 0.0002, |
| "loss": 0.034, |
| "step": 175 |
| }, |
| { |
| "epoch": 7.2, |
| "grad_norm": 0.1112145483493805, |
| "learning_rate": 0.0002, |
| "loss": 0.0196, |
| "step": 180 |
| }, |
| { |
| "epoch": 7.4, |
| "grad_norm": 0.06888523697853088, |
| "learning_rate": 0.0002, |
| "loss": 0.0238, |
| "step": 185 |
| }, |
| { |
| "epoch": 7.6, |
| "grad_norm": 0.1332859843969345, |
| "learning_rate": 0.0002, |
| "loss": 0.0349, |
| "step": 190 |
| }, |
| { |
| "epoch": 7.8, |
| "grad_norm": 0.07381593436002731, |
| "learning_rate": 0.0002, |
| "loss": 0.027, |
| "step": 195 |
| }, |
| { |
| "epoch": 8.0, |
| "grad_norm": 0.12291029840707779, |
| "learning_rate": 0.0002, |
| "loss": 0.0309, |
| "step": 200 |
| }, |
| { |
| "epoch": 8.2, |
| "grad_norm": 0.08110173791646957, |
| "learning_rate": 0.0002, |
| "loss": 0.0192, |
| "step": 205 |
| }, |
| { |
| "epoch": 8.4, |
| "grad_norm": 0.0632745772600174, |
| "learning_rate": 0.0002, |
| "loss": 0.0219, |
| "step": 210 |
| }, |
| { |
| "epoch": 8.6, |
| "grad_norm": 0.07048214226961136, |
| "learning_rate": 0.0002, |
| "loss": 0.0218, |
| "step": 215 |
| }, |
| { |
| "epoch": 8.8, |
| "grad_norm": 0.07824481278657913, |
| "learning_rate": 0.0002, |
| "loss": 0.0673, |
| "step": 220 |
| }, |
| { |
| "epoch": 9.0, |
| "grad_norm": 0.14880645275115967, |
| "learning_rate": 0.0002, |
| "loss": 0.0519, |
| "step": 225 |
| }, |
| { |
| "epoch": 9.2, |
| "grad_norm": 0.07025552541017532, |
| "learning_rate": 0.0002, |
| "loss": 0.0193, |
| "step": 230 |
| }, |
| { |
| "epoch": 9.4, |
| "grad_norm": 0.08039015531539917, |
| "learning_rate": 0.0002, |
| "loss": 0.0234, |
| "step": 235 |
| }, |
| { |
| "epoch": 9.6, |
| "grad_norm": 0.08016083389520645, |
| "learning_rate": 0.0002, |
| "loss": 0.0222, |
| "step": 240 |
| }, |
| { |
| "epoch": 9.8, |
| "grad_norm": 0.06006358191370964, |
| "learning_rate": 0.0002, |
| "loss": 0.0253, |
| "step": 245 |
| }, |
| { |
| "epoch": 10.0, |
| "grad_norm": 0.07456184923648834, |
| "learning_rate": 0.0002, |
| "loss": 0.0279, |
| "step": 250 |
| }, |
| { |
| "epoch": 10.2, |
| "grad_norm": 0.06278860569000244, |
| "learning_rate": 0.0002, |
| "loss": 0.0177, |
| "step": 255 |
| }, |
| { |
| "epoch": 10.4, |
| "grad_norm": 0.07741103321313858, |
| "learning_rate": 0.0002, |
| "loss": 0.0664, |
| "step": 260 |
| }, |
| { |
| "epoch": 10.6, |
| "grad_norm": 0.08807504922151566, |
| "learning_rate": 0.0002, |
| "loss": 0.0208, |
| "step": 265 |
| }, |
| { |
| "epoch": 10.8, |
| "grad_norm": 0.07207378000020981, |
| "learning_rate": 0.0002, |
| "loss": 0.0239, |
| "step": 270 |
| }, |
| { |
| "epoch": 11.0, |
| "grad_norm": 0.25116804242134094, |
| "learning_rate": 0.0002, |
| "loss": 0.0227, |
| "step": 275 |
| }, |
| { |
| "epoch": 11.2, |
| "grad_norm": 0.08251325786113739, |
| "learning_rate": 0.0002, |
| "loss": 0.0172, |
| "step": 280 |
| }, |
| { |
| "epoch": 11.4, |
| "grad_norm": 0.0648980513215065, |
| "learning_rate": 0.0002, |
| "loss": 0.0201, |
| "step": 285 |
| }, |
| { |
| "epoch": 11.6, |
| "grad_norm": 0.06867312639951706, |
| "learning_rate": 0.0002, |
| "loss": 0.0542, |
| "step": 290 |
| }, |
| { |
| "epoch": 11.8, |
| "grad_norm": 0.06138678640127182, |
| "learning_rate": 0.0002, |
| "loss": 0.0253, |
| "step": 295 |
| }, |
| { |
| "epoch": 12.0, |
| "grad_norm": 0.08109026402235031, |
| "learning_rate": 0.0002, |
| "loss": 0.0493, |
| "step": 300 |
| }, |
| { |
| "epoch": 12.2, |
| "grad_norm": 0.07344530522823334, |
| "learning_rate": 0.0002, |
| "loss": 0.0187, |
| "step": 305 |
| }, |
| { |
| "epoch": 12.4, |
| "grad_norm": 0.05648460239171982, |
| "learning_rate": 0.0002, |
| "loss": 0.0216, |
| "step": 310 |
| }, |
| { |
| "epoch": 12.6, |
| "grad_norm": 0.05600782483816147, |
| "learning_rate": 0.0002, |
| "loss": 0.0284, |
| "step": 315 |
| }, |
| { |
| "epoch": 12.8, |
| "grad_norm": 0.04593711718916893, |
| "learning_rate": 0.0002, |
| "loss": 0.0404, |
| "step": 320 |
| }, |
| { |
| "epoch": 13.0, |
| "grad_norm": 0.07701898366212845, |
| "learning_rate": 0.0002, |
| "loss": 0.021, |
| "step": 325 |
| }, |
| { |
| "epoch": 13.2, |
| "grad_norm": 0.0478169247508049, |
| "learning_rate": 0.0002, |
| "loss": 0.0171, |
| "step": 330 |
| }, |
| { |
| "epoch": 13.4, |
| "grad_norm": 0.07693301141262054, |
| "learning_rate": 0.0002, |
| "loss": 0.0182, |
| "step": 335 |
| }, |
| { |
| "epoch": 13.6, |
| "grad_norm": 0.07630215585231781, |
| "learning_rate": 0.0002, |
| "loss": 0.0413, |
| "step": 340 |
| }, |
| { |
| "epoch": 13.8, |
| "grad_norm": 0.07538779824972153, |
| "learning_rate": 0.0002, |
| "loss": 0.0267, |
| "step": 345 |
| }, |
| { |
| "epoch": 14.0, |
| "grad_norm": 0.08219777047634125, |
| "learning_rate": 0.0002, |
| "loss": 0.0379, |
| "step": 350 |
| }, |
| { |
| "epoch": 14.2, |
| "grad_norm": 0.05356384813785553, |
| "learning_rate": 0.0002, |
| "loss": 0.0175, |
| "step": 355 |
| }, |
| { |
| "epoch": 14.4, |
| "grad_norm": 0.08164051175117493, |
| "learning_rate": 0.0002, |
| "loss": 0.0214, |
| "step": 360 |
| }, |
| { |
| "epoch": 14.6, |
| "grad_norm": 0.04210818186402321, |
| "learning_rate": 0.0002, |
| "loss": 0.0224, |
| "step": 365 |
| }, |
| { |
| "epoch": 14.8, |
| "grad_norm": 0.05736362934112549, |
| "learning_rate": 0.0002, |
| "loss": 0.036, |
| "step": 370 |
| }, |
| { |
| "epoch": 15.0, |
| "grad_norm": 0.07131894677877426, |
| "learning_rate": 0.0002, |
| "loss": 0.0253, |
| "step": 375 |
| } |
| ], |
| "logging_steps": 5, |
| "max_steps": 375, |
| "num_input_tokens_seen": 0, |
| "num_train_epochs": 15, |
| "save_steps": 500, |
| "stateful_callbacks": { |
| "TrainerControl": { |
| "args": { |
| "should_epoch_stop": false, |
| "should_evaluate": false, |
| "should_log": false, |
| "should_save": true, |
| "should_training_stop": true |
| }, |
| "attributes": {} |
| } |
| }, |
| "total_flos": 1.204263563618304e+16, |
| "train_batch_size": 2, |
| "trial_name": null, |
| "trial_params": null |
| } |
|
|