| { | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 3.0, | |
| "eval_steps": 500, | |
| "global_step": 39288, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.03817959682345754, | |
| "grad_norm": 47.025718688964844, | |
| "learning_rate": 4.936367338627571e-05, | |
| "loss": 19.8956, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.07635919364691508, | |
| "grad_norm": 39.02047348022461, | |
| "learning_rate": 4.872734677255142e-05, | |
| "loss": 2.0594, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 0.11453879047037263, | |
| "grad_norm": 82.10893249511719, | |
| "learning_rate": 4.8091020158827126e-05, | |
| "loss": 1.924, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 0.15271838729383017, | |
| "grad_norm": 25.970558166503906, | |
| "learning_rate": 4.7454693545102835e-05, | |
| "loss": 1.6354, | |
| "step": 2000 | |
| }, | |
| { | |
| "epoch": 0.19089798411728773, | |
| "grad_norm": 46.208919525146484, | |
| "learning_rate": 4.6818366931378536e-05, | |
| "loss": 1.5016, | |
| "step": 2500 | |
| }, | |
| { | |
| "epoch": 0.22907758094074526, | |
| "grad_norm": 35.68703842163086, | |
| "learning_rate": 4.6182040317654244e-05, | |
| "loss": 1.5022, | |
| "step": 3000 | |
| }, | |
| { | |
| "epoch": 0.2672571777642028, | |
| "grad_norm": 65.33502960205078, | |
| "learning_rate": 4.554571370392995e-05, | |
| "loss": 1.423, | |
| "step": 3500 | |
| }, | |
| { | |
| "epoch": 0.30543677458766033, | |
| "grad_norm": 33.0444450378418, | |
| "learning_rate": 4.490938709020566e-05, | |
| "loss": 1.3971, | |
| "step": 4000 | |
| }, | |
| { | |
| "epoch": 0.3436163714111179, | |
| "grad_norm": 18.407243728637695, | |
| "learning_rate": 4.427306047648137e-05, | |
| "loss": 1.308, | |
| "step": 4500 | |
| }, | |
| { | |
| "epoch": 0.38179596823457546, | |
| "grad_norm": 62.20368194580078, | |
| "learning_rate": 4.3636733862757076e-05, | |
| "loss": 1.3365, | |
| "step": 5000 | |
| }, | |
| { | |
| "epoch": 0.41997556505803296, | |
| "grad_norm": 18.440366744995117, | |
| "learning_rate": 4.3000407249032784e-05, | |
| "loss": 1.1873, | |
| "step": 5500 | |
| }, | |
| { | |
| "epoch": 0.4581551618814905, | |
| "grad_norm": 21.170751571655273, | |
| "learning_rate": 4.236408063530849e-05, | |
| "loss": 1.212, | |
| "step": 6000 | |
| }, | |
| { | |
| "epoch": 0.4963347587049481, | |
| "grad_norm": 38.78303146362305, | |
| "learning_rate": 4.17277540215842e-05, | |
| "loss": 1.2951, | |
| "step": 6500 | |
| }, | |
| { | |
| "epoch": 0.5345143555284056, | |
| "grad_norm": 15.414962768554688, | |
| "learning_rate": 4.109142740785991e-05, | |
| "loss": 1.2291, | |
| "step": 7000 | |
| }, | |
| { | |
| "epoch": 0.5726939523518632, | |
| "grad_norm": 17.313640594482422, | |
| "learning_rate": 4.0455100794135616e-05, | |
| "loss": 1.1234, | |
| "step": 7500 | |
| }, | |
| { | |
| "epoch": 0.6108735491753207, | |
| "grad_norm": 24.39589500427246, | |
| "learning_rate": 3.9818774180411324e-05, | |
| "loss": 1.2496, | |
| "step": 8000 | |
| }, | |
| { | |
| "epoch": 0.6490531459987783, | |
| "grad_norm": 29.411834716796875, | |
| "learning_rate": 3.918244756668703e-05, | |
| "loss": 1.1647, | |
| "step": 8500 | |
| }, | |
| { | |
| "epoch": 0.6872327428222358, | |
| "grad_norm": 24.564817428588867, | |
| "learning_rate": 3.854612095296274e-05, | |
| "loss": 1.1065, | |
| "step": 9000 | |
| }, | |
| { | |
| "epoch": 0.7254123396456933, | |
| "grad_norm": 48.59101104736328, | |
| "learning_rate": 3.790979433923845e-05, | |
| "loss": 1.1279, | |
| "step": 9500 | |
| }, | |
| { | |
| "epoch": 0.7635919364691509, | |
| "grad_norm": 37.30824661254883, | |
| "learning_rate": 3.7273467725514156e-05, | |
| "loss": 1.0621, | |
| "step": 10000 | |
| }, | |
| { | |
| "epoch": 0.8017715332926084, | |
| "grad_norm": 40.20986557006836, | |
| "learning_rate": 3.6637141111789864e-05, | |
| "loss": 1.1039, | |
| "step": 10500 | |
| }, | |
| { | |
| "epoch": 0.8399511301160659, | |
| "grad_norm": 7.365723609924316, | |
| "learning_rate": 3.600081449806557e-05, | |
| "loss": 1.0561, | |
| "step": 11000 | |
| }, | |
| { | |
| "epoch": 0.8781307269395235, | |
| "grad_norm": 10.872908592224121, | |
| "learning_rate": 3.5364487884341273e-05, | |
| "loss": 1.0061, | |
| "step": 11500 | |
| }, | |
| { | |
| "epoch": 0.916310323762981, | |
| "grad_norm": 37.47850799560547, | |
| "learning_rate": 3.472816127061698e-05, | |
| "loss": 1.1034, | |
| "step": 12000 | |
| }, | |
| { | |
| "epoch": 0.9544899205864386, | |
| "grad_norm": 59.97223663330078, | |
| "learning_rate": 3.409183465689269e-05, | |
| "loss": 1.058, | |
| "step": 12500 | |
| }, | |
| { | |
| "epoch": 0.9926695174098962, | |
| "grad_norm": 17.347299575805664, | |
| "learning_rate": 3.34555080431684e-05, | |
| "loss": 0.9714, | |
| "step": 13000 | |
| }, | |
| { | |
| "epoch": 1.0308491142333538, | |
| "grad_norm": 47.64847183227539, | |
| "learning_rate": 3.2819181429444106e-05, | |
| "loss": 0.9232, | |
| "step": 13500 | |
| }, | |
| { | |
| "epoch": 1.0690287110568113, | |
| "grad_norm": 23.9332332611084, | |
| "learning_rate": 3.2182854815719814e-05, | |
| "loss": 0.8723, | |
| "step": 14000 | |
| }, | |
| { | |
| "epoch": 1.1072083078802688, | |
| "grad_norm": 24.243032455444336, | |
| "learning_rate": 3.154652820199552e-05, | |
| "loss": 0.9376, | |
| "step": 14500 | |
| }, | |
| { | |
| "epoch": 1.1453879047037263, | |
| "grad_norm": 18.394193649291992, | |
| "learning_rate": 3.091020158827123e-05, | |
| "loss": 0.9212, | |
| "step": 15000 | |
| }, | |
| { | |
| "epoch": 1.1835675015271838, | |
| "grad_norm": 33.30715560913086, | |
| "learning_rate": 3.027387497454694e-05, | |
| "loss": 0.9196, | |
| "step": 15500 | |
| }, | |
| { | |
| "epoch": 1.2217470983506413, | |
| "grad_norm": 116.27849578857422, | |
| "learning_rate": 2.9637548360822642e-05, | |
| "loss": 0.8325, | |
| "step": 16000 | |
| }, | |
| { | |
| "epoch": 1.2599266951740988, | |
| "grad_norm": 33.555179595947266, | |
| "learning_rate": 2.900122174709835e-05, | |
| "loss": 0.8746, | |
| "step": 16500 | |
| }, | |
| { | |
| "epoch": 1.2981062919975566, | |
| "grad_norm": 35.980411529541016, | |
| "learning_rate": 2.836489513337406e-05, | |
| "loss": 0.8561, | |
| "step": 17000 | |
| }, | |
| { | |
| "epoch": 1.336285888821014, | |
| "grad_norm": 42.04365921020508, | |
| "learning_rate": 2.7728568519649766e-05, | |
| "loss": 0.867, | |
| "step": 17500 | |
| }, | |
| { | |
| "epoch": 1.3744654856444716, | |
| "grad_norm": 43.86580276489258, | |
| "learning_rate": 2.7092241905925474e-05, | |
| "loss": 0.8331, | |
| "step": 18000 | |
| }, | |
| { | |
| "epoch": 1.412645082467929, | |
| "grad_norm": 44.562686920166016, | |
| "learning_rate": 2.6455915292201182e-05, | |
| "loss": 0.8354, | |
| "step": 18500 | |
| }, | |
| { | |
| "epoch": 1.4508246792913866, | |
| "grad_norm": 15.569896697998047, | |
| "learning_rate": 2.581958867847689e-05, | |
| "loss": 0.8514, | |
| "step": 19000 | |
| }, | |
| { | |
| "epoch": 1.4890042761148443, | |
| "grad_norm": 58.31094741821289, | |
| "learning_rate": 2.51832620647526e-05, | |
| "loss": 0.8712, | |
| "step": 19500 | |
| }, | |
| { | |
| "epoch": 1.5271838729383018, | |
| "grad_norm": 15.770406723022461, | |
| "learning_rate": 2.4546935451028307e-05, | |
| "loss": 0.8085, | |
| "step": 20000 | |
| }, | |
| { | |
| "epoch": 1.5653634697617593, | |
| "grad_norm": 29.918487548828125, | |
| "learning_rate": 2.3910608837304015e-05, | |
| "loss": 0.7545, | |
| "step": 20500 | |
| }, | |
| { | |
| "epoch": 1.6035430665852168, | |
| "grad_norm": 112.88292694091797, | |
| "learning_rate": 2.3274282223579723e-05, | |
| "loss": 0.8347, | |
| "step": 21000 | |
| }, | |
| { | |
| "epoch": 1.6417226634086743, | |
| "grad_norm": 16.153079986572266, | |
| "learning_rate": 2.2637955609855427e-05, | |
| "loss": 0.8042, | |
| "step": 21500 | |
| }, | |
| { | |
| "epoch": 1.679902260232132, | |
| "grad_norm": 17.334716796875, | |
| "learning_rate": 2.2001628996131135e-05, | |
| "loss": 0.8368, | |
| "step": 22000 | |
| }, | |
| { | |
| "epoch": 1.7180818570555894, | |
| "grad_norm": 22.649688720703125, | |
| "learning_rate": 2.1365302382406843e-05, | |
| "loss": 0.7438, | |
| "step": 22500 | |
| }, | |
| { | |
| "epoch": 1.756261453879047, | |
| "grad_norm": 23.706069946289062, | |
| "learning_rate": 2.072897576868255e-05, | |
| "loss": 0.8163, | |
| "step": 23000 | |
| }, | |
| { | |
| "epoch": 1.7944410507025046, | |
| "grad_norm": 60.12598419189453, | |
| "learning_rate": 2.0092649154958256e-05, | |
| "loss": 0.8018, | |
| "step": 23500 | |
| }, | |
| { | |
| "epoch": 1.832620647525962, | |
| "grad_norm": 88.19845581054688, | |
| "learning_rate": 1.9456322541233964e-05, | |
| "loss": 0.7776, | |
| "step": 24000 | |
| }, | |
| { | |
| "epoch": 1.8708002443494198, | |
| "grad_norm": 18.326126098632812, | |
| "learning_rate": 1.8819995927509672e-05, | |
| "loss": 0.7694, | |
| "step": 24500 | |
| }, | |
| { | |
| "epoch": 1.9089798411728771, | |
| "grad_norm": 33.15532684326172, | |
| "learning_rate": 1.818366931378538e-05, | |
| "loss": 0.7517, | |
| "step": 25000 | |
| }, | |
| { | |
| "epoch": 1.9471594379963348, | |
| "grad_norm": 11.88496208190918, | |
| "learning_rate": 1.7547342700061088e-05, | |
| "loss": 0.7802, | |
| "step": 25500 | |
| }, | |
| { | |
| "epoch": 1.9853390348197923, | |
| "grad_norm": 53.72198486328125, | |
| "learning_rate": 1.6911016086336796e-05, | |
| "loss": 0.7699, | |
| "step": 26000 | |
| }, | |
| { | |
| "epoch": 2.02351863164325, | |
| "grad_norm": 13.519152641296387, | |
| "learning_rate": 1.6274689472612504e-05, | |
| "loss": 0.7362, | |
| "step": 26500 | |
| }, | |
| { | |
| "epoch": 2.0616982284667076, | |
| "grad_norm": 21.352752685546875, | |
| "learning_rate": 1.5638362858888212e-05, | |
| "loss": 0.7043, | |
| "step": 27000 | |
| }, | |
| { | |
| "epoch": 2.099877825290165, | |
| "grad_norm": 39.630210876464844, | |
| "learning_rate": 1.500203624516392e-05, | |
| "loss": 0.6899, | |
| "step": 27500 | |
| }, | |
| { | |
| "epoch": 2.1380574221136226, | |
| "grad_norm": 15.605965614318848, | |
| "learning_rate": 1.4365709631439625e-05, | |
| "loss": 0.7241, | |
| "step": 28000 | |
| }, | |
| { | |
| "epoch": 2.17623701893708, | |
| "grad_norm": 12.23674201965332, | |
| "learning_rate": 1.3729383017715333e-05, | |
| "loss": 0.6201, | |
| "step": 28500 | |
| }, | |
| { | |
| "epoch": 2.2144166157605376, | |
| "grad_norm": 15.482099533081055, | |
| "learning_rate": 1.309305640399104e-05, | |
| "loss": 0.6966, | |
| "step": 29000 | |
| }, | |
| { | |
| "epoch": 2.2525962125839953, | |
| "grad_norm": 25.2972469329834, | |
| "learning_rate": 1.2456729790266749e-05, | |
| "loss": 0.5878, | |
| "step": 29500 | |
| }, | |
| { | |
| "epoch": 2.2907758094074526, | |
| "grad_norm": 18.335323333740234, | |
| "learning_rate": 1.1820403176542457e-05, | |
| "loss": 0.6468, | |
| "step": 30000 | |
| }, | |
| { | |
| "epoch": 2.3289554062309104, | |
| "grad_norm": 44.75697326660156, | |
| "learning_rate": 1.1184076562818163e-05, | |
| "loss": 0.6888, | |
| "step": 30500 | |
| }, | |
| { | |
| "epoch": 2.3671350030543676, | |
| "grad_norm": 9.713996887207031, | |
| "learning_rate": 1.0547749949093871e-05, | |
| "loss": 0.6584, | |
| "step": 31000 | |
| }, | |
| { | |
| "epoch": 2.4053145998778254, | |
| "grad_norm": 19.04861831665039, | |
| "learning_rate": 9.91142333536958e-06, | |
| "loss": 0.6356, | |
| "step": 31500 | |
| }, | |
| { | |
| "epoch": 2.4434941967012827, | |
| "grad_norm": 16.914947509765625, | |
| "learning_rate": 9.275096721645286e-06, | |
| "loss": 0.6188, | |
| "step": 32000 | |
| }, | |
| { | |
| "epoch": 2.4816737935247404, | |
| "grad_norm": 25.262630462646484, | |
| "learning_rate": 8.638770107920995e-06, | |
| "loss": 0.6062, | |
| "step": 32500 | |
| }, | |
| { | |
| "epoch": 2.5198533903481977, | |
| "grad_norm": 49.74958419799805, | |
| "learning_rate": 8.002443494196702e-06, | |
| "loss": 0.6554, | |
| "step": 33000 | |
| }, | |
| { | |
| "epoch": 2.5580329871716554, | |
| "grad_norm": 35.74587631225586, | |
| "learning_rate": 7.36611688047241e-06, | |
| "loss": 0.6523, | |
| "step": 33500 | |
| }, | |
| { | |
| "epoch": 2.596212583995113, | |
| "grad_norm": 22.55206298828125, | |
| "learning_rate": 6.729790266748116e-06, | |
| "loss": 0.6333, | |
| "step": 34000 | |
| }, | |
| { | |
| "epoch": 2.6343921808185704, | |
| "grad_norm": 19.39708709716797, | |
| "learning_rate": 6.093463653023825e-06, | |
| "loss": 0.6269, | |
| "step": 34500 | |
| }, | |
| { | |
| "epoch": 2.672571777642028, | |
| "grad_norm": 65.50653839111328, | |
| "learning_rate": 5.457137039299532e-06, | |
| "loss": 0.6825, | |
| "step": 35000 | |
| }, | |
| { | |
| "epoch": 2.7107513744654854, | |
| "grad_norm": 10.603687286376953, | |
| "learning_rate": 4.820810425575239e-06, | |
| "loss": 0.6289, | |
| "step": 35500 | |
| }, | |
| { | |
| "epoch": 2.748930971288943, | |
| "grad_norm": 12.743573188781738, | |
| "learning_rate": 4.184483811850947e-06, | |
| "loss": 0.6369, | |
| "step": 36000 | |
| }, | |
| { | |
| "epoch": 2.787110568112401, | |
| "grad_norm": 8.957477569580078, | |
| "learning_rate": 3.5481571981266544e-06, | |
| "loss": 0.6095, | |
| "step": 36500 | |
| }, | |
| { | |
| "epoch": 2.825290164935858, | |
| "grad_norm": 38.0220947265625, | |
| "learning_rate": 2.911830584402362e-06, | |
| "loss": 0.5497, | |
| "step": 37000 | |
| }, | |
| { | |
| "epoch": 2.863469761759316, | |
| "grad_norm": 25.39042854309082, | |
| "learning_rate": 2.27550397067807e-06, | |
| "loss": 0.5781, | |
| "step": 37500 | |
| }, | |
| { | |
| "epoch": 2.901649358582773, | |
| "grad_norm": 11.314273834228516, | |
| "learning_rate": 1.6391773569537775e-06, | |
| "loss": 0.6227, | |
| "step": 38000 | |
| }, | |
| { | |
| "epoch": 2.939828955406231, | |
| "grad_norm": 38.86436080932617, | |
| "learning_rate": 1.0028507432294848e-06, | |
| "loss": 0.5966, | |
| "step": 38500 | |
| }, | |
| { | |
| "epoch": 2.9780085522296886, | |
| "grad_norm": 11.880073547363281, | |
| "learning_rate": 3.6652412950519244e-07, | |
| "loss": 0.5901, | |
| "step": 39000 | |
| }, | |
| { | |
| "epoch": 3.0, | |
| "step": 39288, | |
| "total_flos": 5175762253906176.0, | |
| "train_loss": 1.154250169620813, | |
| "train_runtime": 1582.2568, | |
| "train_samples_per_second": 198.63, | |
| "train_steps_per_second": 24.83 | |
| } | |
| ], | |
| "logging_steps": 500, | |
| "max_steps": 39288, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 3, | |
| "save_steps": 500, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": true | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 5175762253906176.0, | |
| "train_batch_size": 8, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |