| { |
| "best_global_step": null, |
| "best_metric": null, |
| "best_model_checkpoint": null, |
| "epoch": 0.9019165727170236, |
| "eval_steps": 50, |
| "global_step": 200, |
| "is_hyper_param_search": false, |
| "is_local_process_zero": true, |
| "is_world_process_zero": true, |
| "log_history": [ |
| { |
| "epoch": 0.04509582863585118, |
| "grad_norm": 6.924594345036894e-05, |
| "learning_rate": 3.9130434782608694e-07, |
| "loss": 1.6552, |
| "step": 10 |
| }, |
| { |
| "epoch": 0.09019165727170236, |
| "grad_norm": 5.8848174376180395e-05, |
| "learning_rate": 8.260869565217391e-07, |
| "loss": 1.7092, |
| "step": 20 |
| }, |
| { |
| "epoch": 0.13528748590755355, |
| "grad_norm": 6.855162064312026e-05, |
| "learning_rate": 9.97758641300553e-07, |
| "loss": 1.6896, |
| "step": 30 |
| }, |
| { |
| "epoch": 0.18038331454340473, |
| "grad_norm": 5.4469885071739554e-05, |
| "learning_rate": 9.841341526992535e-07, |
| "loss": 1.9092, |
| "step": 40 |
| }, |
| { |
| "epoch": 0.2254791431792559, |
| "grad_norm": 4.8164554755203426e-05, |
| "learning_rate": 9.584688140963944e-07, |
| "loss": 1.5902, |
| "step": 50 |
| }, |
| { |
| "epoch": 0.2254791431792559, |
| "eval_loss": 1.7812083959579468, |
| "eval_runtime": 48.2467, |
| "eval_samples_per_second": 3.876, |
| "eval_steps_per_second": 1.948, |
| "step": 50 |
| }, |
| { |
| "epoch": 0.2705749718151071, |
| "grad_norm": 4.57292226201389e-05, |
| "learning_rate": 9.214009454506752e-07, |
| "loss": 1.5572, |
| "step": 60 |
| }, |
| { |
| "epoch": 0.3156708004509583, |
| "grad_norm": 5.38345702807419e-05, |
| "learning_rate": 8.738524578558546e-07, |
| "loss": 1.6925, |
| "step": 70 |
| }, |
| { |
| "epoch": 0.36076662908680945, |
| "grad_norm": 4.732041998067871e-05, |
| "learning_rate": 8.170059247861193e-07, |
| "loss": 1.6041, |
| "step": 80 |
| }, |
| { |
| "epoch": 0.40586245772266066, |
| "grad_norm": 3.962358459830284e-05, |
| "learning_rate": 7.522751704345887e-07, |
| "loss": 1.4225, |
| "step": 90 |
| }, |
| { |
| "epoch": 0.4509582863585118, |
| "grad_norm": 4.498194175539538e-05, |
| "learning_rate": 6.812701066393123e-07, |
| "loss": 1.5621, |
| "step": 100 |
| }, |
| { |
| "epoch": 0.4509582863585118, |
| "eval_loss": 1.6484575271606445, |
| "eval_runtime": 47.9782, |
| "eval_samples_per_second": 3.898, |
| "eval_steps_per_second": 1.959, |
| "step": 100 |
| }, |
| { |
| "epoch": 0.496054114994363, |
| "grad_norm": 3.846616891678423e-05, |
| "learning_rate": 6.057566929339095e-07, |
| "loss": 1.4544, |
| "step": 110 |
| }, |
| { |
| "epoch": 0.5411499436302142, |
| "grad_norm": 4.5146083721192554e-05, |
| "learning_rate": 5.27613015552254e-07, |
| "loss": 1.5298, |
| "step": 120 |
| }, |
| { |
| "epoch": 0.5862457722660653, |
| "grad_norm": 1.846837403718382e-05, |
| "learning_rate": 4.4878257774169345e-07, |
| "loss": 1.4496, |
| "step": 130 |
| }, |
| { |
| "epoch": 0.6313416009019166, |
| "grad_norm": 3.5699627915164456e-05, |
| "learning_rate": 3.7122596309655174e-07, |
| "loss": 1.4476, |
| "step": 140 |
| }, |
| { |
| "epoch": 0.6764374295377678, |
| "grad_norm": 2.6428568162373267e-05, |
| "learning_rate": 2.9687207408810555e-07, |
| "loss": 1.4961, |
| "step": 150 |
| }, |
| { |
| "epoch": 0.6764374295377678, |
| "eval_loss": 1.5978540182113647, |
| "eval_runtime": 47.947, |
| "eval_samples_per_second": 3.9, |
| "eval_steps_per_second": 1.96, |
| "step": 150 |
| }, |
| { |
| "epoch": 0.7215332581736189, |
| "grad_norm": 3.091954567935318e-05, |
| "learning_rate": 2.275701585324649e-07, |
| "loss": 1.3739, |
| "step": 160 |
| }, |
| { |
| "epoch": 0.7666290868094702, |
| "grad_norm": 3.293905683676712e-05, |
| "learning_rate": 1.6504381714107252e-07, |
| "loss": 1.4553, |
| "step": 170 |
| }, |
| { |
| "epoch": 0.8117249154453213, |
| "grad_norm": 2.837027750501875e-05, |
| "learning_rate": 1.1084813602723514e-07, |
| "loss": 1.4977, |
| "step": 180 |
| }, |
| { |
| "epoch": 0.8568207440811725, |
| "grad_norm": 3.116809602943249e-05, |
| "learning_rate": 6.633101032164273e-08, |
| "loss": 1.5816, |
| "step": 190 |
| }, |
| { |
| "epoch": 0.9019165727170236, |
| "grad_norm": 2.8287462555454113e-05, |
| "learning_rate": 3.2599620813200835e-08, |
| "loss": 1.4839, |
| "step": 200 |
| }, |
| { |
| "epoch": 0.9019165727170236, |
| "eval_loss": 1.5859750509262085, |
| "eval_runtime": 47.9616, |
| "eval_samples_per_second": 3.899, |
| "eval_steps_per_second": 1.96, |
| "step": 200 |
| } |
| ], |
| "logging_steps": 10, |
| "max_steps": 222, |
| "num_input_tokens_seen": 0, |
| "num_train_epochs": 1, |
| "save_steps": 100, |
| "stateful_callbacks": { |
| "TrainerControl": { |
| "args": { |
| "should_epoch_stop": false, |
| "should_evaluate": false, |
| "should_log": false, |
| "should_save": true, |
| "should_training_stop": false |
| }, |
| "attributes": {} |
| } |
| }, |
| "total_flos": 3.1184814522587136e+16, |
| "train_batch_size": 2, |
| "trial_name": null, |
| "trial_params": null |
| } |
|
|