| { | |
| "best_metric": 2.030155658721924, | |
| "best_model_checkpoint": "/uoa/scratch/users/u04ob20/attrib/data/models/google-gemma-2b-it/checkpoint-114", | |
| "epoch": 10.0, | |
| "eval_steps": 57, | |
| "global_step": 1130, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.5, | |
| "grad_norm": 17.672748565673828, | |
| "learning_rate": 4.747787610619469e-05, | |
| "loss": 2.9683, | |
| "step": 57 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "eval_loss": 2.1323869228363037, | |
| "eval_runtime": 14.1951, | |
| "eval_samples_per_second": 15.639, | |
| "eval_steps_per_second": 0.986, | |
| "step": 57 | |
| }, | |
| { | |
| "epoch": 1.01, | |
| "grad_norm": 10.006913185119629, | |
| "learning_rate": 4.495575221238939e-05, | |
| "loss": 1.9812, | |
| "step": 114 | |
| }, | |
| { | |
| "epoch": 1.01, | |
| "eval_loss": 2.030155658721924, | |
| "eval_runtime": 14.0967, | |
| "eval_samples_per_second": 15.748, | |
| "eval_steps_per_second": 0.993, | |
| "step": 114 | |
| }, | |
| { | |
| "epoch": 1.51, | |
| "grad_norm": 7.1079607009887695, | |
| "learning_rate": 4.243362831858407e-05, | |
| "loss": 1.6206, | |
| "step": 171 | |
| }, | |
| { | |
| "epoch": 1.51, | |
| "eval_loss": 2.0554354190826416, | |
| "eval_runtime": 14.0805, | |
| "eval_samples_per_second": 15.766, | |
| "eval_steps_per_second": 0.994, | |
| "step": 171 | |
| }, | |
| { | |
| "epoch": 2.02, | |
| "grad_norm": 8.306120872497559, | |
| "learning_rate": 3.991150442477876e-05, | |
| "loss": 1.6558, | |
| "step": 228 | |
| }, | |
| { | |
| "epoch": 2.02, | |
| "eval_loss": 2.124750852584839, | |
| "eval_runtime": 14.181, | |
| "eval_samples_per_second": 15.655, | |
| "eval_steps_per_second": 0.987, | |
| "step": 228 | |
| }, | |
| { | |
| "epoch": 2.52, | |
| "grad_norm": 5.911388874053955, | |
| "learning_rate": 3.7389380530973455e-05, | |
| "loss": 1.0749, | |
| "step": 285 | |
| }, | |
| { | |
| "epoch": 2.52, | |
| "eval_loss": 2.2880985736846924, | |
| "eval_runtime": 14.046, | |
| "eval_samples_per_second": 15.805, | |
| "eval_steps_per_second": 0.997, | |
| "step": 285 | |
| }, | |
| { | |
| "epoch": 3.03, | |
| "grad_norm": 5.9703569412231445, | |
| "learning_rate": 3.4867256637168145e-05, | |
| "loss": 1.1099, | |
| "step": 342 | |
| }, | |
| { | |
| "epoch": 3.03, | |
| "eval_loss": 2.6607654094696045, | |
| "eval_runtime": 14.1568, | |
| "eval_samples_per_second": 15.681, | |
| "eval_steps_per_second": 0.989, | |
| "step": 342 | |
| }, | |
| { | |
| "epoch": 3.53, | |
| "grad_norm": 7.199086666107178, | |
| "learning_rate": 3.2345132743362834e-05, | |
| "loss": 0.5349, | |
| "step": 399 | |
| }, | |
| { | |
| "epoch": 3.53, | |
| "eval_loss": 2.9111106395721436, | |
| "eval_runtime": 14.1631, | |
| "eval_samples_per_second": 15.675, | |
| "eval_steps_per_second": 0.988, | |
| "step": 399 | |
| }, | |
| { | |
| "epoch": 4.04, | |
| "grad_norm": 3.1459338665008545, | |
| "learning_rate": 2.982300884955752e-05, | |
| "loss": 0.5432, | |
| "step": 456 | |
| }, | |
| { | |
| "epoch": 4.04, | |
| "eval_loss": 3.114436149597168, | |
| "eval_runtime": 14.1302, | |
| "eval_samples_per_second": 15.711, | |
| "eval_steps_per_second": 0.991, | |
| "step": 456 | |
| }, | |
| { | |
| "epoch": 4.54, | |
| "grad_norm": 3.5249204635620117, | |
| "learning_rate": 2.7300884955752216e-05, | |
| "loss": 0.2523, | |
| "step": 513 | |
| }, | |
| { | |
| "epoch": 4.54, | |
| "eval_loss": 3.34505033493042, | |
| "eval_runtime": 14.1883, | |
| "eval_samples_per_second": 15.647, | |
| "eval_steps_per_second": 0.987, | |
| "step": 513 | |
| }, | |
| { | |
| "epoch": 5.04, | |
| "grad_norm": 3.153855085372925, | |
| "learning_rate": 2.4778761061946905e-05, | |
| "loss": 0.2561, | |
| "step": 570 | |
| }, | |
| { | |
| "epoch": 5.04, | |
| "eval_loss": 3.5140204429626465, | |
| "eval_runtime": 14.128, | |
| "eval_samples_per_second": 15.714, | |
| "eval_steps_per_second": 0.991, | |
| "step": 570 | |
| }, | |
| { | |
| "epoch": 5.55, | |
| "grad_norm": 3.072230339050293, | |
| "learning_rate": 2.2256637168141594e-05, | |
| "loss": 0.1508, | |
| "step": 627 | |
| }, | |
| { | |
| "epoch": 5.55, | |
| "eval_loss": 3.5723717212677, | |
| "eval_runtime": 14.0396, | |
| "eval_samples_per_second": 15.812, | |
| "eval_steps_per_second": 0.997, | |
| "step": 627 | |
| }, | |
| { | |
| "epoch": 6.05, | |
| "grad_norm": 1.96257746219635, | |
| "learning_rate": 1.9734513274336283e-05, | |
| "loss": 0.1365, | |
| "step": 684 | |
| }, | |
| { | |
| "epoch": 6.05, | |
| "eval_loss": 3.7443270683288574, | |
| "eval_runtime": 14.1133, | |
| "eval_samples_per_second": 15.73, | |
| "eval_steps_per_second": 0.992, | |
| "step": 684 | |
| }, | |
| { | |
| "epoch": 6.56, | |
| "grad_norm": 2.537320375442505, | |
| "learning_rate": 1.7212389380530976e-05, | |
| "loss": 0.0878, | |
| "step": 741 | |
| }, | |
| { | |
| "epoch": 6.56, | |
| "eval_loss": 3.926490545272827, | |
| "eval_runtime": 14.1716, | |
| "eval_samples_per_second": 15.665, | |
| "eval_steps_per_second": 0.988, | |
| "step": 741 | |
| }, | |
| { | |
| "epoch": 7.06, | |
| "grad_norm": 1.6090797185897827, | |
| "learning_rate": 1.4690265486725665e-05, | |
| "loss": 0.0841, | |
| "step": 798 | |
| }, | |
| { | |
| "epoch": 7.06, | |
| "eval_loss": 3.97700572013855, | |
| "eval_runtime": 14.1937, | |
| "eval_samples_per_second": 15.641, | |
| "eval_steps_per_second": 0.986, | |
| "step": 798 | |
| }, | |
| { | |
| "epoch": 7.57, | |
| "grad_norm": 1.7380380630493164, | |
| "learning_rate": 1.2168141592920354e-05, | |
| "loss": 0.0587, | |
| "step": 855 | |
| }, | |
| { | |
| "epoch": 7.57, | |
| "eval_loss": 4.071342468261719, | |
| "eval_runtime": 14.1098, | |
| "eval_samples_per_second": 15.734, | |
| "eval_steps_per_second": 0.992, | |
| "step": 855 | |
| }, | |
| { | |
| "epoch": 8.07, | |
| "grad_norm": 0.9195989370346069, | |
| "learning_rate": 9.646017699115045e-06, | |
| "loss": 0.0539, | |
| "step": 912 | |
| }, | |
| { | |
| "epoch": 8.07, | |
| "eval_loss": 4.22251558303833, | |
| "eval_runtime": 14.0932, | |
| "eval_samples_per_second": 15.752, | |
| "eval_steps_per_second": 0.993, | |
| "step": 912 | |
| }, | |
| { | |
| "epoch": 8.58, | |
| "grad_norm": 1.6740847826004028, | |
| "learning_rate": 7.123893805309735e-06, | |
| "loss": 0.0404, | |
| "step": 969 | |
| }, | |
| { | |
| "epoch": 8.58, | |
| "eval_loss": 4.388303279876709, | |
| "eval_runtime": 14.0715, | |
| "eval_samples_per_second": 15.777, | |
| "eval_steps_per_second": 0.995, | |
| "step": 969 | |
| }, | |
| { | |
| "epoch": 9.08, | |
| "grad_norm": 0.768718421459198, | |
| "learning_rate": 4.601769911504425e-06, | |
| "loss": 0.0383, | |
| "step": 1026 | |
| }, | |
| { | |
| "epoch": 9.08, | |
| "eval_loss": 4.46160364151001, | |
| "eval_runtime": 14.1199, | |
| "eval_samples_per_second": 15.722, | |
| "eval_steps_per_second": 0.992, | |
| "step": 1026 | |
| }, | |
| { | |
| "epoch": 9.58, | |
| "grad_norm": 0.8811420202255249, | |
| "learning_rate": 2.079646017699115e-06, | |
| "loss": 0.0292, | |
| "step": 1083 | |
| }, | |
| { | |
| "epoch": 9.58, | |
| "eval_loss": 4.561453819274902, | |
| "eval_runtime": 14.1759, | |
| "eval_samples_per_second": 15.66, | |
| "eval_steps_per_second": 0.988, | |
| "step": 1083 | |
| }, | |
| { | |
| "epoch": 10.0, | |
| "step": 1130, | |
| "total_flos": 2.40489782575104e+16, | |
| "train_loss": 0.6407461469152332, | |
| "train_runtime": 3370.5966, | |
| "train_samples_per_second": 4.688, | |
| "train_steps_per_second": 0.335 | |
| } | |
| ], | |
| "logging_steps": 57, | |
| "max_steps": 1130, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 10, | |
| "save_steps": 57, | |
| "total_flos": 2.40489782575104e+16, | |
| "train_batch_size": 14, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |