| { | |
| "best_metric": 1.8410229682922363, | |
| "best_model_checkpoint": "/content/drive/MyDrive/ZephyrAI/zephyr_model_v11_turbo/checkpoint-75", | |
| "epoch": 4.918032786885246, | |
| "eval_steps": 1, | |
| "global_step": 75, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.06557377049180328, | |
| "grad_norm": 1.070236325263977, | |
| "learning_rate": 6.666666666666667e-06, | |
| "loss": 2.4899, | |
| "step": 1 | |
| }, | |
| { | |
| "epoch": 0.06557377049180328, | |
| "eval_loss": 2.437901258468628, | |
| "eval_runtime": 46.6808, | |
| "eval_samples_per_second": 0.557, | |
| "eval_steps_per_second": 0.278, | |
| "step": 1 | |
| }, | |
| { | |
| "epoch": 0.13114754098360656, | |
| "grad_norm": 1.1805078983306885, | |
| "learning_rate": 1.3333333333333333e-05, | |
| "loss": 2.4403, | |
| "step": 2 | |
| }, | |
| { | |
| "epoch": 0.13114754098360656, | |
| "eval_loss": 2.434871196746826, | |
| "eval_runtime": 46.7714, | |
| "eval_samples_per_second": 0.556, | |
| "eval_steps_per_second": 0.278, | |
| "step": 2 | |
| }, | |
| { | |
| "epoch": 0.19672131147540983, | |
| "grad_norm": 1.0782222747802734, | |
| "learning_rate": 2e-05, | |
| "loss": 2.3851, | |
| "step": 3 | |
| }, | |
| { | |
| "epoch": 0.19672131147540983, | |
| "eval_loss": 2.421125650405884, | |
| "eval_runtime": 46.6915, | |
| "eval_samples_per_second": 0.557, | |
| "eval_steps_per_second": 0.278, | |
| "step": 3 | |
| }, | |
| { | |
| "epoch": 0.26229508196721313, | |
| "grad_norm": 1.0158095359802246, | |
| "learning_rate": 1.9722222222222224e-05, | |
| "loss": 2.4596, | |
| "step": 4 | |
| }, | |
| { | |
| "epoch": 0.26229508196721313, | |
| "eval_loss": 2.3971059322357178, | |
| "eval_runtime": 46.6786, | |
| "eval_samples_per_second": 0.557, | |
| "eval_steps_per_second": 0.279, | |
| "step": 4 | |
| }, | |
| { | |
| "epoch": 0.32786885245901637, | |
| "grad_norm": 0.9539541006088257, | |
| "learning_rate": 1.9444444444444445e-05, | |
| "loss": 2.382, | |
| "step": 5 | |
| }, | |
| { | |
| "epoch": 0.32786885245901637, | |
| "eval_loss": 2.3755452632904053, | |
| "eval_runtime": 46.6915, | |
| "eval_samples_per_second": 0.557, | |
| "eval_steps_per_second": 0.278, | |
| "step": 5 | |
| }, | |
| { | |
| "epoch": 0.39344262295081966, | |
| "grad_norm": 0.8312937021255493, | |
| "learning_rate": 1.916666666666667e-05, | |
| "loss": 2.3304, | |
| "step": 6 | |
| }, | |
| { | |
| "epoch": 0.39344262295081966, | |
| "eval_loss": 2.3568789958953857, | |
| "eval_runtime": 46.7831, | |
| "eval_samples_per_second": 0.556, | |
| "eval_steps_per_second": 0.278, | |
| "step": 6 | |
| }, | |
| { | |
| "epoch": 0.45901639344262296, | |
| "grad_norm": 0.7678759694099426, | |
| "learning_rate": 1.888888888888889e-05, | |
| "loss": 2.3028, | |
| "step": 7 | |
| }, | |
| { | |
| "epoch": 0.45901639344262296, | |
| "eval_loss": 2.340029716491699, | |
| "eval_runtime": 46.7542, | |
| "eval_samples_per_second": 0.556, | |
| "eval_steps_per_second": 0.278, | |
| "step": 7 | |
| }, | |
| { | |
| "epoch": 0.5245901639344263, | |
| "grad_norm": 0.6792626976966858, | |
| "learning_rate": 1.8611111111111114e-05, | |
| "loss": 2.3466, | |
| "step": 8 | |
| }, | |
| { | |
| "epoch": 0.5245901639344263, | |
| "eval_loss": 2.3247592449188232, | |
| "eval_runtime": 46.7003, | |
| "eval_samples_per_second": 0.557, | |
| "eval_steps_per_second": 0.278, | |
| "step": 8 | |
| }, | |
| { | |
| "epoch": 0.5901639344262295, | |
| "grad_norm": 0.660761833190918, | |
| "learning_rate": 1.8333333333333333e-05, | |
| "loss": 2.3271, | |
| "step": 9 | |
| }, | |
| { | |
| "epoch": 0.5901639344262295, | |
| "eval_loss": 2.3108019828796387, | |
| "eval_runtime": 46.7515, | |
| "eval_samples_per_second": 0.556, | |
| "eval_steps_per_second": 0.278, | |
| "step": 9 | |
| }, | |
| { | |
| "epoch": 0.6557377049180327, | |
| "grad_norm": 0.6685893535614014, | |
| "learning_rate": 1.8055555555555558e-05, | |
| "loss": 2.2604, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.6557377049180327, | |
| "eval_loss": 2.2978525161743164, | |
| "eval_runtime": 46.6847, | |
| "eval_samples_per_second": 0.557, | |
| "eval_steps_per_second": 0.278, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.7213114754098361, | |
| "grad_norm": 0.5920826196670532, | |
| "learning_rate": 1.7777777777777777e-05, | |
| "loss": 2.2218, | |
| "step": 11 | |
| }, | |
| { | |
| "epoch": 0.7213114754098361, | |
| "eval_loss": 2.2857038974761963, | |
| "eval_runtime": 46.7809, | |
| "eval_samples_per_second": 0.556, | |
| "eval_steps_per_second": 0.278, | |
| "step": 11 | |
| }, | |
| { | |
| "epoch": 0.7868852459016393, | |
| "grad_norm": 0.6445866823196411, | |
| "learning_rate": 1.7500000000000002e-05, | |
| "loss": 2.2014, | |
| "step": 12 | |
| }, | |
| { | |
| "epoch": 0.7868852459016393, | |
| "eval_loss": 2.2739903926849365, | |
| "eval_runtime": 46.6906, | |
| "eval_samples_per_second": 0.557, | |
| "eval_steps_per_second": 0.278, | |
| "step": 12 | |
| }, | |
| { | |
| "epoch": 0.8524590163934426, | |
| "grad_norm": 0.6221867203712463, | |
| "learning_rate": 1.7222222222222224e-05, | |
| "loss": 2.2426, | |
| "step": 13 | |
| }, | |
| { | |
| "epoch": 0.8524590163934426, | |
| "eval_loss": 2.2625272274017334, | |
| "eval_runtime": 46.7451, | |
| "eval_samples_per_second": 0.556, | |
| "eval_steps_per_second": 0.278, | |
| "step": 13 | |
| }, | |
| { | |
| "epoch": 0.9180327868852459, | |
| "grad_norm": 0.6631550192832947, | |
| "learning_rate": 1.6944444444444446e-05, | |
| "loss": 2.2368, | |
| "step": 14 | |
| }, | |
| { | |
| "epoch": 0.9180327868852459, | |
| "eval_loss": 2.2512171268463135, | |
| "eval_runtime": 46.674, | |
| "eval_samples_per_second": 0.557, | |
| "eval_steps_per_second": 0.279, | |
| "step": 14 | |
| }, | |
| { | |
| "epoch": 0.9836065573770492, | |
| "grad_norm": 0.5557384490966797, | |
| "learning_rate": 1.6666666666666667e-05, | |
| "loss": 2.1956, | |
| "step": 15 | |
| }, | |
| { | |
| "epoch": 0.9836065573770492, | |
| "eval_loss": 2.23983097076416, | |
| "eval_runtime": 46.7325, | |
| "eval_samples_per_second": 0.556, | |
| "eval_steps_per_second": 0.278, | |
| "step": 15 | |
| }, | |
| { | |
| "epoch": 1.0491803278688525, | |
| "grad_norm": 0.5928459167480469, | |
| "learning_rate": 1.638888888888889e-05, | |
| "loss": 2.1429, | |
| "step": 16 | |
| }, | |
| { | |
| "epoch": 1.0491803278688525, | |
| "eval_loss": 2.2281932830810547, | |
| "eval_runtime": 46.768, | |
| "eval_samples_per_second": 0.556, | |
| "eval_steps_per_second": 0.278, | |
| "step": 16 | |
| }, | |
| { | |
| "epoch": 1.1147540983606556, | |
| "grad_norm": 0.6136890053749084, | |
| "learning_rate": 1.6111111111111115e-05, | |
| "loss": 2.1862, | |
| "step": 17 | |
| }, | |
| { | |
| "epoch": 1.1147540983606556, | |
| "eval_loss": 2.2162652015686035, | |
| "eval_runtime": 46.7237, | |
| "eval_samples_per_second": 0.556, | |
| "eval_steps_per_second": 0.278, | |
| "step": 17 | |
| }, | |
| { | |
| "epoch": 1.180327868852459, | |
| "grad_norm": 0.5814841389656067, | |
| "learning_rate": 1.5833333333333333e-05, | |
| "loss": 2.2012, | |
| "step": 18 | |
| }, | |
| { | |
| "epoch": 1.180327868852459, | |
| "eval_loss": 2.204092264175415, | |
| "eval_runtime": 46.7753, | |
| "eval_samples_per_second": 0.556, | |
| "eval_steps_per_second": 0.278, | |
| "step": 18 | |
| }, | |
| { | |
| "epoch": 1.2459016393442623, | |
| "grad_norm": 0.5701133608818054, | |
| "learning_rate": 1.555555555555556e-05, | |
| "loss": 2.1545, | |
| "step": 19 | |
| }, | |
| { | |
| "epoch": 1.2459016393442623, | |
| "eval_loss": 2.191829204559326, | |
| "eval_runtime": 46.7098, | |
| "eval_samples_per_second": 0.557, | |
| "eval_steps_per_second": 0.278, | |
| "step": 19 | |
| }, | |
| { | |
| "epoch": 1.3114754098360657, | |
| "grad_norm": 0.5770221948623657, | |
| "learning_rate": 1.5277777777777777e-05, | |
| "loss": 2.1463, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 1.3114754098360657, | |
| "eval_loss": 2.1793527603149414, | |
| "eval_runtime": 46.7538, | |
| "eval_samples_per_second": 0.556, | |
| "eval_steps_per_second": 0.278, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 1.3770491803278688, | |
| "grad_norm": 0.541409969329834, | |
| "learning_rate": 1.5000000000000002e-05, | |
| "loss": 2.062, | |
| "step": 21 | |
| }, | |
| { | |
| "epoch": 1.3770491803278688, | |
| "eval_loss": 2.1669070720672607, | |
| "eval_runtime": 46.6875, | |
| "eval_samples_per_second": 0.557, | |
| "eval_steps_per_second": 0.278, | |
| "step": 21 | |
| }, | |
| { | |
| "epoch": 1.4426229508196722, | |
| "grad_norm": 0.6078354120254517, | |
| "learning_rate": 1.4722222222222224e-05, | |
| "loss": 2.1365, | |
| "step": 22 | |
| }, | |
| { | |
| "epoch": 1.4426229508196722, | |
| "eval_loss": 2.154656410217285, | |
| "eval_runtime": 46.7354, | |
| "eval_samples_per_second": 0.556, | |
| "eval_steps_per_second": 0.278, | |
| "step": 22 | |
| }, | |
| { | |
| "epoch": 1.5081967213114753, | |
| "grad_norm": 0.6025573015213013, | |
| "learning_rate": 1.4444444444444446e-05, | |
| "loss": 2.0426, | |
| "step": 23 | |
| }, | |
| { | |
| "epoch": 1.5081967213114753, | |
| "eval_loss": 2.1424319744110107, | |
| "eval_runtime": 46.7734, | |
| "eval_samples_per_second": 0.556, | |
| "eval_steps_per_second": 0.278, | |
| "step": 23 | |
| }, | |
| { | |
| "epoch": 1.5737704918032787, | |
| "grad_norm": 0.5577809810638428, | |
| "learning_rate": 1.416666666666667e-05, | |
| "loss": 2.0489, | |
| "step": 24 | |
| }, | |
| { | |
| "epoch": 1.5737704918032787, | |
| "eval_loss": 2.1300365924835205, | |
| "eval_runtime": 46.8011, | |
| "eval_samples_per_second": 0.556, | |
| "eval_steps_per_second": 0.278, | |
| "step": 24 | |
| }, | |
| { | |
| "epoch": 1.639344262295082, | |
| "grad_norm": 0.5722355842590332, | |
| "learning_rate": 1.388888888888889e-05, | |
| "loss": 2.0289, | |
| "step": 25 | |
| }, | |
| { | |
| "epoch": 1.639344262295082, | |
| "eval_loss": 2.1180951595306396, | |
| "eval_runtime": 46.6691, | |
| "eval_samples_per_second": 0.557, | |
| "eval_steps_per_second": 0.279, | |
| "step": 25 | |
| }, | |
| { | |
| "epoch": 1.7049180327868854, | |
| "grad_norm": 0.5458038449287415, | |
| "learning_rate": 1.3611111111111113e-05, | |
| "loss": 2.0995, | |
| "step": 26 | |
| }, | |
| { | |
| "epoch": 1.7049180327868854, | |
| "eval_loss": 2.1060945987701416, | |
| "eval_runtime": 46.7807, | |
| "eval_samples_per_second": 0.556, | |
| "eval_steps_per_second": 0.278, | |
| "step": 26 | |
| }, | |
| { | |
| "epoch": 1.7704918032786885, | |
| "grad_norm": 0.5661202073097229, | |
| "learning_rate": 1.3333333333333333e-05, | |
| "loss": 2.0379, | |
| "step": 27 | |
| }, | |
| { | |
| "epoch": 1.7704918032786885, | |
| "eval_loss": 2.0944411754608154, | |
| "eval_runtime": 46.7187, | |
| "eval_samples_per_second": 0.557, | |
| "eval_steps_per_second": 0.278, | |
| "step": 27 | |
| }, | |
| { | |
| "epoch": 1.8360655737704918, | |
| "grad_norm": 0.5147558450698853, | |
| "learning_rate": 1.3055555555555557e-05, | |
| "loss": 2.0116, | |
| "step": 28 | |
| }, | |
| { | |
| "epoch": 1.8360655737704918, | |
| "eval_loss": 2.082529306411743, | |
| "eval_runtime": 46.7106, | |
| "eval_samples_per_second": 0.557, | |
| "eval_steps_per_second": 0.278, | |
| "step": 28 | |
| }, | |
| { | |
| "epoch": 1.901639344262295, | |
| "grad_norm": 0.552470326423645, | |
| "learning_rate": 1.2777777777777777e-05, | |
| "loss": 2.0063, | |
| "step": 29 | |
| }, | |
| { | |
| "epoch": 1.901639344262295, | |
| "eval_loss": 2.071108102798462, | |
| "eval_runtime": 46.7582, | |
| "eval_samples_per_second": 0.556, | |
| "eval_steps_per_second": 0.278, | |
| "step": 29 | |
| }, | |
| { | |
| "epoch": 1.9672131147540983, | |
| "grad_norm": 0.5569080114364624, | |
| "learning_rate": 1.25e-05, | |
| "loss": 2.0361, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 1.9672131147540983, | |
| "eval_loss": 2.059589385986328, | |
| "eval_runtime": 46.7694, | |
| "eval_samples_per_second": 0.556, | |
| "eval_steps_per_second": 0.278, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 2.0327868852459017, | |
| "grad_norm": 0.5699050426483154, | |
| "learning_rate": 1.2222222222222224e-05, | |
| "loss": 1.9885, | |
| "step": 31 | |
| }, | |
| { | |
| "epoch": 2.0327868852459017, | |
| "eval_loss": 2.048337459564209, | |
| "eval_runtime": 46.8642, | |
| "eval_samples_per_second": 0.555, | |
| "eval_steps_per_second": 0.277, | |
| "step": 31 | |
| }, | |
| { | |
| "epoch": 2.098360655737705, | |
| "grad_norm": 0.5750185251235962, | |
| "learning_rate": 1.1944444444444444e-05, | |
| "loss": 2.0413, | |
| "step": 32 | |
| }, | |
| { | |
| "epoch": 2.098360655737705, | |
| "eval_loss": 2.0375266075134277, | |
| "eval_runtime": 46.7785, | |
| "eval_samples_per_second": 0.556, | |
| "eval_steps_per_second": 0.278, | |
| "step": 32 | |
| }, | |
| { | |
| "epoch": 2.1639344262295084, | |
| "grad_norm": 0.5735768675804138, | |
| "learning_rate": 1.1666666666666668e-05, | |
| "loss": 1.9587, | |
| "step": 33 | |
| }, | |
| { | |
| "epoch": 2.1639344262295084, | |
| "eval_loss": 2.026817798614502, | |
| "eval_runtime": 46.7523, | |
| "eval_samples_per_second": 0.556, | |
| "eval_steps_per_second": 0.278, | |
| "step": 33 | |
| }, | |
| { | |
| "epoch": 2.2295081967213113, | |
| "grad_norm": 0.551645040512085, | |
| "learning_rate": 1.138888888888889e-05, | |
| "loss": 1.9807, | |
| "step": 34 | |
| }, | |
| { | |
| "epoch": 2.2295081967213113, | |
| "eval_loss": 2.0163915157318115, | |
| "eval_runtime": 46.8045, | |
| "eval_samples_per_second": 0.556, | |
| "eval_steps_per_second": 0.278, | |
| "step": 34 | |
| }, | |
| { | |
| "epoch": 2.2950819672131146, | |
| "grad_norm": 0.5482759475708008, | |
| "learning_rate": 1.1111111111111113e-05, | |
| "loss": 1.9198, | |
| "step": 35 | |
| }, | |
| { | |
| "epoch": 2.2950819672131146, | |
| "eval_loss": 2.0062360763549805, | |
| "eval_runtime": 46.762, | |
| "eval_samples_per_second": 0.556, | |
| "eval_steps_per_second": 0.278, | |
| "step": 35 | |
| }, | |
| { | |
| "epoch": 2.360655737704918, | |
| "grad_norm": 0.535072922706604, | |
| "learning_rate": 1.0833333333333334e-05, | |
| "loss": 1.9425, | |
| "step": 36 | |
| }, | |
| { | |
| "epoch": 2.360655737704918, | |
| "eval_loss": 1.9964103698730469, | |
| "eval_runtime": 46.7787, | |
| "eval_samples_per_second": 0.556, | |
| "eval_steps_per_second": 0.278, | |
| "step": 36 | |
| }, | |
| { | |
| "epoch": 2.4262295081967213, | |
| "grad_norm": 0.5778989791870117, | |
| "learning_rate": 1.0555555555555557e-05, | |
| "loss": 1.895, | |
| "step": 37 | |
| }, | |
| { | |
| "epoch": 2.4262295081967213, | |
| "eval_loss": 1.9869199991226196, | |
| "eval_runtime": 46.7633, | |
| "eval_samples_per_second": 0.556, | |
| "eval_steps_per_second": 0.278, | |
| "step": 37 | |
| }, | |
| { | |
| "epoch": 2.4918032786885247, | |
| "grad_norm": 0.5230702757835388, | |
| "learning_rate": 1.0277777777777777e-05, | |
| "loss": 1.9428, | |
| "step": 38 | |
| }, | |
| { | |
| "epoch": 2.4918032786885247, | |
| "eval_loss": 1.9776345491409302, | |
| "eval_runtime": 46.7004, | |
| "eval_samples_per_second": 0.557, | |
| "eval_steps_per_second": 0.278, | |
| "step": 38 | |
| }, | |
| { | |
| "epoch": 2.557377049180328, | |
| "grad_norm": 0.5722452402114868, | |
| "learning_rate": 1e-05, | |
| "loss": 1.8718, | |
| "step": 39 | |
| }, | |
| { | |
| "epoch": 2.557377049180328, | |
| "eval_loss": 1.9689091444015503, | |
| "eval_runtime": 46.7098, | |
| "eval_samples_per_second": 0.557, | |
| "eval_steps_per_second": 0.278, | |
| "step": 39 | |
| }, | |
| { | |
| "epoch": 2.6229508196721314, | |
| "grad_norm": 0.5922034382820129, | |
| "learning_rate": 9.722222222222223e-06, | |
| "loss": 1.8852, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 2.6229508196721314, | |
| "eval_loss": 1.9604159593582153, | |
| "eval_runtime": 46.7444, | |
| "eval_samples_per_second": 0.556, | |
| "eval_steps_per_second": 0.278, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 2.6885245901639343, | |
| "grad_norm": 0.6062042117118835, | |
| "learning_rate": 9.444444444444445e-06, | |
| "loss": 1.9446, | |
| "step": 41 | |
| }, | |
| { | |
| "epoch": 2.6885245901639343, | |
| "eval_loss": 1.952351450920105, | |
| "eval_runtime": 46.7995, | |
| "eval_samples_per_second": 0.556, | |
| "eval_steps_per_second": 0.278, | |
| "step": 41 | |
| }, | |
| { | |
| "epoch": 2.7540983606557377, | |
| "grad_norm": 0.5561647415161133, | |
| "learning_rate": 9.166666666666666e-06, | |
| "loss": 1.9389, | |
| "step": 42 | |
| }, | |
| { | |
| "epoch": 2.7540983606557377, | |
| "eval_loss": 1.9445891380310059, | |
| "eval_runtime": 46.7054, | |
| "eval_samples_per_second": 0.557, | |
| "eval_steps_per_second": 0.278, | |
| "step": 42 | |
| }, | |
| { | |
| "epoch": 2.819672131147541, | |
| "grad_norm": 0.5309848189353943, | |
| "learning_rate": 8.888888888888888e-06, | |
| "loss": 1.8356, | |
| "step": 43 | |
| }, | |
| { | |
| "epoch": 2.819672131147541, | |
| "eval_loss": 1.9371622800827026, | |
| "eval_runtime": 46.6887, | |
| "eval_samples_per_second": 0.557, | |
| "eval_steps_per_second": 0.278, | |
| "step": 43 | |
| }, | |
| { | |
| "epoch": 2.8852459016393444, | |
| "grad_norm": 0.5254234075546265, | |
| "learning_rate": 8.611111111111112e-06, | |
| "loss": 1.7884, | |
| "step": 44 | |
| }, | |
| { | |
| "epoch": 2.8852459016393444, | |
| "eval_loss": 1.9301851987838745, | |
| "eval_runtime": 46.6917, | |
| "eval_samples_per_second": 0.557, | |
| "eval_steps_per_second": 0.278, | |
| "step": 44 | |
| }, | |
| { | |
| "epoch": 2.9508196721311473, | |
| "grad_norm": 0.5549314022064209, | |
| "learning_rate": 8.333333333333334e-06, | |
| "loss": 1.8899, | |
| "step": 45 | |
| }, | |
| { | |
| "epoch": 2.9508196721311473, | |
| "eval_loss": 1.923454761505127, | |
| "eval_runtime": 46.7648, | |
| "eval_samples_per_second": 0.556, | |
| "eval_steps_per_second": 0.278, | |
| "step": 45 | |
| }, | |
| { | |
| "epoch": 3.0163934426229506, | |
| "grad_norm": 0.5211459398269653, | |
| "learning_rate": 8.055555555555557e-06, | |
| "loss": 1.8435, | |
| "step": 46 | |
| }, | |
| { | |
| "epoch": 3.0163934426229506, | |
| "eval_loss": 1.9169979095458984, | |
| "eval_runtime": 46.7864, | |
| "eval_samples_per_second": 0.556, | |
| "eval_steps_per_second": 0.278, | |
| "step": 46 | |
| }, | |
| { | |
| "epoch": 3.081967213114754, | |
| "grad_norm": 0.5590901970863342, | |
| "learning_rate": 7.77777777777778e-06, | |
| "loss": 1.8431, | |
| "step": 47 | |
| }, | |
| { | |
| "epoch": 3.081967213114754, | |
| "eval_loss": 1.910991907119751, | |
| "eval_runtime": 46.7269, | |
| "eval_samples_per_second": 0.556, | |
| "eval_steps_per_second": 0.278, | |
| "step": 47 | |
| }, | |
| { | |
| "epoch": 3.1475409836065573, | |
| "grad_norm": 0.5320960283279419, | |
| "learning_rate": 7.500000000000001e-06, | |
| "loss": 1.7688, | |
| "step": 48 | |
| }, | |
| { | |
| "epoch": 3.1475409836065573, | |
| "eval_loss": 1.905371904373169, | |
| "eval_runtime": 46.6869, | |
| "eval_samples_per_second": 0.557, | |
| "eval_steps_per_second": 0.278, | |
| "step": 48 | |
| }, | |
| { | |
| "epoch": 3.2131147540983607, | |
| "grad_norm": 0.49724167585372925, | |
| "learning_rate": 7.222222222222223e-06, | |
| "loss": 1.8617, | |
| "step": 49 | |
| }, | |
| { | |
| "epoch": 3.2131147540983607, | |
| "eval_loss": 1.9000388383865356, | |
| "eval_runtime": 46.7814, | |
| "eval_samples_per_second": 0.556, | |
| "eval_steps_per_second": 0.278, | |
| "step": 49 | |
| }, | |
| { | |
| "epoch": 3.278688524590164, | |
| "grad_norm": 0.5152797102928162, | |
| "learning_rate": 6.944444444444445e-06, | |
| "loss": 1.8381, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 3.278688524590164, | |
| "eval_loss": 1.8951764106750488, | |
| "eval_runtime": 46.7657, | |
| "eval_samples_per_second": 0.556, | |
| "eval_steps_per_second": 0.278, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 3.3442622950819674, | |
| "grad_norm": 0.5719711780548096, | |
| "learning_rate": 6.666666666666667e-06, | |
| "loss": 1.8516, | |
| "step": 51 | |
| }, | |
| { | |
| "epoch": 3.3442622950819674, | |
| "eval_loss": 1.8905354738235474, | |
| "eval_runtime": 46.6775, | |
| "eval_samples_per_second": 0.557, | |
| "eval_steps_per_second": 0.279, | |
| "step": 51 | |
| }, | |
| { | |
| "epoch": 3.4098360655737707, | |
| "grad_norm": 0.5465454459190369, | |
| "learning_rate": 6.3888888888888885e-06, | |
| "loss": 1.8267, | |
| "step": 52 | |
| }, | |
| { | |
| "epoch": 3.4098360655737707, | |
| "eval_loss": 1.88612961769104, | |
| "eval_runtime": 46.6966, | |
| "eval_samples_per_second": 0.557, | |
| "eval_steps_per_second": 0.278, | |
| "step": 52 | |
| }, | |
| { | |
| "epoch": 3.4754098360655736, | |
| "grad_norm": 0.5389249324798584, | |
| "learning_rate": 6.111111111111112e-06, | |
| "loss": 1.8758, | |
| "step": 53 | |
| }, | |
| { | |
| "epoch": 3.4754098360655736, | |
| "eval_loss": 1.8819791078567505, | |
| "eval_runtime": 46.6711, | |
| "eval_samples_per_second": 0.557, | |
| "eval_steps_per_second": 0.279, | |
| "step": 53 | |
| }, | |
| { | |
| "epoch": 3.540983606557377, | |
| "grad_norm": 0.495549738407135, | |
| "learning_rate": 5.833333333333334e-06, | |
| "loss": 1.7792, | |
| "step": 54 | |
| }, | |
| { | |
| "epoch": 3.540983606557377, | |
| "eval_loss": 1.8781176805496216, | |
| "eval_runtime": 46.7024, | |
| "eval_samples_per_second": 0.557, | |
| "eval_steps_per_second": 0.278, | |
| "step": 54 | |
| }, | |
| { | |
| "epoch": 3.6065573770491803, | |
| "grad_norm": 0.49404528737068176, | |
| "learning_rate": 5.555555555555557e-06, | |
| "loss": 1.7878, | |
| "step": 55 | |
| }, | |
| { | |
| "epoch": 3.6065573770491803, | |
| "eval_loss": 1.8743711709976196, | |
| "eval_runtime": 46.725, | |
| "eval_samples_per_second": 0.556, | |
| "eval_steps_per_second": 0.278, | |
| "step": 55 | |
| }, | |
| { | |
| "epoch": 3.6721311475409837, | |
| "grad_norm": 0.5066283941268921, | |
| "learning_rate": 5.2777777777777785e-06, | |
| "loss": 1.763, | |
| "step": 56 | |
| }, | |
| { | |
| "epoch": 3.6721311475409837, | |
| "eval_loss": 1.8708436489105225, | |
| "eval_runtime": 46.7333, | |
| "eval_samples_per_second": 0.556, | |
| "eval_steps_per_second": 0.278, | |
| "step": 56 | |
| }, | |
| { | |
| "epoch": 3.737704918032787, | |
| "grad_norm": 0.5295591950416565, | |
| "learning_rate": 5e-06, | |
| "loss": 1.8013, | |
| "step": 57 | |
| }, | |
| { | |
| "epoch": 3.737704918032787, | |
| "eval_loss": 1.8677045106887817, | |
| "eval_runtime": 46.7182, | |
| "eval_samples_per_second": 0.557, | |
| "eval_steps_per_second": 0.278, | |
| "step": 57 | |
| }, | |
| { | |
| "epoch": 3.80327868852459, | |
| "grad_norm": 0.52997225522995, | |
| "learning_rate": 4.722222222222222e-06, | |
| "loss": 1.8064, | |
| "step": 58 | |
| }, | |
| { | |
| "epoch": 3.80327868852459, | |
| "eval_loss": 1.8646008968353271, | |
| "eval_runtime": 46.7441, | |
| "eval_samples_per_second": 0.556, | |
| "eval_steps_per_second": 0.278, | |
| "step": 58 | |
| }, | |
| { | |
| "epoch": 3.8688524590163933, | |
| "grad_norm": 0.6026036739349365, | |
| "learning_rate": 4.444444444444444e-06, | |
| "loss": 1.7201, | |
| "step": 59 | |
| }, | |
| { | |
| "epoch": 3.8688524590163933, | |
| "eval_loss": 1.8619288206100464, | |
| "eval_runtime": 46.7352, | |
| "eval_samples_per_second": 0.556, | |
| "eval_steps_per_second": 0.278, | |
| "step": 59 | |
| }, | |
| { | |
| "epoch": 3.9344262295081966, | |
| "grad_norm": 0.5750531554222107, | |
| "learning_rate": 4.166666666666667e-06, | |
| "loss": 1.7895, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 3.9344262295081966, | |
| "eval_loss": 1.859217882156372, | |
| "eval_runtime": 46.7301, | |
| "eval_samples_per_second": 0.556, | |
| "eval_steps_per_second": 0.278, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 4.0, | |
| "grad_norm": 0.502881646156311, | |
| "learning_rate": 3.88888888888889e-06, | |
| "loss": 1.7837, | |
| "step": 61 | |
| }, | |
| { | |
| "epoch": 4.0, | |
| "eval_loss": 1.8568609952926636, | |
| "eval_runtime": 46.7362, | |
| "eval_samples_per_second": 0.556, | |
| "eval_steps_per_second": 0.278, | |
| "step": 61 | |
| }, | |
| { | |
| "epoch": 4.065573770491803, | |
| "grad_norm": 0.5146604776382446, | |
| "learning_rate": 3.6111111111111115e-06, | |
| "loss": 1.6988, | |
| "step": 62 | |
| }, | |
| { | |
| "epoch": 4.065573770491803, | |
| "eval_loss": 1.8548047542572021, | |
| "eval_runtime": 46.7365, | |
| "eval_samples_per_second": 0.556, | |
| "eval_steps_per_second": 0.278, | |
| "step": 62 | |
| }, | |
| { | |
| "epoch": 4.131147540983607, | |
| "grad_norm": 0.5581990480422974, | |
| "learning_rate": 3.3333333333333333e-06, | |
| "loss": 1.8038, | |
| "step": 63 | |
| }, | |
| { | |
| "epoch": 4.131147540983607, | |
| "eval_loss": 1.8527864217758179, | |
| "eval_runtime": 46.7975, | |
| "eval_samples_per_second": 0.556, | |
| "eval_steps_per_second": 0.278, | |
| "step": 63 | |
| }, | |
| { | |
| "epoch": 4.19672131147541, | |
| "grad_norm": 0.5218520164489746, | |
| "learning_rate": 3.055555555555556e-06, | |
| "loss": 1.7717, | |
| "step": 64 | |
| }, | |
| { | |
| "epoch": 4.19672131147541, | |
| "eval_loss": 1.850966215133667, | |
| "eval_runtime": 46.7506, | |
| "eval_samples_per_second": 0.556, | |
| "eval_steps_per_second": 0.278, | |
| "step": 64 | |
| }, | |
| { | |
| "epoch": 4.262295081967213, | |
| "grad_norm": 0.5471401214599609, | |
| "learning_rate": 2.7777777777777783e-06, | |
| "loss": 1.7708, | |
| "step": 65 | |
| }, | |
| { | |
| "epoch": 4.262295081967213, | |
| "eval_loss": 1.849280834197998, | |
| "eval_runtime": 46.7943, | |
| "eval_samples_per_second": 0.556, | |
| "eval_steps_per_second": 0.278, | |
| "step": 65 | |
| }, | |
| { | |
| "epoch": 4.327868852459017, | |
| "grad_norm": 0.49679359793663025, | |
| "learning_rate": 2.5e-06, | |
| "loss": 1.7721, | |
| "step": 66 | |
| }, | |
| { | |
| "epoch": 4.327868852459017, | |
| "eval_loss": 1.847765564918518, | |
| "eval_runtime": 46.7913, | |
| "eval_samples_per_second": 0.556, | |
| "eval_steps_per_second": 0.278, | |
| "step": 66 | |
| }, | |
| { | |
| "epoch": 4.39344262295082, | |
| "grad_norm": 0.5536015033721924, | |
| "learning_rate": 2.222222222222222e-06, | |
| "loss": 1.7757, | |
| "step": 67 | |
| }, | |
| { | |
| "epoch": 4.39344262295082, | |
| "eval_loss": 1.8464475870132446, | |
| "eval_runtime": 46.7742, | |
| "eval_samples_per_second": 0.556, | |
| "eval_steps_per_second": 0.278, | |
| "step": 67 | |
| }, | |
| { | |
| "epoch": 4.459016393442623, | |
| "grad_norm": 0.5509213209152222, | |
| "learning_rate": 1.944444444444445e-06, | |
| "loss": 1.8043, | |
| "step": 68 | |
| }, | |
| { | |
| "epoch": 4.459016393442623, | |
| "eval_loss": 1.8451799154281616, | |
| "eval_runtime": 46.7554, | |
| "eval_samples_per_second": 0.556, | |
| "eval_steps_per_second": 0.278, | |
| "step": 68 | |
| }, | |
| { | |
| "epoch": 4.524590163934426, | |
| "grad_norm": 0.49126845598220825, | |
| "learning_rate": 1.6666666666666667e-06, | |
| "loss": 1.7836, | |
| "step": 69 | |
| }, | |
| { | |
| "epoch": 4.524590163934426, | |
| "eval_loss": 1.8441420793533325, | |
| "eval_runtime": 46.7716, | |
| "eval_samples_per_second": 0.556, | |
| "eval_steps_per_second": 0.278, | |
| "step": 69 | |
| }, | |
| { | |
| "epoch": 4.590163934426229, | |
| "grad_norm": 0.5862410664558411, | |
| "learning_rate": 1.3888888888888892e-06, | |
| "loss": 1.8036, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 4.590163934426229, | |
| "eval_loss": 1.8432250022888184, | |
| "eval_runtime": 46.7239, | |
| "eval_samples_per_second": 0.556, | |
| "eval_steps_per_second": 0.278, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 4.655737704918033, | |
| "grad_norm": 0.5185004472732544, | |
| "learning_rate": 1.111111111111111e-06, | |
| "loss": 1.8121, | |
| "step": 71 | |
| }, | |
| { | |
| "epoch": 4.655737704918033, | |
| "eval_loss": 1.8424735069274902, | |
| "eval_runtime": 46.753, | |
| "eval_samples_per_second": 0.556, | |
| "eval_steps_per_second": 0.278, | |
| "step": 71 | |
| }, | |
| { | |
| "epoch": 4.721311475409836, | |
| "grad_norm": 0.5338285565376282, | |
| "learning_rate": 8.333333333333333e-07, | |
| "loss": 1.7523, | |
| "step": 72 | |
| }, | |
| { | |
| "epoch": 4.721311475409836, | |
| "eval_loss": 1.8418388366699219, | |
| "eval_runtime": 46.7365, | |
| "eval_samples_per_second": 0.556, | |
| "eval_steps_per_second": 0.278, | |
| "step": 72 | |
| }, | |
| { | |
| "epoch": 4.786885245901639, | |
| "grad_norm": 0.5311003923416138, | |
| "learning_rate": 5.555555555555555e-07, | |
| "loss": 1.7532, | |
| "step": 73 | |
| }, | |
| { | |
| "epoch": 4.786885245901639, | |
| "eval_loss": 1.841430425643921, | |
| "eval_runtime": 46.7626, | |
| "eval_samples_per_second": 0.556, | |
| "eval_steps_per_second": 0.278, | |
| "step": 73 | |
| }, | |
| { | |
| "epoch": 4.852459016393443, | |
| "grad_norm": 0.551410436630249, | |
| "learning_rate": 2.7777777777777776e-07, | |
| "loss": 1.7593, | |
| "step": 74 | |
| }, | |
| { | |
| "epoch": 4.852459016393443, | |
| "eval_loss": 1.8412089347839355, | |
| "eval_runtime": 46.8159, | |
| "eval_samples_per_second": 0.555, | |
| "eval_steps_per_second": 0.278, | |
| "step": 74 | |
| }, | |
| { | |
| "epoch": 4.918032786885246, | |
| "grad_norm": 0.49288997054100037, | |
| "learning_rate": 0.0, | |
| "loss": 1.7316, | |
| "step": 75 | |
| }, | |
| { | |
| "epoch": 4.918032786885246, | |
| "eval_loss": 1.8410229682922363, | |
| "eval_runtime": 46.8396, | |
| "eval_samples_per_second": 0.555, | |
| "eval_steps_per_second": 0.278, | |
| "step": 75 | |
| } | |
| ], | |
| "logging_steps": 1, | |
| "max_steps": 75, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 5, | |
| "save_steps": 1, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": true | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 2.224575101730816e+17, | |
| "train_batch_size": 2, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |