| { | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 9.999672313792313, | |
| "eval_steps": 500, | |
| "global_step": 76290, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 2.980993577139861e-06, | |
| "loss": 1.8224, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 2.9619871542797224e-06, | |
| "loss": 1.6774, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 2.9429807314195834e-06, | |
| "loss": 1.651, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 2.9239743085594444e-06, | |
| "loss": 1.6331, | |
| "step": 2000 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 2.9049678856993053e-06, | |
| "loss": 1.6218, | |
| "step": 2500 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 2.8859614628391667e-06, | |
| "loss": 1.6129, | |
| "step": 3000 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 2.8669550399790273e-06, | |
| "loss": 1.6042, | |
| "step": 3500 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "learning_rate": 2.8479486171188882e-06, | |
| "loss": 1.596, | |
| "step": 4000 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "learning_rate": 2.8289421942587496e-06, | |
| "loss": 1.5899, | |
| "step": 4500 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "learning_rate": 2.8099357713986106e-06, | |
| "loss": 1.5834, | |
| "step": 5000 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "learning_rate": 2.7909293485384715e-06, | |
| "loss": 1.5772, | |
| "step": 5500 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "learning_rate": 2.771922925678333e-06, | |
| "loss": 1.5719, | |
| "step": 6000 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "learning_rate": 2.752916502818194e-06, | |
| "loss": 1.567, | |
| "step": 6500 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "learning_rate": 2.733910079958055e-06, | |
| "loss": 1.5625, | |
| "step": 7000 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "learning_rate": 2.714903657097916e-06, | |
| "loss": 1.5588, | |
| "step": 7500 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "eval_loss": 1.544177770614624, | |
| "eval_runtime": 7.0475, | |
| "eval_samples_per_second": 139.908, | |
| "eval_steps_per_second": 0.851, | |
| "step": 7629 | |
| }, | |
| { | |
| "epoch": 1.05, | |
| "learning_rate": 2.695897234237777e-06, | |
| "loss": 1.5539, | |
| "step": 8000 | |
| }, | |
| { | |
| "epoch": 1.11, | |
| "learning_rate": 2.6768908113776377e-06, | |
| "loss": 1.5502, | |
| "step": 8500 | |
| }, | |
| { | |
| "epoch": 1.18, | |
| "learning_rate": 2.6578843885174987e-06, | |
| "loss": 1.5473, | |
| "step": 9000 | |
| }, | |
| { | |
| "epoch": 1.25, | |
| "learning_rate": 2.63887796565736e-06, | |
| "loss": 1.5443, | |
| "step": 9500 | |
| }, | |
| { | |
| "epoch": 1.31, | |
| "learning_rate": 2.619871542797221e-06, | |
| "loss": 1.5411, | |
| "step": 10000 | |
| }, | |
| { | |
| "epoch": 1.38, | |
| "learning_rate": 2.600865119937082e-06, | |
| "loss": 1.5391, | |
| "step": 10500 | |
| }, | |
| { | |
| "epoch": 1.44, | |
| "learning_rate": 2.5818586970769434e-06, | |
| "loss": 1.5366, | |
| "step": 11000 | |
| }, | |
| { | |
| "epoch": 1.51, | |
| "learning_rate": 2.5628522742168044e-06, | |
| "loss": 1.5344, | |
| "step": 11500 | |
| }, | |
| { | |
| "epoch": 1.57, | |
| "learning_rate": 2.5438458513566653e-06, | |
| "loss": 1.532, | |
| "step": 12000 | |
| }, | |
| { | |
| "epoch": 1.64, | |
| "learning_rate": 2.5248394284965263e-06, | |
| "loss": 1.5295, | |
| "step": 12500 | |
| }, | |
| { | |
| "epoch": 1.7, | |
| "learning_rate": 2.5058330056363877e-06, | |
| "loss": 1.5286, | |
| "step": 13000 | |
| }, | |
| { | |
| "epoch": 1.77, | |
| "learning_rate": 2.4868265827762487e-06, | |
| "loss": 1.526, | |
| "step": 13500 | |
| }, | |
| { | |
| "epoch": 1.84, | |
| "learning_rate": 2.4678201599161096e-06, | |
| "loss": 1.5247, | |
| "step": 14000 | |
| }, | |
| { | |
| "epoch": 1.9, | |
| "learning_rate": 2.448813737055971e-06, | |
| "loss": 1.5235, | |
| "step": 14500 | |
| }, | |
| { | |
| "epoch": 1.97, | |
| "learning_rate": 2.429807314195832e-06, | |
| "loss": 1.5219, | |
| "step": 15000 | |
| }, | |
| { | |
| "epoch": 2.0, | |
| "eval_loss": 1.5164680480957031, | |
| "eval_runtime": 6.9148, | |
| "eval_samples_per_second": 142.593, | |
| "eval_steps_per_second": 0.868, | |
| "step": 15258 | |
| }, | |
| { | |
| "epoch": 2.03, | |
| "learning_rate": 2.410800891335693e-06, | |
| "loss": 1.5194, | |
| "step": 15500 | |
| }, | |
| { | |
| "epoch": 2.1, | |
| "learning_rate": 2.391794468475554e-06, | |
| "loss": 1.5185, | |
| "step": 16000 | |
| }, | |
| { | |
| "epoch": 2.16, | |
| "learning_rate": 2.372788045615415e-06, | |
| "loss": 1.5166, | |
| "step": 16500 | |
| }, | |
| { | |
| "epoch": 2.23, | |
| "learning_rate": 2.353781622755276e-06, | |
| "loss": 1.516, | |
| "step": 17000 | |
| }, | |
| { | |
| "epoch": 2.29, | |
| "learning_rate": 2.334775199895137e-06, | |
| "loss": 1.5149, | |
| "step": 17500 | |
| }, | |
| { | |
| "epoch": 2.36, | |
| "learning_rate": 2.315768777034998e-06, | |
| "loss": 1.5137, | |
| "step": 18000 | |
| }, | |
| { | |
| "epoch": 2.42, | |
| "learning_rate": 2.296762354174859e-06, | |
| "loss": 1.5127, | |
| "step": 18500 | |
| }, | |
| { | |
| "epoch": 2.49, | |
| "learning_rate": 2.27775593131472e-06, | |
| "loss": 1.5111, | |
| "step": 19000 | |
| }, | |
| { | |
| "epoch": 2.56, | |
| "learning_rate": 2.258749508454581e-06, | |
| "loss": 1.5103, | |
| "step": 19500 | |
| }, | |
| { | |
| "epoch": 2.62, | |
| "learning_rate": 2.239743085594442e-06, | |
| "loss": 1.5095, | |
| "step": 20000 | |
| }, | |
| { | |
| "epoch": 2.69, | |
| "learning_rate": 2.220736662734303e-06, | |
| "loss": 1.5082, | |
| "step": 20500 | |
| }, | |
| { | |
| "epoch": 2.75, | |
| "learning_rate": 2.2017302398741644e-06, | |
| "loss": 1.5075, | |
| "step": 21000 | |
| }, | |
| { | |
| "epoch": 2.82, | |
| "learning_rate": 2.1827238170140254e-06, | |
| "loss": 1.5064, | |
| "step": 21500 | |
| }, | |
| { | |
| "epoch": 2.88, | |
| "learning_rate": 2.1637173941538863e-06, | |
| "loss": 1.5056, | |
| "step": 22000 | |
| }, | |
| { | |
| "epoch": 2.95, | |
| "learning_rate": 2.1447109712937477e-06, | |
| "loss": 1.5053, | |
| "step": 22500 | |
| }, | |
| { | |
| "epoch": 3.0, | |
| "eval_loss": 1.5032986402511597, | |
| "eval_runtime": 7.0258, | |
| "eval_samples_per_second": 140.341, | |
| "eval_steps_per_second": 0.854, | |
| "step": 22887 | |
| }, | |
| { | |
| "epoch": 3.01, | |
| "learning_rate": 2.1257045484336087e-06, | |
| "loss": 1.5047, | |
| "step": 23000 | |
| }, | |
| { | |
| "epoch": 3.08, | |
| "learning_rate": 2.1066981255734696e-06, | |
| "loss": 1.5028, | |
| "step": 23500 | |
| }, | |
| { | |
| "epoch": 3.15, | |
| "learning_rate": 2.0876917027133306e-06, | |
| "loss": 1.5022, | |
| "step": 24000 | |
| }, | |
| { | |
| "epoch": 3.21, | |
| "learning_rate": 2.068685279853192e-06, | |
| "loss": 1.5014, | |
| "step": 24500 | |
| }, | |
| { | |
| "epoch": 3.28, | |
| "learning_rate": 2.049678856993053e-06, | |
| "loss": 1.5005, | |
| "step": 25000 | |
| }, | |
| { | |
| "epoch": 3.34, | |
| "learning_rate": 2.030672434132914e-06, | |
| "loss": 1.5001, | |
| "step": 25500 | |
| }, | |
| { | |
| "epoch": 3.41, | |
| "learning_rate": 2.011666011272775e-06, | |
| "loss": 1.4992, | |
| "step": 26000 | |
| }, | |
| { | |
| "epoch": 3.47, | |
| "learning_rate": 1.9926595884126363e-06, | |
| "loss": 1.4986, | |
| "step": 26500 | |
| }, | |
| { | |
| "epoch": 3.54, | |
| "learning_rate": 1.9736531655524972e-06, | |
| "loss": 1.498, | |
| "step": 27000 | |
| }, | |
| { | |
| "epoch": 3.6, | |
| "learning_rate": 1.9546467426923586e-06, | |
| "loss": 1.4974, | |
| "step": 27500 | |
| }, | |
| { | |
| "epoch": 3.67, | |
| "learning_rate": 1.9356403198322196e-06, | |
| "loss": 1.4969, | |
| "step": 28000 | |
| }, | |
| { | |
| "epoch": 3.74, | |
| "learning_rate": 1.91663389697208e-06, | |
| "loss": 1.4963, | |
| "step": 28500 | |
| }, | |
| { | |
| "epoch": 3.8, | |
| "learning_rate": 1.8976274741119413e-06, | |
| "loss": 1.4957, | |
| "step": 29000 | |
| }, | |
| { | |
| "epoch": 3.87, | |
| "learning_rate": 1.8786210512518025e-06, | |
| "loss": 1.4952, | |
| "step": 29500 | |
| }, | |
| { | |
| "epoch": 3.93, | |
| "learning_rate": 1.8596146283916634e-06, | |
| "loss": 1.4945, | |
| "step": 30000 | |
| }, | |
| { | |
| "epoch": 4.0, | |
| "learning_rate": 1.8406082055315244e-06, | |
| "loss": 1.4935, | |
| "step": 30500 | |
| }, | |
| { | |
| "epoch": 4.0, | |
| "eval_loss": 1.4955875873565674, | |
| "eval_runtime": 6.9086, | |
| "eval_samples_per_second": 142.721, | |
| "eval_steps_per_second": 0.868, | |
| "step": 30517 | |
| }, | |
| { | |
| "epoch": 4.06, | |
| "learning_rate": 1.8216017826713858e-06, | |
| "loss": 1.493, | |
| "step": 31000 | |
| }, | |
| { | |
| "epoch": 4.13, | |
| "learning_rate": 1.8025953598112466e-06, | |
| "loss": 1.4925, | |
| "step": 31500 | |
| }, | |
| { | |
| "epoch": 4.19, | |
| "learning_rate": 1.7835889369511077e-06, | |
| "loss": 1.4921, | |
| "step": 32000 | |
| }, | |
| { | |
| "epoch": 4.26, | |
| "learning_rate": 1.764582514090969e-06, | |
| "loss": 1.4916, | |
| "step": 32500 | |
| }, | |
| { | |
| "epoch": 4.33, | |
| "learning_rate": 1.7455760912308294e-06, | |
| "loss": 1.4914, | |
| "step": 33000 | |
| }, | |
| { | |
| "epoch": 4.39, | |
| "learning_rate": 1.7265696683706906e-06, | |
| "loss": 1.4906, | |
| "step": 33500 | |
| }, | |
| { | |
| "epoch": 4.46, | |
| "learning_rate": 1.7075632455105518e-06, | |
| "loss": 1.4902, | |
| "step": 34000 | |
| }, | |
| { | |
| "epoch": 4.52, | |
| "learning_rate": 1.688556822650413e-06, | |
| "loss": 1.4898, | |
| "step": 34500 | |
| }, | |
| { | |
| "epoch": 4.59, | |
| "learning_rate": 1.669550399790274e-06, | |
| "loss": 1.4892, | |
| "step": 35000 | |
| }, | |
| { | |
| "epoch": 4.65, | |
| "learning_rate": 1.650543976930135e-06, | |
| "loss": 1.4891, | |
| "step": 35500 | |
| }, | |
| { | |
| "epoch": 4.72, | |
| "learning_rate": 1.6315375540699963e-06, | |
| "loss": 1.4886, | |
| "step": 36000 | |
| }, | |
| { | |
| "epoch": 4.78, | |
| "learning_rate": 1.6125311312098573e-06, | |
| "loss": 1.4881, | |
| "step": 36500 | |
| }, | |
| { | |
| "epoch": 4.85, | |
| "learning_rate": 1.5935247083497182e-06, | |
| "loss": 1.4879, | |
| "step": 37000 | |
| }, | |
| { | |
| "epoch": 4.92, | |
| "learning_rate": 1.5745182854895796e-06, | |
| "loss": 1.4878, | |
| "step": 37500 | |
| }, | |
| { | |
| "epoch": 4.98, | |
| "learning_rate": 1.5555118626294406e-06, | |
| "loss": 1.4873, | |
| "step": 38000 | |
| }, | |
| { | |
| "epoch": 5.0, | |
| "eval_loss": 1.4904649257659912, | |
| "eval_runtime": 6.8983, | |
| "eval_samples_per_second": 142.934, | |
| "eval_steps_per_second": 0.87, | |
| "step": 38146 | |
| }, | |
| { | |
| "epoch": 5.05, | |
| "learning_rate": 1.5365054397693011e-06, | |
| "loss": 1.4875, | |
| "step": 38500 | |
| }, | |
| { | |
| "epoch": 5.11, | |
| "learning_rate": 1.5174990169091623e-06, | |
| "loss": 1.4863, | |
| "step": 39000 | |
| }, | |
| { | |
| "epoch": 5.18, | |
| "learning_rate": 1.4984925940490235e-06, | |
| "loss": 1.4857, | |
| "step": 39500 | |
| }, | |
| { | |
| "epoch": 5.24, | |
| "learning_rate": 1.4794861711888844e-06, | |
| "loss": 1.4854, | |
| "step": 40000 | |
| }, | |
| { | |
| "epoch": 5.31, | |
| "learning_rate": 1.4604797483287458e-06, | |
| "loss": 1.4848, | |
| "step": 40500 | |
| }, | |
| { | |
| "epoch": 5.37, | |
| "learning_rate": 1.4414733254686068e-06, | |
| "loss": 1.485, | |
| "step": 41000 | |
| }, | |
| { | |
| "epoch": 5.44, | |
| "learning_rate": 1.4224669026084677e-06, | |
| "loss": 1.4848, | |
| "step": 41500 | |
| }, | |
| { | |
| "epoch": 5.51, | |
| "learning_rate": 1.4034604797483287e-06, | |
| "loss": 1.4842, | |
| "step": 42000 | |
| }, | |
| { | |
| "epoch": 5.57, | |
| "learning_rate": 1.38445405688819e-06, | |
| "loss": 1.4839, | |
| "step": 42500 | |
| }, | |
| { | |
| "epoch": 5.64, | |
| "learning_rate": 1.365447634028051e-06, | |
| "loss": 1.4836, | |
| "step": 43000 | |
| }, | |
| { | |
| "epoch": 5.7, | |
| "learning_rate": 1.346441211167912e-06, | |
| "loss": 1.4834, | |
| "step": 43500 | |
| }, | |
| { | |
| "epoch": 5.77, | |
| "learning_rate": 1.327434788307773e-06, | |
| "loss": 1.4831, | |
| "step": 44000 | |
| }, | |
| { | |
| "epoch": 5.83, | |
| "learning_rate": 1.308428365447634e-06, | |
| "loss": 1.483, | |
| "step": 44500 | |
| }, | |
| { | |
| "epoch": 5.9, | |
| "learning_rate": 1.289421942587495e-06, | |
| "loss": 1.4827, | |
| "step": 45000 | |
| }, | |
| { | |
| "epoch": 5.96, | |
| "learning_rate": 1.2704155197273563e-06, | |
| "loss": 1.4826, | |
| "step": 45500 | |
| }, | |
| { | |
| "epoch": 6.0, | |
| "eval_loss": 1.4877293109893799, | |
| "eval_runtime": 6.98, | |
| "eval_samples_per_second": 141.261, | |
| "eval_steps_per_second": 0.86, | |
| "step": 45775 | |
| }, | |
| { | |
| "epoch": 6.03, | |
| "learning_rate": 1.2514090968672173e-06, | |
| "loss": 1.4818, | |
| "step": 46000 | |
| }, | |
| { | |
| "epoch": 6.09, | |
| "learning_rate": 1.2324026740070782e-06, | |
| "loss": 1.4819, | |
| "step": 46500 | |
| }, | |
| { | |
| "epoch": 6.16, | |
| "learning_rate": 1.2133962511469392e-06, | |
| "loss": 1.4815, | |
| "step": 47000 | |
| }, | |
| { | |
| "epoch": 6.23, | |
| "learning_rate": 1.1943898282868006e-06, | |
| "loss": 1.4814, | |
| "step": 47500 | |
| }, | |
| { | |
| "epoch": 6.29, | |
| "learning_rate": 1.1753834054266616e-06, | |
| "loss": 1.4809, | |
| "step": 48000 | |
| }, | |
| { | |
| "epoch": 6.36, | |
| "learning_rate": 1.1563769825665225e-06, | |
| "loss": 1.4803, | |
| "step": 48500 | |
| }, | |
| { | |
| "epoch": 6.42, | |
| "learning_rate": 1.1373705597063835e-06, | |
| "loss": 1.4808, | |
| "step": 49000 | |
| }, | |
| { | |
| "epoch": 6.49, | |
| "learning_rate": 1.1183641368462449e-06, | |
| "loss": 1.4806, | |
| "step": 49500 | |
| }, | |
| { | |
| "epoch": 6.55, | |
| "learning_rate": 1.0993577139861058e-06, | |
| "loss": 1.4803, | |
| "step": 50000 | |
| }, | |
| { | |
| "epoch": 6.62, | |
| "learning_rate": 1.0803512911259668e-06, | |
| "loss": 1.4801, | |
| "step": 50500 | |
| }, | |
| { | |
| "epoch": 6.68, | |
| "learning_rate": 1.0613448682658278e-06, | |
| "loss": 1.4797, | |
| "step": 51000 | |
| }, | |
| { | |
| "epoch": 6.75, | |
| "learning_rate": 1.0423384454056887e-06, | |
| "loss": 1.4796, | |
| "step": 51500 | |
| }, | |
| { | |
| "epoch": 6.82, | |
| "learning_rate": 1.02333202254555e-06, | |
| "loss": 1.4795, | |
| "step": 52000 | |
| }, | |
| { | |
| "epoch": 6.88, | |
| "learning_rate": 1.004325599685411e-06, | |
| "loss": 1.4793, | |
| "step": 52500 | |
| }, | |
| { | |
| "epoch": 6.95, | |
| "learning_rate": 9.85319176825272e-07, | |
| "loss": 1.479, | |
| "step": 53000 | |
| }, | |
| { | |
| "epoch": 7.0, | |
| "eval_loss": 1.4859654903411865, | |
| "eval_runtime": 7.0683, | |
| "eval_samples_per_second": 139.495, | |
| "eval_steps_per_second": 0.849, | |
| "step": 53404 | |
| }, | |
| { | |
| "epoch": 7.01, | |
| "learning_rate": 9.66312753965133e-07, | |
| "loss": 1.4786, | |
| "step": 53500 | |
| }, | |
| { | |
| "epoch": 7.08, | |
| "learning_rate": 9.473063311049942e-07, | |
| "loss": 1.4788, | |
| "step": 54000 | |
| }, | |
| { | |
| "epoch": 7.14, | |
| "learning_rate": 9.282999082448553e-07, | |
| "loss": 1.4785, | |
| "step": 54500 | |
| }, | |
| { | |
| "epoch": 7.21, | |
| "learning_rate": 9.092934853847164e-07, | |
| "loss": 1.4784, | |
| "step": 55000 | |
| }, | |
| { | |
| "epoch": 7.27, | |
| "learning_rate": 8.902870625245774e-07, | |
| "loss": 1.4781, | |
| "step": 55500 | |
| }, | |
| { | |
| "epoch": 7.34, | |
| "learning_rate": 8.712806396644386e-07, | |
| "loss": 1.478, | |
| "step": 56000 | |
| }, | |
| { | |
| "epoch": 7.41, | |
| "learning_rate": 8.522742168042993e-07, | |
| "loss": 1.4779, | |
| "step": 56500 | |
| }, | |
| { | |
| "epoch": 7.47, | |
| "learning_rate": 8.332677939441605e-07, | |
| "loss": 1.4779, | |
| "step": 57000 | |
| }, | |
| { | |
| "epoch": 7.54, | |
| "learning_rate": 8.142613710840215e-07, | |
| "loss": 1.4774, | |
| "step": 57500 | |
| }, | |
| { | |
| "epoch": 7.6, | |
| "learning_rate": 7.952549482238825e-07, | |
| "loss": 1.4775, | |
| "step": 58000 | |
| }, | |
| { | |
| "epoch": 7.67, | |
| "learning_rate": 7.762485253637437e-07, | |
| "loss": 1.4772, | |
| "step": 58500 | |
| }, | |
| { | |
| "epoch": 7.73, | |
| "learning_rate": 7.572421025036047e-07, | |
| "loss": 1.4771, | |
| "step": 59000 | |
| }, | |
| { | |
| "epoch": 7.8, | |
| "learning_rate": 7.382356796434658e-07, | |
| "loss": 1.4771, | |
| "step": 59500 | |
| }, | |
| { | |
| "epoch": 7.86, | |
| "learning_rate": 7.192292567833268e-07, | |
| "loss": 1.4769, | |
| "step": 60000 | |
| }, | |
| { | |
| "epoch": 7.93, | |
| "learning_rate": 7.00222833923188e-07, | |
| "loss": 1.4769, | |
| "step": 60500 | |
| }, | |
| { | |
| "epoch": 8.0, | |
| "learning_rate": 6.812164110630491e-07, | |
| "loss": 1.4771, | |
| "step": 61000 | |
| }, | |
| { | |
| "epoch": 8.0, | |
| "eval_loss": 1.483931303024292, | |
| "eval_runtime": 6.9286, | |
| "eval_samples_per_second": 142.308, | |
| "eval_steps_per_second": 0.866, | |
| "step": 61034 | |
| }, | |
| { | |
| "epoch": 8.06, | |
| "learning_rate": 6.622099882029101e-07, | |
| "loss": 1.4775, | |
| "step": 61500 | |
| }, | |
| { | |
| "epoch": 8.13, | |
| "learning_rate": 6.432035653427712e-07, | |
| "loss": 1.4765, | |
| "step": 62000 | |
| }, | |
| { | |
| "epoch": 8.19, | |
| "learning_rate": 6.24197142482632e-07, | |
| "loss": 1.4762, | |
| "step": 62500 | |
| }, | |
| { | |
| "epoch": 8.26, | |
| "learning_rate": 6.05190719622493e-07, | |
| "loss": 1.4764, | |
| "step": 63000 | |
| }, | |
| { | |
| "epoch": 8.32, | |
| "learning_rate": 5.861842967623541e-07, | |
| "loss": 1.4762, | |
| "step": 63500 | |
| }, | |
| { | |
| "epoch": 8.39, | |
| "learning_rate": 5.671778739022152e-07, | |
| "loss": 1.4762, | |
| "step": 64000 | |
| }, | |
| { | |
| "epoch": 8.45, | |
| "learning_rate": 5.481714510420763e-07, | |
| "loss": 1.4761, | |
| "step": 64500 | |
| }, | |
| { | |
| "epoch": 8.52, | |
| "learning_rate": 5.291650281819373e-07, | |
| "loss": 1.4759, | |
| "step": 65000 | |
| }, | |
| { | |
| "epoch": 8.59, | |
| "learning_rate": 5.101586053217985e-07, | |
| "loss": 1.4758, | |
| "step": 65500 | |
| }, | |
| { | |
| "epoch": 8.65, | |
| "learning_rate": 4.911521824616596e-07, | |
| "loss": 1.4755, | |
| "step": 66000 | |
| }, | |
| { | |
| "epoch": 8.72, | |
| "learning_rate": 4.7214575960152063e-07, | |
| "loss": 1.4754, | |
| "step": 66500 | |
| }, | |
| { | |
| "epoch": 8.78, | |
| "learning_rate": 4.531393367413817e-07, | |
| "loss": 1.4757, | |
| "step": 67000 | |
| }, | |
| { | |
| "epoch": 8.85, | |
| "learning_rate": 4.341329138812427e-07, | |
| "loss": 1.4758, | |
| "step": 67500 | |
| }, | |
| { | |
| "epoch": 8.91, | |
| "learning_rate": 4.151264910211035e-07, | |
| "loss": 1.4754, | |
| "step": 68000 | |
| }, | |
| { | |
| "epoch": 8.98, | |
| "learning_rate": 3.961200681609647e-07, | |
| "loss": 1.4756, | |
| "step": 68500 | |
| }, | |
| { | |
| "epoch": 9.0, | |
| "eval_loss": 1.4828386306762695, | |
| "eval_runtime": 6.934, | |
| "eval_samples_per_second": 142.198, | |
| "eval_steps_per_second": 0.865, | |
| "step": 68663 | |
| }, | |
| { | |
| "epoch": 9.04, | |
| "learning_rate": 3.771136453008257e-07, | |
| "loss": 1.4747, | |
| "step": 69000 | |
| }, | |
| { | |
| "epoch": 9.11, | |
| "learning_rate": 3.581072224406868e-07, | |
| "loss": 1.4753, | |
| "step": 69500 | |
| }, | |
| { | |
| "epoch": 9.18, | |
| "learning_rate": 3.3910079958054786e-07, | |
| "loss": 1.4752, | |
| "step": 70000 | |
| }, | |
| { | |
| "epoch": 9.24, | |
| "learning_rate": 3.20094376720409e-07, | |
| "loss": 1.4752, | |
| "step": 70500 | |
| }, | |
| { | |
| "epoch": 9.31, | |
| "learning_rate": 3.0108795386027005e-07, | |
| "loss": 1.4751, | |
| "step": 71000 | |
| }, | |
| { | |
| "epoch": 9.37, | |
| "learning_rate": 2.820815310001311e-07, | |
| "loss": 1.4751, | |
| "step": 71500 | |
| }, | |
| { | |
| "epoch": 9.44, | |
| "learning_rate": 2.630751081399922e-07, | |
| "loss": 1.4749, | |
| "step": 72000 | |
| }, | |
| { | |
| "epoch": 9.5, | |
| "learning_rate": 2.4406868527985326e-07, | |
| "loss": 1.4748, | |
| "step": 72500 | |
| }, | |
| { | |
| "epoch": 9.57, | |
| "learning_rate": 2.2506226241971439e-07, | |
| "loss": 1.4751, | |
| "step": 73000 | |
| }, | |
| { | |
| "epoch": 9.63, | |
| "learning_rate": 2.0605583955957543e-07, | |
| "loss": 1.4747, | |
| "step": 73500 | |
| }, | |
| { | |
| "epoch": 9.7, | |
| "learning_rate": 1.870494166994362e-07, | |
| "loss": 1.4748, | |
| "step": 74000 | |
| }, | |
| { | |
| "epoch": 9.77, | |
| "learning_rate": 1.6804299383929728e-07, | |
| "loss": 1.475, | |
| "step": 74500 | |
| }, | |
| { | |
| "epoch": 9.83, | |
| "learning_rate": 1.4903657097915838e-07, | |
| "loss": 1.4746, | |
| "step": 75000 | |
| }, | |
| { | |
| "epoch": 9.9, | |
| "learning_rate": 1.3003014811901947e-07, | |
| "loss": 1.4752, | |
| "step": 75500 | |
| }, | |
| { | |
| "epoch": 9.96, | |
| "learning_rate": 1.1102372525888054e-07, | |
| "loss": 1.4749, | |
| "step": 76000 | |
| }, | |
| { | |
| "epoch": 10.0, | |
| "eval_loss": 1.4823230504989624, | |
| "eval_runtime": 6.8873, | |
| "eval_samples_per_second": 143.161, | |
| "eval_steps_per_second": 0.871, | |
| "step": 76290 | |
| } | |
| ], | |
| "logging_steps": 500, | |
| "max_steps": 76290, | |
| "num_train_epochs": 10, | |
| "save_steps": 500, | |
| "total_flos": 1.609765756918825e+18, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |