| { | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 10.0, | |
| "global_step": 77340, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 9.935350400827516e-05, | |
| "loss": 8.0382, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 9.87070080165503e-05, | |
| "loss": 7.6429, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 9.806051202482544e-05, | |
| "loss": 7.1341, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 9.74140160331006e-05, | |
| "loss": 6.9301, | |
| "step": 2000 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 9.676752004137574e-05, | |
| "loss": 6.8441, | |
| "step": 2500 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 9.61210240496509e-05, | |
| "loss": 6.7735, | |
| "step": 3000 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 9.547452805792605e-05, | |
| "loss": 6.3512, | |
| "step": 3500 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "learning_rate": 9.482803206620118e-05, | |
| "loss": 5.6406, | |
| "step": 4000 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "learning_rate": 9.418153607447634e-05, | |
| "loss": 5.2743, | |
| "step": 4500 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "learning_rate": 9.35350400827515e-05, | |
| "loss": 5.0399, | |
| "step": 5000 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "learning_rate": 9.288854409102664e-05, | |
| "loss": 4.8407, | |
| "step": 5500 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "learning_rate": 9.224204809930179e-05, | |
| "loss": 4.6694, | |
| "step": 6000 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "learning_rate": 9.159555210757695e-05, | |
| "loss": 4.5262, | |
| "step": 6500 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "learning_rate": 9.094905611585208e-05, | |
| "loss": 4.4021, | |
| "step": 7000 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "learning_rate": 9.030256012412723e-05, | |
| "loss": 4.2888, | |
| "step": 7500 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "eval_loss": 4.0826334953308105, | |
| "eval_runtime": 475.8214, | |
| "eval_samples_per_second": 693.498, | |
| "eval_steps_per_second": 1.807, | |
| "step": 7734 | |
| }, | |
| { | |
| "epoch": 1.03, | |
| "learning_rate": 8.965606413240239e-05, | |
| "loss": 4.1903, | |
| "step": 8000 | |
| }, | |
| { | |
| "epoch": 1.1, | |
| "learning_rate": 8.900956814067753e-05, | |
| "loss": 4.1066, | |
| "step": 8500 | |
| }, | |
| { | |
| "epoch": 1.16, | |
| "learning_rate": 8.836307214895269e-05, | |
| "loss": 4.0331, | |
| "step": 9000 | |
| }, | |
| { | |
| "epoch": 1.23, | |
| "learning_rate": 8.771657615722783e-05, | |
| "loss": 3.9644, | |
| "step": 9500 | |
| }, | |
| { | |
| "epoch": 1.29, | |
| "learning_rate": 8.707008016550297e-05, | |
| "loss": 3.9044, | |
| "step": 10000 | |
| }, | |
| { | |
| "epoch": 1.36, | |
| "learning_rate": 8.642358417377813e-05, | |
| "loss": 3.851, | |
| "step": 10500 | |
| }, | |
| { | |
| "epoch": 1.42, | |
| "learning_rate": 8.577708818205328e-05, | |
| "loss": 3.8044, | |
| "step": 11000 | |
| }, | |
| { | |
| "epoch": 1.49, | |
| "learning_rate": 8.513059219032842e-05, | |
| "loss": 3.761, | |
| "step": 11500 | |
| }, | |
| { | |
| "epoch": 1.55, | |
| "learning_rate": 8.448409619860357e-05, | |
| "loss": 3.7197, | |
| "step": 12000 | |
| }, | |
| { | |
| "epoch": 1.62, | |
| "learning_rate": 8.383760020687872e-05, | |
| "loss": 3.6803, | |
| "step": 12500 | |
| }, | |
| { | |
| "epoch": 1.68, | |
| "learning_rate": 8.319110421515386e-05, | |
| "loss": 3.6381, | |
| "step": 13000 | |
| }, | |
| { | |
| "epoch": 1.75, | |
| "learning_rate": 8.254460822342902e-05, | |
| "loss": 3.6119, | |
| "step": 13500 | |
| }, | |
| { | |
| "epoch": 1.81, | |
| "learning_rate": 8.189811223170418e-05, | |
| "loss": 3.577, | |
| "step": 14000 | |
| }, | |
| { | |
| "epoch": 1.87, | |
| "learning_rate": 8.12516162399793e-05, | |
| "loss": 3.5515, | |
| "step": 14500 | |
| }, | |
| { | |
| "epoch": 1.94, | |
| "learning_rate": 8.060512024825446e-05, | |
| "loss": 3.5228, | |
| "step": 15000 | |
| }, | |
| { | |
| "epoch": 2.0, | |
| "eval_loss": 3.361564874649048, | |
| "eval_runtime": 476.9803, | |
| "eval_samples_per_second": 691.813, | |
| "eval_steps_per_second": 1.803, | |
| "step": 15468 | |
| }, | |
| { | |
| "epoch": 2.0, | |
| "learning_rate": 7.995862425652962e-05, | |
| "loss": 3.4922, | |
| "step": 15500 | |
| }, | |
| { | |
| "epoch": 2.07, | |
| "learning_rate": 7.931212826480476e-05, | |
| "loss": 3.4583, | |
| "step": 16000 | |
| }, | |
| { | |
| "epoch": 2.13, | |
| "learning_rate": 7.866563227307991e-05, | |
| "loss": 3.4363, | |
| "step": 16500 | |
| }, | |
| { | |
| "epoch": 2.2, | |
| "learning_rate": 7.801913628135506e-05, | |
| "loss": 3.4124, | |
| "step": 17000 | |
| }, | |
| { | |
| "epoch": 2.26, | |
| "learning_rate": 7.73726402896302e-05, | |
| "loss": 3.3911, | |
| "step": 17500 | |
| }, | |
| { | |
| "epoch": 2.33, | |
| "learning_rate": 7.672614429790535e-05, | |
| "loss": 3.3731, | |
| "step": 18000 | |
| }, | |
| { | |
| "epoch": 2.39, | |
| "learning_rate": 7.607964830618051e-05, | |
| "loss": 3.3487, | |
| "step": 18500 | |
| }, | |
| { | |
| "epoch": 2.46, | |
| "learning_rate": 7.543315231445565e-05, | |
| "loss": 3.333, | |
| "step": 19000 | |
| }, | |
| { | |
| "epoch": 2.52, | |
| "learning_rate": 7.478665632273081e-05, | |
| "loss": 3.3162, | |
| "step": 19500 | |
| }, | |
| { | |
| "epoch": 2.59, | |
| "learning_rate": 7.414016033100595e-05, | |
| "loss": 3.2947, | |
| "step": 20000 | |
| }, | |
| { | |
| "epoch": 2.65, | |
| "learning_rate": 7.349366433928109e-05, | |
| "loss": 3.2783, | |
| "step": 20500 | |
| }, | |
| { | |
| "epoch": 2.72, | |
| "learning_rate": 7.284716834755625e-05, | |
| "loss": 3.2641, | |
| "step": 21000 | |
| }, | |
| { | |
| "epoch": 2.78, | |
| "learning_rate": 7.22006723558314e-05, | |
| "loss": 3.2446, | |
| "step": 21500 | |
| }, | |
| { | |
| "epoch": 2.84, | |
| "learning_rate": 7.155417636410655e-05, | |
| "loss": 3.2326, | |
| "step": 22000 | |
| }, | |
| { | |
| "epoch": 2.91, | |
| "learning_rate": 7.090768037238169e-05, | |
| "loss": 3.2149, | |
| "step": 22500 | |
| }, | |
| { | |
| "epoch": 2.97, | |
| "learning_rate": 7.026118438065684e-05, | |
| "loss": 3.2034, | |
| "step": 23000 | |
| }, | |
| { | |
| "epoch": 3.0, | |
| "eval_loss": 3.070152997970581, | |
| "eval_runtime": 476.4651, | |
| "eval_samples_per_second": 692.561, | |
| "eval_steps_per_second": 1.805, | |
| "step": 23202 | |
| }, | |
| { | |
| "epoch": 3.04, | |
| "learning_rate": 6.961468838893199e-05, | |
| "loss": 3.1826, | |
| "step": 23500 | |
| }, | |
| { | |
| "epoch": 3.1, | |
| "learning_rate": 6.896819239720714e-05, | |
| "loss": 3.1653, | |
| "step": 24000 | |
| }, | |
| { | |
| "epoch": 3.17, | |
| "learning_rate": 6.83216964054823e-05, | |
| "loss": 3.1519, | |
| "step": 24500 | |
| }, | |
| { | |
| "epoch": 3.23, | |
| "learning_rate": 6.767520041375744e-05, | |
| "loss": 3.1387, | |
| "step": 25000 | |
| }, | |
| { | |
| "epoch": 3.3, | |
| "learning_rate": 6.702870442203258e-05, | |
| "loss": 3.1317, | |
| "step": 25500 | |
| }, | |
| { | |
| "epoch": 3.36, | |
| "learning_rate": 6.638220843030774e-05, | |
| "loss": 3.1206, | |
| "step": 26000 | |
| }, | |
| { | |
| "epoch": 3.43, | |
| "learning_rate": 6.573571243858288e-05, | |
| "loss": 3.1074, | |
| "step": 26500 | |
| }, | |
| { | |
| "epoch": 3.49, | |
| "learning_rate": 6.508921644685804e-05, | |
| "loss": 3.0959, | |
| "step": 27000 | |
| }, | |
| { | |
| "epoch": 3.56, | |
| "learning_rate": 6.444272045513318e-05, | |
| "loss": 3.0846, | |
| "step": 27500 | |
| }, | |
| { | |
| "epoch": 3.62, | |
| "learning_rate": 6.379622446340833e-05, | |
| "loss": 3.0691, | |
| "step": 28000 | |
| }, | |
| { | |
| "epoch": 3.69, | |
| "learning_rate": 6.314972847168348e-05, | |
| "loss": 3.0621, | |
| "step": 28500 | |
| }, | |
| { | |
| "epoch": 3.75, | |
| "learning_rate": 6.250323247995863e-05, | |
| "loss": 3.0496, | |
| "step": 29000 | |
| }, | |
| { | |
| "epoch": 3.81, | |
| "learning_rate": 6.185673648823377e-05, | |
| "loss": 3.0406, | |
| "step": 29500 | |
| }, | |
| { | |
| "epoch": 3.88, | |
| "learning_rate": 6.121024049650892e-05, | |
| "loss": 3.0332, | |
| "step": 30000 | |
| }, | |
| { | |
| "epoch": 3.94, | |
| "learning_rate": 6.056374450478407e-05, | |
| "loss": 3.0235, | |
| "step": 30500 | |
| }, | |
| { | |
| "epoch": 4.0, | |
| "eval_loss": 2.890948534011841, | |
| "eval_runtime": 475.9375, | |
| "eval_samples_per_second": 693.328, | |
| "eval_steps_per_second": 1.807, | |
| "step": 30936 | |
| }, | |
| { | |
| "epoch": 4.01, | |
| "learning_rate": 5.991724851305922e-05, | |
| "loss": 3.0144, | |
| "step": 31000 | |
| }, | |
| { | |
| "epoch": 4.07, | |
| "learning_rate": 5.927075252133437e-05, | |
| "loss": 2.9923, | |
| "step": 31500 | |
| }, | |
| { | |
| "epoch": 4.14, | |
| "learning_rate": 5.8624256529609526e-05, | |
| "loss": 2.9902, | |
| "step": 32000 | |
| }, | |
| { | |
| "epoch": 4.2, | |
| "learning_rate": 5.7977760537884675e-05, | |
| "loss": 2.9746, | |
| "step": 32500 | |
| }, | |
| { | |
| "epoch": 4.27, | |
| "learning_rate": 5.733126454615981e-05, | |
| "loss": 2.9708, | |
| "step": 33000 | |
| }, | |
| { | |
| "epoch": 4.33, | |
| "learning_rate": 5.6684768554434966e-05, | |
| "loss": 2.9586, | |
| "step": 33500 | |
| }, | |
| { | |
| "epoch": 4.4, | |
| "learning_rate": 5.6038272562710115e-05, | |
| "loss": 2.9556, | |
| "step": 34000 | |
| }, | |
| { | |
| "epoch": 4.46, | |
| "learning_rate": 5.5391776570985264e-05, | |
| "loss": 2.9471, | |
| "step": 34500 | |
| }, | |
| { | |
| "epoch": 4.53, | |
| "learning_rate": 5.474528057926042e-05, | |
| "loss": 2.941, | |
| "step": 35000 | |
| }, | |
| { | |
| "epoch": 4.59, | |
| "learning_rate": 5.4098784587535555e-05, | |
| "loss": 2.9353, | |
| "step": 35500 | |
| }, | |
| { | |
| "epoch": 4.65, | |
| "learning_rate": 5.3452288595810704e-05, | |
| "loss": 2.9263, | |
| "step": 36000 | |
| }, | |
| { | |
| "epoch": 4.72, | |
| "learning_rate": 5.280579260408586e-05, | |
| "loss": 2.9193, | |
| "step": 36500 | |
| }, | |
| { | |
| "epoch": 4.78, | |
| "learning_rate": 5.215929661236101e-05, | |
| "loss": 2.9101, | |
| "step": 37000 | |
| }, | |
| { | |
| "epoch": 4.85, | |
| "learning_rate": 5.151280062063616e-05, | |
| "loss": 2.9071, | |
| "step": 37500 | |
| }, | |
| { | |
| "epoch": 4.91, | |
| "learning_rate": 5.08663046289113e-05, | |
| "loss": 2.9019, | |
| "step": 38000 | |
| }, | |
| { | |
| "epoch": 4.98, | |
| "learning_rate": 5.021980863718645e-05, | |
| "loss": 2.89, | |
| "step": 38500 | |
| }, | |
| { | |
| "epoch": 5.0, | |
| "eval_loss": 2.768686294555664, | |
| "eval_runtime": 485.6271, | |
| "eval_samples_per_second": 679.495, | |
| "eval_steps_per_second": 1.771, | |
| "step": 38670 | |
| }, | |
| { | |
| "epoch": 5.04, | |
| "learning_rate": 4.95733126454616e-05, | |
| "loss": 2.8791, | |
| "step": 39000 | |
| }, | |
| { | |
| "epoch": 5.11, | |
| "learning_rate": 4.892681665373675e-05, | |
| "loss": 2.87, | |
| "step": 39500 | |
| }, | |
| { | |
| "epoch": 5.17, | |
| "learning_rate": 4.8280320662011896e-05, | |
| "loss": 2.8664, | |
| "step": 40000 | |
| }, | |
| { | |
| "epoch": 5.24, | |
| "learning_rate": 4.7633824670287045e-05, | |
| "loss": 2.859, | |
| "step": 40500 | |
| }, | |
| { | |
| "epoch": 5.3, | |
| "learning_rate": 4.6987328678562193e-05, | |
| "loss": 2.8562, | |
| "step": 41000 | |
| }, | |
| { | |
| "epoch": 5.37, | |
| "learning_rate": 4.634083268683734e-05, | |
| "loss": 2.8489, | |
| "step": 41500 | |
| }, | |
| { | |
| "epoch": 5.43, | |
| "learning_rate": 4.569433669511249e-05, | |
| "loss": 2.846, | |
| "step": 42000 | |
| }, | |
| { | |
| "epoch": 5.5, | |
| "learning_rate": 4.504784070338764e-05, | |
| "loss": 2.8342, | |
| "step": 42500 | |
| }, | |
| { | |
| "epoch": 5.56, | |
| "learning_rate": 4.440134471166279e-05, | |
| "loss": 2.8313, | |
| "step": 43000 | |
| }, | |
| { | |
| "epoch": 5.62, | |
| "learning_rate": 4.375484871993794e-05, | |
| "loss": 2.8217, | |
| "step": 43500 | |
| }, | |
| { | |
| "epoch": 5.69, | |
| "learning_rate": 4.310835272821309e-05, | |
| "loss": 2.8197, | |
| "step": 44000 | |
| }, | |
| { | |
| "epoch": 5.75, | |
| "learning_rate": 4.2461856736488236e-05, | |
| "loss": 2.8158, | |
| "step": 44500 | |
| }, | |
| { | |
| "epoch": 5.82, | |
| "learning_rate": 4.1815360744763385e-05, | |
| "loss": 2.8094, | |
| "step": 45000 | |
| }, | |
| { | |
| "epoch": 5.88, | |
| "learning_rate": 4.1168864753038534e-05, | |
| "loss": 2.8061, | |
| "step": 45500 | |
| }, | |
| { | |
| "epoch": 5.95, | |
| "learning_rate": 4.052236876131368e-05, | |
| "loss": 2.8032, | |
| "step": 46000 | |
| }, | |
| { | |
| "epoch": 6.0, | |
| "eval_loss": 2.681367874145508, | |
| "eval_runtime": 476.3627, | |
| "eval_samples_per_second": 692.71, | |
| "eval_steps_per_second": 1.805, | |
| "step": 46404 | |
| }, | |
| { | |
| "epoch": 6.01, | |
| "learning_rate": 3.987587276958883e-05, | |
| "loss": 2.794, | |
| "step": 46500 | |
| }, | |
| { | |
| "epoch": 6.08, | |
| "learning_rate": 3.922937677786398e-05, | |
| "loss": 2.7858, | |
| "step": 47000 | |
| }, | |
| { | |
| "epoch": 6.14, | |
| "learning_rate": 3.858288078613912e-05, | |
| "loss": 2.7765, | |
| "step": 47500 | |
| }, | |
| { | |
| "epoch": 6.21, | |
| "learning_rate": 3.793638479441428e-05, | |
| "loss": 2.7752, | |
| "step": 48000 | |
| }, | |
| { | |
| "epoch": 6.27, | |
| "learning_rate": 3.728988880268943e-05, | |
| "loss": 2.7725, | |
| "step": 48500 | |
| }, | |
| { | |
| "epoch": 6.34, | |
| "learning_rate": 3.664339281096457e-05, | |
| "loss": 2.7696, | |
| "step": 49000 | |
| }, | |
| { | |
| "epoch": 6.4, | |
| "learning_rate": 3.5996896819239726e-05, | |
| "loss": 2.7665, | |
| "step": 49500 | |
| }, | |
| { | |
| "epoch": 6.46, | |
| "learning_rate": 3.5350400827514875e-05, | |
| "loss": 2.7622, | |
| "step": 50000 | |
| }, | |
| { | |
| "epoch": 6.53, | |
| "learning_rate": 3.470390483579002e-05, | |
| "loss": 2.7566, | |
| "step": 50500 | |
| }, | |
| { | |
| "epoch": 6.59, | |
| "learning_rate": 3.4057408844065166e-05, | |
| "loss": 2.7543, | |
| "step": 51000 | |
| }, | |
| { | |
| "epoch": 6.66, | |
| "learning_rate": 3.3410912852340315e-05, | |
| "loss": 2.7451, | |
| "step": 51500 | |
| }, | |
| { | |
| "epoch": 6.72, | |
| "learning_rate": 3.2764416860615464e-05, | |
| "loss": 2.7392, | |
| "step": 52000 | |
| }, | |
| { | |
| "epoch": 6.79, | |
| "learning_rate": 3.211792086889061e-05, | |
| "loss": 2.7384, | |
| "step": 52500 | |
| }, | |
| { | |
| "epoch": 6.85, | |
| "learning_rate": 3.147142487716576e-05, | |
| "loss": 2.7341, | |
| "step": 53000 | |
| }, | |
| { | |
| "epoch": 6.92, | |
| "learning_rate": 3.082492888544091e-05, | |
| "loss": 2.7317, | |
| "step": 53500 | |
| }, | |
| { | |
| "epoch": 6.98, | |
| "learning_rate": 3.0178432893716057e-05, | |
| "loss": 2.7308, | |
| "step": 54000 | |
| }, | |
| { | |
| "epoch": 7.0, | |
| "eval_loss": 2.6155083179473877, | |
| "eval_runtime": 476.4568, | |
| "eval_samples_per_second": 692.573, | |
| "eval_steps_per_second": 1.805, | |
| "step": 54138 | |
| }, | |
| { | |
| "epoch": 7.05, | |
| "learning_rate": 2.953193690199121e-05, | |
| "loss": 2.7178, | |
| "step": 54500 | |
| }, | |
| { | |
| "epoch": 7.11, | |
| "learning_rate": 2.8885440910266358e-05, | |
| "loss": 2.712, | |
| "step": 55000 | |
| }, | |
| { | |
| "epoch": 7.18, | |
| "learning_rate": 2.8238944918541504e-05, | |
| "loss": 2.7152, | |
| "step": 55500 | |
| }, | |
| { | |
| "epoch": 7.24, | |
| "learning_rate": 2.7592448926816656e-05, | |
| "loss": 2.7076, | |
| "step": 56000 | |
| }, | |
| { | |
| "epoch": 7.31, | |
| "learning_rate": 2.6945952935091805e-05, | |
| "loss": 2.7048, | |
| "step": 56500 | |
| }, | |
| { | |
| "epoch": 7.37, | |
| "learning_rate": 2.629945694336695e-05, | |
| "loss": 2.7038, | |
| "step": 57000 | |
| }, | |
| { | |
| "epoch": 7.43, | |
| "learning_rate": 2.5652960951642103e-05, | |
| "loss": 2.6975, | |
| "step": 57500 | |
| }, | |
| { | |
| "epoch": 7.5, | |
| "learning_rate": 2.500646495991725e-05, | |
| "loss": 2.696, | |
| "step": 58000 | |
| }, | |
| { | |
| "epoch": 7.56, | |
| "learning_rate": 2.4359968968192397e-05, | |
| "loss": 2.6941, | |
| "step": 58500 | |
| }, | |
| { | |
| "epoch": 7.63, | |
| "learning_rate": 2.3713472976467546e-05, | |
| "loss": 2.6884, | |
| "step": 59000 | |
| }, | |
| { | |
| "epoch": 7.69, | |
| "learning_rate": 2.3066976984742695e-05, | |
| "loss": 2.6844, | |
| "step": 59500 | |
| }, | |
| { | |
| "epoch": 7.76, | |
| "learning_rate": 2.2420480993017844e-05, | |
| "loss": 2.6791, | |
| "step": 60000 | |
| }, | |
| { | |
| "epoch": 7.82, | |
| "learning_rate": 2.1773985001292993e-05, | |
| "loss": 2.6759, | |
| "step": 60500 | |
| }, | |
| { | |
| "epoch": 7.89, | |
| "learning_rate": 2.1127489009568142e-05, | |
| "loss": 2.6779, | |
| "step": 61000 | |
| }, | |
| { | |
| "epoch": 7.95, | |
| "learning_rate": 2.0480993017843288e-05, | |
| "loss": 2.6749, | |
| "step": 61500 | |
| }, | |
| { | |
| "epoch": 8.0, | |
| "eval_loss": 2.561079502105713, | |
| "eval_runtime": 476.3937, | |
| "eval_samples_per_second": 692.665, | |
| "eval_steps_per_second": 1.805, | |
| "step": 61872 | |
| }, | |
| { | |
| "epoch": 8.02, | |
| "learning_rate": 1.983449702611844e-05, | |
| "loss": 2.6705, | |
| "step": 62000 | |
| }, | |
| { | |
| "epoch": 8.08, | |
| "learning_rate": 1.918800103439359e-05, | |
| "loss": 2.6613, | |
| "step": 62500 | |
| }, | |
| { | |
| "epoch": 8.15, | |
| "learning_rate": 1.8541505042668735e-05, | |
| "loss": 2.6649, | |
| "step": 63000 | |
| }, | |
| { | |
| "epoch": 8.21, | |
| "learning_rate": 1.7895009050943884e-05, | |
| "loss": 2.6597, | |
| "step": 63500 | |
| }, | |
| { | |
| "epoch": 8.28, | |
| "learning_rate": 1.7248513059219036e-05, | |
| "loss": 2.6568, | |
| "step": 64000 | |
| }, | |
| { | |
| "epoch": 8.34, | |
| "learning_rate": 1.660201706749418e-05, | |
| "loss": 2.6551, | |
| "step": 64500 | |
| }, | |
| { | |
| "epoch": 8.4, | |
| "learning_rate": 1.595552107576933e-05, | |
| "loss": 2.6518, | |
| "step": 65000 | |
| }, | |
| { | |
| "epoch": 8.47, | |
| "learning_rate": 1.530902508404448e-05, | |
| "loss": 2.6501, | |
| "step": 65500 | |
| }, | |
| { | |
| "epoch": 8.53, | |
| "learning_rate": 1.466252909231963e-05, | |
| "loss": 2.6469, | |
| "step": 66000 | |
| }, | |
| { | |
| "epoch": 8.6, | |
| "learning_rate": 1.4016033100594778e-05, | |
| "loss": 2.6465, | |
| "step": 66500 | |
| }, | |
| { | |
| "epoch": 8.66, | |
| "learning_rate": 1.3369537108869925e-05, | |
| "loss": 2.6455, | |
| "step": 67000 | |
| }, | |
| { | |
| "epoch": 8.73, | |
| "learning_rate": 1.2723041117145074e-05, | |
| "loss": 2.6398, | |
| "step": 67500 | |
| }, | |
| { | |
| "epoch": 8.79, | |
| "learning_rate": 1.2076545125420223e-05, | |
| "loss": 2.6436, | |
| "step": 68000 | |
| }, | |
| { | |
| "epoch": 8.86, | |
| "learning_rate": 1.1430049133695372e-05, | |
| "loss": 2.6395, | |
| "step": 68500 | |
| }, | |
| { | |
| "epoch": 8.92, | |
| "learning_rate": 1.078355314197052e-05, | |
| "loss": 2.634, | |
| "step": 69000 | |
| }, | |
| { | |
| "epoch": 8.99, | |
| "learning_rate": 1.0137057150245668e-05, | |
| "loss": 2.6359, | |
| "step": 69500 | |
| }, | |
| { | |
| "epoch": 9.0, | |
| "eval_loss": 2.5289456844329834, | |
| "eval_runtime": 476.5662, | |
| "eval_samples_per_second": 692.414, | |
| "eval_steps_per_second": 1.805, | |
| "step": 69606 | |
| }, | |
| { | |
| "epoch": 9.05, | |
| "learning_rate": 9.490561158520819e-06, | |
| "loss": 2.6261, | |
| "step": 70000 | |
| }, | |
| { | |
| "epoch": 9.12, | |
| "learning_rate": 8.844065166795966e-06, | |
| "loss": 2.6228, | |
| "step": 70500 | |
| }, | |
| { | |
| "epoch": 9.18, | |
| "learning_rate": 8.197569175071115e-06, | |
| "loss": 2.6303, | |
| "step": 71000 | |
| }, | |
| { | |
| "epoch": 9.24, | |
| "learning_rate": 7.551073183346263e-06, | |
| "loss": 2.622, | |
| "step": 71500 | |
| }, | |
| { | |
| "epoch": 9.31, | |
| "learning_rate": 6.904577191621413e-06, | |
| "loss": 2.6257, | |
| "step": 72000 | |
| }, | |
| { | |
| "epoch": 9.37, | |
| "learning_rate": 6.258081199896561e-06, | |
| "loss": 2.6244, | |
| "step": 72500 | |
| }, | |
| { | |
| "epoch": 9.44, | |
| "learning_rate": 5.611585208171709e-06, | |
| "loss": 2.6238, | |
| "step": 73000 | |
| }, | |
| { | |
| "epoch": 9.5, | |
| "learning_rate": 4.965089216446858e-06, | |
| "loss": 2.6207, | |
| "step": 73500 | |
| }, | |
| { | |
| "epoch": 9.57, | |
| "learning_rate": 4.318593224722007e-06, | |
| "loss": 2.6168, | |
| "step": 74000 | |
| }, | |
| { | |
| "epoch": 9.63, | |
| "learning_rate": 3.6720972329971556e-06, | |
| "loss": 2.619, | |
| "step": 74500 | |
| }, | |
| { | |
| "epoch": 9.7, | |
| "learning_rate": 3.025601241272304e-06, | |
| "loss": 2.6185, | |
| "step": 75000 | |
| }, | |
| { | |
| "epoch": 9.76, | |
| "learning_rate": 2.379105249547453e-06, | |
| "loss": 2.6173, | |
| "step": 75500 | |
| }, | |
| { | |
| "epoch": 9.83, | |
| "learning_rate": 1.7326092578226015e-06, | |
| "loss": 2.6114, | |
| "step": 76000 | |
| }, | |
| { | |
| "epoch": 9.89, | |
| "learning_rate": 1.0861132660977503e-06, | |
| "loss": 2.6091, | |
| "step": 76500 | |
| }, | |
| { | |
| "epoch": 9.96, | |
| "learning_rate": 4.396172743728989e-07, | |
| "loss": 2.6133, | |
| "step": 77000 | |
| }, | |
| { | |
| "epoch": 10.0, | |
| "eval_loss": 2.5088417530059814, | |
| "eval_runtime": 484.8466, | |
| "eval_samples_per_second": 680.588, | |
| "eval_steps_per_second": 1.774, | |
| "step": 77340 | |
| } | |
| ], | |
| "max_steps": 77340, | |
| "num_train_epochs": 10, | |
| "total_flos": 1.9542406291981978e+18, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |