| { | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 6.0, | |
| "global_step": 59298, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 4.957840062059429e-05, | |
| "loss": 3.4611, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 4.915680124118858e-05, | |
| "loss": 3.3849, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 4.8735201861782864e-05, | |
| "loss": 3.3355, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 4.831360248237715e-05, | |
| "loss": 3.3087, | |
| "step": 2000 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 4.7892003102971434e-05, | |
| "loss": 3.2799, | |
| "step": 2500 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 4.747040372356572e-05, | |
| "loss": 3.2638, | |
| "step": 3000 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 4.704880434416001e-05, | |
| "loss": 3.2501, | |
| "step": 3500 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 4.6627204964754295e-05, | |
| "loss": 3.2575, | |
| "step": 4000 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 4.620560558534858e-05, | |
| "loss": 3.25, | |
| "step": 4500 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "learning_rate": 4.5784006205942865e-05, | |
| "loss": 3.214, | |
| "step": 5000 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "learning_rate": 4.536240682653715e-05, | |
| "loss": 3.2253, | |
| "step": 5500 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "learning_rate": 4.494080744713144e-05, | |
| "loss": 3.2102, | |
| "step": 6000 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "learning_rate": 4.4519208067725726e-05, | |
| "loss": 3.1872, | |
| "step": 6500 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "learning_rate": 4.409760868832001e-05, | |
| "loss": 3.1946, | |
| "step": 7000 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "learning_rate": 4.36760093089143e-05, | |
| "loss": 3.1831, | |
| "step": 7500 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "learning_rate": 4.325440992950859e-05, | |
| "loss": 3.1851, | |
| "step": 8000 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "learning_rate": 4.283281055010287e-05, | |
| "loss": 3.1794, | |
| "step": 8500 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "learning_rate": 4.241121117069716e-05, | |
| "loss": 3.1698, | |
| "step": 9000 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "learning_rate": 4.198961179129144e-05, | |
| "loss": 3.16, | |
| "step": 9500 | |
| }, | |
| { | |
| "epoch": 1.01, | |
| "learning_rate": 4.1568012411885734e-05, | |
| "loss": 3.1365, | |
| "step": 10000 | |
| }, | |
| { | |
| "epoch": 1.06, | |
| "learning_rate": 4.114641303248002e-05, | |
| "loss": 3.0765, | |
| "step": 10500 | |
| }, | |
| { | |
| "epoch": 1.11, | |
| "learning_rate": 4.0724813653074304e-05, | |
| "loss": 3.0774, | |
| "step": 11000 | |
| }, | |
| { | |
| "epoch": 1.16, | |
| "learning_rate": 4.0303214273668596e-05, | |
| "loss": 3.0747, | |
| "step": 11500 | |
| }, | |
| { | |
| "epoch": 1.21, | |
| "learning_rate": 3.9881614894262874e-05, | |
| "loss": 3.0726, | |
| "step": 12000 | |
| }, | |
| { | |
| "epoch": 1.26, | |
| "learning_rate": 3.9460015514857165e-05, | |
| "loss": 3.0738, | |
| "step": 12500 | |
| }, | |
| { | |
| "epoch": 1.32, | |
| "learning_rate": 3.903841613545145e-05, | |
| "loss": 3.0799, | |
| "step": 13000 | |
| }, | |
| { | |
| "epoch": 1.37, | |
| "learning_rate": 3.8616816756045735e-05, | |
| "loss": 3.059, | |
| "step": 13500 | |
| }, | |
| { | |
| "epoch": 1.42, | |
| "learning_rate": 3.819521737664003e-05, | |
| "loss": 3.0648, | |
| "step": 14000 | |
| }, | |
| { | |
| "epoch": 1.47, | |
| "learning_rate": 3.777361799723431e-05, | |
| "loss": 3.0643, | |
| "step": 14500 | |
| }, | |
| { | |
| "epoch": 1.52, | |
| "learning_rate": 3.7352018617828597e-05, | |
| "loss": 3.0737, | |
| "step": 15000 | |
| }, | |
| { | |
| "epoch": 1.57, | |
| "learning_rate": 3.693041923842288e-05, | |
| "loss": 3.0672, | |
| "step": 15500 | |
| }, | |
| { | |
| "epoch": 1.62, | |
| "learning_rate": 3.6508819859017166e-05, | |
| "loss": 3.071, | |
| "step": 16000 | |
| }, | |
| { | |
| "epoch": 1.67, | |
| "learning_rate": 3.608722047961146e-05, | |
| "loss": 3.058, | |
| "step": 16500 | |
| }, | |
| { | |
| "epoch": 1.72, | |
| "learning_rate": 3.566562110020574e-05, | |
| "loss": 3.0572, | |
| "step": 17000 | |
| }, | |
| { | |
| "epoch": 1.77, | |
| "learning_rate": 3.524402172080003e-05, | |
| "loss": 3.063, | |
| "step": 17500 | |
| }, | |
| { | |
| "epoch": 1.82, | |
| "learning_rate": 3.482242234139432e-05, | |
| "loss": 3.0507, | |
| "step": 18000 | |
| }, | |
| { | |
| "epoch": 1.87, | |
| "learning_rate": 3.44008229619886e-05, | |
| "loss": 3.0435, | |
| "step": 18500 | |
| }, | |
| { | |
| "epoch": 1.92, | |
| "learning_rate": 3.397922358258289e-05, | |
| "loss": 3.0538, | |
| "step": 19000 | |
| }, | |
| { | |
| "epoch": 1.97, | |
| "learning_rate": 3.3557624203177174e-05, | |
| "loss": 3.0411, | |
| "step": 19500 | |
| }, | |
| { | |
| "epoch": 2.02, | |
| "learning_rate": 3.313602482377146e-05, | |
| "loss": 3.0106, | |
| "step": 20000 | |
| }, | |
| { | |
| "epoch": 2.07, | |
| "learning_rate": 3.271442544436575e-05, | |
| "loss": 2.9677, | |
| "step": 20500 | |
| }, | |
| { | |
| "epoch": 2.12, | |
| "learning_rate": 3.2292826064960036e-05, | |
| "loss": 2.9598, | |
| "step": 21000 | |
| }, | |
| { | |
| "epoch": 2.18, | |
| "learning_rate": 3.187122668555432e-05, | |
| "loss": 2.9693, | |
| "step": 21500 | |
| }, | |
| { | |
| "epoch": 2.23, | |
| "learning_rate": 3.1449627306148605e-05, | |
| "loss": 2.9833, | |
| "step": 22000 | |
| }, | |
| { | |
| "epoch": 2.28, | |
| "learning_rate": 3.102802792674289e-05, | |
| "loss": 2.9675, | |
| "step": 22500 | |
| }, | |
| { | |
| "epoch": 2.33, | |
| "learning_rate": 3.060642854733718e-05, | |
| "loss": 2.9713, | |
| "step": 23000 | |
| }, | |
| { | |
| "epoch": 2.38, | |
| "learning_rate": 3.0184829167931467e-05, | |
| "loss": 2.9768, | |
| "step": 23500 | |
| }, | |
| { | |
| "epoch": 2.43, | |
| "learning_rate": 2.9763229788525755e-05, | |
| "loss": 2.9765, | |
| "step": 24000 | |
| }, | |
| { | |
| "epoch": 2.48, | |
| "learning_rate": 2.934163040912004e-05, | |
| "loss": 2.9698, | |
| "step": 24500 | |
| }, | |
| { | |
| "epoch": 2.53, | |
| "learning_rate": 2.892003102971432e-05, | |
| "loss": 2.9934, | |
| "step": 25000 | |
| }, | |
| { | |
| "epoch": 2.58, | |
| "learning_rate": 2.849843165030861e-05, | |
| "loss": 2.9721, | |
| "step": 25500 | |
| }, | |
| { | |
| "epoch": 2.63, | |
| "learning_rate": 2.8076832270902898e-05, | |
| "loss": 2.9766, | |
| "step": 26000 | |
| }, | |
| { | |
| "epoch": 2.68, | |
| "learning_rate": 2.7655232891497186e-05, | |
| "loss": 2.986, | |
| "step": 26500 | |
| }, | |
| { | |
| "epoch": 2.73, | |
| "learning_rate": 2.723363351209147e-05, | |
| "loss": 2.9856, | |
| "step": 27000 | |
| }, | |
| { | |
| "epoch": 2.78, | |
| "learning_rate": 2.681203413268576e-05, | |
| "loss": 2.9851, | |
| "step": 27500 | |
| }, | |
| { | |
| "epoch": 2.83, | |
| "learning_rate": 2.6390434753280048e-05, | |
| "loss": 2.9722, | |
| "step": 28000 | |
| }, | |
| { | |
| "epoch": 2.88, | |
| "learning_rate": 2.596883537387433e-05, | |
| "loss": 2.9713, | |
| "step": 28500 | |
| }, | |
| { | |
| "epoch": 2.93, | |
| "learning_rate": 2.5547235994468614e-05, | |
| "loss": 2.9698, | |
| "step": 29000 | |
| }, | |
| { | |
| "epoch": 2.98, | |
| "learning_rate": 2.5125636615062902e-05, | |
| "loss": 2.9742, | |
| "step": 29500 | |
| }, | |
| { | |
| "epoch": 3.04, | |
| "learning_rate": 2.470403723565719e-05, | |
| "loss": 2.9437, | |
| "step": 30000 | |
| }, | |
| { | |
| "epoch": 3.09, | |
| "learning_rate": 2.428243785625148e-05, | |
| "loss": 2.9012, | |
| "step": 30500 | |
| }, | |
| { | |
| "epoch": 3.14, | |
| "learning_rate": 2.3860838476845764e-05, | |
| "loss": 2.903, | |
| "step": 31000 | |
| }, | |
| { | |
| "epoch": 3.19, | |
| "learning_rate": 2.343923909744005e-05, | |
| "loss": 2.9115, | |
| "step": 31500 | |
| }, | |
| { | |
| "epoch": 3.24, | |
| "learning_rate": 2.3017639718034337e-05, | |
| "loss": 2.9079, | |
| "step": 32000 | |
| }, | |
| { | |
| "epoch": 3.29, | |
| "learning_rate": 2.2596040338628622e-05, | |
| "loss": 2.9209, | |
| "step": 32500 | |
| }, | |
| { | |
| "epoch": 3.34, | |
| "learning_rate": 2.217444095922291e-05, | |
| "loss": 2.9072, | |
| "step": 33000 | |
| }, | |
| { | |
| "epoch": 3.39, | |
| "learning_rate": 2.1752841579817195e-05, | |
| "loss": 2.9141, | |
| "step": 33500 | |
| }, | |
| { | |
| "epoch": 3.44, | |
| "learning_rate": 2.1331242200411483e-05, | |
| "loss": 2.9196, | |
| "step": 34000 | |
| }, | |
| { | |
| "epoch": 3.49, | |
| "learning_rate": 2.0909642821005768e-05, | |
| "loss": 2.9225, | |
| "step": 34500 | |
| }, | |
| { | |
| "epoch": 3.54, | |
| "learning_rate": 2.0488043441600056e-05, | |
| "loss": 2.9171, | |
| "step": 35000 | |
| }, | |
| { | |
| "epoch": 3.59, | |
| "learning_rate": 2.006644406219434e-05, | |
| "loss": 2.9171, | |
| "step": 35500 | |
| }, | |
| { | |
| "epoch": 3.64, | |
| "learning_rate": 1.9644844682788626e-05, | |
| "loss": 2.9088, | |
| "step": 36000 | |
| }, | |
| { | |
| "epoch": 3.69, | |
| "learning_rate": 1.9223245303382915e-05, | |
| "loss": 2.9169, | |
| "step": 36500 | |
| }, | |
| { | |
| "epoch": 3.74, | |
| "learning_rate": 1.8801645923977203e-05, | |
| "loss": 2.9197, | |
| "step": 37000 | |
| }, | |
| { | |
| "epoch": 3.79, | |
| "learning_rate": 1.8380046544571484e-05, | |
| "loss": 2.9118, | |
| "step": 37500 | |
| }, | |
| { | |
| "epoch": 3.84, | |
| "learning_rate": 1.7958447165165773e-05, | |
| "loss": 2.9153, | |
| "step": 38000 | |
| }, | |
| { | |
| "epoch": 3.9, | |
| "learning_rate": 1.753684778576006e-05, | |
| "loss": 2.9104, | |
| "step": 38500 | |
| }, | |
| { | |
| "epoch": 3.95, | |
| "learning_rate": 1.711524840635435e-05, | |
| "loss": 2.9136, | |
| "step": 39000 | |
| }, | |
| { | |
| "epoch": 4.0, | |
| "learning_rate": 1.669364902694863e-05, | |
| "loss": 2.9218, | |
| "step": 39500 | |
| }, | |
| { | |
| "epoch": 4.05, | |
| "learning_rate": 1.627204964754292e-05, | |
| "loss": 2.8708, | |
| "step": 40000 | |
| }, | |
| { | |
| "epoch": 4.1, | |
| "learning_rate": 1.5850450268137207e-05, | |
| "loss": 2.865, | |
| "step": 40500 | |
| }, | |
| { | |
| "epoch": 4.15, | |
| "learning_rate": 1.5428850888731492e-05, | |
| "loss": 2.8641, | |
| "step": 41000 | |
| }, | |
| { | |
| "epoch": 4.2, | |
| "learning_rate": 1.5007251509325779e-05, | |
| "loss": 2.8758, | |
| "step": 41500 | |
| }, | |
| { | |
| "epoch": 4.25, | |
| "learning_rate": 1.4585652129920065e-05, | |
| "loss": 2.866, | |
| "step": 42000 | |
| }, | |
| { | |
| "epoch": 4.3, | |
| "learning_rate": 1.416405275051435e-05, | |
| "loss": 2.8814, | |
| "step": 42500 | |
| }, | |
| { | |
| "epoch": 4.35, | |
| "learning_rate": 1.3742453371108638e-05, | |
| "loss": 2.8632, | |
| "step": 43000 | |
| }, | |
| { | |
| "epoch": 4.4, | |
| "learning_rate": 1.3320853991702925e-05, | |
| "loss": 2.8634, | |
| "step": 43500 | |
| }, | |
| { | |
| "epoch": 4.45, | |
| "learning_rate": 1.2899254612297213e-05, | |
| "loss": 2.8605, | |
| "step": 44000 | |
| }, | |
| { | |
| "epoch": 4.5, | |
| "learning_rate": 1.2477655232891498e-05, | |
| "loss": 2.8665, | |
| "step": 44500 | |
| }, | |
| { | |
| "epoch": 4.55, | |
| "learning_rate": 1.2056055853485785e-05, | |
| "loss": 2.8697, | |
| "step": 45000 | |
| }, | |
| { | |
| "epoch": 4.6, | |
| "learning_rate": 1.163445647408007e-05, | |
| "loss": 2.8756, | |
| "step": 45500 | |
| }, | |
| { | |
| "epoch": 4.65, | |
| "learning_rate": 1.1212857094674358e-05, | |
| "loss": 2.8565, | |
| "step": 46000 | |
| }, | |
| { | |
| "epoch": 4.71, | |
| "learning_rate": 1.0791257715268643e-05, | |
| "loss": 2.8682, | |
| "step": 46500 | |
| }, | |
| { | |
| "epoch": 4.76, | |
| "learning_rate": 1.0369658335862931e-05, | |
| "loss": 2.8639, | |
| "step": 47000 | |
| }, | |
| { | |
| "epoch": 4.81, | |
| "learning_rate": 9.948058956457216e-06, | |
| "loss": 2.8667, | |
| "step": 47500 | |
| }, | |
| { | |
| "epoch": 4.86, | |
| "learning_rate": 9.526459577051503e-06, | |
| "loss": 2.8715, | |
| "step": 48000 | |
| }, | |
| { | |
| "epoch": 4.91, | |
| "learning_rate": 9.104860197645789e-06, | |
| "loss": 2.8704, | |
| "step": 48500 | |
| }, | |
| { | |
| "epoch": 4.96, | |
| "learning_rate": 8.683260818240076e-06, | |
| "loss": 2.8771, | |
| "step": 49000 | |
| }, | |
| { | |
| "epoch": 5.01, | |
| "learning_rate": 8.261661438834362e-06, | |
| "loss": 2.8545, | |
| "step": 49500 | |
| }, | |
| { | |
| "epoch": 5.06, | |
| "learning_rate": 7.840062059428649e-06, | |
| "loss": 2.8237, | |
| "step": 50000 | |
| }, | |
| { | |
| "epoch": 5.11, | |
| "learning_rate": 7.418462680022935e-06, | |
| "loss": 2.832, | |
| "step": 50500 | |
| }, | |
| { | |
| "epoch": 5.16, | |
| "learning_rate": 6.996863300617222e-06, | |
| "loss": 2.8298, | |
| "step": 51000 | |
| }, | |
| { | |
| "epoch": 5.21, | |
| "learning_rate": 6.575263921211508e-06, | |
| "loss": 2.8336, | |
| "step": 51500 | |
| }, | |
| { | |
| "epoch": 5.26, | |
| "learning_rate": 6.153664541805794e-06, | |
| "loss": 2.8441, | |
| "step": 52000 | |
| }, | |
| { | |
| "epoch": 5.31, | |
| "learning_rate": 5.732065162400081e-06, | |
| "loss": 2.8493, | |
| "step": 52500 | |
| }, | |
| { | |
| "epoch": 5.36, | |
| "learning_rate": 5.3104657829943675e-06, | |
| "loss": 2.8324, | |
| "step": 53000 | |
| }, | |
| { | |
| "epoch": 5.41, | |
| "learning_rate": 4.888866403588654e-06, | |
| "loss": 2.8422, | |
| "step": 53500 | |
| }, | |
| { | |
| "epoch": 5.46, | |
| "learning_rate": 4.467267024182941e-06, | |
| "loss": 2.8419, | |
| "step": 54000 | |
| }, | |
| { | |
| "epoch": 5.51, | |
| "learning_rate": 4.045667644777227e-06, | |
| "loss": 2.8283, | |
| "step": 54500 | |
| }, | |
| { | |
| "epoch": 5.57, | |
| "learning_rate": 3.6240682653715134e-06, | |
| "loss": 2.8451, | |
| "step": 55000 | |
| }, | |
| { | |
| "epoch": 5.62, | |
| "learning_rate": 3.2024688859658e-06, | |
| "loss": 2.8294, | |
| "step": 55500 | |
| }, | |
| { | |
| "epoch": 5.67, | |
| "learning_rate": 2.7808695065600866e-06, | |
| "loss": 2.8356, | |
| "step": 56000 | |
| }, | |
| { | |
| "epoch": 5.72, | |
| "learning_rate": 2.359270127154373e-06, | |
| "loss": 2.8375, | |
| "step": 56500 | |
| }, | |
| { | |
| "epoch": 5.77, | |
| "learning_rate": 1.9376707477486594e-06, | |
| "loss": 2.8328, | |
| "step": 57000 | |
| }, | |
| { | |
| "epoch": 5.82, | |
| "learning_rate": 1.516071368342946e-06, | |
| "loss": 2.8397, | |
| "step": 57500 | |
| }, | |
| { | |
| "epoch": 5.87, | |
| "learning_rate": 1.0944719889372323e-06, | |
| "loss": 2.8305, | |
| "step": 58000 | |
| }, | |
| { | |
| "epoch": 5.92, | |
| "learning_rate": 6.728726095315188e-07, | |
| "loss": 2.8395, | |
| "step": 58500 | |
| }, | |
| { | |
| "epoch": 5.97, | |
| "learning_rate": 2.5127323012580527e-07, | |
| "loss": 2.8332, | |
| "step": 59000 | |
| }, | |
| { | |
| "epoch": 6.0, | |
| "step": 59298, | |
| "total_flos": 3.0988189827072e+16, | |
| "train_loss": 2.9834200300580593, | |
| "train_runtime": 15417.1336, | |
| "train_samples_per_second": 3.846, | |
| "train_steps_per_second": 3.846 | |
| } | |
| ], | |
| "max_steps": 59298, | |
| "num_train_epochs": 6, | |
| "total_flos": 3.0988189827072e+16, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |