| { |
| "best_metric": 1.1248719692230225, |
| "best_model_checkpoint": "./enko_mbartLarge_36p_run1/checkpoint-50000", |
| "epoch": 6.50376289138716, |
| "eval_steps": 5000, |
| "global_step": 70000, |
| "is_hyper_param_search": false, |
| "is_local_process_zero": true, |
| "is_world_process_zero": true, |
| "log_history": [ |
| { |
| "epoch": 0.05, |
| "learning_rate": 5e-05, |
| "loss": 1.9707, |
| "step": 500 |
| }, |
| { |
| "epoch": 0.09, |
| "learning_rate": 4.9766638663306266e-05, |
| "loss": 1.5932, |
| "step": 1000 |
| }, |
| { |
| "epoch": 0.14, |
| "learning_rate": 4.953327732661253e-05, |
| "loss": 1.4429, |
| "step": 1500 |
| }, |
| { |
| "epoch": 0.19, |
| "learning_rate": 4.929991598991879e-05, |
| "loss": 1.4161, |
| "step": 2000 |
| }, |
| { |
| "epoch": 0.23, |
| "learning_rate": 4.9066554653225056e-05, |
| "loss": 1.386, |
| "step": 2500 |
| }, |
| { |
| "epoch": 0.28, |
| "learning_rate": 4.883319331653132e-05, |
| "loss": 1.3715, |
| "step": 3000 |
| }, |
| { |
| "epoch": 0.33, |
| "learning_rate": 4.859983197983758e-05, |
| "loss": 1.3461, |
| "step": 3500 |
| }, |
| { |
| "epoch": 0.37, |
| "learning_rate": 4.8366470643143846e-05, |
| "loss": 1.3319, |
| "step": 4000 |
| }, |
| { |
| "epoch": 0.42, |
| "learning_rate": 4.813310930645011e-05, |
| "loss": 1.3196, |
| "step": 4500 |
| }, |
| { |
| "epoch": 0.46, |
| "learning_rate": 4.789974796975637e-05, |
| "loss": 1.3157, |
| "step": 5000 |
| }, |
| { |
| "epoch": 0.46, |
| "eval_bleu": 21.6734, |
| "eval_gen_len": 16.4931, |
| "eval_loss": 1.2895288467407227, |
| "eval_runtime": 1327.3616, |
| "eval_samples_per_second": 16.277, |
| "eval_steps_per_second": 1.018, |
| "step": 5000 |
| }, |
| { |
| "epoch": 0.51, |
| "learning_rate": 4.7666386633062636e-05, |
| "loss": 1.2994, |
| "step": 5500 |
| }, |
| { |
| "epoch": 0.56, |
| "learning_rate": 4.74330252963689e-05, |
| "loss": 1.2914, |
| "step": 6000 |
| }, |
| { |
| "epoch": 0.6, |
| "learning_rate": 4.719966395967516e-05, |
| "loss": 1.2989, |
| "step": 6500 |
| }, |
| { |
| "epoch": 0.65, |
| "learning_rate": 4.6966302622981427e-05, |
| "loss": 1.2784, |
| "step": 7000 |
| }, |
| { |
| "epoch": 0.7, |
| "learning_rate": 4.673294128628769e-05, |
| "loss": 1.2662, |
| "step": 7500 |
| }, |
| { |
| "epoch": 0.74, |
| "learning_rate": 4.649957994959395e-05, |
| "loss": 1.2529, |
| "step": 8000 |
| }, |
| { |
| "epoch": 0.79, |
| "learning_rate": 4.626621861290022e-05, |
| "loss": 1.2614, |
| "step": 8500 |
| }, |
| { |
| "epoch": 0.84, |
| "learning_rate": 4.603285727620648e-05, |
| "loss": 1.2564, |
| "step": 9000 |
| }, |
| { |
| "epoch": 0.88, |
| "learning_rate": 4.5799495939512744e-05, |
| "loss": 1.2435, |
| "step": 9500 |
| }, |
| { |
| "epoch": 0.93, |
| "learning_rate": 4.556613460281901e-05, |
| "loss": 1.2575, |
| "step": 10000 |
| }, |
| { |
| "epoch": 0.93, |
| "eval_bleu": 22.5814, |
| "eval_gen_len": 16.8009, |
| "eval_loss": 1.2279398441314697, |
| "eval_runtime": 1333.527, |
| "eval_samples_per_second": 16.201, |
| "eval_steps_per_second": 1.013, |
| "step": 10000 |
| }, |
| { |
| "epoch": 0.98, |
| "learning_rate": 4.533277326612527e-05, |
| "loss": 1.2473, |
| "step": 10500 |
| }, |
| { |
| "epoch": 1.02, |
| "learning_rate": 4.5099411929431534e-05, |
| "loss": 1.2544, |
| "step": 11000 |
| }, |
| { |
| "epoch": 1.07, |
| "learning_rate": 4.48660505927378e-05, |
| "loss": 1.207, |
| "step": 11500 |
| }, |
| { |
| "epoch": 1.11, |
| "learning_rate": 4.463268925604406e-05, |
| "loss": 1.1671, |
| "step": 12000 |
| }, |
| { |
| "epoch": 1.16, |
| "learning_rate": 4.4399327919350324e-05, |
| "loss": 1.1709, |
| "step": 12500 |
| }, |
| { |
| "epoch": 1.21, |
| "learning_rate": 4.416596658265659e-05, |
| "loss": 1.165, |
| "step": 13000 |
| }, |
| { |
| "epoch": 1.25, |
| "learning_rate": 4.393260524596285e-05, |
| "loss": 1.1727, |
| "step": 13500 |
| }, |
| { |
| "epoch": 1.3, |
| "learning_rate": 4.3699243909269114e-05, |
| "loss": 1.17, |
| "step": 14000 |
| }, |
| { |
| "epoch": 1.35, |
| "learning_rate": 4.346588257257538e-05, |
| "loss": 1.1465, |
| "step": 14500 |
| }, |
| { |
| "epoch": 1.39, |
| "learning_rate": 4.323252123588164e-05, |
| "loss": 1.1578, |
| "step": 15000 |
| }, |
| { |
| "epoch": 1.39, |
| "eval_bleu": 24.2386, |
| "eval_gen_len": 16.5838, |
| "eval_loss": 1.1733436584472656, |
| "eval_runtime": 1308.3261, |
| "eval_samples_per_second": 16.513, |
| "eval_steps_per_second": 1.033, |
| "step": 15000 |
| }, |
| { |
| "epoch": 1.44, |
| "learning_rate": 4.2999159899187904e-05, |
| "loss": 1.1582, |
| "step": 15500 |
| }, |
| { |
| "epoch": 1.49, |
| "learning_rate": 4.276579856249417e-05, |
| "loss": 1.1495, |
| "step": 16000 |
| }, |
| { |
| "epoch": 1.53, |
| "learning_rate": 4.253243722580043e-05, |
| "loss": 1.1349, |
| "step": 16500 |
| }, |
| { |
| "epoch": 1.58, |
| "learning_rate": 4.2299075889106695e-05, |
| "loss": 1.1443, |
| "step": 17000 |
| }, |
| { |
| "epoch": 1.63, |
| "learning_rate": 4.206571455241296e-05, |
| "loss": 1.1326, |
| "step": 17500 |
| }, |
| { |
| "epoch": 1.67, |
| "learning_rate": 4.183235321571922e-05, |
| "loss": 1.1425, |
| "step": 18000 |
| }, |
| { |
| "epoch": 1.72, |
| "learning_rate": 4.1598991879025485e-05, |
| "loss": 1.1196, |
| "step": 18500 |
| }, |
| { |
| "epoch": 1.77, |
| "learning_rate": 4.136563054233175e-05, |
| "loss": 1.1272, |
| "step": 19000 |
| }, |
| { |
| "epoch": 1.81, |
| "learning_rate": 4.113226920563801e-05, |
| "loss": 1.1186, |
| "step": 19500 |
| }, |
| { |
| "epoch": 1.86, |
| "learning_rate": 4.0898907868944275e-05, |
| "loss": 1.0885, |
| "step": 20000 |
| }, |
| { |
| "epoch": 1.86, |
| "eval_bleu": 24.7921, |
| "eval_gen_len": 16.6649, |
| "eval_loss": 1.1464108228683472, |
| "eval_runtime": 1309.2944, |
| "eval_samples_per_second": 16.501, |
| "eval_steps_per_second": 1.032, |
| "step": 20000 |
| }, |
| { |
| "epoch": 1.9, |
| "learning_rate": 4.066554653225054e-05, |
| "loss": 1.0541, |
| "step": 20500 |
| }, |
| { |
| "epoch": 1.95, |
| "learning_rate": 4.04321851955568e-05, |
| "loss": 1.0473, |
| "step": 21000 |
| }, |
| { |
| "epoch": 2.0, |
| "learning_rate": 4.0198823858863065e-05, |
| "loss": 1.0511, |
| "step": 21500 |
| }, |
| { |
| "epoch": 2.04, |
| "learning_rate": 3.996546252216933e-05, |
| "loss": 1.0647, |
| "step": 22000 |
| }, |
| { |
| "epoch": 2.09, |
| "learning_rate": 3.973210118547559e-05, |
| "loss": 1.0747, |
| "step": 22500 |
| }, |
| { |
| "epoch": 2.14, |
| "learning_rate": 3.9498739848781855e-05, |
| "loss": 1.044, |
| "step": 23000 |
| }, |
| { |
| "epoch": 2.18, |
| "learning_rate": 3.926537851208812e-05, |
| "loss": 1.052, |
| "step": 23500 |
| }, |
| { |
| "epoch": 2.23, |
| "learning_rate": 3.903201717539438e-05, |
| "loss": 1.0624, |
| "step": 24000 |
| }, |
| { |
| "epoch": 2.28, |
| "learning_rate": 3.8798655838700645e-05, |
| "loss": 1.0493, |
| "step": 24500 |
| }, |
| { |
| "epoch": 2.32, |
| "learning_rate": 3.856529450200691e-05, |
| "loss": 1.0451, |
| "step": 25000 |
| }, |
| { |
| "epoch": 2.32, |
| "eval_bleu": 24.8186, |
| "eval_gen_len": 16.5188, |
| "eval_loss": 1.1437026262283325, |
| "eval_runtime": 1306.8027, |
| "eval_samples_per_second": 16.533, |
| "eval_steps_per_second": 1.034, |
| "step": 25000 |
| }, |
| { |
| "epoch": 2.37, |
| "learning_rate": 3.833193316531317e-05, |
| "loss": 1.054, |
| "step": 25500 |
| }, |
| { |
| "epoch": 2.42, |
| "learning_rate": 3.8098571828619436e-05, |
| "loss": 1.0726, |
| "step": 26000 |
| }, |
| { |
| "epoch": 2.46, |
| "learning_rate": 3.78652104919257e-05, |
| "loss": 1.032, |
| "step": 26500 |
| }, |
| { |
| "epoch": 2.51, |
| "learning_rate": 3.763184915523196e-05, |
| "loss": 0.9733, |
| "step": 27000 |
| }, |
| { |
| "epoch": 2.56, |
| "learning_rate": 3.7398487818538226e-05, |
| "loss": 0.9602, |
| "step": 27500 |
| }, |
| { |
| "epoch": 2.6, |
| "learning_rate": 3.716512648184449e-05, |
| "loss": 0.9606, |
| "step": 28000 |
| }, |
| { |
| "epoch": 2.65, |
| "learning_rate": 3.693176514515075e-05, |
| "loss": 0.9968, |
| "step": 28500 |
| }, |
| { |
| "epoch": 2.69, |
| "learning_rate": 3.6698403808457016e-05, |
| "loss": 1.036, |
| "step": 29000 |
| }, |
| { |
| "epoch": 2.74, |
| "learning_rate": 3.646504247176328e-05, |
| "loss": 1.0285, |
| "step": 29500 |
| }, |
| { |
| "epoch": 2.79, |
| "learning_rate": 3.623168113506954e-05, |
| "loss": 1.0465, |
| "step": 30000 |
| }, |
| { |
| "epoch": 2.79, |
| "eval_bleu": 24.9684, |
| "eval_gen_len": 16.4987, |
| "eval_loss": 1.1424869298934937, |
| "eval_runtime": 1294.7238, |
| "eval_samples_per_second": 16.687, |
| "eval_steps_per_second": 1.043, |
| "step": 30000 |
| }, |
| { |
| "epoch": 2.83, |
| "learning_rate": 3.5998319798375806e-05, |
| "loss": 1.0458, |
| "step": 30500 |
| }, |
| { |
| "epoch": 2.88, |
| "learning_rate": 3.576495846168207e-05, |
| "loss": 0.9843, |
| "step": 31000 |
| }, |
| { |
| "epoch": 2.93, |
| "learning_rate": 3.553159712498833e-05, |
| "loss": 0.9826, |
| "step": 31500 |
| }, |
| { |
| "epoch": 2.97, |
| "learning_rate": 3.5298235788294596e-05, |
| "loss": 0.9817, |
| "step": 32000 |
| }, |
| { |
| "epoch": 3.02, |
| "learning_rate": 3.506487445160086e-05, |
| "loss": 0.9911, |
| "step": 32500 |
| }, |
| { |
| "epoch": 3.07, |
| "learning_rate": 3.483151311490712e-05, |
| "loss": 1.0227, |
| "step": 33000 |
| }, |
| { |
| "epoch": 3.11, |
| "learning_rate": 3.459815177821339e-05, |
| "loss": 1.0126, |
| "step": 33500 |
| }, |
| { |
| "epoch": 3.16, |
| "learning_rate": 3.436479044151965e-05, |
| "loss": 1.0125, |
| "step": 34000 |
| }, |
| { |
| "epoch": 3.21, |
| "learning_rate": 3.4131429104825913e-05, |
| "loss": 1.0224, |
| "step": 34500 |
| }, |
| { |
| "epoch": 3.25, |
| "learning_rate": 3.389806776813218e-05, |
| "loss": 1.0156, |
| "step": 35000 |
| }, |
| { |
| "epoch": 3.25, |
| "eval_bleu": 25.0127, |
| "eval_gen_len": 16.5515, |
| "eval_loss": 1.1464043855667114, |
| "eval_runtime": 1315.6279, |
| "eval_samples_per_second": 16.422, |
| "eval_steps_per_second": 1.027, |
| "step": 35000 |
| }, |
| { |
| "epoch": 3.3, |
| "learning_rate": 3.366470643143844e-05, |
| "loss": 1.009, |
| "step": 35500 |
| }, |
| { |
| "epoch": 3.34, |
| "learning_rate": 3.3431345094744704e-05, |
| "loss": 0.964, |
| "step": 36000 |
| }, |
| { |
| "epoch": 3.39, |
| "learning_rate": 3.319798375805097e-05, |
| "loss": 0.9596, |
| "step": 36500 |
| }, |
| { |
| "epoch": 3.44, |
| "learning_rate": 3.296462242135723e-05, |
| "loss": 0.9587, |
| "step": 37000 |
| }, |
| { |
| "epoch": 3.48, |
| "learning_rate": 3.2731261084663494e-05, |
| "loss": 0.9024, |
| "step": 37500 |
| }, |
| { |
| "epoch": 3.53, |
| "learning_rate": 3.249789974796976e-05, |
| "loss": 0.9277, |
| "step": 38000 |
| }, |
| { |
| "epoch": 3.58, |
| "learning_rate": 3.226453841127603e-05, |
| "loss": 0.9486, |
| "step": 38500 |
| }, |
| { |
| "epoch": 3.62, |
| "learning_rate": 3.2031177074582284e-05, |
| "loss": 0.9543, |
| "step": 39000 |
| }, |
| { |
| "epoch": 3.67, |
| "learning_rate": 3.179781573788855e-05, |
| "loss": 1.0035, |
| "step": 39500 |
| }, |
| { |
| "epoch": 3.72, |
| "learning_rate": 3.156445440119481e-05, |
| "loss": 0.9893, |
| "step": 40000 |
| }, |
| { |
| "epoch": 3.72, |
| "eval_bleu": 24.6323, |
| "eval_gen_len": 16.6096, |
| "eval_loss": 1.1543793678283691, |
| "eval_runtime": 1310.7246, |
| "eval_samples_per_second": 16.483, |
| "eval_steps_per_second": 1.031, |
| "step": 40000 |
| }, |
| { |
| "epoch": 3.76, |
| "learning_rate": 3.1331093064501074e-05, |
| "loss": 0.9877, |
| "step": 40500 |
| }, |
| { |
| "epoch": 3.81, |
| "learning_rate": 3.109773172780734e-05, |
| "loss": 0.9901, |
| "step": 41000 |
| }, |
| { |
| "epoch": 3.86, |
| "learning_rate": 3.08643703911136e-05, |
| "loss": 0.9699, |
| "step": 41500 |
| }, |
| { |
| "epoch": 3.9, |
| "learning_rate": 3.0631009054419864e-05, |
| "loss": 0.9425, |
| "step": 42000 |
| }, |
| { |
| "epoch": 3.95, |
| "learning_rate": 3.039764771772613e-05, |
| "loss": 0.9494, |
| "step": 42500 |
| }, |
| { |
| "epoch": 4.0, |
| "learning_rate": 3.016428638103239e-05, |
| "loss": 0.9508, |
| "step": 43000 |
| }, |
| { |
| "epoch": 4.04, |
| "learning_rate": 2.9930925044338655e-05, |
| "loss": 0.9409, |
| "step": 43500 |
| }, |
| { |
| "epoch": 4.09, |
| "learning_rate": 2.969756370764492e-05, |
| "loss": 0.9012, |
| "step": 44000 |
| }, |
| { |
| "epoch": 4.13, |
| "learning_rate": 2.946420237095118e-05, |
| "loss": 0.8723, |
| "step": 44500 |
| }, |
| { |
| "epoch": 4.18, |
| "learning_rate": 2.9230841034257445e-05, |
| "loss": 0.8779, |
| "step": 45000 |
| }, |
| { |
| "epoch": 4.18, |
| "eval_bleu": 25.3982, |
| "eval_gen_len": 16.457, |
| "eval_loss": 1.1419273614883423, |
| "eval_runtime": 1289.3715, |
| "eval_samples_per_second": 16.756, |
| "eval_steps_per_second": 1.048, |
| "step": 45000 |
| }, |
| { |
| "epoch": 4.23, |
| "learning_rate": 2.8997479697563705e-05, |
| "loss": 0.8922, |
| "step": 45500 |
| }, |
| { |
| "epoch": 4.27, |
| "learning_rate": 2.8764118360869975e-05, |
| "loss": 0.8749, |
| "step": 46000 |
| }, |
| { |
| "epoch": 4.32, |
| "learning_rate": 2.8530757024176235e-05, |
| "loss": 0.8625, |
| "step": 46500 |
| }, |
| { |
| "epoch": 4.37, |
| "learning_rate": 2.82973956874825e-05, |
| "loss": 0.8248, |
| "step": 47000 |
| }, |
| { |
| "epoch": 4.41, |
| "learning_rate": 2.8064034350788765e-05, |
| "loss": 0.8554, |
| "step": 47500 |
| }, |
| { |
| "epoch": 4.46, |
| "learning_rate": 2.7830673014095025e-05, |
| "loss": 0.8822, |
| "step": 48000 |
| }, |
| { |
| "epoch": 4.51, |
| "learning_rate": 2.759731167740129e-05, |
| "loss": 0.7873, |
| "step": 48500 |
| }, |
| { |
| "epoch": 4.55, |
| "learning_rate": 2.7363950340707552e-05, |
| "loss": 0.8181, |
| "step": 49000 |
| }, |
| { |
| "epoch": 4.6, |
| "learning_rate": 2.713058900401382e-05, |
| "loss": 0.8259, |
| "step": 49500 |
| }, |
| { |
| "epoch": 4.65, |
| "learning_rate": 2.689722766732008e-05, |
| "loss": 0.8565, |
| "step": 50000 |
| }, |
| { |
| "epoch": 4.65, |
| "eval_bleu": 26.0152, |
| "eval_gen_len": 16.4749, |
| "eval_loss": 1.1248719692230225, |
| "eval_runtime": 1285.2488, |
| "eval_samples_per_second": 16.81, |
| "eval_steps_per_second": 1.051, |
| "step": 50000 |
| }, |
| { |
| "epoch": 4.69, |
| "learning_rate": 2.6663866330626342e-05, |
| "loss": 0.9321, |
| "step": 50500 |
| }, |
| { |
| "epoch": 4.74, |
| "learning_rate": 2.643050499393261e-05, |
| "loss": 0.8021, |
| "step": 51000 |
| }, |
| { |
| "epoch": 4.78, |
| "learning_rate": 2.6197143657238872e-05, |
| "loss": 0.7701, |
| "step": 51500 |
| }, |
| { |
| "epoch": 4.83, |
| "learning_rate": 2.5963782320545132e-05, |
| "loss": 0.7838, |
| "step": 52000 |
| }, |
| { |
| "epoch": 4.88, |
| "learning_rate": 2.5730420983851396e-05, |
| "loss": 0.7467, |
| "step": 52500 |
| }, |
| { |
| "epoch": 4.92, |
| "learning_rate": 2.5497059647157663e-05, |
| "loss": 0.7778, |
| "step": 53000 |
| }, |
| { |
| "epoch": 4.97, |
| "learning_rate": 2.5263698310463923e-05, |
| "loss": 0.7451, |
| "step": 53500 |
| }, |
| { |
| "epoch": 5.02, |
| "learning_rate": 2.5030336973770186e-05, |
| "loss": 0.7485, |
| "step": 54000 |
| }, |
| { |
| "epoch": 5.06, |
| "learning_rate": 2.479697563707645e-05, |
| "loss": 0.7287, |
| "step": 54500 |
| }, |
| { |
| "epoch": 5.11, |
| "learning_rate": 2.4563614300382716e-05, |
| "loss": 0.7293, |
| "step": 55000 |
| }, |
| { |
| "epoch": 5.11, |
| "eval_bleu": 25.8209, |
| "eval_gen_len": 16.3462, |
| "eval_loss": 1.1566184759140015, |
| "eval_runtime": 1291.0651, |
| "eval_samples_per_second": 16.734, |
| "eval_steps_per_second": 1.046, |
| "step": 55000 |
| }, |
| { |
| "epoch": 5.16, |
| "learning_rate": 2.4330252963688976e-05, |
| "loss": 0.7792, |
| "step": 55500 |
| }, |
| { |
| "epoch": 5.2, |
| "learning_rate": 2.409689162699524e-05, |
| "loss": 0.7879, |
| "step": 56000 |
| }, |
| { |
| "epoch": 5.25, |
| "learning_rate": 2.3863530290301503e-05, |
| "loss": 0.7784, |
| "step": 56500 |
| }, |
| { |
| "epoch": 5.3, |
| "learning_rate": 2.3630168953607766e-05, |
| "loss": 0.7922, |
| "step": 57000 |
| }, |
| { |
| "epoch": 5.34, |
| "learning_rate": 2.3396807616914033e-05, |
| "loss": 0.7878, |
| "step": 57500 |
| }, |
| { |
| "epoch": 5.39, |
| "learning_rate": 2.3163446280220293e-05, |
| "loss": 0.7396, |
| "step": 58000 |
| }, |
| { |
| "epoch": 5.44, |
| "learning_rate": 2.293008494352656e-05, |
| "loss": 0.7602, |
| "step": 58500 |
| }, |
| { |
| "epoch": 5.48, |
| "learning_rate": 2.269672360683282e-05, |
| "loss": 0.7061, |
| "step": 59000 |
| }, |
| { |
| "epoch": 5.53, |
| "learning_rate": 2.2463362270139087e-05, |
| "loss": 0.6956, |
| "step": 59500 |
| }, |
| { |
| "epoch": 5.57, |
| "learning_rate": 2.2230000933445347e-05, |
| "loss": 0.7294, |
| "step": 60000 |
| }, |
| { |
| "epoch": 5.57, |
| "eval_bleu": 25.061, |
| "eval_gen_len": 16.3295, |
| "eval_loss": 1.1824171543121338, |
| "eval_runtime": 1280.888, |
| "eval_samples_per_second": 16.867, |
| "eval_steps_per_second": 1.055, |
| "step": 60000 |
| }, |
| { |
| "epoch": 5.62, |
| "learning_rate": 2.199663959675161e-05, |
| "loss": 0.7407, |
| "step": 60500 |
| }, |
| { |
| "epoch": 5.67, |
| "learning_rate": 2.1763278260057874e-05, |
| "loss": 0.8116, |
| "step": 61000 |
| }, |
| { |
| "epoch": 5.71, |
| "learning_rate": 2.1529916923364137e-05, |
| "loss": 0.8055, |
| "step": 61500 |
| }, |
| { |
| "epoch": 5.76, |
| "learning_rate": 2.1296555586670404e-05, |
| "loss": 0.7065, |
| "step": 62000 |
| }, |
| { |
| "epoch": 5.81, |
| "learning_rate": 2.1063194249976664e-05, |
| "loss": 0.688, |
| "step": 62500 |
| }, |
| { |
| "epoch": 5.85, |
| "learning_rate": 2.082983291328293e-05, |
| "loss": 0.6955, |
| "step": 63000 |
| }, |
| { |
| "epoch": 5.9, |
| "learning_rate": 2.059647157658919e-05, |
| "loss": 0.6782, |
| "step": 63500 |
| }, |
| { |
| "epoch": 5.95, |
| "learning_rate": 2.0363110239895454e-05, |
| "loss": 0.7089, |
| "step": 64000 |
| }, |
| { |
| "epoch": 5.99, |
| "learning_rate": 2.0129748903201717e-05, |
| "loss": 0.7563, |
| "step": 64500 |
| }, |
| { |
| "epoch": 6.04, |
| "learning_rate": 1.989638756650798e-05, |
| "loss": 0.7254, |
| "step": 65000 |
| }, |
| { |
| "epoch": 6.04, |
| "eval_bleu": 24.8712, |
| "eval_gen_len": 16.381, |
| "eval_loss": 1.2153156995773315, |
| "eval_runtime": 1285.2485, |
| "eval_samples_per_second": 16.81, |
| "eval_steps_per_second": 1.051, |
| "step": 65000 |
| }, |
| { |
| "epoch": 6.09, |
| "learning_rate": 1.9663026229814247e-05, |
| "loss": 0.6709, |
| "step": 65500 |
| }, |
| { |
| "epoch": 6.13, |
| "learning_rate": 1.9429664893120507e-05, |
| "loss": 0.7066, |
| "step": 66000 |
| }, |
| { |
| "epoch": 6.18, |
| "learning_rate": 1.9196303556426774e-05, |
| "loss": 0.7173, |
| "step": 66500 |
| }, |
| { |
| "epoch": 6.23, |
| "learning_rate": 1.8962942219733034e-05, |
| "loss": 0.7251, |
| "step": 67000 |
| }, |
| { |
| "epoch": 6.27, |
| "learning_rate": 1.8729580883039298e-05, |
| "loss": 0.723, |
| "step": 67500 |
| }, |
| { |
| "epoch": 6.32, |
| "learning_rate": 1.849621954634556e-05, |
| "loss": 0.7624, |
| "step": 68000 |
| }, |
| { |
| "epoch": 6.36, |
| "learning_rate": 1.8262858209651824e-05, |
| "loss": 0.7217, |
| "step": 68500 |
| }, |
| { |
| "epoch": 6.41, |
| "learning_rate": 1.802949687295809e-05, |
| "loss": 0.7277, |
| "step": 69000 |
| }, |
| { |
| "epoch": 6.46, |
| "learning_rate": 1.779613553626435e-05, |
| "loss": 0.7644, |
| "step": 69500 |
| }, |
| { |
| "epoch": 6.5, |
| "learning_rate": 1.7562774199570618e-05, |
| "loss": 0.6469, |
| "step": 70000 |
| }, |
| { |
| "epoch": 6.5, |
| "eval_bleu": 25.3326, |
| "eval_gen_len": 16.5447, |
| "eval_loss": 1.1965116262435913, |
| "eval_runtime": 1299.2068, |
| "eval_samples_per_second": 16.629, |
| "eval_steps_per_second": 1.04, |
| "step": 70000 |
| }, |
| { |
| "epoch": 6.5, |
| "step": 70000, |
| "total_flos": 2.4274280583538934e+18, |
| "train_loss": 0.9888105150495257, |
| "train_runtime": 69154.692, |
| "train_samples_per_second": 24.901, |
| "train_steps_per_second": 1.556 |
| } |
| ], |
| "logging_steps": 500, |
| "max_steps": 107630, |
| "num_train_epochs": 10, |
| "save_steps": 5000, |
| "total_flos": 2.4274280583538934e+18, |
| "trial_name": null, |
| "trial_params": null |
| } |
|
|