| { | |
| "best_global_step": null, | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 35.0, | |
| "eval_steps": 500, | |
| "global_step": 7665, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.045662100456621, | |
| "grad_norm": 7.013392925262451, | |
| "learning_rate": 4.994129158512721e-05, | |
| "loss": 2.6272, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.091324200913242, | |
| "grad_norm": 2.9674174785614014, | |
| "learning_rate": 4.987606001304632e-05, | |
| "loss": 0.6549, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.136986301369863, | |
| "grad_norm": 2.2170982360839844, | |
| "learning_rate": 4.981082844096543e-05, | |
| "loss": 0.274, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.182648401826484, | |
| "grad_norm": 1.2071510553359985, | |
| "learning_rate": 4.974559686888454e-05, | |
| "loss": 0.1543, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.228310502283105, | |
| "grad_norm": 0.8103004693984985, | |
| "learning_rate": 4.9680365296803655e-05, | |
| "loss": 0.146, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.273972602739726, | |
| "grad_norm": 0.8910240530967712, | |
| "learning_rate": 4.961513372472277e-05, | |
| "loss": 0.1213, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.319634703196347, | |
| "grad_norm": 1.551806926727295, | |
| "learning_rate": 4.954990215264188e-05, | |
| "loss": 0.1473, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 0.365296803652968, | |
| "grad_norm": 0.5918143391609192, | |
| "learning_rate": 4.948467058056099e-05, | |
| "loss": 0.0856, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.410958904109589, | |
| "grad_norm": 1.8692866563796997, | |
| "learning_rate": 4.9419439008480104e-05, | |
| "loss": 0.0968, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 0.45662100456621, | |
| "grad_norm": 1.902163028717041, | |
| "learning_rate": 4.9354207436399216e-05, | |
| "loss": 0.0993, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.502283105022831, | |
| "grad_norm": 0.6189514398574829, | |
| "learning_rate": 4.9288975864318335e-05, | |
| "loss": 0.0896, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 0.547945205479452, | |
| "grad_norm": 0.9506679773330688, | |
| "learning_rate": 4.922374429223745e-05, | |
| "loss": 0.0816, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 0.593607305936073, | |
| "grad_norm": 1.173135757446289, | |
| "learning_rate": 4.915851272015656e-05, | |
| "loss": 0.0794, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 0.639269406392694, | |
| "grad_norm": 0.889118492603302, | |
| "learning_rate": 4.909328114807567e-05, | |
| "loss": 0.0592, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 0.684931506849315, | |
| "grad_norm": 0.6385554671287537, | |
| "learning_rate": 4.9028049575994784e-05, | |
| "loss": 0.052, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 0.730593607305936, | |
| "grad_norm": 0.889135479927063, | |
| "learning_rate": 4.8962818003913896e-05, | |
| "loss": 0.0607, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 0.776255707762557, | |
| "grad_norm": 5.079071998596191, | |
| "learning_rate": 4.889758643183301e-05, | |
| "loss": 0.0888, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 0.821917808219178, | |
| "grad_norm": 4.792678356170654, | |
| "learning_rate": 4.883235485975212e-05, | |
| "loss": 0.0767, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 0.867579908675799, | |
| "grad_norm": 0.6212195754051208, | |
| "learning_rate": 4.876712328767123e-05, | |
| "loss": 0.121, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 0.91324200913242, | |
| "grad_norm": 0.9692168235778809, | |
| "learning_rate": 4.8701891715590345e-05, | |
| "loss": 0.0716, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.958904109589041, | |
| "grad_norm": 0.5348126888275146, | |
| "learning_rate": 4.8636660143509464e-05, | |
| "loss": 0.0828, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "eval_bertscore_f1": 0.8785514626912252, | |
| "eval_bleu": 0.6709298648976769, | |
| "eval_loss": 0.059928230941295624, | |
| "eval_rougeL": 0.2992309920195003, | |
| "eval_runtime": 78.4856, | |
| "eval_samples_per_second": 19.15, | |
| "eval_steps_per_second": 1.198, | |
| "step": 219 | |
| }, | |
| { | |
| "epoch": 1.004566210045662, | |
| "grad_norm": 0.592220664024353, | |
| "learning_rate": 4.8571428571428576e-05, | |
| "loss": 0.1077, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 1.0502283105022832, | |
| "grad_norm": 0.6266194581985474, | |
| "learning_rate": 4.850619699934769e-05, | |
| "loss": 0.0584, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 1.095890410958904, | |
| "grad_norm": 0.7085559964179993, | |
| "learning_rate": 4.84409654272668e-05, | |
| "loss": 0.0663, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 1.1415525114155252, | |
| "grad_norm": 0.4316001832485199, | |
| "learning_rate": 4.837573385518591e-05, | |
| "loss": 0.0528, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 1.187214611872146, | |
| "grad_norm": 0.4922642707824707, | |
| "learning_rate": 4.8310502283105025e-05, | |
| "loss": 0.0502, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 1.2328767123287672, | |
| "grad_norm": 0.4204493761062622, | |
| "learning_rate": 4.824527071102414e-05, | |
| "loss": 0.0567, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 1.278538812785388, | |
| "grad_norm": 0.3499491810798645, | |
| "learning_rate": 4.818003913894325e-05, | |
| "loss": 0.055, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 1.3242009132420092, | |
| "grad_norm": 0.4341481328010559, | |
| "learning_rate": 4.811480756686236e-05, | |
| "loss": 0.0462, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 1.36986301369863, | |
| "grad_norm": 0.7632847428321838, | |
| "learning_rate": 4.804957599478147e-05, | |
| "loss": 0.0578, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 1.4155251141552512, | |
| "grad_norm": 0.9663671255111694, | |
| "learning_rate": 4.798434442270059e-05, | |
| "loss": 0.0421, | |
| "step": 310 | |
| }, | |
| { | |
| "epoch": 1.461187214611872, | |
| "grad_norm": 0.42978063225746155, | |
| "learning_rate": 4.7919112850619704e-05, | |
| "loss": 0.0489, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 1.5068493150684932, | |
| "grad_norm": 0.4420313239097595, | |
| "learning_rate": 4.7853881278538817e-05, | |
| "loss": 0.0572, | |
| "step": 330 | |
| }, | |
| { | |
| "epoch": 1.5525114155251143, | |
| "grad_norm": 0.5001460909843445, | |
| "learning_rate": 4.778864970645793e-05, | |
| "loss": 0.0526, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 1.5981735159817352, | |
| "grad_norm": 0.3782072961330414, | |
| "learning_rate": 4.772341813437704e-05, | |
| "loss": 0.0656, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 1.643835616438356, | |
| "grad_norm": 0.7828798294067383, | |
| "learning_rate": 4.765818656229615e-05, | |
| "loss": 0.0545, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 1.6894977168949772, | |
| "grad_norm": 0.477966845035553, | |
| "learning_rate": 4.7592954990215265e-05, | |
| "loss": 0.0595, | |
| "step": 370 | |
| }, | |
| { | |
| "epoch": 1.7351598173515983, | |
| "grad_norm": 0.3790488541126251, | |
| "learning_rate": 4.752772341813438e-05, | |
| "loss": 0.0784, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 1.7808219178082192, | |
| "grad_norm": 0.551754355430603, | |
| "learning_rate": 4.746249184605349e-05, | |
| "loss": 0.071, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 1.82648401826484, | |
| "grad_norm": 0.34281566739082336, | |
| "learning_rate": 4.73972602739726e-05, | |
| "loss": 0.0405, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 1.8721461187214612, | |
| "grad_norm": 1.1913338899612427, | |
| "learning_rate": 4.733202870189172e-05, | |
| "loss": 0.0731, | |
| "step": 410 | |
| }, | |
| { | |
| "epoch": 1.9178082191780823, | |
| "grad_norm": 0.34058138728141785, | |
| "learning_rate": 4.726679712981083e-05, | |
| "loss": 0.0564, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 1.9634703196347032, | |
| "grad_norm": 0.311128169298172, | |
| "learning_rate": 4.7201565557729945e-05, | |
| "loss": 0.0534, | |
| "step": 430 | |
| }, | |
| { | |
| "epoch": 2.0, | |
| "eval_bertscore_f1": 0.8737317200351062, | |
| "eval_bleu": 0.4862796407603182, | |
| "eval_loss": 0.04997913911938667, | |
| "eval_rougeL": 0.3070874128312525, | |
| "eval_runtime": 65.3099, | |
| "eval_samples_per_second": 23.013, | |
| "eval_steps_per_second": 1.439, | |
| "step": 438 | |
| }, | |
| { | |
| "epoch": 2.009132420091324, | |
| "grad_norm": 0.5538708567619324, | |
| "learning_rate": 4.713633398564906e-05, | |
| "loss": 0.0767, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 2.0547945205479454, | |
| "grad_norm": 0.40191444754600525, | |
| "learning_rate": 4.707110241356817e-05, | |
| "loss": 0.0413, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 2.1004566210045663, | |
| "grad_norm": 0.5775645971298218, | |
| "learning_rate": 4.700587084148728e-05, | |
| "loss": 0.0527, | |
| "step": 460 | |
| }, | |
| { | |
| "epoch": 2.146118721461187, | |
| "grad_norm": 0.6470644474029541, | |
| "learning_rate": 4.6940639269406394e-05, | |
| "loss": 0.0467, | |
| "step": 470 | |
| }, | |
| { | |
| "epoch": 2.191780821917808, | |
| "grad_norm": 1.177239179611206, | |
| "learning_rate": 4.6875407697325506e-05, | |
| "loss": 0.0526, | |
| "step": 480 | |
| }, | |
| { | |
| "epoch": 2.237442922374429, | |
| "grad_norm": 0.42161795496940613, | |
| "learning_rate": 4.681017612524462e-05, | |
| "loss": 0.0549, | |
| "step": 490 | |
| }, | |
| { | |
| "epoch": 2.2831050228310503, | |
| "grad_norm": 0.24745285511016846, | |
| "learning_rate": 4.674494455316373e-05, | |
| "loss": 0.04, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 2.328767123287671, | |
| "grad_norm": 0.301503449678421, | |
| "learning_rate": 4.667971298108285e-05, | |
| "loss": 0.0599, | |
| "step": 510 | |
| }, | |
| { | |
| "epoch": 2.374429223744292, | |
| "grad_norm": 0.8956980109214783, | |
| "learning_rate": 4.661448140900196e-05, | |
| "loss": 0.0622, | |
| "step": 520 | |
| }, | |
| { | |
| "epoch": 2.4200913242009134, | |
| "grad_norm": 0.32675889134407043, | |
| "learning_rate": 4.6549249836921074e-05, | |
| "loss": 0.0398, | |
| "step": 530 | |
| }, | |
| { | |
| "epoch": 2.4657534246575343, | |
| "grad_norm": 0.32268226146698, | |
| "learning_rate": 4.6484018264840186e-05, | |
| "loss": 0.0423, | |
| "step": 540 | |
| }, | |
| { | |
| "epoch": 2.5114155251141552, | |
| "grad_norm": 0.3572223484516144, | |
| "learning_rate": 4.64187866927593e-05, | |
| "loss": 0.0433, | |
| "step": 550 | |
| }, | |
| { | |
| "epoch": 2.557077625570776, | |
| "grad_norm": 0.32862910628318787, | |
| "learning_rate": 4.635355512067841e-05, | |
| "loss": 0.0471, | |
| "step": 560 | |
| }, | |
| { | |
| "epoch": 2.602739726027397, | |
| "grad_norm": 0.413512259721756, | |
| "learning_rate": 4.628832354859752e-05, | |
| "loss": 0.0575, | |
| "step": 570 | |
| }, | |
| { | |
| "epoch": 2.6484018264840183, | |
| "grad_norm": 0.6657458543777466, | |
| "learning_rate": 4.6223091976516634e-05, | |
| "loss": 0.0463, | |
| "step": 580 | |
| }, | |
| { | |
| "epoch": 2.6940639269406392, | |
| "grad_norm": 0.3851383626461029, | |
| "learning_rate": 4.6157860404435747e-05, | |
| "loss": 0.0543, | |
| "step": 590 | |
| }, | |
| { | |
| "epoch": 2.73972602739726, | |
| "grad_norm": 0.5988168716430664, | |
| "learning_rate": 4.609262883235486e-05, | |
| "loss": 0.0546, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 2.7853881278538815, | |
| "grad_norm": 0.45219168066978455, | |
| "learning_rate": 4.602739726027398e-05, | |
| "loss": 0.0472, | |
| "step": 610 | |
| }, | |
| { | |
| "epoch": 2.8310502283105023, | |
| "grad_norm": 0.3951447010040283, | |
| "learning_rate": 4.596216568819309e-05, | |
| "loss": 0.0478, | |
| "step": 620 | |
| }, | |
| { | |
| "epoch": 2.8767123287671232, | |
| "grad_norm": 0.2301015555858612, | |
| "learning_rate": 4.58969341161122e-05, | |
| "loss": 0.041, | |
| "step": 630 | |
| }, | |
| { | |
| "epoch": 2.922374429223744, | |
| "grad_norm": 0.556537926197052, | |
| "learning_rate": 4.5831702544031314e-05, | |
| "loss": 0.0386, | |
| "step": 640 | |
| }, | |
| { | |
| "epoch": 2.968036529680365, | |
| "grad_norm": 1.080448865890503, | |
| "learning_rate": 4.5766470971950426e-05, | |
| "loss": 0.0499, | |
| "step": 650 | |
| }, | |
| { | |
| "epoch": 3.0, | |
| "eval_bertscore_f1": 0.8748933539260488, | |
| "eval_bleu": 0.9425275721886326, | |
| "eval_loss": 0.04728687182068825, | |
| "eval_rougeL": 0.28689011243059326, | |
| "eval_runtime": 65.6139, | |
| "eval_samples_per_second": 22.907, | |
| "eval_steps_per_second": 1.433, | |
| "step": 657 | |
| }, | |
| { | |
| "epoch": 3.0136986301369864, | |
| "grad_norm": 0.2687000334262848, | |
| "learning_rate": 4.570123939986954e-05, | |
| "loss": 0.0616, | |
| "step": 660 | |
| }, | |
| { | |
| "epoch": 3.0593607305936072, | |
| "grad_norm": 0.45356887578964233, | |
| "learning_rate": 4.563600782778865e-05, | |
| "loss": 0.0462, | |
| "step": 670 | |
| }, | |
| { | |
| "epoch": 3.105022831050228, | |
| "grad_norm": 0.24934875965118408, | |
| "learning_rate": 4.557077625570776e-05, | |
| "loss": 0.0412, | |
| "step": 680 | |
| }, | |
| { | |
| "epoch": 3.1506849315068495, | |
| "grad_norm": 0.3431055545806885, | |
| "learning_rate": 4.5505544683626875e-05, | |
| "loss": 0.0391, | |
| "step": 690 | |
| }, | |
| { | |
| "epoch": 3.1963470319634704, | |
| "grad_norm": 0.46532484889030457, | |
| "learning_rate": 4.544031311154599e-05, | |
| "loss": 0.0571, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 3.2420091324200913, | |
| "grad_norm": 0.24521034955978394, | |
| "learning_rate": 4.5375081539465106e-05, | |
| "loss": 0.0479, | |
| "step": 710 | |
| }, | |
| { | |
| "epoch": 3.287671232876712, | |
| "grad_norm": 0.22706200182437897, | |
| "learning_rate": 4.530984996738422e-05, | |
| "loss": 0.04, | |
| "step": 720 | |
| }, | |
| { | |
| "epoch": 3.3333333333333335, | |
| "grad_norm": 0.4912814199924469, | |
| "learning_rate": 4.524461839530333e-05, | |
| "loss": 0.0399, | |
| "step": 730 | |
| }, | |
| { | |
| "epoch": 3.3789954337899544, | |
| "grad_norm": 0.3373461365699768, | |
| "learning_rate": 4.517938682322244e-05, | |
| "loss": 0.0402, | |
| "step": 740 | |
| }, | |
| { | |
| "epoch": 3.4246575342465753, | |
| "grad_norm": 0.6649588942527771, | |
| "learning_rate": 4.5114155251141555e-05, | |
| "loss": 0.0391, | |
| "step": 750 | |
| }, | |
| { | |
| "epoch": 3.470319634703196, | |
| "grad_norm": 0.34905117750167847, | |
| "learning_rate": 4.504892367906067e-05, | |
| "loss": 0.0482, | |
| "step": 760 | |
| }, | |
| { | |
| "epoch": 3.5159817351598175, | |
| "grad_norm": 0.24569763243198395, | |
| "learning_rate": 4.498369210697978e-05, | |
| "loss": 0.0359, | |
| "step": 770 | |
| }, | |
| { | |
| "epoch": 3.5616438356164384, | |
| "grad_norm": 0.294628769159317, | |
| "learning_rate": 4.491846053489889e-05, | |
| "loss": 0.0429, | |
| "step": 780 | |
| }, | |
| { | |
| "epoch": 3.6073059360730593, | |
| "grad_norm": 0.23211832344532013, | |
| "learning_rate": 4.4853228962818004e-05, | |
| "loss": 0.0328, | |
| "step": 790 | |
| }, | |
| { | |
| "epoch": 3.65296803652968, | |
| "grad_norm": 0.19963033497333527, | |
| "learning_rate": 4.4787997390737116e-05, | |
| "loss": 0.0397, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 3.6986301369863015, | |
| "grad_norm": 0.2943974733352661, | |
| "learning_rate": 4.4722765818656235e-05, | |
| "loss": 0.0423, | |
| "step": 810 | |
| }, | |
| { | |
| "epoch": 3.7442922374429224, | |
| "grad_norm": 0.20433223247528076, | |
| "learning_rate": 4.465753424657535e-05, | |
| "loss": 0.0375, | |
| "step": 820 | |
| }, | |
| { | |
| "epoch": 3.7899543378995433, | |
| "grad_norm": 0.22242455184459686, | |
| "learning_rate": 4.459230267449446e-05, | |
| "loss": 0.0384, | |
| "step": 830 | |
| }, | |
| { | |
| "epoch": 3.8356164383561646, | |
| "grad_norm": 0.3566916882991791, | |
| "learning_rate": 4.452707110241357e-05, | |
| "loss": 0.0383, | |
| "step": 840 | |
| }, | |
| { | |
| "epoch": 3.8812785388127855, | |
| "grad_norm": 0.9326187372207642, | |
| "learning_rate": 4.446183953033268e-05, | |
| "loss": 0.0486, | |
| "step": 850 | |
| }, | |
| { | |
| "epoch": 3.9269406392694064, | |
| "grad_norm": 0.547247588634491, | |
| "learning_rate": 4.4396607958251795e-05, | |
| "loss": 0.0478, | |
| "step": 860 | |
| }, | |
| { | |
| "epoch": 3.9726027397260273, | |
| "grad_norm": 0.5731884241104126, | |
| "learning_rate": 4.433137638617091e-05, | |
| "loss": 0.038, | |
| "step": 870 | |
| }, | |
| { | |
| "epoch": 4.0, | |
| "eval_bertscore_f1": 0.8795713798966475, | |
| "eval_bleu": 0.6914713325727705, | |
| "eval_loss": 0.04532833769917488, | |
| "eval_rougeL": 0.30243117527372954, | |
| "eval_runtime": 65.1517, | |
| "eval_samples_per_second": 23.069, | |
| "eval_steps_per_second": 1.443, | |
| "step": 876 | |
| }, | |
| { | |
| "epoch": 4.018264840182648, | |
| "grad_norm": 0.7525059580802917, | |
| "learning_rate": 4.426614481409002e-05, | |
| "loss": 0.0482, | |
| "step": 880 | |
| }, | |
| { | |
| "epoch": 4.063926940639269, | |
| "grad_norm": 0.4412573277950287, | |
| "learning_rate": 4.420091324200913e-05, | |
| "loss": 0.0383, | |
| "step": 890 | |
| }, | |
| { | |
| "epoch": 4.109589041095891, | |
| "grad_norm": 0.4645381569862366, | |
| "learning_rate": 4.4135681669928244e-05, | |
| "loss": 0.041, | |
| "step": 900 | |
| }, | |
| { | |
| "epoch": 4.155251141552512, | |
| "grad_norm": 0.2361067682504654, | |
| "learning_rate": 4.407045009784736e-05, | |
| "loss": 0.0412, | |
| "step": 910 | |
| }, | |
| { | |
| "epoch": 4.200913242009133, | |
| "grad_norm": 0.2725714445114136, | |
| "learning_rate": 4.4005218525766475e-05, | |
| "loss": 0.0409, | |
| "step": 920 | |
| }, | |
| { | |
| "epoch": 4.2465753424657535, | |
| "grad_norm": 0.25887736678123474, | |
| "learning_rate": 4.393998695368559e-05, | |
| "loss": 0.0391, | |
| "step": 930 | |
| }, | |
| { | |
| "epoch": 4.292237442922374, | |
| "grad_norm": 0.22646628320217133, | |
| "learning_rate": 4.38747553816047e-05, | |
| "loss": 0.0416, | |
| "step": 940 | |
| }, | |
| { | |
| "epoch": 4.337899543378995, | |
| "grad_norm": 0.30533650517463684, | |
| "learning_rate": 4.380952380952381e-05, | |
| "loss": 0.0412, | |
| "step": 950 | |
| }, | |
| { | |
| "epoch": 4.383561643835616, | |
| "grad_norm": 0.2850722670555115, | |
| "learning_rate": 4.3744292237442924e-05, | |
| "loss": 0.0341, | |
| "step": 960 | |
| }, | |
| { | |
| "epoch": 4.429223744292237, | |
| "grad_norm": 0.6640023589134216, | |
| "learning_rate": 4.3679060665362036e-05, | |
| "loss": 0.0448, | |
| "step": 970 | |
| }, | |
| { | |
| "epoch": 4.474885844748858, | |
| "grad_norm": 0.4189640283584595, | |
| "learning_rate": 4.361382909328115e-05, | |
| "loss": 0.0364, | |
| "step": 980 | |
| }, | |
| { | |
| "epoch": 4.52054794520548, | |
| "grad_norm": 0.27891141176223755, | |
| "learning_rate": 4.354859752120026e-05, | |
| "loss": 0.0377, | |
| "step": 990 | |
| }, | |
| { | |
| "epoch": 4.566210045662101, | |
| "grad_norm": 0.24627180397510529, | |
| "learning_rate": 4.348336594911937e-05, | |
| "loss": 0.0401, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 4.6118721461187215, | |
| "grad_norm": 0.23066160082817078, | |
| "learning_rate": 4.341813437703849e-05, | |
| "loss": 0.0366, | |
| "step": 1010 | |
| }, | |
| { | |
| "epoch": 4.657534246575342, | |
| "grad_norm": 0.2346690148115158, | |
| "learning_rate": 4.3352902804957604e-05, | |
| "loss": 0.0441, | |
| "step": 1020 | |
| }, | |
| { | |
| "epoch": 4.703196347031963, | |
| "grad_norm": 0.2368006706237793, | |
| "learning_rate": 4.3287671232876716e-05, | |
| "loss": 0.0406, | |
| "step": 1030 | |
| }, | |
| { | |
| "epoch": 4.748858447488584, | |
| "grad_norm": 0.19953212141990662, | |
| "learning_rate": 4.322243966079583e-05, | |
| "loss": 0.0398, | |
| "step": 1040 | |
| }, | |
| { | |
| "epoch": 4.794520547945205, | |
| "grad_norm": 0.8159717321395874, | |
| "learning_rate": 4.315720808871494e-05, | |
| "loss": 0.0426, | |
| "step": 1050 | |
| }, | |
| { | |
| "epoch": 4.840182648401827, | |
| "grad_norm": 0.29258477687835693, | |
| "learning_rate": 4.309197651663405e-05, | |
| "loss": 0.0376, | |
| "step": 1060 | |
| }, | |
| { | |
| "epoch": 4.885844748858448, | |
| "grad_norm": 0.22379961609840393, | |
| "learning_rate": 4.3026744944553165e-05, | |
| "loss": 0.0461, | |
| "step": 1070 | |
| }, | |
| { | |
| "epoch": 4.931506849315069, | |
| "grad_norm": 0.36883947253227234, | |
| "learning_rate": 4.296151337247228e-05, | |
| "loss": 0.0464, | |
| "step": 1080 | |
| }, | |
| { | |
| "epoch": 4.9771689497716896, | |
| "grad_norm": 0.26361939311027527, | |
| "learning_rate": 4.289628180039139e-05, | |
| "loss": 0.0425, | |
| "step": 1090 | |
| }, | |
| { | |
| "epoch": 5.0, | |
| "eval_bertscore_f1": 0.874491546682255, | |
| "eval_bleu": 0.5863769810718094, | |
| "eval_loss": 0.044686976820230484, | |
| "eval_rougeL": 0.30493134131703603, | |
| "eval_runtime": 65.1008, | |
| "eval_samples_per_second": 23.087, | |
| "eval_steps_per_second": 1.444, | |
| "step": 1095 | |
| }, | |
| { | |
| "epoch": 5.0228310502283104, | |
| "grad_norm": 0.29032638669013977, | |
| "learning_rate": 4.28310502283105e-05, | |
| "loss": 0.0407, | |
| "step": 1100 | |
| }, | |
| { | |
| "epoch": 5.068493150684931, | |
| "grad_norm": 0.1798933893442154, | |
| "learning_rate": 4.276581865622962e-05, | |
| "loss": 0.0369, | |
| "step": 1110 | |
| }, | |
| { | |
| "epoch": 5.114155251141552, | |
| "grad_norm": 0.20421646535396576, | |
| "learning_rate": 4.270058708414873e-05, | |
| "loss": 0.0366, | |
| "step": 1120 | |
| }, | |
| { | |
| "epoch": 5.159817351598173, | |
| "grad_norm": 0.4460963308811188, | |
| "learning_rate": 4.2635355512067844e-05, | |
| "loss": 0.037, | |
| "step": 1130 | |
| }, | |
| { | |
| "epoch": 5.205479452054795, | |
| "grad_norm": 0.19764983654022217, | |
| "learning_rate": 4.257012393998696e-05, | |
| "loss": 0.0369, | |
| "step": 1140 | |
| }, | |
| { | |
| "epoch": 5.251141552511416, | |
| "grad_norm": 0.15257489681243896, | |
| "learning_rate": 4.250489236790607e-05, | |
| "loss": 0.0357, | |
| "step": 1150 | |
| }, | |
| { | |
| "epoch": 5.296803652968037, | |
| "grad_norm": 0.9983281493186951, | |
| "learning_rate": 4.243966079582518e-05, | |
| "loss": 0.0406, | |
| "step": 1160 | |
| }, | |
| { | |
| "epoch": 5.342465753424658, | |
| "grad_norm": 0.253713995218277, | |
| "learning_rate": 4.237442922374429e-05, | |
| "loss": 0.0353, | |
| "step": 1170 | |
| }, | |
| { | |
| "epoch": 5.3881278538812785, | |
| "grad_norm": 0.24854066967964172, | |
| "learning_rate": 4.2309197651663405e-05, | |
| "loss": 0.0352, | |
| "step": 1180 | |
| }, | |
| { | |
| "epoch": 5.433789954337899, | |
| "grad_norm": 0.29193824529647827, | |
| "learning_rate": 4.224396607958252e-05, | |
| "loss": 0.0452, | |
| "step": 1190 | |
| }, | |
| { | |
| "epoch": 5.47945205479452, | |
| "grad_norm": 0.40590381622314453, | |
| "learning_rate": 4.217873450750163e-05, | |
| "loss": 0.0378, | |
| "step": 1200 | |
| }, | |
| { | |
| "epoch": 5.525114155251142, | |
| "grad_norm": 0.18774063885211945, | |
| "learning_rate": 4.211350293542075e-05, | |
| "loss": 0.0425, | |
| "step": 1210 | |
| }, | |
| { | |
| "epoch": 5.570776255707763, | |
| "grad_norm": 0.26123547554016113, | |
| "learning_rate": 4.204827136333986e-05, | |
| "loss": 0.0421, | |
| "step": 1220 | |
| }, | |
| { | |
| "epoch": 5.616438356164384, | |
| "grad_norm": 0.6077787280082703, | |
| "learning_rate": 4.198303979125897e-05, | |
| "loss": 0.0342, | |
| "step": 1230 | |
| }, | |
| { | |
| "epoch": 5.662100456621005, | |
| "grad_norm": 0.3458491563796997, | |
| "learning_rate": 4.1917808219178085e-05, | |
| "loss": 0.0357, | |
| "step": 1240 | |
| }, | |
| { | |
| "epoch": 5.707762557077626, | |
| "grad_norm": 0.4896857440471649, | |
| "learning_rate": 4.18525766470972e-05, | |
| "loss": 0.0396, | |
| "step": 1250 | |
| }, | |
| { | |
| "epoch": 5.7534246575342465, | |
| "grad_norm": 0.16601338982582092, | |
| "learning_rate": 4.178734507501631e-05, | |
| "loss": 0.0389, | |
| "step": 1260 | |
| }, | |
| { | |
| "epoch": 5.799086757990867, | |
| "grad_norm": 0.2273034006357193, | |
| "learning_rate": 4.172211350293542e-05, | |
| "loss": 0.0434, | |
| "step": 1270 | |
| }, | |
| { | |
| "epoch": 5.844748858447488, | |
| "grad_norm": 0.3890644907951355, | |
| "learning_rate": 4.1656881930854534e-05, | |
| "loss": 0.0437, | |
| "step": 1280 | |
| }, | |
| { | |
| "epoch": 5.890410958904109, | |
| "grad_norm": 0.20492355525493622, | |
| "learning_rate": 4.1591650358773646e-05, | |
| "loss": 0.0382, | |
| "step": 1290 | |
| }, | |
| { | |
| "epoch": 5.936073059360731, | |
| "grad_norm": 0.34492093324661255, | |
| "learning_rate": 4.152641878669276e-05, | |
| "loss": 0.0334, | |
| "step": 1300 | |
| }, | |
| { | |
| "epoch": 5.981735159817352, | |
| "grad_norm": 0.20600809156894684, | |
| "learning_rate": 4.146118721461188e-05, | |
| "loss": 0.0345, | |
| "step": 1310 | |
| }, | |
| { | |
| "epoch": 6.0, | |
| "eval_bertscore_f1": 0.8773880335226586, | |
| "eval_bleu": 0.7957797365038125, | |
| "eval_loss": 0.04304474964737892, | |
| "eval_rougeL": 0.298324896779863, | |
| "eval_runtime": 65.1124, | |
| "eval_samples_per_second": 23.083, | |
| "eval_steps_per_second": 1.444, | |
| "step": 1314 | |
| }, | |
| { | |
| "epoch": 6.027397260273973, | |
| "grad_norm": 0.28056013584136963, | |
| "learning_rate": 4.139595564253099e-05, | |
| "loss": 0.0351, | |
| "step": 1320 | |
| }, | |
| { | |
| "epoch": 6.073059360730594, | |
| "grad_norm": 12.815491676330566, | |
| "learning_rate": 4.13307240704501e-05, | |
| "loss": 0.0398, | |
| "step": 1330 | |
| }, | |
| { | |
| "epoch": 6.1187214611872145, | |
| "grad_norm": 0.3111923635005951, | |
| "learning_rate": 4.1265492498369214e-05, | |
| "loss": 0.0428, | |
| "step": 1340 | |
| }, | |
| { | |
| "epoch": 6.164383561643835, | |
| "grad_norm": 0.23117075860500336, | |
| "learning_rate": 4.1200260926288326e-05, | |
| "loss": 0.0351, | |
| "step": 1350 | |
| }, | |
| { | |
| "epoch": 6.210045662100456, | |
| "grad_norm": 0.22561274468898773, | |
| "learning_rate": 4.113502935420744e-05, | |
| "loss": 0.0385, | |
| "step": 1360 | |
| }, | |
| { | |
| "epoch": 6.255707762557078, | |
| "grad_norm": 0.16258157789707184, | |
| "learning_rate": 4.106979778212655e-05, | |
| "loss": 0.0339, | |
| "step": 1370 | |
| }, | |
| { | |
| "epoch": 6.301369863013699, | |
| "grad_norm": 0.6510432362556458, | |
| "learning_rate": 4.100456621004566e-05, | |
| "loss": 0.036, | |
| "step": 1380 | |
| }, | |
| { | |
| "epoch": 6.34703196347032, | |
| "grad_norm": 0.24988140165805817, | |
| "learning_rate": 4.0939334637964774e-05, | |
| "loss": 0.0355, | |
| "step": 1390 | |
| }, | |
| { | |
| "epoch": 6.392694063926941, | |
| "grad_norm": 0.3678286671638489, | |
| "learning_rate": 4.087410306588389e-05, | |
| "loss": 0.0338, | |
| "step": 1400 | |
| }, | |
| { | |
| "epoch": 6.438356164383562, | |
| "grad_norm": 0.2946147918701172, | |
| "learning_rate": 4.0808871493803006e-05, | |
| "loss": 0.0419, | |
| "step": 1410 | |
| }, | |
| { | |
| "epoch": 6.4840182648401825, | |
| "grad_norm": 0.21736349165439606, | |
| "learning_rate": 4.074363992172212e-05, | |
| "loss": 0.0344, | |
| "step": 1420 | |
| }, | |
| { | |
| "epoch": 6.529680365296803, | |
| "grad_norm": 0.3145897388458252, | |
| "learning_rate": 4.067840834964123e-05, | |
| "loss": 0.0385, | |
| "step": 1430 | |
| }, | |
| { | |
| "epoch": 6.575342465753424, | |
| "grad_norm": 0.24647307395935059, | |
| "learning_rate": 4.061317677756034e-05, | |
| "loss": 0.0384, | |
| "step": 1440 | |
| }, | |
| { | |
| "epoch": 6.621004566210045, | |
| "grad_norm": 0.31906986236572266, | |
| "learning_rate": 4.0547945205479454e-05, | |
| "loss": 0.0361, | |
| "step": 1450 | |
| }, | |
| { | |
| "epoch": 6.666666666666667, | |
| "grad_norm": 0.23453353345394135, | |
| "learning_rate": 4.0482713633398566e-05, | |
| "loss": 0.0353, | |
| "step": 1460 | |
| }, | |
| { | |
| "epoch": 6.712328767123288, | |
| "grad_norm": 3.637096405029297, | |
| "learning_rate": 4.041748206131768e-05, | |
| "loss": 0.0412, | |
| "step": 1470 | |
| }, | |
| { | |
| "epoch": 6.757990867579909, | |
| "grad_norm": 0.21208681166172028, | |
| "learning_rate": 4.035225048923679e-05, | |
| "loss": 0.0354, | |
| "step": 1480 | |
| }, | |
| { | |
| "epoch": 6.80365296803653, | |
| "grad_norm": 0.29757997393608093, | |
| "learning_rate": 4.02870189171559e-05, | |
| "loss": 0.0372, | |
| "step": 1490 | |
| }, | |
| { | |
| "epoch": 6.8493150684931505, | |
| "grad_norm": 0.21638333797454834, | |
| "learning_rate": 4.0221787345075015e-05, | |
| "loss": 0.0343, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 6.894977168949771, | |
| "grad_norm": 0.24771864712238312, | |
| "learning_rate": 4.0156555772994134e-05, | |
| "loss": 0.0384, | |
| "step": 1510 | |
| }, | |
| { | |
| "epoch": 6.940639269406392, | |
| "grad_norm": 0.6341008543968201, | |
| "learning_rate": 4.0091324200913246e-05, | |
| "loss": 0.0457, | |
| "step": 1520 | |
| }, | |
| { | |
| "epoch": 6.986301369863014, | |
| "grad_norm": 0.2649524509906769, | |
| "learning_rate": 4.002609262883236e-05, | |
| "loss": 0.0362, | |
| "step": 1530 | |
| }, | |
| { | |
| "epoch": 7.0, | |
| "eval_bertscore_f1": 0.8794697234809835, | |
| "eval_bleu": 0.7000552060314711, | |
| "eval_loss": 0.0426379069685936, | |
| "eval_rougeL": 0.30292468350769075, | |
| "eval_runtime": 65.6416, | |
| "eval_samples_per_second": 22.897, | |
| "eval_steps_per_second": 1.432, | |
| "step": 1533 | |
| }, | |
| { | |
| "epoch": 7.031963470319635, | |
| "grad_norm": 0.2106972336769104, | |
| "learning_rate": 3.996086105675147e-05, | |
| "loss": 0.0376, | |
| "step": 1540 | |
| }, | |
| { | |
| "epoch": 7.077625570776256, | |
| "grad_norm": 0.19316372275352478, | |
| "learning_rate": 3.989562948467058e-05, | |
| "loss": 0.0357, | |
| "step": 1550 | |
| }, | |
| { | |
| "epoch": 7.123287671232877, | |
| "grad_norm": 0.2257992923259735, | |
| "learning_rate": 3.9830397912589695e-05, | |
| "loss": 0.037, | |
| "step": 1560 | |
| }, | |
| { | |
| "epoch": 7.168949771689498, | |
| "grad_norm": 0.2736322283744812, | |
| "learning_rate": 3.976516634050881e-05, | |
| "loss": 0.0424, | |
| "step": 1570 | |
| }, | |
| { | |
| "epoch": 7.2146118721461185, | |
| "grad_norm": 0.32095804810523987, | |
| "learning_rate": 3.969993476842792e-05, | |
| "loss": 0.0364, | |
| "step": 1580 | |
| }, | |
| { | |
| "epoch": 7.260273972602739, | |
| "grad_norm": 0.17005418241024017, | |
| "learning_rate": 3.963470319634703e-05, | |
| "loss": 0.0365, | |
| "step": 1590 | |
| }, | |
| { | |
| "epoch": 7.30593607305936, | |
| "grad_norm": 0.5590267777442932, | |
| "learning_rate": 3.9569471624266144e-05, | |
| "loss": 0.0345, | |
| "step": 1600 | |
| }, | |
| { | |
| "epoch": 7.351598173515982, | |
| "grad_norm": 0.3150683641433716, | |
| "learning_rate": 3.950424005218526e-05, | |
| "loss": 0.0382, | |
| "step": 1610 | |
| }, | |
| { | |
| "epoch": 7.397260273972603, | |
| "grad_norm": 0.22695161402225494, | |
| "learning_rate": 3.9439008480104375e-05, | |
| "loss": 0.036, | |
| "step": 1620 | |
| }, | |
| { | |
| "epoch": 7.442922374429224, | |
| "grad_norm": 0.428966760635376, | |
| "learning_rate": 3.937377690802349e-05, | |
| "loss": 0.0362, | |
| "step": 1630 | |
| }, | |
| { | |
| "epoch": 7.488584474885845, | |
| "grad_norm": 0.45022791624069214, | |
| "learning_rate": 3.93085453359426e-05, | |
| "loss": 0.0359, | |
| "step": 1640 | |
| }, | |
| { | |
| "epoch": 7.534246575342466, | |
| "grad_norm": 1.2701294422149658, | |
| "learning_rate": 3.924331376386171e-05, | |
| "loss": 0.0326, | |
| "step": 1650 | |
| }, | |
| { | |
| "epoch": 7.579908675799087, | |
| "grad_norm": 0.2580886483192444, | |
| "learning_rate": 3.9178082191780823e-05, | |
| "loss": 0.0333, | |
| "step": 1660 | |
| }, | |
| { | |
| "epoch": 7.6255707762557075, | |
| "grad_norm": 0.3360602855682373, | |
| "learning_rate": 3.9112850619699936e-05, | |
| "loss": 0.0339, | |
| "step": 1670 | |
| }, | |
| { | |
| "epoch": 7.671232876712329, | |
| "grad_norm": 0.39842474460601807, | |
| "learning_rate": 3.904761904761905e-05, | |
| "loss": 0.0378, | |
| "step": 1680 | |
| }, | |
| { | |
| "epoch": 7.71689497716895, | |
| "grad_norm": 0.8376930356025696, | |
| "learning_rate": 3.898238747553816e-05, | |
| "loss": 0.0389, | |
| "step": 1690 | |
| }, | |
| { | |
| "epoch": 7.762557077625571, | |
| "grad_norm": 0.23643815517425537, | |
| "learning_rate": 3.891715590345727e-05, | |
| "loss": 0.038, | |
| "step": 1700 | |
| }, | |
| { | |
| "epoch": 7.808219178082192, | |
| "grad_norm": 0.26686662435531616, | |
| "learning_rate": 3.885192433137639e-05, | |
| "loss": 0.0339, | |
| "step": 1710 | |
| }, | |
| { | |
| "epoch": 7.853881278538813, | |
| "grad_norm": 0.16813237965106964, | |
| "learning_rate": 3.87866927592955e-05, | |
| "loss": 0.0328, | |
| "step": 1720 | |
| }, | |
| { | |
| "epoch": 7.899543378995434, | |
| "grad_norm": 0.28077274560928345, | |
| "learning_rate": 3.8721461187214615e-05, | |
| "loss": 0.0326, | |
| "step": 1730 | |
| }, | |
| { | |
| "epoch": 7.945205479452055, | |
| "grad_norm": 0.3392268717288971, | |
| "learning_rate": 3.865622961513373e-05, | |
| "loss": 0.0416, | |
| "step": 1740 | |
| }, | |
| { | |
| "epoch": 7.9908675799086755, | |
| "grad_norm": 0.24137546122074127, | |
| "learning_rate": 3.859099804305284e-05, | |
| "loss": 0.036, | |
| "step": 1750 | |
| }, | |
| { | |
| "epoch": 8.0, | |
| "eval_bertscore_f1": 0.8751992449630361, | |
| "eval_bleu": 0.7231744684160546, | |
| "eval_loss": 0.042749155312776566, | |
| "eval_rougeL": 0.2994421875548862, | |
| "eval_runtime": 65.5673, | |
| "eval_samples_per_second": 22.923, | |
| "eval_steps_per_second": 1.434, | |
| "step": 1752 | |
| }, | |
| { | |
| "epoch": 8.036529680365296, | |
| "grad_norm": 0.3088216185569763, | |
| "learning_rate": 3.852576647097195e-05, | |
| "loss": 0.0347, | |
| "step": 1760 | |
| }, | |
| { | |
| "epoch": 8.082191780821917, | |
| "grad_norm": 0.3175930678844452, | |
| "learning_rate": 3.8460534898891064e-05, | |
| "loss": 0.0348, | |
| "step": 1770 | |
| }, | |
| { | |
| "epoch": 8.127853881278538, | |
| "grad_norm": 0.4923621416091919, | |
| "learning_rate": 3.8395303326810176e-05, | |
| "loss": 0.0354, | |
| "step": 1780 | |
| }, | |
| { | |
| "epoch": 8.173515981735159, | |
| "grad_norm": 0.23712977766990662, | |
| "learning_rate": 3.833007175472929e-05, | |
| "loss": 0.039, | |
| "step": 1790 | |
| }, | |
| { | |
| "epoch": 8.219178082191782, | |
| "grad_norm": 0.24011415243148804, | |
| "learning_rate": 3.82648401826484e-05, | |
| "loss": 0.036, | |
| "step": 1800 | |
| }, | |
| { | |
| "epoch": 8.264840182648403, | |
| "grad_norm": 0.23044519126415253, | |
| "learning_rate": 3.819960861056752e-05, | |
| "loss": 0.0317, | |
| "step": 1810 | |
| }, | |
| { | |
| "epoch": 8.310502283105023, | |
| "grad_norm": 0.25046950578689575, | |
| "learning_rate": 3.813437703848663e-05, | |
| "loss": 0.0361, | |
| "step": 1820 | |
| }, | |
| { | |
| "epoch": 8.356164383561644, | |
| "grad_norm": 0.2792360782623291, | |
| "learning_rate": 3.8069145466405744e-05, | |
| "loss": 0.0372, | |
| "step": 1830 | |
| }, | |
| { | |
| "epoch": 8.401826484018265, | |
| "grad_norm": 0.7518793940544128, | |
| "learning_rate": 3.8003913894324856e-05, | |
| "loss": 0.0372, | |
| "step": 1840 | |
| }, | |
| { | |
| "epoch": 8.447488584474886, | |
| "grad_norm": 0.2685680389404297, | |
| "learning_rate": 3.793868232224397e-05, | |
| "loss": 0.0338, | |
| "step": 1850 | |
| }, | |
| { | |
| "epoch": 8.493150684931507, | |
| "grad_norm": 0.19953973591327667, | |
| "learning_rate": 3.787345075016308e-05, | |
| "loss": 0.0342, | |
| "step": 1860 | |
| }, | |
| { | |
| "epoch": 8.538812785388128, | |
| "grad_norm": 0.2889842987060547, | |
| "learning_rate": 3.780821917808219e-05, | |
| "loss": 0.0347, | |
| "step": 1870 | |
| }, | |
| { | |
| "epoch": 8.584474885844749, | |
| "grad_norm": 0.2976756691932678, | |
| "learning_rate": 3.7742987606001305e-05, | |
| "loss": 0.0357, | |
| "step": 1880 | |
| }, | |
| { | |
| "epoch": 8.63013698630137, | |
| "grad_norm": 0.2278045117855072, | |
| "learning_rate": 3.767775603392042e-05, | |
| "loss": 0.0367, | |
| "step": 1890 | |
| }, | |
| { | |
| "epoch": 8.67579908675799, | |
| "grad_norm": 0.14058364927768707, | |
| "learning_rate": 3.761252446183953e-05, | |
| "loss": 0.0323, | |
| "step": 1900 | |
| }, | |
| { | |
| "epoch": 8.721461187214611, | |
| "grad_norm": 0.28625041246414185, | |
| "learning_rate": 3.754729288975865e-05, | |
| "loss": 0.035, | |
| "step": 1910 | |
| }, | |
| { | |
| "epoch": 8.767123287671232, | |
| "grad_norm": 0.2872960567474365, | |
| "learning_rate": 3.748206131767776e-05, | |
| "loss": 0.0402, | |
| "step": 1920 | |
| }, | |
| { | |
| "epoch": 8.812785388127853, | |
| "grad_norm": 0.34051889181137085, | |
| "learning_rate": 3.741682974559687e-05, | |
| "loss": 0.0342, | |
| "step": 1930 | |
| }, | |
| { | |
| "epoch": 8.858447488584474, | |
| "grad_norm": 0.23217986524105072, | |
| "learning_rate": 3.7351598173515985e-05, | |
| "loss": 0.0375, | |
| "step": 1940 | |
| }, | |
| { | |
| "epoch": 8.904109589041095, | |
| "grad_norm": 0.38106223940849304, | |
| "learning_rate": 3.72863666014351e-05, | |
| "loss": 0.0363, | |
| "step": 1950 | |
| }, | |
| { | |
| "epoch": 8.949771689497716, | |
| "grad_norm": 0.17972268164157867, | |
| "learning_rate": 3.722113502935421e-05, | |
| "loss": 0.0367, | |
| "step": 1960 | |
| }, | |
| { | |
| "epoch": 8.995433789954339, | |
| "grad_norm": 0.27785488963127136, | |
| "learning_rate": 3.715590345727332e-05, | |
| "loss": 0.0341, | |
| "step": 1970 | |
| }, | |
| { | |
| "epoch": 9.0, | |
| "eval_bertscore_f1": 0.8794770974836902, | |
| "eval_bleu": 0.701948780010586, | |
| "eval_loss": 0.043190885335206985, | |
| "eval_rougeL": 0.30334992897033064, | |
| "eval_runtime": 64.8862, | |
| "eval_samples_per_second": 23.164, | |
| "eval_steps_per_second": 1.449, | |
| "step": 1971 | |
| }, | |
| { | |
| "epoch": 9.04109589041096, | |
| "grad_norm": 0.2479076385498047, | |
| "learning_rate": 3.709067188519243e-05, | |
| "loss": 0.0314, | |
| "step": 1980 | |
| }, | |
| { | |
| "epoch": 9.08675799086758, | |
| "grad_norm": 0.3738096058368683, | |
| "learning_rate": 3.7025440313111545e-05, | |
| "loss": 0.0381, | |
| "step": 1990 | |
| }, | |
| { | |
| "epoch": 9.132420091324201, | |
| "grad_norm": 0.2466166913509369, | |
| "learning_rate": 3.696020874103066e-05, | |
| "loss": 0.0322, | |
| "step": 2000 | |
| }, | |
| { | |
| "epoch": 9.178082191780822, | |
| "grad_norm": 0.5537144541740417, | |
| "learning_rate": 3.6894977168949777e-05, | |
| "loss": 0.0362, | |
| "step": 2010 | |
| }, | |
| { | |
| "epoch": 9.223744292237443, | |
| "grad_norm": 0.711529552936554, | |
| "learning_rate": 3.682974559686889e-05, | |
| "loss": 0.0373, | |
| "step": 2020 | |
| }, | |
| { | |
| "epoch": 9.269406392694064, | |
| "grad_norm": 0.15315942466259003, | |
| "learning_rate": 3.6764514024788e-05, | |
| "loss": 0.0375, | |
| "step": 2030 | |
| }, | |
| { | |
| "epoch": 9.315068493150685, | |
| "grad_norm": 0.2318936139345169, | |
| "learning_rate": 3.669928245270711e-05, | |
| "loss": 0.0337, | |
| "step": 2040 | |
| }, | |
| { | |
| "epoch": 9.360730593607306, | |
| "grad_norm": 0.15388214588165283, | |
| "learning_rate": 3.6634050880626225e-05, | |
| "loss": 0.0363, | |
| "step": 2050 | |
| }, | |
| { | |
| "epoch": 9.406392694063927, | |
| "grad_norm": 0.2665681540966034, | |
| "learning_rate": 3.656881930854534e-05, | |
| "loss": 0.0321, | |
| "step": 2060 | |
| }, | |
| { | |
| "epoch": 9.452054794520548, | |
| "grad_norm": 0.3937646150588989, | |
| "learning_rate": 3.650358773646445e-05, | |
| "loss": 0.0372, | |
| "step": 2070 | |
| }, | |
| { | |
| "epoch": 9.497716894977168, | |
| "grad_norm": 0.6450703144073486, | |
| "learning_rate": 3.643835616438356e-05, | |
| "loss": 0.0327, | |
| "step": 2080 | |
| }, | |
| { | |
| "epoch": 9.54337899543379, | |
| "grad_norm": 0.8227304816246033, | |
| "learning_rate": 3.6373124592302674e-05, | |
| "loss": 0.0328, | |
| "step": 2090 | |
| }, | |
| { | |
| "epoch": 9.58904109589041, | |
| "grad_norm": 0.5075876712799072, | |
| "learning_rate": 3.6307893020221786e-05, | |
| "loss": 0.0344, | |
| "step": 2100 | |
| }, | |
| { | |
| "epoch": 9.634703196347033, | |
| "grad_norm": 0.25154736638069153, | |
| "learning_rate": 3.6242661448140905e-05, | |
| "loss": 0.0345, | |
| "step": 2110 | |
| }, | |
| { | |
| "epoch": 9.680365296803654, | |
| "grad_norm": 0.4204483926296234, | |
| "learning_rate": 3.617742987606002e-05, | |
| "loss": 0.0342, | |
| "step": 2120 | |
| }, | |
| { | |
| "epoch": 9.726027397260275, | |
| "grad_norm": 0.3545476198196411, | |
| "learning_rate": 3.611219830397913e-05, | |
| "loss": 0.0358, | |
| "step": 2130 | |
| }, | |
| { | |
| "epoch": 9.771689497716896, | |
| "grad_norm": 0.26954710483551025, | |
| "learning_rate": 3.604696673189824e-05, | |
| "loss": 0.0375, | |
| "step": 2140 | |
| }, | |
| { | |
| "epoch": 9.817351598173516, | |
| "grad_norm": 0.1699344962835312, | |
| "learning_rate": 3.5981735159817354e-05, | |
| "loss": 0.0338, | |
| "step": 2150 | |
| }, | |
| { | |
| "epoch": 9.863013698630137, | |
| "grad_norm": 0.24426551163196564, | |
| "learning_rate": 3.5916503587736466e-05, | |
| "loss": 0.0366, | |
| "step": 2160 | |
| }, | |
| { | |
| "epoch": 9.908675799086758, | |
| "grad_norm": 0.17183125019073486, | |
| "learning_rate": 3.585127201565558e-05, | |
| "loss": 0.0316, | |
| "step": 2170 | |
| }, | |
| { | |
| "epoch": 9.954337899543379, | |
| "grad_norm": 0.21993212401866913, | |
| "learning_rate": 3.578604044357469e-05, | |
| "loss": 0.0314, | |
| "step": 2180 | |
| }, | |
| { | |
| "epoch": 10.0, | |
| "grad_norm": 0.24478712677955627, | |
| "learning_rate": 3.57208088714938e-05, | |
| "loss": 0.0359, | |
| "step": 2190 | |
| }, | |
| { | |
| "epoch": 10.0, | |
| "eval_bertscore_f1": 0.8741392707792981, | |
| "eval_bleu": 0.5581555403302182, | |
| "eval_loss": 0.041961919516325, | |
| "eval_rougeL": 0.3069551250109942, | |
| "eval_runtime": 65.0323, | |
| "eval_samples_per_second": 23.112, | |
| "eval_steps_per_second": 1.445, | |
| "step": 2190 | |
| }, | |
| { | |
| "epoch": 10.045662100456621, | |
| "grad_norm": 0.3946993947029114, | |
| "learning_rate": 3.5655577299412915e-05, | |
| "loss": 0.0349, | |
| "step": 2200 | |
| }, | |
| { | |
| "epoch": 10.091324200913242, | |
| "grad_norm": 0.15859711170196533, | |
| "learning_rate": 3.5590345727332033e-05, | |
| "loss": 0.0322, | |
| "step": 2210 | |
| }, | |
| { | |
| "epoch": 10.136986301369863, | |
| "grad_norm": 0.766708493232727, | |
| "learning_rate": 3.5525114155251146e-05, | |
| "loss": 0.0343, | |
| "step": 2220 | |
| }, | |
| { | |
| "epoch": 10.182648401826484, | |
| "grad_norm": 0.17286263406276703, | |
| "learning_rate": 3.545988258317026e-05, | |
| "loss": 0.0341, | |
| "step": 2230 | |
| }, | |
| { | |
| "epoch": 10.228310502283104, | |
| "grad_norm": 0.2328823357820511, | |
| "learning_rate": 3.539465101108937e-05, | |
| "loss": 0.0343, | |
| "step": 2240 | |
| }, | |
| { | |
| "epoch": 10.273972602739725, | |
| "grad_norm": 0.23741354048252106, | |
| "learning_rate": 3.532941943900848e-05, | |
| "loss": 0.0334, | |
| "step": 2250 | |
| }, | |
| { | |
| "epoch": 10.319634703196346, | |
| "grad_norm": 0.23443607985973358, | |
| "learning_rate": 3.5264187866927594e-05, | |
| "loss": 0.0317, | |
| "step": 2260 | |
| }, | |
| { | |
| "epoch": 10.365296803652967, | |
| "grad_norm": 0.29927483201026917, | |
| "learning_rate": 3.5198956294846707e-05, | |
| "loss": 0.0334, | |
| "step": 2270 | |
| }, | |
| { | |
| "epoch": 10.41095890410959, | |
| "grad_norm": 0.5162585377693176, | |
| "learning_rate": 3.513372472276582e-05, | |
| "loss": 0.0368, | |
| "step": 2280 | |
| }, | |
| { | |
| "epoch": 10.45662100456621, | |
| "grad_norm": 0.20458880066871643, | |
| "learning_rate": 3.506849315068493e-05, | |
| "loss": 0.0333, | |
| "step": 2290 | |
| }, | |
| { | |
| "epoch": 10.502283105022832, | |
| "grad_norm": 0.6160792708396912, | |
| "learning_rate": 3.500326157860404e-05, | |
| "loss": 0.0358, | |
| "step": 2300 | |
| }, | |
| { | |
| "epoch": 10.547945205479452, | |
| "grad_norm": 0.34151145815849304, | |
| "learning_rate": 3.493803000652316e-05, | |
| "loss": 0.0332, | |
| "step": 2310 | |
| }, | |
| { | |
| "epoch": 10.593607305936073, | |
| "grad_norm": 0.2050519436597824, | |
| "learning_rate": 3.4872798434442274e-05, | |
| "loss": 0.0333, | |
| "step": 2320 | |
| }, | |
| { | |
| "epoch": 10.639269406392694, | |
| "grad_norm": 0.33147433400154114, | |
| "learning_rate": 3.4807566862361386e-05, | |
| "loss": 0.034, | |
| "step": 2330 | |
| }, | |
| { | |
| "epoch": 10.684931506849315, | |
| "grad_norm": 0.20570585131645203, | |
| "learning_rate": 3.47423352902805e-05, | |
| "loss": 0.0315, | |
| "step": 2340 | |
| }, | |
| { | |
| "epoch": 10.730593607305936, | |
| "grad_norm": 0.21926531195640564, | |
| "learning_rate": 3.467710371819961e-05, | |
| "loss": 0.0335, | |
| "step": 2350 | |
| }, | |
| { | |
| "epoch": 10.776255707762557, | |
| "grad_norm": 0.20111322402954102, | |
| "learning_rate": 3.461187214611872e-05, | |
| "loss": 0.038, | |
| "step": 2360 | |
| }, | |
| { | |
| "epoch": 10.821917808219178, | |
| "grad_norm": 0.37514954805374146, | |
| "learning_rate": 3.4546640574037835e-05, | |
| "loss": 0.0346, | |
| "step": 2370 | |
| }, | |
| { | |
| "epoch": 10.867579908675799, | |
| "grad_norm": 0.19650527834892273, | |
| "learning_rate": 3.448140900195695e-05, | |
| "loss": 0.0323, | |
| "step": 2380 | |
| }, | |
| { | |
| "epoch": 10.91324200913242, | |
| "grad_norm": 0.3676673471927643, | |
| "learning_rate": 3.441617742987606e-05, | |
| "loss": 0.0322, | |
| "step": 2390 | |
| }, | |
| { | |
| "epoch": 10.95890410958904, | |
| "grad_norm": 0.2045547366142273, | |
| "learning_rate": 3.435094585779517e-05, | |
| "loss": 0.036, | |
| "step": 2400 | |
| }, | |
| { | |
| "epoch": 11.0, | |
| "eval_bertscore_f1": 0.8752922487909288, | |
| "eval_bleu": 0.9581851182952255, | |
| "eval_loss": 0.0423484668135643, | |
| "eval_rougeL": 0.28949663545817383, | |
| "eval_runtime": 64.9513, | |
| "eval_samples_per_second": 23.14, | |
| "eval_steps_per_second": 1.447, | |
| "step": 2409 | |
| }, | |
| { | |
| "epoch": 11.004566210045661, | |
| "grad_norm": 0.21690687537193298, | |
| "learning_rate": 3.428571428571429e-05, | |
| "loss": 0.0348, | |
| "step": 2410 | |
| }, | |
| { | |
| "epoch": 11.050228310502282, | |
| "grad_norm": 0.5675901174545288, | |
| "learning_rate": 3.42204827136334e-05, | |
| "loss": 0.0339, | |
| "step": 2420 | |
| }, | |
| { | |
| "epoch": 11.095890410958905, | |
| "grad_norm": 0.18212980031967163, | |
| "learning_rate": 3.4155251141552515e-05, | |
| "loss": 0.0361, | |
| "step": 2430 | |
| }, | |
| { | |
| "epoch": 11.141552511415526, | |
| "grad_norm": 0.23558968305587769, | |
| "learning_rate": 3.409001956947163e-05, | |
| "loss": 0.0376, | |
| "step": 2440 | |
| }, | |
| { | |
| "epoch": 11.187214611872147, | |
| "grad_norm": 0.1530950367450714, | |
| "learning_rate": 3.402478799739074e-05, | |
| "loss": 0.0382, | |
| "step": 2450 | |
| }, | |
| { | |
| "epoch": 11.232876712328768, | |
| "grad_norm": 0.2215454876422882, | |
| "learning_rate": 3.395955642530985e-05, | |
| "loss": 0.0335, | |
| "step": 2460 | |
| }, | |
| { | |
| "epoch": 11.278538812785389, | |
| "grad_norm": 0.31875351071357727, | |
| "learning_rate": 3.3894324853228963e-05, | |
| "loss": 0.0336, | |
| "step": 2470 | |
| }, | |
| { | |
| "epoch": 11.32420091324201, | |
| "grad_norm": 0.18925060331821442, | |
| "learning_rate": 3.3829093281148076e-05, | |
| "loss": 0.033, | |
| "step": 2480 | |
| }, | |
| { | |
| "epoch": 11.36986301369863, | |
| "grad_norm": 0.1896040290594101, | |
| "learning_rate": 3.376386170906719e-05, | |
| "loss": 0.0366, | |
| "step": 2490 | |
| }, | |
| { | |
| "epoch": 11.415525114155251, | |
| "grad_norm": 0.18136650323867798, | |
| "learning_rate": 3.36986301369863e-05, | |
| "loss": 0.0325, | |
| "step": 2500 | |
| }, | |
| { | |
| "epoch": 11.461187214611872, | |
| "grad_norm": 0.1443028599023819, | |
| "learning_rate": 3.363339856490542e-05, | |
| "loss": 0.0342, | |
| "step": 2510 | |
| }, | |
| { | |
| "epoch": 11.506849315068493, | |
| "grad_norm": 0.3814453184604645, | |
| "learning_rate": 3.356816699282453e-05, | |
| "loss": 0.0342, | |
| "step": 2520 | |
| }, | |
| { | |
| "epoch": 11.552511415525114, | |
| "grad_norm": 0.19911924004554749, | |
| "learning_rate": 3.350293542074364e-05, | |
| "loss": 0.034, | |
| "step": 2530 | |
| }, | |
| { | |
| "epoch": 11.598173515981735, | |
| "grad_norm": 0.16410912573337555, | |
| "learning_rate": 3.3437703848662755e-05, | |
| "loss": 0.034, | |
| "step": 2540 | |
| }, | |
| { | |
| "epoch": 11.643835616438356, | |
| "grad_norm": 0.18103031814098358, | |
| "learning_rate": 3.337247227658187e-05, | |
| "loss": 0.0309, | |
| "step": 2550 | |
| }, | |
| { | |
| "epoch": 11.689497716894977, | |
| "grad_norm": 0.25762510299682617, | |
| "learning_rate": 3.330724070450098e-05, | |
| "loss": 0.0325, | |
| "step": 2560 | |
| }, | |
| { | |
| "epoch": 11.735159817351597, | |
| "grad_norm": 0.20280788838863373, | |
| "learning_rate": 3.324200913242009e-05, | |
| "loss": 0.0326, | |
| "step": 2570 | |
| }, | |
| { | |
| "epoch": 11.780821917808218, | |
| "grad_norm": 0.2774021029472351, | |
| "learning_rate": 3.3176777560339204e-05, | |
| "loss": 0.0331, | |
| "step": 2580 | |
| }, | |
| { | |
| "epoch": 11.826484018264841, | |
| "grad_norm": 0.187673881649971, | |
| "learning_rate": 3.3111545988258316e-05, | |
| "loss": 0.0342, | |
| "step": 2590 | |
| }, | |
| { | |
| "epoch": 11.872146118721462, | |
| "grad_norm": 0.1484459787607193, | |
| "learning_rate": 3.304631441617743e-05, | |
| "loss": 0.0312, | |
| "step": 2600 | |
| }, | |
| { | |
| "epoch": 11.917808219178083, | |
| "grad_norm": 0.25383421778678894, | |
| "learning_rate": 3.298108284409655e-05, | |
| "loss": 0.0331, | |
| "step": 2610 | |
| }, | |
| { | |
| "epoch": 11.963470319634704, | |
| "grad_norm": 0.1816657930612564, | |
| "learning_rate": 3.291585127201566e-05, | |
| "loss": 0.0324, | |
| "step": 2620 | |
| }, | |
| { | |
| "epoch": 12.0, | |
| "eval_bertscore_f1": 0.8740238581430254, | |
| "eval_bleu": 0.9457621299589393, | |
| "eval_loss": 0.04242411255836487, | |
| "eval_rougeL": 0.2864058958363045, | |
| "eval_runtime": 65.3663, | |
| "eval_samples_per_second": 22.994, | |
| "eval_steps_per_second": 1.438, | |
| "step": 2628 | |
| }, | |
| { | |
| "epoch": 12.009132420091325, | |
| "grad_norm": 0.22657038271427155, | |
| "learning_rate": 3.285061969993477e-05, | |
| "loss": 0.0328, | |
| "step": 2630 | |
| }, | |
| { | |
| "epoch": 12.054794520547945, | |
| "grad_norm": 0.36884865164756775, | |
| "learning_rate": 3.2785388127853884e-05, | |
| "loss": 0.0344, | |
| "step": 2640 | |
| }, | |
| { | |
| "epoch": 12.100456621004566, | |
| "grad_norm": 0.24654433131217957, | |
| "learning_rate": 3.2720156555772996e-05, | |
| "loss": 0.0335, | |
| "step": 2650 | |
| }, | |
| { | |
| "epoch": 12.146118721461187, | |
| "grad_norm": 0.21425046026706696, | |
| "learning_rate": 3.265492498369211e-05, | |
| "loss": 0.0353, | |
| "step": 2660 | |
| }, | |
| { | |
| "epoch": 12.191780821917808, | |
| "grad_norm": 0.1591794192790985, | |
| "learning_rate": 3.258969341161122e-05, | |
| "loss": 0.03, | |
| "step": 2670 | |
| }, | |
| { | |
| "epoch": 12.237442922374429, | |
| "grad_norm": 0.16401417553424835, | |
| "learning_rate": 3.252446183953033e-05, | |
| "loss": 0.0319, | |
| "step": 2680 | |
| }, | |
| { | |
| "epoch": 12.28310502283105, | |
| "grad_norm": 0.37971609830856323, | |
| "learning_rate": 3.2459230267449445e-05, | |
| "loss": 0.0318, | |
| "step": 2690 | |
| }, | |
| { | |
| "epoch": 12.32876712328767, | |
| "grad_norm": 0.17105679214000702, | |
| "learning_rate": 3.239399869536856e-05, | |
| "loss": 0.0329, | |
| "step": 2700 | |
| }, | |
| { | |
| "epoch": 12.374429223744292, | |
| "grad_norm": 0.1671990305185318, | |
| "learning_rate": 3.2328767123287676e-05, | |
| "loss": 0.0321, | |
| "step": 2710 | |
| }, | |
| { | |
| "epoch": 12.420091324200913, | |
| "grad_norm": 0.18382422626018524, | |
| "learning_rate": 3.226353555120679e-05, | |
| "loss": 0.0344, | |
| "step": 2720 | |
| }, | |
| { | |
| "epoch": 12.465753424657533, | |
| "grad_norm": 0.4917747378349304, | |
| "learning_rate": 3.21983039791259e-05, | |
| "loss": 0.0368, | |
| "step": 2730 | |
| }, | |
| { | |
| "epoch": 12.511415525114156, | |
| "grad_norm": 0.274912565946579, | |
| "learning_rate": 3.213307240704501e-05, | |
| "loss": 0.0341, | |
| "step": 2740 | |
| }, | |
| { | |
| "epoch": 12.557077625570777, | |
| "grad_norm": 0.34153324365615845, | |
| "learning_rate": 3.2067840834964125e-05, | |
| "loss": 0.0343, | |
| "step": 2750 | |
| }, | |
| { | |
| "epoch": 12.602739726027398, | |
| "grad_norm": 0.5951138138771057, | |
| "learning_rate": 3.200260926288324e-05, | |
| "loss": 0.0314, | |
| "step": 2760 | |
| }, | |
| { | |
| "epoch": 12.648401826484019, | |
| "grad_norm": 0.1907246708869934, | |
| "learning_rate": 3.193737769080235e-05, | |
| "loss": 0.0298, | |
| "step": 2770 | |
| }, | |
| { | |
| "epoch": 12.69406392694064, | |
| "grad_norm": 0.2845422625541687, | |
| "learning_rate": 3.187214611872146e-05, | |
| "loss": 0.0315, | |
| "step": 2780 | |
| }, | |
| { | |
| "epoch": 12.73972602739726, | |
| "grad_norm": 0.5828273892402649, | |
| "learning_rate": 3.180691454664057e-05, | |
| "loss": 0.0339, | |
| "step": 2790 | |
| }, | |
| { | |
| "epoch": 12.785388127853881, | |
| "grad_norm": 0.12505371868610382, | |
| "learning_rate": 3.174168297455969e-05, | |
| "loss": 0.0329, | |
| "step": 2800 | |
| }, | |
| { | |
| "epoch": 12.831050228310502, | |
| "grad_norm": 0.18723076581954956, | |
| "learning_rate": 3.1676451402478804e-05, | |
| "loss": 0.0296, | |
| "step": 2810 | |
| }, | |
| { | |
| "epoch": 12.876712328767123, | |
| "grad_norm": 0.20346902310848236, | |
| "learning_rate": 3.1611219830397917e-05, | |
| "loss": 0.034, | |
| "step": 2820 | |
| }, | |
| { | |
| "epoch": 12.922374429223744, | |
| "grad_norm": 0.21218132972717285, | |
| "learning_rate": 3.154598825831703e-05, | |
| "loss": 0.0308, | |
| "step": 2830 | |
| }, | |
| { | |
| "epoch": 12.968036529680365, | |
| "grad_norm": 0.2717015743255615, | |
| "learning_rate": 3.148075668623614e-05, | |
| "loss": 0.0331, | |
| "step": 2840 | |
| }, | |
| { | |
| "epoch": 13.0, | |
| "eval_bertscore_f1": 0.8743115136088805, | |
| "eval_bleu": 0.5546674313556451, | |
| "eval_loss": 0.041579220443964005, | |
| "eval_rougeL": 0.30863203533209793, | |
| "eval_runtime": 65.4384, | |
| "eval_samples_per_second": 22.968, | |
| "eval_steps_per_second": 1.436, | |
| "step": 2847 | |
| }, | |
| { | |
| "epoch": 13.013698630136986, | |
| "grad_norm": 0.45697784423828125, | |
| "learning_rate": 3.141552511415525e-05, | |
| "loss": 0.034, | |
| "step": 2850 | |
| }, | |
| { | |
| "epoch": 13.059360730593607, | |
| "grad_norm": 0.2796197235584259, | |
| "learning_rate": 3.1350293542074365e-05, | |
| "loss": 0.0349, | |
| "step": 2860 | |
| }, | |
| { | |
| "epoch": 13.105022831050228, | |
| "grad_norm": 0.2153574377298355, | |
| "learning_rate": 3.128506196999348e-05, | |
| "loss": 0.0323, | |
| "step": 2870 | |
| }, | |
| { | |
| "epoch": 13.150684931506849, | |
| "grad_norm": 0.19059832394123077, | |
| "learning_rate": 3.121983039791259e-05, | |
| "loss": 0.0342, | |
| "step": 2880 | |
| }, | |
| { | |
| "epoch": 13.19634703196347, | |
| "grad_norm": 0.2336965799331665, | |
| "learning_rate": 3.11545988258317e-05, | |
| "loss": 0.0335, | |
| "step": 2890 | |
| }, | |
| { | |
| "epoch": 13.242009132420092, | |
| "grad_norm": 0.2201681286096573, | |
| "learning_rate": 3.108936725375082e-05, | |
| "loss": 0.0324, | |
| "step": 2900 | |
| }, | |
| { | |
| "epoch": 13.287671232876713, | |
| "grad_norm": 0.4124416708946228, | |
| "learning_rate": 3.102413568166993e-05, | |
| "loss": 0.0332, | |
| "step": 2910 | |
| }, | |
| { | |
| "epoch": 13.333333333333334, | |
| "grad_norm": 0.24998348951339722, | |
| "learning_rate": 3.0958904109589045e-05, | |
| "loss": 0.0318, | |
| "step": 2920 | |
| }, | |
| { | |
| "epoch": 13.378995433789955, | |
| "grad_norm": 0.22465960681438446, | |
| "learning_rate": 3.089367253750816e-05, | |
| "loss": 0.0335, | |
| "step": 2930 | |
| }, | |
| { | |
| "epoch": 13.424657534246576, | |
| "grad_norm": 0.21678076684474945, | |
| "learning_rate": 3.082844096542727e-05, | |
| "loss": 0.0339, | |
| "step": 2940 | |
| }, | |
| { | |
| "epoch": 13.470319634703197, | |
| "grad_norm": 0.1709524393081665, | |
| "learning_rate": 3.076320939334638e-05, | |
| "loss": 0.0331, | |
| "step": 2950 | |
| }, | |
| { | |
| "epoch": 13.515981735159817, | |
| "grad_norm": 0.1303175985813141, | |
| "learning_rate": 3.0697977821265494e-05, | |
| "loss": 0.0288, | |
| "step": 2960 | |
| }, | |
| { | |
| "epoch": 13.561643835616438, | |
| "grad_norm": 0.2762468159198761, | |
| "learning_rate": 3.0632746249184606e-05, | |
| "loss": 0.0303, | |
| "step": 2970 | |
| }, | |
| { | |
| "epoch": 13.60730593607306, | |
| "grad_norm": 0.12408158928155899, | |
| "learning_rate": 3.056751467710372e-05, | |
| "loss": 0.0311, | |
| "step": 2980 | |
| }, | |
| { | |
| "epoch": 13.65296803652968, | |
| "grad_norm": 0.1823064684867859, | |
| "learning_rate": 3.0502283105022834e-05, | |
| "loss": 0.0294, | |
| "step": 2990 | |
| }, | |
| { | |
| "epoch": 13.698630136986301, | |
| "grad_norm": 0.23568011820316315, | |
| "learning_rate": 3.0437051532941946e-05, | |
| "loss": 0.0329, | |
| "step": 3000 | |
| }, | |
| { | |
| "epoch": 13.744292237442922, | |
| "grad_norm": 0.19115784764289856, | |
| "learning_rate": 3.0371819960861058e-05, | |
| "loss": 0.0317, | |
| "step": 3010 | |
| }, | |
| { | |
| "epoch": 13.789954337899543, | |
| "grad_norm": 0.149480938911438, | |
| "learning_rate": 3.030658838878017e-05, | |
| "loss": 0.0332, | |
| "step": 3020 | |
| }, | |
| { | |
| "epoch": 13.835616438356164, | |
| "grad_norm": 0.21910887956619263, | |
| "learning_rate": 3.0241356816699286e-05, | |
| "loss": 0.0352, | |
| "step": 3030 | |
| }, | |
| { | |
| "epoch": 13.881278538812785, | |
| "grad_norm": 0.20704659819602966, | |
| "learning_rate": 3.0176125244618398e-05, | |
| "loss": 0.0312, | |
| "step": 3040 | |
| }, | |
| { | |
| "epoch": 13.926940639269407, | |
| "grad_norm": 0.15948213636875153, | |
| "learning_rate": 3.011089367253751e-05, | |
| "loss": 0.0317, | |
| "step": 3050 | |
| }, | |
| { | |
| "epoch": 13.972602739726028, | |
| "grad_norm": 0.2612301707267761, | |
| "learning_rate": 3.0045662100456622e-05, | |
| "loss": 0.0337, | |
| "step": 3060 | |
| }, | |
| { | |
| "epoch": 14.0, | |
| "eval_bertscore_f1": 0.8789519665246001, | |
| "eval_bleu": 0.7617727244561832, | |
| "eval_loss": 0.042337022721767426, | |
| "eval_rougeL": 0.30510897093069805, | |
| "eval_runtime": 65.1106, | |
| "eval_samples_per_second": 23.084, | |
| "eval_steps_per_second": 1.444, | |
| "step": 3066 | |
| }, | |
| { | |
| "epoch": 14.018264840182649, | |
| "grad_norm": 0.1484575718641281, | |
| "learning_rate": 2.9980430528375734e-05, | |
| "loss": 0.0331, | |
| "step": 3070 | |
| }, | |
| { | |
| "epoch": 14.06392694063927, | |
| "grad_norm": 0.2205040156841278, | |
| "learning_rate": 2.991519895629485e-05, | |
| "loss": 0.0318, | |
| "step": 3080 | |
| }, | |
| { | |
| "epoch": 14.10958904109589, | |
| "grad_norm": 0.18296614289283752, | |
| "learning_rate": 2.9849967384213962e-05, | |
| "loss": 0.0315, | |
| "step": 3090 | |
| }, | |
| { | |
| "epoch": 14.155251141552512, | |
| "grad_norm": 0.38665133714675903, | |
| "learning_rate": 2.9784735812133074e-05, | |
| "loss": 0.0329, | |
| "step": 3100 | |
| }, | |
| { | |
| "epoch": 14.200913242009133, | |
| "grad_norm": 0.21823015809059143, | |
| "learning_rate": 2.9719504240052186e-05, | |
| "loss": 0.0319, | |
| "step": 3110 | |
| }, | |
| { | |
| "epoch": 14.246575342465754, | |
| "grad_norm": 0.16656526923179626, | |
| "learning_rate": 2.96542726679713e-05, | |
| "loss": 0.0349, | |
| "step": 3120 | |
| }, | |
| { | |
| "epoch": 14.292237442922374, | |
| "grad_norm": 0.15548262000083923, | |
| "learning_rate": 2.9589041095890414e-05, | |
| "loss": 0.0322, | |
| "step": 3130 | |
| }, | |
| { | |
| "epoch": 14.337899543378995, | |
| "grad_norm": 0.15261350572109222, | |
| "learning_rate": 2.9523809523809526e-05, | |
| "loss": 0.0327, | |
| "step": 3140 | |
| }, | |
| { | |
| "epoch": 14.383561643835616, | |
| "grad_norm": 0.37383410334587097, | |
| "learning_rate": 2.945857795172864e-05, | |
| "loss": 0.0305, | |
| "step": 3150 | |
| }, | |
| { | |
| "epoch": 14.429223744292237, | |
| "grad_norm": 0.25296688079833984, | |
| "learning_rate": 2.939334637964775e-05, | |
| "loss": 0.0331, | |
| "step": 3160 | |
| }, | |
| { | |
| "epoch": 14.474885844748858, | |
| "grad_norm": 0.1821436882019043, | |
| "learning_rate": 2.9328114807566863e-05, | |
| "loss": 0.0323, | |
| "step": 3170 | |
| }, | |
| { | |
| "epoch": 14.520547945205479, | |
| "grad_norm": 0.24707984924316406, | |
| "learning_rate": 2.926288323548598e-05, | |
| "loss": 0.0379, | |
| "step": 3180 | |
| }, | |
| { | |
| "epoch": 14.5662100456621, | |
| "grad_norm": 0.47426894307136536, | |
| "learning_rate": 2.919765166340509e-05, | |
| "loss": 0.0297, | |
| "step": 3190 | |
| }, | |
| { | |
| "epoch": 14.61187214611872, | |
| "grad_norm": 1.039271593093872, | |
| "learning_rate": 2.9132420091324203e-05, | |
| "loss": 0.0333, | |
| "step": 3200 | |
| }, | |
| { | |
| "epoch": 14.657534246575342, | |
| "grad_norm": 0.3582177460193634, | |
| "learning_rate": 2.9067188519243315e-05, | |
| "loss": 0.0319, | |
| "step": 3210 | |
| }, | |
| { | |
| "epoch": 14.703196347031964, | |
| "grad_norm": 0.23270468413829803, | |
| "learning_rate": 2.9001956947162427e-05, | |
| "loss": 0.0313, | |
| "step": 3220 | |
| }, | |
| { | |
| "epoch": 14.748858447488585, | |
| "grad_norm": 0.18566164374351501, | |
| "learning_rate": 2.8936725375081543e-05, | |
| "loss": 0.0313, | |
| "step": 3230 | |
| }, | |
| { | |
| "epoch": 14.794520547945206, | |
| "grad_norm": 0.1770157665014267, | |
| "learning_rate": 2.8871493803000655e-05, | |
| "loss": 0.0328, | |
| "step": 3240 | |
| }, | |
| { | |
| "epoch": 14.840182648401827, | |
| "grad_norm": 0.18902353942394257, | |
| "learning_rate": 2.8806262230919767e-05, | |
| "loss": 0.0305, | |
| "step": 3250 | |
| }, | |
| { | |
| "epoch": 14.885844748858448, | |
| "grad_norm": 0.21729163825511932, | |
| "learning_rate": 2.874103065883888e-05, | |
| "loss": 0.0348, | |
| "step": 3260 | |
| }, | |
| { | |
| "epoch": 14.931506849315069, | |
| "grad_norm": 0.1655968874692917, | |
| "learning_rate": 2.867579908675799e-05, | |
| "loss": 0.0359, | |
| "step": 3270 | |
| }, | |
| { | |
| "epoch": 14.97716894977169, | |
| "grad_norm": 0.21756260097026825, | |
| "learning_rate": 2.8610567514677107e-05, | |
| "loss": 0.0342, | |
| "step": 3280 | |
| }, | |
| { | |
| "epoch": 15.0, | |
| "eval_bertscore_f1": 0.874911447445393, | |
| "eval_bleu": 0.7648419907762714, | |
| "eval_loss": 0.042606160044670105, | |
| "eval_rougeL": 0.29841511614623334, | |
| "eval_runtime": 65.1464, | |
| "eval_samples_per_second": 23.071, | |
| "eval_steps_per_second": 1.443, | |
| "step": 3285 | |
| }, | |
| { | |
| "epoch": 15.02283105022831, | |
| "grad_norm": 0.23614604771137238, | |
| "learning_rate": 2.854533594259622e-05, | |
| "loss": 0.0325, | |
| "step": 3290 | |
| }, | |
| { | |
| "epoch": 15.068493150684931, | |
| "grad_norm": 0.25791773200035095, | |
| "learning_rate": 2.848010437051533e-05, | |
| "loss": 0.0328, | |
| "step": 3300 | |
| }, | |
| { | |
| "epoch": 15.114155251141552, | |
| "grad_norm": 0.23111367225646973, | |
| "learning_rate": 2.8414872798434443e-05, | |
| "loss": 0.033, | |
| "step": 3310 | |
| }, | |
| { | |
| "epoch": 15.159817351598173, | |
| "grad_norm": 0.1846420168876648, | |
| "learning_rate": 2.8349641226353556e-05, | |
| "loss": 0.0307, | |
| "step": 3320 | |
| }, | |
| { | |
| "epoch": 15.205479452054794, | |
| "grad_norm": 0.18300461769104004, | |
| "learning_rate": 2.828440965427267e-05, | |
| "loss": 0.0338, | |
| "step": 3330 | |
| }, | |
| { | |
| "epoch": 15.251141552511415, | |
| "grad_norm": 0.18972980976104736, | |
| "learning_rate": 2.8219178082191783e-05, | |
| "loss": 0.0328, | |
| "step": 3340 | |
| }, | |
| { | |
| "epoch": 15.296803652968036, | |
| "grad_norm": 0.19306117296218872, | |
| "learning_rate": 2.8153946510110896e-05, | |
| "loss": 0.0303, | |
| "step": 3350 | |
| }, | |
| { | |
| "epoch": 15.342465753424657, | |
| "grad_norm": 0.1415315568447113, | |
| "learning_rate": 2.8088714938030008e-05, | |
| "loss": 0.0314, | |
| "step": 3360 | |
| }, | |
| { | |
| "epoch": 15.38812785388128, | |
| "grad_norm": 0.17149841785430908, | |
| "learning_rate": 2.802348336594912e-05, | |
| "loss": 0.0318, | |
| "step": 3370 | |
| }, | |
| { | |
| "epoch": 15.4337899543379, | |
| "grad_norm": 0.19151033461093903, | |
| "learning_rate": 2.7958251793868235e-05, | |
| "loss": 0.0319, | |
| "step": 3380 | |
| }, | |
| { | |
| "epoch": 15.479452054794521, | |
| "grad_norm": 0.24431711435317993, | |
| "learning_rate": 2.7893020221787348e-05, | |
| "loss": 0.0318, | |
| "step": 3390 | |
| }, | |
| { | |
| "epoch": 15.525114155251142, | |
| "grad_norm": 0.21643275022506714, | |
| "learning_rate": 2.782778864970646e-05, | |
| "loss": 0.0305, | |
| "step": 3400 | |
| }, | |
| { | |
| "epoch": 15.570776255707763, | |
| "grad_norm": 0.4153854250907898, | |
| "learning_rate": 2.7762557077625572e-05, | |
| "loss": 0.0319, | |
| "step": 3410 | |
| }, | |
| { | |
| "epoch": 15.616438356164384, | |
| "grad_norm": 0.3891514539718628, | |
| "learning_rate": 2.7697325505544684e-05, | |
| "loss": 0.0306, | |
| "step": 3420 | |
| }, | |
| { | |
| "epoch": 15.662100456621005, | |
| "grad_norm": 0.23639145493507385, | |
| "learning_rate": 2.76320939334638e-05, | |
| "loss": 0.0321, | |
| "step": 3430 | |
| }, | |
| { | |
| "epoch": 15.707762557077626, | |
| "grad_norm": 0.19188259541988373, | |
| "learning_rate": 2.7566862361382912e-05, | |
| "loss": 0.0348, | |
| "step": 3440 | |
| }, | |
| { | |
| "epoch": 15.753424657534246, | |
| "grad_norm": 0.1795605719089508, | |
| "learning_rate": 2.7501630789302024e-05, | |
| "loss": 0.0332, | |
| "step": 3450 | |
| }, | |
| { | |
| "epoch": 15.799086757990867, | |
| "grad_norm": 0.1636660099029541, | |
| "learning_rate": 2.7436399217221136e-05, | |
| "loss": 0.0315, | |
| "step": 3460 | |
| }, | |
| { | |
| "epoch": 15.844748858447488, | |
| "grad_norm": 0.19387169182300568, | |
| "learning_rate": 2.737116764514025e-05, | |
| "loss": 0.0321, | |
| "step": 3470 | |
| }, | |
| { | |
| "epoch": 15.89041095890411, | |
| "grad_norm": 0.19701571762561798, | |
| "learning_rate": 2.7305936073059364e-05, | |
| "loss": 0.0306, | |
| "step": 3480 | |
| }, | |
| { | |
| "epoch": 15.93607305936073, | |
| "grad_norm": 0.11701110750436783, | |
| "learning_rate": 2.7240704500978476e-05, | |
| "loss": 0.0318, | |
| "step": 3490 | |
| }, | |
| { | |
| "epoch": 15.981735159817351, | |
| "grad_norm": 0.18387459218502045, | |
| "learning_rate": 2.7175472928897588e-05, | |
| "loss": 0.0318, | |
| "step": 3500 | |
| }, | |
| { | |
| "epoch": 16.0, | |
| "eval_bertscore_f1": 0.8746434982744599, | |
| "eval_bleu": 0.5087846068780025, | |
| "eval_loss": 0.04214566573500633, | |
| "eval_rougeL": 0.3110731653536748, | |
| "eval_runtime": 65.1075, | |
| "eval_samples_per_second": 23.085, | |
| "eval_steps_per_second": 1.444, | |
| "step": 3504 | |
| }, | |
| { | |
| "epoch": 16.027397260273972, | |
| "grad_norm": 0.23977895081043243, | |
| "learning_rate": 2.71102413568167e-05, | |
| "loss": 0.0313, | |
| "step": 3510 | |
| }, | |
| { | |
| "epoch": 16.073059360730593, | |
| "grad_norm": 0.1618836671113968, | |
| "learning_rate": 2.7045009784735813e-05, | |
| "loss": 0.0332, | |
| "step": 3520 | |
| }, | |
| { | |
| "epoch": 16.118721461187214, | |
| "grad_norm": 0.16630738973617554, | |
| "learning_rate": 2.6979778212654928e-05, | |
| "loss": 0.0314, | |
| "step": 3530 | |
| }, | |
| { | |
| "epoch": 16.164383561643834, | |
| "grad_norm": 0.12681370973587036, | |
| "learning_rate": 2.691454664057404e-05, | |
| "loss": 0.0358, | |
| "step": 3540 | |
| }, | |
| { | |
| "epoch": 16.210045662100455, | |
| "grad_norm": 0.2200222760438919, | |
| "learning_rate": 2.6849315068493153e-05, | |
| "loss": 0.0313, | |
| "step": 3550 | |
| }, | |
| { | |
| "epoch": 16.255707762557076, | |
| "grad_norm": 0.13934072852134705, | |
| "learning_rate": 2.6784083496412265e-05, | |
| "loss": 0.0317, | |
| "step": 3560 | |
| }, | |
| { | |
| "epoch": 16.301369863013697, | |
| "grad_norm": 0.2129465490579605, | |
| "learning_rate": 2.6718851924331377e-05, | |
| "loss": 0.0317, | |
| "step": 3570 | |
| }, | |
| { | |
| "epoch": 16.347031963470318, | |
| "grad_norm": 0.19109800457954407, | |
| "learning_rate": 2.6653620352250492e-05, | |
| "loss": 0.0326, | |
| "step": 3580 | |
| }, | |
| { | |
| "epoch": 16.39269406392694, | |
| "grad_norm": 0.20696400105953217, | |
| "learning_rate": 2.6588388780169605e-05, | |
| "loss": 0.0319, | |
| "step": 3590 | |
| }, | |
| { | |
| "epoch": 16.438356164383563, | |
| "grad_norm": 0.21380408108234406, | |
| "learning_rate": 2.6523157208088717e-05, | |
| "loss": 0.0329, | |
| "step": 3600 | |
| }, | |
| { | |
| "epoch": 16.484018264840184, | |
| "grad_norm": 0.2107531875371933, | |
| "learning_rate": 2.645792563600783e-05, | |
| "loss": 0.0328, | |
| "step": 3610 | |
| }, | |
| { | |
| "epoch": 16.529680365296805, | |
| "grad_norm": 0.1593928039073944, | |
| "learning_rate": 2.639269406392694e-05, | |
| "loss": 0.0334, | |
| "step": 3620 | |
| }, | |
| { | |
| "epoch": 16.575342465753426, | |
| "grad_norm": 0.2697247862815857, | |
| "learning_rate": 2.6327462491846057e-05, | |
| "loss": 0.0315, | |
| "step": 3630 | |
| }, | |
| { | |
| "epoch": 16.621004566210047, | |
| "grad_norm": 0.18833734095096588, | |
| "learning_rate": 2.626223091976517e-05, | |
| "loss": 0.0317, | |
| "step": 3640 | |
| }, | |
| { | |
| "epoch": 16.666666666666668, | |
| "grad_norm": 0.13584983348846436, | |
| "learning_rate": 2.619699934768428e-05, | |
| "loss": 0.0316, | |
| "step": 3650 | |
| }, | |
| { | |
| "epoch": 16.71232876712329, | |
| "grad_norm": 0.15701667964458466, | |
| "learning_rate": 2.6131767775603393e-05, | |
| "loss": 0.033, | |
| "step": 3660 | |
| }, | |
| { | |
| "epoch": 16.75799086757991, | |
| "grad_norm": 0.233077272772789, | |
| "learning_rate": 2.6066536203522505e-05, | |
| "loss": 0.0323, | |
| "step": 3670 | |
| }, | |
| { | |
| "epoch": 16.80365296803653, | |
| "grad_norm": 0.3123913109302521, | |
| "learning_rate": 2.600130463144162e-05, | |
| "loss": 0.0316, | |
| "step": 3680 | |
| }, | |
| { | |
| "epoch": 16.84931506849315, | |
| "grad_norm": 0.13571007549762726, | |
| "learning_rate": 2.5936073059360733e-05, | |
| "loss": 0.0327, | |
| "step": 3690 | |
| }, | |
| { | |
| "epoch": 16.894977168949772, | |
| "grad_norm": 0.2144760936498642, | |
| "learning_rate": 2.5870841487279845e-05, | |
| "loss": 0.0419, | |
| "step": 3700 | |
| }, | |
| { | |
| "epoch": 16.940639269406393, | |
| "grad_norm": 0.17451731860637665, | |
| "learning_rate": 2.5805609915198957e-05, | |
| "loss": 0.0315, | |
| "step": 3710 | |
| }, | |
| { | |
| "epoch": 16.986301369863014, | |
| "grad_norm": 0.18419808149337769, | |
| "learning_rate": 2.574037834311807e-05, | |
| "loss": 0.036, | |
| "step": 3720 | |
| }, | |
| { | |
| "epoch": 17.0, | |
| "eval_bertscore_f1": 0.8797636342223136, | |
| "eval_bleu": 0.706349516211386, | |
| "eval_loss": 0.041200146079063416, | |
| "eval_rougeL": 0.3065169488015689, | |
| "eval_runtime": 65.4904, | |
| "eval_samples_per_second": 22.95, | |
| "eval_steps_per_second": 1.435, | |
| "step": 3723 | |
| }, | |
| { | |
| "epoch": 17.031963470319635, | |
| "grad_norm": 0.19375832378864288, | |
| "learning_rate": 2.5675146771037185e-05, | |
| "loss": 0.0318, | |
| "step": 3730 | |
| }, | |
| { | |
| "epoch": 17.077625570776256, | |
| "grad_norm": 0.18487004935741425, | |
| "learning_rate": 2.5609915198956297e-05, | |
| "loss": 0.0295, | |
| "step": 3740 | |
| }, | |
| { | |
| "epoch": 17.123287671232877, | |
| "grad_norm": 0.25660863518714905, | |
| "learning_rate": 2.554468362687541e-05, | |
| "loss": 0.0308, | |
| "step": 3750 | |
| }, | |
| { | |
| "epoch": 17.168949771689498, | |
| "grad_norm": 0.16052749752998352, | |
| "learning_rate": 2.547945205479452e-05, | |
| "loss": 0.0325, | |
| "step": 3760 | |
| }, | |
| { | |
| "epoch": 17.21461187214612, | |
| "grad_norm": 0.22260412573814392, | |
| "learning_rate": 2.5414220482713634e-05, | |
| "loss": 0.0326, | |
| "step": 3770 | |
| }, | |
| { | |
| "epoch": 17.26027397260274, | |
| "grad_norm": 0.1811288297176361, | |
| "learning_rate": 2.534898891063275e-05, | |
| "loss": 0.032, | |
| "step": 3780 | |
| }, | |
| { | |
| "epoch": 17.30593607305936, | |
| "grad_norm": 0.24194012582302094, | |
| "learning_rate": 2.528375733855186e-05, | |
| "loss": 0.0318, | |
| "step": 3790 | |
| }, | |
| { | |
| "epoch": 17.35159817351598, | |
| "grad_norm": 0.14382942020893097, | |
| "learning_rate": 2.5218525766470974e-05, | |
| "loss": 0.0307, | |
| "step": 3800 | |
| }, | |
| { | |
| "epoch": 17.397260273972602, | |
| "grad_norm": 0.1479591429233551, | |
| "learning_rate": 2.5153294194390086e-05, | |
| "loss": 0.0297, | |
| "step": 3810 | |
| }, | |
| { | |
| "epoch": 17.442922374429223, | |
| "grad_norm": 0.18002314865589142, | |
| "learning_rate": 2.5088062622309198e-05, | |
| "loss": 0.0299, | |
| "step": 3820 | |
| }, | |
| { | |
| "epoch": 17.488584474885844, | |
| "grad_norm": 0.20746544003486633, | |
| "learning_rate": 2.5022831050228314e-05, | |
| "loss": 0.0301, | |
| "step": 3830 | |
| }, | |
| { | |
| "epoch": 17.534246575342465, | |
| "grad_norm": 0.18796005845069885, | |
| "learning_rate": 2.4957599478147426e-05, | |
| "loss": 0.0313, | |
| "step": 3840 | |
| }, | |
| { | |
| "epoch": 17.579908675799086, | |
| "grad_norm": 0.1633671671152115, | |
| "learning_rate": 2.4892367906066538e-05, | |
| "loss": 0.031, | |
| "step": 3850 | |
| }, | |
| { | |
| "epoch": 17.625570776255707, | |
| "grad_norm": 0.20464873313903809, | |
| "learning_rate": 2.482713633398565e-05, | |
| "loss": 0.0294, | |
| "step": 3860 | |
| }, | |
| { | |
| "epoch": 17.671232876712327, | |
| "grad_norm": 0.1803971529006958, | |
| "learning_rate": 2.4761904761904762e-05, | |
| "loss": 0.0322, | |
| "step": 3870 | |
| }, | |
| { | |
| "epoch": 17.71689497716895, | |
| "grad_norm": 0.2384488433599472, | |
| "learning_rate": 2.4696673189823878e-05, | |
| "loss": 0.0319, | |
| "step": 3880 | |
| }, | |
| { | |
| "epoch": 17.76255707762557, | |
| "grad_norm": 0.11598934233188629, | |
| "learning_rate": 2.463144161774299e-05, | |
| "loss": 0.0334, | |
| "step": 3890 | |
| }, | |
| { | |
| "epoch": 17.80821917808219, | |
| "grad_norm": 0.2237691730260849, | |
| "learning_rate": 2.4566210045662102e-05, | |
| "loss": 0.0317, | |
| "step": 3900 | |
| }, | |
| { | |
| "epoch": 17.853881278538815, | |
| "grad_norm": 0.23967674374580383, | |
| "learning_rate": 2.4500978473581214e-05, | |
| "loss": 0.0325, | |
| "step": 3910 | |
| }, | |
| { | |
| "epoch": 17.899543378995435, | |
| "grad_norm": 0.15783822536468506, | |
| "learning_rate": 2.4435746901500327e-05, | |
| "loss": 0.0318, | |
| "step": 3920 | |
| }, | |
| { | |
| "epoch": 17.945205479452056, | |
| "grad_norm": 0.1800832897424698, | |
| "learning_rate": 2.4370515329419442e-05, | |
| "loss": 0.0317, | |
| "step": 3930 | |
| }, | |
| { | |
| "epoch": 17.990867579908677, | |
| "grad_norm": 0.5130301713943481, | |
| "learning_rate": 2.4305283757338554e-05, | |
| "loss": 0.0319, | |
| "step": 3940 | |
| }, | |
| { | |
| "epoch": 18.0, | |
| "eval_bertscore_f1": 0.8798081558780201, | |
| "eval_bleu": 0.7220797778715596, | |
| "eval_loss": 0.042013321071863174, | |
| "eval_rougeL": 0.30594782179206126, | |
| "eval_runtime": 65.524, | |
| "eval_samples_per_second": 22.938, | |
| "eval_steps_per_second": 1.435, | |
| "step": 3942 | |
| }, | |
| { | |
| "epoch": 18.036529680365298, | |
| "grad_norm": 0.16154253482818604, | |
| "learning_rate": 2.4240052185257666e-05, | |
| "loss": 0.03, | |
| "step": 3950 | |
| }, | |
| { | |
| "epoch": 18.08219178082192, | |
| "grad_norm": 0.16206321120262146, | |
| "learning_rate": 2.417482061317678e-05, | |
| "loss": 0.0297, | |
| "step": 3960 | |
| }, | |
| { | |
| "epoch": 18.12785388127854, | |
| "grad_norm": 0.22228258848190308, | |
| "learning_rate": 2.410958904109589e-05, | |
| "loss": 0.0302, | |
| "step": 3970 | |
| }, | |
| { | |
| "epoch": 18.17351598173516, | |
| "grad_norm": 0.2068241685628891, | |
| "learning_rate": 2.4044357469015006e-05, | |
| "loss": 0.0322, | |
| "step": 3980 | |
| }, | |
| { | |
| "epoch": 18.21917808219178, | |
| "grad_norm": 0.14935638010501862, | |
| "learning_rate": 2.397912589693412e-05, | |
| "loss": 0.0319, | |
| "step": 3990 | |
| }, | |
| { | |
| "epoch": 18.264840182648403, | |
| "grad_norm": 0.16079388558864594, | |
| "learning_rate": 2.391389432485323e-05, | |
| "loss": 0.0312, | |
| "step": 4000 | |
| }, | |
| { | |
| "epoch": 18.310502283105023, | |
| "grad_norm": 0.1624707281589508, | |
| "learning_rate": 2.3848662752772343e-05, | |
| "loss": 0.0316, | |
| "step": 4010 | |
| }, | |
| { | |
| "epoch": 18.356164383561644, | |
| "grad_norm": 0.1676281988620758, | |
| "learning_rate": 2.3783431180691455e-05, | |
| "loss": 0.0305, | |
| "step": 4020 | |
| }, | |
| { | |
| "epoch": 18.401826484018265, | |
| "grad_norm": 0.16586793959140778, | |
| "learning_rate": 2.371819960861057e-05, | |
| "loss": 0.0323, | |
| "step": 4030 | |
| }, | |
| { | |
| "epoch": 18.447488584474886, | |
| "grad_norm": 0.1340363472700119, | |
| "learning_rate": 2.3652968036529683e-05, | |
| "loss": 0.0289, | |
| "step": 4040 | |
| }, | |
| { | |
| "epoch": 18.493150684931507, | |
| "grad_norm": 0.14134575426578522, | |
| "learning_rate": 2.3587736464448795e-05, | |
| "loss": 0.0318, | |
| "step": 4050 | |
| }, | |
| { | |
| "epoch": 18.538812785388128, | |
| "grad_norm": 0.16164334118366241, | |
| "learning_rate": 2.3522504892367907e-05, | |
| "loss": 0.0303, | |
| "step": 4060 | |
| }, | |
| { | |
| "epoch": 18.58447488584475, | |
| "grad_norm": 0.3421485722064972, | |
| "learning_rate": 2.345727332028702e-05, | |
| "loss": 0.0313, | |
| "step": 4070 | |
| }, | |
| { | |
| "epoch": 18.63013698630137, | |
| "grad_norm": 0.17548483610153198, | |
| "learning_rate": 2.3392041748206135e-05, | |
| "loss": 0.0322, | |
| "step": 4080 | |
| }, | |
| { | |
| "epoch": 18.67579908675799, | |
| "grad_norm": 0.11101207137107849, | |
| "learning_rate": 2.3326810176125247e-05, | |
| "loss": 0.033, | |
| "step": 4090 | |
| }, | |
| { | |
| "epoch": 18.72146118721461, | |
| "grad_norm": 0.19600647687911987, | |
| "learning_rate": 2.326157860404436e-05, | |
| "loss": 0.0314, | |
| "step": 4100 | |
| }, | |
| { | |
| "epoch": 18.767123287671232, | |
| "grad_norm": 0.4979359805583954, | |
| "learning_rate": 2.319634703196347e-05, | |
| "loss": 0.0304, | |
| "step": 4110 | |
| }, | |
| { | |
| "epoch": 18.812785388127853, | |
| "grad_norm": 0.16611818969249725, | |
| "learning_rate": 2.3131115459882584e-05, | |
| "loss": 0.0326, | |
| "step": 4120 | |
| }, | |
| { | |
| "epoch": 18.858447488584474, | |
| "grad_norm": 0.2194747030735016, | |
| "learning_rate": 2.30658838878017e-05, | |
| "loss": 0.0336, | |
| "step": 4130 | |
| }, | |
| { | |
| "epoch": 18.904109589041095, | |
| "grad_norm": 0.2581609785556793, | |
| "learning_rate": 2.300065231572081e-05, | |
| "loss": 0.0306, | |
| "step": 4140 | |
| }, | |
| { | |
| "epoch": 18.949771689497716, | |
| "grad_norm": 0.20705465972423553, | |
| "learning_rate": 2.2935420743639923e-05, | |
| "loss": 0.031, | |
| "step": 4150 | |
| }, | |
| { | |
| "epoch": 18.995433789954337, | |
| "grad_norm": 0.18591341376304626, | |
| "learning_rate": 2.2870189171559036e-05, | |
| "loss": 0.0331, | |
| "step": 4160 | |
| }, | |
| { | |
| "epoch": 19.0, | |
| "eval_bertscore_f1": 0.8795998138502606, | |
| "eval_bleu": 0.7183983246376581, | |
| "eval_loss": 0.04113374277949333, | |
| "eval_rougeL": 0.3053785040223903, | |
| "eval_runtime": 64.8761, | |
| "eval_samples_per_second": 23.167, | |
| "eval_steps_per_second": 1.449, | |
| "step": 4161 | |
| }, | |
| { | |
| "epoch": 19.041095890410958, | |
| "grad_norm": 0.2786708176136017, | |
| "learning_rate": 2.2804957599478148e-05, | |
| "loss": 0.0326, | |
| "step": 4170 | |
| }, | |
| { | |
| "epoch": 19.08675799086758, | |
| "grad_norm": 0.17057956755161285, | |
| "learning_rate": 2.2739726027397263e-05, | |
| "loss": 0.0306, | |
| "step": 4180 | |
| }, | |
| { | |
| "epoch": 19.1324200913242, | |
| "grad_norm": 0.15491874516010284, | |
| "learning_rate": 2.2674494455316376e-05, | |
| "loss": 0.0313, | |
| "step": 4190 | |
| }, | |
| { | |
| "epoch": 19.17808219178082, | |
| "grad_norm": 0.319887638092041, | |
| "learning_rate": 2.2609262883235488e-05, | |
| "loss": 0.0316, | |
| "step": 4200 | |
| }, | |
| { | |
| "epoch": 19.22374429223744, | |
| "grad_norm": 0.18852587044239044, | |
| "learning_rate": 2.25440313111546e-05, | |
| "loss": 0.0299, | |
| "step": 4210 | |
| }, | |
| { | |
| "epoch": 19.269406392694062, | |
| "grad_norm": 0.17552684247493744, | |
| "learning_rate": 2.2478799739073715e-05, | |
| "loss": 0.0306, | |
| "step": 4220 | |
| }, | |
| { | |
| "epoch": 19.315068493150687, | |
| "grad_norm": 0.25381070375442505, | |
| "learning_rate": 2.2413568166992828e-05, | |
| "loss": 0.0325, | |
| "step": 4230 | |
| }, | |
| { | |
| "epoch": 19.360730593607308, | |
| "grad_norm": 0.23695605993270874, | |
| "learning_rate": 2.234833659491194e-05, | |
| "loss": 0.0317, | |
| "step": 4240 | |
| }, | |
| { | |
| "epoch": 19.40639269406393, | |
| "grad_norm": 0.15467983484268188, | |
| "learning_rate": 2.2283105022831052e-05, | |
| "loss": 0.0307, | |
| "step": 4250 | |
| }, | |
| { | |
| "epoch": 19.45205479452055, | |
| "grad_norm": 0.19283008575439453, | |
| "learning_rate": 2.2217873450750164e-05, | |
| "loss": 0.0312, | |
| "step": 4260 | |
| }, | |
| { | |
| "epoch": 19.49771689497717, | |
| "grad_norm": 0.130938321352005, | |
| "learning_rate": 2.215264187866928e-05, | |
| "loss": 0.0304, | |
| "step": 4270 | |
| }, | |
| { | |
| "epoch": 19.54337899543379, | |
| "grad_norm": 0.17690137028694153, | |
| "learning_rate": 2.2087410306588392e-05, | |
| "loss": 0.0326, | |
| "step": 4280 | |
| }, | |
| { | |
| "epoch": 19.589041095890412, | |
| "grad_norm": 0.15000636875629425, | |
| "learning_rate": 2.2022178734507504e-05, | |
| "loss": 0.0299, | |
| "step": 4290 | |
| }, | |
| { | |
| "epoch": 19.634703196347033, | |
| "grad_norm": 0.19072392582893372, | |
| "learning_rate": 2.1956947162426616e-05, | |
| "loss": 0.0311, | |
| "step": 4300 | |
| }, | |
| { | |
| "epoch": 19.680365296803654, | |
| "grad_norm": 0.24786625802516937, | |
| "learning_rate": 2.189171559034573e-05, | |
| "loss": 0.0325, | |
| "step": 4310 | |
| }, | |
| { | |
| "epoch": 19.726027397260275, | |
| "grad_norm": 1.5711880922317505, | |
| "learning_rate": 2.1826484018264844e-05, | |
| "loss": 0.0354, | |
| "step": 4320 | |
| }, | |
| { | |
| "epoch": 19.771689497716896, | |
| "grad_norm": 0.2299468070268631, | |
| "learning_rate": 2.1761252446183956e-05, | |
| "loss": 0.0308, | |
| "step": 4330 | |
| }, | |
| { | |
| "epoch": 19.817351598173516, | |
| "grad_norm": 0.14930671453475952, | |
| "learning_rate": 2.1696020874103068e-05, | |
| "loss": 0.0295, | |
| "step": 4340 | |
| }, | |
| { | |
| "epoch": 19.863013698630137, | |
| "grad_norm": 0.263234406709671, | |
| "learning_rate": 2.163078930202218e-05, | |
| "loss": 0.0305, | |
| "step": 4350 | |
| }, | |
| { | |
| "epoch": 19.908675799086758, | |
| "grad_norm": 0.2260236144065857, | |
| "learning_rate": 2.1565557729941293e-05, | |
| "loss": 0.0332, | |
| "step": 4360 | |
| }, | |
| { | |
| "epoch": 19.95433789954338, | |
| "grad_norm": 0.23456306755542755, | |
| "learning_rate": 2.1500326157860408e-05, | |
| "loss": 0.0303, | |
| "step": 4370 | |
| }, | |
| { | |
| "epoch": 20.0, | |
| "grad_norm": 0.18612632155418396, | |
| "learning_rate": 2.143509458577952e-05, | |
| "loss": 0.0305, | |
| "step": 4380 | |
| }, | |
| { | |
| "epoch": 20.0, | |
| "eval_bertscore_f1": 0.8794884783223876, | |
| "eval_bleu": 0.7197986607742484, | |
| "eval_loss": 0.04177827760577202, | |
| "eval_rougeL": 0.30584228187550916, | |
| "eval_runtime": 64.9233, | |
| "eval_samples_per_second": 23.15, | |
| "eval_steps_per_second": 1.448, | |
| "step": 4380 | |
| }, | |
| { | |
| "epoch": 20.04566210045662, | |
| "grad_norm": 1.3157190084457397, | |
| "learning_rate": 2.1369863013698632e-05, | |
| "loss": 0.0319, | |
| "step": 4390 | |
| }, | |
| { | |
| "epoch": 20.091324200913242, | |
| "grad_norm": 0.2511720061302185, | |
| "learning_rate": 2.1304631441617745e-05, | |
| "loss": 0.0316, | |
| "step": 4400 | |
| }, | |
| { | |
| "epoch": 20.136986301369863, | |
| "grad_norm": 0.15252014994621277, | |
| "learning_rate": 2.1239399869536857e-05, | |
| "loss": 0.0304, | |
| "step": 4410 | |
| }, | |
| { | |
| "epoch": 20.182648401826484, | |
| "grad_norm": 0.27835097908973694, | |
| "learning_rate": 2.1174168297455972e-05, | |
| "loss": 0.0318, | |
| "step": 4420 | |
| }, | |
| { | |
| "epoch": 20.228310502283104, | |
| "grad_norm": 0.1549067199230194, | |
| "learning_rate": 2.1108936725375085e-05, | |
| "loss": 0.0317, | |
| "step": 4430 | |
| }, | |
| { | |
| "epoch": 20.273972602739725, | |
| "grad_norm": 0.28859424591064453, | |
| "learning_rate": 2.1043705153294197e-05, | |
| "loss": 0.0321, | |
| "step": 4440 | |
| }, | |
| { | |
| "epoch": 20.319634703196346, | |
| "grad_norm": 0.18468542397022247, | |
| "learning_rate": 2.097847358121331e-05, | |
| "loss": 0.0319, | |
| "step": 4450 | |
| }, | |
| { | |
| "epoch": 20.365296803652967, | |
| "grad_norm": 0.14522314071655273, | |
| "learning_rate": 2.091324200913242e-05, | |
| "loss": 0.0298, | |
| "step": 4460 | |
| }, | |
| { | |
| "epoch": 20.410958904109588, | |
| "grad_norm": 0.2167425900697708, | |
| "learning_rate": 2.0848010437051537e-05, | |
| "loss": 0.0301, | |
| "step": 4470 | |
| }, | |
| { | |
| "epoch": 20.45662100456621, | |
| "grad_norm": 0.17523987591266632, | |
| "learning_rate": 2.078277886497065e-05, | |
| "loss": 0.0299, | |
| "step": 4480 | |
| }, | |
| { | |
| "epoch": 20.50228310502283, | |
| "grad_norm": 0.2222181111574173, | |
| "learning_rate": 2.071754729288976e-05, | |
| "loss": 0.0318, | |
| "step": 4490 | |
| }, | |
| { | |
| "epoch": 20.54794520547945, | |
| "grad_norm": 0.15766695141792297, | |
| "learning_rate": 2.0652315720808873e-05, | |
| "loss": 0.0315, | |
| "step": 4500 | |
| }, | |
| { | |
| "epoch": 20.59360730593607, | |
| "grad_norm": 0.23727627098560333, | |
| "learning_rate": 2.0587084148727985e-05, | |
| "loss": 0.0293, | |
| "step": 4510 | |
| }, | |
| { | |
| "epoch": 20.639269406392692, | |
| "grad_norm": 0.18761341273784637, | |
| "learning_rate": 2.05218525766471e-05, | |
| "loss": 0.0298, | |
| "step": 4520 | |
| }, | |
| { | |
| "epoch": 20.684931506849313, | |
| "grad_norm": 0.22776687145233154, | |
| "learning_rate": 2.0456621004566213e-05, | |
| "loss": 0.0326, | |
| "step": 4530 | |
| }, | |
| { | |
| "epoch": 20.730593607305934, | |
| "grad_norm": 0.17745907604694366, | |
| "learning_rate": 2.0391389432485325e-05, | |
| "loss": 0.0302, | |
| "step": 4540 | |
| }, | |
| { | |
| "epoch": 20.77625570776256, | |
| "grad_norm": 0.19407926499843597, | |
| "learning_rate": 2.0326157860404437e-05, | |
| "loss": 0.0323, | |
| "step": 4550 | |
| }, | |
| { | |
| "epoch": 20.82191780821918, | |
| "grad_norm": 0.1550801694393158, | |
| "learning_rate": 2.026092628832355e-05, | |
| "loss": 0.0301, | |
| "step": 4560 | |
| }, | |
| { | |
| "epoch": 20.8675799086758, | |
| "grad_norm": 0.2518145740032196, | |
| "learning_rate": 2.0195694716242665e-05, | |
| "loss": 0.0313, | |
| "step": 4570 | |
| }, | |
| { | |
| "epoch": 20.91324200913242, | |
| "grad_norm": 0.18234769999980927, | |
| "learning_rate": 2.0130463144161777e-05, | |
| "loss": 0.0313, | |
| "step": 4580 | |
| }, | |
| { | |
| "epoch": 20.958904109589042, | |
| "grad_norm": 0.17262135446071625, | |
| "learning_rate": 2.006523157208089e-05, | |
| "loss": 0.0302, | |
| "step": 4590 | |
| }, | |
| { | |
| "epoch": 21.0, | |
| "eval_bertscore_f1": 0.8762413332720875, | |
| "eval_bleu": 0.8743328548000596, | |
| "eval_loss": 0.04259444400668144, | |
| "eval_rougeL": 0.29478954803373814, | |
| "eval_runtime": 65.0506, | |
| "eval_samples_per_second": 23.105, | |
| "eval_steps_per_second": 1.445, | |
| "step": 4599 | |
| }, | |
| { | |
| "epoch": 21.004566210045663, | |
| "grad_norm": 0.22366954386234283, | |
| "learning_rate": 2e-05, | |
| "loss": 0.0327, | |
| "step": 4600 | |
| }, | |
| { | |
| "epoch": 21.050228310502284, | |
| "grad_norm": 0.19344310462474823, | |
| "learning_rate": 1.9934768427919114e-05, | |
| "loss": 0.0333, | |
| "step": 4610 | |
| }, | |
| { | |
| "epoch": 21.095890410958905, | |
| "grad_norm": 0.18375934660434723, | |
| "learning_rate": 1.986953685583823e-05, | |
| "loss": 0.0345, | |
| "step": 4620 | |
| }, | |
| { | |
| "epoch": 21.141552511415526, | |
| "grad_norm": 0.15804964303970337, | |
| "learning_rate": 1.980430528375734e-05, | |
| "loss": 0.0291, | |
| "step": 4630 | |
| }, | |
| { | |
| "epoch": 21.187214611872147, | |
| "grad_norm": 0.17671050131320953, | |
| "learning_rate": 1.9739073711676454e-05, | |
| "loss": 0.0316, | |
| "step": 4640 | |
| }, | |
| { | |
| "epoch": 21.232876712328768, | |
| "grad_norm": 0.6683818101882935, | |
| "learning_rate": 1.9673842139595566e-05, | |
| "loss": 0.03, | |
| "step": 4650 | |
| }, | |
| { | |
| "epoch": 21.27853881278539, | |
| "grad_norm": 0.16015271842479706, | |
| "learning_rate": 1.9608610567514678e-05, | |
| "loss": 0.0292, | |
| "step": 4660 | |
| }, | |
| { | |
| "epoch": 21.32420091324201, | |
| "grad_norm": 0.25762075185775757, | |
| "learning_rate": 1.9543378995433794e-05, | |
| "loss": 0.0324, | |
| "step": 4670 | |
| }, | |
| { | |
| "epoch": 21.36986301369863, | |
| "grad_norm": 0.15417823195457458, | |
| "learning_rate": 1.9478147423352906e-05, | |
| "loss": 0.0321, | |
| "step": 4680 | |
| }, | |
| { | |
| "epoch": 21.41552511415525, | |
| "grad_norm": 0.1617811620235443, | |
| "learning_rate": 1.9412915851272018e-05, | |
| "loss": 0.0334, | |
| "step": 4690 | |
| }, | |
| { | |
| "epoch": 21.461187214611872, | |
| "grad_norm": 0.1719217449426651, | |
| "learning_rate": 1.934768427919113e-05, | |
| "loss": 0.0294, | |
| "step": 4700 | |
| }, | |
| { | |
| "epoch": 21.506849315068493, | |
| "grad_norm": 0.1514827162027359, | |
| "learning_rate": 1.9282452707110242e-05, | |
| "loss": 0.0304, | |
| "step": 4710 | |
| }, | |
| { | |
| "epoch": 21.552511415525114, | |
| "grad_norm": 0.14078788459300995, | |
| "learning_rate": 1.9217221135029358e-05, | |
| "loss": 0.0302, | |
| "step": 4720 | |
| }, | |
| { | |
| "epoch": 21.598173515981735, | |
| "grad_norm": 0.1602380871772766, | |
| "learning_rate": 1.915198956294847e-05, | |
| "loss": 0.0309, | |
| "step": 4730 | |
| }, | |
| { | |
| "epoch": 21.643835616438356, | |
| "grad_norm": 0.23968474566936493, | |
| "learning_rate": 1.9086757990867582e-05, | |
| "loss": 0.031, | |
| "step": 4740 | |
| }, | |
| { | |
| "epoch": 21.689497716894977, | |
| "grad_norm": 0.10929176211357117, | |
| "learning_rate": 1.9021526418786694e-05, | |
| "loss": 0.0293, | |
| "step": 4750 | |
| }, | |
| { | |
| "epoch": 21.735159817351597, | |
| "grad_norm": 0.18268796801567078, | |
| "learning_rate": 1.8956294846705807e-05, | |
| "loss": 0.0319, | |
| "step": 4760 | |
| }, | |
| { | |
| "epoch": 21.78082191780822, | |
| "grad_norm": 0.260498583316803, | |
| "learning_rate": 1.8891063274624922e-05, | |
| "loss": 0.0318, | |
| "step": 4770 | |
| }, | |
| { | |
| "epoch": 21.82648401826484, | |
| "grad_norm": 0.18598124384880066, | |
| "learning_rate": 1.8825831702544034e-05, | |
| "loss": 0.0309, | |
| "step": 4780 | |
| }, | |
| { | |
| "epoch": 21.87214611872146, | |
| "grad_norm": 0.14917334914207458, | |
| "learning_rate": 1.8760600130463146e-05, | |
| "loss": 0.0303, | |
| "step": 4790 | |
| }, | |
| { | |
| "epoch": 21.91780821917808, | |
| "grad_norm": 0.17310568690299988, | |
| "learning_rate": 1.869536855838226e-05, | |
| "loss": 0.0311, | |
| "step": 4800 | |
| }, | |
| { | |
| "epoch": 21.963470319634702, | |
| "grad_norm": 0.16899165511131287, | |
| "learning_rate": 1.863013698630137e-05, | |
| "loss": 0.0302, | |
| "step": 4810 | |
| }, | |
| { | |
| "epoch": 22.0, | |
| "eval_bertscore_f1": 0.8747094868027363, | |
| "eval_bleu": 0.9393280067525835, | |
| "eval_loss": 0.041989196091890335, | |
| "eval_rougeL": 0.2900498588163605, | |
| "eval_runtime": 65.3354, | |
| "eval_samples_per_second": 23.004, | |
| "eval_steps_per_second": 1.439, | |
| "step": 4818 | |
| }, | |
| { | |
| "epoch": 22.009132420091323, | |
| "grad_norm": 0.19090060889720917, | |
| "learning_rate": 1.8564905414220486e-05, | |
| "loss": 0.0354, | |
| "step": 4820 | |
| }, | |
| { | |
| "epoch": 22.054794520547944, | |
| "grad_norm": 0.19969524443149567, | |
| "learning_rate": 1.84996738421396e-05, | |
| "loss": 0.0299, | |
| "step": 4830 | |
| }, | |
| { | |
| "epoch": 22.100456621004565, | |
| "grad_norm": 0.2311638593673706, | |
| "learning_rate": 1.843444227005871e-05, | |
| "loss": 0.0301, | |
| "step": 4840 | |
| }, | |
| { | |
| "epoch": 22.146118721461185, | |
| "grad_norm": 0.1671016663312912, | |
| "learning_rate": 1.8369210697977823e-05, | |
| "loss": 0.0302, | |
| "step": 4850 | |
| }, | |
| { | |
| "epoch": 22.19178082191781, | |
| "grad_norm": 0.14974653720855713, | |
| "learning_rate": 1.8303979125896935e-05, | |
| "loss": 0.0296, | |
| "step": 4860 | |
| }, | |
| { | |
| "epoch": 22.23744292237443, | |
| "grad_norm": 0.20775803923606873, | |
| "learning_rate": 1.823874755381605e-05, | |
| "loss": 0.0325, | |
| "step": 4870 | |
| }, | |
| { | |
| "epoch": 22.28310502283105, | |
| "grad_norm": 0.18768472969532013, | |
| "learning_rate": 1.8173515981735163e-05, | |
| "loss": 0.0317, | |
| "step": 4880 | |
| }, | |
| { | |
| "epoch": 22.328767123287673, | |
| "grad_norm": 0.18841847777366638, | |
| "learning_rate": 1.8108284409654275e-05, | |
| "loss": 0.0286, | |
| "step": 4890 | |
| }, | |
| { | |
| "epoch": 22.374429223744293, | |
| "grad_norm": 0.15577343106269836, | |
| "learning_rate": 1.8043052837573387e-05, | |
| "loss": 0.0312, | |
| "step": 4900 | |
| }, | |
| { | |
| "epoch": 22.420091324200914, | |
| "grad_norm": 0.17433637380599976, | |
| "learning_rate": 1.79778212654925e-05, | |
| "loss": 0.0303, | |
| "step": 4910 | |
| }, | |
| { | |
| "epoch": 22.465753424657535, | |
| "grad_norm": 0.1529913693666458, | |
| "learning_rate": 1.7912589693411615e-05, | |
| "loss": 0.0295, | |
| "step": 4920 | |
| }, | |
| { | |
| "epoch": 22.511415525114156, | |
| "grad_norm": 0.14870087802410126, | |
| "learning_rate": 1.7847358121330727e-05, | |
| "loss": 0.0316, | |
| "step": 4930 | |
| }, | |
| { | |
| "epoch": 22.557077625570777, | |
| "grad_norm": 0.12427645176649094, | |
| "learning_rate": 1.778212654924984e-05, | |
| "loss": 0.0279, | |
| "step": 4940 | |
| }, | |
| { | |
| "epoch": 22.602739726027398, | |
| "grad_norm": 0.2489793747663498, | |
| "learning_rate": 1.771689497716895e-05, | |
| "loss": 0.032, | |
| "step": 4950 | |
| }, | |
| { | |
| "epoch": 22.64840182648402, | |
| "grad_norm": 0.15614229440689087, | |
| "learning_rate": 1.7651663405088064e-05, | |
| "loss": 0.0303, | |
| "step": 4960 | |
| }, | |
| { | |
| "epoch": 22.69406392694064, | |
| "grad_norm": 0.4469321072101593, | |
| "learning_rate": 1.758643183300718e-05, | |
| "loss": 0.0314, | |
| "step": 4970 | |
| }, | |
| { | |
| "epoch": 22.73972602739726, | |
| "grad_norm": 0.123336561024189, | |
| "learning_rate": 1.752120026092629e-05, | |
| "loss": 0.0303, | |
| "step": 4980 | |
| }, | |
| { | |
| "epoch": 22.78538812785388, | |
| "grad_norm": 0.1903766691684723, | |
| "learning_rate": 1.7455968688845403e-05, | |
| "loss": 0.03, | |
| "step": 4990 | |
| }, | |
| { | |
| "epoch": 22.831050228310502, | |
| "grad_norm": 0.1910991221666336, | |
| "learning_rate": 1.7390737116764516e-05, | |
| "loss": 0.0297, | |
| "step": 5000 | |
| }, | |
| { | |
| "epoch": 22.876712328767123, | |
| "grad_norm": 0.16593395173549652, | |
| "learning_rate": 1.7325505544683628e-05, | |
| "loss": 0.0309, | |
| "step": 5010 | |
| }, | |
| { | |
| "epoch": 22.922374429223744, | |
| "grad_norm": 0.19862234592437744, | |
| "learning_rate": 1.7260273972602743e-05, | |
| "loss": 0.0294, | |
| "step": 5020 | |
| }, | |
| { | |
| "epoch": 22.968036529680365, | |
| "grad_norm": 0.5439543128013611, | |
| "learning_rate": 1.7195042400521855e-05, | |
| "loss": 0.031, | |
| "step": 5030 | |
| }, | |
| { | |
| "epoch": 23.0, | |
| "eval_bertscore_f1": 0.8795151574089776, | |
| "eval_bleu": 0.726659433737333, | |
| "eval_loss": 0.041650936007499695, | |
| "eval_rougeL": 0.3058748502333608, | |
| "eval_runtime": 65.391, | |
| "eval_samples_per_second": 22.985, | |
| "eval_steps_per_second": 1.438, | |
| "step": 5037 | |
| }, | |
| { | |
| "epoch": 23.013698630136986, | |
| "grad_norm": 0.16144615411758423, | |
| "learning_rate": 1.7129810828440968e-05, | |
| "loss": 0.0308, | |
| "step": 5040 | |
| }, | |
| { | |
| "epoch": 23.059360730593607, | |
| "grad_norm": 0.21443626284599304, | |
| "learning_rate": 1.706457925636008e-05, | |
| "loss": 0.0318, | |
| "step": 5050 | |
| }, | |
| { | |
| "epoch": 23.105022831050228, | |
| "grad_norm": 0.1895105540752411, | |
| "learning_rate": 1.6999347684279192e-05, | |
| "loss": 0.0311, | |
| "step": 5060 | |
| }, | |
| { | |
| "epoch": 23.15068493150685, | |
| "grad_norm": 0.17296628654003143, | |
| "learning_rate": 1.6934116112198308e-05, | |
| "loss": 0.0297, | |
| "step": 5070 | |
| }, | |
| { | |
| "epoch": 23.19634703196347, | |
| "grad_norm": 0.22210587561130524, | |
| "learning_rate": 1.686888454011742e-05, | |
| "loss": 0.0286, | |
| "step": 5080 | |
| }, | |
| { | |
| "epoch": 23.24200913242009, | |
| "grad_norm": 0.1891639083623886, | |
| "learning_rate": 1.6803652968036532e-05, | |
| "loss": 0.0301, | |
| "step": 5090 | |
| }, | |
| { | |
| "epoch": 23.28767123287671, | |
| "grad_norm": 0.17124980688095093, | |
| "learning_rate": 1.6738421395955644e-05, | |
| "loss": 0.0321, | |
| "step": 5100 | |
| }, | |
| { | |
| "epoch": 23.333333333333332, | |
| "grad_norm": 0.16418173909187317, | |
| "learning_rate": 1.6673189823874756e-05, | |
| "loss": 0.0304, | |
| "step": 5110 | |
| }, | |
| { | |
| "epoch": 23.378995433789953, | |
| "grad_norm": 0.17459054291248322, | |
| "learning_rate": 1.6607958251793872e-05, | |
| "loss": 0.0311, | |
| "step": 5120 | |
| }, | |
| { | |
| "epoch": 23.424657534246574, | |
| "grad_norm": 0.15232379734516144, | |
| "learning_rate": 1.6542726679712984e-05, | |
| "loss": 0.0315, | |
| "step": 5130 | |
| }, | |
| { | |
| "epoch": 23.470319634703195, | |
| "grad_norm": 0.25604724884033203, | |
| "learning_rate": 1.6477495107632096e-05, | |
| "loss": 0.0308, | |
| "step": 5140 | |
| }, | |
| { | |
| "epoch": 23.515981735159816, | |
| "grad_norm": 0.3925830125808716, | |
| "learning_rate": 1.641226353555121e-05, | |
| "loss": 0.0283, | |
| "step": 5150 | |
| }, | |
| { | |
| "epoch": 23.561643835616437, | |
| "grad_norm": 0.1935456395149231, | |
| "learning_rate": 1.634703196347032e-05, | |
| "loss": 0.0276, | |
| "step": 5160 | |
| }, | |
| { | |
| "epoch": 23.60730593607306, | |
| "grad_norm": 0.16616645455360413, | |
| "learning_rate": 1.6281800391389436e-05, | |
| "loss": 0.0309, | |
| "step": 5170 | |
| }, | |
| { | |
| "epoch": 23.652968036529682, | |
| "grad_norm": 0.19446122646331787, | |
| "learning_rate": 1.6216568819308548e-05, | |
| "loss": 0.0291, | |
| "step": 5180 | |
| }, | |
| { | |
| "epoch": 23.698630136986303, | |
| "grad_norm": 0.20272910594940186, | |
| "learning_rate": 1.615133724722766e-05, | |
| "loss": 0.0314, | |
| "step": 5190 | |
| }, | |
| { | |
| "epoch": 23.744292237442924, | |
| "grad_norm": 0.21669697761535645, | |
| "learning_rate": 1.6086105675146773e-05, | |
| "loss": 0.0322, | |
| "step": 5200 | |
| }, | |
| { | |
| "epoch": 23.789954337899545, | |
| "grad_norm": 0.19931752979755402, | |
| "learning_rate": 1.6020874103065885e-05, | |
| "loss": 0.0296, | |
| "step": 5210 | |
| }, | |
| { | |
| "epoch": 23.835616438356166, | |
| "grad_norm": 0.20344951748847961, | |
| "learning_rate": 1.5955642530985e-05, | |
| "loss": 0.0296, | |
| "step": 5220 | |
| }, | |
| { | |
| "epoch": 23.881278538812786, | |
| "grad_norm": 0.15517154335975647, | |
| "learning_rate": 1.5890410958904112e-05, | |
| "loss": 0.0309, | |
| "step": 5230 | |
| }, | |
| { | |
| "epoch": 23.926940639269407, | |
| "grad_norm": 0.22952044010162354, | |
| "learning_rate": 1.5825179386823225e-05, | |
| "loss": 0.0312, | |
| "step": 5240 | |
| }, | |
| { | |
| "epoch": 23.972602739726028, | |
| "grad_norm": 0.2000703364610672, | |
| "learning_rate": 1.5759947814742337e-05, | |
| "loss": 0.0321, | |
| "step": 5250 | |
| }, | |
| { | |
| "epoch": 24.0, | |
| "eval_bertscore_f1": 0.8752217368927306, | |
| "eval_bleu": 0.931234723225367, | |
| "eval_loss": 0.04182644933462143, | |
| "eval_rougeL": 0.29302986167189593, | |
| "eval_runtime": 64.8994, | |
| "eval_samples_per_second": 23.159, | |
| "eval_steps_per_second": 1.448, | |
| "step": 5256 | |
| }, | |
| { | |
| "epoch": 24.01826484018265, | |
| "grad_norm": 0.12788528203964233, | |
| "learning_rate": 1.569471624266145e-05, | |
| "loss": 0.0288, | |
| "step": 5260 | |
| }, | |
| { | |
| "epoch": 24.06392694063927, | |
| "grad_norm": 0.1858115792274475, | |
| "learning_rate": 1.5629484670580565e-05, | |
| "loss": 0.0319, | |
| "step": 5270 | |
| }, | |
| { | |
| "epoch": 24.10958904109589, | |
| "grad_norm": 0.20606492459774017, | |
| "learning_rate": 1.5564253098499677e-05, | |
| "loss": 0.0327, | |
| "step": 5280 | |
| }, | |
| { | |
| "epoch": 24.15525114155251, | |
| "grad_norm": 0.1898220181465149, | |
| "learning_rate": 1.549902152641879e-05, | |
| "loss": 0.0307, | |
| "step": 5290 | |
| }, | |
| { | |
| "epoch": 24.200913242009133, | |
| "grad_norm": 0.20800863206386566, | |
| "learning_rate": 1.54337899543379e-05, | |
| "loss": 0.0316, | |
| "step": 5300 | |
| }, | |
| { | |
| "epoch": 24.246575342465754, | |
| "grad_norm": 0.1523219794034958, | |
| "learning_rate": 1.5368558382257013e-05, | |
| "loss": 0.0298, | |
| "step": 5310 | |
| }, | |
| { | |
| "epoch": 24.292237442922374, | |
| "grad_norm": 0.1901863068342209, | |
| "learning_rate": 1.530332681017613e-05, | |
| "loss": 0.0298, | |
| "step": 5320 | |
| }, | |
| { | |
| "epoch": 24.337899543378995, | |
| "grad_norm": 0.212814062833786, | |
| "learning_rate": 1.5238095238095241e-05, | |
| "loss": 0.0322, | |
| "step": 5330 | |
| }, | |
| { | |
| "epoch": 24.383561643835616, | |
| "grad_norm": 0.17524544894695282, | |
| "learning_rate": 1.5172863666014353e-05, | |
| "loss": 0.0313, | |
| "step": 5340 | |
| }, | |
| { | |
| "epoch": 24.429223744292237, | |
| "grad_norm": 0.10430175811052322, | |
| "learning_rate": 1.5107632093933465e-05, | |
| "loss": 0.0292, | |
| "step": 5350 | |
| }, | |
| { | |
| "epoch": 24.474885844748858, | |
| "grad_norm": 0.17600584030151367, | |
| "learning_rate": 1.504240052185258e-05, | |
| "loss": 0.0291, | |
| "step": 5360 | |
| }, | |
| { | |
| "epoch": 24.52054794520548, | |
| "grad_norm": 0.18452323973178864, | |
| "learning_rate": 1.497716894977169e-05, | |
| "loss": 0.0285, | |
| "step": 5370 | |
| }, | |
| { | |
| "epoch": 24.5662100456621, | |
| "grad_norm": 0.19338224828243256, | |
| "learning_rate": 1.4911937377690802e-05, | |
| "loss": 0.0309, | |
| "step": 5380 | |
| }, | |
| { | |
| "epoch": 24.61187214611872, | |
| "grad_norm": 0.10477393120527267, | |
| "learning_rate": 1.4846705805609914e-05, | |
| "loss": 0.0275, | |
| "step": 5390 | |
| }, | |
| { | |
| "epoch": 24.65753424657534, | |
| "grad_norm": 0.15825144946575165, | |
| "learning_rate": 1.4781474233529028e-05, | |
| "loss": 0.0307, | |
| "step": 5400 | |
| }, | |
| { | |
| "epoch": 24.703196347031962, | |
| "grad_norm": 0.1530427634716034, | |
| "learning_rate": 1.471624266144814e-05, | |
| "loss": 0.0282, | |
| "step": 5410 | |
| }, | |
| { | |
| "epoch": 24.748858447488583, | |
| "grad_norm": 0.17281366884708405, | |
| "learning_rate": 1.4651011089367254e-05, | |
| "loss": 0.0298, | |
| "step": 5420 | |
| }, | |
| { | |
| "epoch": 24.794520547945204, | |
| "grad_norm": 0.22076232731342316, | |
| "learning_rate": 1.4585779517286366e-05, | |
| "loss": 0.0294, | |
| "step": 5430 | |
| }, | |
| { | |
| "epoch": 24.840182648401825, | |
| "grad_norm": 0.16596472263336182, | |
| "learning_rate": 1.4520547945205478e-05, | |
| "loss": 0.0307, | |
| "step": 5440 | |
| }, | |
| { | |
| "epoch": 24.885844748858446, | |
| "grad_norm": 0.22020399570465088, | |
| "learning_rate": 1.4455316373124592e-05, | |
| "loss": 0.0286, | |
| "step": 5450 | |
| }, | |
| { | |
| "epoch": 24.931506849315067, | |
| "grad_norm": 0.15291544795036316, | |
| "learning_rate": 1.4390084801043704e-05, | |
| "loss": 0.0292, | |
| "step": 5460 | |
| }, | |
| { | |
| "epoch": 24.977168949771688, | |
| "grad_norm": 0.23781470954418182, | |
| "learning_rate": 1.4324853228962818e-05, | |
| "loss": 0.0309, | |
| "step": 5470 | |
| }, | |
| { | |
| "epoch": 25.0, | |
| "eval_bertscore_f1": 0.8788308417567078, | |
| "eval_bleu": 0.7686779949835236, | |
| "eval_loss": 0.04270849749445915, | |
| "eval_rougeL": 0.3022423509007887, | |
| "eval_runtime": 64.9101, | |
| "eval_samples_per_second": 23.155, | |
| "eval_steps_per_second": 1.448, | |
| "step": 5475 | |
| }, | |
| { | |
| "epoch": 25.022831050228312, | |
| "grad_norm": 0.20918726921081543, | |
| "learning_rate": 1.425962165688193e-05, | |
| "loss": 0.0306, | |
| "step": 5480 | |
| }, | |
| { | |
| "epoch": 25.068493150684933, | |
| "grad_norm": 0.11463820934295654, | |
| "learning_rate": 1.4194390084801042e-05, | |
| "loss": 0.0299, | |
| "step": 5490 | |
| }, | |
| { | |
| "epoch": 25.114155251141554, | |
| "grad_norm": 0.1545298844575882, | |
| "learning_rate": 1.4129158512720156e-05, | |
| "loss": 0.0311, | |
| "step": 5500 | |
| }, | |
| { | |
| "epoch": 25.159817351598175, | |
| "grad_norm": 0.13323558866977692, | |
| "learning_rate": 1.4063926940639269e-05, | |
| "loss": 0.0297, | |
| "step": 5510 | |
| }, | |
| { | |
| "epoch": 25.205479452054796, | |
| "grad_norm": 0.16467635333538055, | |
| "learning_rate": 1.3998695368558382e-05, | |
| "loss": 0.0313, | |
| "step": 5520 | |
| }, | |
| { | |
| "epoch": 25.251141552511417, | |
| "grad_norm": 0.19874680042266846, | |
| "learning_rate": 1.3933463796477495e-05, | |
| "loss": 0.0311, | |
| "step": 5530 | |
| }, | |
| { | |
| "epoch": 25.296803652968038, | |
| "grad_norm": 0.1181570440530777, | |
| "learning_rate": 1.3868232224396607e-05, | |
| "loss": 0.0284, | |
| "step": 5540 | |
| }, | |
| { | |
| "epoch": 25.34246575342466, | |
| "grad_norm": 0.13484494388103485, | |
| "learning_rate": 1.380300065231572e-05, | |
| "loss": 0.0321, | |
| "step": 5550 | |
| }, | |
| { | |
| "epoch": 25.38812785388128, | |
| "grad_norm": 0.17654231190681458, | |
| "learning_rate": 1.3737769080234833e-05, | |
| "loss": 0.0297, | |
| "step": 5560 | |
| }, | |
| { | |
| "epoch": 25.4337899543379, | |
| "grad_norm": 0.3322019875049591, | |
| "learning_rate": 1.3672537508153947e-05, | |
| "loss": 0.0297, | |
| "step": 5570 | |
| }, | |
| { | |
| "epoch": 25.47945205479452, | |
| "grad_norm": 0.12533101439476013, | |
| "learning_rate": 1.3607305936073059e-05, | |
| "loss": 0.0332, | |
| "step": 5580 | |
| }, | |
| { | |
| "epoch": 25.525114155251142, | |
| "grad_norm": 0.15482525527477264, | |
| "learning_rate": 1.3542074363992171e-05, | |
| "loss": 0.0292, | |
| "step": 5590 | |
| }, | |
| { | |
| "epoch": 25.570776255707763, | |
| "grad_norm": 0.16881971061229706, | |
| "learning_rate": 1.3476842791911285e-05, | |
| "loss": 0.0313, | |
| "step": 5600 | |
| }, | |
| { | |
| "epoch": 25.616438356164384, | |
| "grad_norm": 0.17732620239257812, | |
| "learning_rate": 1.3411611219830397e-05, | |
| "loss": 0.0293, | |
| "step": 5610 | |
| }, | |
| { | |
| "epoch": 25.662100456621005, | |
| "grad_norm": 0.184401273727417, | |
| "learning_rate": 1.3346379647749511e-05, | |
| "loss": 0.0293, | |
| "step": 5620 | |
| }, | |
| { | |
| "epoch": 25.707762557077626, | |
| "grad_norm": 0.16654548048973083, | |
| "learning_rate": 1.3281148075668623e-05, | |
| "loss": 0.0295, | |
| "step": 5630 | |
| }, | |
| { | |
| "epoch": 25.753424657534246, | |
| "grad_norm": 0.19625544548034668, | |
| "learning_rate": 1.3215916503587735e-05, | |
| "loss": 0.0299, | |
| "step": 5640 | |
| }, | |
| { | |
| "epoch": 25.799086757990867, | |
| "grad_norm": 0.18322430551052094, | |
| "learning_rate": 1.3150684931506849e-05, | |
| "loss": 0.0306, | |
| "step": 5650 | |
| }, | |
| { | |
| "epoch": 25.84474885844749, | |
| "grad_norm": 0.16067004203796387, | |
| "learning_rate": 1.3085453359425961e-05, | |
| "loss": 0.0303, | |
| "step": 5660 | |
| }, | |
| { | |
| "epoch": 25.89041095890411, | |
| "grad_norm": 0.13349705934524536, | |
| "learning_rate": 1.3020221787345075e-05, | |
| "loss": 0.0279, | |
| "step": 5670 | |
| }, | |
| { | |
| "epoch": 25.93607305936073, | |
| "grad_norm": 0.19681212306022644, | |
| "learning_rate": 1.2954990215264187e-05, | |
| "loss": 0.0304, | |
| "step": 5680 | |
| }, | |
| { | |
| "epoch": 25.98173515981735, | |
| "grad_norm": 0.305829256772995, | |
| "learning_rate": 1.28897586431833e-05, | |
| "loss": 0.0292, | |
| "step": 5690 | |
| }, | |
| { | |
| "epoch": 26.0, | |
| "eval_bertscore_f1": 0.8774845823794306, | |
| "eval_bleu": 0.6514875461109689, | |
| "eval_loss": 0.042685650289058685, | |
| "eval_rougeL": 0.3085123872128879, | |
| "eval_runtime": 65.3563, | |
| "eval_samples_per_second": 22.997, | |
| "eval_steps_per_second": 1.438, | |
| "step": 5694 | |
| }, | |
| { | |
| "epoch": 26.027397260273972, | |
| "grad_norm": 0.227409228682518, | |
| "learning_rate": 1.2824527071102413e-05, | |
| "loss": 0.0299, | |
| "step": 5700 | |
| }, | |
| { | |
| "epoch": 26.073059360730593, | |
| "grad_norm": 0.17818889021873474, | |
| "learning_rate": 1.2759295499021525e-05, | |
| "loss": 0.03, | |
| "step": 5710 | |
| }, | |
| { | |
| "epoch": 26.118721461187214, | |
| "grad_norm": 0.13791914284229279, | |
| "learning_rate": 1.269406392694064e-05, | |
| "loss": 0.029, | |
| "step": 5720 | |
| }, | |
| { | |
| "epoch": 26.164383561643834, | |
| "grad_norm": 0.16791503131389618, | |
| "learning_rate": 1.2628832354859752e-05, | |
| "loss": 0.0308, | |
| "step": 5730 | |
| }, | |
| { | |
| "epoch": 26.210045662100455, | |
| "grad_norm": 0.2646624445915222, | |
| "learning_rate": 1.2563600782778864e-05, | |
| "loss": 0.0303, | |
| "step": 5740 | |
| }, | |
| { | |
| "epoch": 26.255707762557076, | |
| "grad_norm": 0.19949132204055786, | |
| "learning_rate": 1.249836921069798e-05, | |
| "loss": 0.0304, | |
| "step": 5750 | |
| }, | |
| { | |
| "epoch": 26.301369863013697, | |
| "grad_norm": 0.14131474494934082, | |
| "learning_rate": 1.2433137638617091e-05, | |
| "loss": 0.0304, | |
| "step": 5760 | |
| }, | |
| { | |
| "epoch": 26.347031963470318, | |
| "grad_norm": 0.13681089878082275, | |
| "learning_rate": 1.2367906066536204e-05, | |
| "loss": 0.0293, | |
| "step": 5770 | |
| }, | |
| { | |
| "epoch": 26.39269406392694, | |
| "grad_norm": 0.18042202293872833, | |
| "learning_rate": 1.2302674494455317e-05, | |
| "loss": 0.0298, | |
| "step": 5780 | |
| }, | |
| { | |
| "epoch": 26.438356164383563, | |
| "grad_norm": 0.10717923194169998, | |
| "learning_rate": 1.223744292237443e-05, | |
| "loss": 0.0275, | |
| "step": 5790 | |
| }, | |
| { | |
| "epoch": 26.484018264840184, | |
| "grad_norm": 0.1440982073545456, | |
| "learning_rate": 1.2172211350293543e-05, | |
| "loss": 0.0313, | |
| "step": 5800 | |
| }, | |
| { | |
| "epoch": 26.529680365296805, | |
| "grad_norm": 0.21595457196235657, | |
| "learning_rate": 1.2106979778212656e-05, | |
| "loss": 0.0306, | |
| "step": 5810 | |
| }, | |
| { | |
| "epoch": 26.575342465753426, | |
| "grad_norm": 0.18737435340881348, | |
| "learning_rate": 1.2041748206131768e-05, | |
| "loss": 0.0295, | |
| "step": 5820 | |
| }, | |
| { | |
| "epoch": 26.621004566210047, | |
| "grad_norm": 0.14051635563373566, | |
| "learning_rate": 1.1976516634050882e-05, | |
| "loss": 0.031, | |
| "step": 5830 | |
| }, | |
| { | |
| "epoch": 26.666666666666668, | |
| "grad_norm": 0.23766343295574188, | |
| "learning_rate": 1.1911285061969994e-05, | |
| "loss": 0.0298, | |
| "step": 5840 | |
| }, | |
| { | |
| "epoch": 26.71232876712329, | |
| "grad_norm": 0.19277657568454742, | |
| "learning_rate": 1.1846053489889108e-05, | |
| "loss": 0.0292, | |
| "step": 5850 | |
| }, | |
| { | |
| "epoch": 26.75799086757991, | |
| "grad_norm": 0.2688007354736328, | |
| "learning_rate": 1.178082191780822e-05, | |
| "loss": 0.0304, | |
| "step": 5860 | |
| }, | |
| { | |
| "epoch": 26.80365296803653, | |
| "grad_norm": 0.1929878294467926, | |
| "learning_rate": 1.1715590345727332e-05, | |
| "loss": 0.028, | |
| "step": 5870 | |
| }, | |
| { | |
| "epoch": 26.84931506849315, | |
| "grad_norm": 0.15954306721687317, | |
| "learning_rate": 1.1650358773646446e-05, | |
| "loss": 0.0297, | |
| "step": 5880 | |
| }, | |
| { | |
| "epoch": 26.894977168949772, | |
| "grad_norm": 0.19489558041095734, | |
| "learning_rate": 1.1585127201565558e-05, | |
| "loss": 0.0322, | |
| "step": 5890 | |
| }, | |
| { | |
| "epoch": 26.940639269406393, | |
| "grad_norm": 0.1502341479063034, | |
| "learning_rate": 1.1519895629484672e-05, | |
| "loss": 0.0286, | |
| "step": 5900 | |
| }, | |
| { | |
| "epoch": 26.986301369863014, | |
| "grad_norm": 0.2791164219379425, | |
| "learning_rate": 1.1454664057403784e-05, | |
| "loss": 0.0318, | |
| "step": 5910 | |
| }, | |
| { | |
| "epoch": 27.0, | |
| "eval_bertscore_f1": 0.8776480831944459, | |
| "eval_bleu": 0.6814688964486258, | |
| "eval_loss": 0.0420895554125309, | |
| "eval_rougeL": 0.30520018097920826, | |
| "eval_runtime": 65.2587, | |
| "eval_samples_per_second": 23.031, | |
| "eval_steps_per_second": 1.44, | |
| "step": 5913 | |
| }, | |
| { | |
| "epoch": 27.031963470319635, | |
| "grad_norm": 0.17120176553726196, | |
| "learning_rate": 1.1389432485322896e-05, | |
| "loss": 0.0291, | |
| "step": 5920 | |
| }, | |
| { | |
| "epoch": 27.077625570776256, | |
| "grad_norm": 0.15285202860832214, | |
| "learning_rate": 1.132420091324201e-05, | |
| "loss": 0.0287, | |
| "step": 5930 | |
| }, | |
| { | |
| "epoch": 27.123287671232877, | |
| "grad_norm": 0.15016716718673706, | |
| "learning_rate": 1.1258969341161122e-05, | |
| "loss": 0.0297, | |
| "step": 5940 | |
| }, | |
| { | |
| "epoch": 27.168949771689498, | |
| "grad_norm": 0.16641657054424286, | |
| "learning_rate": 1.1193737769080236e-05, | |
| "loss": 0.0311, | |
| "step": 5950 | |
| }, | |
| { | |
| "epoch": 27.21461187214612, | |
| "grad_norm": 0.24323779344558716, | |
| "learning_rate": 1.1128506196999348e-05, | |
| "loss": 0.0309, | |
| "step": 5960 | |
| }, | |
| { | |
| "epoch": 27.26027397260274, | |
| "grad_norm": 0.18576470017433167, | |
| "learning_rate": 1.106327462491846e-05, | |
| "loss": 0.0285, | |
| "step": 5970 | |
| }, | |
| { | |
| "epoch": 27.30593607305936, | |
| "grad_norm": 0.19552282989025116, | |
| "learning_rate": 1.0998043052837574e-05, | |
| "loss": 0.0281, | |
| "step": 5980 | |
| }, | |
| { | |
| "epoch": 27.35159817351598, | |
| "grad_norm": 0.192937433719635, | |
| "learning_rate": 1.0932811480756687e-05, | |
| "loss": 0.0292, | |
| "step": 5990 | |
| }, | |
| { | |
| "epoch": 27.397260273972602, | |
| "grad_norm": 0.19229499995708466, | |
| "learning_rate": 1.08675799086758e-05, | |
| "loss": 0.0287, | |
| "step": 6000 | |
| }, | |
| { | |
| "epoch": 27.442922374429223, | |
| "grad_norm": 0.2223975658416748, | |
| "learning_rate": 1.0802348336594913e-05, | |
| "loss": 0.0309, | |
| "step": 6010 | |
| }, | |
| { | |
| "epoch": 27.488584474885844, | |
| "grad_norm": 0.41968482732772827, | |
| "learning_rate": 1.0737116764514025e-05, | |
| "loss": 0.0332, | |
| "step": 6020 | |
| }, | |
| { | |
| "epoch": 27.534246575342465, | |
| "grad_norm": 0.36680567264556885, | |
| "learning_rate": 1.0671885192433139e-05, | |
| "loss": 0.0291, | |
| "step": 6030 | |
| }, | |
| { | |
| "epoch": 27.579908675799086, | |
| "grad_norm": 0.1873111128807068, | |
| "learning_rate": 1.0606653620352251e-05, | |
| "loss": 0.0282, | |
| "step": 6040 | |
| }, | |
| { | |
| "epoch": 27.625570776255707, | |
| "grad_norm": 0.21136601269245148, | |
| "learning_rate": 1.0541422048271365e-05, | |
| "loss": 0.0285, | |
| "step": 6050 | |
| }, | |
| { | |
| "epoch": 27.671232876712327, | |
| "grad_norm": 0.15826620161533356, | |
| "learning_rate": 1.0476190476190477e-05, | |
| "loss": 0.0291, | |
| "step": 6060 | |
| }, | |
| { | |
| "epoch": 27.71689497716895, | |
| "grad_norm": 0.16135139763355255, | |
| "learning_rate": 1.0410958904109589e-05, | |
| "loss": 0.0312, | |
| "step": 6070 | |
| }, | |
| { | |
| "epoch": 27.76255707762557, | |
| "grad_norm": 0.18225039541721344, | |
| "learning_rate": 1.0345727332028703e-05, | |
| "loss": 0.03, | |
| "step": 6080 | |
| }, | |
| { | |
| "epoch": 27.80821917808219, | |
| "grad_norm": 0.19162273406982422, | |
| "learning_rate": 1.0280495759947815e-05, | |
| "loss": 0.0285, | |
| "step": 6090 | |
| }, | |
| { | |
| "epoch": 27.853881278538815, | |
| "grad_norm": 0.2653762102127075, | |
| "learning_rate": 1.0215264187866929e-05, | |
| "loss": 0.0318, | |
| "step": 6100 | |
| }, | |
| { | |
| "epoch": 27.899543378995435, | |
| "grad_norm": 0.258783757686615, | |
| "learning_rate": 1.0150032615786041e-05, | |
| "loss": 0.0286, | |
| "step": 6110 | |
| }, | |
| { | |
| "epoch": 27.945205479452056, | |
| "grad_norm": 0.18329358100891113, | |
| "learning_rate": 1.0084801043705153e-05, | |
| "loss": 0.0299, | |
| "step": 6120 | |
| }, | |
| { | |
| "epoch": 27.990867579908677, | |
| "grad_norm": 0.19685736298561096, | |
| "learning_rate": 1.0019569471624267e-05, | |
| "loss": 0.0317, | |
| "step": 6130 | |
| }, | |
| { | |
| "epoch": 28.0, | |
| "eval_bertscore_f1": 0.8782506199653991, | |
| "eval_bleu": 0.7977746926944412, | |
| "eval_loss": 0.042202867567539215, | |
| "eval_rougeL": 0.2995364849769159, | |
| "eval_runtime": 65.0364, | |
| "eval_samples_per_second": 23.11, | |
| "eval_steps_per_second": 1.445, | |
| "step": 6132 | |
| }, | |
| { | |
| "epoch": 28.036529680365298, | |
| "grad_norm": 0.1952570676803589, | |
| "learning_rate": 9.95433789954338e-06, | |
| "loss": 0.0308, | |
| "step": 6140 | |
| }, | |
| { | |
| "epoch": 28.08219178082192, | |
| "grad_norm": 0.14125581085681915, | |
| "learning_rate": 9.889106327462493e-06, | |
| "loss": 0.0283, | |
| "step": 6150 | |
| }, | |
| { | |
| "epoch": 28.12785388127854, | |
| "grad_norm": 0.1538328379392624, | |
| "learning_rate": 9.823874755381605e-06, | |
| "loss": 0.0307, | |
| "step": 6160 | |
| }, | |
| { | |
| "epoch": 28.17351598173516, | |
| "grad_norm": 0.16571789979934692, | |
| "learning_rate": 9.758643183300718e-06, | |
| "loss": 0.0306, | |
| "step": 6170 | |
| }, | |
| { | |
| "epoch": 28.21917808219178, | |
| "grad_norm": 0.2275710552930832, | |
| "learning_rate": 9.693411611219831e-06, | |
| "loss": 0.0298, | |
| "step": 6180 | |
| }, | |
| { | |
| "epoch": 28.264840182648403, | |
| "grad_norm": 0.16344133019447327, | |
| "learning_rate": 9.628180039138944e-06, | |
| "loss": 0.0292, | |
| "step": 6190 | |
| }, | |
| { | |
| "epoch": 28.310502283105023, | |
| "grad_norm": 0.15537413954734802, | |
| "learning_rate": 9.562948467058057e-06, | |
| "loss": 0.0297, | |
| "step": 6200 | |
| }, | |
| { | |
| "epoch": 28.356164383561644, | |
| "grad_norm": 0.17001962661743164, | |
| "learning_rate": 9.49771689497717e-06, | |
| "loss": 0.0313, | |
| "step": 6210 | |
| }, | |
| { | |
| "epoch": 28.401826484018265, | |
| "grad_norm": 0.17238697409629822, | |
| "learning_rate": 9.432485322896282e-06, | |
| "loss": 0.0303, | |
| "step": 6220 | |
| }, | |
| { | |
| "epoch": 28.447488584474886, | |
| "grad_norm": 0.34354132413864136, | |
| "learning_rate": 9.367253750815396e-06, | |
| "loss": 0.0289, | |
| "step": 6230 | |
| }, | |
| { | |
| "epoch": 28.493150684931507, | |
| "grad_norm": 0.10073091834783554, | |
| "learning_rate": 9.302022178734508e-06, | |
| "loss": 0.0303, | |
| "step": 6240 | |
| }, | |
| { | |
| "epoch": 28.538812785388128, | |
| "grad_norm": 0.1798088103532791, | |
| "learning_rate": 9.236790606653622e-06, | |
| "loss": 0.0283, | |
| "step": 6250 | |
| }, | |
| { | |
| "epoch": 28.58447488584475, | |
| "grad_norm": 0.1783922165632248, | |
| "learning_rate": 9.171559034572734e-06, | |
| "loss": 0.0294, | |
| "step": 6260 | |
| }, | |
| { | |
| "epoch": 28.63013698630137, | |
| "grad_norm": 0.177333265542984, | |
| "learning_rate": 9.106327462491846e-06, | |
| "loss": 0.0308, | |
| "step": 6270 | |
| }, | |
| { | |
| "epoch": 28.67579908675799, | |
| "grad_norm": 0.1782556176185608, | |
| "learning_rate": 9.04109589041096e-06, | |
| "loss": 0.0313, | |
| "step": 6280 | |
| }, | |
| { | |
| "epoch": 28.72146118721461, | |
| "grad_norm": 0.20494292676448822, | |
| "learning_rate": 8.975864318330072e-06, | |
| "loss": 0.0272, | |
| "step": 6290 | |
| }, | |
| { | |
| "epoch": 28.767123287671232, | |
| "grad_norm": 0.32170209288597107, | |
| "learning_rate": 8.910632746249186e-06, | |
| "loss": 0.0289, | |
| "step": 6300 | |
| }, | |
| { | |
| "epoch": 28.812785388127853, | |
| "grad_norm": 0.14453481137752533, | |
| "learning_rate": 8.845401174168298e-06, | |
| "loss": 0.0288, | |
| "step": 6310 | |
| }, | |
| { | |
| "epoch": 28.858447488584474, | |
| "grad_norm": 0.15864479541778564, | |
| "learning_rate": 8.78016960208741e-06, | |
| "loss": 0.0294, | |
| "step": 6320 | |
| }, | |
| { | |
| "epoch": 28.904109589041095, | |
| "grad_norm": 0.1791839748620987, | |
| "learning_rate": 8.714938030006524e-06, | |
| "loss": 0.029, | |
| "step": 6330 | |
| }, | |
| { | |
| "epoch": 28.949771689497716, | |
| "grad_norm": 0.21326766908168793, | |
| "learning_rate": 8.649706457925636e-06, | |
| "loss": 0.0288, | |
| "step": 6340 | |
| }, | |
| { | |
| "epoch": 28.995433789954337, | |
| "grad_norm": 0.18078750371932983, | |
| "learning_rate": 8.58447488584475e-06, | |
| "loss": 0.0284, | |
| "step": 6350 | |
| }, | |
| { | |
| "epoch": 29.0, | |
| "eval_bertscore_f1": 0.8773477768311085, | |
| "eval_bleu": 0.8007310925138926, | |
| "eval_loss": 0.04212983325123787, | |
| "eval_rougeL": 0.29792348880513014, | |
| "eval_runtime": 64.8542, | |
| "eval_samples_per_second": 23.175, | |
| "eval_steps_per_second": 1.449, | |
| "step": 6351 | |
| }, | |
| { | |
| "epoch": 29.041095890410958, | |
| "grad_norm": 0.3029780983924866, | |
| "learning_rate": 8.519243313763862e-06, | |
| "loss": 0.0278, | |
| "step": 6360 | |
| }, | |
| { | |
| "epoch": 29.08675799086758, | |
| "grad_norm": 0.21092386543750763, | |
| "learning_rate": 8.454011741682975e-06, | |
| "loss": 0.0302, | |
| "step": 6370 | |
| }, | |
| { | |
| "epoch": 29.1324200913242, | |
| "grad_norm": 0.20935019850730896, | |
| "learning_rate": 8.388780169602088e-06, | |
| "loss": 0.0283, | |
| "step": 6380 | |
| }, | |
| { | |
| "epoch": 29.17808219178082, | |
| "grad_norm": 0.1683349758386612, | |
| "learning_rate": 8.3235485975212e-06, | |
| "loss": 0.0296, | |
| "step": 6390 | |
| }, | |
| { | |
| "epoch": 29.22374429223744, | |
| "grad_norm": 0.15146371722221375, | |
| "learning_rate": 8.258317025440314e-06, | |
| "loss": 0.0281, | |
| "step": 6400 | |
| }, | |
| { | |
| "epoch": 29.269406392694062, | |
| "grad_norm": 0.1463659703731537, | |
| "learning_rate": 8.193085453359427e-06, | |
| "loss": 0.0306, | |
| "step": 6410 | |
| }, | |
| { | |
| "epoch": 29.315068493150687, | |
| "grad_norm": 0.22994855046272278, | |
| "learning_rate": 8.127853881278539e-06, | |
| "loss": 0.0326, | |
| "step": 6420 | |
| }, | |
| { | |
| "epoch": 29.360730593607308, | |
| "grad_norm": 0.18445008993148804, | |
| "learning_rate": 8.062622309197653e-06, | |
| "loss": 0.0297, | |
| "step": 6430 | |
| }, | |
| { | |
| "epoch": 29.40639269406393, | |
| "grad_norm": 0.18058903515338898, | |
| "learning_rate": 7.997390737116765e-06, | |
| "loss": 0.0302, | |
| "step": 6440 | |
| }, | |
| { | |
| "epoch": 29.45205479452055, | |
| "grad_norm": 0.17116139829158783, | |
| "learning_rate": 7.932159165035879e-06, | |
| "loss": 0.0303, | |
| "step": 6450 | |
| }, | |
| { | |
| "epoch": 29.49771689497717, | |
| "grad_norm": 0.18373791873455048, | |
| "learning_rate": 7.86692759295499e-06, | |
| "loss": 0.0266, | |
| "step": 6460 | |
| }, | |
| { | |
| "epoch": 29.54337899543379, | |
| "grad_norm": 0.18657907843589783, | |
| "learning_rate": 7.801696020874103e-06, | |
| "loss": 0.0303, | |
| "step": 6470 | |
| }, | |
| { | |
| "epoch": 29.589041095890412, | |
| "grad_norm": 0.17647583782672882, | |
| "learning_rate": 7.736464448793217e-06, | |
| "loss": 0.0277, | |
| "step": 6480 | |
| }, | |
| { | |
| "epoch": 29.634703196347033, | |
| "grad_norm": 0.14070016145706177, | |
| "learning_rate": 7.671232876712329e-06, | |
| "loss": 0.0295, | |
| "step": 6490 | |
| }, | |
| { | |
| "epoch": 29.680365296803654, | |
| "grad_norm": 0.1862981617450714, | |
| "learning_rate": 7.606001304631442e-06, | |
| "loss": 0.0284, | |
| "step": 6500 | |
| }, | |
| { | |
| "epoch": 29.726027397260275, | |
| "grad_norm": 0.22975805401802063, | |
| "learning_rate": 7.540769732550555e-06, | |
| "loss": 0.0292, | |
| "step": 6510 | |
| }, | |
| { | |
| "epoch": 29.771689497716896, | |
| "grad_norm": 0.31355002522468567, | |
| "learning_rate": 7.475538160469668e-06, | |
| "loss": 0.0293, | |
| "step": 6520 | |
| }, | |
| { | |
| "epoch": 29.817351598173516, | |
| "grad_norm": 0.18986040353775024, | |
| "learning_rate": 7.410306588388781e-06, | |
| "loss": 0.0279, | |
| "step": 6530 | |
| }, | |
| { | |
| "epoch": 29.863013698630137, | |
| "grad_norm": 0.21212980151176453, | |
| "learning_rate": 7.345075016307894e-06, | |
| "loss": 0.029, | |
| "step": 6540 | |
| }, | |
| { | |
| "epoch": 29.908675799086758, | |
| "grad_norm": 0.23085573315620422, | |
| "learning_rate": 7.279843444227006e-06, | |
| "loss": 0.0293, | |
| "step": 6550 | |
| }, | |
| { | |
| "epoch": 29.95433789954338, | |
| "grad_norm": 0.19988933205604553, | |
| "learning_rate": 7.214611872146119e-06, | |
| "loss": 0.0305, | |
| "step": 6560 | |
| }, | |
| { | |
| "epoch": 30.0, | |
| "grad_norm": 0.19826553761959076, | |
| "learning_rate": 7.149380300065232e-06, | |
| "loss": 0.028, | |
| "step": 6570 | |
| }, | |
| { | |
| "epoch": 30.0, | |
| "eval_bertscore_f1": 0.8785085355291665, | |
| "eval_bleu": 0.702881238096362, | |
| "eval_loss": 0.04208363965153694, | |
| "eval_rougeL": 0.3061862473643293, | |
| "eval_runtime": 64.8873, | |
| "eval_samples_per_second": 23.163, | |
| "eval_steps_per_second": 1.449, | |
| "step": 6570 | |
| }, | |
| { | |
| "epoch": 30.04566210045662, | |
| "grad_norm": 0.14557534456253052, | |
| "learning_rate": 7.084148727984345e-06, | |
| "loss": 0.0288, | |
| "step": 6580 | |
| }, | |
| { | |
| "epoch": 30.091324200913242, | |
| "grad_norm": 0.1587667614221573, | |
| "learning_rate": 7.018917155903458e-06, | |
| "loss": 0.0296, | |
| "step": 6590 | |
| }, | |
| { | |
| "epoch": 30.136986301369863, | |
| "grad_norm": 0.13343170285224915, | |
| "learning_rate": 6.9536855838225706e-06, | |
| "loss": 0.0289, | |
| "step": 6600 | |
| }, | |
| { | |
| "epoch": 30.182648401826484, | |
| "grad_norm": 0.1827743649482727, | |
| "learning_rate": 6.8884540117416836e-06, | |
| "loss": 0.0289, | |
| "step": 6610 | |
| }, | |
| { | |
| "epoch": 30.228310502283104, | |
| "grad_norm": 0.26161623001098633, | |
| "learning_rate": 6.823222439660797e-06, | |
| "loss": 0.0305, | |
| "step": 6620 | |
| }, | |
| { | |
| "epoch": 30.273972602739725, | |
| "grad_norm": 0.1480785757303238, | |
| "learning_rate": 6.75799086757991e-06, | |
| "loss": 0.0277, | |
| "step": 6630 | |
| }, | |
| { | |
| "epoch": 30.319634703196346, | |
| "grad_norm": 0.21986277401447296, | |
| "learning_rate": 6.692759295499023e-06, | |
| "loss": 0.0301, | |
| "step": 6640 | |
| }, | |
| { | |
| "epoch": 30.365296803652967, | |
| "grad_norm": 0.21434344351291656, | |
| "learning_rate": 6.627527723418135e-06, | |
| "loss": 0.0307, | |
| "step": 6650 | |
| }, | |
| { | |
| "epoch": 30.410958904109588, | |
| "grad_norm": 0.1664654165506363, | |
| "learning_rate": 6.562296151337248e-06, | |
| "loss": 0.0296, | |
| "step": 6660 | |
| }, | |
| { | |
| "epoch": 30.45662100456621, | |
| "grad_norm": 0.21170803904533386, | |
| "learning_rate": 6.497064579256361e-06, | |
| "loss": 0.0281, | |
| "step": 6670 | |
| }, | |
| { | |
| "epoch": 30.50228310502283, | |
| "grad_norm": 0.2334418147802353, | |
| "learning_rate": 6.431833007175474e-06, | |
| "loss": 0.0295, | |
| "step": 6680 | |
| }, | |
| { | |
| "epoch": 30.54794520547945, | |
| "grad_norm": 0.1512773334980011, | |
| "learning_rate": 6.366601435094587e-06, | |
| "loss": 0.0293, | |
| "step": 6690 | |
| }, | |
| { | |
| "epoch": 30.59360730593607, | |
| "grad_norm": 0.1553216278553009, | |
| "learning_rate": 6.301369863013699e-06, | |
| "loss": 0.0292, | |
| "step": 6700 | |
| }, | |
| { | |
| "epoch": 30.639269406392692, | |
| "grad_norm": 0.14663270115852356, | |
| "learning_rate": 6.236138290932811e-06, | |
| "loss": 0.0277, | |
| "step": 6710 | |
| }, | |
| { | |
| "epoch": 30.684931506849313, | |
| "grad_norm": 0.22883833944797516, | |
| "learning_rate": 6.170906718851924e-06, | |
| "loss": 0.0294, | |
| "step": 6720 | |
| }, | |
| { | |
| "epoch": 30.730593607305934, | |
| "grad_norm": 0.19154661893844604, | |
| "learning_rate": 6.105675146771037e-06, | |
| "loss": 0.0275, | |
| "step": 6730 | |
| }, | |
| { | |
| "epoch": 30.77625570776256, | |
| "grad_norm": 0.1744658201932907, | |
| "learning_rate": 6.04044357469015e-06, | |
| "loss": 0.0282, | |
| "step": 6740 | |
| }, | |
| { | |
| "epoch": 30.82191780821918, | |
| "grad_norm": 0.13991741836071014, | |
| "learning_rate": 5.975212002609263e-06, | |
| "loss": 0.0301, | |
| "step": 6750 | |
| }, | |
| { | |
| "epoch": 30.8675799086758, | |
| "grad_norm": 0.2106420397758484, | |
| "learning_rate": 5.9099804305283755e-06, | |
| "loss": 0.0297, | |
| "step": 6760 | |
| }, | |
| { | |
| "epoch": 30.91324200913242, | |
| "grad_norm": 0.1574612706899643, | |
| "learning_rate": 5.8447488584474885e-06, | |
| "loss": 0.0291, | |
| "step": 6770 | |
| }, | |
| { | |
| "epoch": 30.958904109589042, | |
| "grad_norm": 0.2075018733739853, | |
| "learning_rate": 5.7795172863666015e-06, | |
| "loss": 0.0268, | |
| "step": 6780 | |
| }, | |
| { | |
| "epoch": 31.0, | |
| "eval_bertscore_f1": 0.8759105591399624, | |
| "eval_bleu": 0.6233413223757481, | |
| "eval_loss": 0.04237150773406029, | |
| "eval_rougeL": 0.3068191295790029, | |
| "eval_runtime": 65.4719, | |
| "eval_samples_per_second": 22.956, | |
| "eval_steps_per_second": 1.436, | |
| "step": 6789 | |
| }, | |
| { | |
| "epoch": 31.004566210045663, | |
| "grad_norm": 0.2037985920906067, | |
| "learning_rate": 5.7142857142857145e-06, | |
| "loss": 0.03, | |
| "step": 6790 | |
| }, | |
| { | |
| "epoch": 31.050228310502284, | |
| "grad_norm": 0.2403508424758911, | |
| "learning_rate": 5.6490541422048275e-06, | |
| "loss": 0.0287, | |
| "step": 6800 | |
| }, | |
| { | |
| "epoch": 31.095890410958905, | |
| "grad_norm": 0.16700613498687744, | |
| "learning_rate": 5.58382257012394e-06, | |
| "loss": 0.0305, | |
| "step": 6810 | |
| }, | |
| { | |
| "epoch": 31.141552511415526, | |
| "grad_norm": 0.18683423101902008, | |
| "learning_rate": 5.518590998043053e-06, | |
| "loss": 0.0297, | |
| "step": 6820 | |
| }, | |
| { | |
| "epoch": 31.187214611872147, | |
| "grad_norm": 0.1736307293176651, | |
| "learning_rate": 5.453359425962166e-06, | |
| "loss": 0.0287, | |
| "step": 6830 | |
| }, | |
| { | |
| "epoch": 31.232876712328768, | |
| "grad_norm": 0.4329126477241516, | |
| "learning_rate": 5.388127853881279e-06, | |
| "loss": 0.0286, | |
| "step": 6840 | |
| }, | |
| { | |
| "epoch": 31.27853881278539, | |
| "grad_norm": 0.19322291016578674, | |
| "learning_rate": 5.322896281800392e-06, | |
| "loss": 0.0299, | |
| "step": 6850 | |
| }, | |
| { | |
| "epoch": 31.32420091324201, | |
| "grad_norm": 0.12365171313285828, | |
| "learning_rate": 5.257664709719504e-06, | |
| "loss": 0.0288, | |
| "step": 6860 | |
| }, | |
| { | |
| "epoch": 31.36986301369863, | |
| "grad_norm": 0.2190866321325302, | |
| "learning_rate": 5.192433137638617e-06, | |
| "loss": 0.0287, | |
| "step": 6870 | |
| }, | |
| { | |
| "epoch": 31.41552511415525, | |
| "grad_norm": 0.1890123039484024, | |
| "learning_rate": 5.12720156555773e-06, | |
| "loss": 0.0292, | |
| "step": 6880 | |
| }, | |
| { | |
| "epoch": 31.461187214611872, | |
| "grad_norm": 0.21521024405956268, | |
| "learning_rate": 5.061969993476843e-06, | |
| "loss": 0.0286, | |
| "step": 6890 | |
| }, | |
| { | |
| "epoch": 31.506849315068493, | |
| "grad_norm": 0.15392209589481354, | |
| "learning_rate": 4.996738421395956e-06, | |
| "loss": 0.0279, | |
| "step": 6900 | |
| }, | |
| { | |
| "epoch": 31.552511415525114, | |
| "grad_norm": 0.1591772735118866, | |
| "learning_rate": 4.931506849315068e-06, | |
| "loss": 0.0304, | |
| "step": 6910 | |
| }, | |
| { | |
| "epoch": 31.598173515981735, | |
| "grad_norm": 0.18570680916309357, | |
| "learning_rate": 4.866275277234181e-06, | |
| "loss": 0.0292, | |
| "step": 6920 | |
| }, | |
| { | |
| "epoch": 31.643835616438356, | |
| "grad_norm": 0.21021844446659088, | |
| "learning_rate": 4.801043705153294e-06, | |
| "loss": 0.03, | |
| "step": 6930 | |
| }, | |
| { | |
| "epoch": 31.689497716894977, | |
| "grad_norm": 0.11242254078388214, | |
| "learning_rate": 4.735812133072407e-06, | |
| "loss": 0.0284, | |
| "step": 6940 | |
| }, | |
| { | |
| "epoch": 31.735159817351597, | |
| "grad_norm": 0.12586183845996857, | |
| "learning_rate": 4.67058056099152e-06, | |
| "loss": 0.0315, | |
| "step": 6950 | |
| }, | |
| { | |
| "epoch": 31.78082191780822, | |
| "grad_norm": 0.2376343458890915, | |
| "learning_rate": 4.6053489889106324e-06, | |
| "loss": 0.0268, | |
| "step": 6960 | |
| }, | |
| { | |
| "epoch": 31.82648401826484, | |
| "grad_norm": 0.10612351447343826, | |
| "learning_rate": 4.5401174168297455e-06, | |
| "loss": 0.0274, | |
| "step": 6970 | |
| }, | |
| { | |
| "epoch": 31.87214611872146, | |
| "grad_norm": 0.16219161450862885, | |
| "learning_rate": 4.4748858447488585e-06, | |
| "loss": 0.0291, | |
| "step": 6980 | |
| }, | |
| { | |
| "epoch": 31.91780821917808, | |
| "grad_norm": 0.22465045750141144, | |
| "learning_rate": 4.4096542726679715e-06, | |
| "loss": 0.0293, | |
| "step": 6990 | |
| }, | |
| { | |
| "epoch": 31.963470319634702, | |
| "grad_norm": 0.17077922821044922, | |
| "learning_rate": 4.3444227005870845e-06, | |
| "loss": 0.0278, | |
| "step": 7000 | |
| }, | |
| { | |
| "epoch": 32.0, | |
| "eval_bertscore_f1": 0.8782741467078685, | |
| "eval_bleu": 0.7519832650589673, | |
| "eval_loss": 0.04253660514950752, | |
| "eval_rougeL": 0.30201395092264444, | |
| "eval_runtime": 64.8813, | |
| "eval_samples_per_second": 23.165, | |
| "eval_steps_per_second": 1.449, | |
| "step": 7008 | |
| }, | |
| { | |
| "epoch": 32.009132420091326, | |
| "grad_norm": 0.1892678290605545, | |
| "learning_rate": 4.2791911285061975e-06, | |
| "loss": 0.0277, | |
| "step": 7010 | |
| }, | |
| { | |
| "epoch": 32.054794520547944, | |
| "grad_norm": 0.22672104835510254, | |
| "learning_rate": 4.21395955642531e-06, | |
| "loss": 0.0303, | |
| "step": 7020 | |
| }, | |
| { | |
| "epoch": 32.10045662100457, | |
| "grad_norm": 0.12975597381591797, | |
| "learning_rate": 4.148727984344423e-06, | |
| "loss": 0.0281, | |
| "step": 7030 | |
| }, | |
| { | |
| "epoch": 32.146118721461185, | |
| "grad_norm": 0.17574161291122437, | |
| "learning_rate": 4.083496412263536e-06, | |
| "loss": 0.0301, | |
| "step": 7040 | |
| }, | |
| { | |
| "epoch": 32.19178082191781, | |
| "grad_norm": 0.2801176607608795, | |
| "learning_rate": 4.018264840182649e-06, | |
| "loss": 0.0278, | |
| "step": 7050 | |
| }, | |
| { | |
| "epoch": 32.23744292237443, | |
| "grad_norm": 0.17608405649662018, | |
| "learning_rate": 3.953033268101762e-06, | |
| "loss": 0.0287, | |
| "step": 7060 | |
| }, | |
| { | |
| "epoch": 32.28310502283105, | |
| "grad_norm": 0.2439558506011963, | |
| "learning_rate": 3.887801696020874e-06, | |
| "loss": 0.0288, | |
| "step": 7070 | |
| }, | |
| { | |
| "epoch": 32.32876712328767, | |
| "grad_norm": 0.1893412321805954, | |
| "learning_rate": 3.822570123939987e-06, | |
| "loss": 0.0297, | |
| "step": 7080 | |
| }, | |
| { | |
| "epoch": 32.37442922374429, | |
| "grad_norm": 0.1842028945684433, | |
| "learning_rate": 3.7573385518591e-06, | |
| "loss": 0.0286, | |
| "step": 7090 | |
| }, | |
| { | |
| "epoch": 32.42009132420091, | |
| "grad_norm": 0.14635466039180756, | |
| "learning_rate": 3.6921069797782126e-06, | |
| "loss": 0.0268, | |
| "step": 7100 | |
| }, | |
| { | |
| "epoch": 32.465753424657535, | |
| "grad_norm": 0.16718840599060059, | |
| "learning_rate": 3.6268754076973256e-06, | |
| "loss": 0.0294, | |
| "step": 7110 | |
| }, | |
| { | |
| "epoch": 32.51141552511415, | |
| "grad_norm": 0.19221429526805878, | |
| "learning_rate": 3.5616438356164386e-06, | |
| "loss": 0.0285, | |
| "step": 7120 | |
| }, | |
| { | |
| "epoch": 32.55707762557078, | |
| "grad_norm": 0.15354396402835846, | |
| "learning_rate": 3.496412263535551e-06, | |
| "loss": 0.0284, | |
| "step": 7130 | |
| }, | |
| { | |
| "epoch": 32.602739726027394, | |
| "grad_norm": 0.15389499068260193, | |
| "learning_rate": 3.4311806914546642e-06, | |
| "loss": 0.028, | |
| "step": 7140 | |
| }, | |
| { | |
| "epoch": 32.64840182648402, | |
| "grad_norm": 0.15369181334972382, | |
| "learning_rate": 3.365949119373777e-06, | |
| "loss": 0.0301, | |
| "step": 7150 | |
| }, | |
| { | |
| "epoch": 32.694063926940636, | |
| "grad_norm": 0.21850866079330444, | |
| "learning_rate": 3.30071754729289e-06, | |
| "loss": 0.0294, | |
| "step": 7160 | |
| }, | |
| { | |
| "epoch": 32.73972602739726, | |
| "grad_norm": 0.1314105987548828, | |
| "learning_rate": 3.235485975212003e-06, | |
| "loss": 0.0295, | |
| "step": 7170 | |
| }, | |
| { | |
| "epoch": 32.78538812785388, | |
| "grad_norm": 0.15347546339035034, | |
| "learning_rate": 3.1702544031311154e-06, | |
| "loss": 0.0284, | |
| "step": 7180 | |
| }, | |
| { | |
| "epoch": 32.8310502283105, | |
| "grad_norm": 0.191708505153656, | |
| "learning_rate": 3.1050228310502285e-06, | |
| "loss": 0.0297, | |
| "step": 7190 | |
| }, | |
| { | |
| "epoch": 32.87671232876713, | |
| "grad_norm": 0.21553701162338257, | |
| "learning_rate": 3.0397912589693415e-06, | |
| "loss": 0.0277, | |
| "step": 7200 | |
| }, | |
| { | |
| "epoch": 32.922374429223744, | |
| "grad_norm": 0.1383250206708908, | |
| "learning_rate": 2.974559686888454e-06, | |
| "loss": 0.0292, | |
| "step": 7210 | |
| }, | |
| { | |
| "epoch": 32.96803652968037, | |
| "grad_norm": 0.1352762132883072, | |
| "learning_rate": 2.909328114807567e-06, | |
| "loss": 0.0285, | |
| "step": 7220 | |
| }, | |
| { | |
| "epoch": 33.0, | |
| "eval_bertscore_f1": 0.8765576011485762, | |
| "eval_bleu": 0.6868938752836411, | |
| "eval_loss": 0.0425349660217762, | |
| "eval_rougeL": 0.3051064502374255, | |
| "eval_runtime": 65.0494, | |
| "eval_samples_per_second": 23.106, | |
| "eval_steps_per_second": 1.445, | |
| "step": 7227 | |
| }, | |
| { | |
| "epoch": 33.013698630136986, | |
| "grad_norm": 0.2675996422767639, | |
| "learning_rate": 2.8440965427266797e-06, | |
| "loss": 0.028, | |
| "step": 7230 | |
| }, | |
| { | |
| "epoch": 33.05936073059361, | |
| "grad_norm": 0.19094286859035492, | |
| "learning_rate": 2.7788649706457927e-06, | |
| "loss": 0.0282, | |
| "step": 7240 | |
| }, | |
| { | |
| "epoch": 33.10502283105023, | |
| "grad_norm": 0.18227194249629974, | |
| "learning_rate": 2.7136333985649057e-06, | |
| "loss": 0.0286, | |
| "step": 7250 | |
| }, | |
| { | |
| "epoch": 33.15068493150685, | |
| "grad_norm": 0.1970142275094986, | |
| "learning_rate": 2.6484018264840183e-06, | |
| "loss": 0.0261, | |
| "step": 7260 | |
| }, | |
| { | |
| "epoch": 33.19634703196347, | |
| "grad_norm": 0.14779959619045258, | |
| "learning_rate": 2.5831702544031313e-06, | |
| "loss": 0.0296, | |
| "step": 7270 | |
| }, | |
| { | |
| "epoch": 33.242009132420094, | |
| "grad_norm": 0.1774580031633377, | |
| "learning_rate": 2.517938682322244e-06, | |
| "loss": 0.031, | |
| "step": 7280 | |
| }, | |
| { | |
| "epoch": 33.28767123287671, | |
| "grad_norm": 0.20665834844112396, | |
| "learning_rate": 2.452707110241357e-06, | |
| "loss": 0.028, | |
| "step": 7290 | |
| }, | |
| { | |
| "epoch": 33.333333333333336, | |
| "grad_norm": 0.17283357679843903, | |
| "learning_rate": 2.38747553816047e-06, | |
| "loss": 0.0276, | |
| "step": 7300 | |
| }, | |
| { | |
| "epoch": 33.37899543378995, | |
| "grad_norm": 0.14525634050369263, | |
| "learning_rate": 2.3222439660795826e-06, | |
| "loss": 0.0281, | |
| "step": 7310 | |
| }, | |
| { | |
| "epoch": 33.42465753424658, | |
| "grad_norm": 0.16242118179798126, | |
| "learning_rate": 2.2570123939986956e-06, | |
| "loss": 0.0283, | |
| "step": 7320 | |
| }, | |
| { | |
| "epoch": 33.470319634703195, | |
| "grad_norm": 0.20483170449733734, | |
| "learning_rate": 2.191780821917808e-06, | |
| "loss": 0.029, | |
| "step": 7330 | |
| }, | |
| { | |
| "epoch": 33.51598173515982, | |
| "grad_norm": 0.18850266933441162, | |
| "learning_rate": 2.126549249836921e-06, | |
| "loss": 0.0312, | |
| "step": 7340 | |
| }, | |
| { | |
| "epoch": 33.56164383561644, | |
| "grad_norm": 0.24169522523880005, | |
| "learning_rate": 2.0613176777560342e-06, | |
| "loss": 0.0292, | |
| "step": 7350 | |
| }, | |
| { | |
| "epoch": 33.60730593607306, | |
| "grad_norm": 0.1359410136938095, | |
| "learning_rate": 1.996086105675147e-06, | |
| "loss": 0.0279, | |
| "step": 7360 | |
| }, | |
| { | |
| "epoch": 33.65296803652968, | |
| "grad_norm": 0.16504965722560883, | |
| "learning_rate": 1.93085453359426e-06, | |
| "loss": 0.0276, | |
| "step": 7370 | |
| }, | |
| { | |
| "epoch": 33.6986301369863, | |
| "grad_norm": 0.12197738885879517, | |
| "learning_rate": 1.8656229615133726e-06, | |
| "loss": 0.028, | |
| "step": 7380 | |
| }, | |
| { | |
| "epoch": 33.74429223744292, | |
| "grad_norm": 0.1623210310935974, | |
| "learning_rate": 1.8003913894324854e-06, | |
| "loss": 0.0295, | |
| "step": 7390 | |
| }, | |
| { | |
| "epoch": 33.789954337899545, | |
| "grad_norm": 0.20221665501594543, | |
| "learning_rate": 1.7351598173515982e-06, | |
| "loss": 0.0265, | |
| "step": 7400 | |
| }, | |
| { | |
| "epoch": 33.83561643835616, | |
| "grad_norm": 0.19585196673870087, | |
| "learning_rate": 1.669928245270711e-06, | |
| "loss": 0.0292, | |
| "step": 7410 | |
| }, | |
| { | |
| "epoch": 33.881278538812786, | |
| "grad_norm": 0.2633768320083618, | |
| "learning_rate": 1.604696673189824e-06, | |
| "loss": 0.0293, | |
| "step": 7420 | |
| }, | |
| { | |
| "epoch": 33.926940639269404, | |
| "grad_norm": 0.1686674803495407, | |
| "learning_rate": 1.5394651011089369e-06, | |
| "loss": 0.0298, | |
| "step": 7430 | |
| }, | |
| { | |
| "epoch": 33.97260273972603, | |
| "grad_norm": 0.14225494861602783, | |
| "learning_rate": 1.4742335290280497e-06, | |
| "loss": 0.0296, | |
| "step": 7440 | |
| }, | |
| { | |
| "epoch": 34.0, | |
| "eval_bertscore_f1": 0.8776290599933404, | |
| "eval_bleu": 0.7360000570941625, | |
| "eval_loss": 0.04238814115524292, | |
| "eval_rougeL": 0.3019882967022193, | |
| "eval_runtime": 65.4669, | |
| "eval_samples_per_second": 22.958, | |
| "eval_steps_per_second": 1.436, | |
| "step": 7446 | |
| }, | |
| { | |
| "epoch": 34.018264840182646, | |
| "grad_norm": 0.1806912124156952, | |
| "learning_rate": 1.4090019569471625e-06, | |
| "loss": 0.027, | |
| "step": 7450 | |
| }, | |
| { | |
| "epoch": 34.06392694063927, | |
| "grad_norm": 0.14808008074760437, | |
| "learning_rate": 1.3437703848662755e-06, | |
| "loss": 0.0292, | |
| "step": 7460 | |
| }, | |
| { | |
| "epoch": 34.10958904109589, | |
| "grad_norm": 0.1517217606306076, | |
| "learning_rate": 1.2785388127853883e-06, | |
| "loss": 0.0296, | |
| "step": 7470 | |
| }, | |
| { | |
| "epoch": 34.15525114155251, | |
| "grad_norm": 0.24545730650424957, | |
| "learning_rate": 1.2133072407045011e-06, | |
| "loss": 0.0277, | |
| "step": 7480 | |
| }, | |
| { | |
| "epoch": 34.20091324200913, | |
| "grad_norm": 0.1936875879764557, | |
| "learning_rate": 1.148075668623614e-06, | |
| "loss": 0.028, | |
| "step": 7490 | |
| }, | |
| { | |
| "epoch": 34.24657534246575, | |
| "grad_norm": 0.16787278652191162, | |
| "learning_rate": 1.0828440965427267e-06, | |
| "loss": 0.0287, | |
| "step": 7500 | |
| }, | |
| { | |
| "epoch": 34.29223744292237, | |
| "grad_norm": 0.15620262920856476, | |
| "learning_rate": 1.0176125244618398e-06, | |
| "loss": 0.0281, | |
| "step": 7510 | |
| }, | |
| { | |
| "epoch": 34.337899543378995, | |
| "grad_norm": 0.20259173214435577, | |
| "learning_rate": 9.523809523809526e-07, | |
| "loss": 0.0293, | |
| "step": 7520 | |
| }, | |
| { | |
| "epoch": 34.38356164383562, | |
| "grad_norm": 0.22473689913749695, | |
| "learning_rate": 8.871493803000652e-07, | |
| "loss": 0.0289, | |
| "step": 7530 | |
| }, | |
| { | |
| "epoch": 34.42922374429224, | |
| "grad_norm": 0.1960284262895584, | |
| "learning_rate": 8.219178082191781e-07, | |
| "loss": 0.0285, | |
| "step": 7540 | |
| }, | |
| { | |
| "epoch": 34.47488584474886, | |
| "grad_norm": 0.18783798813819885, | |
| "learning_rate": 7.56686236138291e-07, | |
| "loss": 0.0282, | |
| "step": 7550 | |
| }, | |
| { | |
| "epoch": 34.52054794520548, | |
| "grad_norm": 0.22345979511737823, | |
| "learning_rate": 6.914546640574038e-07, | |
| "loss": 0.0283, | |
| "step": 7560 | |
| }, | |
| { | |
| "epoch": 34.5662100456621, | |
| "grad_norm": 0.1570662409067154, | |
| "learning_rate": 6.262230919765167e-07, | |
| "loss": 0.028, | |
| "step": 7570 | |
| }, | |
| { | |
| "epoch": 34.61187214611872, | |
| "grad_norm": 0.1383802741765976, | |
| "learning_rate": 5.609915198956295e-07, | |
| "loss": 0.0296, | |
| "step": 7580 | |
| }, | |
| { | |
| "epoch": 34.657534246575345, | |
| "grad_norm": 0.1836177259683609, | |
| "learning_rate": 4.957599478147424e-07, | |
| "loss": 0.0273, | |
| "step": 7590 | |
| }, | |
| { | |
| "epoch": 34.70319634703196, | |
| "grad_norm": 0.3105820417404175, | |
| "learning_rate": 4.305283757338552e-07, | |
| "loss": 0.0285, | |
| "step": 7600 | |
| }, | |
| { | |
| "epoch": 34.74885844748859, | |
| "grad_norm": 0.23880144953727722, | |
| "learning_rate": 3.6529680365296803e-07, | |
| "loss": 0.0284, | |
| "step": 7610 | |
| }, | |
| { | |
| "epoch": 34.794520547945204, | |
| "grad_norm": 0.20711283385753632, | |
| "learning_rate": 3.000652315720809e-07, | |
| "loss": 0.0272, | |
| "step": 7620 | |
| }, | |
| { | |
| "epoch": 34.84018264840183, | |
| "grad_norm": 0.1828579604625702, | |
| "learning_rate": 2.3483365949119375e-07, | |
| "loss": 0.031, | |
| "step": 7630 | |
| }, | |
| { | |
| "epoch": 34.885844748858446, | |
| "grad_norm": 0.15802130103111267, | |
| "learning_rate": 1.696020874103066e-07, | |
| "loss": 0.0266, | |
| "step": 7640 | |
| }, | |
| { | |
| "epoch": 34.93150684931507, | |
| "grad_norm": 0.19433271884918213, | |
| "learning_rate": 1.0437051532941944e-07, | |
| "loss": 0.0281, | |
| "step": 7650 | |
| }, | |
| { | |
| "epoch": 34.97716894977169, | |
| "grad_norm": 0.15197888016700745, | |
| "learning_rate": 3.9138943248532294e-08, | |
| "loss": 0.0279, | |
| "step": 7660 | |
| }, | |
| { | |
| "epoch": 35.0, | |
| "eval_bertscore_f1": 0.8782449891308666, | |
| "eval_bleu": 0.7455200374263401, | |
| "eval_loss": 0.04245344549417496, | |
| "eval_rougeL": 0.30427156803355015, | |
| "eval_runtime": 65.2903, | |
| "eval_samples_per_second": 23.02, | |
| "eval_steps_per_second": 1.44, | |
| "step": 7665 | |
| } | |
| ], | |
| "logging_steps": 10, | |
| "max_steps": 7665, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 35, | |
| "save_steps": 500, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": true | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 7469843241369600.0, | |
| "train_batch_size": 16, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |