| { |
| "best_global_step": 6000, |
| "best_metric": 0.6719915270805359, |
| "best_model_checkpoint": "/workspace/rails-finetune/adapters-qwen3-coder-30b/checkpoint-6000", |
| "epoch": 0.9581794590278471, |
| "eval_steps": 500, |
| "global_step": 6000, |
| "is_hyper_param_search": false, |
| "is_local_process_zero": true, |
| "is_world_process_zero": true, |
| "log_history": [ |
| { |
| "epoch": 0.0015969657650464117, |
| "grad_norm": 0.15089154243469238, |
| "learning_rate": 6.000000000000001e-07, |
| "loss": 1.5709858894348145, |
| "step": 10 |
| }, |
| { |
| "epoch": 0.0031939315300928235, |
| "grad_norm": 0.1435190588235855, |
| "learning_rate": 1.2666666666666669e-06, |
| "loss": 1.6188209533691407, |
| "step": 20 |
| }, |
| { |
| "epoch": 0.004790897295139236, |
| "grad_norm": 0.14615978300571442, |
| "learning_rate": 1.9333333333333336e-06, |
| "loss": 1.6381675720214843, |
| "step": 30 |
| }, |
| { |
| "epoch": 0.006387863060185647, |
| "grad_norm": 0.1281592696905136, |
| "learning_rate": 2.6e-06, |
| "loss": 1.5563851356506349, |
| "step": 40 |
| }, |
| { |
| "epoch": 0.00798482882523206, |
| "grad_norm": 0.16032098233699799, |
| "learning_rate": 3.266666666666667e-06, |
| "loss": 1.6382104873657226, |
| "step": 50 |
| }, |
| { |
| "epoch": 0.009581794590278471, |
| "grad_norm": 0.14609354734420776, |
| "learning_rate": 3.9333333333333335e-06, |
| "loss": 1.6256795883178712, |
| "step": 60 |
| }, |
| { |
| "epoch": 0.011178760355324883, |
| "grad_norm": 0.13070641458034515, |
| "learning_rate": 4.600000000000001e-06, |
| "loss": 1.5222463607788086, |
| "step": 70 |
| }, |
| { |
| "epoch": 0.012775726120371294, |
| "grad_norm": 0.1499311476945877, |
| "learning_rate": 5.2666666666666665e-06, |
| "loss": 1.518197727203369, |
| "step": 80 |
| }, |
| { |
| "epoch": 0.014372691885417706, |
| "grad_norm": 0.12919697165489197, |
| "learning_rate": 5.933333333333335e-06, |
| "loss": 1.4710905075073242, |
| "step": 90 |
| }, |
| { |
| "epoch": 0.01596965765046412, |
| "grad_norm": 0.11437301337718964, |
| "learning_rate": 6.600000000000001e-06, |
| "loss": 1.4049152374267577, |
| "step": 100 |
| }, |
| { |
| "epoch": 0.01756662341551053, |
| "grad_norm": 0.13434389233589172, |
| "learning_rate": 7.266666666666668e-06, |
| "loss": 1.3853497505187988, |
| "step": 110 |
| }, |
| { |
| "epoch": 0.019163589180556943, |
| "grad_norm": 0.08636850863695145, |
| "learning_rate": 7.933333333333334e-06, |
| "loss": 1.3351438522338868, |
| "step": 120 |
| }, |
| { |
| "epoch": 0.020760554945603353, |
| "grad_norm": 0.08471965044736862, |
| "learning_rate": 8.6e-06, |
| "loss": 1.2937309265136718, |
| "step": 130 |
| }, |
| { |
| "epoch": 0.022357520710649767, |
| "grad_norm": 0.058983951807022095, |
| "learning_rate": 9.266666666666667e-06, |
| "loss": 1.2116679191589355, |
| "step": 140 |
| }, |
| { |
| "epoch": 0.023954486475696177, |
| "grad_norm": 0.06733114272356033, |
| "learning_rate": 9.933333333333334e-06, |
| "loss": 1.1680314064025878, |
| "step": 150 |
| }, |
| { |
| "epoch": 0.025551452240742588, |
| "grad_norm": 0.045156873762607574, |
| "learning_rate": 1.0600000000000002e-05, |
| "loss": 1.1280742645263673, |
| "step": 160 |
| }, |
| { |
| "epoch": 0.027148418005789002, |
| "grad_norm": 0.0415460579097271, |
| "learning_rate": 1.1266666666666668e-05, |
| "loss": 1.0879782676696776, |
| "step": 170 |
| }, |
| { |
| "epoch": 0.028745383770835412, |
| "grad_norm": 0.05221620574593544, |
| "learning_rate": 1.1933333333333335e-05, |
| "loss": 1.0641801834106446, |
| "step": 180 |
| }, |
| { |
| "epoch": 0.030342349535881826, |
| "grad_norm": 0.036492571234703064, |
| "learning_rate": 1.2600000000000001e-05, |
| "loss": 1.0208759307861328, |
| "step": 190 |
| }, |
| { |
| "epoch": 0.03193931530092824, |
| "grad_norm": 0.03153146430850029, |
| "learning_rate": 1.3266666666666668e-05, |
| "loss": 1.0204400062561034, |
| "step": 200 |
| }, |
| { |
| "epoch": 0.03353628106597465, |
| "grad_norm": 0.03343074768781662, |
| "learning_rate": 1.3933333333333334e-05, |
| "loss": 0.9532640457153321, |
| "step": 210 |
| }, |
| { |
| "epoch": 0.03513324683102106, |
| "grad_norm": 0.034289468079805374, |
| "learning_rate": 1.46e-05, |
| "loss": 0.9065071105957031, |
| "step": 220 |
| }, |
| { |
| "epoch": 0.03673021259606747, |
| "grad_norm": 0.040245093405246735, |
| "learning_rate": 1.5266666666666667e-05, |
| "loss": 0.895290470123291, |
| "step": 230 |
| }, |
| { |
| "epoch": 0.038327178361113885, |
| "grad_norm": 0.03090902790427208, |
| "learning_rate": 1.5933333333333336e-05, |
| "loss": 0.89828462600708, |
| "step": 240 |
| }, |
| { |
| "epoch": 0.03992414412616029, |
| "grad_norm": 0.023231355473399162, |
| "learning_rate": 1.66e-05, |
| "loss": 0.867582893371582, |
| "step": 250 |
| }, |
| { |
| "epoch": 0.041521109891206706, |
| "grad_norm": 0.022748615592718124, |
| "learning_rate": 1.726666666666667e-05, |
| "loss": 0.8476214408874512, |
| "step": 260 |
| }, |
| { |
| "epoch": 0.04311807565625312, |
| "grad_norm": 0.025588329881429672, |
| "learning_rate": 1.7933333333333333e-05, |
| "loss": 0.8639037132263183, |
| "step": 270 |
| }, |
| { |
| "epoch": 0.044715041421299534, |
| "grad_norm": 0.029987547546625137, |
| "learning_rate": 1.86e-05, |
| "loss": 0.8353777885437011, |
| "step": 280 |
| }, |
| { |
| "epoch": 0.04631200718634594, |
| "grad_norm": 0.020994428545236588, |
| "learning_rate": 1.926666666666667e-05, |
| "loss": 0.8189143180847168, |
| "step": 290 |
| }, |
| { |
| "epoch": 0.047908972951392355, |
| "grad_norm": 0.023545585572719574, |
| "learning_rate": 1.9933333333333334e-05, |
| "loss": 0.7846664905548095, |
| "step": 300 |
| }, |
| { |
| "epoch": 0.04950593871643877, |
| "grad_norm": 0.025134805589914322, |
| "learning_rate": 1.999988754726792e-05, |
| "loss": 0.7751604557037354, |
| "step": 310 |
| }, |
| { |
| "epoch": 0.051102904481485176, |
| "grad_norm": 0.021275486797094345, |
| "learning_rate": 1.9999498825021314e-05, |
| "loss": 0.7967638969421387, |
| "step": 320 |
| }, |
| { |
| "epoch": 0.05269987024653159, |
| "grad_norm": 0.019746815785765648, |
| "learning_rate": 1.9998832455745586e-05, |
| "loss": 0.8192337036132813, |
| "step": 330 |
| }, |
| { |
| "epoch": 0.054296836011578004, |
| "grad_norm": 0.02567191608250141, |
| "learning_rate": 1.9997888457943197e-05, |
| "loss": 0.7604565620422363, |
| "step": 340 |
| }, |
| { |
| "epoch": 0.05589380177662441, |
| "grad_norm": 0.024295911192893982, |
| "learning_rate": 1.9996666857825287e-05, |
| "loss": 0.7550592422485352, |
| "step": 350 |
| }, |
| { |
| "epoch": 0.057490767541670824, |
| "grad_norm": 0.02220323495566845, |
| "learning_rate": 1.9995167689310917e-05, |
| "loss": 0.7514551639556885, |
| "step": 360 |
| }, |
| { |
| "epoch": 0.05908773330671724, |
| "grad_norm": 0.024802017956972122, |
| "learning_rate": 1.9993390994026144e-05, |
| "loss": 0.7904987812042237, |
| "step": 370 |
| }, |
| { |
| "epoch": 0.06068469907176365, |
| "grad_norm": 0.027685435488820076, |
| "learning_rate": 1.9991336821302856e-05, |
| "loss": 0.7770297050476074, |
| "step": 380 |
| }, |
| { |
| "epoch": 0.06228166483681006, |
| "grad_norm": 0.019939841702580452, |
| "learning_rate": 1.9989005228177406e-05, |
| "loss": 0.7634893417358398, |
| "step": 390 |
| }, |
| { |
| "epoch": 0.06387863060185647, |
| "grad_norm": 0.028217511251568794, |
| "learning_rate": 1.9986396279389028e-05, |
| "loss": 0.8059945106506348, |
| "step": 400 |
| }, |
| { |
| "epoch": 0.06547559636690288, |
| "grad_norm": 0.019439322873950005, |
| "learning_rate": 1.9983510047378046e-05, |
| "loss": 0.7787755489349365, |
| "step": 410 |
| }, |
| { |
| "epoch": 0.0670725621319493, |
| "grad_norm": 0.02309611812233925, |
| "learning_rate": 1.9980346612283842e-05, |
| "loss": 0.7745323657989502, |
| "step": 420 |
| }, |
| { |
| "epoch": 0.06866952789699571, |
| "grad_norm": 0.020378535613417625, |
| "learning_rate": 1.997690606194266e-05, |
| "loss": 0.7243520736694335, |
| "step": 430 |
| }, |
| { |
| "epoch": 0.07026649366204211, |
| "grad_norm": 0.021776653826236725, |
| "learning_rate": 1.9973188491885146e-05, |
| "loss": 0.7424108028411865, |
| "step": 440 |
| }, |
| { |
| "epoch": 0.07186345942708854, |
| "grad_norm": 0.03435182571411133, |
| "learning_rate": 1.9969194005333697e-05, |
| "loss": 0.7804229736328125, |
| "step": 450 |
| }, |
| { |
| "epoch": 0.07346042519213494, |
| "grad_norm": 0.026880595833063126, |
| "learning_rate": 1.9964922713199613e-05, |
| "loss": 0.7963083267211915, |
| "step": 460 |
| }, |
| { |
| "epoch": 0.07505739095718135, |
| "grad_norm": 0.02643360197544098, |
| "learning_rate": 1.9960374734079984e-05, |
| "loss": 0.7517959117889405, |
| "step": 470 |
| }, |
| { |
| "epoch": 0.07665435672222777, |
| "grad_norm": 0.036808405071496964, |
| "learning_rate": 1.9955550194254436e-05, |
| "loss": 0.7511186122894287, |
| "step": 480 |
| }, |
| { |
| "epoch": 0.07825132248727418, |
| "grad_norm": 0.03115280158817768, |
| "learning_rate": 1.995044922768159e-05, |
| "loss": 0.765751314163208, |
| "step": 490 |
| }, |
| { |
| "epoch": 0.07984828825232058, |
| "grad_norm": 0.025536708533763885, |
| "learning_rate": 1.994507197599537e-05, |
| "loss": 0.7346895217895508, |
| "step": 500 |
| }, |
| { |
| "epoch": 0.07984828825232058, |
| "eval_loss": 0.7535409331321716, |
| "eval_runtime": 1705.9703, |
| "eval_samples_per_second": 3.263, |
| "eval_steps_per_second": 1.631, |
| "step": 500 |
| }, |
| { |
| "epoch": 0.081445254017367, |
| "grad_norm": 0.03875249624252319, |
| "learning_rate": 1.993941858850106e-05, |
| "loss": 0.7961405277252197, |
| "step": 510 |
| }, |
| { |
| "epoch": 0.08304221978241341, |
| "grad_norm": 0.027557438239455223, |
| "learning_rate": 1.993348922217114e-05, |
| "loss": 0.7667338371276855, |
| "step": 520 |
| }, |
| { |
| "epoch": 0.08463918554745983, |
| "grad_norm": 0.026045992970466614, |
| "learning_rate": 1.9927284041640967e-05, |
| "loss": 0.7275139808654785, |
| "step": 530 |
| }, |
| { |
| "epoch": 0.08623615131250624, |
| "grad_norm": 0.026213862001895905, |
| "learning_rate": 1.992080321920416e-05, |
| "loss": 0.7866159439086914, |
| "step": 540 |
| }, |
| { |
| "epoch": 0.08783311707755265, |
| "grad_norm": 0.028538137674331665, |
| "learning_rate": 1.9914046934807853e-05, |
| "loss": 0.7298189640045166, |
| "step": 550 |
| }, |
| { |
| "epoch": 0.08943008284259907, |
| "grad_norm": 0.028126219287514687, |
| "learning_rate": 1.9907015376047675e-05, |
| "loss": 0.749655294418335, |
| "step": 560 |
| }, |
| { |
| "epoch": 0.09102704860764547, |
| "grad_norm": 0.027085473760962486, |
| "learning_rate": 1.9899708738162553e-05, |
| "loss": 0.8086295127868652, |
| "step": 570 |
| }, |
| { |
| "epoch": 0.09262401437269188, |
| "grad_norm": 0.034640420228242874, |
| "learning_rate": 1.989212722402928e-05, |
| "loss": 0.7653923511505127, |
| "step": 580 |
| }, |
| { |
| "epoch": 0.0942209801377383, |
| "grad_norm": 0.03858828917145729, |
| "learning_rate": 1.98842710441569e-05, |
| "loss": 0.8174265861511231, |
| "step": 590 |
| }, |
| { |
| "epoch": 0.09581794590278471, |
| "grad_norm": 0.04248794540762901, |
| "learning_rate": 1.987614041668084e-05, |
| "loss": 0.7357111930847168, |
| "step": 600 |
| }, |
| { |
| "epoch": 0.09741491166783112, |
| "grad_norm": 0.03530848026275635, |
| "learning_rate": 1.9867735567356876e-05, |
| "loss": 0.7747371196746826, |
| "step": 610 |
| }, |
| { |
| "epoch": 0.09901187743287754, |
| "grad_norm": 0.027734428644180298, |
| "learning_rate": 1.9859056729554845e-05, |
| "loss": 0.73280348777771, |
| "step": 620 |
| }, |
| { |
| "epoch": 0.10060884319792394, |
| "grad_norm": 0.02657356671988964, |
| "learning_rate": 1.9850104144252177e-05, |
| "loss": 0.749216365814209, |
| "step": 630 |
| }, |
| { |
| "epoch": 0.10220580896297035, |
| "grad_norm": 0.021906374022364616, |
| "learning_rate": 1.98408780600272e-05, |
| "loss": 0.7534349441528321, |
| "step": 640 |
| }, |
| { |
| "epoch": 0.10380277472801677, |
| "grad_norm": 0.025684406980872154, |
| "learning_rate": 1.9831378733052244e-05, |
| "loss": 0.7199561595916748, |
| "step": 650 |
| }, |
| { |
| "epoch": 0.10539974049306318, |
| "grad_norm": 0.025311095640063286, |
| "learning_rate": 1.982160642708652e-05, |
| "loss": 0.7382417678833008, |
| "step": 660 |
| }, |
| { |
| "epoch": 0.10699670625810959, |
| "grad_norm": 0.035563357174396515, |
| "learning_rate": 1.9811561413468794e-05, |
| "loss": 0.7394683837890625, |
| "step": 670 |
| }, |
| { |
| "epoch": 0.10859367202315601, |
| "grad_norm": 0.023977380245923996, |
| "learning_rate": 1.9801243971109868e-05, |
| "loss": 0.7267738819122315, |
| "step": 680 |
| }, |
| { |
| "epoch": 0.11019063778820241, |
| "grad_norm": 0.02275015600025654, |
| "learning_rate": 1.9790654386484818e-05, |
| "loss": 0.7240358829498291, |
| "step": 690 |
| }, |
| { |
| "epoch": 0.11178760355324882, |
| "grad_norm": 0.036198344081640244, |
| "learning_rate": 1.9779792953625052e-05, |
| "loss": 0.7299670696258544, |
| "step": 700 |
| }, |
| { |
| "epoch": 0.11338456931829524, |
| "grad_norm": 0.027184955775737762, |
| "learning_rate": 1.976865997411014e-05, |
| "loss": 0.695775318145752, |
| "step": 710 |
| }, |
| { |
| "epoch": 0.11498153508334165, |
| "grad_norm": 0.023749997839331627, |
| "learning_rate": 1.9757255757059446e-05, |
| "loss": 0.7071991920471191, |
| "step": 720 |
| }, |
| { |
| "epoch": 0.11657850084838806, |
| "grad_norm": 0.027117466554045677, |
| "learning_rate": 1.9745580619123535e-05, |
| "loss": 0.7466438293457032, |
| "step": 730 |
| }, |
| { |
| "epoch": 0.11817546661343448, |
| "grad_norm": 0.027798349037766457, |
| "learning_rate": 1.9733634884475395e-05, |
| "loss": 0.7503840923309326, |
| "step": 740 |
| }, |
| { |
| "epoch": 0.11977243237848088, |
| "grad_norm": 0.026327304542064667, |
| "learning_rate": 1.9721418884801414e-05, |
| "loss": 0.724392032623291, |
| "step": 750 |
| }, |
| { |
| "epoch": 0.1213693981435273, |
| "grad_norm": 0.029995381832122803, |
| "learning_rate": 1.97089329592922e-05, |
| "loss": 0.7100958824157715, |
| "step": 760 |
| }, |
| { |
| "epoch": 0.12296636390857371, |
| "grad_norm": 0.03063913807272911, |
| "learning_rate": 1.969617745463314e-05, |
| "loss": 0.7130911827087403, |
| "step": 770 |
| }, |
| { |
| "epoch": 0.12456332967362012, |
| "grad_norm": 0.021976860240101814, |
| "learning_rate": 1.968315272499478e-05, |
| "loss": 0.7231676578521729, |
| "step": 780 |
| }, |
| { |
| "epoch": 0.12616029543866653, |
| "grad_norm": 0.02538118325173855, |
| "learning_rate": 1.9669859132022994e-05, |
| "loss": 0.7253612995147705, |
| "step": 790 |
| }, |
| { |
| "epoch": 0.12775726120371295, |
| "grad_norm": 0.02788228541612625, |
| "learning_rate": 1.9656297044828943e-05, |
| "loss": 0.7297886848449707, |
| "step": 800 |
| }, |
| { |
| "epoch": 0.12935422696875937, |
| "grad_norm": 0.028525004163384438, |
| "learning_rate": 1.9642466839978814e-05, |
| "loss": 0.7109212875366211, |
| "step": 810 |
| }, |
| { |
| "epoch": 0.13095119273380576, |
| "grad_norm": 0.028368208557367325, |
| "learning_rate": 1.962836890148339e-05, |
| "loss": 0.7552286624908447, |
| "step": 820 |
| }, |
| { |
| "epoch": 0.13254815849885218, |
| "grad_norm": 0.03732667118310928, |
| "learning_rate": 1.9614003620787358e-05, |
| "loss": 0.74849853515625, |
| "step": 830 |
| }, |
| { |
| "epoch": 0.1341451242638986, |
| "grad_norm": 0.026267215609550476, |
| "learning_rate": 1.9599371396758457e-05, |
| "loss": 0.7323933124542237, |
| "step": 840 |
| }, |
| { |
| "epoch": 0.135742090028945, |
| "grad_norm": 0.031789544969797134, |
| "learning_rate": 1.958447263567641e-05, |
| "loss": 0.7394798755645752, |
| "step": 850 |
| }, |
| { |
| "epoch": 0.13733905579399142, |
| "grad_norm": 0.03209487721323967, |
| "learning_rate": 1.956930775122162e-05, |
| "loss": 0.7479897499084472, |
| "step": 860 |
| }, |
| { |
| "epoch": 0.13893602155903784, |
| "grad_norm": 0.026272661983966827, |
| "learning_rate": 1.9553877164463698e-05, |
| "loss": 0.7194801807403565, |
| "step": 870 |
| }, |
| { |
| "epoch": 0.14053298732408423, |
| "grad_norm": 0.02558542974293232, |
| "learning_rate": 1.953818130384978e-05, |
| "loss": 0.6967973709106445, |
| "step": 880 |
| }, |
| { |
| "epoch": 0.14212995308913065, |
| "grad_norm": 0.03761237859725952, |
| "learning_rate": 1.9522220605192615e-05, |
| "loss": 0.6678271770477295, |
| "step": 890 |
| }, |
| { |
| "epoch": 0.14372691885417707, |
| "grad_norm": 0.029605882242321968, |
| "learning_rate": 1.9505995511658464e-05, |
| "loss": 0.6922338008880615, |
| "step": 900 |
| }, |
| { |
| "epoch": 0.14532388461922346, |
| "grad_norm": 0.03535737469792366, |
| "learning_rate": 1.948950647375481e-05, |
| "loss": 0.7159334182739258, |
| "step": 910 |
| }, |
| { |
| "epoch": 0.14692085038426989, |
| "grad_norm": 0.030855044722557068, |
| "learning_rate": 1.9472753949317843e-05, |
| "loss": 0.7333884716033936, |
| "step": 920 |
| }, |
| { |
| "epoch": 0.1485178161493163, |
| "grad_norm": 0.05045896768569946, |
| "learning_rate": 1.9455738403499728e-05, |
| "loss": 0.6935329914093018, |
| "step": 930 |
| }, |
| { |
| "epoch": 0.1501147819143627, |
| "grad_norm": 0.033187806606292725, |
| "learning_rate": 1.9438460308755724e-05, |
| "loss": 0.7046589851379395, |
| "step": 940 |
| }, |
| { |
| "epoch": 0.15171174767940912, |
| "grad_norm": 0.03167016804218292, |
| "learning_rate": 1.9420920144831044e-05, |
| "loss": 0.710863447189331, |
| "step": 950 |
| }, |
| { |
| "epoch": 0.15330871344445554, |
| "grad_norm": 0.025539802387356758, |
| "learning_rate": 1.9403118398747533e-05, |
| "loss": 0.6936647891998291, |
| "step": 960 |
| }, |
| { |
| "epoch": 0.15490567920950193, |
| "grad_norm": 0.04223870858550072, |
| "learning_rate": 1.9385055564790157e-05, |
| "loss": 0.6985628128051757, |
| "step": 970 |
| }, |
| { |
| "epoch": 0.15650264497454835, |
| "grad_norm": 0.030605314299464226, |
| "learning_rate": 1.9366732144493266e-05, |
| "loss": 0.7294198989868164, |
| "step": 980 |
| }, |
| { |
| "epoch": 0.15809961073959478, |
| "grad_norm": 0.03095085918903351, |
| "learning_rate": 1.9348148646626687e-05, |
| "loss": 0.7009531021118164, |
| "step": 990 |
| }, |
| { |
| "epoch": 0.15969657650464117, |
| "grad_norm": 0.02993757091462612, |
| "learning_rate": 1.9329305587181574e-05, |
| "loss": 0.7266313552856445, |
| "step": 1000 |
| }, |
| { |
| "epoch": 0.15969657650464117, |
| "eval_loss": 0.7159722447395325, |
| "eval_runtime": 1703.1621, |
| "eval_samples_per_second": 3.268, |
| "eval_steps_per_second": 1.634, |
| "step": 1000 |
| }, |
| { |
| "epoch": 0.1612935422696876, |
| "grad_norm": 0.035543542355298996, |
| "learning_rate": 1.9310203489356092e-05, |
| "loss": 0.74082350730896, |
| "step": 1010 |
| }, |
| { |
| "epoch": 0.162890508034734, |
| "grad_norm": 0.03636915981769562, |
| "learning_rate": 1.9290842883540897e-05, |
| "loss": 0.717669153213501, |
| "step": 1020 |
| }, |
| { |
| "epoch": 0.1644874737997804, |
| "grad_norm": 0.04510757327079773, |
| "learning_rate": 1.92712243073044e-05, |
| "loss": 0.7365827560424805, |
| "step": 1030 |
| }, |
| { |
| "epoch": 0.16608443956482682, |
| "grad_norm": 0.03563699871301651, |
| "learning_rate": 1.925134830537784e-05, |
| "loss": 0.6981801986694336, |
| "step": 1040 |
| }, |
| { |
| "epoch": 0.16768140532987325, |
| "grad_norm": 0.04541337490081787, |
| "learning_rate": 1.9231215429640167e-05, |
| "loss": 0.754232931137085, |
| "step": 1050 |
| }, |
| { |
| "epoch": 0.16927837109491967, |
| "grad_norm": 0.033323634415864944, |
| "learning_rate": 1.921082623910271e-05, |
| "loss": 0.691849946975708, |
| "step": 1060 |
| }, |
| { |
| "epoch": 0.17087533685996606, |
| "grad_norm": 0.03559419885277748, |
| "learning_rate": 1.919018129989366e-05, |
| "loss": 0.6994197845458985, |
| "step": 1070 |
| }, |
| { |
| "epoch": 0.17247230262501248, |
| "grad_norm": 0.03182852268218994, |
| "learning_rate": 1.916928118524235e-05, |
| "loss": 0.6645867824554443, |
| "step": 1080 |
| }, |
| { |
| "epoch": 0.1740692683900589, |
| "grad_norm": 0.02616371586918831, |
| "learning_rate": 1.9148126475463336e-05, |
| "loss": 0.7270137786865234, |
| "step": 1090 |
| }, |
| { |
| "epoch": 0.1756662341551053, |
| "grad_norm": 0.03250862658023834, |
| "learning_rate": 1.9126717757940288e-05, |
| "loss": 0.7272531509399414, |
| "step": 1100 |
| }, |
| { |
| "epoch": 0.17726319992015172, |
| "grad_norm": 0.036741774529218674, |
| "learning_rate": 1.9105055627109683e-05, |
| "loss": 0.7251851558685303, |
| "step": 1110 |
| }, |
| { |
| "epoch": 0.17886016568519814, |
| "grad_norm": 0.03240974619984627, |
| "learning_rate": 1.908314068444429e-05, |
| "loss": 0.7145021915435791, |
| "step": 1120 |
| }, |
| { |
| "epoch": 0.18045713145024453, |
| "grad_norm": 0.0324835442006588, |
| "learning_rate": 1.9060973538436478e-05, |
| "loss": 0.7045553684234619, |
| "step": 1130 |
| }, |
| { |
| "epoch": 0.18205409721529095, |
| "grad_norm": 0.029804598540067673, |
| "learning_rate": 1.9038554804581318e-05, |
| "loss": 0.7342820644378663, |
| "step": 1140 |
| }, |
| { |
| "epoch": 0.18365106298033737, |
| "grad_norm": 0.03631270304322243, |
| "learning_rate": 1.9015885105359492e-05, |
| "loss": 0.7435333728790283, |
| "step": 1150 |
| }, |
| { |
| "epoch": 0.18524802874538376, |
| "grad_norm": 0.031312599778175354, |
| "learning_rate": 1.8992965070220007e-05, |
| "loss": 0.7185348033905029, |
| "step": 1160 |
| }, |
| { |
| "epoch": 0.18684499451043018, |
| "grad_norm": 0.03553950414061546, |
| "learning_rate": 1.896979533556273e-05, |
| "loss": 0.7191666603088379, |
| "step": 1170 |
| }, |
| { |
| "epoch": 0.1884419602754766, |
| "grad_norm": 0.024733861908316612, |
| "learning_rate": 1.8946376544720698e-05, |
| "loss": 0.7352997779846191, |
| "step": 1180 |
| }, |
| { |
| "epoch": 0.190038926040523, |
| "grad_norm": 0.030409259721636772, |
| "learning_rate": 1.8922709347942275e-05, |
| "loss": 0.7265621185302734, |
| "step": 1190 |
| }, |
| { |
| "epoch": 0.19163589180556942, |
| "grad_norm": 0.030736226588487625, |
| "learning_rate": 1.8898794402373077e-05, |
| "loss": 0.7150910377502442, |
| "step": 1200 |
| }, |
| { |
| "epoch": 0.19323285757061584, |
| "grad_norm": 0.038398947566747665, |
| "learning_rate": 1.887463237203775e-05, |
| "loss": 0.7055376052856446, |
| "step": 1210 |
| }, |
| { |
| "epoch": 0.19482982333566223, |
| "grad_norm": 0.03485625982284546, |
| "learning_rate": 1.88502239278215e-05, |
| "loss": 0.6555115699768066, |
| "step": 1220 |
| }, |
| { |
| "epoch": 0.19642678910070865, |
| "grad_norm": 0.03220400586724281, |
| "learning_rate": 1.8825569747451505e-05, |
| "loss": 0.6946470737457275, |
| "step": 1230 |
| }, |
| { |
| "epoch": 0.19802375486575508, |
| "grad_norm": 0.03333678096532822, |
| "learning_rate": 1.880067051547806e-05, |
| "loss": 0.7297664642333984, |
| "step": 1240 |
| }, |
| { |
| "epoch": 0.19962072063080147, |
| "grad_norm": 0.0367787703871727, |
| "learning_rate": 1.8775526923255597e-05, |
| "loss": 0.7051557064056396, |
| "step": 1250 |
| }, |
| { |
| "epoch": 0.2012176863958479, |
| "grad_norm": 0.03103001043200493, |
| "learning_rate": 1.8750139668923472e-05, |
| "loss": 0.7219597339630127, |
| "step": 1260 |
| }, |
| { |
| "epoch": 0.2028146521608943, |
| "grad_norm": 0.04533790051937103, |
| "learning_rate": 1.872450945738659e-05, |
| "loss": 0.6916751384735107, |
| "step": 1270 |
| }, |
| { |
| "epoch": 0.2044116179259407, |
| "grad_norm": 0.03510045260190964, |
| "learning_rate": 1.8698637000295816e-05, |
| "loss": 0.7119457721710205, |
| "step": 1280 |
| }, |
| { |
| "epoch": 0.20600858369098712, |
| "grad_norm": 0.029134295880794525, |
| "learning_rate": 1.867252301602825e-05, |
| "loss": 0.6870355129241943, |
| "step": 1290 |
| }, |
| { |
| "epoch": 0.20760554945603354, |
| "grad_norm": 0.03194071725010872, |
| "learning_rate": 1.8646168229667238e-05, |
| "loss": 0.734464693069458, |
| "step": 1300 |
| }, |
| { |
| "epoch": 0.20920251522107994, |
| "grad_norm": 0.026471436023712158, |
| "learning_rate": 1.861957337298227e-05, |
| "loss": 0.7240866184234619, |
| "step": 1310 |
| }, |
| { |
| "epoch": 0.21079948098612636, |
| "grad_norm": 0.0347883440554142, |
| "learning_rate": 1.8592739184408657e-05, |
| "loss": 0.6553101062774658, |
| "step": 1320 |
| }, |
| { |
| "epoch": 0.21239644675117278, |
| "grad_norm": 0.026658106595277786, |
| "learning_rate": 1.8565666409027004e-05, |
| "loss": 0.7519384384155273, |
| "step": 1330 |
| }, |
| { |
| "epoch": 0.21399341251621917, |
| "grad_norm": 0.030675504356622696, |
| "learning_rate": 1.8538355798542556e-05, |
| "loss": 0.6963082790374756, |
| "step": 1340 |
| }, |
| { |
| "epoch": 0.2155903782812656, |
| "grad_norm": 0.04593832045793533, |
| "learning_rate": 1.85108081112643e-05, |
| "loss": 0.7145741939544678, |
| "step": 1350 |
| }, |
| { |
| "epoch": 0.21718734404631201, |
| "grad_norm": 0.04163511469960213, |
| "learning_rate": 1.8483024112083928e-05, |
| "loss": 0.7342512130737304, |
| "step": 1360 |
| }, |
| { |
| "epoch": 0.2187843098113584, |
| "grad_norm": 0.025683345273137093, |
| "learning_rate": 1.8455004572454583e-05, |
| "loss": 0.7134137630462647, |
| "step": 1370 |
| }, |
| { |
| "epoch": 0.22038127557640483, |
| "grad_norm": 0.026864832267165184, |
| "learning_rate": 1.8426750270369452e-05, |
| "loss": 0.6854844570159913, |
| "step": 1380 |
| }, |
| { |
| "epoch": 0.22197824134145125, |
| "grad_norm": 0.03378361091017723, |
| "learning_rate": 1.839826199034015e-05, |
| "loss": 0.6904460430145264, |
| "step": 1390 |
| }, |
| { |
| "epoch": 0.22357520710649764, |
| "grad_norm": 0.03360961750149727, |
| "learning_rate": 1.8369540523374963e-05, |
| "loss": 0.7089653491973877, |
| "step": 1400 |
| }, |
| { |
| "epoch": 0.22517217287154406, |
| "grad_norm": 0.033883776515722275, |
| "learning_rate": 1.8340586666956846e-05, |
| "loss": 0.7053616523742676, |
| "step": 1410 |
| }, |
| { |
| "epoch": 0.22676913863659048, |
| "grad_norm": 0.03643488511443138, |
| "learning_rate": 1.8311401225021318e-05, |
| "loss": 0.7411230087280274, |
| "step": 1420 |
| }, |
| { |
| "epoch": 0.22836610440163688, |
| "grad_norm": 0.030240802094340324, |
| "learning_rate": 1.8281985007934115e-05, |
| "loss": 0.7020374298095703, |
| "step": 1430 |
| }, |
| { |
| "epoch": 0.2299630701666833, |
| "grad_norm": 0.03315526619553566, |
| "learning_rate": 1.8252338832468702e-05, |
| "loss": 0.7227590084075928, |
| "step": 1440 |
| }, |
| { |
| "epoch": 0.23156003593172972, |
| "grad_norm": 0.04955840855836868, |
| "learning_rate": 1.8222463521783584e-05, |
| "loss": 0.7004672527313233, |
| "step": 1450 |
| }, |
| { |
| "epoch": 0.2331570016967761, |
| "grad_norm": 0.03338323533535004, |
| "learning_rate": 1.819235990539946e-05, |
| "loss": 0.7032230377197266, |
| "step": 1460 |
| }, |
| { |
| "epoch": 0.23475396746182253, |
| "grad_norm": 0.03659350797533989, |
| "learning_rate": 1.8162028819176192e-05, |
| "loss": 0.7022134780883789, |
| "step": 1470 |
| }, |
| { |
| "epoch": 0.23635093322686895, |
| "grad_norm": 0.029659852385520935, |
| "learning_rate": 1.813147110528958e-05, |
| "loss": 0.7326688289642334, |
| "step": 1480 |
| }, |
| { |
| "epoch": 0.23794789899191537, |
| "grad_norm": 0.03314507007598877, |
| "learning_rate": 1.8100687612208e-05, |
| "loss": 0.7101527690887451, |
| "step": 1490 |
| }, |
| { |
| "epoch": 0.23954486475696177, |
| "grad_norm": 0.048770975321531296, |
| "learning_rate": 1.806967919466883e-05, |
| "loss": 0.7013855457305909, |
| "step": 1500 |
| }, |
| { |
| "epoch": 0.23954486475696177, |
| "eval_loss": 0.7003746032714844, |
| "eval_runtime": 1694.4125, |
| "eval_samples_per_second": 3.285, |
| "eval_steps_per_second": 1.642, |
| "step": 1500 |
| }, |
| { |
| "epoch": 0.2411418305220082, |
| "grad_norm": 0.031342763453722, |
| "learning_rate": 1.803844671365471e-05, |
| "loss": 0.6915247917175293, |
| "step": 1510 |
| }, |
| { |
| "epoch": 0.2427387962870546, |
| "grad_norm": 0.029329324141144753, |
| "learning_rate": 1.800699103636967e-05, |
| "loss": 0.721204423904419, |
| "step": 1520 |
| }, |
| { |
| "epoch": 0.244335762052101, |
| "grad_norm": 0.03263983875513077, |
| "learning_rate": 1.7975313036215015e-05, |
| "loss": 0.7468688011169433, |
| "step": 1530 |
| }, |
| { |
| "epoch": 0.24593272781714742, |
| "grad_norm": 0.035935308784246445, |
| "learning_rate": 1.794341359276509e-05, |
| "loss": 0.6986902713775635, |
| "step": 1540 |
| }, |
| { |
| "epoch": 0.24752969358219384, |
| "grad_norm": 0.03797990456223488, |
| "learning_rate": 1.7911293591742855e-05, |
| "loss": 0.7199019908905029, |
| "step": 1550 |
| }, |
| { |
| "epoch": 0.24912665934724024, |
| "grad_norm": 0.024390801787376404, |
| "learning_rate": 1.787895392499529e-05, |
| "loss": 0.6761057853698731, |
| "step": 1560 |
| }, |
| { |
| "epoch": 0.25072362511228663, |
| "grad_norm": 0.0313715860247612, |
| "learning_rate": 1.7846395490468643e-05, |
| "loss": 0.6731560707092286, |
| "step": 1570 |
| }, |
| { |
| "epoch": 0.25232059087733305, |
| "grad_norm": 0.0433608703315258, |
| "learning_rate": 1.781361919218348e-05, |
| "loss": 0.685046911239624, |
| "step": 1580 |
| }, |
| { |
| "epoch": 0.25391755664237947, |
| "grad_norm": 0.050231028348207474, |
| "learning_rate": 1.7780625940209596e-05, |
| "loss": 0.7280925273895263, |
| "step": 1590 |
| }, |
| { |
| "epoch": 0.2555145224074259, |
| "grad_norm": 0.0476137213408947, |
| "learning_rate": 1.774741665064074e-05, |
| "loss": 0.728914451599121, |
| "step": 1600 |
| }, |
| { |
| "epoch": 0.2571114881724723, |
| "grad_norm": 0.03540361300110817, |
| "learning_rate": 1.771399224556919e-05, |
| "loss": 0.6937174320220947, |
| "step": 1610 |
| }, |
| { |
| "epoch": 0.25870845393751873, |
| "grad_norm": 0.038516197353601456, |
| "learning_rate": 1.7680353653060135e-05, |
| "loss": 0.6788946151733398, |
| "step": 1620 |
| }, |
| { |
| "epoch": 0.2603054197025651, |
| "grad_norm": 0.0472414456307888, |
| "learning_rate": 1.7646501807125905e-05, |
| "loss": 0.7246061325073242, |
| "step": 1630 |
| }, |
| { |
| "epoch": 0.2619023854676115, |
| "grad_norm": 0.026418814435601234, |
| "learning_rate": 1.7612437647700056e-05, |
| "loss": 0.6792353630065918, |
| "step": 1640 |
| }, |
| { |
| "epoch": 0.26349935123265794, |
| "grad_norm": 0.025900904089212418, |
| "learning_rate": 1.757816212061126e-05, |
| "loss": 0.6991750240325928, |
| "step": 1650 |
| }, |
| { |
| "epoch": 0.26509631699770436, |
| "grad_norm": 0.036637961864471436, |
| "learning_rate": 1.7543676177557042e-05, |
| "loss": 0.7213243007659912, |
| "step": 1660 |
| }, |
| { |
| "epoch": 0.2666932827627508, |
| "grad_norm": 0.031415775418281555, |
| "learning_rate": 1.750898077607735e-05, |
| "loss": 0.6988609790802002, |
| "step": 1670 |
| }, |
| { |
| "epoch": 0.2682902485277972, |
| "grad_norm": 0.036175280809402466, |
| "learning_rate": 1.7474076879527977e-05, |
| "loss": 0.6812397003173828, |
| "step": 1680 |
| }, |
| { |
| "epoch": 0.2698872142928436, |
| "grad_norm": 0.03462570160627365, |
| "learning_rate": 1.743896545705382e-05, |
| "loss": 0.6791361331939697, |
| "step": 1690 |
| }, |
| { |
| "epoch": 0.27148418005789, |
| "grad_norm": 0.03334948793053627, |
| "learning_rate": 1.740364748356195e-05, |
| "loss": 0.6668567180633544, |
| "step": 1700 |
| }, |
| { |
| "epoch": 0.2730811458229364, |
| "grad_norm": 0.03270823508501053, |
| "learning_rate": 1.7368123939694554e-05, |
| "loss": 0.6887404441833496, |
| "step": 1710 |
| }, |
| { |
| "epoch": 0.27467811158798283, |
| "grad_norm": 0.03120650351047516, |
| "learning_rate": 1.7332395811801706e-05, |
| "loss": 0.6659372806549072, |
| "step": 1720 |
| }, |
| { |
| "epoch": 0.27627507735302925, |
| "grad_norm": 0.042766936123371124, |
| "learning_rate": 1.7296464091913986e-05, |
| "loss": 0.7515771389007568, |
| "step": 1730 |
| }, |
| { |
| "epoch": 0.2778720431180757, |
| "grad_norm": 0.027079230174422264, |
| "learning_rate": 1.7260329777714923e-05, |
| "loss": 0.7227123260498047, |
| "step": 1740 |
| }, |
| { |
| "epoch": 0.2794690088831221, |
| "grad_norm": 0.03032068908214569, |
| "learning_rate": 1.722399387251329e-05, |
| "loss": 0.695145559310913, |
| "step": 1750 |
| }, |
| { |
| "epoch": 0.28106597464816846, |
| "grad_norm": 0.03420598804950714, |
| "learning_rate": 1.7187457385215274e-05, |
| "loss": 0.6715566158294678, |
| "step": 1760 |
| }, |
| { |
| "epoch": 0.2826629404132149, |
| "grad_norm": 0.037743836641311646, |
| "learning_rate": 1.7150721330296428e-05, |
| "loss": 0.6833428382873535, |
| "step": 1770 |
| }, |
| { |
| "epoch": 0.2842599061782613, |
| "grad_norm": 0.035041823983192444, |
| "learning_rate": 1.7113786727773528e-05, |
| "loss": 0.7029437065124512, |
| "step": 1780 |
| }, |
| { |
| "epoch": 0.2858568719433077, |
| "grad_norm": 0.03538225591182709, |
| "learning_rate": 1.7076654603176234e-05, |
| "loss": 0.7168496131896973, |
| "step": 1790 |
| }, |
| { |
| "epoch": 0.28745383770835414, |
| "grad_norm": 0.03331568092107773, |
| "learning_rate": 1.7039325987518623e-05, |
| "loss": 0.7009388923645019, |
| "step": 1800 |
| }, |
| { |
| "epoch": 0.28905080347340056, |
| "grad_norm": 0.033853549510240555, |
| "learning_rate": 1.700180191727057e-05, |
| "loss": 0.7036603450775146, |
| "step": 1810 |
| }, |
| { |
| "epoch": 0.29064776923844693, |
| "grad_norm": 0.03764677420258522, |
| "learning_rate": 1.696408343432895e-05, |
| "loss": 0.6595663547515869, |
| "step": 1820 |
| }, |
| { |
| "epoch": 0.29224473500349335, |
| "grad_norm": 0.028797749429941177, |
| "learning_rate": 1.6926171585988728e-05, |
| "loss": 0.6624193668365479, |
| "step": 1830 |
| }, |
| { |
| "epoch": 0.29384170076853977, |
| "grad_norm": 0.03950833901762962, |
| "learning_rate": 1.6888067424913863e-05, |
| "loss": 0.7282033920288086, |
| "step": 1840 |
| }, |
| { |
| "epoch": 0.2954386665335862, |
| "grad_norm": 0.032418642193078995, |
| "learning_rate": 1.6849772009108094e-05, |
| "loss": 0.6922025680541992, |
| "step": 1850 |
| }, |
| { |
| "epoch": 0.2970356322986326, |
| "grad_norm": 0.04753576219081879, |
| "learning_rate": 1.6811286401885554e-05, |
| "loss": 0.6543939590454102, |
| "step": 1860 |
| }, |
| { |
| "epoch": 0.29863259806367903, |
| "grad_norm": 0.03233656659722328, |
| "learning_rate": 1.677261167184125e-05, |
| "loss": 0.7344549655914306, |
| "step": 1870 |
| }, |
| { |
| "epoch": 0.3002295638287254, |
| "grad_norm": 0.030969245359301567, |
| "learning_rate": 1.673374889282139e-05, |
| "loss": 0.6734820365905761, |
| "step": 1880 |
| }, |
| { |
| "epoch": 0.3018265295937718, |
| "grad_norm": 0.03292040154337883, |
| "learning_rate": 1.6694699143893566e-05, |
| "loss": 0.6728582382202148, |
| "step": 1890 |
| }, |
| { |
| "epoch": 0.30342349535881824, |
| "grad_norm": 0.05033507198095322, |
| "learning_rate": 1.6655463509316797e-05, |
| "loss": 0.7132375240325928, |
| "step": 1900 |
| }, |
| { |
| "epoch": 0.30502046112386466, |
| "grad_norm": 0.03168455511331558, |
| "learning_rate": 1.6616043078511425e-05, |
| "loss": 0.7131926536560058, |
| "step": 1910 |
| }, |
| { |
| "epoch": 0.3066174268889111, |
| "grad_norm": 0.03114555962383747, |
| "learning_rate": 1.657643894602885e-05, |
| "loss": 0.7184583187103272, |
| "step": 1920 |
| }, |
| { |
| "epoch": 0.3082143926539575, |
| "grad_norm": 0.031201248988509178, |
| "learning_rate": 1.6536652211521155e-05, |
| "loss": 0.6606316566467285, |
| "step": 1930 |
| }, |
| { |
| "epoch": 0.30981135841900387, |
| "grad_norm": 0.035202883183956146, |
| "learning_rate": 1.6496683979710576e-05, |
| "loss": 0.7029526233673096, |
| "step": 1940 |
| }, |
| { |
| "epoch": 0.3114083241840503, |
| "grad_norm": 0.03515475615859032, |
| "learning_rate": 1.6456535360358807e-05, |
| "loss": 0.6629764080047608, |
| "step": 1950 |
| }, |
| { |
| "epoch": 0.3130052899490967, |
| "grad_norm": 0.03606207296252251, |
| "learning_rate": 1.6416207468236208e-05, |
| "loss": 0.6876577377319336, |
| "step": 1960 |
| }, |
| { |
| "epoch": 0.31460225571414313, |
| "grad_norm": 0.03580459579825401, |
| "learning_rate": 1.6375701423090846e-05, |
| "loss": 0.6879127979278564, |
| "step": 1970 |
| }, |
| { |
| "epoch": 0.31619922147918955, |
| "grad_norm": 0.048258859664201736, |
| "learning_rate": 1.6335018349617394e-05, |
| "loss": 0.7255002498626709, |
| "step": 1980 |
| }, |
| { |
| "epoch": 0.317796187244236, |
| "grad_norm": 0.046948857605457306, |
| "learning_rate": 1.629415937742591e-05, |
| "loss": 0.6919849395751954, |
| "step": 1990 |
| }, |
| { |
| "epoch": 0.31939315300928234, |
| "grad_norm": 0.03351585939526558, |
| "learning_rate": 1.625312564101049e-05, |
| "loss": 0.6886905193328857, |
| "step": 2000 |
| }, |
| { |
| "epoch": 0.31939315300928234, |
| "eval_loss": 0.6917140483856201, |
| "eval_runtime": 1706.5142, |
| "eval_samples_per_second": 3.262, |
| "eval_steps_per_second": 1.631, |
| "step": 2000 |
| }, |
| { |
| "epoch": 0.32099011877432876, |
| "grad_norm": 0.028385179117321968, |
| "learning_rate": 1.621191827971772e-05, |
| "loss": 0.7103724956512452, |
| "step": 2010 |
| }, |
| { |
| "epoch": 0.3225870845393752, |
| "grad_norm": 0.02977878600358963, |
| "learning_rate": 1.617053843771509e-05, |
| "loss": 0.7012845516204834, |
| "step": 2020 |
| }, |
| { |
| "epoch": 0.3241840503044216, |
| "grad_norm": 0.03950609639286995, |
| "learning_rate": 1.6128987263959212e-05, |
| "loss": 0.670585298538208, |
| "step": 2030 |
| }, |
| { |
| "epoch": 0.325781016069468, |
| "grad_norm": 0.03531676530838013, |
| "learning_rate": 1.6087265912163898e-05, |
| "loss": 0.6601489543914795, |
| "step": 2040 |
| }, |
| { |
| "epoch": 0.32737798183451444, |
| "grad_norm": 0.030362443998456, |
| "learning_rate": 1.6045375540768136e-05, |
| "loss": 0.7151473522186279, |
| "step": 2050 |
| }, |
| { |
| "epoch": 0.3289749475995608, |
| "grad_norm": 0.03205285966396332, |
| "learning_rate": 1.600331731290395e-05, |
| "loss": 0.6645094394683838, |
| "step": 2060 |
| }, |
| { |
| "epoch": 0.33057191336460723, |
| "grad_norm": 0.028163282200694084, |
| "learning_rate": 1.5961092396364053e-05, |
| "loss": 0.6868838310241699, |
| "step": 2070 |
| }, |
| { |
| "epoch": 0.33216887912965365, |
| "grad_norm": 0.03369533643126488, |
| "learning_rate": 1.5918701963569475e-05, |
| "loss": 0.7029123783111573, |
| "step": 2080 |
| }, |
| { |
| "epoch": 0.33376584489470007, |
| "grad_norm": 0.0379788838326931, |
| "learning_rate": 1.5876147191536983e-05, |
| "loss": 0.7173356056213379, |
| "step": 2090 |
| }, |
| { |
| "epoch": 0.3353628106597465, |
| "grad_norm": 0.0344555526971817, |
| "learning_rate": 1.583342926184639e-05, |
| "loss": 0.7149466037750244, |
| "step": 2100 |
| }, |
| { |
| "epoch": 0.3369597764247929, |
| "grad_norm": 0.03237446770071983, |
| "learning_rate": 1.5790549360607772e-05, |
| "loss": 0.6467396259307862, |
| "step": 2110 |
| }, |
| { |
| "epoch": 0.33855674218983933, |
| "grad_norm": 0.032042454928159714, |
| "learning_rate": 1.5747508678428515e-05, |
| "loss": 0.6929362297058106, |
| "step": 2120 |
| }, |
| { |
| "epoch": 0.3401537079548857, |
| "grad_norm": 0.042478688061237335, |
| "learning_rate": 1.570430841038027e-05, |
| "loss": 0.6702353000640869, |
| "step": 2130 |
| }, |
| { |
| "epoch": 0.3417506737199321, |
| "grad_norm": 0.04559747502207756, |
| "learning_rate": 1.5660949755965757e-05, |
| "loss": 0.6600518226623535, |
| "step": 2140 |
| }, |
| { |
| "epoch": 0.34334763948497854, |
| "grad_norm": 0.03681457042694092, |
| "learning_rate": 1.5617433919085474e-05, |
| "loss": 0.6836549758911132, |
| "step": 2150 |
| }, |
| { |
| "epoch": 0.34494460525002496, |
| "grad_norm": 0.02465015836060047, |
| "learning_rate": 1.5573762108004262e-05, |
| "loss": 0.7304904460906982, |
| "step": 2160 |
| }, |
| { |
| "epoch": 0.3465415710150714, |
| "grad_norm": 0.03969291225075722, |
| "learning_rate": 1.5529935535317746e-05, |
| "loss": 0.6897444725036621, |
| "step": 2170 |
| }, |
| { |
| "epoch": 0.3481385367801178, |
| "grad_norm": 0.03233686834573746, |
| "learning_rate": 1.548595541791869e-05, |
| "loss": 0.684897518157959, |
| "step": 2180 |
| }, |
| { |
| "epoch": 0.34973550254516417, |
| "grad_norm": 0.03138101100921631, |
| "learning_rate": 1.544182297696319e-05, |
| "loss": 0.7246172428131104, |
| "step": 2190 |
| }, |
| { |
| "epoch": 0.3513324683102106, |
| "grad_norm": 0.03732126206159592, |
| "learning_rate": 1.5397539437836765e-05, |
| "loss": 0.6686253070831298, |
| "step": 2200 |
| }, |
| { |
| "epoch": 0.352929434075257, |
| "grad_norm": 0.028484191745519638, |
| "learning_rate": 1.535310603012035e-05, |
| "loss": 0.6723896503448487, |
| "step": 2210 |
| }, |
| { |
| "epoch": 0.35452639984030343, |
| "grad_norm": 0.04804018512368202, |
| "learning_rate": 1.5308523987556144e-05, |
| "loss": 0.7210356712341308, |
| "step": 2220 |
| }, |
| { |
| "epoch": 0.35612336560534985, |
| "grad_norm": 0.035007379949092865, |
| "learning_rate": 1.5263794548013347e-05, |
| "loss": 0.6699332237243653, |
| "step": 2230 |
| }, |
| { |
| "epoch": 0.35772033137039627, |
| "grad_norm": 0.034309230744838715, |
| "learning_rate": 1.5218918953453808e-05, |
| "loss": 0.6913293838500977, |
| "step": 2240 |
| }, |
| { |
| "epoch": 0.35931729713544264, |
| "grad_norm": 0.05658178776502609, |
| "learning_rate": 1.5173898449897527e-05, |
| "loss": 0.6920335292816162, |
| "step": 2250 |
| }, |
| { |
| "epoch": 0.36091426290048906, |
| "grad_norm": 0.03614016994833946, |
| "learning_rate": 1.5128734287388051e-05, |
| "loss": 0.6855293273925781, |
| "step": 2260 |
| }, |
| { |
| "epoch": 0.3625112286655355, |
| "grad_norm": 0.0348992682993412, |
| "learning_rate": 1.5083427719957792e-05, |
| "loss": 0.7068172454833984, |
| "step": 2270 |
| }, |
| { |
| "epoch": 0.3641081944305819, |
| "grad_norm": 0.037384193390607834, |
| "learning_rate": 1.5037980005593178e-05, |
| "loss": 0.7026480197906494, |
| "step": 2280 |
| }, |
| { |
| "epoch": 0.3657051601956283, |
| "grad_norm": 0.03162362426519394, |
| "learning_rate": 1.4992392406199741e-05, |
| "loss": 0.7369038581848144, |
| "step": 2290 |
| }, |
| { |
| "epoch": 0.36730212596067474, |
| "grad_norm": 0.03644828870892525, |
| "learning_rate": 1.4946666187567073e-05, |
| "loss": 0.7133200168609619, |
| "step": 2300 |
| }, |
| { |
| "epoch": 0.3688990917257211, |
| "grad_norm": 0.05075901374220848, |
| "learning_rate": 1.490080261933368e-05, |
| "loss": 0.7146055221557617, |
| "step": 2310 |
| }, |
| { |
| "epoch": 0.3704960574907675, |
| "grad_norm": 0.029153743758797646, |
| "learning_rate": 1.4854802974951732e-05, |
| "loss": 0.6756022453308106, |
| "step": 2320 |
| }, |
| { |
| "epoch": 0.37209302325581395, |
| "grad_norm": 0.030742282047867775, |
| "learning_rate": 1.4808668531651698e-05, |
| "loss": 0.6855404376983643, |
| "step": 2330 |
| }, |
| { |
| "epoch": 0.37368998902086037, |
| "grad_norm": 0.03525373339653015, |
| "learning_rate": 1.4762400570406889e-05, |
| "loss": 0.6626357078552246, |
| "step": 2340 |
| }, |
| { |
| "epoch": 0.3752869547859068, |
| "grad_norm": 0.036246221512556076, |
| "learning_rate": 1.4716000375897897e-05, |
| "loss": 0.714638090133667, |
| "step": 2350 |
| }, |
| { |
| "epoch": 0.3768839205509532, |
| "grad_norm": 0.03430463373661041, |
| "learning_rate": 1.4669469236476897e-05, |
| "loss": 0.6736037731170654, |
| "step": 2360 |
| }, |
| { |
| "epoch": 0.3784808863159996, |
| "grad_norm": 0.035327013581991196, |
| "learning_rate": 1.4622808444131904e-05, |
| "loss": 0.6907155036926269, |
| "step": 2370 |
| }, |
| { |
| "epoch": 0.380077852081046, |
| "grad_norm": 0.039804793894290924, |
| "learning_rate": 1.457601929445089e-05, |
| "loss": 0.6726024627685547, |
| "step": 2380 |
| }, |
| { |
| "epoch": 0.3816748178460924, |
| "grad_norm": 0.03689780458807945, |
| "learning_rate": 1.4529103086585803e-05, |
| "loss": 0.6932793617248535, |
| "step": 2390 |
| }, |
| { |
| "epoch": 0.38327178361113884, |
| "grad_norm": 0.03002229705452919, |
| "learning_rate": 1.4482061123216507e-05, |
| "loss": 0.6448677062988282, |
| "step": 2400 |
| }, |
| { |
| "epoch": 0.38486874937618526, |
| "grad_norm": 0.04287157952785492, |
| "learning_rate": 1.4434894710514596e-05, |
| "loss": 0.6573331356048584, |
| "step": 2410 |
| }, |
| { |
| "epoch": 0.3864657151412317, |
| "grad_norm": 0.033608682453632355, |
| "learning_rate": 1.4387605158107146e-05, |
| "loss": 0.6761981010437011, |
| "step": 2420 |
| }, |
| { |
| "epoch": 0.38806268090627805, |
| "grad_norm": 0.0340590700507164, |
| "learning_rate": 1.4340193779040335e-05, |
| "loss": 0.7119544982910156, |
| "step": 2430 |
| }, |
| { |
| "epoch": 0.38965964667132447, |
| "grad_norm": 0.04576168581843376, |
| "learning_rate": 1.429266188974299e-05, |
| "loss": 0.722407054901123, |
| "step": 2440 |
| }, |
| { |
| "epoch": 0.3912566124363709, |
| "grad_norm": 0.030462034046649933, |
| "learning_rate": 1.4245010809990052e-05, |
| "loss": 0.6741881847381592, |
| "step": 2450 |
| }, |
| { |
| "epoch": 0.3928535782014173, |
| "grad_norm": 0.024982597678899765, |
| "learning_rate": 1.419724186286589e-05, |
| "loss": 0.6923768997192383, |
| "step": 2460 |
| }, |
| { |
| "epoch": 0.39445054396646373, |
| "grad_norm": 0.028493596240878105, |
| "learning_rate": 1.414935637472761e-05, |
| "loss": 0.7135504722595215, |
| "step": 2470 |
| }, |
| { |
| "epoch": 0.39604750973151015, |
| "grad_norm": 0.03309721499681473, |
| "learning_rate": 1.4101355675168197e-05, |
| "loss": 0.6870269298553466, |
| "step": 2480 |
| }, |
| { |
| "epoch": 0.3976444754965565, |
| "grad_norm": 0.033349502831697464, |
| "learning_rate": 1.4053241096979611e-05, |
| "loss": 0.7310259819030762, |
| "step": 2490 |
| }, |
| { |
| "epoch": 0.39924144126160294, |
| "grad_norm": 0.0456295944750309, |
| "learning_rate": 1.4005013976115777e-05, |
| "loss": 0.6850334644317627, |
| "step": 2500 |
| }, |
| { |
| "epoch": 0.39924144126160294, |
| "eval_loss": 0.6847082376480103, |
| "eval_runtime": 1727.6141, |
| "eval_samples_per_second": 3.222, |
| "eval_steps_per_second": 1.611, |
| "step": 2500 |
| }, |
| { |
| "epoch": 0.40083840702664936, |
| "grad_norm": 0.04150133952498436, |
| "learning_rate": 1.3956675651655489e-05, |
| "loss": 0.6748110771179199, |
| "step": 2510 |
| }, |
| { |
| "epoch": 0.4024353727916958, |
| "grad_norm": 0.05328802019357681, |
| "learning_rate": 1.390822746576523e-05, |
| "loss": 0.6966916561126709, |
| "step": 2520 |
| }, |
| { |
| "epoch": 0.4040323385567422, |
| "grad_norm": 0.04510056599974632, |
| "learning_rate": 1.3859670763661908e-05, |
| "loss": 0.6688936710357666, |
| "step": 2530 |
| }, |
| { |
| "epoch": 0.4056293043217886, |
| "grad_norm": 0.03648371994495392, |
| "learning_rate": 1.38110068935755e-05, |
| "loss": 0.6503877162933349, |
| "step": 2540 |
| }, |
| { |
| "epoch": 0.40722627008683504, |
| "grad_norm": 0.03958805277943611, |
| "learning_rate": 1.3762237206711624e-05, |
| "loss": 0.6709461212158203, |
| "step": 2550 |
| }, |
| { |
| "epoch": 0.4088232358518814, |
| "grad_norm": 0.03620138764381409, |
| "learning_rate": 1.3713363057214006e-05, |
| "loss": 0.6775450706481934, |
| "step": 2560 |
| }, |
| { |
| "epoch": 0.4104202016169278, |
| "grad_norm": 0.036233801394701004, |
| "learning_rate": 1.3664385802126904e-05, |
| "loss": 0.7039198875427246, |
| "step": 2570 |
| }, |
| { |
| "epoch": 0.41201716738197425, |
| "grad_norm": 0.034267205744981766, |
| "learning_rate": 1.3615306801357413e-05, |
| "loss": 0.6893480777740478, |
| "step": 2580 |
| }, |
| { |
| "epoch": 0.41361413314702067, |
| "grad_norm": 0.04056892171502113, |
| "learning_rate": 1.3566127417637698e-05, |
| "loss": 0.6852858543395997, |
| "step": 2590 |
| }, |
| { |
| "epoch": 0.4152110989120671, |
| "grad_norm": 0.04654235392808914, |
| "learning_rate": 1.351684901648718e-05, |
| "loss": 0.6932817459106445, |
| "step": 2600 |
| }, |
| { |
| "epoch": 0.4168080646771135, |
| "grad_norm": 0.05177690088748932, |
| "learning_rate": 1.3467472966174595e-05, |
| "loss": 0.663325309753418, |
| "step": 2610 |
| }, |
| { |
| "epoch": 0.4184050304421599, |
| "grad_norm": 0.04320165514945984, |
| "learning_rate": 1.3418000637680026e-05, |
| "loss": 0.6278695106506348, |
| "step": 2620 |
| }, |
| { |
| "epoch": 0.4200019962072063, |
| "grad_norm": 0.0365922674536705, |
| "learning_rate": 1.336843340465682e-05, |
| "loss": 0.6674781799316406, |
| "step": 2630 |
| }, |
| { |
| "epoch": 0.4215989619722527, |
| "grad_norm": 0.03406934812664986, |
| "learning_rate": 1.3318772643393447e-05, |
| "loss": 0.6559438705444336, |
| "step": 2640 |
| }, |
| { |
| "epoch": 0.42319592773729914, |
| "grad_norm": 0.030784646049141884, |
| "learning_rate": 1.3269019732775304e-05, |
| "loss": 0.7225832462310791, |
| "step": 2650 |
| }, |
| { |
| "epoch": 0.42479289350234556, |
| "grad_norm": 0.0360528789460659, |
| "learning_rate": 1.3219176054246404e-05, |
| "loss": 0.7170959949493408, |
| "step": 2660 |
| }, |
| { |
| "epoch": 0.426389859267392, |
| "grad_norm": 0.0355147160589695, |
| "learning_rate": 1.316924299177103e-05, |
| "loss": 0.683486795425415, |
| "step": 2670 |
| }, |
| { |
| "epoch": 0.42798682503243834, |
| "grad_norm": 0.046759720891714096, |
| "learning_rate": 1.3119221931795321e-05, |
| "loss": 0.7131591320037842, |
| "step": 2680 |
| }, |
| { |
| "epoch": 0.42958379079748477, |
| "grad_norm": 0.03554374352097511, |
| "learning_rate": 1.3069114263208746e-05, |
| "loss": 0.6715425491333008, |
| "step": 2690 |
| }, |
| { |
| "epoch": 0.4311807565625312, |
| "grad_norm": 0.04263368248939514, |
| "learning_rate": 1.3018921377305574e-05, |
| "loss": 0.6871604442596435, |
| "step": 2700 |
| }, |
| { |
| "epoch": 0.4327777223275776, |
| "grad_norm": 0.04496333748102188, |
| "learning_rate": 1.2968644667746207e-05, |
| "loss": 0.6588253021240235, |
| "step": 2710 |
| }, |
| { |
| "epoch": 0.43437468809262403, |
| "grad_norm": 0.036086685955524445, |
| "learning_rate": 1.2918285530518519e-05, |
| "loss": 0.6757514953613282, |
| "step": 2720 |
| }, |
| { |
| "epoch": 0.43597165385767045, |
| "grad_norm": 0.03308691456913948, |
| "learning_rate": 1.2867845363899075e-05, |
| "loss": 0.6658079624176025, |
| "step": 2730 |
| }, |
| { |
| "epoch": 0.4375686196227168, |
| "grad_norm": 0.04500787332653999, |
| "learning_rate": 1.2817325568414299e-05, |
| "loss": 0.6812032222747803, |
| "step": 2740 |
| }, |
| { |
| "epoch": 0.43916558538776324, |
| "grad_norm": 0.03841916844248772, |
| "learning_rate": 1.2766727546801612e-05, |
| "loss": 0.6864322662353516, |
| "step": 2750 |
| }, |
| { |
| "epoch": 0.44076255115280966, |
| "grad_norm": 0.037914253771305084, |
| "learning_rate": 1.2716052703970459e-05, |
| "loss": 0.7053091526031494, |
| "step": 2760 |
| }, |
| { |
| "epoch": 0.4423595169178561, |
| "grad_norm": 0.03686266764998436, |
| "learning_rate": 1.2665302446963312e-05, |
| "loss": 0.7154440879821777, |
| "step": 2770 |
| }, |
| { |
| "epoch": 0.4439564826829025, |
| "grad_norm": 0.04227980226278305, |
| "learning_rate": 1.261447818491661e-05, |
| "loss": 0.716032886505127, |
| "step": 2780 |
| }, |
| { |
| "epoch": 0.4455534484479489, |
| "grad_norm": 0.04683419317007065, |
| "learning_rate": 1.2563581329021608e-05, |
| "loss": 0.6738180637359619, |
| "step": 2790 |
| }, |
| { |
| "epoch": 0.4471504142129953, |
| "grad_norm": 0.05264711007475853, |
| "learning_rate": 1.2512613292485218e-05, |
| "loss": 0.7129797458648681, |
| "step": 2800 |
| }, |
| { |
| "epoch": 0.4487473799780417, |
| "grad_norm": 0.04030587151646614, |
| "learning_rate": 1.2461575490490753e-05, |
| "loss": 0.727331018447876, |
| "step": 2810 |
| }, |
| { |
| "epoch": 0.4503443457430881, |
| "grad_norm": 0.04141924902796745, |
| "learning_rate": 1.2410469340158655e-05, |
| "loss": 0.676334810256958, |
| "step": 2820 |
| }, |
| { |
| "epoch": 0.45194131150813455, |
| "grad_norm": 0.04476945474743843, |
| "learning_rate": 1.2359296260507117e-05, |
| "loss": 0.6823254108428956, |
| "step": 2830 |
| }, |
| { |
| "epoch": 0.45353827727318097, |
| "grad_norm": 0.03869614750146866, |
| "learning_rate": 1.23080576724127e-05, |
| "loss": 0.6882941722869873, |
| "step": 2840 |
| }, |
| { |
| "epoch": 0.4551352430382274, |
| "grad_norm": 0.031092161312699318, |
| "learning_rate": 1.2256754998570887e-05, |
| "loss": 0.6744899272918701, |
| "step": 2850 |
| }, |
| { |
| "epoch": 0.45673220880327375, |
| "grad_norm": 0.03507015109062195, |
| "learning_rate": 1.2205389663456566e-05, |
| "loss": 0.7103989124298096, |
| "step": 2860 |
| }, |
| { |
| "epoch": 0.4583291745683202, |
| "grad_norm": 0.032851576805114746, |
| "learning_rate": 1.2153963093284483e-05, |
| "loss": 0.7116181373596191, |
| "step": 2870 |
| }, |
| { |
| "epoch": 0.4599261403333666, |
| "grad_norm": 0.05045435577630997, |
| "learning_rate": 1.210247671596965e-05, |
| "loss": 0.6489015579223633, |
| "step": 2880 |
| }, |
| { |
| "epoch": 0.461523106098413, |
| "grad_norm": 0.032028377056121826, |
| "learning_rate": 1.2050931961087673e-05, |
| "loss": 0.6941986560821534, |
| "step": 2890 |
| }, |
| { |
| "epoch": 0.46312007186345944, |
| "grad_norm": 0.038908179849386215, |
| "learning_rate": 1.1999330259835096e-05, |
| "loss": 0.6908363342285156, |
| "step": 2900 |
| }, |
| { |
| "epoch": 0.46471703762850586, |
| "grad_norm": 0.03464297205209732, |
| "learning_rate": 1.1947673044989627e-05, |
| "loss": 0.7050665378570556, |
| "step": 2910 |
| }, |
| { |
| "epoch": 0.4663140033935522, |
| "grad_norm": 0.028774775564670563, |
| "learning_rate": 1.1895961750870375e-05, |
| "loss": 0.6700276851654052, |
| "step": 2920 |
| }, |
| { |
| "epoch": 0.46791096915859864, |
| "grad_norm": 0.03465382754802704, |
| "learning_rate": 1.1844197813298018e-05, |
| "loss": 0.6821481227874756, |
| "step": 2930 |
| }, |
| { |
| "epoch": 0.46950793492364506, |
| "grad_norm": 0.03162831813097, |
| "learning_rate": 1.1792382669554938e-05, |
| "loss": 0.6927790164947509, |
| "step": 2940 |
| }, |
| { |
| "epoch": 0.4711049006886915, |
| "grad_norm": 0.03354581072926521, |
| "learning_rate": 1.1740517758345312e-05, |
| "loss": 0.7100222110748291, |
| "step": 2950 |
| }, |
| { |
| "epoch": 0.4727018664537379, |
| "grad_norm": 0.033408552408218384, |
| "learning_rate": 1.168860451975516e-05, |
| "loss": 0.6876598358154297, |
| "step": 2960 |
| }, |
| { |
| "epoch": 0.4742988322187843, |
| "grad_norm": 0.04334475100040436, |
| "learning_rate": 1.1636644395212375e-05, |
| "loss": 0.6866564273834228, |
| "step": 2970 |
| }, |
| { |
| "epoch": 0.47589579798383075, |
| "grad_norm": 0.040464501827955246, |
| "learning_rate": 1.1584638827446684e-05, |
| "loss": 0.6588007926940918, |
| "step": 2980 |
| }, |
| { |
| "epoch": 0.4774927637488771, |
| "grad_norm": 0.04957546666264534, |
| "learning_rate": 1.1532589260449589e-05, |
| "loss": 0.7070925235748291, |
| "step": 2990 |
| }, |
| { |
| "epoch": 0.47908972951392353, |
| "grad_norm": 0.038624707609415054, |
| "learning_rate": 1.1480497139434294e-05, |
| "loss": 0.6919936180114746, |
| "step": 3000 |
| }, |
| { |
| "epoch": 0.47908972951392353, |
| "eval_loss": 0.6803271770477295, |
| "eval_runtime": 1741.9445, |
| "eval_samples_per_second": 3.195, |
| "eval_steps_per_second": 1.598, |
| "step": 3000 |
| }, |
| { |
| "epoch": 0.48068669527896996, |
| "grad_norm": 0.04100416228175163, |
| "learning_rate": 1.1428363910795547e-05, |
| "loss": 0.6852492809295654, |
| "step": 3010 |
| }, |
| { |
| "epoch": 0.4822836610440164, |
| "grad_norm": 0.033158283680677414, |
| "learning_rate": 1.1376191022069501e-05, |
| "loss": 0.667516565322876, |
| "step": 3020 |
| }, |
| { |
| "epoch": 0.4838806268090628, |
| "grad_norm": 0.03429755941033363, |
| "learning_rate": 1.1323979921893514e-05, |
| "loss": 0.6956333160400391, |
| "step": 3030 |
| }, |
| { |
| "epoch": 0.4854775925741092, |
| "grad_norm": 0.038232944905757904, |
| "learning_rate": 1.1271732059965925e-05, |
| "loss": 0.6716940402984619, |
| "step": 3040 |
| }, |
| { |
| "epoch": 0.4870745583391556, |
| "grad_norm": 0.039248283952474594, |
| "learning_rate": 1.1219448887005805e-05, |
| "loss": 0.6984889507293701, |
| "step": 3050 |
| }, |
| { |
| "epoch": 0.488671524104202, |
| "grad_norm": 0.03670873865485191, |
| "learning_rate": 1.1167131854712676e-05, |
| "loss": 0.6709868907928467, |
| "step": 3060 |
| }, |
| { |
| "epoch": 0.4902684898692484, |
| "grad_norm": 0.038310978561639786, |
| "learning_rate": 1.11147824157262e-05, |
| "loss": 0.6672838211059571, |
| "step": 3070 |
| }, |
| { |
| "epoch": 0.49186545563429485, |
| "grad_norm": 0.049189358949661255, |
| "learning_rate": 1.1062402023585846e-05, |
| "loss": 0.6705825805664063, |
| "step": 3080 |
| }, |
| { |
| "epoch": 0.49346242139934127, |
| "grad_norm": 0.0355689600110054, |
| "learning_rate": 1.1009992132690523e-05, |
| "loss": 0.6964725017547607, |
| "step": 3090 |
| }, |
| { |
| "epoch": 0.4950593871643877, |
| "grad_norm": 0.03348139300942421, |
| "learning_rate": 1.0957554198258225e-05, |
| "loss": 0.6852751255035401, |
| "step": 3100 |
| }, |
| { |
| "epoch": 0.49665635292943405, |
| "grad_norm": 0.03575340285897255, |
| "learning_rate": 1.0905089676285588e-05, |
| "loss": 0.670794153213501, |
| "step": 3110 |
| }, |
| { |
| "epoch": 0.4982533186944805, |
| "grad_norm": 0.04119575023651123, |
| "learning_rate": 1.0852600023507482e-05, |
| "loss": 0.6651177406311035, |
| "step": 3120 |
| }, |
| { |
| "epoch": 0.4998502844595269, |
| "grad_norm": 0.04122074693441391, |
| "learning_rate": 1.0800086697356576e-05, |
| "loss": 0.6976897239685058, |
| "step": 3130 |
| }, |
| { |
| "epoch": 0.5014472502245733, |
| "grad_norm": 0.034223176538944244, |
| "learning_rate": 1.0747551155922836e-05, |
| "loss": 0.6890422821044921, |
| "step": 3140 |
| }, |
| { |
| "epoch": 0.5030442159896197, |
| "grad_norm": 0.03588424623012543, |
| "learning_rate": 1.069499485791307e-05, |
| "loss": 0.702085018157959, |
| "step": 3150 |
| }, |
| { |
| "epoch": 0.5046411817546661, |
| "grad_norm": 0.03808250650763512, |
| "learning_rate": 1.0642419262610417e-05, |
| "loss": 0.6910345554351807, |
| "step": 3160 |
| }, |
| { |
| "epoch": 0.5062381475197125, |
| "grad_norm": 0.03852913901209831, |
| "learning_rate": 1.0589825829833825e-05, |
| "loss": 0.6685952186584473, |
| "step": 3170 |
| }, |
| { |
| "epoch": 0.5078351132847589, |
| "grad_norm": 0.04528075084090233, |
| "learning_rate": 1.0537216019897514e-05, |
| "loss": 0.6971333980560303, |
| "step": 3180 |
| }, |
| { |
| "epoch": 0.5094320790498054, |
| "grad_norm": 0.048296891152858734, |
| "learning_rate": 1.0484591293570437e-05, |
| "loss": 0.6506364345550537, |
| "step": 3190 |
| }, |
| { |
| "epoch": 0.5110290448148518, |
| "grad_norm": 0.03280564025044441, |
| "learning_rate": 1.0431953112035722e-05, |
| "loss": 0.7106664657592774, |
| "step": 3200 |
| }, |
| { |
| "epoch": 0.5126260105798982, |
| "grad_norm": 0.03674250841140747, |
| "learning_rate": 1.0379302936850083e-05, |
| "loss": 0.6798049449920655, |
| "step": 3210 |
| }, |
| { |
| "epoch": 0.5142229763449446, |
| "grad_norm": 0.02851015329360962, |
| "learning_rate": 1.0326642229903265e-05, |
| "loss": 0.6674058914184571, |
| "step": 3220 |
| }, |
| { |
| "epoch": 0.515819942109991, |
| "grad_norm": 0.04876931011676788, |
| "learning_rate": 1.0273972453377433e-05, |
| "loss": 0.6956850051879883, |
| "step": 3230 |
| }, |
| { |
| "epoch": 0.5174169078750375, |
| "grad_norm": 0.04633413255214691, |
| "learning_rate": 1.022129506970657e-05, |
| "loss": 0.6859352111816406, |
| "step": 3240 |
| }, |
| { |
| "epoch": 0.5190138736400839, |
| "grad_norm": 0.03311692923307419, |
| "learning_rate": 1.0168611541535898e-05, |
| "loss": 0.7103584289550782, |
| "step": 3250 |
| }, |
| { |
| "epoch": 0.5206108394051302, |
| "grad_norm": 0.047599345445632935, |
| "learning_rate": 1.0115923331681231e-05, |
| "loss": 0.6435458183288574, |
| "step": 3260 |
| }, |
| { |
| "epoch": 0.5222078051701766, |
| "grad_norm": 0.033676277846097946, |
| "learning_rate": 1.0063231903088384e-05, |
| "loss": 0.6581607341766358, |
| "step": 3270 |
| }, |
| { |
| "epoch": 0.523804770935223, |
| "grad_norm": 0.04461989924311638, |
| "learning_rate": 1.0010538718792544e-05, |
| "loss": 0.7152835369110108, |
| "step": 3280 |
| }, |
| { |
| "epoch": 0.5254017367002695, |
| "grad_norm": 0.04492766410112381, |
| "learning_rate": 9.957845241877639e-06, |
| "loss": 0.7056149482727051, |
| "step": 3290 |
| }, |
| { |
| "epoch": 0.5269987024653159, |
| "grad_norm": 0.02856948785483837, |
| "learning_rate": 9.905152935435733e-06, |
| "loss": 0.6655986785888672, |
| "step": 3300 |
| }, |
| { |
| "epoch": 0.5285956682303623, |
| "grad_norm": 0.0438731387257576, |
| "learning_rate": 9.852463262526383e-06, |
| "loss": 0.7045872688293457, |
| "step": 3310 |
| }, |
| { |
| "epoch": 0.5301926339954087, |
| "grad_norm": 0.04025017097592354, |
| "learning_rate": 9.799777686136025e-06, |
| "loss": 0.6844919204711915, |
| "step": 3320 |
| }, |
| { |
| "epoch": 0.5317895997604551, |
| "grad_norm": 0.04453453794121742, |
| "learning_rate": 9.747097669137352e-06, |
| "loss": 0.6857856750488281, |
| "step": 3330 |
| }, |
| { |
| "epoch": 0.5333865655255016, |
| "grad_norm": 0.034065935760736465, |
| "learning_rate": 9.694424674248687e-06, |
| "loss": 0.6782355785369873, |
| "step": 3340 |
| }, |
| { |
| "epoch": 0.534983531290548, |
| "grad_norm": 0.03808142989873886, |
| "learning_rate": 9.641760163993394e-06, |
| "loss": 0.7040555000305175, |
| "step": 3350 |
| }, |
| { |
| "epoch": 0.5365804970555944, |
| "grad_norm": 0.03482348844408989, |
| "learning_rate": 9.58910560065923e-06, |
| "loss": 0.6730469226837158, |
| "step": 3360 |
| }, |
| { |
| "epoch": 0.5381774628206408, |
| "grad_norm": 0.04541369527578354, |
| "learning_rate": 9.536462446257777e-06, |
| "loss": 0.7102465152740478, |
| "step": 3370 |
| }, |
| { |
| "epoch": 0.5397744285856872, |
| "grad_norm": 0.043802615255117416, |
| "learning_rate": 9.483832162483841e-06, |
| "loss": 0.6941490173339844, |
| "step": 3380 |
| }, |
| { |
| "epoch": 0.5413713943507336, |
| "grad_norm": 0.03652375563979149, |
| "learning_rate": 9.431216210674849e-06, |
| "loss": 0.704205322265625, |
| "step": 3390 |
| }, |
| { |
| "epoch": 0.54296836011578, |
| "grad_norm": 0.03886988013982773, |
| "learning_rate": 9.378616051770287e-06, |
| "loss": 0.6582465171813965, |
| "step": 3400 |
| }, |
| { |
| "epoch": 0.5445653258808264, |
| "grad_norm": 0.044551681727170944, |
| "learning_rate": 9.326033146271144e-06, |
| "loss": 0.7078223705291748, |
| "step": 3410 |
| }, |
| { |
| "epoch": 0.5461622916458728, |
| "grad_norm": 0.03762778267264366, |
| "learning_rate": 9.273468954199333e-06, |
| "loss": 0.6956658840179444, |
| "step": 3420 |
| }, |
| { |
| "epoch": 0.5477592574109192, |
| "grad_norm": 0.03794670104980469, |
| "learning_rate": 9.220924935057184e-06, |
| "loss": 0.6778861522674561, |
| "step": 3430 |
| }, |
| { |
| "epoch": 0.5493562231759657, |
| "grad_norm": 0.03766616806387901, |
| "learning_rate": 9.168402547786897e-06, |
| "loss": 0.6759885311126709, |
| "step": 3440 |
| }, |
| { |
| "epoch": 0.5509531889410121, |
| "grad_norm": 0.04172630235552788, |
| "learning_rate": 9.115903250730035e-06, |
| "loss": 0.6739662170410157, |
| "step": 3450 |
| }, |
| { |
| "epoch": 0.5525501547060585, |
| "grad_norm": 0.04936975613236427, |
| "learning_rate": 9.063428501587043e-06, |
| "loss": 0.6581857681274415, |
| "step": 3460 |
| }, |
| { |
| "epoch": 0.5541471204711049, |
| "grad_norm": 0.029507510364055634, |
| "learning_rate": 9.010979757376766e-06, |
| "loss": 0.6508080005645752, |
| "step": 3470 |
| }, |
| { |
| "epoch": 0.5557440862361513, |
| "grad_norm": 0.03853528946638107, |
| "learning_rate": 8.958558474395987e-06, |
| "loss": 0.6520394325256348, |
| "step": 3480 |
| }, |
| { |
| "epoch": 0.5573410520011978, |
| "grad_norm": 0.04168470576405525, |
| "learning_rate": 8.906166108178999e-06, |
| "loss": 0.6652052402496338, |
| "step": 3490 |
| }, |
| { |
| "epoch": 0.5589380177662442, |
| "grad_norm": 0.05398240312933922, |
| "learning_rate": 8.853804113457203e-06, |
| "loss": 0.6777332782745361, |
| "step": 3500 |
| }, |
| { |
| "epoch": 0.5589380177662442, |
| "eval_loss": 0.677163302898407, |
| "eval_runtime": 1733.3138, |
| "eval_samples_per_second": 3.211, |
| "eval_steps_per_second": 1.606, |
| "step": 3500 |
| }, |
| { |
| "epoch": 0.5605349835312905, |
| "grad_norm": 0.03181201219558716, |
| "learning_rate": 8.801473944118683e-06, |
| "loss": 0.687204122543335, |
| "step": 3510 |
| }, |
| { |
| "epoch": 0.5621319492963369, |
| "grad_norm": 0.03610098734498024, |
| "learning_rate": 8.749177053167865e-06, |
| "loss": 0.7028386116027832, |
| "step": 3520 |
| }, |
| { |
| "epoch": 0.5637289150613833, |
| "grad_norm": 0.0450638048350811, |
| "learning_rate": 8.696914892685172e-06, |
| "loss": 0.6494070529937744, |
| "step": 3530 |
| }, |
| { |
| "epoch": 0.5653258808264298, |
| "grad_norm": 0.053478166460990906, |
| "learning_rate": 8.644688913786678e-06, |
| "loss": 0.6962610721588135, |
| "step": 3540 |
| }, |
| { |
| "epoch": 0.5669228465914762, |
| "grad_norm": 0.03924020007252693, |
| "learning_rate": 8.592500566583856e-06, |
| "loss": 0.6941705703735351, |
| "step": 3550 |
| }, |
| { |
| "epoch": 0.5685198123565226, |
| "grad_norm": 0.037506818771362305, |
| "learning_rate": 8.540351300143284e-06, |
| "loss": 0.6824637413024902, |
| "step": 3560 |
| }, |
| { |
| "epoch": 0.570116778121569, |
| "grad_norm": 0.03883660212159157, |
| "learning_rate": 8.488242562446416e-06, |
| "loss": 0.7028771877288819, |
| "step": 3570 |
| }, |
| { |
| "epoch": 0.5717137438866154, |
| "grad_norm": 0.047728635370731354, |
| "learning_rate": 8.436175800349393e-06, |
| "loss": 0.7102296829223633, |
| "step": 3580 |
| }, |
| { |
| "epoch": 0.5733107096516619, |
| "grad_norm": 0.0423489511013031, |
| "learning_rate": 8.384152459542849e-06, |
| "loss": 0.7113842487335205, |
| "step": 3590 |
| }, |
| { |
| "epoch": 0.5749076754167083, |
| "grad_norm": 0.034012071788311005, |
| "learning_rate": 8.332173984511774e-06, |
| "loss": 0.6995905876159668, |
| "step": 3600 |
| }, |
| { |
| "epoch": 0.5765046411817547, |
| "grad_norm": 0.02976168505847454, |
| "learning_rate": 8.280241818495423e-06, |
| "loss": 0.7066914558410644, |
| "step": 3610 |
| }, |
| { |
| "epoch": 0.5781016069468011, |
| "grad_norm": 0.04333082213997841, |
| "learning_rate": 8.228357403447225e-06, |
| "loss": 0.6559367656707764, |
| "step": 3620 |
| }, |
| { |
| "epoch": 0.5796985727118474, |
| "grad_norm": 0.0726407989859581, |
| "learning_rate": 8.176522179994744e-06, |
| "loss": 0.691002082824707, |
| "step": 3630 |
| }, |
| { |
| "epoch": 0.5812955384768939, |
| "grad_norm": 0.035537876188755035, |
| "learning_rate": 8.124737587399697e-06, |
| "loss": 0.7173010349273682, |
| "step": 3640 |
| }, |
| { |
| "epoch": 0.5828925042419403, |
| "grad_norm": 0.036615390330553055, |
| "learning_rate": 8.073005063517973e-06, |
| "loss": 0.712274169921875, |
| "step": 3650 |
| }, |
| { |
| "epoch": 0.5844894700069867, |
| "grad_norm": 0.02762191742658615, |
| "learning_rate": 8.021326044759718e-06, |
| "loss": 0.7043851852416992, |
| "step": 3660 |
| }, |
| { |
| "epoch": 0.5860864357720331, |
| "grad_norm": 0.033354733139276505, |
| "learning_rate": 7.969701966049447e-06, |
| "loss": 0.6736726284027099, |
| "step": 3670 |
| }, |
| { |
| "epoch": 0.5876834015370795, |
| "grad_norm": 0.04251250997185707, |
| "learning_rate": 7.918134260786214e-06, |
| "loss": 0.679281759262085, |
| "step": 3680 |
| }, |
| { |
| "epoch": 0.589280367302126, |
| "grad_norm": 0.03284266218543053, |
| "learning_rate": 7.866624360803786e-06, |
| "loss": 0.6851751804351807, |
| "step": 3690 |
| }, |
| { |
| "epoch": 0.5908773330671724, |
| "grad_norm": 0.050126783549785614, |
| "learning_rate": 7.81517369633092e-06, |
| "loss": 0.6820148468017578, |
| "step": 3700 |
| }, |
| { |
| "epoch": 0.5924742988322188, |
| "grad_norm": 0.042198847979307175, |
| "learning_rate": 7.763783695951623e-06, |
| "loss": 0.6824104309082031, |
| "step": 3710 |
| }, |
| { |
| "epoch": 0.5940712645972652, |
| "grad_norm": 0.03443041443824768, |
| "learning_rate": 7.7124557865655e-06, |
| "loss": 0.670203685760498, |
| "step": 3720 |
| }, |
| { |
| "epoch": 0.5956682303623116, |
| "grad_norm": 0.036490943282842636, |
| "learning_rate": 7.661191393348136e-06, |
| "loss": 0.7197192192077637, |
| "step": 3730 |
| }, |
| { |
| "epoch": 0.5972651961273581, |
| "grad_norm": 0.03636472299695015, |
| "learning_rate": 7.609991939711517e-06, |
| "loss": 0.6888130187988282, |
| "step": 3740 |
| }, |
| { |
| "epoch": 0.5988621618924044, |
| "grad_norm": 0.040862370282411575, |
| "learning_rate": 7.558858847264502e-06, |
| "loss": 0.6817411422729492, |
| "step": 3750 |
| }, |
| { |
| "epoch": 0.6004591276574508, |
| "grad_norm": 0.036771345883607864, |
| "learning_rate": 7.507793535773377e-06, |
| "loss": 0.7020951271057129, |
| "step": 3760 |
| }, |
| { |
| "epoch": 0.6020560934224972, |
| "grad_norm": 0.036549679934978485, |
| "learning_rate": 7.456797423122401e-06, |
| "loss": 0.6926814079284668, |
| "step": 3770 |
| }, |
| { |
| "epoch": 0.6036530591875436, |
| "grad_norm": 0.03661128878593445, |
| "learning_rate": 7.405871925274457e-06, |
| "loss": 0.7070647716522217, |
| "step": 3780 |
| }, |
| { |
| "epoch": 0.6052500249525901, |
| "grad_norm": 0.05932987853884697, |
| "learning_rate": 7.3550184562317295e-06, |
| "loss": 0.7125431060791015, |
| "step": 3790 |
| }, |
| { |
| "epoch": 0.6068469907176365, |
| "grad_norm": 0.03799287602305412, |
| "learning_rate": 7.3042384279964394e-06, |
| "loss": 0.6607150554656982, |
| "step": 3800 |
| }, |
| { |
| "epoch": 0.6084439564826829, |
| "grad_norm": 0.031429167836904526, |
| "learning_rate": 7.253533250531656e-06, |
| "loss": 0.6806385517120361, |
| "step": 3810 |
| }, |
| { |
| "epoch": 0.6100409222477293, |
| "grad_norm": 0.05243317410349846, |
| "learning_rate": 7.202904331722127e-06, |
| "loss": 0.6834061145782471, |
| "step": 3820 |
| }, |
| { |
| "epoch": 0.6116378880127757, |
| "grad_norm": 0.0379512719810009, |
| "learning_rate": 7.152353077335189e-06, |
| "loss": 0.665333604812622, |
| "step": 3830 |
| }, |
| { |
| "epoch": 0.6132348537778222, |
| "grad_norm": 0.03895451873540878, |
| "learning_rate": 7.101880890981751e-06, |
| "loss": 0.6566737651824951, |
| "step": 3840 |
| }, |
| { |
| "epoch": 0.6148318195428686, |
| "grad_norm": 0.03916362673044205, |
| "learning_rate": 7.0514891740773105e-06, |
| "loss": 0.7027400970458985, |
| "step": 3850 |
| }, |
| { |
| "epoch": 0.616428785307915, |
| "grad_norm": 0.03747180104255676, |
| "learning_rate": 7.001179325803037e-06, |
| "loss": 0.6675007820129395, |
| "step": 3860 |
| }, |
| { |
| "epoch": 0.6180257510729614, |
| "grad_norm": 0.04694506898522377, |
| "learning_rate": 6.95095274306693e-06, |
| "loss": 0.6993120193481446, |
| "step": 3870 |
| }, |
| { |
| "epoch": 0.6196227168380077, |
| "grad_norm": 0.035580579191446304, |
| "learning_rate": 6.900810820465036e-06, |
| "loss": 0.7056044578552246, |
| "step": 3880 |
| }, |
| { |
| "epoch": 0.6212196826030542, |
| "grad_norm": 0.03560471907258034, |
| "learning_rate": 6.8507549502427105e-06, |
| "loss": 0.7067145824432373, |
| "step": 3890 |
| }, |
| { |
| "epoch": 0.6228166483681006, |
| "grad_norm": 0.03395050764083862, |
| "learning_rate": 6.800786522255974e-06, |
| "loss": 0.6799113750457764, |
| "step": 3900 |
| }, |
| { |
| "epoch": 0.624413614133147, |
| "grad_norm": 0.04874083772301674, |
| "learning_rate": 6.7509069239329295e-06, |
| "loss": 0.6687004089355468, |
| "step": 3910 |
| }, |
| { |
| "epoch": 0.6260105798981934, |
| "grad_norm": 0.03185804560780525, |
| "learning_rate": 6.701117540235204e-06, |
| "loss": 0.6309501171112061, |
| "step": 3920 |
| }, |
| { |
| "epoch": 0.6276075456632398, |
| "grad_norm": 0.04211097210645676, |
| "learning_rate": 6.651419753619535e-06, |
| "loss": 0.7135450839996338, |
| "step": 3930 |
| }, |
| { |
| "epoch": 0.6292045114282863, |
| "grad_norm": 0.03911832720041275, |
| "learning_rate": 6.601814943999363e-06, |
| "loss": 0.7103962421417236, |
| "step": 3940 |
| }, |
| { |
| "epoch": 0.6308014771933327, |
| "grad_norm": 0.033491455018520355, |
| "learning_rate": 6.552304488706512e-06, |
| "loss": 0.7135549068450928, |
| "step": 3950 |
| }, |
| { |
| "epoch": 0.6323984429583791, |
| "grad_norm": 0.031158218160271645, |
| "learning_rate": 6.502889762452969e-06, |
| "loss": 0.6667858600616455, |
| "step": 3960 |
| }, |
| { |
| "epoch": 0.6339954087234255, |
| "grad_norm": 0.037267692387104034, |
| "learning_rate": 6.453572137292689e-06, |
| "loss": 0.7077006816864013, |
| "step": 3970 |
| }, |
| { |
| "epoch": 0.635592374488472, |
| "grad_norm": 0.03646966442465782, |
| "learning_rate": 6.404352982583511e-06, |
| "loss": 0.663360595703125, |
| "step": 3980 |
| }, |
| { |
| "epoch": 0.6371893402535184, |
| "grad_norm": 0.0357484444975853, |
| "learning_rate": 6.355233664949132e-06, |
| "loss": 0.6508517742156983, |
| "step": 3990 |
| }, |
| { |
| "epoch": 0.6387863060185647, |
| "grad_norm": 0.03508693352341652, |
| "learning_rate": 6.306215548241174e-06, |
| "loss": 0.669168758392334, |
| "step": 4000 |
| }, |
| { |
| "epoch": 0.6387863060185647, |
| "eval_loss": 0.6750220060348511, |
| "eval_runtime": 1753.5267, |
| "eval_samples_per_second": 3.174, |
| "eval_steps_per_second": 1.587, |
| "step": 4000 |
| }, |
| { |
| "epoch": 0.6403832717836111, |
| "grad_norm": 0.03497602418065071, |
| "learning_rate": 6.257299993501289e-06, |
| "loss": 0.6791082859039307, |
| "step": 4010 |
| }, |
| { |
| "epoch": 0.6419802375486575, |
| "grad_norm": 0.04195614159107208, |
| "learning_rate": 6.208488358923393e-06, |
| "loss": 0.6831603050231934, |
| "step": 4020 |
| }, |
| { |
| "epoch": 0.6435772033137039, |
| "grad_norm": 0.049627698957920074, |
| "learning_rate": 6.1597819998159506e-06, |
| "loss": 0.7127910137176514, |
| "step": 4030 |
| }, |
| { |
| "epoch": 0.6451741690787504, |
| "grad_norm": 0.03807014226913452, |
| "learning_rate": 6.111182268564322e-06, |
| "loss": 0.6687079429626465, |
| "step": 4040 |
| }, |
| { |
| "epoch": 0.6467711348437968, |
| "grad_norm": 0.0430082343518734, |
| "learning_rate": 6.062690514593237e-06, |
| "loss": 0.6742999076843261, |
| "step": 4050 |
| }, |
| { |
| "epoch": 0.6483681006088432, |
| "grad_norm": 0.032011304050683975, |
| "learning_rate": 6.014308084329327e-06, |
| "loss": 0.693959903717041, |
| "step": 4060 |
| }, |
| { |
| "epoch": 0.6499650663738896, |
| "grad_norm": 0.0382765531539917, |
| "learning_rate": 5.966036321163709e-06, |
| "loss": 0.6869026184082031, |
| "step": 4070 |
| }, |
| { |
| "epoch": 0.651562032138936, |
| "grad_norm": 0.038456711918115616, |
| "learning_rate": 5.917876565414729e-06, |
| "loss": 0.6802726268768311, |
| "step": 4080 |
| }, |
| { |
| "epoch": 0.6531589979039825, |
| "grad_norm": 0.05505223199725151, |
| "learning_rate": 5.8698301542907144e-06, |
| "loss": 0.6540178775787353, |
| "step": 4090 |
| }, |
| { |
| "epoch": 0.6547559636690289, |
| "grad_norm": 0.06022993102669716, |
| "learning_rate": 5.8218984218528514e-06, |
| "loss": 0.6731237888336181, |
| "step": 4100 |
| }, |
| { |
| "epoch": 0.6563529294340753, |
| "grad_norm": 0.04032977297902107, |
| "learning_rate": 5.774082698978154e-06, |
| "loss": 0.7120064258575439, |
| "step": 4110 |
| }, |
| { |
| "epoch": 0.6579498951991216, |
| "grad_norm": 0.04511059820652008, |
| "learning_rate": 5.726384313322504e-06, |
| "loss": 0.6774604797363282, |
| "step": 4120 |
| }, |
| { |
| "epoch": 0.659546860964168, |
| "grad_norm": 0.04290354251861572, |
| "learning_rate": 5.678804589283777e-06, |
| "loss": 0.6660148143768311, |
| "step": 4130 |
| }, |
| { |
| "epoch": 0.6611438267292145, |
| "grad_norm": 0.03994971513748169, |
| "learning_rate": 5.6313448479650944e-06, |
| "loss": 0.6876091957092285, |
| "step": 4140 |
| }, |
| { |
| "epoch": 0.6627407924942609, |
| "grad_norm": 0.03699030727148056, |
| "learning_rate": 5.584006407138114e-06, |
| "loss": 0.6590198516845703, |
| "step": 4150 |
| }, |
| { |
| "epoch": 0.6643377582593073, |
| "grad_norm": 0.041105858981609344, |
| "learning_rate": 5.536790581206451e-06, |
| "loss": 0.6922510147094727, |
| "step": 4160 |
| }, |
| { |
| "epoch": 0.6659347240243537, |
| "grad_norm": 0.036437924951314926, |
| "learning_rate": 5.489698681169196e-06, |
| "loss": 0.7056612968444824, |
| "step": 4170 |
| }, |
| { |
| "epoch": 0.6675316897894001, |
| "grad_norm": 0.03376461938023567, |
| "learning_rate": 5.442732014584495e-06, |
| "loss": 0.6548415184020996, |
| "step": 4180 |
| }, |
| { |
| "epoch": 0.6691286555544466, |
| "grad_norm": 0.042223311960697174, |
| "learning_rate": 5.395891885533244e-06, |
| "loss": 0.6739361763000489, |
| "step": 4190 |
| }, |
| { |
| "epoch": 0.670725621319493, |
| "grad_norm": 0.03633170202374458, |
| "learning_rate": 5.349179594582898e-06, |
| "loss": 0.7088988304138184, |
| "step": 4200 |
| }, |
| { |
| "epoch": 0.6723225870845394, |
| "grad_norm": 0.037932299077510834, |
| "learning_rate": 5.302596438751339e-06, |
| "loss": 0.693552827835083, |
| "step": 4210 |
| }, |
| { |
| "epoch": 0.6739195528495858, |
| "grad_norm": 0.03990362584590912, |
| "learning_rate": 5.256143711470869e-06, |
| "loss": 0.6921378135681152, |
| "step": 4220 |
| }, |
| { |
| "epoch": 0.6755165186146322, |
| "grad_norm": 0.037905026227235794, |
| "learning_rate": 5.20982270255231e-06, |
| "loss": 0.6850985050201416, |
| "step": 4230 |
| }, |
| { |
| "epoch": 0.6771134843796787, |
| "grad_norm": 0.0379810705780983, |
| "learning_rate": 5.163634698149163e-06, |
| "loss": 0.6650650978088379, |
| "step": 4240 |
| }, |
| { |
| "epoch": 0.678710450144725, |
| "grad_norm": 0.02918444201350212, |
| "learning_rate": 5.11758098072193e-06, |
| "loss": 0.6373987674713135, |
| "step": 4250 |
| }, |
| { |
| "epoch": 0.6803074159097714, |
| "grad_norm": 0.03856576979160309, |
| "learning_rate": 5.071662829002477e-06, |
| "loss": 0.7125391483306884, |
| "step": 4260 |
| }, |
| { |
| "epoch": 0.6819043816748178, |
| "grad_norm": 0.03352683037519455, |
| "learning_rate": 5.02588151795854e-06, |
| "loss": 0.6727629661560058, |
| "step": 4270 |
| }, |
| { |
| "epoch": 0.6835013474398642, |
| "grad_norm": 0.04891199618577957, |
| "learning_rate": 4.98023831875833e-06, |
| "loss": 0.722879695892334, |
| "step": 4280 |
| }, |
| { |
| "epoch": 0.6850983132049107, |
| "grad_norm": 0.04167689383029938, |
| "learning_rate": 4.9347344987352305e-06, |
| "loss": 0.6740260124206543, |
| "step": 4290 |
| }, |
| { |
| "epoch": 0.6866952789699571, |
| "grad_norm": 0.03634532913565636, |
| "learning_rate": 4.889371321352607e-06, |
| "loss": 0.649868106842041, |
| "step": 4300 |
| }, |
| { |
| "epoch": 0.6882922447350035, |
| "grad_norm": 0.04563901200890541, |
| "learning_rate": 4.844150046168722e-06, |
| "loss": 0.6745593547821045, |
| "step": 4310 |
| }, |
| { |
| "epoch": 0.6898892105000499, |
| "grad_norm": 0.04015154391527176, |
| "learning_rate": 4.799071928801784e-06, |
| "loss": 0.7058079719543457, |
| "step": 4320 |
| }, |
| { |
| "epoch": 0.6914861762650963, |
| "grad_norm": 0.03595186024904251, |
| "learning_rate": 4.7541382208950505e-06, |
| "loss": 0.6532906532287598, |
| "step": 4330 |
| }, |
| { |
| "epoch": 0.6930831420301428, |
| "grad_norm": 0.030780136585235596, |
| "learning_rate": 4.709350170082103e-06, |
| "loss": 0.6842369079589844, |
| "step": 4340 |
| }, |
| { |
| "epoch": 0.6946801077951892, |
| "grad_norm": 0.04692146182060242, |
| "learning_rate": 4.6647090199522025e-06, |
| "loss": 0.7000434398651123, |
| "step": 4350 |
| }, |
| { |
| "epoch": 0.6962770735602356, |
| "grad_norm": 0.0310053788125515, |
| "learning_rate": 4.620216010015725e-06, |
| "loss": 0.6727443218231202, |
| "step": 4360 |
| }, |
| { |
| "epoch": 0.6978740393252819, |
| "grad_norm": 0.03317731246352196, |
| "learning_rate": 4.575872375669793e-06, |
| "loss": 0.6653543949127197, |
| "step": 4370 |
| }, |
| { |
| "epoch": 0.6994710050903283, |
| "grad_norm": 0.05431032553315163, |
| "learning_rate": 4.531679348163955e-06, |
| "loss": 0.691961145401001, |
| "step": 4380 |
| }, |
| { |
| "epoch": 0.7010679708553748, |
| "grad_norm": 0.037718210369348526, |
| "learning_rate": 4.487638154565978e-06, |
| "loss": 0.6595858573913574, |
| "step": 4390 |
| }, |
| { |
| "epoch": 0.7026649366204212, |
| "grad_norm": 0.03125704079866409, |
| "learning_rate": 4.443750017727814e-06, |
| "loss": 0.7066734790802002, |
| "step": 4400 |
| }, |
| { |
| "epoch": 0.7042619023854676, |
| "grad_norm": 0.050476983189582825, |
| "learning_rate": 4.400016156251611e-06, |
| "loss": 0.6906495094299316, |
| "step": 4410 |
| }, |
| { |
| "epoch": 0.705858868150514, |
| "grad_norm": 0.034239206463098526, |
| "learning_rate": 4.356437784455896e-06, |
| "loss": 0.698778486251831, |
| "step": 4420 |
| }, |
| { |
| "epoch": 0.7074558339155604, |
| "grad_norm": 0.03537076339125633, |
| "learning_rate": 4.313016112341861e-06, |
| "loss": 0.6730549335479736, |
| "step": 4430 |
| }, |
| { |
| "epoch": 0.7090527996806069, |
| "grad_norm": 0.0366840697824955, |
| "learning_rate": 4.269752345559761e-06, |
| "loss": 0.6689012527465821, |
| "step": 4440 |
| }, |
| { |
| "epoch": 0.7106497654456533, |
| "grad_norm": 0.04745423048734665, |
| "learning_rate": 4.226647685375428e-06, |
| "loss": 0.6668778419494629, |
| "step": 4450 |
| }, |
| { |
| "epoch": 0.7122467312106997, |
| "grad_norm": 0.043418820947408676, |
| "learning_rate": 4.183703328636924e-06, |
| "loss": 0.6641871929168701, |
| "step": 4460 |
| }, |
| { |
| "epoch": 0.7138436969757461, |
| "grad_norm": 0.0369136743247509, |
| "learning_rate": 4.140920467741325e-06, |
| "loss": 0.6592461585998535, |
| "step": 4470 |
| }, |
| { |
| "epoch": 0.7154406627407925, |
| "grad_norm": 0.034368593245744705, |
| "learning_rate": 4.098300290601581e-06, |
| "loss": 0.6683683395385742, |
| "step": 4480 |
| }, |
| { |
| "epoch": 0.7170376285058389, |
| "grad_norm": 0.042594823986291885, |
| "learning_rate": 4.055843980613561e-06, |
| "loss": 0.6916409015655518, |
| "step": 4490 |
| }, |
| { |
| "epoch": 0.7186345942708853, |
| "grad_norm": 0.042385537177324295, |
| "learning_rate": 4.013552716623185e-06, |
| "loss": 0.6435957431793213, |
| "step": 4500 |
| }, |
| { |
| "epoch": 0.7186345942708853, |
| "eval_loss": 0.6736165881156921, |
| "eval_runtime": 1722.8539, |
| "eval_samples_per_second": 3.231, |
| "eval_steps_per_second": 1.615, |
| "step": 4500 |
| }, |
| { |
| "epoch": 0.7202315600359317, |
| "grad_norm": 0.04700292646884918, |
| "learning_rate": 3.971427672893672e-06, |
| "loss": 0.688172435760498, |
| "step": 4510 |
| }, |
| { |
| "epoch": 0.7218285258009781, |
| "grad_norm": 0.040195874869823456, |
| "learning_rate": 3.929470019072972e-06, |
| "loss": 0.6703804016113282, |
| "step": 4520 |
| }, |
| { |
| "epoch": 0.7234254915660245, |
| "grad_norm": 0.036519281566143036, |
| "learning_rate": 3.8876809201612695e-06, |
| "loss": 0.6743984222412109, |
| "step": 4530 |
| }, |
| { |
| "epoch": 0.725022457331071, |
| "grad_norm": 0.03687750920653343, |
| "learning_rate": 3.846061536478626e-06, |
| "loss": 0.6919309139251709, |
| "step": 4540 |
| }, |
| { |
| "epoch": 0.7266194230961174, |
| "grad_norm": 0.034661829471588135, |
| "learning_rate": 3.804613023632788e-06, |
| "loss": 0.6852948665618896, |
| "step": 4550 |
| }, |
| { |
| "epoch": 0.7282163888611638, |
| "grad_norm": 0.04090991988778114, |
| "learning_rate": 3.763336532487076e-06, |
| "loss": 0.6639271259307862, |
| "step": 4560 |
| }, |
| { |
| "epoch": 0.7298133546262102, |
| "grad_norm": 0.053740449249744415, |
| "learning_rate": 3.7222332091284384e-06, |
| "loss": 0.7086254119873047, |
| "step": 4570 |
| }, |
| { |
| "epoch": 0.7314103203912566, |
| "grad_norm": 0.05317610129714012, |
| "learning_rate": 3.6813041948356408e-06, |
| "loss": 0.688640546798706, |
| "step": 4580 |
| }, |
| { |
| "epoch": 0.7330072861563031, |
| "grad_norm": 0.04328848421573639, |
| "learning_rate": 3.640550626047562e-06, |
| "loss": 0.6996944904327392, |
| "step": 4590 |
| }, |
| { |
| "epoch": 0.7346042519213495, |
| "grad_norm": 0.03426145389676094, |
| "learning_rate": 3.5999736343316406e-06, |
| "loss": 0.7036723613739013, |
| "step": 4600 |
| }, |
| { |
| "epoch": 0.7362012176863958, |
| "grad_norm": 0.04445657134056091, |
| "learning_rate": 3.559574346352459e-06, |
| "loss": 0.6727498054504395, |
| "step": 4610 |
| }, |
| { |
| "epoch": 0.7377981834514422, |
| "grad_norm": 0.034136440604925156, |
| "learning_rate": 3.5193538838404716e-06, |
| "loss": 0.6709301471710205, |
| "step": 4620 |
| }, |
| { |
| "epoch": 0.7393951492164886, |
| "grad_norm": 0.03670594096183777, |
| "learning_rate": 3.4793133635608334e-06, |
| "loss": 0.7098387241363525, |
| "step": 4630 |
| }, |
| { |
| "epoch": 0.740992114981535, |
| "grad_norm": 0.04140735790133476, |
| "learning_rate": 3.4394538972824167e-06, |
| "loss": 0.6899067401885987, |
| "step": 4640 |
| }, |
| { |
| "epoch": 0.7425890807465815, |
| "grad_norm": 0.03637846186757088, |
| "learning_rate": 3.3997765917469348e-06, |
| "loss": 0.6745570659637451, |
| "step": 4650 |
| }, |
| { |
| "epoch": 0.7441860465116279, |
| "grad_norm": 0.03964189440011978, |
| "learning_rate": 3.3602825486381886e-06, |
| "loss": 0.6715593338012695, |
| "step": 4660 |
| }, |
| { |
| "epoch": 0.7457830122766743, |
| "grad_norm": 0.036266524344682693, |
| "learning_rate": 3.3209728645515204e-06, |
| "loss": 0.6992376804351806, |
| "step": 4670 |
| }, |
| { |
| "epoch": 0.7473799780417207, |
| "grad_norm": 0.050657473504543304, |
| "learning_rate": 3.281848630963327e-06, |
| "loss": 0.689194917678833, |
| "step": 4680 |
| }, |
| { |
| "epoch": 0.7489769438067672, |
| "grad_norm": 0.03800279274582863, |
| "learning_rate": 3.2429109342007747e-06, |
| "loss": 0.7068089962005615, |
| "step": 4690 |
| }, |
| { |
| "epoch": 0.7505739095718136, |
| "grad_norm": 0.034397054463624954, |
| "learning_rate": 3.204160855411638e-06, |
| "loss": 0.7055879592895508, |
| "step": 4700 |
| }, |
| { |
| "epoch": 0.75217087533686, |
| "grad_norm": 0.03663492202758789, |
| "learning_rate": 3.1655994705342595e-06, |
| "loss": 0.6516903400421142, |
| "step": 4710 |
| }, |
| { |
| "epoch": 0.7537678411019064, |
| "grad_norm": 0.10431419312953949, |
| "learning_rate": 3.1272278502676932e-06, |
| "loss": 0.6833277702331543, |
| "step": 4720 |
| }, |
| { |
| "epoch": 0.7553648068669528, |
| "grad_norm": 0.04803668335080147, |
| "learning_rate": 3.0890470600419785e-06, |
| "loss": 0.7252533435821533, |
| "step": 4730 |
| }, |
| { |
| "epoch": 0.7569617726319992, |
| "grad_norm": 0.052451301366090775, |
| "learning_rate": 3.051058159988539e-06, |
| "loss": 0.6383802413940429, |
| "step": 4740 |
| }, |
| { |
| "epoch": 0.7585587383970456, |
| "grad_norm": 0.04214588552713394, |
| "learning_rate": 3.0132622049107672e-06, |
| "loss": 0.6385763645172119, |
| "step": 4750 |
| }, |
| { |
| "epoch": 0.760155704162092, |
| "grad_norm": 0.04353920742869377, |
| "learning_rate": 2.975660244254721e-06, |
| "loss": 0.6771360874176026, |
| "step": 4760 |
| }, |
| { |
| "epoch": 0.7617526699271384, |
| "grad_norm": 0.034283965826034546, |
| "learning_rate": 2.9382533220799893e-06, |
| "loss": 0.6694541931152344, |
| "step": 4770 |
| }, |
| { |
| "epoch": 0.7633496356921848, |
| "grad_norm": 0.04156842082738876, |
| "learning_rate": 2.901042477030709e-06, |
| "loss": 0.719163179397583, |
| "step": 4780 |
| }, |
| { |
| "epoch": 0.7649466014572313, |
| "grad_norm": 0.038868360221385956, |
| "learning_rate": 2.864028742306725e-06, |
| "loss": 0.6658177852630616, |
| "step": 4790 |
| }, |
| { |
| "epoch": 0.7665435672222777, |
| "grad_norm": 0.03701276332139969, |
| "learning_rate": 2.827213145634887e-06, |
| "loss": 0.6825634956359863, |
| "step": 4800 |
| }, |
| { |
| "epoch": 0.7681405329873241, |
| "grad_norm": 0.043009012937545776, |
| "learning_rate": 2.7905967092405305e-06, |
| "loss": 0.6813971519470214, |
| "step": 4810 |
| }, |
| { |
| "epoch": 0.7697374987523705, |
| "grad_norm": 0.03427146002650261, |
| "learning_rate": 2.754180449819092e-06, |
| "loss": 0.6715961933135987, |
| "step": 4820 |
| }, |
| { |
| "epoch": 0.7713344645174169, |
| "grad_norm": 0.04296226054430008, |
| "learning_rate": 2.717965378507864e-06, |
| "loss": 0.6857268810272217, |
| "step": 4830 |
| }, |
| { |
| "epoch": 0.7729314302824634, |
| "grad_norm": 0.0429498665034771, |
| "learning_rate": 2.681952500857944e-06, |
| "loss": 0.6708025932312012, |
| "step": 4840 |
| }, |
| { |
| "epoch": 0.7745283960475098, |
| "grad_norm": 0.03906711935997009, |
| "learning_rate": 2.6461428168062973e-06, |
| "loss": 0.6925381660461426, |
| "step": 4850 |
| }, |
| { |
| "epoch": 0.7761253618125561, |
| "grad_norm": 0.04651563614606857, |
| "learning_rate": 2.6105373206479913e-06, |
| "loss": 0.6757769107818603, |
| "step": 4860 |
| }, |
| { |
| "epoch": 0.7777223275776025, |
| "grad_norm": 0.04317731410264969, |
| "learning_rate": 2.575137001008592e-06, |
| "loss": 0.665839433670044, |
| "step": 4870 |
| }, |
| { |
| "epoch": 0.7793192933426489, |
| "grad_norm": 0.03744199126958847, |
| "learning_rate": 2.539942840816727e-06, |
| "loss": 0.7045305252075196, |
| "step": 4880 |
| }, |
| { |
| "epoch": 0.7809162591076954, |
| "grad_norm": 0.04245394840836525, |
| "learning_rate": 2.5049558172767675e-06, |
| "loss": 0.6957573890686035, |
| "step": 4890 |
| }, |
| { |
| "epoch": 0.7825132248727418, |
| "grad_norm": 0.03942755609750748, |
| "learning_rate": 2.47017690184172e-06, |
| "loss": 0.6565013408660889, |
| "step": 4900 |
| }, |
| { |
| "epoch": 0.7841101906377882, |
| "grad_norm": 0.04654339328408241, |
| "learning_rate": 2.4356070601862327e-06, |
| "loss": 0.6680932998657226, |
| "step": 4910 |
| }, |
| { |
| "epoch": 0.7857071564028346, |
| "grad_norm": 0.05248269438743591, |
| "learning_rate": 2.4012472521797923e-06, |
| "loss": 0.6607920169830322, |
| "step": 4920 |
| }, |
| { |
| "epoch": 0.787304122167881, |
| "grad_norm": 0.04624265059828758, |
| "learning_rate": 2.3670984318600786e-06, |
| "loss": 0.6683568954467773, |
| "step": 4930 |
| }, |
| { |
| "epoch": 0.7889010879329275, |
| "grad_norm": 0.04104055091738701, |
| "learning_rate": 2.333161547406464e-06, |
| "loss": 0.6837501049041748, |
| "step": 4940 |
| }, |
| { |
| "epoch": 0.7904980536979739, |
| "grad_norm": 0.039029769599437714, |
| "learning_rate": 2.2994375411136884e-06, |
| "loss": 0.6702389240264892, |
| "step": 4950 |
| }, |
| { |
| "epoch": 0.7920950194630203, |
| "grad_norm": 0.03849470987915993, |
| "learning_rate": 2.265927349365691e-06, |
| "loss": 0.6521383285522461, |
| "step": 4960 |
| }, |
| { |
| "epoch": 0.7936919852280667, |
| "grad_norm": 0.05400224030017853, |
| "learning_rate": 2.232631902609629e-06, |
| "loss": 0.6533032894134522, |
| "step": 4970 |
| }, |
| { |
| "epoch": 0.795288950993113, |
| "grad_norm": 0.043887361884117126, |
| "learning_rate": 2.1995521253300167e-06, |
| "loss": 0.6794767379760742, |
| "step": 4980 |
| }, |
| { |
| "epoch": 0.7968859167581595, |
| "grad_norm": 0.04501941800117493, |
| "learning_rate": 2.166688936023078e-06, |
| "loss": 0.6698923587799073, |
| "step": 4990 |
| }, |
| { |
| "epoch": 0.7984828825232059, |
| "grad_norm": 0.045139558613300323, |
| "learning_rate": 2.1340432471712367e-06, |
| "loss": 0.661639404296875, |
| "step": 5000 |
| }, |
| { |
| "epoch": 0.7984828825232059, |
| "eval_loss": 0.6726189255714417, |
| "eval_runtime": 1709.319, |
| "eval_samples_per_second": 3.256, |
| "eval_steps_per_second": 1.628, |
| "step": 5000 |
| }, |
| { |
| "epoch": 0.8000798482882523, |
| "grad_norm": 0.04251272603869438, |
| "learning_rate": 2.101615965217768e-06, |
| "loss": 0.7006937503814697, |
| "step": 5010 |
| }, |
| { |
| "epoch": 0.8016768140532987, |
| "grad_norm": 0.040616679936647415, |
| "learning_rate": 2.0694079905416475e-06, |
| "loss": 0.6587623119354248, |
| "step": 5020 |
| }, |
| { |
| "epoch": 0.8032737798183451, |
| "grad_norm": 0.04379381984472275, |
| "learning_rate": 2.0374202174325443e-06, |
| "loss": 0.6809040069580078, |
| "step": 5030 |
| }, |
| { |
| "epoch": 0.8048707455833916, |
| "grad_norm": 0.042080674320459366, |
| "learning_rate": 2.0056535340659867e-06, |
| "loss": 0.6683236598968506, |
| "step": 5040 |
| }, |
| { |
| "epoch": 0.806467711348438, |
| "grad_norm": 0.038132891058921814, |
| "learning_rate": 1.97410882247871e-06, |
| "loss": 0.6637454032897949, |
| "step": 5050 |
| }, |
| { |
| "epoch": 0.8080646771134844, |
| "grad_norm": 0.029925232753157616, |
| "learning_rate": 1.942786958544154e-06, |
| "loss": 0.6917686939239502, |
| "step": 5060 |
| }, |
| { |
| "epoch": 0.8096616428785308, |
| "grad_norm": 0.041264407336711884, |
| "learning_rate": 1.91168881194815e-06, |
| "loss": 0.7052114486694336, |
| "step": 5070 |
| }, |
| { |
| "epoch": 0.8112586086435772, |
| "grad_norm": 0.03905145451426506, |
| "learning_rate": 1.8808152461647766e-06, |
| "loss": 0.667028284072876, |
| "step": 5080 |
| }, |
| { |
| "epoch": 0.8128555744086237, |
| "grad_norm": 0.047729961574077606, |
| "learning_rate": 1.8501671184323811e-06, |
| "loss": 0.691020917892456, |
| "step": 5090 |
| }, |
| { |
| "epoch": 0.8144525401736701, |
| "grad_norm": 0.034812819212675095, |
| "learning_rate": 1.8197452797297699e-06, |
| "loss": 0.7124451637268067, |
| "step": 5100 |
| }, |
| { |
| "epoch": 0.8160495059387164, |
| "grad_norm": 0.044796060770750046, |
| "learning_rate": 1.7895505747525933e-06, |
| "loss": 0.7617973804473877, |
| "step": 5110 |
| }, |
| { |
| "epoch": 0.8176464717037628, |
| "grad_norm": 0.039166904985904694, |
| "learning_rate": 1.759583841889878e-06, |
| "loss": 0.711750841140747, |
| "step": 5120 |
| }, |
| { |
| "epoch": 0.8192434374688092, |
| "grad_norm": 0.03871138393878937, |
| "learning_rate": 1.7298459132007628e-06, |
| "loss": 0.6736746788024902, |
| "step": 5130 |
| }, |
| { |
| "epoch": 0.8208404032338557, |
| "grad_norm": 0.046535786241292953, |
| "learning_rate": 1.7003376143913886e-06, |
| "loss": 0.7094714164733886, |
| "step": 5140 |
| }, |
| { |
| "epoch": 0.8224373689989021, |
| "grad_norm": 0.04319629445672035, |
| "learning_rate": 1.6710597647919613e-06, |
| "loss": 0.6693609237670899, |
| "step": 5150 |
| }, |
| { |
| "epoch": 0.8240343347639485, |
| "grad_norm": 0.04204268380999565, |
| "learning_rate": 1.6420131773340263e-06, |
| "loss": 0.6906074047088623, |
| "step": 5160 |
| }, |
| { |
| "epoch": 0.8256313005289949, |
| "grad_norm": 0.041262589395046234, |
| "learning_rate": 1.6131986585278714e-06, |
| "loss": 0.6685310363769531, |
| "step": 5170 |
| }, |
| { |
| "epoch": 0.8272282662940413, |
| "grad_norm": 0.03515459597110748, |
| "learning_rate": 1.5846170084401458e-06, |
| "loss": 0.6744651317596435, |
| "step": 5180 |
| }, |
| { |
| "epoch": 0.8288252320590878, |
| "grad_norm": 0.044764138758182526, |
| "learning_rate": 1.5562690206716479e-06, |
| "loss": 0.6720765113830567, |
| "step": 5190 |
| }, |
| { |
| "epoch": 0.8304221978241342, |
| "grad_norm": 0.03811672702431679, |
| "learning_rate": 1.5281554823352896e-06, |
| "loss": 0.6512242317199707, |
| "step": 5200 |
| }, |
| { |
| "epoch": 0.8320191635891806, |
| "grad_norm": 0.03850555047392845, |
| "learning_rate": 1.500277174034227e-06, |
| "loss": 0.6925489425659179, |
| "step": 5210 |
| }, |
| { |
| "epoch": 0.833616129354227, |
| "grad_norm": 0.039821285754442215, |
| "learning_rate": 1.4726348698402015e-06, |
| "loss": 0.6698274612426758, |
| "step": 5220 |
| }, |
| { |
| "epoch": 0.8352130951192733, |
| "grad_norm": 0.042530301958322525, |
| "learning_rate": 1.4452293372720483e-06, |
| "loss": 0.657553243637085, |
| "step": 5230 |
| }, |
| { |
| "epoch": 0.8368100608843198, |
| "grad_norm": 0.03285897895693779, |
| "learning_rate": 1.4180613372743678e-06, |
| "loss": 0.6513785362243653, |
| "step": 5240 |
| }, |
| { |
| "epoch": 0.8384070266493662, |
| "grad_norm": 0.0418255515396595, |
| "learning_rate": 1.39113162419642e-06, |
| "loss": 0.6803596019744873, |
| "step": 5250 |
| }, |
| { |
| "epoch": 0.8400039924144126, |
| "grad_norm": 0.03861439600586891, |
| "learning_rate": 1.3644409457711604e-06, |
| "loss": 0.7275567531585694, |
| "step": 5260 |
| }, |
| { |
| "epoch": 0.841600958179459, |
| "grad_norm": 0.035479992628097534, |
| "learning_rate": 1.3379900430944858e-06, |
| "loss": 0.7078885555267334, |
| "step": 5270 |
| }, |
| { |
| "epoch": 0.8431979239445054, |
| "grad_norm": 0.036275558173656464, |
| "learning_rate": 1.31177965060466e-06, |
| "loss": 0.6733812808990478, |
| "step": 5280 |
| }, |
| { |
| "epoch": 0.8447948897095519, |
| "grad_norm": 0.037248894572257996, |
| "learning_rate": 1.285810496061921e-06, |
| "loss": 0.6728686809539794, |
| "step": 5290 |
| }, |
| { |
| "epoch": 0.8463918554745983, |
| "grad_norm": 0.03526445850729942, |
| "learning_rate": 1.2600833005282631e-06, |
| "loss": 0.6712977409362793, |
| "step": 5300 |
| }, |
| { |
| "epoch": 0.8479888212396447, |
| "grad_norm": 0.03622904419898987, |
| "learning_rate": 1.2345987783474311e-06, |
| "loss": 0.6881800651550293, |
| "step": 5310 |
| }, |
| { |
| "epoch": 0.8495857870046911, |
| "grad_norm": 0.05652482062578201, |
| "learning_rate": 1.2093576371250737e-06, |
| "loss": 0.709586763381958, |
| "step": 5320 |
| }, |
| { |
| "epoch": 0.8511827527697375, |
| "grad_norm": 0.05090609937906265, |
| "learning_rate": 1.1843605777091028e-06, |
| "loss": 0.7400435447692871, |
| "step": 5330 |
| }, |
| { |
| "epoch": 0.852779718534784, |
| "grad_norm": 0.038989149034023285, |
| "learning_rate": 1.1596082941702313e-06, |
| "loss": 0.695259952545166, |
| "step": 5340 |
| }, |
| { |
| "epoch": 0.8543766842998303, |
| "grad_norm": 0.04729596897959709, |
| "learning_rate": 1.135101473782706e-06, |
| "loss": 0.681358528137207, |
| "step": 5350 |
| }, |
| { |
| "epoch": 0.8559736500648767, |
| "grad_norm": 0.0379403792321682, |
| "learning_rate": 1.1108407970052149e-06, |
| "loss": 0.6542623043060303, |
| "step": 5360 |
| }, |
| { |
| "epoch": 0.8575706158299231, |
| "grad_norm": 0.044038884341716766, |
| "learning_rate": 1.086826937462e-06, |
| "loss": 0.6564295291900635, |
| "step": 5370 |
| }, |
| { |
| "epoch": 0.8591675815949695, |
| "grad_norm": 0.04039445519447327, |
| "learning_rate": 1.0630605619241564e-06, |
| "loss": 0.6783324241638183, |
| "step": 5380 |
| }, |
| { |
| "epoch": 0.860764547360016, |
| "grad_norm": 0.034845978021621704, |
| "learning_rate": 1.0395423302911112e-06, |
| "loss": 0.6549112796783447, |
| "step": 5390 |
| }, |
| { |
| "epoch": 0.8623615131250624, |
| "grad_norm": 0.048201173543930054, |
| "learning_rate": 1.016272895572309e-06, |
| "loss": 0.6641885757446289, |
| "step": 5400 |
| }, |
| { |
| "epoch": 0.8639584788901088, |
| "grad_norm": 0.037778809666633606, |
| "learning_rate": 9.932529038690696e-07, |
| "loss": 0.7034367561340332, |
| "step": 5410 |
| }, |
| { |
| "epoch": 0.8655554446551552, |
| "grad_norm": 0.0427137166261673, |
| "learning_rate": 9.704829943566574e-07, |
| "loss": 0.6777073383331299, |
| "step": 5420 |
| }, |
| { |
| "epoch": 0.8671524104202016, |
| "grad_norm": 0.037193115800619125, |
| "learning_rate": 9.479637992665313e-07, |
| "loss": 0.6748072147369385, |
| "step": 5430 |
| }, |
| { |
| "epoch": 0.8687493761852481, |
| "grad_norm": 0.0337526798248291, |
| "learning_rate": 9.256959438687929e-07, |
| "loss": 0.6617064952850342, |
| "step": 5440 |
| }, |
| { |
| "epoch": 0.8703463419502945, |
| "grad_norm": 0.039619434624910355, |
| "learning_rate": 9.036800464548157e-07, |
| "loss": 0.6612413883209228, |
| "step": 5450 |
| }, |
| { |
| "epoch": 0.8719433077153409, |
| "grad_norm": 0.03671206533908844, |
| "learning_rate": 8.819167183200905e-07, |
| "loss": 0.6929868698120117, |
| "step": 5460 |
| }, |
| { |
| "epoch": 0.8735402734803873, |
| "grad_norm": 0.034324243664741516, |
| "learning_rate": 8.604065637472392e-07, |
| "loss": 0.685936689376831, |
| "step": 5470 |
| }, |
| { |
| "epoch": 0.8751372392454336, |
| "grad_norm": 0.035913076251745224, |
| "learning_rate": 8.391501799892466e-07, |
| "loss": 0.7082894802093506, |
| "step": 5480 |
| }, |
| { |
| "epoch": 0.87673420501048, |
| "grad_norm": 0.04239356890320778, |
| "learning_rate": 8.181481572528726e-07, |
| "loss": 0.7099148750305175, |
| "step": 5490 |
| }, |
| { |
| "epoch": 0.8783311707755265, |
| "grad_norm": 0.04144662618637085, |
| "learning_rate": 7.974010786822684e-07, |
| "loss": 0.6640007972717286, |
| "step": 5500 |
| }, |
| { |
| "epoch": 0.8783311707755265, |
| "eval_loss": 0.672135055065155, |
| "eval_runtime": 1721.4336, |
| "eval_samples_per_second": 3.233, |
| "eval_steps_per_second": 1.617, |
| "step": 5500 |
| }, |
| { |
| "epoch": 0.8799281365405729, |
| "grad_norm": 0.030981754884123802, |
| "learning_rate": 7.769095203427734e-07, |
| "loss": 0.6549492359161377, |
| "step": 5510 |
| }, |
| { |
| "epoch": 0.8815251023056193, |
| "grad_norm": 0.04086959362030029, |
| "learning_rate": 7.56674051204932e-07, |
| "loss": 0.6875146389007568, |
| "step": 5520 |
| }, |
| { |
| "epoch": 0.8831220680706657, |
| "grad_norm": 0.0377834253013134, |
| "learning_rate": 7.366952331286958e-07, |
| "loss": 0.7493057250976562, |
| "step": 5530 |
| }, |
| { |
| "epoch": 0.8847190338357122, |
| "grad_norm": 0.03834075480699539, |
| "learning_rate": 7.169736208478118e-07, |
| "loss": 0.6828866481781006, |
| "step": 5540 |
| }, |
| { |
| "epoch": 0.8863159996007586, |
| "grad_norm": 0.03951543942093849, |
| "learning_rate": 6.975097619544324e-07, |
| "loss": 0.68464674949646, |
| "step": 5550 |
| }, |
| { |
| "epoch": 0.887912965365805, |
| "grad_norm": 0.044246163219213486, |
| "learning_rate": 6.78304196883901e-07, |
| "loss": 0.6737260341644287, |
| "step": 5560 |
| }, |
| { |
| "epoch": 0.8895099311308514, |
| "grad_norm": 0.04266819730401039, |
| "learning_rate": 6.59357458899752e-07, |
| "loss": 0.7164577960968017, |
| "step": 5570 |
| }, |
| { |
| "epoch": 0.8911068968958978, |
| "grad_norm": 0.0311746709048748, |
| "learning_rate": 6.40670074078904e-07, |
| "loss": 0.680039358139038, |
| "step": 5580 |
| }, |
| { |
| "epoch": 0.8927038626609443, |
| "grad_norm": 0.035248082131147385, |
| "learning_rate": 6.222425612970462e-07, |
| "loss": 0.6851954936981202, |
| "step": 5590 |
| }, |
| { |
| "epoch": 0.8943008284259906, |
| "grad_norm": 0.03986518457531929, |
| "learning_rate": 6.040754322142417e-07, |
| "loss": 0.6760111331939698, |
| "step": 5600 |
| }, |
| { |
| "epoch": 0.895897794191037, |
| "grad_norm": 0.03682507202029228, |
| "learning_rate": 5.861691912607159e-07, |
| "loss": 0.6883102416992187, |
| "step": 5610 |
| }, |
| { |
| "epoch": 0.8974947599560834, |
| "grad_norm": 0.03000503219664097, |
| "learning_rate": 5.685243356228409e-07, |
| "loss": 0.6542916774749756, |
| "step": 5620 |
| }, |
| { |
| "epoch": 0.8990917257211298, |
| "grad_norm": 0.044614020735025406, |
| "learning_rate": 5.511413552293476e-07, |
| "loss": 0.6680744647979736, |
| "step": 5630 |
| }, |
| { |
| "epoch": 0.9006886914861763, |
| "grad_norm": 0.03846392408013344, |
| "learning_rate": 5.340207327377156e-07, |
| "loss": 0.7028224468231201, |
| "step": 5640 |
| }, |
| { |
| "epoch": 0.9022856572512227, |
| "grad_norm": 0.038470469415187836, |
| "learning_rate": 5.171629435207615e-07, |
| "loss": 0.6939930438995361, |
| "step": 5650 |
| }, |
| { |
| "epoch": 0.9038826230162691, |
| "grad_norm": 0.03757188841700554, |
| "learning_rate": 5.005684556534574e-07, |
| "loss": 0.656277084350586, |
| "step": 5660 |
| }, |
| { |
| "epoch": 0.9054795887813155, |
| "grad_norm": 0.04011685401201248, |
| "learning_rate": 4.842377298999191e-07, |
| "loss": 0.638321828842163, |
| "step": 5670 |
| }, |
| { |
| "epoch": 0.9070765545463619, |
| "grad_norm": 0.039341870695352554, |
| "learning_rate": 4.681712197006205e-07, |
| "loss": 0.6716189384460449, |
| "step": 5680 |
| }, |
| { |
| "epoch": 0.9086735203114084, |
| "grad_norm": 0.046330180019140244, |
| "learning_rate": 4.5236937115980204e-07, |
| "loss": 0.6909313678741456, |
| "step": 5690 |
| }, |
| { |
| "epoch": 0.9102704860764548, |
| "grad_norm": 0.041009679436683655, |
| "learning_rate": 4.368326230330855e-07, |
| "loss": 0.6272893428802491, |
| "step": 5700 |
| }, |
| { |
| "epoch": 0.9118674518415012, |
| "grad_norm": 0.03455239534378052, |
| "learning_rate": 4.215614067152851e-07, |
| "loss": 0.6927034854888916, |
| "step": 5710 |
| }, |
| { |
| "epoch": 0.9134644176065475, |
| "grad_norm": 0.03742032125592232, |
| "learning_rate": 4.065561462284351e-07, |
| "loss": 0.664287805557251, |
| "step": 5720 |
| }, |
| { |
| "epoch": 0.9150613833715939, |
| "grad_norm": 0.03536292910575867, |
| "learning_rate": 3.918172582100177e-07, |
| "loss": 0.664261770248413, |
| "step": 5730 |
| }, |
| { |
| "epoch": 0.9166583491366403, |
| "grad_norm": 0.044113483279943466, |
| "learning_rate": 3.7734515190138934e-07, |
| "loss": 0.7117574691772461, |
| "step": 5740 |
| }, |
| { |
| "epoch": 0.9182553149016868, |
| "grad_norm": 0.04183537885546684, |
| "learning_rate": 3.631402291364206e-07, |
| "loss": 0.6857941627502442, |
| "step": 5750 |
| }, |
| { |
| "epoch": 0.9198522806667332, |
| "grad_norm": 0.02898181416094303, |
| "learning_rate": 3.492028843303419e-07, |
| "loss": 0.6839983463287354, |
| "step": 5760 |
| }, |
| { |
| "epoch": 0.9214492464317796, |
| "grad_norm": 0.040342144668102264, |
| "learning_rate": 3.3553350446878244e-07, |
| "loss": 0.6938966751098633, |
| "step": 5770 |
| }, |
| { |
| "epoch": 0.923046212196826, |
| "grad_norm": 0.05028856173157692, |
| "learning_rate": 3.221324690970351e-07, |
| "loss": 0.6682254791259765, |
| "step": 5780 |
| }, |
| { |
| "epoch": 0.9246431779618725, |
| "grad_norm": 0.033745042979717255, |
| "learning_rate": 3.090001503095175e-07, |
| "loss": 0.6571940422058106, |
| "step": 5790 |
| }, |
| { |
| "epoch": 0.9262401437269189, |
| "grad_norm": 0.036207910627126694, |
| "learning_rate": 2.9613691273942977e-07, |
| "loss": 0.7331238269805909, |
| "step": 5800 |
| }, |
| { |
| "epoch": 0.9278371094919653, |
| "grad_norm": 0.04358496144413948, |
| "learning_rate": 2.835431135486455e-07, |
| "loss": 0.6768048763275146, |
| "step": 5810 |
| }, |
| { |
| "epoch": 0.9294340752570117, |
| "grad_norm": 0.051529187709093094, |
| "learning_rate": 2.7121910241778014e-07, |
| "loss": 0.6914044857025147, |
| "step": 5820 |
| }, |
| { |
| "epoch": 0.9310310410220581, |
| "grad_norm": 0.03387345373630524, |
| "learning_rate": 2.591652215364904e-07, |
| "loss": 0.67225022315979, |
| "step": 5830 |
| }, |
| { |
| "epoch": 0.9326280067871044, |
| "grad_norm": 0.043374814093112946, |
| "learning_rate": 2.473818055939747e-07, |
| "loss": 0.6870133876800537, |
| "step": 5840 |
| }, |
| { |
| "epoch": 0.9342249725521509, |
| "grad_norm": 0.04162263497710228, |
| "learning_rate": 2.358691817696712e-07, |
| "loss": 0.6274053573608398, |
| "step": 5850 |
| }, |
| { |
| "epoch": 0.9358219383171973, |
| "grad_norm": 0.03655523061752319, |
| "learning_rate": 2.2462766972418226e-07, |
| "loss": 0.6559162139892578, |
| "step": 5860 |
| }, |
| { |
| "epoch": 0.9374189040822437, |
| "grad_norm": 0.036134812980890274, |
| "learning_rate": 2.1365758159039073e-07, |
| "loss": 0.6783801078796386, |
| "step": 5870 |
| }, |
| { |
| "epoch": 0.9390158698472901, |
| "grad_norm": 0.03370088338851929, |
| "learning_rate": 2.0295922196480355e-07, |
| "loss": 0.679423189163208, |
| "step": 5880 |
| }, |
| { |
| "epoch": 0.9406128356123366, |
| "grad_norm": 0.044190868735313416, |
| "learning_rate": 1.9253288789908287e-07, |
| "loss": 0.6669793128967285, |
| "step": 5890 |
| }, |
| { |
| "epoch": 0.942209801377383, |
| "grad_norm": 0.03765581548213959, |
| "learning_rate": 1.823788688918049e-07, |
| "loss": 0.6280838966369628, |
| "step": 5900 |
| }, |
| { |
| "epoch": 0.9438067671424294, |
| "grad_norm": 0.05461205169558525, |
| "learning_rate": 1.7249744688042414e-07, |
| "loss": 0.6410412788391113, |
| "step": 5910 |
| }, |
| { |
| "epoch": 0.9454037329074758, |
| "grad_norm": 0.046921294182538986, |
| "learning_rate": 1.6288889623343295e-07, |
| "loss": 0.6527444839477539, |
| "step": 5920 |
| }, |
| { |
| "epoch": 0.9470006986725222, |
| "grad_norm": 0.03945396840572357, |
| "learning_rate": 1.535534837427588e-07, |
| "loss": 0.694933271408081, |
| "step": 5930 |
| }, |
| { |
| "epoch": 0.9485976644375687, |
| "grad_norm": 0.03962003439664841, |
| "learning_rate": 1.44491468616349e-07, |
| "loss": 0.6714414596557617, |
| "step": 5940 |
| }, |
| { |
| "epoch": 0.9501946302026151, |
| "grad_norm": 0.07099556922912598, |
| "learning_rate": 1.3570310247096764e-07, |
| "loss": 0.7129894733428955, |
| "step": 5950 |
| }, |
| { |
| "epoch": 0.9517915959676615, |
| "grad_norm": 0.03700670972466469, |
| "learning_rate": 1.2718862932522224e-07, |
| "loss": 0.6442203044891357, |
| "step": 5960 |
| }, |
| { |
| "epoch": 0.9533885617327078, |
| "grad_norm": 0.04066568240523338, |
| "learning_rate": 1.1894828559277816e-07, |
| "loss": 0.6612192153930664, |
| "step": 5970 |
| }, |
| { |
| "epoch": 0.9549855274977542, |
| "grad_norm": 0.04404400289058685, |
| "learning_rate": 1.1098230007579702e-07, |
| "loss": 0.6766380310058594, |
| "step": 5980 |
| }, |
| { |
| "epoch": 0.9565824932628006, |
| "grad_norm": 0.04611429572105408, |
| "learning_rate": 1.0329089395858527e-07, |
| "loss": 0.6801462650299073, |
| "step": 5990 |
| }, |
| { |
| "epoch": 0.9581794590278471, |
| "grad_norm": 0.03326278179883957, |
| "learning_rate": 9.587428080145344e-08, |
| "loss": 0.6539367198944092, |
| "step": 6000 |
| }, |
| { |
| "epoch": 0.9581794590278471, |
| "eval_loss": 0.6719915270805359, |
| "eval_runtime": 1743.0303, |
| "eval_samples_per_second": 3.193, |
| "eval_steps_per_second": 1.597, |
| "step": 6000 |
| } |
| ], |
| "logging_steps": 10, |
| "max_steps": 6262, |
| "num_input_tokens_seen": 0, |
| "num_train_epochs": 1, |
| "save_steps": 1000, |
| "stateful_callbacks": { |
| "TrainerControl": { |
| "args": { |
| "should_epoch_stop": false, |
| "should_evaluate": false, |
| "should_log": false, |
| "should_save": true, |
| "should_training_stop": false |
| }, |
| "attributes": {} |
| } |
| }, |
| "total_flos": 1.3174132430797578e+19, |
| "train_batch_size": 2, |
| "trial_name": null, |
| "trial_params": null |
| } |
|
|