| { | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 3.0, | |
| "global_step": 101625, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.0, | |
| "learning_rate": 3.0831802676462875e-07, | |
| "loss": 4.3168, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 6.363159275780635e-07, | |
| "loss": 4.155, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 9.643138283914982e-07, | |
| "loss": 3.7804, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 1.2890317501967988e-06, | |
| "loss": 3.8055, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "learning_rate": 1.6170296510102337e-06, | |
| "loss": 3.6927, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 1.9450275518236685e-06, | |
| "loss": 3.5158, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 2.273025452637103e-06, | |
| "loss": 3.6069, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 2.6010233534505377e-06, | |
| "loss": 3.6429, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 2.929021254263973e-06, | |
| "loss": 3.6292, | |
| "step": 900 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 3.257019155077408e-06, | |
| "loss": 3.4969, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 3.585017055890842e-06, | |
| "loss": 3.5082, | |
| "step": 1100 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 3.9130149567042775e-06, | |
| "loss": 3.5132, | |
| "step": 1200 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 4.2410128575177125e-06, | |
| "loss": 3.5413, | |
| "step": 1300 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 4.569010758331147e-06, | |
| "loss": 3.4286, | |
| "step": 1400 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "learning_rate": 4.897008659144582e-06, | |
| "loss": 3.4228, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 5.225006559958016e-06, | |
| "loss": 3.4743, | |
| "step": 1600 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 5.553004460771451e-06, | |
| "loss": 3.4929, | |
| "step": 1700 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 5.881002361584886e-06, | |
| "loss": 3.4324, | |
| "step": 1800 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 6.20900026239832e-06, | |
| "loss": 3.4384, | |
| "step": 1900 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 6.536998163211756e-06, | |
| "loss": 3.488, | |
| "step": 2000 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 6.86499606402519e-06, | |
| "loss": 3.3966, | |
| "step": 2100 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 7.1929939648386246e-06, | |
| "loss": 3.4662, | |
| "step": 2200 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 7.5209918656520605e-06, | |
| "loss": 3.5339, | |
| "step": 2300 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 7.848989766465495e-06, | |
| "loss": 3.3541, | |
| "step": 2400 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "learning_rate": 8.17698766727893e-06, | |
| "loss": 3.3954, | |
| "step": 2500 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 8.504985568092365e-06, | |
| "loss": 3.3898, | |
| "step": 2600 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 8.8329834689058e-06, | |
| "loss": 3.431, | |
| "step": 2700 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 9.160981369719235e-06, | |
| "loss": 3.4336, | |
| "step": 2800 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 9.48897927053267e-06, | |
| "loss": 3.4402, | |
| "step": 2900 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 9.816977171346103e-06, | |
| "loss": 3.3912, | |
| "step": 3000 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 1.0144975072159538e-05, | |
| "loss": 3.3787, | |
| "step": 3100 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 1.0472972972972973e-05, | |
| "loss": 3.3998, | |
| "step": 3200 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 1.0800970873786408e-05, | |
| "loss": 3.4397, | |
| "step": 3300 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 1.1125688795591708e-05, | |
| "loss": 3.313, | |
| "step": 3400 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "learning_rate": 1.1453686696405144e-05, | |
| "loss": 3.3708, | |
| "step": 3500 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 1.1781684597218578e-05, | |
| "loss": 3.3924, | |
| "step": 3600 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 1.2109682498032013e-05, | |
| "loss": 3.3792, | |
| "step": 3700 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 1.2437680398845448e-05, | |
| "loss": 3.3953, | |
| "step": 3800 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 1.2765678299658885e-05, | |
| "loss": 3.4286, | |
| "step": 3900 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 1.3093676200472318e-05, | |
| "loss": 3.3751, | |
| "step": 4000 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 1.3421674101285753e-05, | |
| "loss": 3.4209, | |
| "step": 4100 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 1.3749672002099186e-05, | |
| "loss": 3.3963, | |
| "step": 4200 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 1.4077669902912621e-05, | |
| "loss": 3.4004, | |
| "step": 4300 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 1.4405667803726055e-05, | |
| "loss": 3.3787, | |
| "step": 4400 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "learning_rate": 1.4733665704539493e-05, | |
| "loss": 3.3521, | |
| "step": 4500 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 1.5061663605352927e-05, | |
| "loss": 3.3486, | |
| "step": 4600 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 1.538966150616636e-05, | |
| "loss": 3.3643, | |
| "step": 4700 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 1.5717659406979797e-05, | |
| "loss": 3.4276, | |
| "step": 4800 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 1.604565730779323e-05, | |
| "loss": 3.3843, | |
| "step": 4900 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 1.6373655208606667e-05, | |
| "loss": 3.3808, | |
| "step": 5000 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 1.67016531094201e-05, | |
| "loss": 3.4145, | |
| "step": 5100 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "learning_rate": 1.7029651010233537e-05, | |
| "loss": 3.4281, | |
| "step": 5200 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 1.735764891104697e-05, | |
| "loss": 3.3493, | |
| "step": 5300 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 1.7685646811860404e-05, | |
| "loss": 3.3366, | |
| "step": 5400 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 1.801364471267384e-05, | |
| "loss": 3.3874, | |
| "step": 5500 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 1.8341642613487277e-05, | |
| "loss": 3.3963, | |
| "step": 5600 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 1.866964051430071e-05, | |
| "loss": 3.4015, | |
| "step": 5700 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 1.8997638415114144e-05, | |
| "loss": 3.4283, | |
| "step": 5800 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 1.9325636315927577e-05, | |
| "loss": 3.3884, | |
| "step": 5900 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 1.9653634216741014e-05, | |
| "loss": 3.3508, | |
| "step": 6000 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 1.998163211755445e-05, | |
| "loss": 3.3748, | |
| "step": 6100 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "learning_rate": 2.0309630018367884e-05, | |
| "loss": 3.3285, | |
| "step": 6200 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 2.0637627919181317e-05, | |
| "loss": 3.3309, | |
| "step": 6300 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 2.0965625819994754e-05, | |
| "loss": 3.3261, | |
| "step": 6400 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 2.1293623720808187e-05, | |
| "loss": 3.3698, | |
| "step": 6500 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 2.1621621621621624e-05, | |
| "loss": 3.4237, | |
| "step": 6600 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 2.1949619522435057e-05, | |
| "loss": 3.3434, | |
| "step": 6700 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 2.227761742324849e-05, | |
| "loss": 3.3825, | |
| "step": 6800 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 2.2605615324061928e-05, | |
| "loss": 3.4024, | |
| "step": 6900 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 2.293361322487536e-05, | |
| "loss": 3.4105, | |
| "step": 7000 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 2.3261611125688798e-05, | |
| "loss": 3.3623, | |
| "step": 7100 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "learning_rate": 2.358960902650223e-05, | |
| "loss": 3.3796, | |
| "step": 7200 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 2.3917606927315668e-05, | |
| "loss": 3.4282, | |
| "step": 7300 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 2.42456048281291e-05, | |
| "loss": 3.3423, | |
| "step": 7400 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 2.4573602728942534e-05, | |
| "loss": 3.3964, | |
| "step": 7500 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 2.490160062975597e-05, | |
| "loss": 3.3467, | |
| "step": 7600 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 2.5229598530569405e-05, | |
| "loss": 3.3749, | |
| "step": 7700 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 2.555759643138284e-05, | |
| "loss": 3.3759, | |
| "step": 7800 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 2.5885594332196278e-05, | |
| "loss": 3.3565, | |
| "step": 7900 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 2.6213592233009708e-05, | |
| "loss": 3.3313, | |
| "step": 8000 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 2.6541590133823148e-05, | |
| "loss": 3.4114, | |
| "step": 8100 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "learning_rate": 2.6869588034636578e-05, | |
| "loss": 3.4161, | |
| "step": 8200 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 2.7197585935450015e-05, | |
| "loss": 3.4302, | |
| "step": 8300 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 2.7525583836263448e-05, | |
| "loss": 3.443, | |
| "step": 8400 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 2.7853581737076885e-05, | |
| "loss": 3.358, | |
| "step": 8500 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 2.818157963789032e-05, | |
| "loss": 3.381, | |
| "step": 8600 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 2.850957753870375e-05, | |
| "loss": 3.3726, | |
| "step": 8700 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 2.883757543951719e-05, | |
| "loss": 3.494, | |
| "step": 8800 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 2.9165573340330622e-05, | |
| "loss": 3.4925, | |
| "step": 8900 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 2.949357124114406e-05, | |
| "loss": 3.4443, | |
| "step": 9000 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 2.9821569141957495e-05, | |
| "loss": 3.476, | |
| "step": 9100 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 3.0149567042770925e-05, | |
| "loss": 3.4143, | |
| "step": 9200 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "learning_rate": 3.0474284964576228e-05, | |
| "loss": 3.3816, | |
| "step": 9300 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 3.080228286538966e-05, | |
| "loss": 3.493, | |
| "step": 9400 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 3.1130280766203095e-05, | |
| "loss": 3.4608, | |
| "step": 9500 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 3.1458278667016535e-05, | |
| "loss": 3.3971, | |
| "step": 9600 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 3.178627656782997e-05, | |
| "loss": 3.4157, | |
| "step": 9700 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 3.21142744686434e-05, | |
| "loss": 3.3921, | |
| "step": 9800 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "learning_rate": 3.244227236945684e-05, | |
| "loss": 3.3785, | |
| "step": 9900 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 3.277027027027027e-05, | |
| "loss": 3.3634, | |
| "step": 10000 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 3.309826817108371e-05, | |
| "loss": 3.3881, | |
| "step": 10100 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 3.342626607189714e-05, | |
| "loss": 3.362, | |
| "step": 10200 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 3.3754263972710575e-05, | |
| "loss": 3.4395, | |
| "step": 10300 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 3.4082261873524015e-05, | |
| "loss": 3.4151, | |
| "step": 10400 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 3.441025977433744e-05, | |
| "loss": 3.3813, | |
| "step": 10500 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 3.473825767515088e-05, | |
| "loss": 3.3624, | |
| "step": 10600 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 3.5066255575964315e-05, | |
| "loss": 3.5004, | |
| "step": 10700 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 3.539425347677775e-05, | |
| "loss": 3.373, | |
| "step": 10800 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 3.572225137759119e-05, | |
| "loss": 3.4307, | |
| "step": 10900 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "learning_rate": 3.605024927840462e-05, | |
| "loss": 3.451, | |
| "step": 11000 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 3.6378247179218055e-05, | |
| "loss": 3.3964, | |
| "step": 11100 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 3.670624508003149e-05, | |
| "loss": 3.3537, | |
| "step": 11200 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 3.703424298084492e-05, | |
| "loss": 3.469, | |
| "step": 11300 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 3.736224088165836e-05, | |
| "loss": 3.4307, | |
| "step": 11400 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 3.7690238782471796e-05, | |
| "loss": 3.4911, | |
| "step": 11500 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 3.801823668328523e-05, | |
| "loss": 3.3976, | |
| "step": 11600 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 3.834623458409866e-05, | |
| "loss": 3.4258, | |
| "step": 11700 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 3.8674232484912096e-05, | |
| "loss": 3.4217, | |
| "step": 11800 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 3.9002230385725536e-05, | |
| "loss": 3.4396, | |
| "step": 11900 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "learning_rate": 3.933022828653897e-05, | |
| "loss": 3.4841, | |
| "step": 12000 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 3.96582261873524e-05, | |
| "loss": 3.4864, | |
| "step": 12100 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 3.9986224088165836e-05, | |
| "loss": 3.4933, | |
| "step": 12200 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 4.0314221988979276e-05, | |
| "loss": 3.4383, | |
| "step": 12300 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 4.064221988979271e-05, | |
| "loss": 3.3793, | |
| "step": 12400 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 4.097021779060614e-05, | |
| "loss": 3.4841, | |
| "step": 12500 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 4.1298215691419576e-05, | |
| "loss": 3.482, | |
| "step": 12600 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 4.162621359223301e-05, | |
| "loss": 3.4459, | |
| "step": 12700 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 4.195421149304645e-05, | |
| "loss": 3.4549, | |
| "step": 12800 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 4.228220939385988e-05, | |
| "loss": 3.5482, | |
| "step": 12900 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "learning_rate": 4.2610207294673316e-05, | |
| "loss": 3.4926, | |
| "step": 13000 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 4.293820519548675e-05, | |
| "loss": 3.5195, | |
| "step": 13100 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 4.326620309630018e-05, | |
| "loss": 3.5106, | |
| "step": 13200 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 4.359420099711362e-05, | |
| "loss": 3.4536, | |
| "step": 13300 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 4.392219889792705e-05, | |
| "loss": 3.4985, | |
| "step": 13400 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 4.425019679874049e-05, | |
| "loss": 3.4308, | |
| "step": 13500 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 4.457491472054579e-05, | |
| "loss": 3.4666, | |
| "step": 13600 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 4.4902912621359226e-05, | |
| "loss": 3.5624, | |
| "step": 13700 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 4.523091052217266e-05, | |
| "loss": 3.4941, | |
| "step": 13800 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 4.55589084229861e-05, | |
| "loss": 3.4589, | |
| "step": 13900 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "learning_rate": 4.5886906323799526e-05, | |
| "loss": 3.5033, | |
| "step": 14000 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 4.6214904224612966e-05, | |
| "loss": 3.4525, | |
| "step": 14100 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 4.65429021254264e-05, | |
| "loss": 3.5153, | |
| "step": 14200 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 4.687090002623983e-05, | |
| "loss": 3.5456, | |
| "step": 14300 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 4.719889792705327e-05, | |
| "loss": 3.4878, | |
| "step": 14400 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 4.75268958278667e-05, | |
| "loss": 3.4686, | |
| "step": 14500 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 4.785489372868014e-05, | |
| "loss": 3.4891, | |
| "step": 14600 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "learning_rate": 4.818289162949357e-05, | |
| "loss": 3.4157, | |
| "step": 14700 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 4.8510889530307006e-05, | |
| "loss": 3.444, | |
| "step": 14800 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 4.8838887431120446e-05, | |
| "loss": 3.5387, | |
| "step": 14900 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 4.916688533193387e-05, | |
| "loss": 3.5091, | |
| "step": 15000 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 4.949488323274731e-05, | |
| "loss": 3.5356, | |
| "step": 15100 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 4.9822881133560746e-05, | |
| "loss": 3.4978, | |
| "step": 15200 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 4.997337377432538e-05, | |
| "loss": 3.4566, | |
| "step": 15300 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 4.991549067503271e-05, | |
| "loss": 3.5653, | |
| "step": 15400 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 4.985760757574004e-05, | |
| "loss": 3.5342, | |
| "step": 15500 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 4.979972447644737e-05, | |
| "loss": 3.5189, | |
| "step": 15600 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "learning_rate": 4.97418413771547e-05, | |
| "loss": 3.5025, | |
| "step": 15700 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 4.968395827786203e-05, | |
| "loss": 3.5216, | |
| "step": 15800 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 4.9626075178569366e-05, | |
| "loss": 3.5428, | |
| "step": 15900 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 4.9568192079276696e-05, | |
| "loss": 3.5112, | |
| "step": 16000 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 4.9510308979984026e-05, | |
| "loss": 3.4538, | |
| "step": 16100 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 4.9452425880691356e-05, | |
| "loss": 3.5476, | |
| "step": 16200 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 4.9394542781398686e-05, | |
| "loss": 3.5039, | |
| "step": 16300 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 4.933665968210602e-05, | |
| "loss": 3.4226, | |
| "step": 16400 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 4.927877658281335e-05, | |
| "loss": 3.5409, | |
| "step": 16500 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 4.922089348352068e-05, | |
| "loss": 3.4823, | |
| "step": 16600 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "learning_rate": 4.916301038422801e-05, | |
| "loss": 3.5517, | |
| "step": 16700 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "learning_rate": 4.910512728493535e-05, | |
| "loss": 3.5905, | |
| "step": 16800 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "learning_rate": 4.9047244185642673e-05, | |
| "loss": 3.5193, | |
| "step": 16900 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "learning_rate": 4.898936108635001e-05, | |
| "loss": 3.5474, | |
| "step": 17000 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "learning_rate": 4.893147798705734e-05, | |
| "loss": 3.5083, | |
| "step": 17100 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "learning_rate": 4.887359488776467e-05, | |
| "loss": 3.5096, | |
| "step": 17200 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "learning_rate": 4.8815711788472e-05, | |
| "loss": 3.4946, | |
| "step": 17300 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "learning_rate": 4.875782868917934e-05, | |
| "loss": 3.5561, | |
| "step": 17400 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "learning_rate": 4.869994558988667e-05, | |
| "loss": 3.5484, | |
| "step": 17500 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "learning_rate": 4.864264132158692e-05, | |
| "loss": 3.473, | |
| "step": 17600 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "learning_rate": 4.858475822229426e-05, | |
| "loss": 3.4777, | |
| "step": 17700 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "learning_rate": 4.852687512300159e-05, | |
| "loss": 3.5096, | |
| "step": 17800 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "learning_rate": 4.846899202370892e-05, | |
| "loss": 3.5513, | |
| "step": 17900 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "learning_rate": 4.841110892441625e-05, | |
| "loss": 3.54, | |
| "step": 18000 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "learning_rate": 4.8353225825123584e-05, | |
| "loss": 3.4855, | |
| "step": 18100 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "learning_rate": 4.8295342725830915e-05, | |
| "loss": 3.4912, | |
| "step": 18200 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "learning_rate": 4.823803845753117e-05, | |
| "loss": 3.4644, | |
| "step": 18300 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "learning_rate": 4.8180155358238504e-05, | |
| "loss": 3.5219, | |
| "step": 18400 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "learning_rate": 4.8122272258945834e-05, | |
| "loss": 3.4587, | |
| "step": 18500 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "learning_rate": 4.8064389159653164e-05, | |
| "loss": 3.5399, | |
| "step": 18600 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "learning_rate": 4.8006506060360494e-05, | |
| "loss": 3.5016, | |
| "step": 18700 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "learning_rate": 4.794862296106783e-05, | |
| "loss": 3.4807, | |
| "step": 18800 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "learning_rate": 4.789073986177516e-05, | |
| "loss": 3.5076, | |
| "step": 18900 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "learning_rate": 4.783285676248249e-05, | |
| "loss": 3.4874, | |
| "step": 19000 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "learning_rate": 4.777555249418275e-05, | |
| "loss": 3.4773, | |
| "step": 19100 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "learning_rate": 4.771766939489008e-05, | |
| "loss": 3.5177, | |
| "step": 19200 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "learning_rate": 4.765978629559741e-05, | |
| "loss": 3.4338, | |
| "step": 19300 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "learning_rate": 4.760190319630475e-05, | |
| "loss": 3.4832, | |
| "step": 19400 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "learning_rate": 4.754402009701208e-05, | |
| "loss": 3.3966, | |
| "step": 19500 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "learning_rate": 4.748613699771941e-05, | |
| "loss": 3.5248, | |
| "step": 19600 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "learning_rate": 4.742825389842674e-05, | |
| "loss": 3.5104, | |
| "step": 19700 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "learning_rate": 4.7370370799134075e-05, | |
| "loss": 3.4758, | |
| "step": 19800 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "learning_rate": 4.73124876998414e-05, | |
| "loss": 3.468, | |
| "step": 19900 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "learning_rate": 4.7254604600548735e-05, | |
| "loss": 3.5482, | |
| "step": 20000 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "learning_rate": 4.7196721501256066e-05, | |
| "loss": 3.4968, | |
| "step": 20100 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "learning_rate": 4.7138838401963396e-05, | |
| "loss": 3.4743, | |
| "step": 20200 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "learning_rate": 4.7080955302670726e-05, | |
| "loss": 3.4734, | |
| "step": 20300 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "learning_rate": 4.702307220337806e-05, | |
| "loss": 3.4668, | |
| "step": 20400 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "learning_rate": 4.6965189104085386e-05, | |
| "loss": 3.4479, | |
| "step": 20500 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "learning_rate": 4.690730600479272e-05, | |
| "loss": 3.4609, | |
| "step": 20600 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "learning_rate": 4.684942290550006e-05, | |
| "loss": 3.5061, | |
| "step": 20700 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "learning_rate": 4.679153980620738e-05, | |
| "loss": 3.474, | |
| "step": 20800 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "learning_rate": 4.673365670691472e-05, | |
| "loss": 3.46, | |
| "step": 20900 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "learning_rate": 4.667577360762205e-05, | |
| "loss": 3.4596, | |
| "step": 21000 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "learning_rate": 4.661789050832938e-05, | |
| "loss": 3.5117, | |
| "step": 21100 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "learning_rate": 4.656000740903671e-05, | |
| "loss": 3.5576, | |
| "step": 21200 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "learning_rate": 4.650212430974405e-05, | |
| "loss": 3.5696, | |
| "step": 21300 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "learning_rate": 4.644424121045137e-05, | |
| "loss": 3.4833, | |
| "step": 21400 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "learning_rate": 4.638635811115871e-05, | |
| "loss": 3.4939, | |
| "step": 21500 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "learning_rate": 4.632905384285897e-05, | |
| "loss": 3.4656, | |
| "step": 21600 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "learning_rate": 4.62711707435663e-05, | |
| "loss": 3.4646, | |
| "step": 21700 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "learning_rate": 4.621328764427363e-05, | |
| "loss": 3.5053, | |
| "step": 21800 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "learning_rate": 4.615540454498096e-05, | |
| "loss": 3.426, | |
| "step": 21900 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "learning_rate": 4.6097521445688294e-05, | |
| "loss": 3.5011, | |
| "step": 22000 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "learning_rate": 4.603963834639562e-05, | |
| "loss": 3.5055, | |
| "step": 22100 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "learning_rate": 4.5981755247102954e-05, | |
| "loss": 3.5099, | |
| "step": 22200 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "learning_rate": 4.5923872147810284e-05, | |
| "loss": 3.4436, | |
| "step": 22300 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "learning_rate": 4.5865989048517614e-05, | |
| "loss": 3.4445, | |
| "step": 22400 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "learning_rate": 4.580810594922495e-05, | |
| "loss": 3.4688, | |
| "step": 22500 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "learning_rate": 4.575022284993228e-05, | |
| "loss": 3.4393, | |
| "step": 22600 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "learning_rate": 4.569233975063961e-05, | |
| "loss": 3.455, | |
| "step": 22700 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "learning_rate": 4.563445665134694e-05, | |
| "loss": 3.47, | |
| "step": 22800 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "learning_rate": 4.557657355205427e-05, | |
| "loss": 3.4016, | |
| "step": 22900 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "learning_rate": 4.55186904527616e-05, | |
| "loss": 3.4426, | |
| "step": 23000 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "learning_rate": 4.546080735346894e-05, | |
| "loss": 3.4976, | |
| "step": 23100 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "learning_rate": 4.540292425417627e-05, | |
| "loss": 3.4368, | |
| "step": 23200 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "learning_rate": 4.53450411548836e-05, | |
| "loss": 3.439, | |
| "step": 23300 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "learning_rate": 4.528715805559093e-05, | |
| "loss": 3.5062, | |
| "step": 23400 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "learning_rate": 4.522927495629826e-05, | |
| "loss": 3.4495, | |
| "step": 23500 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "learning_rate": 4.5171391857005596e-05, | |
| "loss": 3.4425, | |
| "step": 23600 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "learning_rate": 4.5113508757712926e-05, | |
| "loss": 3.4741, | |
| "step": 23700 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "learning_rate": 4.5055625658420256e-05, | |
| "loss": 3.4746, | |
| "step": 23800 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "learning_rate": 4.4997742559127586e-05, | |
| "loss": 3.3827, | |
| "step": 23900 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "learning_rate": 4.493985945983492e-05, | |
| "loss": 3.3663, | |
| "step": 24000 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "learning_rate": 4.4881976360542246e-05, | |
| "loss": 3.4349, | |
| "step": 24100 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "learning_rate": 4.482409326124958e-05, | |
| "loss": 3.4367, | |
| "step": 24200 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "learning_rate": 4.4766210161956913e-05, | |
| "loss": 3.3795, | |
| "step": 24300 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "learning_rate": 4.4708327062664244e-05, | |
| "loss": 3.4992, | |
| "step": 24400 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "learning_rate": 4.4650443963371574e-05, | |
| "loss": 3.4647, | |
| "step": 24500 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "learning_rate": 4.459256086407891e-05, | |
| "loss": 3.441, | |
| "step": 24600 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "learning_rate": 4.453467776478624e-05, | |
| "loss": 3.4718, | |
| "step": 24700 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "learning_rate": 4.447679466549357e-05, | |
| "loss": 3.4318, | |
| "step": 24800 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "learning_rate": 4.441891156620091e-05, | |
| "loss": 3.4683, | |
| "step": 24900 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "learning_rate": 4.436102846690823e-05, | |
| "loss": 3.4339, | |
| "step": 25000 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "learning_rate": 4.430314536761557e-05, | |
| "loss": 3.4285, | |
| "step": 25100 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "learning_rate": 4.42452622683229e-05, | |
| "loss": 3.3831, | |
| "step": 25200 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "learning_rate": 4.418737916903023e-05, | |
| "loss": 3.4561, | |
| "step": 25300 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "learning_rate": 4.412949606973756e-05, | |
| "loss": 3.4521, | |
| "step": 25400 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "learning_rate": 4.4071612970444895e-05, | |
| "loss": 3.4461, | |
| "step": 25500 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "learning_rate": 4.401372987115222e-05, | |
| "loss": 3.52, | |
| "step": 25600 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "learning_rate": 4.3955846771859555e-05, | |
| "loss": 3.4293, | |
| "step": 25700 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "learning_rate": 4.3897963672566885e-05, | |
| "loss": 3.3902, | |
| "step": 25800 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "learning_rate": 4.3840080573274215e-05, | |
| "loss": 3.4207, | |
| "step": 25900 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "learning_rate": 4.378219747398155e-05, | |
| "loss": 3.5207, | |
| "step": 26000 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "learning_rate": 4.372431437468888e-05, | |
| "loss": 3.3999, | |
| "step": 26100 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "learning_rate": 4.366643127539621e-05, | |
| "loss": 3.4113, | |
| "step": 26200 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "learning_rate": 4.360854817610354e-05, | |
| "loss": 3.4914, | |
| "step": 26300 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "learning_rate": 4.355066507681087e-05, | |
| "loss": 3.4771, | |
| "step": 26400 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "learning_rate": 4.349336080851113e-05, | |
| "loss": 3.4838, | |
| "step": 26500 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "learning_rate": 4.343547770921846e-05, | |
| "loss": 3.4307, | |
| "step": 26600 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "learning_rate": 4.33775946099258e-05, | |
| "loss": 3.539, | |
| "step": 26700 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "learning_rate": 4.331971151063313e-05, | |
| "loss": 3.4617, | |
| "step": 26800 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "learning_rate": 4.326182841134046e-05, | |
| "loss": 3.4023, | |
| "step": 26900 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "learning_rate": 4.320394531204779e-05, | |
| "loss": 3.4677, | |
| "step": 27000 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "learning_rate": 4.314606221275512e-05, | |
| "loss": 3.4065, | |
| "step": 27100 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "learning_rate": 4.308817911346245e-05, | |
| "loss": 3.4339, | |
| "step": 27200 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "learning_rate": 4.3030296014169787e-05, | |
| "loss": 3.4403, | |
| "step": 27300 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "learning_rate": 4.297241291487712e-05, | |
| "loss": 3.4626, | |
| "step": 27400 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "learning_rate": 4.291452981558445e-05, | |
| "loss": 3.4698, | |
| "step": 27500 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "learning_rate": 4.2856646716291784e-05, | |
| "loss": 3.4339, | |
| "step": 27600 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "learning_rate": 4.279876361699911e-05, | |
| "loss": 3.3525, | |
| "step": 27700 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "learning_rate": 4.2740880517706444e-05, | |
| "loss": 3.4385, | |
| "step": 27800 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "learning_rate": 4.2682997418413774e-05, | |
| "loss": 3.3924, | |
| "step": 27900 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "learning_rate": 4.2625114319121104e-05, | |
| "loss": 3.4542, | |
| "step": 28000 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "learning_rate": 4.2567810050821364e-05, | |
| "loss": 3.439, | |
| "step": 28100 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "learning_rate": 4.2509926951528694e-05, | |
| "loss": 3.4499, | |
| "step": 28200 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "learning_rate": 4.245204385223603e-05, | |
| "loss": 3.4655, | |
| "step": 28300 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "learning_rate": 4.2394160752943354e-05, | |
| "loss": 3.4605, | |
| "step": 28400 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "learning_rate": 4.233627765365069e-05, | |
| "loss": 3.413, | |
| "step": 28500 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "learning_rate": 4.227839455435802e-05, | |
| "loss": 3.4802, | |
| "step": 28600 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "learning_rate": 4.222051145506535e-05, | |
| "loss": 3.4359, | |
| "step": 28700 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "learning_rate": 4.216262835577268e-05, | |
| "loss": 3.3859, | |
| "step": 28800 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "learning_rate": 4.210474525648002e-05, | |
| "loss": 3.3875, | |
| "step": 28900 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "learning_rate": 4.204686215718734e-05, | |
| "loss": 3.3715, | |
| "step": 29000 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "learning_rate": 4.198897905789468e-05, | |
| "loss": 3.3754, | |
| "step": 29100 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "learning_rate": 4.193109595860201e-05, | |
| "loss": 3.3918, | |
| "step": 29200 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "learning_rate": 4.187321285930934e-05, | |
| "loss": 3.4793, | |
| "step": 29300 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "learning_rate": 4.1815329760016675e-05, | |
| "loss": 3.4157, | |
| "step": 29400 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "learning_rate": 4.1757446660724005e-05, | |
| "loss": 3.4226, | |
| "step": 29500 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "learning_rate": 4.1699563561431335e-05, | |
| "loss": 3.4078, | |
| "step": 29600 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "learning_rate": 4.1641680462138665e-05, | |
| "loss": 3.4142, | |
| "step": 29700 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "learning_rate": 4.1583797362846e-05, | |
| "loss": 3.4222, | |
| "step": 29800 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "learning_rate": 4.1525914263553326e-05, | |
| "loss": 3.4325, | |
| "step": 29900 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "learning_rate": 4.146803116426066e-05, | |
| "loss": 3.4555, | |
| "step": 30000 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "learning_rate": 4.141014806496799e-05, | |
| "loss": 3.432, | |
| "step": 30100 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "learning_rate": 4.135226496567532e-05, | |
| "loss": 3.3894, | |
| "step": 30200 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "learning_rate": 4.129438186638265e-05, | |
| "loss": 3.4094, | |
| "step": 30300 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "learning_rate": 4.123649876708999e-05, | |
| "loss": 3.4366, | |
| "step": 30400 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "learning_rate": 4.117861566779732e-05, | |
| "loss": 3.4109, | |
| "step": 30500 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "learning_rate": 4.112073256850465e-05, | |
| "loss": 3.4778, | |
| "step": 30600 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "learning_rate": 4.106284946921199e-05, | |
| "loss": 3.4626, | |
| "step": 30700 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "learning_rate": 4.100496636991931e-05, | |
| "loss": 3.3753, | |
| "step": 30800 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "learning_rate": 4.094708327062665e-05, | |
| "loss": 3.4158, | |
| "step": 30900 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "learning_rate": 4.088920017133398e-05, | |
| "loss": 3.4925, | |
| "step": 31000 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "learning_rate": 4.083131707204131e-05, | |
| "loss": 3.4383, | |
| "step": 31100 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "learning_rate": 4.077343397274864e-05, | |
| "loss": 3.3489, | |
| "step": 31200 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "learning_rate": 4.0715550873455974e-05, | |
| "loss": 3.3988, | |
| "step": 31300 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "learning_rate": 4.06576677741633e-05, | |
| "loss": 3.4295, | |
| "step": 31400 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "learning_rate": 4.0599784674870634e-05, | |
| "loss": 3.4186, | |
| "step": 31500 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "learning_rate": 4.0541901575577965e-05, | |
| "loss": 3.4509, | |
| "step": 31600 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "learning_rate": 4.0484018476285295e-05, | |
| "loss": 3.3621, | |
| "step": 31700 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "learning_rate": 4.042613537699263e-05, | |
| "loss": 3.4542, | |
| "step": 31800 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "learning_rate": 4.036825227769996e-05, | |
| "loss": 3.3612, | |
| "step": 31900 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "learning_rate": 4.031036917840729e-05, | |
| "loss": 3.3997, | |
| "step": 32000 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "learning_rate": 4.025248607911462e-05, | |
| "loss": 3.4172, | |
| "step": 32100 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "learning_rate": 4.019460297982195e-05, | |
| "loss": 3.4839, | |
| "step": 32200 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "learning_rate": 4.013671988052928e-05, | |
| "loss": 3.3381, | |
| "step": 32300 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "learning_rate": 4.007883678123662e-05, | |
| "loss": 3.4331, | |
| "step": 32400 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "learning_rate": 4.002095368194395e-05, | |
| "loss": 3.3947, | |
| "step": 32500 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "learning_rate": 3.996307058265128e-05, | |
| "loss": 3.3557, | |
| "step": 32600 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "learning_rate": 3.990518748335861e-05, | |
| "loss": 3.3993, | |
| "step": 32700 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "learning_rate": 3.984730438406594e-05, | |
| "loss": 3.3265, | |
| "step": 32800 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "learning_rate": 3.9789421284773276e-05, | |
| "loss": 3.3464, | |
| "step": 32900 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "learning_rate": 3.9731538185480606e-05, | |
| "loss": 3.354, | |
| "step": 33000 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "learning_rate": 3.9673655086187936e-05, | |
| "loss": 3.4181, | |
| "step": 33100 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "learning_rate": 3.9615771986895266e-05, | |
| "loss": 3.3932, | |
| "step": 33200 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "learning_rate": 3.9558467718595526e-05, | |
| "loss": 3.3718, | |
| "step": 33300 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "learning_rate": 3.950058461930286e-05, | |
| "loss": 3.3526, | |
| "step": 33400 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "learning_rate": 3.9442701520010186e-05, | |
| "loss": 3.4861, | |
| "step": 33500 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "learning_rate": 3.938481842071752e-05, | |
| "loss": 3.4428, | |
| "step": 33600 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "learning_rate": 3.932693532142485e-05, | |
| "loss": 3.4784, | |
| "step": 33700 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "learning_rate": 3.926905222213218e-05, | |
| "loss": 3.3259, | |
| "step": 33800 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "eval_gen_len": 19.3901, | |
| "eval_loss": 3.25349760055542, | |
| "eval_rouge1": 17.942, | |
| "eval_rouge2": 4.5143, | |
| "eval_rougeL": 14.2766, | |
| "eval_rougeLsum": 15.582, | |
| "eval_runtime": 746.4697, | |
| "eval_samples_per_second": 10.083, | |
| "eval_steps_per_second": 2.521, | |
| "step": 33875 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "learning_rate": 3.921116912283951e-05, | |
| "loss": 3.3341, | |
| "step": 33900 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "learning_rate": 3.915328602354685e-05, | |
| "loss": 3.2272, | |
| "step": 34000 | |
| }, | |
| { | |
| "epoch": 1.01, | |
| "learning_rate": 3.9095402924254174e-05, | |
| "loss": 3.1016, | |
| "step": 34100 | |
| }, | |
| { | |
| "epoch": 1.01, | |
| "learning_rate": 3.903751982496151e-05, | |
| "loss": 3.0869, | |
| "step": 34200 | |
| }, | |
| { | |
| "epoch": 1.01, | |
| "learning_rate": 3.897963672566884e-05, | |
| "loss": 3.0574, | |
| "step": 34300 | |
| }, | |
| { | |
| "epoch": 1.02, | |
| "learning_rate": 3.892175362637617e-05, | |
| "loss": 3.0656, | |
| "step": 34400 | |
| }, | |
| { | |
| "epoch": 1.02, | |
| "learning_rate": 3.88638705270835e-05, | |
| "loss": 3.0487, | |
| "step": 34500 | |
| }, | |
| { | |
| "epoch": 1.02, | |
| "learning_rate": 3.880598742779084e-05, | |
| "loss": 3.0965, | |
| "step": 34600 | |
| }, | |
| { | |
| "epoch": 1.02, | |
| "learning_rate": 3.874810432849817e-05, | |
| "loss": 3.1261, | |
| "step": 34700 | |
| }, | |
| { | |
| "epoch": 1.03, | |
| "learning_rate": 3.86902212292055e-05, | |
| "loss": 3.0195, | |
| "step": 34800 | |
| }, | |
| { | |
| "epoch": 1.03, | |
| "learning_rate": 3.8632338129912835e-05, | |
| "loss": 3.1102, | |
| "step": 34900 | |
| }, | |
| { | |
| "epoch": 1.03, | |
| "learning_rate": 3.857445503062016e-05, | |
| "loss": 3.0478, | |
| "step": 35000 | |
| }, | |
| { | |
| "epoch": 1.04, | |
| "learning_rate": 3.8516571931327495e-05, | |
| "loss": 3.0921, | |
| "step": 35100 | |
| }, | |
| { | |
| "epoch": 1.04, | |
| "learning_rate": 3.8458688832034825e-05, | |
| "loss": 3.0725, | |
| "step": 35200 | |
| }, | |
| { | |
| "epoch": 1.04, | |
| "learning_rate": 3.8400805732742155e-05, | |
| "loss": 3.0383, | |
| "step": 35300 | |
| }, | |
| { | |
| "epoch": 1.05, | |
| "learning_rate": 3.8342922633449485e-05, | |
| "loss": 3.0486, | |
| "step": 35400 | |
| }, | |
| { | |
| "epoch": 1.05, | |
| "learning_rate": 3.828503953415682e-05, | |
| "loss": 3.0608, | |
| "step": 35500 | |
| }, | |
| { | |
| "epoch": 1.05, | |
| "learning_rate": 3.8227156434864145e-05, | |
| "loss": 3.09, | |
| "step": 35600 | |
| }, | |
| { | |
| "epoch": 1.05, | |
| "learning_rate": 3.8169852166564405e-05, | |
| "loss": 3.0758, | |
| "step": 35700 | |
| }, | |
| { | |
| "epoch": 1.06, | |
| "learning_rate": 3.811196906727174e-05, | |
| "loss": 3.0034, | |
| "step": 35800 | |
| }, | |
| { | |
| "epoch": 1.06, | |
| "learning_rate": 3.805408596797907e-05, | |
| "loss": 3.1317, | |
| "step": 35900 | |
| }, | |
| { | |
| "epoch": 1.06, | |
| "learning_rate": 3.79962028686864e-05, | |
| "loss": 3.09, | |
| "step": 36000 | |
| }, | |
| { | |
| "epoch": 1.07, | |
| "learning_rate": 3.793831976939373e-05, | |
| "loss": 3.1168, | |
| "step": 36100 | |
| }, | |
| { | |
| "epoch": 1.07, | |
| "learning_rate": 3.788043667010107e-05, | |
| "loss": 3.115, | |
| "step": 36200 | |
| }, | |
| { | |
| "epoch": 1.07, | |
| "learning_rate": 3.782255357080839e-05, | |
| "loss": 3.0612, | |
| "step": 36300 | |
| }, | |
| { | |
| "epoch": 1.07, | |
| "learning_rate": 3.776467047151573e-05, | |
| "loss": 3.0553, | |
| "step": 36400 | |
| }, | |
| { | |
| "epoch": 1.08, | |
| "learning_rate": 3.770678737222306e-05, | |
| "loss": 3.0838, | |
| "step": 36500 | |
| }, | |
| { | |
| "epoch": 1.08, | |
| "learning_rate": 3.764890427293039e-05, | |
| "loss": 3.13, | |
| "step": 36600 | |
| }, | |
| { | |
| "epoch": 1.08, | |
| "learning_rate": 3.7591021173637726e-05, | |
| "loss": 3.0685, | |
| "step": 36700 | |
| }, | |
| { | |
| "epoch": 1.09, | |
| "learning_rate": 3.7533138074345056e-05, | |
| "loss": 3.0328, | |
| "step": 36800 | |
| }, | |
| { | |
| "epoch": 1.09, | |
| "learning_rate": 3.7475254975052386e-05, | |
| "loss": 3.0607, | |
| "step": 36900 | |
| }, | |
| { | |
| "epoch": 1.09, | |
| "learning_rate": 3.7417371875759717e-05, | |
| "loss": 3.0557, | |
| "step": 37000 | |
| }, | |
| { | |
| "epoch": 1.1, | |
| "learning_rate": 3.735948877646705e-05, | |
| "loss": 3.1025, | |
| "step": 37100 | |
| }, | |
| { | |
| "epoch": 1.1, | |
| "learning_rate": 3.730160567717438e-05, | |
| "loss": 3.0598, | |
| "step": 37200 | |
| }, | |
| { | |
| "epoch": 1.1, | |
| "learning_rate": 3.7243722577881714e-05, | |
| "loss": 3.1292, | |
| "step": 37300 | |
| }, | |
| { | |
| "epoch": 1.1, | |
| "learning_rate": 3.7185839478589044e-05, | |
| "loss": 3.0739, | |
| "step": 37400 | |
| }, | |
| { | |
| "epoch": 1.11, | |
| "learning_rate": 3.7127956379296374e-05, | |
| "loss": 3.1366, | |
| "step": 37500 | |
| }, | |
| { | |
| "epoch": 1.11, | |
| "learning_rate": 3.707007328000371e-05, | |
| "loss": 3.0558, | |
| "step": 37600 | |
| }, | |
| { | |
| "epoch": 1.11, | |
| "learning_rate": 3.7012190180711034e-05, | |
| "loss": 3.0975, | |
| "step": 37700 | |
| }, | |
| { | |
| "epoch": 1.12, | |
| "learning_rate": 3.695430708141837e-05, | |
| "loss": 3.0948, | |
| "step": 37800 | |
| }, | |
| { | |
| "epoch": 1.12, | |
| "learning_rate": 3.68964239821257e-05, | |
| "loss": 3.1102, | |
| "step": 37900 | |
| }, | |
| { | |
| "epoch": 1.12, | |
| "learning_rate": 3.683854088283303e-05, | |
| "loss": 3.1117, | |
| "step": 38000 | |
| }, | |
| { | |
| "epoch": 1.12, | |
| "learning_rate": 3.678065778354036e-05, | |
| "loss": 3.007, | |
| "step": 38100 | |
| }, | |
| { | |
| "epoch": 1.13, | |
| "learning_rate": 3.67227746842477e-05, | |
| "loss": 2.9965, | |
| "step": 38200 | |
| }, | |
| { | |
| "epoch": 1.13, | |
| "learning_rate": 3.666489158495502e-05, | |
| "loss": 3.0942, | |
| "step": 38300 | |
| }, | |
| { | |
| "epoch": 1.13, | |
| "learning_rate": 3.660700848566236e-05, | |
| "loss": 3.0544, | |
| "step": 38400 | |
| }, | |
| { | |
| "epoch": 1.14, | |
| "learning_rate": 3.654912538636969e-05, | |
| "loss": 3.071, | |
| "step": 38500 | |
| }, | |
| { | |
| "epoch": 1.14, | |
| "learning_rate": 3.649124228707702e-05, | |
| "loss": 3.0174, | |
| "step": 38600 | |
| }, | |
| { | |
| "epoch": 1.14, | |
| "learning_rate": 3.6433359187784355e-05, | |
| "loss": 3.0719, | |
| "step": 38700 | |
| }, | |
| { | |
| "epoch": 1.15, | |
| "learning_rate": 3.6375476088491685e-05, | |
| "loss": 3.0904, | |
| "step": 38800 | |
| }, | |
| { | |
| "epoch": 1.15, | |
| "learning_rate": 3.6317592989199016e-05, | |
| "loss": 3.0974, | |
| "step": 38900 | |
| }, | |
| { | |
| "epoch": 1.15, | |
| "learning_rate": 3.6259709889906346e-05, | |
| "loss": 3.0983, | |
| "step": 39000 | |
| }, | |
| { | |
| "epoch": 1.15, | |
| "learning_rate": 3.620182679061368e-05, | |
| "loss": 3.092, | |
| "step": 39100 | |
| }, | |
| { | |
| "epoch": 1.16, | |
| "learning_rate": 3.6143943691321006e-05, | |
| "loss": 3.0146, | |
| "step": 39200 | |
| }, | |
| { | |
| "epoch": 1.16, | |
| "learning_rate": 3.608606059202834e-05, | |
| "loss": 3.1821, | |
| "step": 39300 | |
| }, | |
| { | |
| "epoch": 1.16, | |
| "learning_rate": 3.602817749273567e-05, | |
| "loss": 3.1287, | |
| "step": 39400 | |
| }, | |
| { | |
| "epoch": 1.17, | |
| "learning_rate": 3.5970294393443e-05, | |
| "loss": 3.0698, | |
| "step": 39500 | |
| }, | |
| { | |
| "epoch": 1.17, | |
| "learning_rate": 3.591241129415033e-05, | |
| "loss": 3.0935, | |
| "step": 39600 | |
| }, | |
| { | |
| "epoch": 1.17, | |
| "learning_rate": 3.585510702585059e-05, | |
| "loss": 3.0959, | |
| "step": 39700 | |
| }, | |
| { | |
| "epoch": 1.17, | |
| "learning_rate": 3.579722392655793e-05, | |
| "loss": 3.0677, | |
| "step": 39800 | |
| }, | |
| { | |
| "epoch": 1.18, | |
| "learning_rate": 3.573934082726525e-05, | |
| "loss": 3.12, | |
| "step": 39900 | |
| }, | |
| { | |
| "epoch": 1.18, | |
| "learning_rate": 3.568145772797259e-05, | |
| "loss": 3.0438, | |
| "step": 40000 | |
| }, | |
| { | |
| "epoch": 1.18, | |
| "learning_rate": 3.562357462867992e-05, | |
| "loss": 3.1065, | |
| "step": 40100 | |
| }, | |
| { | |
| "epoch": 1.19, | |
| "learning_rate": 3.556569152938725e-05, | |
| "loss": 3.122, | |
| "step": 40200 | |
| }, | |
| { | |
| "epoch": 1.19, | |
| "learning_rate": 3.550780843009458e-05, | |
| "loss": 3.0528, | |
| "step": 40300 | |
| }, | |
| { | |
| "epoch": 1.19, | |
| "learning_rate": 3.544992533080192e-05, | |
| "loss": 3.0939, | |
| "step": 40400 | |
| }, | |
| { | |
| "epoch": 1.2, | |
| "learning_rate": 3.539204223150925e-05, | |
| "loss": 3.0964, | |
| "step": 40500 | |
| }, | |
| { | |
| "epoch": 1.2, | |
| "learning_rate": 3.533415913221658e-05, | |
| "loss": 3.0712, | |
| "step": 40600 | |
| }, | |
| { | |
| "epoch": 1.2, | |
| "learning_rate": 3.5276276032923914e-05, | |
| "loss": 3.0369, | |
| "step": 40700 | |
| }, | |
| { | |
| "epoch": 1.2, | |
| "learning_rate": 3.521839293363124e-05, | |
| "loss": 3.08, | |
| "step": 40800 | |
| }, | |
| { | |
| "epoch": 1.21, | |
| "learning_rate": 3.5160509834338574e-05, | |
| "loss": 3.1407, | |
| "step": 40900 | |
| }, | |
| { | |
| "epoch": 1.21, | |
| "learning_rate": 3.5102626735045904e-05, | |
| "loss": 3.0949, | |
| "step": 41000 | |
| }, | |
| { | |
| "epoch": 1.21, | |
| "learning_rate": 3.5044743635753234e-05, | |
| "loss": 3.157, | |
| "step": 41100 | |
| }, | |
| { | |
| "epoch": 1.22, | |
| "learning_rate": 3.4986860536460564e-05, | |
| "loss": 3.0251, | |
| "step": 41200 | |
| }, | |
| { | |
| "epoch": 1.22, | |
| "learning_rate": 3.49289774371679e-05, | |
| "loss": 3.1212, | |
| "step": 41300 | |
| }, | |
| { | |
| "epoch": 1.22, | |
| "learning_rate": 3.4871094337875225e-05, | |
| "loss": 3.1201, | |
| "step": 41400 | |
| }, | |
| { | |
| "epoch": 1.23, | |
| "learning_rate": 3.481321123858256e-05, | |
| "loss": 3.1373, | |
| "step": 41500 | |
| }, | |
| { | |
| "epoch": 1.23, | |
| "learning_rate": 3.475532813928989e-05, | |
| "loss": 3.1454, | |
| "step": 41600 | |
| }, | |
| { | |
| "epoch": 1.23, | |
| "learning_rate": 3.469802387099015e-05, | |
| "loss": 3.1846, | |
| "step": 41700 | |
| }, | |
| { | |
| "epoch": 1.23, | |
| "learning_rate": 3.464014077169748e-05, | |
| "loss": 3.075, | |
| "step": 41800 | |
| }, | |
| { | |
| "epoch": 1.24, | |
| "learning_rate": 3.458225767240481e-05, | |
| "loss": 3.1067, | |
| "step": 41900 | |
| }, | |
| { | |
| "epoch": 1.24, | |
| "learning_rate": 3.452437457311215e-05, | |
| "loss": 3.0698, | |
| "step": 42000 | |
| }, | |
| { | |
| "epoch": 1.24, | |
| "learning_rate": 3.446649147381947e-05, | |
| "loss": 3.197, | |
| "step": 42100 | |
| }, | |
| { | |
| "epoch": 1.25, | |
| "learning_rate": 3.440860837452681e-05, | |
| "loss": 3.0678, | |
| "step": 42200 | |
| }, | |
| { | |
| "epoch": 1.25, | |
| "learning_rate": 3.435072527523414e-05, | |
| "loss": 3.111, | |
| "step": 42300 | |
| }, | |
| { | |
| "epoch": 1.25, | |
| "learning_rate": 3.429284217594147e-05, | |
| "loss": 3.1098, | |
| "step": 42400 | |
| }, | |
| { | |
| "epoch": 1.25, | |
| "learning_rate": 3.4234959076648805e-05, | |
| "loss": 3.0416, | |
| "step": 42500 | |
| }, | |
| { | |
| "epoch": 1.26, | |
| "learning_rate": 3.4177075977356136e-05, | |
| "loss": 3.0468, | |
| "step": 42600 | |
| }, | |
| { | |
| "epoch": 1.26, | |
| "learning_rate": 3.4119192878063466e-05, | |
| "loss": 3.1422, | |
| "step": 42700 | |
| }, | |
| { | |
| "epoch": 1.26, | |
| "learning_rate": 3.4061309778770796e-05, | |
| "loss": 3.1238, | |
| "step": 42800 | |
| }, | |
| { | |
| "epoch": 1.27, | |
| "learning_rate": 3.4003426679478126e-05, | |
| "loss": 3.0876, | |
| "step": 42900 | |
| }, | |
| { | |
| "epoch": 1.27, | |
| "learning_rate": 3.3945543580185456e-05, | |
| "loss": 3.0583, | |
| "step": 43000 | |
| }, | |
| { | |
| "epoch": 1.27, | |
| "learning_rate": 3.388766048089279e-05, | |
| "loss": 3.1087, | |
| "step": 43100 | |
| }, | |
| { | |
| "epoch": 1.28, | |
| "learning_rate": 3.382977738160012e-05, | |
| "loss": 3.0599, | |
| "step": 43200 | |
| }, | |
| { | |
| "epoch": 1.28, | |
| "learning_rate": 3.377189428230745e-05, | |
| "loss": 3.0935, | |
| "step": 43300 | |
| }, | |
| { | |
| "epoch": 1.28, | |
| "learning_rate": 3.371401118301479e-05, | |
| "loss": 3.1727, | |
| "step": 43400 | |
| }, | |
| { | |
| "epoch": 1.28, | |
| "learning_rate": 3.365612808372211e-05, | |
| "loss": 3.0115, | |
| "step": 43500 | |
| }, | |
| { | |
| "epoch": 1.29, | |
| "learning_rate": 3.359824498442945e-05, | |
| "loss": 3.1088, | |
| "step": 43600 | |
| }, | |
| { | |
| "epoch": 1.29, | |
| "learning_rate": 3.354036188513678e-05, | |
| "loss": 3.0485, | |
| "step": 43700 | |
| }, | |
| { | |
| "epoch": 1.29, | |
| "learning_rate": 3.348247878584411e-05, | |
| "loss": 3.102, | |
| "step": 43800 | |
| }, | |
| { | |
| "epoch": 1.3, | |
| "learning_rate": 3.342459568655144e-05, | |
| "loss": 3.0692, | |
| "step": 43900 | |
| }, | |
| { | |
| "epoch": 1.3, | |
| "learning_rate": 3.336671258725878e-05, | |
| "loss": 3.1147, | |
| "step": 44000 | |
| }, | |
| { | |
| "epoch": 1.3, | |
| "learning_rate": 3.33088294879661e-05, | |
| "loss": 3.1878, | |
| "step": 44100 | |
| }, | |
| { | |
| "epoch": 1.3, | |
| "learning_rate": 3.325152521966636e-05, | |
| "loss": 3.0811, | |
| "step": 44200 | |
| }, | |
| { | |
| "epoch": 1.31, | |
| "learning_rate": 3.31936421203737e-05, | |
| "loss": 3.0288, | |
| "step": 44300 | |
| }, | |
| { | |
| "epoch": 1.31, | |
| "learning_rate": 3.313575902108103e-05, | |
| "loss": 3.0671, | |
| "step": 44400 | |
| }, | |
| { | |
| "epoch": 1.31, | |
| "learning_rate": 3.307787592178836e-05, | |
| "loss": 3.0503, | |
| "step": 44500 | |
| }, | |
| { | |
| "epoch": 1.32, | |
| "learning_rate": 3.301999282249569e-05, | |
| "loss": 3.0966, | |
| "step": 44600 | |
| }, | |
| { | |
| "epoch": 1.32, | |
| "learning_rate": 3.2962109723203024e-05, | |
| "loss": 3.0395, | |
| "step": 44700 | |
| }, | |
| { | |
| "epoch": 1.32, | |
| "learning_rate": 3.290422662391035e-05, | |
| "loss": 3.0602, | |
| "step": 44800 | |
| }, | |
| { | |
| "epoch": 1.33, | |
| "learning_rate": 3.2846343524617684e-05, | |
| "loss": 3.075, | |
| "step": 44900 | |
| }, | |
| { | |
| "epoch": 1.33, | |
| "learning_rate": 3.2788460425325014e-05, | |
| "loss": 3.1158, | |
| "step": 45000 | |
| }, | |
| { | |
| "epoch": 1.33, | |
| "learning_rate": 3.2730577326032345e-05, | |
| "loss": 3.0903, | |
| "step": 45100 | |
| }, | |
| { | |
| "epoch": 1.33, | |
| "learning_rate": 3.267269422673968e-05, | |
| "loss": 3.0556, | |
| "step": 45200 | |
| }, | |
| { | |
| "epoch": 1.34, | |
| "learning_rate": 3.261481112744701e-05, | |
| "loss": 3.1016, | |
| "step": 45300 | |
| }, | |
| { | |
| "epoch": 1.34, | |
| "learning_rate": 3.255692802815434e-05, | |
| "loss": 3.1105, | |
| "step": 45400 | |
| }, | |
| { | |
| "epoch": 1.34, | |
| "learning_rate": 3.249904492886167e-05, | |
| "loss": 3.1088, | |
| "step": 45500 | |
| }, | |
| { | |
| "epoch": 1.35, | |
| "learning_rate": 3.244116182956901e-05, | |
| "loss": 3.1597, | |
| "step": 45600 | |
| }, | |
| { | |
| "epoch": 1.35, | |
| "learning_rate": 3.238327873027633e-05, | |
| "loss": 3.0724, | |
| "step": 45700 | |
| }, | |
| { | |
| "epoch": 1.35, | |
| "learning_rate": 3.232539563098367e-05, | |
| "loss": 3.0631, | |
| "step": 45800 | |
| }, | |
| { | |
| "epoch": 1.35, | |
| "learning_rate": 3.2267512531691e-05, | |
| "loss": 3.0182, | |
| "step": 45900 | |
| }, | |
| { | |
| "epoch": 1.36, | |
| "learning_rate": 3.220962943239833e-05, | |
| "loss": 3.0879, | |
| "step": 46000 | |
| }, | |
| { | |
| "epoch": 1.36, | |
| "learning_rate": 3.215174633310566e-05, | |
| "loss": 3.1382, | |
| "step": 46100 | |
| }, | |
| { | |
| "epoch": 1.36, | |
| "learning_rate": 3.2093863233812996e-05, | |
| "loss": 3.1101, | |
| "step": 46200 | |
| }, | |
| { | |
| "epoch": 1.37, | |
| "learning_rate": 3.2035980134520326e-05, | |
| "loss": 3.0471, | |
| "step": 46300 | |
| }, | |
| { | |
| "epoch": 1.37, | |
| "learning_rate": 3.1978097035227656e-05, | |
| "loss": 3.0564, | |
| "step": 46400 | |
| }, | |
| { | |
| "epoch": 1.37, | |
| "learning_rate": 3.1920213935934986e-05, | |
| "loss": 3.1499, | |
| "step": 46500 | |
| }, | |
| { | |
| "epoch": 1.38, | |
| "learning_rate": 3.1862330836642316e-05, | |
| "loss": 3.092, | |
| "step": 46600 | |
| }, | |
| { | |
| "epoch": 1.38, | |
| "learning_rate": 3.180444773734965e-05, | |
| "loss": 3.0732, | |
| "step": 46700 | |
| }, | |
| { | |
| "epoch": 1.38, | |
| "learning_rate": 3.1746564638056983e-05, | |
| "loss": 3.0483, | |
| "step": 46800 | |
| }, | |
| { | |
| "epoch": 1.38, | |
| "learning_rate": 3.1688681538764314e-05, | |
| "loss": 3.0419, | |
| "step": 46900 | |
| }, | |
| { | |
| "epoch": 1.39, | |
| "learning_rate": 3.1630798439471644e-05, | |
| "loss": 3.0499, | |
| "step": 47000 | |
| }, | |
| { | |
| "epoch": 1.39, | |
| "learning_rate": 3.1572915340178974e-05, | |
| "loss": 3.1122, | |
| "step": 47100 | |
| }, | |
| { | |
| "epoch": 1.39, | |
| "learning_rate": 3.1515032240886304e-05, | |
| "loss": 3.1401, | |
| "step": 47200 | |
| }, | |
| { | |
| "epoch": 1.4, | |
| "learning_rate": 3.145714914159364e-05, | |
| "loss": 3.072, | |
| "step": 47300 | |
| }, | |
| { | |
| "epoch": 1.4, | |
| "learning_rate": 3.139926604230097e-05, | |
| "loss": 3.0722, | |
| "step": 47400 | |
| }, | |
| { | |
| "epoch": 1.4, | |
| "learning_rate": 3.13413829430083e-05, | |
| "loss": 3.0141, | |
| "step": 47500 | |
| }, | |
| { | |
| "epoch": 1.41, | |
| "learning_rate": 3.128349984371564e-05, | |
| "loss": 3.0253, | |
| "step": 47600 | |
| }, | |
| { | |
| "epoch": 1.41, | |
| "learning_rate": 3.122561674442296e-05, | |
| "loss": 3.037, | |
| "step": 47700 | |
| }, | |
| { | |
| "epoch": 1.41, | |
| "learning_rate": 3.11677336451303e-05, | |
| "loss": 3.0667, | |
| "step": 47800 | |
| }, | |
| { | |
| "epoch": 1.41, | |
| "learning_rate": 3.110985054583763e-05, | |
| "loss": 3.0826, | |
| "step": 47900 | |
| }, | |
| { | |
| "epoch": 1.42, | |
| "learning_rate": 3.105196744654496e-05, | |
| "loss": 3.0285, | |
| "step": 48000 | |
| }, | |
| { | |
| "epoch": 1.42, | |
| "learning_rate": 3.099408434725229e-05, | |
| "loss": 3.1176, | |
| "step": 48100 | |
| }, | |
| { | |
| "epoch": 1.42, | |
| "learning_rate": 3.0936201247959625e-05, | |
| "loss": 3.0716, | |
| "step": 48200 | |
| }, | |
| { | |
| "epoch": 1.43, | |
| "learning_rate": 3.0878896979659885e-05, | |
| "loss": 2.9875, | |
| "step": 48300 | |
| }, | |
| { | |
| "epoch": 1.43, | |
| "learning_rate": 3.082101388036721e-05, | |
| "loss": 3.0571, | |
| "step": 48400 | |
| }, | |
| { | |
| "epoch": 1.43, | |
| "learning_rate": 3.0763130781074545e-05, | |
| "loss": 3.0559, | |
| "step": 48500 | |
| }, | |
| { | |
| "epoch": 1.43, | |
| "learning_rate": 3.0705247681781875e-05, | |
| "loss": 3.072, | |
| "step": 48600 | |
| }, | |
| { | |
| "epoch": 1.44, | |
| "learning_rate": 3.0647364582489205e-05, | |
| "loss": 3.0742, | |
| "step": 48700 | |
| }, | |
| { | |
| "epoch": 1.44, | |
| "learning_rate": 3.0589481483196535e-05, | |
| "loss": 3.0804, | |
| "step": 48800 | |
| }, | |
| { | |
| "epoch": 1.44, | |
| "learning_rate": 3.053159838390387e-05, | |
| "loss": 3.1331, | |
| "step": 48900 | |
| }, | |
| { | |
| "epoch": 1.45, | |
| "learning_rate": 3.04737152846112e-05, | |
| "loss": 3.0802, | |
| "step": 49000 | |
| }, | |
| { | |
| "epoch": 1.45, | |
| "learning_rate": 3.0415832185318532e-05, | |
| "loss": 3.0288, | |
| "step": 49100 | |
| }, | |
| { | |
| "epoch": 1.45, | |
| "learning_rate": 3.0357949086025862e-05, | |
| "loss": 3.0775, | |
| "step": 49200 | |
| }, | |
| { | |
| "epoch": 1.46, | |
| "learning_rate": 3.0300065986733196e-05, | |
| "loss": 3.0955, | |
| "step": 49300 | |
| }, | |
| { | |
| "epoch": 1.46, | |
| "learning_rate": 3.024218288744053e-05, | |
| "loss": 3.1154, | |
| "step": 49400 | |
| }, | |
| { | |
| "epoch": 1.46, | |
| "learning_rate": 3.0184299788147856e-05, | |
| "loss": 3.0906, | |
| "step": 49500 | |
| }, | |
| { | |
| "epoch": 1.46, | |
| "learning_rate": 3.0126416688855193e-05, | |
| "loss": 3.0599, | |
| "step": 49600 | |
| }, | |
| { | |
| "epoch": 1.47, | |
| "learning_rate": 3.006853358956252e-05, | |
| "loss": 3.0214, | |
| "step": 49700 | |
| }, | |
| { | |
| "epoch": 1.47, | |
| "learning_rate": 3.0010650490269853e-05, | |
| "loss": 3.0938, | |
| "step": 49800 | |
| }, | |
| { | |
| "epoch": 1.47, | |
| "learning_rate": 2.9953346221970113e-05, | |
| "loss": 3.0698, | |
| "step": 49900 | |
| }, | |
| { | |
| "epoch": 1.48, | |
| "learning_rate": 2.9895463122677443e-05, | |
| "loss": 3.1009, | |
| "step": 50000 | |
| }, | |
| { | |
| "epoch": 1.48, | |
| "learning_rate": 2.9837580023384776e-05, | |
| "loss": 3.0987, | |
| "step": 50100 | |
| }, | |
| { | |
| "epoch": 1.48, | |
| "learning_rate": 2.9779696924092103e-05, | |
| "loss": 3.0924, | |
| "step": 50200 | |
| }, | |
| { | |
| "epoch": 1.48, | |
| "learning_rate": 2.972181382479944e-05, | |
| "loss": 3.0633, | |
| "step": 50300 | |
| }, | |
| { | |
| "epoch": 1.49, | |
| "learning_rate": 2.9663930725506767e-05, | |
| "loss": 3.0363, | |
| "step": 50400 | |
| }, | |
| { | |
| "epoch": 1.49, | |
| "learning_rate": 2.96060476262141e-05, | |
| "loss": 3.0711, | |
| "step": 50500 | |
| }, | |
| { | |
| "epoch": 1.49, | |
| "learning_rate": 2.954816452692143e-05, | |
| "loss": 3.1192, | |
| "step": 50600 | |
| }, | |
| { | |
| "epoch": 1.5, | |
| "learning_rate": 2.9490281427628764e-05, | |
| "loss": 3.0765, | |
| "step": 50700 | |
| }, | |
| { | |
| "epoch": 1.5, | |
| "learning_rate": 2.943239832833609e-05, | |
| "loss": 3.1171, | |
| "step": 50800 | |
| }, | |
| { | |
| "epoch": 1.5, | |
| "learning_rate": 2.9374515229043427e-05, | |
| "loss": 3.1118, | |
| "step": 50900 | |
| }, | |
| { | |
| "epoch": 1.51, | |
| "learning_rate": 2.9316632129750754e-05, | |
| "loss": 2.9891, | |
| "step": 51000 | |
| }, | |
| { | |
| "epoch": 1.51, | |
| "learning_rate": 2.9258749030458087e-05, | |
| "loss": 3.1239, | |
| "step": 51100 | |
| }, | |
| { | |
| "epoch": 1.51, | |
| "learning_rate": 2.920086593116542e-05, | |
| "loss": 3.0719, | |
| "step": 51200 | |
| }, | |
| { | |
| "epoch": 1.51, | |
| "learning_rate": 2.914298283187275e-05, | |
| "loss": 3.0945, | |
| "step": 51300 | |
| }, | |
| { | |
| "epoch": 1.52, | |
| "learning_rate": 2.9085099732580084e-05, | |
| "loss": 3.1234, | |
| "step": 51400 | |
| }, | |
| { | |
| "epoch": 1.52, | |
| "learning_rate": 2.9027216633287415e-05, | |
| "loss": 3.1663, | |
| "step": 51500 | |
| }, | |
| { | |
| "epoch": 1.52, | |
| "learning_rate": 2.8969333533994748e-05, | |
| "loss": 3.0325, | |
| "step": 51600 | |
| }, | |
| { | |
| "epoch": 1.53, | |
| "learning_rate": 2.8911450434702075e-05, | |
| "loss": 3.0971, | |
| "step": 51700 | |
| }, | |
| { | |
| "epoch": 1.53, | |
| "learning_rate": 2.8853567335409408e-05, | |
| "loss": 3.0374, | |
| "step": 51800 | |
| }, | |
| { | |
| "epoch": 1.53, | |
| "learning_rate": 2.879568423611674e-05, | |
| "loss": 3.0485, | |
| "step": 51900 | |
| }, | |
| { | |
| "epoch": 1.54, | |
| "learning_rate": 2.8737801136824072e-05, | |
| "loss": 3.0894, | |
| "step": 52000 | |
| }, | |
| { | |
| "epoch": 1.54, | |
| "learning_rate": 2.8679918037531402e-05, | |
| "loss": 3.1166, | |
| "step": 52100 | |
| }, | |
| { | |
| "epoch": 1.54, | |
| "learning_rate": 2.8622034938238735e-05, | |
| "loss": 3.0091, | |
| "step": 52200 | |
| }, | |
| { | |
| "epoch": 1.54, | |
| "learning_rate": 2.856415183894607e-05, | |
| "loss": 3.0695, | |
| "step": 52300 | |
| }, | |
| { | |
| "epoch": 1.55, | |
| "learning_rate": 2.8506268739653396e-05, | |
| "loss": 3.0978, | |
| "step": 52400 | |
| }, | |
| { | |
| "epoch": 1.55, | |
| "learning_rate": 2.844838564036073e-05, | |
| "loss": 3.0559, | |
| "step": 52500 | |
| }, | |
| { | |
| "epoch": 1.55, | |
| "learning_rate": 2.839050254106806e-05, | |
| "loss": 3.0492, | |
| "step": 52600 | |
| }, | |
| { | |
| "epoch": 1.56, | |
| "learning_rate": 2.8332619441775393e-05, | |
| "loss": 3.0304, | |
| "step": 52700 | |
| }, | |
| { | |
| "epoch": 1.56, | |
| "learning_rate": 2.8274736342482723e-05, | |
| "loss": 3.1342, | |
| "step": 52800 | |
| }, | |
| { | |
| "epoch": 1.56, | |
| "learning_rate": 2.8216853243190056e-05, | |
| "loss": 3.076, | |
| "step": 52900 | |
| }, | |
| { | |
| "epoch": 1.56, | |
| "learning_rate": 2.8158970143897383e-05, | |
| "loss": 2.9751, | |
| "step": 53000 | |
| }, | |
| { | |
| "epoch": 1.57, | |
| "learning_rate": 2.8101087044604717e-05, | |
| "loss": 3.0754, | |
| "step": 53100 | |
| }, | |
| { | |
| "epoch": 1.57, | |
| "learning_rate": 2.8043203945312047e-05, | |
| "loss": 3.0723, | |
| "step": 53200 | |
| }, | |
| { | |
| "epoch": 1.57, | |
| "learning_rate": 2.798532084601938e-05, | |
| "loss": 3.0367, | |
| "step": 53300 | |
| }, | |
| { | |
| "epoch": 1.58, | |
| "learning_rate": 2.7927437746726714e-05, | |
| "loss": 3.0745, | |
| "step": 53400 | |
| }, | |
| { | |
| "epoch": 1.58, | |
| "learning_rate": 2.7869554647434044e-05, | |
| "loss": 3.0561, | |
| "step": 53500 | |
| }, | |
| { | |
| "epoch": 1.58, | |
| "learning_rate": 2.7811671548141377e-05, | |
| "loss": 3.1498, | |
| "step": 53600 | |
| }, | |
| { | |
| "epoch": 1.59, | |
| "learning_rate": 2.7753788448848704e-05, | |
| "loss": 3.0278, | |
| "step": 53700 | |
| }, | |
| { | |
| "epoch": 1.59, | |
| "learning_rate": 2.769590534955604e-05, | |
| "loss": 3.0985, | |
| "step": 53800 | |
| }, | |
| { | |
| "epoch": 1.59, | |
| "learning_rate": 2.7638022250263367e-05, | |
| "loss": 3.0816, | |
| "step": 53900 | |
| }, | |
| { | |
| "epoch": 1.59, | |
| "learning_rate": 2.75801391509707e-05, | |
| "loss": 3.0399, | |
| "step": 54000 | |
| }, | |
| { | |
| "epoch": 1.6, | |
| "learning_rate": 2.752283488267096e-05, | |
| "loss": 3.0389, | |
| "step": 54100 | |
| }, | |
| { | |
| "epoch": 1.6, | |
| "learning_rate": 2.746495178337829e-05, | |
| "loss": 3.0626, | |
| "step": 54200 | |
| }, | |
| { | |
| "epoch": 1.6, | |
| "learning_rate": 2.7407068684085624e-05, | |
| "loss": 3.0901, | |
| "step": 54300 | |
| }, | |
| { | |
| "epoch": 1.61, | |
| "learning_rate": 2.734918558479295e-05, | |
| "loss": 3.0882, | |
| "step": 54400 | |
| }, | |
| { | |
| "epoch": 1.61, | |
| "learning_rate": 2.7291302485500288e-05, | |
| "loss": 3.1266, | |
| "step": 54500 | |
| }, | |
| { | |
| "epoch": 1.61, | |
| "learning_rate": 2.7233419386207614e-05, | |
| "loss": 3.0028, | |
| "step": 54600 | |
| }, | |
| { | |
| "epoch": 1.61, | |
| "learning_rate": 2.7175536286914948e-05, | |
| "loss": 3.0538, | |
| "step": 54700 | |
| }, | |
| { | |
| "epoch": 1.62, | |
| "learning_rate": 2.7117653187622278e-05, | |
| "loss": 3.0825, | |
| "step": 54800 | |
| }, | |
| { | |
| "epoch": 1.62, | |
| "learning_rate": 2.705977008832961e-05, | |
| "loss": 3.0424, | |
| "step": 54900 | |
| }, | |
| { | |
| "epoch": 1.62, | |
| "learning_rate": 2.7001886989036938e-05, | |
| "loss": 3.1286, | |
| "step": 55000 | |
| }, | |
| { | |
| "epoch": 1.63, | |
| "learning_rate": 2.6944003889744275e-05, | |
| "loss": 3.1161, | |
| "step": 55100 | |
| }, | |
| { | |
| "epoch": 1.63, | |
| "learning_rate": 2.688612079045161e-05, | |
| "loss": 3.0712, | |
| "step": 55200 | |
| }, | |
| { | |
| "epoch": 1.63, | |
| "learning_rate": 2.6828237691158935e-05, | |
| "loss": 3.0373, | |
| "step": 55300 | |
| }, | |
| { | |
| "epoch": 1.64, | |
| "learning_rate": 2.677035459186627e-05, | |
| "loss": 3.0883, | |
| "step": 55400 | |
| }, | |
| { | |
| "epoch": 1.64, | |
| "learning_rate": 2.67124714925736e-05, | |
| "loss": 3.041, | |
| "step": 55500 | |
| }, | |
| { | |
| "epoch": 1.64, | |
| "learning_rate": 2.6654588393280932e-05, | |
| "loss": 3.0623, | |
| "step": 55600 | |
| }, | |
| { | |
| "epoch": 1.64, | |
| "learning_rate": 2.6596705293988262e-05, | |
| "loss": 3.1202, | |
| "step": 55700 | |
| }, | |
| { | |
| "epoch": 1.65, | |
| "learning_rate": 2.6538822194695596e-05, | |
| "loss": 3.0211, | |
| "step": 55800 | |
| }, | |
| { | |
| "epoch": 1.65, | |
| "learning_rate": 2.6480939095402923e-05, | |
| "loss": 3.1307, | |
| "step": 55900 | |
| }, | |
| { | |
| "epoch": 1.65, | |
| "learning_rate": 2.6423055996110256e-05, | |
| "loss": 3.0234, | |
| "step": 56000 | |
| }, | |
| { | |
| "epoch": 1.66, | |
| "learning_rate": 2.6365172896817586e-05, | |
| "loss": 3.013, | |
| "step": 56100 | |
| }, | |
| { | |
| "epoch": 1.66, | |
| "learning_rate": 2.6307868628517846e-05, | |
| "loss": 3.0643, | |
| "step": 56200 | |
| }, | |
| { | |
| "epoch": 1.66, | |
| "learning_rate": 2.624998552922518e-05, | |
| "loss": 3.046, | |
| "step": 56300 | |
| }, | |
| { | |
| "epoch": 1.66, | |
| "learning_rate": 2.619210242993251e-05, | |
| "loss": 2.9987, | |
| "step": 56400 | |
| }, | |
| { | |
| "epoch": 1.67, | |
| "learning_rate": 2.6134219330639843e-05, | |
| "loss": 3.1, | |
| "step": 56500 | |
| }, | |
| { | |
| "epoch": 1.67, | |
| "learning_rate": 2.6076915062340102e-05, | |
| "loss": 3.0732, | |
| "step": 56600 | |
| }, | |
| { | |
| "epoch": 1.67, | |
| "learning_rate": 2.601903196304743e-05, | |
| "loss": 2.996, | |
| "step": 56700 | |
| }, | |
| { | |
| "epoch": 1.68, | |
| "learning_rate": 2.5961148863754763e-05, | |
| "loss": 2.9735, | |
| "step": 56800 | |
| }, | |
| { | |
| "epoch": 1.68, | |
| "learning_rate": 2.5903265764462093e-05, | |
| "loss": 3.1598, | |
| "step": 56900 | |
| }, | |
| { | |
| "epoch": 1.68, | |
| "learning_rate": 2.5845382665169426e-05, | |
| "loss": 3.0743, | |
| "step": 57000 | |
| }, | |
| { | |
| "epoch": 1.69, | |
| "learning_rate": 2.5787499565876756e-05, | |
| "loss": 3.0798, | |
| "step": 57100 | |
| }, | |
| { | |
| "epoch": 1.69, | |
| "learning_rate": 2.572961646658409e-05, | |
| "loss": 3.0895, | |
| "step": 57200 | |
| }, | |
| { | |
| "epoch": 1.69, | |
| "learning_rate": 2.5671733367291416e-05, | |
| "loss": 3.051, | |
| "step": 57300 | |
| }, | |
| { | |
| "epoch": 1.69, | |
| "learning_rate": 2.561385026799875e-05, | |
| "loss": 3.0544, | |
| "step": 57400 | |
| }, | |
| { | |
| "epoch": 1.7, | |
| "learning_rate": 2.555596716870608e-05, | |
| "loss": 3.0848, | |
| "step": 57500 | |
| }, | |
| { | |
| "epoch": 1.7, | |
| "learning_rate": 2.5498084069413413e-05, | |
| "loss": 3.0787, | |
| "step": 57600 | |
| }, | |
| { | |
| "epoch": 1.7, | |
| "learning_rate": 2.5440779801113673e-05, | |
| "loss": 3.1021, | |
| "step": 57700 | |
| }, | |
| { | |
| "epoch": 1.71, | |
| "learning_rate": 2.5382896701821003e-05, | |
| "loss": 3.0226, | |
| "step": 57800 | |
| }, | |
| { | |
| "epoch": 1.71, | |
| "learning_rate": 2.5325013602528337e-05, | |
| "loss": 3.0637, | |
| "step": 57900 | |
| }, | |
| { | |
| "epoch": 1.71, | |
| "learning_rate": 2.5267130503235663e-05, | |
| "loss": 3.1298, | |
| "step": 58000 | |
| }, | |
| { | |
| "epoch": 1.72, | |
| "learning_rate": 2.5209247403942997e-05, | |
| "loss": 3.0767, | |
| "step": 58100 | |
| }, | |
| { | |
| "epoch": 1.72, | |
| "learning_rate": 2.5151364304650327e-05, | |
| "loss": 2.9812, | |
| "step": 58200 | |
| }, | |
| { | |
| "epoch": 1.72, | |
| "learning_rate": 2.509348120535766e-05, | |
| "loss": 3.0861, | |
| "step": 58300 | |
| }, | |
| { | |
| "epoch": 1.72, | |
| "learning_rate": 2.5035598106064994e-05, | |
| "loss": 3.1261, | |
| "step": 58400 | |
| }, | |
| { | |
| "epoch": 1.73, | |
| "learning_rate": 2.4977715006772324e-05, | |
| "loss": 3.0393, | |
| "step": 58500 | |
| }, | |
| { | |
| "epoch": 1.73, | |
| "learning_rate": 2.4919831907479654e-05, | |
| "loss": 3.0622, | |
| "step": 58600 | |
| }, | |
| { | |
| "epoch": 1.73, | |
| "learning_rate": 2.4861948808186984e-05, | |
| "loss": 3.0922, | |
| "step": 58700 | |
| }, | |
| { | |
| "epoch": 1.74, | |
| "learning_rate": 2.480406570889432e-05, | |
| "loss": 3.0503, | |
| "step": 58800 | |
| }, | |
| { | |
| "epoch": 1.74, | |
| "learning_rate": 2.474618260960165e-05, | |
| "loss": 3.0453, | |
| "step": 58900 | |
| }, | |
| { | |
| "epoch": 1.74, | |
| "learning_rate": 2.468829951030898e-05, | |
| "loss": 3.0043, | |
| "step": 59000 | |
| }, | |
| { | |
| "epoch": 1.74, | |
| "learning_rate": 2.4630416411016315e-05, | |
| "loss": 3.0512, | |
| "step": 59100 | |
| }, | |
| { | |
| "epoch": 1.75, | |
| "learning_rate": 2.4572533311723645e-05, | |
| "loss": 3.0278, | |
| "step": 59200 | |
| }, | |
| { | |
| "epoch": 1.75, | |
| "learning_rate": 2.4514650212430975e-05, | |
| "loss": 3.0435, | |
| "step": 59300 | |
| }, | |
| { | |
| "epoch": 1.75, | |
| "learning_rate": 2.445676711313831e-05, | |
| "loss": 3.0182, | |
| "step": 59400 | |
| }, | |
| { | |
| "epoch": 1.76, | |
| "learning_rate": 2.439888401384564e-05, | |
| "loss": 3.013, | |
| "step": 59500 | |
| }, | |
| { | |
| "epoch": 1.76, | |
| "learning_rate": 2.434100091455297e-05, | |
| "loss": 3.1001, | |
| "step": 59600 | |
| }, | |
| { | |
| "epoch": 1.76, | |
| "learning_rate": 2.4283117815260302e-05, | |
| "loss": 3.0871, | |
| "step": 59700 | |
| }, | |
| { | |
| "epoch": 1.77, | |
| "learning_rate": 2.4225234715967632e-05, | |
| "loss": 3.0644, | |
| "step": 59800 | |
| }, | |
| { | |
| "epoch": 1.77, | |
| "learning_rate": 2.4167351616674962e-05, | |
| "loss": 3.0736, | |
| "step": 59900 | |
| }, | |
| { | |
| "epoch": 1.77, | |
| "learning_rate": 2.4109468517382296e-05, | |
| "loss": 3.0558, | |
| "step": 60000 | |
| }, | |
| { | |
| "epoch": 1.77, | |
| "learning_rate": 2.405158541808963e-05, | |
| "loss": 3.0976, | |
| "step": 60100 | |
| }, | |
| { | |
| "epoch": 1.78, | |
| "learning_rate": 2.399370231879696e-05, | |
| "loss": 3.0441, | |
| "step": 60200 | |
| }, | |
| { | |
| "epoch": 1.78, | |
| "learning_rate": 2.393581921950429e-05, | |
| "loss": 3.0623, | |
| "step": 60300 | |
| }, | |
| { | |
| "epoch": 1.78, | |
| "learning_rate": 2.3877936120211623e-05, | |
| "loss": 3.0432, | |
| "step": 60400 | |
| }, | |
| { | |
| "epoch": 1.79, | |
| "learning_rate": 2.3820053020918953e-05, | |
| "loss": 3.031, | |
| "step": 60500 | |
| }, | |
| { | |
| "epoch": 1.79, | |
| "learning_rate": 2.3762169921626283e-05, | |
| "loss": 3.0308, | |
| "step": 60600 | |
| }, | |
| { | |
| "epoch": 1.79, | |
| "learning_rate": 2.3704286822333617e-05, | |
| "loss": 3.0621, | |
| "step": 60700 | |
| }, | |
| { | |
| "epoch": 1.79, | |
| "learning_rate": 2.3646403723040947e-05, | |
| "loss": 2.9949, | |
| "step": 60800 | |
| }, | |
| { | |
| "epoch": 1.8, | |
| "learning_rate": 2.3588520623748277e-05, | |
| "loss": 3.0625, | |
| "step": 60900 | |
| }, | |
| { | |
| "epoch": 1.8, | |
| "learning_rate": 2.353063752445561e-05, | |
| "loss": 3.0175, | |
| "step": 61000 | |
| }, | |
| { | |
| "epoch": 1.8, | |
| "learning_rate": 2.3472754425162944e-05, | |
| "loss": 2.9987, | |
| "step": 61100 | |
| }, | |
| { | |
| "epoch": 1.81, | |
| "learning_rate": 2.3414871325870274e-05, | |
| "loss": 2.998, | |
| "step": 61200 | |
| }, | |
| { | |
| "epoch": 1.81, | |
| "learning_rate": 2.3356988226577607e-05, | |
| "loss": 3.0248, | |
| "step": 61300 | |
| }, | |
| { | |
| "epoch": 1.81, | |
| "learning_rate": 2.3299105127284938e-05, | |
| "loss": 3.0506, | |
| "step": 61400 | |
| }, | |
| { | |
| "epoch": 1.82, | |
| "learning_rate": 2.3241222027992268e-05, | |
| "loss": 3.1104, | |
| "step": 61500 | |
| }, | |
| { | |
| "epoch": 1.82, | |
| "learning_rate": 2.31833389286996e-05, | |
| "loss": 3.0363, | |
| "step": 61600 | |
| }, | |
| { | |
| "epoch": 1.82, | |
| "learning_rate": 2.312545582940693e-05, | |
| "loss": 3.0894, | |
| "step": 61700 | |
| }, | |
| { | |
| "epoch": 1.82, | |
| "learning_rate": 2.306757273011426e-05, | |
| "loss": 3.0058, | |
| "step": 61800 | |
| }, | |
| { | |
| "epoch": 1.83, | |
| "learning_rate": 2.300968963082159e-05, | |
| "loss": 2.9939, | |
| "step": 61900 | |
| }, | |
| { | |
| "epoch": 1.83, | |
| "learning_rate": 2.2951806531528925e-05, | |
| "loss": 2.9725, | |
| "step": 62000 | |
| }, | |
| { | |
| "epoch": 1.83, | |
| "learning_rate": 2.2893923432236255e-05, | |
| "loss": 3.0223, | |
| "step": 62100 | |
| }, | |
| { | |
| "epoch": 1.84, | |
| "learning_rate": 2.283604033294359e-05, | |
| "loss": 3.0144, | |
| "step": 62200 | |
| }, | |
| { | |
| "epoch": 1.84, | |
| "learning_rate": 2.2778157233650922e-05, | |
| "loss": 3.0021, | |
| "step": 62300 | |
| }, | |
| { | |
| "epoch": 1.84, | |
| "learning_rate": 2.2720274134358252e-05, | |
| "loss": 2.9874, | |
| "step": 62400 | |
| }, | |
| { | |
| "epoch": 1.85, | |
| "learning_rate": 2.2662391035065582e-05, | |
| "loss": 3.0061, | |
| "step": 62500 | |
| }, | |
| { | |
| "epoch": 1.85, | |
| "learning_rate": 2.2604507935772916e-05, | |
| "loss": 2.9589, | |
| "step": 62600 | |
| }, | |
| { | |
| "epoch": 1.85, | |
| "learning_rate": 2.2546624836480246e-05, | |
| "loss": 3.0054, | |
| "step": 62700 | |
| }, | |
| { | |
| "epoch": 1.85, | |
| "learning_rate": 2.2488741737187576e-05, | |
| "loss": 3.0209, | |
| "step": 62800 | |
| }, | |
| { | |
| "epoch": 1.86, | |
| "learning_rate": 2.243085863789491e-05, | |
| "loss": 3.0615, | |
| "step": 62900 | |
| }, | |
| { | |
| "epoch": 1.86, | |
| "learning_rate": 2.237297553860224e-05, | |
| "loss": 2.9817, | |
| "step": 63000 | |
| }, | |
| { | |
| "epoch": 1.86, | |
| "learning_rate": 2.231509243930957e-05, | |
| "loss": 3.0172, | |
| "step": 63100 | |
| }, | |
| { | |
| "epoch": 1.87, | |
| "learning_rate": 2.2257209340016903e-05, | |
| "loss": 3.0459, | |
| "step": 63200 | |
| }, | |
| { | |
| "epoch": 1.87, | |
| "learning_rate": 2.2199326240724233e-05, | |
| "loss": 3.0939, | |
| "step": 63300 | |
| }, | |
| { | |
| "epoch": 1.87, | |
| "learning_rate": 2.2141443141431567e-05, | |
| "loss": 3.0101, | |
| "step": 63400 | |
| }, | |
| { | |
| "epoch": 1.87, | |
| "learning_rate": 2.2083560042138897e-05, | |
| "loss": 3.0242, | |
| "step": 63500 | |
| }, | |
| { | |
| "epoch": 1.88, | |
| "learning_rate": 2.202567694284623e-05, | |
| "loss": 3.0988, | |
| "step": 63600 | |
| }, | |
| { | |
| "epoch": 1.88, | |
| "learning_rate": 2.196779384355356e-05, | |
| "loss": 3.089, | |
| "step": 63700 | |
| }, | |
| { | |
| "epoch": 1.88, | |
| "learning_rate": 2.190991074426089e-05, | |
| "loss": 3.1095, | |
| "step": 63800 | |
| }, | |
| { | |
| "epoch": 1.89, | |
| "learning_rate": 2.185260647596115e-05, | |
| "loss": 3.0332, | |
| "step": 63900 | |
| }, | |
| { | |
| "epoch": 1.89, | |
| "learning_rate": 2.1794723376668483e-05, | |
| "loss": 3.0994, | |
| "step": 64000 | |
| }, | |
| { | |
| "epoch": 1.89, | |
| "learning_rate": 2.1736840277375814e-05, | |
| "loss": 3.0328, | |
| "step": 64100 | |
| }, | |
| { | |
| "epoch": 1.9, | |
| "learning_rate": 2.1678957178083144e-05, | |
| "loss": 3.0795, | |
| "step": 64200 | |
| }, | |
| { | |
| "epoch": 1.9, | |
| "learning_rate": 2.1621074078790477e-05, | |
| "loss": 3.0247, | |
| "step": 64300 | |
| }, | |
| { | |
| "epoch": 1.9, | |
| "learning_rate": 2.1563190979497807e-05, | |
| "loss": 3.0352, | |
| "step": 64400 | |
| }, | |
| { | |
| "epoch": 1.9, | |
| "learning_rate": 2.1505307880205137e-05, | |
| "loss": 3.0503, | |
| "step": 64500 | |
| }, | |
| { | |
| "epoch": 1.91, | |
| "learning_rate": 2.144742478091247e-05, | |
| "loss": 3.0047, | |
| "step": 64600 | |
| }, | |
| { | |
| "epoch": 1.91, | |
| "learning_rate": 2.13895416816198e-05, | |
| "loss": 3.0388, | |
| "step": 64700 | |
| }, | |
| { | |
| "epoch": 1.91, | |
| "learning_rate": 2.133165858232713e-05, | |
| "loss": 3.0377, | |
| "step": 64800 | |
| }, | |
| { | |
| "epoch": 1.92, | |
| "learning_rate": 2.1273775483034465e-05, | |
| "loss": 3.0398, | |
| "step": 64900 | |
| }, | |
| { | |
| "epoch": 1.92, | |
| "learning_rate": 2.1215892383741795e-05, | |
| "loss": 3.0546, | |
| "step": 65000 | |
| }, | |
| { | |
| "epoch": 1.92, | |
| "learning_rate": 2.1158009284449125e-05, | |
| "loss": 3.019, | |
| "step": 65100 | |
| }, | |
| { | |
| "epoch": 1.92, | |
| "learning_rate": 2.110012618515646e-05, | |
| "loss": 3.0192, | |
| "step": 65200 | |
| }, | |
| { | |
| "epoch": 1.93, | |
| "learning_rate": 2.1042243085863792e-05, | |
| "loss": 2.9797, | |
| "step": 65300 | |
| }, | |
| { | |
| "epoch": 1.93, | |
| "learning_rate": 2.0984359986571122e-05, | |
| "loss": 3.1248, | |
| "step": 65400 | |
| }, | |
| { | |
| "epoch": 1.93, | |
| "learning_rate": 2.0926476887278455e-05, | |
| "loss": 3.0729, | |
| "step": 65500 | |
| }, | |
| { | |
| "epoch": 1.94, | |
| "learning_rate": 2.0868593787985785e-05, | |
| "loss": 3.0108, | |
| "step": 65600 | |
| }, | |
| { | |
| "epoch": 1.94, | |
| "learning_rate": 2.0810710688693116e-05, | |
| "loss": 2.9739, | |
| "step": 65700 | |
| }, | |
| { | |
| "epoch": 1.94, | |
| "learning_rate": 2.075282758940045e-05, | |
| "loss": 3.058, | |
| "step": 65800 | |
| }, | |
| { | |
| "epoch": 1.95, | |
| "learning_rate": 2.069494449010778e-05, | |
| "loss": 3.0111, | |
| "step": 65900 | |
| }, | |
| { | |
| "epoch": 1.95, | |
| "learning_rate": 2.063706139081511e-05, | |
| "loss": 3.075, | |
| "step": 66000 | |
| }, | |
| { | |
| "epoch": 1.95, | |
| "learning_rate": 2.0579178291522443e-05, | |
| "loss": 3.0493, | |
| "step": 66100 | |
| }, | |
| { | |
| "epoch": 1.95, | |
| "learning_rate": 2.0521295192229773e-05, | |
| "loss": 2.9718, | |
| "step": 66200 | |
| }, | |
| { | |
| "epoch": 1.96, | |
| "learning_rate": 2.0463412092937106e-05, | |
| "loss": 3.0471, | |
| "step": 66300 | |
| }, | |
| { | |
| "epoch": 1.96, | |
| "learning_rate": 2.0405528993644436e-05, | |
| "loss": 3.0502, | |
| "step": 66400 | |
| }, | |
| { | |
| "epoch": 1.96, | |
| "learning_rate": 2.0348803556337622e-05, | |
| "loss": 3.0482, | |
| "step": 66500 | |
| }, | |
| { | |
| "epoch": 1.97, | |
| "learning_rate": 2.0290920457044955e-05, | |
| "loss": 3.0391, | |
| "step": 66600 | |
| }, | |
| { | |
| "epoch": 1.97, | |
| "learning_rate": 2.0233037357752285e-05, | |
| "loss": 2.9687, | |
| "step": 66700 | |
| }, | |
| { | |
| "epoch": 1.97, | |
| "learning_rate": 2.0175154258459616e-05, | |
| "loss": 3.0298, | |
| "step": 66800 | |
| }, | |
| { | |
| "epoch": 1.97, | |
| "learning_rate": 2.011727115916695e-05, | |
| "loss": 3.0962, | |
| "step": 66900 | |
| }, | |
| { | |
| "epoch": 1.98, | |
| "learning_rate": 2.005938805987428e-05, | |
| "loss": 3.0824, | |
| "step": 67000 | |
| }, | |
| { | |
| "epoch": 1.98, | |
| "learning_rate": 2.000150496058161e-05, | |
| "loss": 3.0595, | |
| "step": 67100 | |
| }, | |
| { | |
| "epoch": 1.98, | |
| "learning_rate": 1.9943621861288943e-05, | |
| "loss": 2.9599, | |
| "step": 67200 | |
| }, | |
| { | |
| "epoch": 1.99, | |
| "learning_rate": 1.9885738761996273e-05, | |
| "loss": 2.9756, | |
| "step": 67300 | |
| }, | |
| { | |
| "epoch": 1.99, | |
| "learning_rate": 1.9827855662703603e-05, | |
| "loss": 2.9989, | |
| "step": 67400 | |
| }, | |
| { | |
| "epoch": 1.99, | |
| "learning_rate": 1.9769972563410936e-05, | |
| "loss": 3.0489, | |
| "step": 67500 | |
| }, | |
| { | |
| "epoch": 2.0, | |
| "learning_rate": 1.971208946411827e-05, | |
| "loss": 3.0045, | |
| "step": 67600 | |
| }, | |
| { | |
| "epoch": 2.0, | |
| "learning_rate": 1.96542063648256e-05, | |
| "loss": 2.9764, | |
| "step": 67700 | |
| }, | |
| { | |
| "epoch": 2.0, | |
| "eval_gen_len": 19.9174, | |
| "eval_loss": 3.12784481048584, | |
| "eval_rouge1": 18.6558, | |
| "eval_rouge2": 5.1844, | |
| "eval_rougeL": 15.0939, | |
| "eval_rougeLsum": 16.3367, | |
| "eval_runtime": 747.2706, | |
| "eval_samples_per_second": 10.073, | |
| "eval_steps_per_second": 2.518, | |
| "step": 67750 | |
| }, | |
| { | |
| "epoch": 2.0, | |
| "learning_rate": 1.959632326553293e-05, | |
| "loss": 2.7806, | |
| "step": 67800 | |
| }, | |
| { | |
| "epoch": 2.0, | |
| "learning_rate": 1.9538440166240264e-05, | |
| "loss": 2.6776, | |
| "step": 67900 | |
| }, | |
| { | |
| "epoch": 2.01, | |
| "learning_rate": 1.9480557066947594e-05, | |
| "loss": 2.6714, | |
| "step": 68000 | |
| }, | |
| { | |
| "epoch": 2.01, | |
| "learning_rate": 1.9422673967654924e-05, | |
| "loss": 2.6082, | |
| "step": 68100 | |
| }, | |
| { | |
| "epoch": 2.01, | |
| "learning_rate": 1.9364790868362257e-05, | |
| "loss": 2.6635, | |
| "step": 68200 | |
| }, | |
| { | |
| "epoch": 2.02, | |
| "learning_rate": 1.9306907769069587e-05, | |
| "loss": 2.6254, | |
| "step": 68300 | |
| }, | |
| { | |
| "epoch": 2.02, | |
| "learning_rate": 1.9249024669776918e-05, | |
| "loss": 2.6649, | |
| "step": 68400 | |
| }, | |
| { | |
| "epoch": 2.02, | |
| "learning_rate": 1.919114157048425e-05, | |
| "loss": 2.6291, | |
| "step": 68500 | |
| }, | |
| { | |
| "epoch": 2.03, | |
| "learning_rate": 1.913325847119158e-05, | |
| "loss": 2.6979, | |
| "step": 68600 | |
| }, | |
| { | |
| "epoch": 2.03, | |
| "learning_rate": 1.907537537189891e-05, | |
| "loss": 2.6635, | |
| "step": 68700 | |
| }, | |
| { | |
| "epoch": 2.03, | |
| "learning_rate": 1.9017492272606248e-05, | |
| "loss": 2.6288, | |
| "step": 68800 | |
| }, | |
| { | |
| "epoch": 2.03, | |
| "learning_rate": 1.8959609173313578e-05, | |
| "loss": 2.6401, | |
| "step": 68900 | |
| }, | |
| { | |
| "epoch": 2.04, | |
| "learning_rate": 1.890172607402091e-05, | |
| "loss": 2.6152, | |
| "step": 69000 | |
| }, | |
| { | |
| "epoch": 2.04, | |
| "learning_rate": 1.8843842974728242e-05, | |
| "loss": 2.6912, | |
| "step": 69100 | |
| }, | |
| { | |
| "epoch": 2.04, | |
| "learning_rate": 1.8785959875435572e-05, | |
| "loss": 2.6413, | |
| "step": 69200 | |
| }, | |
| { | |
| "epoch": 2.05, | |
| "learning_rate": 1.8728076776142902e-05, | |
| "loss": 2.6168, | |
| "step": 69300 | |
| }, | |
| { | |
| "epoch": 2.05, | |
| "learning_rate": 1.8670193676850235e-05, | |
| "loss": 2.661, | |
| "step": 69400 | |
| }, | |
| { | |
| "epoch": 2.05, | |
| "learning_rate": 1.8612310577557566e-05, | |
| "loss": 2.6739, | |
| "step": 69500 | |
| }, | |
| { | |
| "epoch": 2.05, | |
| "learning_rate": 1.8554427478264896e-05, | |
| "loss": 2.6745, | |
| "step": 69600 | |
| }, | |
| { | |
| "epoch": 2.06, | |
| "learning_rate": 1.849654437897223e-05, | |
| "loss": 2.7476, | |
| "step": 69700 | |
| }, | |
| { | |
| "epoch": 2.06, | |
| "learning_rate": 1.843924011067249e-05, | |
| "loss": 2.6416, | |
| "step": 69800 | |
| }, | |
| { | |
| "epoch": 2.06, | |
| "learning_rate": 1.838135701137982e-05, | |
| "loss": 2.6431, | |
| "step": 69900 | |
| }, | |
| { | |
| "epoch": 2.07, | |
| "learning_rate": 1.832347391208715e-05, | |
| "loss": 2.6663, | |
| "step": 70000 | |
| }, | |
| { | |
| "epoch": 2.07, | |
| "learning_rate": 1.8265590812794482e-05, | |
| "loss": 2.5952, | |
| "step": 70100 | |
| }, | |
| { | |
| "epoch": 2.07, | |
| "learning_rate": 1.8207707713501812e-05, | |
| "loss": 2.6954, | |
| "step": 70200 | |
| }, | |
| { | |
| "epoch": 2.08, | |
| "learning_rate": 1.8149824614209143e-05, | |
| "loss": 2.6513, | |
| "step": 70300 | |
| }, | |
| { | |
| "epoch": 2.08, | |
| "learning_rate": 1.8091941514916476e-05, | |
| "loss": 2.6924, | |
| "step": 70400 | |
| }, | |
| { | |
| "epoch": 2.08, | |
| "learning_rate": 1.8034058415623806e-05, | |
| "loss": 2.7011, | |
| "step": 70500 | |
| }, | |
| { | |
| "epoch": 2.08, | |
| "learning_rate": 1.797617531633114e-05, | |
| "loss": 2.734, | |
| "step": 70600 | |
| }, | |
| { | |
| "epoch": 2.09, | |
| "learning_rate": 1.791829221703847e-05, | |
| "loss": 2.6728, | |
| "step": 70700 | |
| }, | |
| { | |
| "epoch": 2.09, | |
| "learning_rate": 1.7860409117745803e-05, | |
| "loss": 2.6891, | |
| "step": 70800 | |
| }, | |
| { | |
| "epoch": 2.09, | |
| "learning_rate": 1.7802526018453133e-05, | |
| "loss": 2.6044, | |
| "step": 70900 | |
| }, | |
| { | |
| "epoch": 2.1, | |
| "learning_rate": 1.7744642919160463e-05, | |
| "loss": 2.7256, | |
| "step": 71000 | |
| }, | |
| { | |
| "epoch": 2.1, | |
| "learning_rate": 1.7686759819867797e-05, | |
| "loss": 2.5696, | |
| "step": 71100 | |
| }, | |
| { | |
| "epoch": 2.1, | |
| "learning_rate": 1.7628876720575127e-05, | |
| "loss": 2.6064, | |
| "step": 71200 | |
| }, | |
| { | |
| "epoch": 2.1, | |
| "learning_rate": 1.7570993621282457e-05, | |
| "loss": 2.6711, | |
| "step": 71300 | |
| }, | |
| { | |
| "epoch": 2.11, | |
| "learning_rate": 1.751311052198979e-05, | |
| "loss": 2.6586, | |
| "step": 71400 | |
| }, | |
| { | |
| "epoch": 2.11, | |
| "learning_rate": 1.745522742269712e-05, | |
| "loss": 2.6367, | |
| "step": 71500 | |
| }, | |
| { | |
| "epoch": 2.11, | |
| "learning_rate": 1.739734432340445e-05, | |
| "loss": 2.5933, | |
| "step": 71600 | |
| }, | |
| { | |
| "epoch": 2.12, | |
| "learning_rate": 1.7339461224111784e-05, | |
| "loss": 2.7044, | |
| "step": 71700 | |
| }, | |
| { | |
| "epoch": 2.12, | |
| "learning_rate": 1.7281578124819118e-05, | |
| "loss": 2.6311, | |
| "step": 71800 | |
| }, | |
| { | |
| "epoch": 2.12, | |
| "learning_rate": 1.7223695025526448e-05, | |
| "loss": 2.6212, | |
| "step": 71900 | |
| }, | |
| { | |
| "epoch": 2.13, | |
| "learning_rate": 1.7165811926233778e-05, | |
| "loss": 2.6314, | |
| "step": 72000 | |
| }, | |
| { | |
| "epoch": 2.13, | |
| "learning_rate": 1.710792882694111e-05, | |
| "loss": 2.7652, | |
| "step": 72100 | |
| }, | |
| { | |
| "epoch": 2.13, | |
| "learning_rate": 1.705004572764844e-05, | |
| "loss": 2.6811, | |
| "step": 72200 | |
| }, | |
| { | |
| "epoch": 2.13, | |
| "learning_rate": 1.6992741459348698e-05, | |
| "loss": 2.6055, | |
| "step": 72300 | |
| }, | |
| { | |
| "epoch": 2.14, | |
| "learning_rate": 1.693485836005603e-05, | |
| "loss": 2.6991, | |
| "step": 72400 | |
| }, | |
| { | |
| "epoch": 2.14, | |
| "learning_rate": 1.6876975260763365e-05, | |
| "loss": 2.6355, | |
| "step": 72500 | |
| }, | |
| { | |
| "epoch": 2.14, | |
| "learning_rate": 1.6819092161470695e-05, | |
| "loss": 2.6464, | |
| "step": 72600 | |
| }, | |
| { | |
| "epoch": 2.15, | |
| "learning_rate": 1.6761209062178025e-05, | |
| "loss": 2.6495, | |
| "step": 72700 | |
| }, | |
| { | |
| "epoch": 2.15, | |
| "learning_rate": 1.670332596288536e-05, | |
| "loss": 2.6867, | |
| "step": 72800 | |
| }, | |
| { | |
| "epoch": 2.15, | |
| "learning_rate": 1.664544286359269e-05, | |
| "loss": 2.6519, | |
| "step": 72900 | |
| }, | |
| { | |
| "epoch": 2.15, | |
| "learning_rate": 1.658755976430002e-05, | |
| "loss": 2.6794, | |
| "step": 73000 | |
| }, | |
| { | |
| "epoch": 2.16, | |
| "learning_rate": 1.6529676665007352e-05, | |
| "loss": 2.6281, | |
| "step": 73100 | |
| }, | |
| { | |
| "epoch": 2.16, | |
| "learning_rate": 1.6471793565714682e-05, | |
| "loss": 2.6679, | |
| "step": 73200 | |
| }, | |
| { | |
| "epoch": 2.16, | |
| "learning_rate": 1.6413910466422012e-05, | |
| "loss": 2.6788, | |
| "step": 73300 | |
| }, | |
| { | |
| "epoch": 2.17, | |
| "learning_rate": 1.6356027367129346e-05, | |
| "loss": 2.6201, | |
| "step": 73400 | |
| }, | |
| { | |
| "epoch": 2.17, | |
| "learning_rate": 1.629814426783668e-05, | |
| "loss": 2.6763, | |
| "step": 73500 | |
| }, | |
| { | |
| "epoch": 2.17, | |
| "learning_rate": 1.624026116854401e-05, | |
| "loss": 2.6572, | |
| "step": 73600 | |
| }, | |
| { | |
| "epoch": 2.18, | |
| "learning_rate": 1.6182378069251343e-05, | |
| "loss": 2.6733, | |
| "step": 73700 | |
| }, | |
| { | |
| "epoch": 2.18, | |
| "learning_rate": 1.6124494969958673e-05, | |
| "loss": 2.6219, | |
| "step": 73800 | |
| }, | |
| { | |
| "epoch": 2.18, | |
| "learning_rate": 1.6066611870666003e-05, | |
| "loss": 2.5916, | |
| "step": 73900 | |
| }, | |
| { | |
| "epoch": 2.18, | |
| "learning_rate": 1.6008728771373337e-05, | |
| "loss": 2.689, | |
| "step": 74000 | |
| }, | |
| { | |
| "epoch": 2.19, | |
| "learning_rate": 1.5950845672080667e-05, | |
| "loss": 2.6972, | |
| "step": 74100 | |
| }, | |
| { | |
| "epoch": 2.19, | |
| "learning_rate": 1.5892962572787997e-05, | |
| "loss": 2.6629, | |
| "step": 74200 | |
| }, | |
| { | |
| "epoch": 2.19, | |
| "learning_rate": 1.583507947349533e-05, | |
| "loss": 2.6508, | |
| "step": 74300 | |
| }, | |
| { | |
| "epoch": 2.2, | |
| "learning_rate": 1.577719637420266e-05, | |
| "loss": 2.7088, | |
| "step": 74400 | |
| }, | |
| { | |
| "epoch": 2.2, | |
| "learning_rate": 1.571931327490999e-05, | |
| "loss": 2.644, | |
| "step": 74500 | |
| }, | |
| { | |
| "epoch": 2.2, | |
| "learning_rate": 1.5661430175617324e-05, | |
| "loss": 2.6248, | |
| "step": 74600 | |
| }, | |
| { | |
| "epoch": 2.21, | |
| "learning_rate": 1.5603547076324657e-05, | |
| "loss": 2.5826, | |
| "step": 74700 | |
| }, | |
| { | |
| "epoch": 2.21, | |
| "learning_rate": 1.5545663977031988e-05, | |
| "loss": 2.6165, | |
| "step": 74800 | |
| }, | |
| { | |
| "epoch": 2.21, | |
| "learning_rate": 1.5487780877739318e-05, | |
| "loss": 2.6644, | |
| "step": 74900 | |
| }, | |
| { | |
| "epoch": 2.21, | |
| "learning_rate": 1.542989777844665e-05, | |
| "loss": 2.7307, | |
| "step": 75000 | |
| }, | |
| { | |
| "epoch": 2.22, | |
| "learning_rate": 1.537201467915398e-05, | |
| "loss": 2.664, | |
| "step": 75100 | |
| }, | |
| { | |
| "epoch": 2.22, | |
| "learning_rate": 1.531413157986131e-05, | |
| "loss": 2.6484, | |
| "step": 75200 | |
| }, | |
| { | |
| "epoch": 2.22, | |
| "learning_rate": 1.5256248480568645e-05, | |
| "loss": 2.6445, | |
| "step": 75300 | |
| }, | |
| { | |
| "epoch": 2.23, | |
| "learning_rate": 1.5198365381275975e-05, | |
| "loss": 2.6681, | |
| "step": 75400 | |
| }, | |
| { | |
| "epoch": 2.23, | |
| "learning_rate": 1.5140482281983307e-05, | |
| "loss": 2.6783, | |
| "step": 75500 | |
| }, | |
| { | |
| "epoch": 2.23, | |
| "learning_rate": 1.5082599182690638e-05, | |
| "loss": 2.6748, | |
| "step": 75600 | |
| }, | |
| { | |
| "epoch": 2.23, | |
| "learning_rate": 1.5024716083397969e-05, | |
| "loss": 2.6884, | |
| "step": 75700 | |
| }, | |
| { | |
| "epoch": 2.24, | |
| "learning_rate": 1.4966832984105302e-05, | |
| "loss": 2.6373, | |
| "step": 75800 | |
| }, | |
| { | |
| "epoch": 2.24, | |
| "learning_rate": 1.4908949884812634e-05, | |
| "loss": 2.6803, | |
| "step": 75900 | |
| }, | |
| { | |
| "epoch": 2.24, | |
| "learning_rate": 1.4851066785519966e-05, | |
| "loss": 2.6763, | |
| "step": 76000 | |
| }, | |
| { | |
| "epoch": 2.25, | |
| "learning_rate": 1.4793183686227296e-05, | |
| "loss": 2.6724, | |
| "step": 76100 | |
| }, | |
| { | |
| "epoch": 2.25, | |
| "learning_rate": 1.4735300586934628e-05, | |
| "loss": 2.6505, | |
| "step": 76200 | |
| }, | |
| { | |
| "epoch": 2.25, | |
| "learning_rate": 1.467741748764196e-05, | |
| "loss": 2.6467, | |
| "step": 76300 | |
| }, | |
| { | |
| "epoch": 2.26, | |
| "learning_rate": 1.4620113219342215e-05, | |
| "loss": 2.6609, | |
| "step": 76400 | |
| }, | |
| { | |
| "epoch": 2.26, | |
| "learning_rate": 1.4562230120049549e-05, | |
| "loss": 2.6014, | |
| "step": 76500 | |
| }, | |
| { | |
| "epoch": 2.26, | |
| "learning_rate": 1.450434702075688e-05, | |
| "loss": 2.6619, | |
| "step": 76600 | |
| }, | |
| { | |
| "epoch": 2.26, | |
| "learning_rate": 1.4446463921464213e-05, | |
| "loss": 2.6916, | |
| "step": 76700 | |
| }, | |
| { | |
| "epoch": 2.27, | |
| "learning_rate": 1.4388580822171543e-05, | |
| "loss": 2.7405, | |
| "step": 76800 | |
| }, | |
| { | |
| "epoch": 2.27, | |
| "learning_rate": 1.4330697722878874e-05, | |
| "loss": 2.6346, | |
| "step": 76900 | |
| }, | |
| { | |
| "epoch": 2.27, | |
| "learning_rate": 1.4272814623586206e-05, | |
| "loss": 2.6796, | |
| "step": 77000 | |
| }, | |
| { | |
| "epoch": 2.28, | |
| "learning_rate": 1.4214931524293536e-05, | |
| "loss": 2.6257, | |
| "step": 77100 | |
| }, | |
| { | |
| "epoch": 2.28, | |
| "learning_rate": 1.4157048425000868e-05, | |
| "loss": 2.622, | |
| "step": 77200 | |
| }, | |
| { | |
| "epoch": 2.28, | |
| "learning_rate": 1.40991653257082e-05, | |
| "loss": 2.6802, | |
| "step": 77300 | |
| }, | |
| { | |
| "epoch": 2.28, | |
| "learning_rate": 1.404128222641553e-05, | |
| "loss": 2.6289, | |
| "step": 77400 | |
| }, | |
| { | |
| "epoch": 2.29, | |
| "learning_rate": 1.3983399127122862e-05, | |
| "loss": 2.6751, | |
| "step": 77500 | |
| }, | |
| { | |
| "epoch": 2.29, | |
| "learning_rate": 1.3925516027830195e-05, | |
| "loss": 2.6744, | |
| "step": 77600 | |
| }, | |
| { | |
| "epoch": 2.29, | |
| "learning_rate": 1.3867632928537527e-05, | |
| "loss": 2.6766, | |
| "step": 77700 | |
| }, | |
| { | |
| "epoch": 2.3, | |
| "learning_rate": 1.3809749829244859e-05, | |
| "loss": 2.7145, | |
| "step": 77800 | |
| }, | |
| { | |
| "epoch": 2.3, | |
| "learning_rate": 1.3751866729952189e-05, | |
| "loss": 2.727, | |
| "step": 77900 | |
| }, | |
| { | |
| "epoch": 2.3, | |
| "learning_rate": 1.369398363065952e-05, | |
| "loss": 2.647, | |
| "step": 78000 | |
| }, | |
| { | |
| "epoch": 2.31, | |
| "learning_rate": 1.3636100531366853e-05, | |
| "loss": 2.7302, | |
| "step": 78100 | |
| }, | |
| { | |
| "epoch": 2.31, | |
| "learning_rate": 1.3578217432074183e-05, | |
| "loss": 2.7183, | |
| "step": 78200 | |
| }, | |
| { | |
| "epoch": 2.31, | |
| "learning_rate": 1.3520334332781515e-05, | |
| "loss": 2.5994, | |
| "step": 78300 | |
| }, | |
| { | |
| "epoch": 2.31, | |
| "learning_rate": 1.3462451233488846e-05, | |
| "loss": 2.632, | |
| "step": 78400 | |
| }, | |
| { | |
| "epoch": 2.32, | |
| "learning_rate": 1.3404568134196176e-05, | |
| "loss": 2.6969, | |
| "step": 78500 | |
| }, | |
| { | |
| "epoch": 2.32, | |
| "learning_rate": 1.3346685034903508e-05, | |
| "loss": 2.6856, | |
| "step": 78600 | |
| }, | |
| { | |
| "epoch": 2.32, | |
| "learning_rate": 1.3289380766603768e-05, | |
| "loss": 2.6106, | |
| "step": 78700 | |
| }, | |
| { | |
| "epoch": 2.33, | |
| "learning_rate": 1.32314976673111e-05, | |
| "loss": 2.642, | |
| "step": 78800 | |
| }, | |
| { | |
| "epoch": 2.33, | |
| "learning_rate": 1.317361456801843e-05, | |
| "loss": 2.6484, | |
| "step": 78900 | |
| }, | |
| { | |
| "epoch": 2.33, | |
| "learning_rate": 1.3115731468725761e-05, | |
| "loss": 2.6598, | |
| "step": 79000 | |
| }, | |
| { | |
| "epoch": 2.34, | |
| "learning_rate": 1.3057848369433093e-05, | |
| "loss": 2.5861, | |
| "step": 79100 | |
| }, | |
| { | |
| "epoch": 2.34, | |
| "learning_rate": 1.2999965270140423e-05, | |
| "loss": 2.6432, | |
| "step": 79200 | |
| }, | |
| { | |
| "epoch": 2.34, | |
| "learning_rate": 1.2942082170847755e-05, | |
| "loss": 2.6069, | |
| "step": 79300 | |
| }, | |
| { | |
| "epoch": 2.34, | |
| "learning_rate": 1.2884199071555089e-05, | |
| "loss": 2.7582, | |
| "step": 79400 | |
| }, | |
| { | |
| "epoch": 2.35, | |
| "learning_rate": 1.282631597226242e-05, | |
| "loss": 2.7005, | |
| "step": 79500 | |
| }, | |
| { | |
| "epoch": 2.35, | |
| "learning_rate": 1.2768432872969752e-05, | |
| "loss": 2.5888, | |
| "step": 79600 | |
| }, | |
| { | |
| "epoch": 2.35, | |
| "learning_rate": 1.2711128604670008e-05, | |
| "loss": 2.7044, | |
| "step": 79700 | |
| }, | |
| { | |
| "epoch": 2.36, | |
| "learning_rate": 1.265324550537734e-05, | |
| "loss": 2.6257, | |
| "step": 79800 | |
| }, | |
| { | |
| "epoch": 2.36, | |
| "learning_rate": 1.259536240608467e-05, | |
| "loss": 2.6742, | |
| "step": 79900 | |
| }, | |
| { | |
| "epoch": 2.36, | |
| "learning_rate": 1.2537479306792002e-05, | |
| "loss": 2.675, | |
| "step": 80000 | |
| }, | |
| { | |
| "epoch": 2.36, | |
| "learning_rate": 1.2479596207499335e-05, | |
| "loss": 2.6145, | |
| "step": 80100 | |
| }, | |
| { | |
| "epoch": 2.37, | |
| "learning_rate": 1.2421713108206666e-05, | |
| "loss": 2.68, | |
| "step": 80200 | |
| }, | |
| { | |
| "epoch": 2.37, | |
| "learning_rate": 1.2363830008913999e-05, | |
| "loss": 2.6717, | |
| "step": 80300 | |
| }, | |
| { | |
| "epoch": 2.37, | |
| "learning_rate": 1.230594690962133e-05, | |
| "loss": 2.6563, | |
| "step": 80400 | |
| }, | |
| { | |
| "epoch": 2.38, | |
| "learning_rate": 1.2248063810328661e-05, | |
| "loss": 2.7272, | |
| "step": 80500 | |
| }, | |
| { | |
| "epoch": 2.38, | |
| "learning_rate": 1.2190180711035993e-05, | |
| "loss": 2.6457, | |
| "step": 80600 | |
| }, | |
| { | |
| "epoch": 2.38, | |
| "learning_rate": 1.2132297611743323e-05, | |
| "loss": 2.7066, | |
| "step": 80700 | |
| }, | |
| { | |
| "epoch": 2.39, | |
| "learning_rate": 1.2074414512450655e-05, | |
| "loss": 2.6104, | |
| "step": 80800 | |
| }, | |
| { | |
| "epoch": 2.39, | |
| "learning_rate": 1.2016531413157988e-05, | |
| "loss": 2.6494, | |
| "step": 80900 | |
| }, | |
| { | |
| "epoch": 2.39, | |
| "learning_rate": 1.1958648313865318e-05, | |
| "loss": 2.656, | |
| "step": 81000 | |
| }, | |
| { | |
| "epoch": 2.39, | |
| "learning_rate": 1.190076521457265e-05, | |
| "loss": 2.7509, | |
| "step": 81100 | |
| }, | |
| { | |
| "epoch": 2.4, | |
| "learning_rate": 1.184288211527998e-05, | |
| "loss": 2.6835, | |
| "step": 81200 | |
| }, | |
| { | |
| "epoch": 2.4, | |
| "learning_rate": 1.1784999015987312e-05, | |
| "loss": 2.6426, | |
| "step": 81300 | |
| }, | |
| { | |
| "epoch": 2.4, | |
| "learning_rate": 1.1727115916694644e-05, | |
| "loss": 2.6237, | |
| "step": 81400 | |
| }, | |
| { | |
| "epoch": 2.41, | |
| "learning_rate": 1.1669232817401976e-05, | |
| "loss": 2.6958, | |
| "step": 81500 | |
| }, | |
| { | |
| "epoch": 2.41, | |
| "learning_rate": 1.1611349718109307e-05, | |
| "loss": 2.6218, | |
| "step": 81600 | |
| }, | |
| { | |
| "epoch": 2.41, | |
| "learning_rate": 1.1553466618816639e-05, | |
| "loss": 2.5956, | |
| "step": 81700 | |
| }, | |
| { | |
| "epoch": 2.41, | |
| "learning_rate": 1.149558351952397e-05, | |
| "loss": 2.6071, | |
| "step": 81800 | |
| }, | |
| { | |
| "epoch": 2.42, | |
| "learning_rate": 1.1437700420231301e-05, | |
| "loss": 2.6936, | |
| "step": 81900 | |
| }, | |
| { | |
| "epoch": 2.42, | |
| "learning_rate": 1.1379817320938633e-05, | |
| "loss": 2.7179, | |
| "step": 82000 | |
| }, | |
| { | |
| "epoch": 2.42, | |
| "learning_rate": 1.1321934221645965e-05, | |
| "loss": 2.7001, | |
| "step": 82100 | |
| }, | |
| { | |
| "epoch": 2.43, | |
| "learning_rate": 1.1264051122353296e-05, | |
| "loss": 2.5915, | |
| "step": 82200 | |
| }, | |
| { | |
| "epoch": 2.43, | |
| "learning_rate": 1.1206168023060626e-05, | |
| "loss": 2.6884, | |
| "step": 82300 | |
| }, | |
| { | |
| "epoch": 2.43, | |
| "learning_rate": 1.1148284923767958e-05, | |
| "loss": 2.647, | |
| "step": 82400 | |
| }, | |
| { | |
| "epoch": 2.44, | |
| "learning_rate": 1.109040182447529e-05, | |
| "loss": 2.6499, | |
| "step": 82500 | |
| }, | |
| { | |
| "epoch": 2.44, | |
| "learning_rate": 1.1032518725182622e-05, | |
| "loss": 2.654, | |
| "step": 82600 | |
| }, | |
| { | |
| "epoch": 2.44, | |
| "learning_rate": 1.0974635625889954e-05, | |
| "loss": 2.6469, | |
| "step": 82700 | |
| }, | |
| { | |
| "epoch": 2.44, | |
| "learning_rate": 1.0916752526597285e-05, | |
| "loss": 2.6167, | |
| "step": 82800 | |
| }, | |
| { | |
| "epoch": 2.45, | |
| "learning_rate": 1.0858869427304616e-05, | |
| "loss": 2.6403, | |
| "step": 82900 | |
| }, | |
| { | |
| "epoch": 2.45, | |
| "learning_rate": 1.0800986328011947e-05, | |
| "loss": 2.632, | |
| "step": 83000 | |
| }, | |
| { | |
| "epoch": 2.45, | |
| "learning_rate": 1.074310322871928e-05, | |
| "loss": 2.6953, | |
| "step": 83100 | |
| }, | |
| { | |
| "epoch": 2.46, | |
| "learning_rate": 1.0685220129426611e-05, | |
| "loss": 2.7353, | |
| "step": 83200 | |
| }, | |
| { | |
| "epoch": 2.46, | |
| "learning_rate": 1.0627337030133943e-05, | |
| "loss": 2.6768, | |
| "step": 83300 | |
| }, | |
| { | |
| "epoch": 2.46, | |
| "learning_rate": 1.0569453930841273e-05, | |
| "loss": 2.6685, | |
| "step": 83400 | |
| }, | |
| { | |
| "epoch": 2.46, | |
| "learning_rate": 1.0511570831548605e-05, | |
| "loss": 2.7646, | |
| "step": 83500 | |
| }, | |
| { | |
| "epoch": 2.47, | |
| "learning_rate": 1.0453687732255936e-05, | |
| "loss": 2.5976, | |
| "step": 83600 | |
| }, | |
| { | |
| "epoch": 2.47, | |
| "learning_rate": 1.0395804632963268e-05, | |
| "loss": 2.6466, | |
| "step": 83700 | |
| }, | |
| { | |
| "epoch": 2.47, | |
| "learning_rate": 1.03379215336706e-05, | |
| "loss": 2.7244, | |
| "step": 83800 | |
| }, | |
| { | |
| "epoch": 2.48, | |
| "learning_rate": 1.0280038434377932e-05, | |
| "loss": 2.6485, | |
| "step": 83900 | |
| }, | |
| { | |
| "epoch": 2.48, | |
| "learning_rate": 1.0222155335085262e-05, | |
| "loss": 2.6187, | |
| "step": 84000 | |
| }, | |
| { | |
| "epoch": 2.48, | |
| "learning_rate": 1.016485106678552e-05, | |
| "loss": 2.5851, | |
| "step": 84100 | |
| }, | |
| { | |
| "epoch": 2.49, | |
| "learning_rate": 1.0106967967492852e-05, | |
| "loss": 2.5945, | |
| "step": 84200 | |
| }, | |
| { | |
| "epoch": 2.49, | |
| "learning_rate": 1.0049084868200183e-05, | |
| "loss": 2.6874, | |
| "step": 84300 | |
| }, | |
| { | |
| "epoch": 2.49, | |
| "learning_rate": 9.991780599900441e-06, | |
| "loss": 2.6016, | |
| "step": 84400 | |
| }, | |
| { | |
| "epoch": 2.49, | |
| "learning_rate": 9.933897500607773e-06, | |
| "loss": 2.7371, | |
| "step": 84500 | |
| }, | |
| { | |
| "epoch": 2.5, | |
| "learning_rate": 9.876014401315105e-06, | |
| "loss": 2.6355, | |
| "step": 84600 | |
| }, | |
| { | |
| "epoch": 2.5, | |
| "learning_rate": 9.818131302022437e-06, | |
| "loss": 2.6629, | |
| "step": 84700 | |
| }, | |
| { | |
| "epoch": 2.5, | |
| "learning_rate": 9.760248202729767e-06, | |
| "loss": 2.6944, | |
| "step": 84800 | |
| }, | |
| { | |
| "epoch": 2.51, | |
| "learning_rate": 9.702365103437098e-06, | |
| "loss": 2.6812, | |
| "step": 84900 | |
| }, | |
| { | |
| "epoch": 2.51, | |
| "learning_rate": 9.64448200414443e-06, | |
| "loss": 2.6876, | |
| "step": 85000 | |
| }, | |
| { | |
| "epoch": 2.51, | |
| "learning_rate": 9.586598904851762e-06, | |
| "loss": 2.6602, | |
| "step": 85100 | |
| }, | |
| { | |
| "epoch": 2.52, | |
| "learning_rate": 9.528715805559094e-06, | |
| "loss": 2.6384, | |
| "step": 85200 | |
| }, | |
| { | |
| "epoch": 2.52, | |
| "learning_rate": 9.470832706266426e-06, | |
| "loss": 2.6552, | |
| "step": 85300 | |
| }, | |
| { | |
| "epoch": 2.52, | |
| "learning_rate": 9.412949606973756e-06, | |
| "loss": 2.6656, | |
| "step": 85400 | |
| }, | |
| { | |
| "epoch": 2.52, | |
| "learning_rate": 9.355066507681088e-06, | |
| "loss": 2.6523, | |
| "step": 85500 | |
| }, | |
| { | |
| "epoch": 2.53, | |
| "learning_rate": 9.29718340838842e-06, | |
| "loss": 2.741, | |
| "step": 85600 | |
| }, | |
| { | |
| "epoch": 2.53, | |
| "learning_rate": 9.239300309095751e-06, | |
| "loss": 2.6532, | |
| "step": 85700 | |
| }, | |
| { | |
| "epoch": 2.53, | |
| "learning_rate": 9.181417209803083e-06, | |
| "loss": 2.6828, | |
| "step": 85800 | |
| }, | |
| { | |
| "epoch": 2.54, | |
| "learning_rate": 9.123534110510413e-06, | |
| "loss": 2.7142, | |
| "step": 85900 | |
| }, | |
| { | |
| "epoch": 2.54, | |
| "learning_rate": 9.065651011217745e-06, | |
| "loss": 2.6404, | |
| "step": 86000 | |
| }, | |
| { | |
| "epoch": 2.54, | |
| "learning_rate": 9.007767911925077e-06, | |
| "loss": 2.661, | |
| "step": 86100 | |
| }, | |
| { | |
| "epoch": 2.54, | |
| "learning_rate": 8.949884812632408e-06, | |
| "loss": 2.676, | |
| "step": 86200 | |
| }, | |
| { | |
| "epoch": 2.55, | |
| "learning_rate": 8.89200171333974e-06, | |
| "loss": 2.6487, | |
| "step": 86300 | |
| }, | |
| { | |
| "epoch": 2.55, | |
| "learning_rate": 8.83411861404707e-06, | |
| "loss": 2.6235, | |
| "step": 86400 | |
| }, | |
| { | |
| "epoch": 2.55, | |
| "learning_rate": 8.776235514754402e-06, | |
| "loss": 2.6581, | |
| "step": 86500 | |
| }, | |
| { | |
| "epoch": 2.56, | |
| "learning_rate": 8.718352415461734e-06, | |
| "loss": 2.6467, | |
| "step": 86600 | |
| }, | |
| { | |
| "epoch": 2.56, | |
| "learning_rate": 8.660469316169064e-06, | |
| "loss": 2.6874, | |
| "step": 86700 | |
| }, | |
| { | |
| "epoch": 2.56, | |
| "learning_rate": 8.602586216876397e-06, | |
| "loss": 2.5841, | |
| "step": 86800 | |
| }, | |
| { | |
| "epoch": 2.57, | |
| "learning_rate": 8.54470311758373e-06, | |
| "loss": 2.6598, | |
| "step": 86900 | |
| }, | |
| { | |
| "epoch": 2.57, | |
| "learning_rate": 8.48682001829106e-06, | |
| "loss": 2.6999, | |
| "step": 87000 | |
| }, | |
| { | |
| "epoch": 2.57, | |
| "learning_rate": 8.428936918998391e-06, | |
| "loss": 2.6375, | |
| "step": 87100 | |
| }, | |
| { | |
| "epoch": 2.57, | |
| "learning_rate": 8.371053819705723e-06, | |
| "loss": 2.6541, | |
| "step": 87200 | |
| }, | |
| { | |
| "epoch": 2.58, | |
| "learning_rate": 8.313170720413053e-06, | |
| "loss": 2.6914, | |
| "step": 87300 | |
| }, | |
| { | |
| "epoch": 2.58, | |
| "learning_rate": 8.255287621120387e-06, | |
| "loss": 2.6972, | |
| "step": 87400 | |
| }, | |
| { | |
| "epoch": 2.58, | |
| "learning_rate": 8.197404521827717e-06, | |
| "loss": 2.6217, | |
| "step": 87500 | |
| }, | |
| { | |
| "epoch": 2.59, | |
| "learning_rate": 8.139521422535048e-06, | |
| "loss": 2.6914, | |
| "step": 87600 | |
| }, | |
| { | |
| "epoch": 2.59, | |
| "learning_rate": 8.08163832324238e-06, | |
| "loss": 2.6, | |
| "step": 87700 | |
| }, | |
| { | |
| "epoch": 2.59, | |
| "learning_rate": 8.02375522394971e-06, | |
| "loss": 2.6203, | |
| "step": 87800 | |
| }, | |
| { | |
| "epoch": 2.59, | |
| "learning_rate": 7.965872124657044e-06, | |
| "loss": 2.6175, | |
| "step": 87900 | |
| }, | |
| { | |
| "epoch": 2.6, | |
| "learning_rate": 7.907989025364376e-06, | |
| "loss": 2.6322, | |
| "step": 88000 | |
| }, | |
| { | |
| "epoch": 2.6, | |
| "learning_rate": 7.850105926071706e-06, | |
| "loss": 2.6456, | |
| "step": 88100 | |
| }, | |
| { | |
| "epoch": 2.6, | |
| "learning_rate": 7.792222826779038e-06, | |
| "loss": 2.6561, | |
| "step": 88200 | |
| }, | |
| { | |
| "epoch": 2.61, | |
| "learning_rate": 7.73433972748637e-06, | |
| "loss": 2.5974, | |
| "step": 88300 | |
| }, | |
| { | |
| "epoch": 2.61, | |
| "learning_rate": 7.6764566281937e-06, | |
| "loss": 2.6547, | |
| "step": 88400 | |
| }, | |
| { | |
| "epoch": 2.61, | |
| "learning_rate": 7.618573528901032e-06, | |
| "loss": 2.6182, | |
| "step": 88500 | |
| }, | |
| { | |
| "epoch": 2.62, | |
| "learning_rate": 7.560690429608364e-06, | |
| "loss": 2.5736, | |
| "step": 88600 | |
| }, | |
| { | |
| "epoch": 2.62, | |
| "learning_rate": 7.502807330315695e-06, | |
| "loss": 2.6473, | |
| "step": 88700 | |
| }, | |
| { | |
| "epoch": 2.62, | |
| "learning_rate": 7.445503062015953e-06, | |
| "loss": 2.6361, | |
| "step": 88800 | |
| }, | |
| { | |
| "epoch": 2.62, | |
| "learning_rate": 7.387619962723284e-06, | |
| "loss": 2.6453, | |
| "step": 88900 | |
| }, | |
| { | |
| "epoch": 2.63, | |
| "learning_rate": 7.329736863430615e-06, | |
| "loss": 2.6865, | |
| "step": 89000 | |
| }, | |
| { | |
| "epoch": 2.63, | |
| "learning_rate": 7.271853764137946e-06, | |
| "loss": 2.5792, | |
| "step": 89100 | |
| }, | |
| { | |
| "epoch": 2.63, | |
| "learning_rate": 7.213970664845279e-06, | |
| "loss": 2.6216, | |
| "step": 89200 | |
| }, | |
| { | |
| "epoch": 2.64, | |
| "learning_rate": 7.156666396545538e-06, | |
| "loss": 2.6165, | |
| "step": 89300 | |
| }, | |
| { | |
| "epoch": 2.64, | |
| "learning_rate": 7.0987832972528685e-06, | |
| "loss": 2.6615, | |
| "step": 89400 | |
| }, | |
| { | |
| "epoch": 2.64, | |
| "learning_rate": 7.0409001979601995e-06, | |
| "loss": 2.6344, | |
| "step": 89500 | |
| }, | |
| { | |
| "epoch": 2.65, | |
| "learning_rate": 6.983017098667531e-06, | |
| "loss": 2.6097, | |
| "step": 89600 | |
| }, | |
| { | |
| "epoch": 2.65, | |
| "learning_rate": 6.925133999374862e-06, | |
| "loss": 2.6498, | |
| "step": 89700 | |
| }, | |
| { | |
| "epoch": 2.65, | |
| "learning_rate": 6.867250900082195e-06, | |
| "loss": 2.7191, | |
| "step": 89800 | |
| }, | |
| { | |
| "epoch": 2.65, | |
| "learning_rate": 6.809367800789526e-06, | |
| "loss": 2.6154, | |
| "step": 89900 | |
| }, | |
| { | |
| "epoch": 2.66, | |
| "learning_rate": 6.751484701496858e-06, | |
| "loss": 2.5856, | |
| "step": 90000 | |
| }, | |
| { | |
| "epoch": 2.66, | |
| "learning_rate": 6.6936016022041886e-06, | |
| "loss": 2.6058, | |
| "step": 90100 | |
| }, | |
| { | |
| "epoch": 2.66, | |
| "learning_rate": 6.6357185029115195e-06, | |
| "loss": 2.7033, | |
| "step": 90200 | |
| }, | |
| { | |
| "epoch": 2.67, | |
| "learning_rate": 6.577835403618851e-06, | |
| "loss": 2.6258, | |
| "step": 90300 | |
| }, | |
| { | |
| "epoch": 2.67, | |
| "learning_rate": 6.519952304326184e-06, | |
| "loss": 2.6289, | |
| "step": 90400 | |
| }, | |
| { | |
| "epoch": 2.67, | |
| "learning_rate": 6.462069205033515e-06, | |
| "loss": 2.6527, | |
| "step": 90500 | |
| }, | |
| { | |
| "epoch": 2.67, | |
| "learning_rate": 6.404186105740846e-06, | |
| "loss": 2.5685, | |
| "step": 90600 | |
| }, | |
| { | |
| "epoch": 2.68, | |
| "learning_rate": 6.346303006448178e-06, | |
| "loss": 2.6636, | |
| "step": 90700 | |
| }, | |
| { | |
| "epoch": 2.68, | |
| "learning_rate": 6.288419907155509e-06, | |
| "loss": 2.6443, | |
| "step": 90800 | |
| }, | |
| { | |
| "epoch": 2.68, | |
| "learning_rate": 6.23053680786284e-06, | |
| "loss": 2.6078, | |
| "step": 90900 | |
| }, | |
| { | |
| "epoch": 2.69, | |
| "learning_rate": 6.172653708570171e-06, | |
| "loss": 2.6654, | |
| "step": 91000 | |
| }, | |
| { | |
| "epoch": 2.69, | |
| "learning_rate": 6.114770609277504e-06, | |
| "loss": 2.6124, | |
| "step": 91100 | |
| }, | |
| { | |
| "epoch": 2.69, | |
| "learning_rate": 6.056887509984835e-06, | |
| "loss": 2.6233, | |
| "step": 91200 | |
| }, | |
| { | |
| "epoch": 2.7, | |
| "learning_rate": 5.999004410692166e-06, | |
| "loss": 2.614, | |
| "step": 91300 | |
| }, | |
| { | |
| "epoch": 2.7, | |
| "learning_rate": 5.941121311399498e-06, | |
| "loss": 2.6472, | |
| "step": 91400 | |
| }, | |
| { | |
| "epoch": 2.7, | |
| "learning_rate": 5.8832382121068295e-06, | |
| "loss": 2.6337, | |
| "step": 91500 | |
| }, | |
| { | |
| "epoch": 2.7, | |
| "learning_rate": 5.82535511281416e-06, | |
| "loss": 2.5766, | |
| "step": 91600 | |
| }, | |
| { | |
| "epoch": 2.71, | |
| "learning_rate": 5.767472013521492e-06, | |
| "loss": 2.6287, | |
| "step": 91700 | |
| }, | |
| { | |
| "epoch": 2.71, | |
| "learning_rate": 5.709588914228824e-06, | |
| "loss": 2.5166, | |
| "step": 91800 | |
| }, | |
| { | |
| "epoch": 2.71, | |
| "learning_rate": 5.651705814936156e-06, | |
| "loss": 2.6885, | |
| "step": 91900 | |
| }, | |
| { | |
| "epoch": 2.72, | |
| "learning_rate": 5.593822715643487e-06, | |
| "loss": 2.6536, | |
| "step": 92000 | |
| }, | |
| { | |
| "epoch": 2.72, | |
| "learning_rate": 5.535939616350818e-06, | |
| "loss": 2.6554, | |
| "step": 92100 | |
| }, | |
| { | |
| "epoch": 2.72, | |
| "learning_rate": 5.47805651705815e-06, | |
| "loss": 2.5428, | |
| "step": 92200 | |
| }, | |
| { | |
| "epoch": 2.72, | |
| "learning_rate": 5.420173417765481e-06, | |
| "loss": 2.6454, | |
| "step": 92300 | |
| }, | |
| { | |
| "epoch": 2.73, | |
| "learning_rate": 5.362290318472812e-06, | |
| "loss": 2.6419, | |
| "step": 92400 | |
| }, | |
| { | |
| "epoch": 2.73, | |
| "learning_rate": 5.304407219180144e-06, | |
| "loss": 2.6316, | |
| "step": 92500 | |
| }, | |
| { | |
| "epoch": 2.73, | |
| "learning_rate": 5.247102950880403e-06, | |
| "loss": 2.6203, | |
| "step": 92600 | |
| }, | |
| { | |
| "epoch": 2.74, | |
| "learning_rate": 5.189219851587734e-06, | |
| "loss": 2.604, | |
| "step": 92700 | |
| }, | |
| { | |
| "epoch": 2.74, | |
| "learning_rate": 5.131336752295065e-06, | |
| "loss": 2.6726, | |
| "step": 92800 | |
| }, | |
| { | |
| "epoch": 2.74, | |
| "learning_rate": 5.073453653002397e-06, | |
| "loss": 2.629, | |
| "step": 92900 | |
| }, | |
| { | |
| "epoch": 2.75, | |
| "learning_rate": 5.015570553709728e-06, | |
| "loss": 2.6562, | |
| "step": 93000 | |
| }, | |
| { | |
| "epoch": 2.75, | |
| "learning_rate": 4.957687454417059e-06, | |
| "loss": 2.637, | |
| "step": 93100 | |
| }, | |
| { | |
| "epoch": 2.75, | |
| "learning_rate": 4.899804355124391e-06, | |
| "loss": 2.6057, | |
| "step": 93200 | |
| }, | |
| { | |
| "epoch": 2.75, | |
| "learning_rate": 4.841921255831723e-06, | |
| "loss": 2.7188, | |
| "step": 93300 | |
| }, | |
| { | |
| "epoch": 2.76, | |
| "learning_rate": 4.784038156539054e-06, | |
| "loss": 2.6679, | |
| "step": 93400 | |
| }, | |
| { | |
| "epoch": 2.76, | |
| "learning_rate": 4.7261550572463855e-06, | |
| "loss": 2.658, | |
| "step": 93500 | |
| }, | |
| { | |
| "epoch": 2.76, | |
| "learning_rate": 4.668271957953716e-06, | |
| "loss": 2.7056, | |
| "step": 93600 | |
| }, | |
| { | |
| "epoch": 2.77, | |
| "learning_rate": 4.610388858661048e-06, | |
| "loss": 2.568, | |
| "step": 93700 | |
| }, | |
| { | |
| "epoch": 2.77, | |
| "learning_rate": 4.55250575936838e-06, | |
| "loss": 2.6135, | |
| "step": 93800 | |
| }, | |
| { | |
| "epoch": 2.77, | |
| "learning_rate": 4.494622660075711e-06, | |
| "loss": 2.6412, | |
| "step": 93900 | |
| }, | |
| { | |
| "epoch": 2.77, | |
| "learning_rate": 4.436739560783043e-06, | |
| "loss": 2.6006, | |
| "step": 94000 | |
| }, | |
| { | |
| "epoch": 2.78, | |
| "learning_rate": 4.3788564614903745e-06, | |
| "loss": 2.6539, | |
| "step": 94100 | |
| }, | |
| { | |
| "epoch": 2.78, | |
| "learning_rate": 4.3209733621977055e-06, | |
| "loss": 2.5896, | |
| "step": 94200 | |
| }, | |
| { | |
| "epoch": 2.78, | |
| "learning_rate": 4.263090262905037e-06, | |
| "loss": 2.6026, | |
| "step": 94300 | |
| }, | |
| { | |
| "epoch": 2.79, | |
| "learning_rate": 4.205207163612369e-06, | |
| "loss": 2.6195, | |
| "step": 94400 | |
| }, | |
| { | |
| "epoch": 2.79, | |
| "learning_rate": 4.1473240643197e-06, | |
| "loss": 2.6127, | |
| "step": 94500 | |
| }, | |
| { | |
| "epoch": 2.79, | |
| "learning_rate": 4.089440965027032e-06, | |
| "loss": 2.6983, | |
| "step": 94600 | |
| }, | |
| { | |
| "epoch": 2.8, | |
| "learning_rate": 4.031557865734363e-06, | |
| "loss": 2.5699, | |
| "step": 94700 | |
| }, | |
| { | |
| "epoch": 2.8, | |
| "learning_rate": 3.9736747664416946e-06, | |
| "loss": 2.6127, | |
| "step": 94800 | |
| }, | |
| { | |
| "epoch": 2.8, | |
| "learning_rate": 3.915791667149026e-06, | |
| "loss": 2.6966, | |
| "step": 94900 | |
| }, | |
| { | |
| "epoch": 2.8, | |
| "learning_rate": 3.857908567856357e-06, | |
| "loss": 2.6416, | |
| "step": 95000 | |
| }, | |
| { | |
| "epoch": 2.81, | |
| "learning_rate": 3.8000254685636887e-06, | |
| "loss": 2.6861, | |
| "step": 95100 | |
| }, | |
| { | |
| "epoch": 2.81, | |
| "learning_rate": 3.7421423692710205e-06, | |
| "loss": 2.5394, | |
| "step": 95200 | |
| }, | |
| { | |
| "epoch": 2.81, | |
| "learning_rate": 3.684259269978352e-06, | |
| "loss": 2.5937, | |
| "step": 95300 | |
| }, | |
| { | |
| "epoch": 2.82, | |
| "learning_rate": 3.626376170685683e-06, | |
| "loss": 2.6151, | |
| "step": 95400 | |
| }, | |
| { | |
| "epoch": 2.82, | |
| "learning_rate": 3.568493071393015e-06, | |
| "loss": 2.6567, | |
| "step": 95500 | |
| }, | |
| { | |
| "epoch": 2.82, | |
| "learning_rate": 3.5106099721003464e-06, | |
| "loss": 2.5833, | |
| "step": 95600 | |
| }, | |
| { | |
| "epoch": 2.83, | |
| "learning_rate": 3.452726872807678e-06, | |
| "loss": 2.6624, | |
| "step": 95700 | |
| }, | |
| { | |
| "epoch": 2.83, | |
| "learning_rate": 3.394843773515009e-06, | |
| "loss": 2.6202, | |
| "step": 95800 | |
| }, | |
| { | |
| "epoch": 2.83, | |
| "learning_rate": 3.3369606742223405e-06, | |
| "loss": 2.5676, | |
| "step": 95900 | |
| }, | |
| { | |
| "epoch": 2.83, | |
| "learning_rate": 3.2790775749296723e-06, | |
| "loss": 2.6249, | |
| "step": 96000 | |
| }, | |
| { | |
| "epoch": 2.84, | |
| "learning_rate": 3.2211944756370037e-06, | |
| "loss": 2.6144, | |
| "step": 96100 | |
| }, | |
| { | |
| "epoch": 2.84, | |
| "learning_rate": 3.163311376344335e-06, | |
| "loss": 2.6163, | |
| "step": 96200 | |
| }, | |
| { | |
| "epoch": 2.84, | |
| "learning_rate": 3.1054282770516664e-06, | |
| "loss": 2.5738, | |
| "step": 96300 | |
| }, | |
| { | |
| "epoch": 2.85, | |
| "learning_rate": 3.047545177758998e-06, | |
| "loss": 2.6832, | |
| "step": 96400 | |
| }, | |
| { | |
| "epoch": 2.85, | |
| "learning_rate": 2.9896620784663296e-06, | |
| "loss": 2.5983, | |
| "step": 96500 | |
| }, | |
| { | |
| "epoch": 2.85, | |
| "learning_rate": 2.932357810166588e-06, | |
| "loss": 2.6802, | |
| "step": 96600 | |
| }, | |
| { | |
| "epoch": 2.85, | |
| "learning_rate": 2.875053541866846e-06, | |
| "loss": 2.6392, | |
| "step": 96700 | |
| }, | |
| { | |
| "epoch": 2.86, | |
| "learning_rate": 2.817170442574177e-06, | |
| "loss": 2.627, | |
| "step": 96800 | |
| }, | |
| { | |
| "epoch": 2.86, | |
| "learning_rate": 2.7592873432815088e-06, | |
| "loss": 2.6258, | |
| "step": 96900 | |
| }, | |
| { | |
| "epoch": 2.86, | |
| "learning_rate": 2.70140424398884e-06, | |
| "loss": 2.5845, | |
| "step": 97000 | |
| }, | |
| { | |
| "epoch": 2.87, | |
| "learning_rate": 2.643521144696172e-06, | |
| "loss": 2.6863, | |
| "step": 97100 | |
| }, | |
| { | |
| "epoch": 2.87, | |
| "learning_rate": 2.5856380454035033e-06, | |
| "loss": 2.5749, | |
| "step": 97200 | |
| }, | |
| { | |
| "epoch": 2.87, | |
| "learning_rate": 2.5277549461108347e-06, | |
| "loss": 2.5582, | |
| "step": 97300 | |
| }, | |
| { | |
| "epoch": 2.88, | |
| "learning_rate": 2.469871846818166e-06, | |
| "loss": 2.6249, | |
| "step": 97400 | |
| }, | |
| { | |
| "epoch": 2.88, | |
| "learning_rate": 2.4119887475254974e-06, | |
| "loss": 2.6195, | |
| "step": 97500 | |
| }, | |
| { | |
| "epoch": 2.88, | |
| "learning_rate": 2.3541056482328292e-06, | |
| "loss": 2.5851, | |
| "step": 97600 | |
| }, | |
| { | |
| "epoch": 2.88, | |
| "learning_rate": 2.2962225489401606e-06, | |
| "loss": 2.6251, | |
| "step": 97700 | |
| }, | |
| { | |
| "epoch": 2.89, | |
| "learning_rate": 2.238339449647492e-06, | |
| "loss": 2.6062, | |
| "step": 97800 | |
| }, | |
| { | |
| "epoch": 2.89, | |
| "learning_rate": 2.1804563503548233e-06, | |
| "loss": 2.6002, | |
| "step": 97900 | |
| }, | |
| { | |
| "epoch": 2.89, | |
| "learning_rate": 2.122573251062155e-06, | |
| "loss": 2.5926, | |
| "step": 98000 | |
| }, | |
| { | |
| "epoch": 2.9, | |
| "learning_rate": 2.0646901517694865e-06, | |
| "loss": 2.5646, | |
| "step": 98100 | |
| }, | |
| { | |
| "epoch": 2.9, | |
| "learning_rate": 2.006807052476818e-06, | |
| "loss": 2.6043, | |
| "step": 98200 | |
| }, | |
| { | |
| "epoch": 2.9, | |
| "learning_rate": 1.9489239531841493e-06, | |
| "loss": 2.6333, | |
| "step": 98300 | |
| }, | |
| { | |
| "epoch": 2.9, | |
| "learning_rate": 1.891040853891481e-06, | |
| "loss": 2.6832, | |
| "step": 98400 | |
| }, | |
| { | |
| "epoch": 2.91, | |
| "learning_rate": 1.8331577545988122e-06, | |
| "loss": 2.5962, | |
| "step": 98500 | |
| }, | |
| { | |
| "epoch": 2.91, | |
| "learning_rate": 1.7752746553061438e-06, | |
| "loss": 2.5622, | |
| "step": 98600 | |
| }, | |
| { | |
| "epoch": 2.91, | |
| "learning_rate": 1.7173915560134754e-06, | |
| "loss": 2.5946, | |
| "step": 98700 | |
| }, | |
| { | |
| "epoch": 2.92, | |
| "learning_rate": 1.659508456720807e-06, | |
| "loss": 2.5805, | |
| "step": 98800 | |
| }, | |
| { | |
| "epoch": 2.92, | |
| "learning_rate": 1.6016253574281381e-06, | |
| "loss": 2.6398, | |
| "step": 98900 | |
| }, | |
| { | |
| "epoch": 2.92, | |
| "learning_rate": 1.5437422581354697e-06, | |
| "loss": 2.5873, | |
| "step": 99000 | |
| }, | |
| { | |
| "epoch": 2.93, | |
| "learning_rate": 1.485859158842801e-06, | |
| "loss": 2.5955, | |
| "step": 99100 | |
| }, | |
| { | |
| "epoch": 2.93, | |
| "learning_rate": 1.4279760595501327e-06, | |
| "loss": 2.6957, | |
| "step": 99200 | |
| }, | |
| { | |
| "epoch": 2.93, | |
| "learning_rate": 1.370092960257464e-06, | |
| "loss": 2.5805, | |
| "step": 99300 | |
| }, | |
| { | |
| "epoch": 2.93, | |
| "learning_rate": 1.3122098609647956e-06, | |
| "loss": 2.5943, | |
| "step": 99400 | |
| }, | |
| { | |
| "epoch": 2.94, | |
| "learning_rate": 1.254326761672127e-06, | |
| "loss": 2.652, | |
| "step": 99500 | |
| }, | |
| { | |
| "epoch": 2.94, | |
| "learning_rate": 1.1964436623794584e-06, | |
| "loss": 2.656, | |
| "step": 99600 | |
| }, | |
| { | |
| "epoch": 2.94, | |
| "learning_rate": 1.13856056308679e-06, | |
| "loss": 2.5938, | |
| "step": 99700 | |
| }, | |
| { | |
| "epoch": 2.95, | |
| "learning_rate": 1.0806774637941215e-06, | |
| "loss": 2.5805, | |
| "step": 99800 | |
| }, | |
| { | |
| "epoch": 2.95, | |
| "learning_rate": 1.0227943645014529e-06, | |
| "loss": 2.6259, | |
| "step": 99900 | |
| }, | |
| { | |
| "epoch": 2.95, | |
| "learning_rate": 9.649112652087845e-07, | |
| "loss": 2.6734, | |
| "step": 100000 | |
| }, | |
| { | |
| "epoch": 2.95, | |
| "learning_rate": 9.070281659161157e-07, | |
| "loss": 2.6134, | |
| "step": 100100 | |
| }, | |
| { | |
| "epoch": 2.96, | |
| "learning_rate": 8.491450666234473e-07, | |
| "loss": 2.6146, | |
| "step": 100200 | |
| }, | |
| { | |
| "epoch": 2.96, | |
| "learning_rate": 7.912619673307788e-07, | |
| "loss": 2.6645, | |
| "step": 100300 | |
| }, | |
| { | |
| "epoch": 2.96, | |
| "learning_rate": 7.333788680381103e-07, | |
| "loss": 2.6002, | |
| "step": 100400 | |
| }, | |
| { | |
| "epoch": 2.97, | |
| "learning_rate": 6.754957687454418e-07, | |
| "loss": 2.6512, | |
| "step": 100500 | |
| }, | |
| { | |
| "epoch": 2.97, | |
| "learning_rate": 6.176126694527732e-07, | |
| "loss": 2.6164, | |
| "step": 100600 | |
| }, | |
| { | |
| "epoch": 2.97, | |
| "learning_rate": 5.597295701601047e-07, | |
| "loss": 2.5653, | |
| "step": 100700 | |
| }, | |
| { | |
| "epoch": 2.98, | |
| "learning_rate": 5.024253018603628e-07, | |
| "loss": 2.5852, | |
| "step": 100800 | |
| }, | |
| { | |
| "epoch": 2.98, | |
| "learning_rate": 4.4454220256769427e-07, | |
| "loss": 2.6347, | |
| "step": 100900 | |
| }, | |
| { | |
| "epoch": 2.98, | |
| "learning_rate": 3.866591032750258e-07, | |
| "loss": 2.589, | |
| "step": 101000 | |
| }, | |
| { | |
| "epoch": 2.98, | |
| "learning_rate": 3.287760039823572e-07, | |
| "loss": 2.5523, | |
| "step": 101100 | |
| }, | |
| { | |
| "epoch": 2.99, | |
| "learning_rate": 2.708929046896887e-07, | |
| "loss": 2.6118, | |
| "step": 101200 | |
| }, | |
| { | |
| "epoch": 2.99, | |
| "learning_rate": 2.130098053970202e-07, | |
| "loss": 2.6409, | |
| "step": 101300 | |
| }, | |
| { | |
| "epoch": 2.99, | |
| "learning_rate": 1.5570553709727834e-07, | |
| "loss": 2.5427, | |
| "step": 101400 | |
| }, | |
| { | |
| "epoch": 3.0, | |
| "learning_rate": 9.782243780460982e-08, | |
| "loss": 2.6227, | |
| "step": 101500 | |
| }, | |
| { | |
| "epoch": 3.0, | |
| "learning_rate": 3.993933851194129e-08, | |
| "loss": 2.5889, | |
| "step": 101600 | |
| }, | |
| { | |
| "epoch": 3.0, | |
| "eval_gen_len": 19.8855, | |
| "eval_loss": 3.09702730178833, | |
| "eval_rouge1": 19.1763, | |
| "eval_rouge2": 5.4517, | |
| "eval_rougeL": 15.5342, | |
| "eval_rougeLsum": 16.7186, | |
| "eval_runtime": 748.2031, | |
| "eval_samples_per_second": 10.06, | |
| "eval_steps_per_second": 2.515, | |
| "step": 101625 | |
| }, | |
| { | |
| "epoch": 3.0, | |
| "step": 101625, | |
| "total_flos": 7.175645735046021e+17, | |
| "train_loss": 3.0545432459982558, | |
| "train_runtime": 34500.7445, | |
| "train_samples_per_second": 11.782, | |
| "train_steps_per_second": 2.946 | |
| } | |
| ], | |
| "max_steps": 101625, | |
| "num_train_epochs": 3, | |
| "total_flos": 7.175645735046021e+17, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |