| { | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 3.0, | |
| "eval_steps": 1, | |
| "global_step": 61188, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.004902922141596391, | |
| "grad_norm": 0.5712823271751404, | |
| "learning_rate": 4.902922141596392e-07, | |
| "loss": 0.3911, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.009805844283192783, | |
| "grad_norm": 0.7707542181015015, | |
| "learning_rate": 9.805844283192783e-07, | |
| "loss": 0.3924, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.014708766424789174, | |
| "grad_norm": 0.7901243567466736, | |
| "learning_rate": 1.4708766424789175e-06, | |
| "loss": 0.338, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.019611688566385566, | |
| "grad_norm": 0.4191139042377472, | |
| "learning_rate": 1.9611688566385566e-06, | |
| "loss": 0.2563, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 0.024514610707981957, | |
| "grad_norm": 0.5541927218437195, | |
| "learning_rate": 2.4514610707981958e-06, | |
| "loss": 0.1763, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.02941753284957835, | |
| "grad_norm": 0.5028297305107117, | |
| "learning_rate": 2.941753284957835e-06, | |
| "loss": 0.1511, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 0.03432045499117474, | |
| "grad_norm": 0.6625041961669922, | |
| "learning_rate": 3.432045499117474e-06, | |
| "loss": 0.142, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 0.03922337713277113, | |
| "grad_norm": 0.42951473593711853, | |
| "learning_rate": 3.922337713277113e-06, | |
| "loss": 0.1381, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 0.04412629927436752, | |
| "grad_norm": 0.6672629714012146, | |
| "learning_rate": 4.412629927436753e-06, | |
| "loss": 0.132, | |
| "step": 900 | |
| }, | |
| { | |
| "epoch": 0.049029221415963914, | |
| "grad_norm": 0.4937724173069, | |
| "learning_rate": 4.9029221415963915e-06, | |
| "loss": 0.1418, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 0.053932143557560305, | |
| "grad_norm": 0.3915172815322876, | |
| "learning_rate": 5.393214355756031e-06, | |
| "loss": 0.1254, | |
| "step": 1100 | |
| }, | |
| { | |
| "epoch": 0.0588350656991567, | |
| "grad_norm": 0.39539510011672974, | |
| "learning_rate": 5.88350656991567e-06, | |
| "loss": 0.1252, | |
| "step": 1200 | |
| }, | |
| { | |
| "epoch": 0.06373798784075309, | |
| "grad_norm": 0.7038196325302124, | |
| "learning_rate": 6.373798784075309e-06, | |
| "loss": 0.121, | |
| "step": 1300 | |
| }, | |
| { | |
| "epoch": 0.06864090998234948, | |
| "grad_norm": 0.6034905910491943, | |
| "learning_rate": 6.864090998234948e-06, | |
| "loss": 0.1231, | |
| "step": 1400 | |
| }, | |
| { | |
| "epoch": 0.07354383212394587, | |
| "grad_norm": 0.5692676305770874, | |
| "learning_rate": 7.354383212394588e-06, | |
| "loss": 0.1138, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 0.07844675426554226, | |
| "grad_norm": 0.8084454536437988, | |
| "learning_rate": 7.844675426554226e-06, | |
| "loss": 0.1149, | |
| "step": 1600 | |
| }, | |
| { | |
| "epoch": 0.08334967640713865, | |
| "grad_norm": 0.4154703915119171, | |
| "learning_rate": 8.334967640713865e-06, | |
| "loss": 0.1069, | |
| "step": 1700 | |
| }, | |
| { | |
| "epoch": 0.08825259854873505, | |
| "grad_norm": 0.8582517504692078, | |
| "learning_rate": 8.825259854873506e-06, | |
| "loss": 0.1121, | |
| "step": 1800 | |
| }, | |
| { | |
| "epoch": 0.09315552069033144, | |
| "grad_norm": 0.5646160244941711, | |
| "learning_rate": 9.315552069033144e-06, | |
| "loss": 0.1018, | |
| "step": 1900 | |
| }, | |
| { | |
| "epoch": 0.09805844283192783, | |
| "grad_norm": 0.5309118032455444, | |
| "learning_rate": 9.805844283192783e-06, | |
| "loss": 0.1027, | |
| "step": 2000 | |
| }, | |
| { | |
| "epoch": 0.10296136497352422, | |
| "grad_norm": 0.6853476166725159, | |
| "learning_rate": 1.0296136497352422e-05, | |
| "loss": 0.1084, | |
| "step": 2100 | |
| }, | |
| { | |
| "epoch": 0.10786428711512061, | |
| "grad_norm": 0.6206139922142029, | |
| "learning_rate": 1.0786428711512062e-05, | |
| "loss": 0.1008, | |
| "step": 2200 | |
| }, | |
| { | |
| "epoch": 0.112767209256717, | |
| "grad_norm": 0.5461515784263611, | |
| "learning_rate": 1.1276720925671701e-05, | |
| "loss": 0.0988, | |
| "step": 2300 | |
| }, | |
| { | |
| "epoch": 0.1176701313983134, | |
| "grad_norm": 0.8931940793991089, | |
| "learning_rate": 1.176701313983134e-05, | |
| "loss": 0.0915, | |
| "step": 2400 | |
| }, | |
| { | |
| "epoch": 0.12257305353990978, | |
| "grad_norm": 0.779230535030365, | |
| "learning_rate": 1.2257305353990978e-05, | |
| "loss": 0.0978, | |
| "step": 2500 | |
| }, | |
| { | |
| "epoch": 0.12747597568150618, | |
| "grad_norm": 0.6850101947784424, | |
| "learning_rate": 1.2747597568150619e-05, | |
| "loss": 0.095, | |
| "step": 2600 | |
| }, | |
| { | |
| "epoch": 0.13237889782310258, | |
| "grad_norm": 0.9090009331703186, | |
| "learning_rate": 1.323788978231026e-05, | |
| "loss": 0.0915, | |
| "step": 2700 | |
| }, | |
| { | |
| "epoch": 0.13728181996469896, | |
| "grad_norm": 0.7030491828918457, | |
| "learning_rate": 1.3728181996469896e-05, | |
| "loss": 0.0884, | |
| "step": 2800 | |
| }, | |
| { | |
| "epoch": 0.14218474210629536, | |
| "grad_norm": 0.7005199790000916, | |
| "learning_rate": 1.4218474210629537e-05, | |
| "loss": 0.0846, | |
| "step": 2900 | |
| }, | |
| { | |
| "epoch": 0.14708766424789174, | |
| "grad_norm": 0.8862398266792297, | |
| "learning_rate": 1.4708766424789175e-05, | |
| "loss": 0.0871, | |
| "step": 3000 | |
| }, | |
| { | |
| "epoch": 0.15199058638948815, | |
| "grad_norm": 0.667682409286499, | |
| "learning_rate": 1.5199058638948816e-05, | |
| "loss": 0.0852, | |
| "step": 3100 | |
| }, | |
| { | |
| "epoch": 0.15689350853108452, | |
| "grad_norm": 0.6546317338943481, | |
| "learning_rate": 1.5689350853108453e-05, | |
| "loss": 0.0843, | |
| "step": 3200 | |
| }, | |
| { | |
| "epoch": 0.16179643067268093, | |
| "grad_norm": 1.0147327184677124, | |
| "learning_rate": 1.6179643067268095e-05, | |
| "loss": 0.0824, | |
| "step": 3300 | |
| }, | |
| { | |
| "epoch": 0.1666993528142773, | |
| "grad_norm": 0.9607380628585815, | |
| "learning_rate": 1.666993528142773e-05, | |
| "loss": 0.0783, | |
| "step": 3400 | |
| }, | |
| { | |
| "epoch": 0.1716022749558737, | |
| "grad_norm": 0.6415452361106873, | |
| "learning_rate": 1.7160227495587372e-05, | |
| "loss": 0.0804, | |
| "step": 3500 | |
| }, | |
| { | |
| "epoch": 0.1765051970974701, | |
| "grad_norm": 0.5318363904953003, | |
| "learning_rate": 1.765051970974701e-05, | |
| "loss": 0.0808, | |
| "step": 3600 | |
| }, | |
| { | |
| "epoch": 0.1814081192390665, | |
| "grad_norm": 0.7230063080787659, | |
| "learning_rate": 1.814081192390665e-05, | |
| "loss": 0.0775, | |
| "step": 3700 | |
| }, | |
| { | |
| "epoch": 0.18631104138066287, | |
| "grad_norm": 0.9152194857597351, | |
| "learning_rate": 1.863110413806629e-05, | |
| "loss": 0.0746, | |
| "step": 3800 | |
| }, | |
| { | |
| "epoch": 0.19121396352225928, | |
| "grad_norm": 0.769027054309845, | |
| "learning_rate": 1.9121396352225927e-05, | |
| "loss": 0.0774, | |
| "step": 3900 | |
| }, | |
| { | |
| "epoch": 0.19611688566385566, | |
| "grad_norm": 0.8306779861450195, | |
| "learning_rate": 1.9611688566385566e-05, | |
| "loss": 0.0735, | |
| "step": 4000 | |
| }, | |
| { | |
| "epoch": 0.20101980780545206, | |
| "grad_norm": 0.502835750579834, | |
| "learning_rate": 2.0101980780545208e-05, | |
| "loss": 0.076, | |
| "step": 4100 | |
| }, | |
| { | |
| "epoch": 0.20592272994704844, | |
| "grad_norm": 0.7486907839775085, | |
| "learning_rate": 2.0592272994704843e-05, | |
| "loss": 0.0786, | |
| "step": 4200 | |
| }, | |
| { | |
| "epoch": 0.21082565208864484, | |
| "grad_norm": 0.6800108551979065, | |
| "learning_rate": 2.1082565208864486e-05, | |
| "loss": 0.0724, | |
| "step": 4300 | |
| }, | |
| { | |
| "epoch": 0.21572857423024122, | |
| "grad_norm": 0.8575353026390076, | |
| "learning_rate": 2.1572857423024124e-05, | |
| "loss": 0.0726, | |
| "step": 4400 | |
| }, | |
| { | |
| "epoch": 0.22063149637183763, | |
| "grad_norm": 1.0145362615585327, | |
| "learning_rate": 2.2063149637183763e-05, | |
| "loss": 0.0659, | |
| "step": 4500 | |
| }, | |
| { | |
| "epoch": 0.225534418513434, | |
| "grad_norm": 0.7558209896087646, | |
| "learning_rate": 2.2553441851343402e-05, | |
| "loss": 0.0688, | |
| "step": 4600 | |
| }, | |
| { | |
| "epoch": 0.2304373406550304, | |
| "grad_norm": 0.8182774186134338, | |
| "learning_rate": 2.304373406550304e-05, | |
| "loss": 0.0718, | |
| "step": 4700 | |
| }, | |
| { | |
| "epoch": 0.2353402627966268, | |
| "grad_norm": 0.5569434762001038, | |
| "learning_rate": 2.353402627966268e-05, | |
| "loss": 0.0704, | |
| "step": 4800 | |
| }, | |
| { | |
| "epoch": 0.2402431849382232, | |
| "grad_norm": 0.7403717041015625, | |
| "learning_rate": 2.402431849382232e-05, | |
| "loss": 0.0729, | |
| "step": 4900 | |
| }, | |
| { | |
| "epoch": 0.24514610707981957, | |
| "grad_norm": 1.4814640283584595, | |
| "learning_rate": 2.4514610707981957e-05, | |
| "loss": 0.0731, | |
| "step": 5000 | |
| }, | |
| { | |
| "epoch": 0.250049029221416, | |
| "grad_norm": 0.5557782649993896, | |
| "learning_rate": 2.50049029221416e-05, | |
| "loss": 0.0676, | |
| "step": 5100 | |
| }, | |
| { | |
| "epoch": 0.25495195136301235, | |
| "grad_norm": 0.5674753785133362, | |
| "learning_rate": 2.5495195136301238e-05, | |
| "loss": 0.0675, | |
| "step": 5200 | |
| }, | |
| { | |
| "epoch": 0.25985487350460873, | |
| "grad_norm": 0.8610398769378662, | |
| "learning_rate": 2.5985487350460873e-05, | |
| "loss": 0.0676, | |
| "step": 5300 | |
| }, | |
| { | |
| "epoch": 0.26475779564620516, | |
| "grad_norm": 0.6987388134002686, | |
| "learning_rate": 2.647577956462052e-05, | |
| "loss": 0.0683, | |
| "step": 5400 | |
| }, | |
| { | |
| "epoch": 0.26966071778780154, | |
| "grad_norm": 1.1044350862503052, | |
| "learning_rate": 2.6966071778780154e-05, | |
| "loss": 0.0684, | |
| "step": 5500 | |
| }, | |
| { | |
| "epoch": 0.2745636399293979, | |
| "grad_norm": 0.7976281046867371, | |
| "learning_rate": 2.7456363992939792e-05, | |
| "loss": 0.0666, | |
| "step": 5600 | |
| }, | |
| { | |
| "epoch": 0.2794665620709943, | |
| "grad_norm": 0.7889923453330994, | |
| "learning_rate": 2.794665620709943e-05, | |
| "loss": 0.0688, | |
| "step": 5700 | |
| }, | |
| { | |
| "epoch": 0.28436948421259073, | |
| "grad_norm": 1.111566424369812, | |
| "learning_rate": 2.8436948421259073e-05, | |
| "loss": 0.0637, | |
| "step": 5800 | |
| }, | |
| { | |
| "epoch": 0.2892724063541871, | |
| "grad_norm": 0.8932578563690186, | |
| "learning_rate": 2.8927240635418712e-05, | |
| "loss": 0.0636, | |
| "step": 5900 | |
| }, | |
| { | |
| "epoch": 0.2941753284957835, | |
| "grad_norm": 1.0494166612625122, | |
| "learning_rate": 2.941753284957835e-05, | |
| "loss": 0.0641, | |
| "step": 6000 | |
| }, | |
| { | |
| "epoch": 0.29907825063737986, | |
| "grad_norm": 0.7820980548858643, | |
| "learning_rate": 2.9907825063737986e-05, | |
| "loss": 0.0715, | |
| "step": 6100 | |
| }, | |
| { | |
| "epoch": 0.3039811727789763, | |
| "grad_norm": 0.9710949659347534, | |
| "learning_rate": 3.039811727789763e-05, | |
| "loss": 0.066, | |
| "step": 6200 | |
| }, | |
| { | |
| "epoch": 0.30888409492057267, | |
| "grad_norm": 0.807854950428009, | |
| "learning_rate": 3.088840949205727e-05, | |
| "loss": 0.0629, | |
| "step": 6300 | |
| }, | |
| { | |
| "epoch": 0.31378701706216905, | |
| "grad_norm": 0.8420621156692505, | |
| "learning_rate": 3.1378701706216906e-05, | |
| "loss": 0.0692, | |
| "step": 6400 | |
| }, | |
| { | |
| "epoch": 0.3186899392037654, | |
| "grad_norm": 0.776466965675354, | |
| "learning_rate": 3.1868993920376544e-05, | |
| "loss": 0.0625, | |
| "step": 6500 | |
| }, | |
| { | |
| "epoch": 0.32359286134536186, | |
| "grad_norm": 0.8149374127388, | |
| "learning_rate": 3.235928613453619e-05, | |
| "loss": 0.0612, | |
| "step": 6600 | |
| }, | |
| { | |
| "epoch": 0.32849578348695824, | |
| "grad_norm": 1.0844323635101318, | |
| "learning_rate": 3.284957834869583e-05, | |
| "loss": 0.0641, | |
| "step": 6700 | |
| }, | |
| { | |
| "epoch": 0.3333987056285546, | |
| "grad_norm": 0.6500579118728638, | |
| "learning_rate": 3.333987056285546e-05, | |
| "loss": 0.0597, | |
| "step": 6800 | |
| }, | |
| { | |
| "epoch": 0.338301627770151, | |
| "grad_norm": 0.7218449115753174, | |
| "learning_rate": 3.38301627770151e-05, | |
| "loss": 0.0603, | |
| "step": 6900 | |
| }, | |
| { | |
| "epoch": 0.3432045499117474, | |
| "grad_norm": 1.0777571201324463, | |
| "learning_rate": 3.4320454991174745e-05, | |
| "loss": 0.0572, | |
| "step": 7000 | |
| }, | |
| { | |
| "epoch": 0.3481074720533438, | |
| "grad_norm": 0.786324143409729, | |
| "learning_rate": 3.4810747205334384e-05, | |
| "loss": 0.06, | |
| "step": 7100 | |
| }, | |
| { | |
| "epoch": 0.3530103941949402, | |
| "grad_norm": 0.44759607315063477, | |
| "learning_rate": 3.530103941949402e-05, | |
| "loss": 0.064, | |
| "step": 7200 | |
| }, | |
| { | |
| "epoch": 0.35791331633653656, | |
| "grad_norm": 0.4074469208717346, | |
| "learning_rate": 3.5791331633653654e-05, | |
| "loss": 0.06, | |
| "step": 7300 | |
| }, | |
| { | |
| "epoch": 0.362816238478133, | |
| "grad_norm": 0.9225257039070129, | |
| "learning_rate": 3.62816238478133e-05, | |
| "loss": 0.061, | |
| "step": 7400 | |
| }, | |
| { | |
| "epoch": 0.36771916061972937, | |
| "grad_norm": 0.6077998280525208, | |
| "learning_rate": 3.677191606197294e-05, | |
| "loss": 0.053, | |
| "step": 7500 | |
| }, | |
| { | |
| "epoch": 0.37262208276132575, | |
| "grad_norm": 0.9564809799194336, | |
| "learning_rate": 3.726220827613258e-05, | |
| "loss": 0.0558, | |
| "step": 7600 | |
| }, | |
| { | |
| "epoch": 0.3775250049029221, | |
| "grad_norm": 0.6612119078636169, | |
| "learning_rate": 3.7752500490292216e-05, | |
| "loss": 0.0621, | |
| "step": 7700 | |
| }, | |
| { | |
| "epoch": 0.38242792704451856, | |
| "grad_norm": 0.7367020845413208, | |
| "learning_rate": 3.8242792704451855e-05, | |
| "loss": 0.0596, | |
| "step": 7800 | |
| }, | |
| { | |
| "epoch": 0.38733084918611493, | |
| "grad_norm": 0.595927357673645, | |
| "learning_rate": 3.873308491861149e-05, | |
| "loss": 0.0576, | |
| "step": 7900 | |
| }, | |
| { | |
| "epoch": 0.3922337713277113, | |
| "grad_norm": 0.7036033272743225, | |
| "learning_rate": 3.922337713277113e-05, | |
| "loss": 0.059, | |
| "step": 8000 | |
| }, | |
| { | |
| "epoch": 0.3971366934693077, | |
| "grad_norm": 0.8807573914527893, | |
| "learning_rate": 3.971366934693077e-05, | |
| "loss": 0.0587, | |
| "step": 8100 | |
| }, | |
| { | |
| "epoch": 0.4020396156109041, | |
| "grad_norm": 0.8724229335784912, | |
| "learning_rate": 4.0203961561090416e-05, | |
| "loss": 0.0588, | |
| "step": 8200 | |
| }, | |
| { | |
| "epoch": 0.4069425377525005, | |
| "grad_norm": 0.7618996500968933, | |
| "learning_rate": 4.0694253775250055e-05, | |
| "loss": 0.0576, | |
| "step": 8300 | |
| }, | |
| { | |
| "epoch": 0.4118454598940969, | |
| "grad_norm": 0.743467390537262, | |
| "learning_rate": 4.118454598940969e-05, | |
| "loss": 0.0581, | |
| "step": 8400 | |
| }, | |
| { | |
| "epoch": 0.41674838203569325, | |
| "grad_norm": 0.6151107549667358, | |
| "learning_rate": 4.1674838203569326e-05, | |
| "loss": 0.0633, | |
| "step": 8500 | |
| }, | |
| { | |
| "epoch": 0.4216513041772897, | |
| "grad_norm": 1.3272536993026733, | |
| "learning_rate": 4.216513041772897e-05, | |
| "loss": 0.0579, | |
| "step": 8600 | |
| }, | |
| { | |
| "epoch": 0.42655422631888606, | |
| "grad_norm": 0.7883033752441406, | |
| "learning_rate": 4.265542263188861e-05, | |
| "loss": 0.0579, | |
| "step": 8700 | |
| }, | |
| { | |
| "epoch": 0.43145714846048244, | |
| "grad_norm": 0.6852431297302246, | |
| "learning_rate": 4.314571484604825e-05, | |
| "loss": 0.0594, | |
| "step": 8800 | |
| }, | |
| { | |
| "epoch": 0.4363600706020788, | |
| "grad_norm": 0.5753522515296936, | |
| "learning_rate": 4.363600706020789e-05, | |
| "loss": 0.0564, | |
| "step": 8900 | |
| }, | |
| { | |
| "epoch": 0.44126299274367525, | |
| "grad_norm": 0.510944128036499, | |
| "learning_rate": 4.4126299274367526e-05, | |
| "loss": 0.0571, | |
| "step": 9000 | |
| }, | |
| { | |
| "epoch": 0.44616591488527163, | |
| "grad_norm": 0.7688953876495361, | |
| "learning_rate": 4.4616591488527165e-05, | |
| "loss": 0.0544, | |
| "step": 9100 | |
| }, | |
| { | |
| "epoch": 0.451068837026868, | |
| "grad_norm": 0.6209575533866882, | |
| "learning_rate": 4.5106883702686804e-05, | |
| "loss": 0.054, | |
| "step": 9200 | |
| }, | |
| { | |
| "epoch": 0.4559717591684644, | |
| "grad_norm": 0.42720097303390503, | |
| "learning_rate": 4.559717591684644e-05, | |
| "loss": 0.0581, | |
| "step": 9300 | |
| }, | |
| { | |
| "epoch": 0.4608746813100608, | |
| "grad_norm": 0.43942296504974365, | |
| "learning_rate": 4.608746813100608e-05, | |
| "loss": 0.0527, | |
| "step": 9400 | |
| }, | |
| { | |
| "epoch": 0.4657776034516572, | |
| "grad_norm": 0.8481646776199341, | |
| "learning_rate": 4.657776034516572e-05, | |
| "loss": 0.0553, | |
| "step": 9500 | |
| }, | |
| { | |
| "epoch": 0.4706805255932536, | |
| "grad_norm": 0.670299232006073, | |
| "learning_rate": 4.706805255932536e-05, | |
| "loss": 0.0474, | |
| "step": 9600 | |
| }, | |
| { | |
| "epoch": 0.47558344773484995, | |
| "grad_norm": 0.7548401951789856, | |
| "learning_rate": 4.7558344773485e-05, | |
| "loss": 0.0527, | |
| "step": 9700 | |
| }, | |
| { | |
| "epoch": 0.4804863698764464, | |
| "grad_norm": 0.49902454018592834, | |
| "learning_rate": 4.804863698764464e-05, | |
| "loss": 0.0563, | |
| "step": 9800 | |
| }, | |
| { | |
| "epoch": 0.48538929201804276, | |
| "grad_norm": 0.6898229718208313, | |
| "learning_rate": 4.853892920180428e-05, | |
| "loss": 0.054, | |
| "step": 9900 | |
| }, | |
| { | |
| "epoch": 0.49029221415963914, | |
| "grad_norm": 0.7754454612731934, | |
| "learning_rate": 4.902922141596391e-05, | |
| "loss": 0.0569, | |
| "step": 10000 | |
| }, | |
| { | |
| "epoch": 0.4951951363012355, | |
| "grad_norm": 0.5083482265472412, | |
| "learning_rate": 4.951951363012355e-05, | |
| "loss": 0.0524, | |
| "step": 10100 | |
| }, | |
| { | |
| "epoch": 0.500098058442832, | |
| "grad_norm": 0.5904260277748108, | |
| "learning_rate": 5.00098058442832e-05, | |
| "loss": 0.0544, | |
| "step": 10200 | |
| }, | |
| { | |
| "epoch": 0.5050009805844283, | |
| "grad_norm": 0.3116986155509949, | |
| "learning_rate": 5.0500098058442836e-05, | |
| "loss": 0.0569, | |
| "step": 10300 | |
| }, | |
| { | |
| "epoch": 0.5099039027260247, | |
| "grad_norm": 0.4503382444381714, | |
| "learning_rate": 5.0990390272602475e-05, | |
| "loss": 0.0493, | |
| "step": 10400 | |
| }, | |
| { | |
| "epoch": 0.5148068248676211, | |
| "grad_norm": 0.4754750728607178, | |
| "learning_rate": 5.1480682486762114e-05, | |
| "loss": 0.0549, | |
| "step": 10500 | |
| }, | |
| { | |
| "epoch": 0.5197097470092175, | |
| "grad_norm": 0.5342922806739807, | |
| "learning_rate": 5.1970974700921746e-05, | |
| "loss": 0.0544, | |
| "step": 10600 | |
| }, | |
| { | |
| "epoch": 0.5246126691508138, | |
| "grad_norm": 0.8138884902000427, | |
| "learning_rate": 5.2461266915081384e-05, | |
| "loss": 0.0524, | |
| "step": 10700 | |
| }, | |
| { | |
| "epoch": 0.5295155912924103, | |
| "grad_norm": 0.8454159498214722, | |
| "learning_rate": 5.295155912924104e-05, | |
| "loss": 0.0508, | |
| "step": 10800 | |
| }, | |
| { | |
| "epoch": 0.5344185134340067, | |
| "grad_norm": 0.5879427790641785, | |
| "learning_rate": 5.3441851343400675e-05, | |
| "loss": 0.0512, | |
| "step": 10900 | |
| }, | |
| { | |
| "epoch": 0.5393214355756031, | |
| "grad_norm": 0.6766416430473328, | |
| "learning_rate": 5.393214355756031e-05, | |
| "loss": 0.0517, | |
| "step": 11000 | |
| }, | |
| { | |
| "epoch": 0.5442243577171995, | |
| "grad_norm": 0.5555431842803955, | |
| "learning_rate": 5.4422435771719946e-05, | |
| "loss": 0.0528, | |
| "step": 11100 | |
| }, | |
| { | |
| "epoch": 0.5491272798587958, | |
| "grad_norm": 0.43258973956108093, | |
| "learning_rate": 5.4912727985879585e-05, | |
| "loss": 0.0536, | |
| "step": 11200 | |
| }, | |
| { | |
| "epoch": 0.5540302020003922, | |
| "grad_norm": 0.3768501877784729, | |
| "learning_rate": 5.5403020200039224e-05, | |
| "loss": 0.0514, | |
| "step": 11300 | |
| }, | |
| { | |
| "epoch": 0.5589331241419886, | |
| "grad_norm": 0.4958087205886841, | |
| "learning_rate": 5.589331241419886e-05, | |
| "loss": 0.0524, | |
| "step": 11400 | |
| }, | |
| { | |
| "epoch": 0.563836046283585, | |
| "grad_norm": 0.5648083686828613, | |
| "learning_rate": 5.63836046283585e-05, | |
| "loss": 0.0556, | |
| "step": 11500 | |
| }, | |
| { | |
| "epoch": 0.5687389684251815, | |
| "grad_norm": 0.39568981528282166, | |
| "learning_rate": 5.6873896842518147e-05, | |
| "loss": 0.0497, | |
| "step": 11600 | |
| }, | |
| { | |
| "epoch": 0.5736418905667778, | |
| "grad_norm": 0.3252398371696472, | |
| "learning_rate": 5.7364189056677785e-05, | |
| "loss": 0.0547, | |
| "step": 11700 | |
| }, | |
| { | |
| "epoch": 0.5785448127083742, | |
| "grad_norm": 0.24218697845935822, | |
| "learning_rate": 5.7854481270837424e-05, | |
| "loss": 0.0448, | |
| "step": 11800 | |
| }, | |
| { | |
| "epoch": 0.5834477348499706, | |
| "grad_norm": 0.4660034775733948, | |
| "learning_rate": 5.834477348499706e-05, | |
| "loss": 0.0509, | |
| "step": 11900 | |
| }, | |
| { | |
| "epoch": 0.588350656991567, | |
| "grad_norm": 0.14836396276950836, | |
| "learning_rate": 5.88350656991567e-05, | |
| "loss": 0.0494, | |
| "step": 12000 | |
| }, | |
| { | |
| "epoch": 0.5932535791331633, | |
| "grad_norm": 0.3045461177825928, | |
| "learning_rate": 5.932535791331634e-05, | |
| "loss": 0.0524, | |
| "step": 12100 | |
| }, | |
| { | |
| "epoch": 0.5981565012747597, | |
| "grad_norm": 0.3134712874889374, | |
| "learning_rate": 5.981565012747597e-05, | |
| "loss": 0.0461, | |
| "step": 12200 | |
| }, | |
| { | |
| "epoch": 0.6030594234163561, | |
| "grad_norm": 0.3141169250011444, | |
| "learning_rate": 6.030594234163561e-05, | |
| "loss": 0.0511, | |
| "step": 12300 | |
| }, | |
| { | |
| "epoch": 0.6079623455579526, | |
| "grad_norm": 0.22318390011787415, | |
| "learning_rate": 6.079623455579526e-05, | |
| "loss": 0.0495, | |
| "step": 12400 | |
| }, | |
| { | |
| "epoch": 0.612865267699549, | |
| "grad_norm": 0.5111287832260132, | |
| "learning_rate": 6.12865267699549e-05, | |
| "loss": 0.0532, | |
| "step": 12500 | |
| }, | |
| { | |
| "epoch": 0.6177681898411453, | |
| "grad_norm": 0.3595598042011261, | |
| "learning_rate": 6.177681898411453e-05, | |
| "loss": 0.0482, | |
| "step": 12600 | |
| }, | |
| { | |
| "epoch": 0.6226711119827417, | |
| "grad_norm": 0.294145405292511, | |
| "learning_rate": 6.226711119827417e-05, | |
| "loss": 0.047, | |
| "step": 12700 | |
| }, | |
| { | |
| "epoch": 0.6275740341243381, | |
| "grad_norm": 0.39084282517433167, | |
| "learning_rate": 6.275740341243381e-05, | |
| "loss": 0.0509, | |
| "step": 12800 | |
| }, | |
| { | |
| "epoch": 0.6324769562659345, | |
| "grad_norm": 0.36329174041748047, | |
| "learning_rate": 6.324769562659345e-05, | |
| "loss": 0.0466, | |
| "step": 12900 | |
| }, | |
| { | |
| "epoch": 0.6373798784075309, | |
| "grad_norm": 0.48088952898979187, | |
| "learning_rate": 6.373798784075309e-05, | |
| "loss": 0.049, | |
| "step": 13000 | |
| }, | |
| { | |
| "epoch": 0.6422828005491272, | |
| "grad_norm": 0.3504548966884613, | |
| "learning_rate": 6.422828005491273e-05, | |
| "loss": 0.0565, | |
| "step": 13100 | |
| }, | |
| { | |
| "epoch": 0.6471857226907237, | |
| "grad_norm": 0.4509230852127075, | |
| "learning_rate": 6.471857226907238e-05, | |
| "loss": 0.0476, | |
| "step": 13200 | |
| }, | |
| { | |
| "epoch": 0.6520886448323201, | |
| "grad_norm": 0.49520692229270935, | |
| "learning_rate": 6.520886448323202e-05, | |
| "loss": 0.0495, | |
| "step": 13300 | |
| }, | |
| { | |
| "epoch": 0.6569915669739165, | |
| "grad_norm": 0.2963676154613495, | |
| "learning_rate": 6.569915669739166e-05, | |
| "loss": 0.0486, | |
| "step": 13400 | |
| }, | |
| { | |
| "epoch": 0.6618944891155129, | |
| "grad_norm": 0.279095321893692, | |
| "learning_rate": 6.618944891155128e-05, | |
| "loss": 0.0495, | |
| "step": 13500 | |
| }, | |
| { | |
| "epoch": 0.6667974112571092, | |
| "grad_norm": 0.3021456003189087, | |
| "learning_rate": 6.667974112571092e-05, | |
| "loss": 0.0497, | |
| "step": 13600 | |
| }, | |
| { | |
| "epoch": 0.6717003333987056, | |
| "grad_norm": 0.49535706639289856, | |
| "learning_rate": 6.717003333987056e-05, | |
| "loss": 0.0501, | |
| "step": 13700 | |
| }, | |
| { | |
| "epoch": 0.676603255540302, | |
| "grad_norm": 0.38932666182518005, | |
| "learning_rate": 6.76603255540302e-05, | |
| "loss": 0.0492, | |
| "step": 13800 | |
| }, | |
| { | |
| "epoch": 0.6815061776818984, | |
| "grad_norm": 0.33593711256980896, | |
| "learning_rate": 6.815061776818984e-05, | |
| "loss": 0.0466, | |
| "step": 13900 | |
| }, | |
| { | |
| "epoch": 0.6864090998234949, | |
| "grad_norm": 0.5009463429450989, | |
| "learning_rate": 6.864090998234949e-05, | |
| "loss": 0.0496, | |
| "step": 14000 | |
| }, | |
| { | |
| "epoch": 0.6913120219650912, | |
| "grad_norm": 0.28159981966018677, | |
| "learning_rate": 6.913120219650913e-05, | |
| "loss": 0.0495, | |
| "step": 14100 | |
| }, | |
| { | |
| "epoch": 0.6962149441066876, | |
| "grad_norm": 0.18720287084579468, | |
| "learning_rate": 6.962149441066877e-05, | |
| "loss": 0.0495, | |
| "step": 14200 | |
| }, | |
| { | |
| "epoch": 0.701117866248284, | |
| "grad_norm": 0.3543257713317871, | |
| "learning_rate": 7.01117866248284e-05, | |
| "loss": 0.0474, | |
| "step": 14300 | |
| }, | |
| { | |
| "epoch": 0.7060207883898804, | |
| "grad_norm": 0.3448733389377594, | |
| "learning_rate": 7.060207883898804e-05, | |
| "loss": 0.0465, | |
| "step": 14400 | |
| }, | |
| { | |
| "epoch": 0.7109237105314767, | |
| "grad_norm": 0.47061529755592346, | |
| "learning_rate": 7.109237105314768e-05, | |
| "loss": 0.0479, | |
| "step": 14500 | |
| }, | |
| { | |
| "epoch": 0.7158266326730731, | |
| "grad_norm": 0.18680918216705322, | |
| "learning_rate": 7.158266326730731e-05, | |
| "loss": 0.0478, | |
| "step": 14600 | |
| }, | |
| { | |
| "epoch": 0.7207295548146695, | |
| "grad_norm": 0.2860680818557739, | |
| "learning_rate": 7.207295548146695e-05, | |
| "loss": 0.0416, | |
| "step": 14700 | |
| }, | |
| { | |
| "epoch": 0.725632476956266, | |
| "grad_norm": 0.5014829635620117, | |
| "learning_rate": 7.25632476956266e-05, | |
| "loss": 0.0515, | |
| "step": 14800 | |
| }, | |
| { | |
| "epoch": 0.7305353990978624, | |
| "grad_norm": 0.624320924282074, | |
| "learning_rate": 7.305353990978624e-05, | |
| "loss": 0.0448, | |
| "step": 14900 | |
| }, | |
| { | |
| "epoch": 0.7354383212394587, | |
| "grad_norm": 0.37180984020233154, | |
| "learning_rate": 7.354383212394588e-05, | |
| "loss": 0.0445, | |
| "step": 15000 | |
| }, | |
| { | |
| "epoch": 0.7403412433810551, | |
| "grad_norm": 0.3159758448600769, | |
| "learning_rate": 7.403412433810552e-05, | |
| "loss": 0.0462, | |
| "step": 15100 | |
| }, | |
| { | |
| "epoch": 0.7452441655226515, | |
| "grad_norm": 0.4232478439807892, | |
| "learning_rate": 7.452441655226515e-05, | |
| "loss": 0.0455, | |
| "step": 15200 | |
| }, | |
| { | |
| "epoch": 0.7501470876642479, | |
| "grad_norm": 0.18599086999893188, | |
| "learning_rate": 7.501470876642479e-05, | |
| "loss": 0.0452, | |
| "step": 15300 | |
| }, | |
| { | |
| "epoch": 0.7550500098058442, | |
| "grad_norm": 0.37140703201293945, | |
| "learning_rate": 7.550500098058443e-05, | |
| "loss": 0.048, | |
| "step": 15400 | |
| }, | |
| { | |
| "epoch": 0.7599529319474406, | |
| "grad_norm": 0.45977580547332764, | |
| "learning_rate": 7.599529319474407e-05, | |
| "loss": 0.0449, | |
| "step": 15500 | |
| }, | |
| { | |
| "epoch": 0.7648558540890371, | |
| "grad_norm": 0.2783525586128235, | |
| "learning_rate": 7.648558540890371e-05, | |
| "loss": 0.0408, | |
| "step": 15600 | |
| }, | |
| { | |
| "epoch": 0.7697587762306335, | |
| "grad_norm": 0.3637586832046509, | |
| "learning_rate": 7.697587762306335e-05, | |
| "loss": 0.0461, | |
| "step": 15700 | |
| }, | |
| { | |
| "epoch": 0.7746616983722299, | |
| "grad_norm": 0.3235738277435303, | |
| "learning_rate": 7.746616983722299e-05, | |
| "loss": 0.0462, | |
| "step": 15800 | |
| }, | |
| { | |
| "epoch": 0.7795646205138262, | |
| "grad_norm": 0.28826725482940674, | |
| "learning_rate": 7.795646205138263e-05, | |
| "loss": 0.0446, | |
| "step": 15900 | |
| }, | |
| { | |
| "epoch": 0.7844675426554226, | |
| "grad_norm": 0.4955115020275116, | |
| "learning_rate": 7.844675426554226e-05, | |
| "loss": 0.0469, | |
| "step": 16000 | |
| }, | |
| { | |
| "epoch": 0.789370464797019, | |
| "grad_norm": 0.4371892809867859, | |
| "learning_rate": 7.89370464797019e-05, | |
| "loss": 0.0446, | |
| "step": 16100 | |
| }, | |
| { | |
| "epoch": 0.7942733869386154, | |
| "grad_norm": 0.42605656385421753, | |
| "learning_rate": 7.942733869386154e-05, | |
| "loss": 0.0488, | |
| "step": 16200 | |
| }, | |
| { | |
| "epoch": 0.7991763090802118, | |
| "grad_norm": 0.2385357916355133, | |
| "learning_rate": 7.991763090802118e-05, | |
| "loss": 0.0466, | |
| "step": 16300 | |
| }, | |
| { | |
| "epoch": 0.8040792312218082, | |
| "grad_norm": 0.34842029213905334, | |
| "learning_rate": 8.040792312218083e-05, | |
| "loss": 0.0473, | |
| "step": 16400 | |
| }, | |
| { | |
| "epoch": 0.8089821533634046, | |
| "grad_norm": 0.348670095205307, | |
| "learning_rate": 8.089821533634047e-05, | |
| "loss": 0.046, | |
| "step": 16500 | |
| }, | |
| { | |
| "epoch": 0.813885075505001, | |
| "grad_norm": 0.0661313384771347, | |
| "learning_rate": 8.138850755050011e-05, | |
| "loss": 0.0424, | |
| "step": 16600 | |
| }, | |
| { | |
| "epoch": 0.8187879976465974, | |
| "grad_norm": 0.27380064129829407, | |
| "learning_rate": 8.187879976465974e-05, | |
| "loss": 0.0396, | |
| "step": 16700 | |
| }, | |
| { | |
| "epoch": 0.8236909197881938, | |
| "grad_norm": 0.2240746170282364, | |
| "learning_rate": 8.236909197881937e-05, | |
| "loss": 0.043, | |
| "step": 16800 | |
| }, | |
| { | |
| "epoch": 0.8285938419297901, | |
| "grad_norm": 0.39140066504478455, | |
| "learning_rate": 8.285938419297901e-05, | |
| "loss": 0.0403, | |
| "step": 16900 | |
| }, | |
| { | |
| "epoch": 0.8334967640713865, | |
| "grad_norm": 0.31604424118995667, | |
| "learning_rate": 8.334967640713865e-05, | |
| "loss": 0.0458, | |
| "step": 17000 | |
| }, | |
| { | |
| "epoch": 0.8383996862129829, | |
| "grad_norm": 0.29236578941345215, | |
| "learning_rate": 8.383996862129829e-05, | |
| "loss": 0.0477, | |
| "step": 17100 | |
| }, | |
| { | |
| "epoch": 0.8433026083545794, | |
| "grad_norm": 0.40906059741973877, | |
| "learning_rate": 8.433026083545794e-05, | |
| "loss": 0.0401, | |
| "step": 17200 | |
| }, | |
| { | |
| "epoch": 0.8482055304961758, | |
| "grad_norm": 0.2863566279411316, | |
| "learning_rate": 8.482055304961758e-05, | |
| "loss": 0.0388, | |
| "step": 17300 | |
| }, | |
| { | |
| "epoch": 0.8531084526377721, | |
| "grad_norm": 0.437642902135849, | |
| "learning_rate": 8.531084526377722e-05, | |
| "loss": 0.045, | |
| "step": 17400 | |
| }, | |
| { | |
| "epoch": 0.8580113747793685, | |
| "grad_norm": 0.29014888405799866, | |
| "learning_rate": 8.580113747793686e-05, | |
| "loss": 0.0418, | |
| "step": 17500 | |
| }, | |
| { | |
| "epoch": 0.8629142969209649, | |
| "grad_norm": 0.27143239974975586, | |
| "learning_rate": 8.62914296920965e-05, | |
| "loss": 0.0443, | |
| "step": 17600 | |
| }, | |
| { | |
| "epoch": 0.8678172190625613, | |
| "grad_norm": 0.14953571557998657, | |
| "learning_rate": 8.678172190625614e-05, | |
| "loss": 0.0431, | |
| "step": 17700 | |
| }, | |
| { | |
| "epoch": 0.8727201412041576, | |
| "grad_norm": 0.37678760290145874, | |
| "learning_rate": 8.727201412041577e-05, | |
| "loss": 0.0441, | |
| "step": 17800 | |
| }, | |
| { | |
| "epoch": 0.877623063345754, | |
| "grad_norm": 0.22045595943927765, | |
| "learning_rate": 8.77623063345754e-05, | |
| "loss": 0.0469, | |
| "step": 17900 | |
| }, | |
| { | |
| "epoch": 0.8825259854873505, | |
| "grad_norm": 0.2513475716114044, | |
| "learning_rate": 8.825259854873505e-05, | |
| "loss": 0.0397, | |
| "step": 18000 | |
| }, | |
| { | |
| "epoch": 0.8874289076289469, | |
| "grad_norm": 0.21822652220726013, | |
| "learning_rate": 8.874289076289469e-05, | |
| "loss": 0.0461, | |
| "step": 18100 | |
| }, | |
| { | |
| "epoch": 0.8923318297705433, | |
| "grad_norm": 0.31887462735176086, | |
| "learning_rate": 8.923318297705433e-05, | |
| "loss": 0.0404, | |
| "step": 18200 | |
| }, | |
| { | |
| "epoch": 0.8972347519121396, | |
| "grad_norm": 0.3159906268119812, | |
| "learning_rate": 8.972347519121397e-05, | |
| "loss": 0.041, | |
| "step": 18300 | |
| }, | |
| { | |
| "epoch": 0.902137674053736, | |
| "grad_norm": 0.20168788731098175, | |
| "learning_rate": 9.021376740537361e-05, | |
| "loss": 0.0437, | |
| "step": 18400 | |
| }, | |
| { | |
| "epoch": 0.9070405961953324, | |
| "grad_norm": 0.45709195733070374, | |
| "learning_rate": 9.070405961953325e-05, | |
| "loss": 0.0425, | |
| "step": 18500 | |
| }, | |
| { | |
| "epoch": 0.9119435183369288, | |
| "grad_norm": 0.2685320973396301, | |
| "learning_rate": 9.119435183369288e-05, | |
| "loss": 0.0418, | |
| "step": 18600 | |
| }, | |
| { | |
| "epoch": 0.9168464404785251, | |
| "grad_norm": 0.14851762354373932, | |
| "learning_rate": 9.168464404785252e-05, | |
| "loss": 0.042, | |
| "step": 18700 | |
| }, | |
| { | |
| "epoch": 0.9217493626201216, | |
| "grad_norm": 0.1795952171087265, | |
| "learning_rate": 9.217493626201216e-05, | |
| "loss": 0.0417, | |
| "step": 18800 | |
| }, | |
| { | |
| "epoch": 0.926652284761718, | |
| "grad_norm": 0.11718897521495819, | |
| "learning_rate": 9.26652284761718e-05, | |
| "loss": 0.0456, | |
| "step": 18900 | |
| }, | |
| { | |
| "epoch": 0.9315552069033144, | |
| "grad_norm": 0.11578471958637238, | |
| "learning_rate": 9.315552069033144e-05, | |
| "loss": 0.0421, | |
| "step": 19000 | |
| }, | |
| { | |
| "epoch": 0.9364581290449108, | |
| "grad_norm": 0.2343079149723053, | |
| "learning_rate": 9.364581290449108e-05, | |
| "loss": 0.0406, | |
| "step": 19100 | |
| }, | |
| { | |
| "epoch": 0.9413610511865071, | |
| "grad_norm": 0.29279735684394836, | |
| "learning_rate": 9.413610511865072e-05, | |
| "loss": 0.0395, | |
| "step": 19200 | |
| }, | |
| { | |
| "epoch": 0.9462639733281035, | |
| "grad_norm": 0.2973973751068115, | |
| "learning_rate": 9.462639733281036e-05, | |
| "loss": 0.0343, | |
| "step": 19300 | |
| }, | |
| { | |
| "epoch": 0.9511668954696999, | |
| "grad_norm": 0.4062161147594452, | |
| "learning_rate": 9.511668954697e-05, | |
| "loss": 0.0442, | |
| "step": 19400 | |
| }, | |
| { | |
| "epoch": 0.9560698176112963, | |
| "grad_norm": 0.4501809775829315, | |
| "learning_rate": 9.560698176112963e-05, | |
| "loss": 0.0425, | |
| "step": 19500 | |
| }, | |
| { | |
| "epoch": 0.9609727397528928, | |
| "grad_norm": 0.2221497744321823, | |
| "learning_rate": 9.609727397528929e-05, | |
| "loss": 0.0415, | |
| "step": 19600 | |
| }, | |
| { | |
| "epoch": 0.9658756618944891, | |
| "grad_norm": 0.2941090762615204, | |
| "learning_rate": 9.658756618944892e-05, | |
| "loss": 0.0409, | |
| "step": 19700 | |
| }, | |
| { | |
| "epoch": 0.9707785840360855, | |
| "grad_norm": 0.32935085892677307, | |
| "learning_rate": 9.707785840360856e-05, | |
| "loss": 0.0436, | |
| "step": 19800 | |
| }, | |
| { | |
| "epoch": 0.9756815061776819, | |
| "grad_norm": 0.31291815638542175, | |
| "learning_rate": 9.756815061776819e-05, | |
| "loss": 0.0407, | |
| "step": 19900 | |
| }, | |
| { | |
| "epoch": 0.9805844283192783, | |
| "grad_norm": 0.3670090138912201, | |
| "learning_rate": 9.805844283192783e-05, | |
| "loss": 0.0415, | |
| "step": 20000 | |
| }, | |
| { | |
| "epoch": 0.9854873504608747, | |
| "grad_norm": 0.29986053705215454, | |
| "learning_rate": 9.854873504608747e-05, | |
| "loss": 0.0453, | |
| "step": 20100 | |
| }, | |
| { | |
| "epoch": 0.990390272602471, | |
| "grad_norm": 0.4767267405986786, | |
| "learning_rate": 9.90390272602471e-05, | |
| "loss": 0.0422, | |
| "step": 20200 | |
| }, | |
| { | |
| "epoch": 0.9952931947440674, | |
| "grad_norm": 0.45798006653785706, | |
| "learning_rate": 9.952931947440674e-05, | |
| "loss": 0.0422, | |
| "step": 20300 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "eval_loss": 0.03229685500264168, | |
| "eval_runtime": 221.6122, | |
| "eval_samples_per_second": 22.562, | |
| "eval_steps_per_second": 22.562, | |
| "step": 20396 | |
| }, | |
| { | |
| "epoch": 1.000196116885664, | |
| "grad_norm": 0.3639116585254669, | |
| "learning_rate": 9.999999988283857e-05, | |
| "loss": 0.0414, | |
| "step": 20400 | |
| }, | |
| { | |
| "epoch": 1.0050990390272603, | |
| "grad_norm": 0.2748640179634094, | |
| "learning_rate": 9.999992079888559e-05, | |
| "loss": 0.0335, | |
| "step": 20500 | |
| }, | |
| { | |
| "epoch": 1.0100019611688567, | |
| "grad_norm": 0.1961342841386795, | |
| "learning_rate": 9.999969526339681e-05, | |
| "loss": 0.0269, | |
| "step": 20600 | |
| }, | |
| { | |
| "epoch": 1.014904883310453, | |
| "grad_norm": 0.15799719095230103, | |
| "learning_rate": 9.999932327703287e-05, | |
| "loss": 0.0286, | |
| "step": 20700 | |
| }, | |
| { | |
| "epoch": 1.0198078054520494, | |
| "grad_norm": 0.26234665513038635, | |
| "learning_rate": 9.999880484088331e-05, | |
| "loss": 0.0293, | |
| "step": 20800 | |
| }, | |
| { | |
| "epoch": 1.0247107275936458, | |
| "grad_norm": 0.24818925559520721, | |
| "learning_rate": 9.999813995646665e-05, | |
| "loss": 0.0305, | |
| "step": 20900 | |
| }, | |
| { | |
| "epoch": 1.0296136497352422, | |
| "grad_norm": 0.39123162627220154, | |
| "learning_rate": 9.999732862573036e-05, | |
| "loss": 0.0309, | |
| "step": 21000 | |
| }, | |
| { | |
| "epoch": 1.0345165718768385, | |
| "grad_norm": 0.27162662148475647, | |
| "learning_rate": 9.999637085105086e-05, | |
| "loss": 0.0292, | |
| "step": 21100 | |
| }, | |
| { | |
| "epoch": 1.039419494018435, | |
| "grad_norm": 0.22811737656593323, | |
| "learning_rate": 9.99952666352335e-05, | |
| "loss": 0.0257, | |
| "step": 21200 | |
| }, | |
| { | |
| "epoch": 1.0443224161600313, | |
| "grad_norm": 0.19570066034793854, | |
| "learning_rate": 9.999401598151257e-05, | |
| "loss": 0.0301, | |
| "step": 21300 | |
| }, | |
| { | |
| "epoch": 1.0492253383016277, | |
| "grad_norm": 0.2861987054347992, | |
| "learning_rate": 9.999261889355128e-05, | |
| "loss": 0.0311, | |
| "step": 21400 | |
| }, | |
| { | |
| "epoch": 1.0541282604432243, | |
| "grad_norm": 0.1091461256146431, | |
| "learning_rate": 9.999107537544174e-05, | |
| "loss": 0.0295, | |
| "step": 21500 | |
| }, | |
| { | |
| "epoch": 1.0590311825848207, | |
| "grad_norm": 0.1704551875591278, | |
| "learning_rate": 9.998938543170499e-05, | |
| "loss": 0.0303, | |
| "step": 21600 | |
| }, | |
| { | |
| "epoch": 1.063934104726417, | |
| "grad_norm": 0.25216275453567505, | |
| "learning_rate": 9.998754906729093e-05, | |
| "loss": 0.0342, | |
| "step": 21700 | |
| }, | |
| { | |
| "epoch": 1.0688370268680134, | |
| "grad_norm": 0.17220348119735718, | |
| "learning_rate": 9.998556628757832e-05, | |
| "loss": 0.0293, | |
| "step": 21800 | |
| }, | |
| { | |
| "epoch": 1.0737399490096098, | |
| "grad_norm": 0.46171247959136963, | |
| "learning_rate": 9.998343709837481e-05, | |
| "loss": 0.0298, | |
| "step": 21900 | |
| }, | |
| { | |
| "epoch": 1.0786428711512062, | |
| "grad_norm": 0.23441363871097565, | |
| "learning_rate": 9.998116150591685e-05, | |
| "loss": 0.0295, | |
| "step": 22000 | |
| }, | |
| { | |
| "epoch": 1.0835457932928025, | |
| "grad_norm": 0.3156551718711853, | |
| "learning_rate": 9.997873951686976e-05, | |
| "loss": 0.0286, | |
| "step": 22100 | |
| }, | |
| { | |
| "epoch": 1.088448715434399, | |
| "grad_norm": 0.43607401847839355, | |
| "learning_rate": 9.99761711383276e-05, | |
| "loss": 0.029, | |
| "step": 22200 | |
| }, | |
| { | |
| "epoch": 1.0933516375759953, | |
| "grad_norm": 0.11727584898471832, | |
| "learning_rate": 9.997345637781325e-05, | |
| "loss": 0.0288, | |
| "step": 22300 | |
| }, | |
| { | |
| "epoch": 1.0982545597175917, | |
| "grad_norm": 0.09930656105279922, | |
| "learning_rate": 9.997059524327837e-05, | |
| "loss": 0.0298, | |
| "step": 22400 | |
| }, | |
| { | |
| "epoch": 1.103157481859188, | |
| "grad_norm": 0.28237900137901306, | |
| "learning_rate": 9.996758774310326e-05, | |
| "loss": 0.0305, | |
| "step": 22500 | |
| }, | |
| { | |
| "epoch": 1.1080604040007844, | |
| "grad_norm": 0.318491667509079, | |
| "learning_rate": 9.996443388609706e-05, | |
| "loss": 0.0315, | |
| "step": 22600 | |
| }, | |
| { | |
| "epoch": 1.1129633261423808, | |
| "grad_norm": 0.24862946569919586, | |
| "learning_rate": 9.996113368149751e-05, | |
| "loss": 0.0308, | |
| "step": 22700 | |
| }, | |
| { | |
| "epoch": 1.1178662482839772, | |
| "grad_norm": 0.2209358662366867, | |
| "learning_rate": 9.9957687138971e-05, | |
| "loss": 0.0272, | |
| "step": 22800 | |
| }, | |
| { | |
| "epoch": 1.1227691704255736, | |
| "grad_norm": 0.20178958773612976, | |
| "learning_rate": 9.995409426861261e-05, | |
| "loss": 0.0301, | |
| "step": 22900 | |
| }, | |
| { | |
| "epoch": 1.12767209256717, | |
| "grad_norm": 0.05385458096861839, | |
| "learning_rate": 9.995035508094596e-05, | |
| "loss": 0.0293, | |
| "step": 23000 | |
| }, | |
| { | |
| "epoch": 1.1325750147087663, | |
| "grad_norm": 0.42040759325027466, | |
| "learning_rate": 9.994646958692327e-05, | |
| "loss": 0.0278, | |
| "step": 23100 | |
| }, | |
| { | |
| "epoch": 1.1374779368503627, | |
| "grad_norm": 0.2339663952589035, | |
| "learning_rate": 9.99424377979253e-05, | |
| "loss": 0.0276, | |
| "step": 23200 | |
| }, | |
| { | |
| "epoch": 1.1423808589919593, | |
| "grad_norm": 0.3861044645309448, | |
| "learning_rate": 9.993825972576127e-05, | |
| "loss": 0.0288, | |
| "step": 23300 | |
| }, | |
| { | |
| "epoch": 1.1472837811335557, | |
| "grad_norm": 0.27347856760025024, | |
| "learning_rate": 9.993393538266892e-05, | |
| "loss": 0.0301, | |
| "step": 23400 | |
| }, | |
| { | |
| "epoch": 1.152186703275152, | |
| "grad_norm": 0.19038136303424835, | |
| "learning_rate": 9.992946478131444e-05, | |
| "loss": 0.0303, | |
| "step": 23500 | |
| }, | |
| { | |
| "epoch": 1.1570896254167484, | |
| "grad_norm": 0.23598672449588776, | |
| "learning_rate": 9.992484793479233e-05, | |
| "loss": 0.0309, | |
| "step": 23600 | |
| }, | |
| { | |
| "epoch": 1.1619925475583448, | |
| "grad_norm": 0.28980758786201477, | |
| "learning_rate": 9.992008485662552e-05, | |
| "loss": 0.0296, | |
| "step": 23700 | |
| }, | |
| { | |
| "epoch": 1.1668954696999412, | |
| "grad_norm": 0.19490975141525269, | |
| "learning_rate": 9.991517556076522e-05, | |
| "loss": 0.0296, | |
| "step": 23800 | |
| }, | |
| { | |
| "epoch": 1.1717983918415376, | |
| "grad_norm": 0.12678834795951843, | |
| "learning_rate": 9.991012006159094e-05, | |
| "loss": 0.0265, | |
| "step": 23900 | |
| }, | |
| { | |
| "epoch": 1.176701313983134, | |
| "grad_norm": 0.29768842458724976, | |
| "learning_rate": 9.990491837391042e-05, | |
| "loss": 0.03, | |
| "step": 24000 | |
| }, | |
| { | |
| "epoch": 1.1816042361247303, | |
| "grad_norm": 0.33209484815597534, | |
| "learning_rate": 9.989957051295957e-05, | |
| "loss": 0.0275, | |
| "step": 24100 | |
| }, | |
| { | |
| "epoch": 1.1865071582663267, | |
| "grad_norm": 0.24054765701293945, | |
| "learning_rate": 9.989407649440249e-05, | |
| "loss": 0.0262, | |
| "step": 24200 | |
| }, | |
| { | |
| "epoch": 1.191410080407923, | |
| "grad_norm": 0.1172015443444252, | |
| "learning_rate": 9.988843633433134e-05, | |
| "loss": 0.0306, | |
| "step": 24300 | |
| }, | |
| { | |
| "epoch": 1.1963130025495194, | |
| "grad_norm": 0.22456437349319458, | |
| "learning_rate": 9.988265004926633e-05, | |
| "loss": 0.0308, | |
| "step": 24400 | |
| }, | |
| { | |
| "epoch": 1.2012159246911158, | |
| "grad_norm": 0.17739421129226685, | |
| "learning_rate": 9.987671765615574e-05, | |
| "loss": 0.0311, | |
| "step": 24500 | |
| }, | |
| { | |
| "epoch": 1.2061188468327124, | |
| "grad_norm": 0.39153727889060974, | |
| "learning_rate": 9.987063917237572e-05, | |
| "loss": 0.0318, | |
| "step": 24600 | |
| }, | |
| { | |
| "epoch": 1.2110217689743088, | |
| "grad_norm": 0.1307838261127472, | |
| "learning_rate": 9.986441461573036e-05, | |
| "loss": 0.0267, | |
| "step": 24700 | |
| }, | |
| { | |
| "epoch": 1.2159246911159052, | |
| "grad_norm": 0.35021230578422546, | |
| "learning_rate": 9.985804400445163e-05, | |
| "loss": 0.0314, | |
| "step": 24800 | |
| }, | |
| { | |
| "epoch": 1.2208276132575016, | |
| "grad_norm": 0.20624665915966034, | |
| "learning_rate": 9.985152735719927e-05, | |
| "loss": 0.0332, | |
| "step": 24900 | |
| }, | |
| { | |
| "epoch": 1.225730535399098, | |
| "grad_norm": 0.32110336422920227, | |
| "learning_rate": 9.984486469306078e-05, | |
| "loss": 0.0281, | |
| "step": 25000 | |
| }, | |
| { | |
| "epoch": 1.2306334575406943, | |
| "grad_norm": 0.2653130888938904, | |
| "learning_rate": 9.983805603155131e-05, | |
| "loss": 0.0287, | |
| "step": 25100 | |
| }, | |
| { | |
| "epoch": 1.2355363796822907, | |
| "grad_norm": 0.11658957600593567, | |
| "learning_rate": 9.983110139261369e-05, | |
| "loss": 0.0315, | |
| "step": 25200 | |
| }, | |
| { | |
| "epoch": 1.240439301823887, | |
| "grad_norm": 0.5068228244781494, | |
| "learning_rate": 9.98240007966183e-05, | |
| "loss": 0.0251, | |
| "step": 25300 | |
| }, | |
| { | |
| "epoch": 1.2453422239654834, | |
| "grad_norm": 0.27341774106025696, | |
| "learning_rate": 9.981675426436304e-05, | |
| "loss": 0.0311, | |
| "step": 25400 | |
| }, | |
| { | |
| "epoch": 1.2502451461070798, | |
| "grad_norm": 0.22564496099948883, | |
| "learning_rate": 9.980936181707326e-05, | |
| "loss": 0.027, | |
| "step": 25500 | |
| }, | |
| { | |
| "epoch": 1.2551480682486762, | |
| "grad_norm": 0.33587944507598877, | |
| "learning_rate": 9.980182347640168e-05, | |
| "loss": 0.0285, | |
| "step": 25600 | |
| }, | |
| { | |
| "epoch": 1.2600509903902726, | |
| "grad_norm": 0.19802583754062653, | |
| "learning_rate": 9.97941392644284e-05, | |
| "loss": 0.0251, | |
| "step": 25700 | |
| }, | |
| { | |
| "epoch": 1.264953912531869, | |
| "grad_norm": 0.36870622634887695, | |
| "learning_rate": 9.978630920366072e-05, | |
| "loss": 0.0294, | |
| "step": 25800 | |
| }, | |
| { | |
| "epoch": 1.2698568346734653, | |
| "grad_norm": 0.2961844801902771, | |
| "learning_rate": 9.977833331703317e-05, | |
| "loss": 0.0294, | |
| "step": 25900 | |
| }, | |
| { | |
| "epoch": 1.2747597568150617, | |
| "grad_norm": 0.31916147470474243, | |
| "learning_rate": 9.977021162790743e-05, | |
| "loss": 0.0318, | |
| "step": 26000 | |
| }, | |
| { | |
| "epoch": 1.279662678956658, | |
| "grad_norm": 0.27534112334251404, | |
| "learning_rate": 9.976194416007217e-05, | |
| "loss": 0.0297, | |
| "step": 26100 | |
| }, | |
| { | |
| "epoch": 1.2845656010982545, | |
| "grad_norm": 0.30862635374069214, | |
| "learning_rate": 9.975353093774313e-05, | |
| "loss": 0.028, | |
| "step": 26200 | |
| }, | |
| { | |
| "epoch": 1.2894685232398508, | |
| "grad_norm": 0.364801824092865, | |
| "learning_rate": 9.974497198556292e-05, | |
| "loss": 0.0286, | |
| "step": 26300 | |
| }, | |
| { | |
| "epoch": 1.2943714453814472, | |
| "grad_norm": 0.39874985814094543, | |
| "learning_rate": 9.973626732860102e-05, | |
| "loss": 0.0302, | |
| "step": 26400 | |
| }, | |
| { | |
| "epoch": 1.2992743675230438, | |
| "grad_norm": 0.3267800509929657, | |
| "learning_rate": 9.972741699235368e-05, | |
| "loss": 0.0314, | |
| "step": 26500 | |
| }, | |
| { | |
| "epoch": 1.3041772896646402, | |
| "grad_norm": 0.24671390652656555, | |
| "learning_rate": 9.971842100274384e-05, | |
| "loss": 0.0258, | |
| "step": 26600 | |
| }, | |
| { | |
| "epoch": 1.3090802118062366, | |
| "grad_norm": 0.3735731840133667, | |
| "learning_rate": 9.97092793861211e-05, | |
| "loss": 0.0291, | |
| "step": 26700 | |
| }, | |
| { | |
| "epoch": 1.313983133947833, | |
| "grad_norm": 0.13338430225849152, | |
| "learning_rate": 9.969999216926151e-05, | |
| "loss": 0.0301, | |
| "step": 26800 | |
| }, | |
| { | |
| "epoch": 1.3188860560894293, | |
| "grad_norm": 0.23581305146217346, | |
| "learning_rate": 9.969055937936773e-05, | |
| "loss": 0.0309, | |
| "step": 26900 | |
| }, | |
| { | |
| "epoch": 1.3237889782310257, | |
| "grad_norm": 0.09043537825345993, | |
| "learning_rate": 9.96809810440687e-05, | |
| "loss": 0.0274, | |
| "step": 27000 | |
| }, | |
| { | |
| "epoch": 1.328691900372622, | |
| "grad_norm": 0.36778169870376587, | |
| "learning_rate": 9.96712571914197e-05, | |
| "loss": 0.0306, | |
| "step": 27100 | |
| }, | |
| { | |
| "epoch": 1.3335948225142185, | |
| "grad_norm": 0.12855975329875946, | |
| "learning_rate": 9.966138784990227e-05, | |
| "loss": 0.023, | |
| "step": 27200 | |
| }, | |
| { | |
| "epoch": 1.3384977446558148, | |
| "grad_norm": 0.5465484261512756, | |
| "learning_rate": 9.965137304842402e-05, | |
| "loss": 0.0287, | |
| "step": 27300 | |
| }, | |
| { | |
| "epoch": 1.3434006667974112, | |
| "grad_norm": 0.22656790912151337, | |
| "learning_rate": 9.964121281631866e-05, | |
| "loss": 0.0303, | |
| "step": 27400 | |
| }, | |
| { | |
| "epoch": 1.3483035889390076, | |
| "grad_norm": 0.334923654794693, | |
| "learning_rate": 9.96309071833459e-05, | |
| "loss": 0.0265, | |
| "step": 27500 | |
| }, | |
| { | |
| "epoch": 1.353206511080604, | |
| "grad_norm": 0.26427406072616577, | |
| "learning_rate": 9.96204561796913e-05, | |
| "loss": 0.0299, | |
| "step": 27600 | |
| }, | |
| { | |
| "epoch": 1.3581094332222006, | |
| "grad_norm": 0.1943475604057312, | |
| "learning_rate": 9.960985983596618e-05, | |
| "loss": 0.028, | |
| "step": 27700 | |
| }, | |
| { | |
| "epoch": 1.363012355363797, | |
| "grad_norm": 0.3269706070423126, | |
| "learning_rate": 9.959911818320764e-05, | |
| "loss": 0.0271, | |
| "step": 27800 | |
| }, | |
| { | |
| "epoch": 1.3679152775053933, | |
| "grad_norm": 0.2163461148738861, | |
| "learning_rate": 9.958823125287836e-05, | |
| "loss": 0.0295, | |
| "step": 27900 | |
| }, | |
| { | |
| "epoch": 1.3728181996469897, | |
| "grad_norm": 0.24624493718147278, | |
| "learning_rate": 9.957719907686654e-05, | |
| "loss": 0.0343, | |
| "step": 28000 | |
| }, | |
| { | |
| "epoch": 1.377721121788586, | |
| "grad_norm": 0.15970952808856964, | |
| "learning_rate": 9.95660216874858e-05, | |
| "loss": 0.0301, | |
| "step": 28100 | |
| }, | |
| { | |
| "epoch": 1.3826240439301825, | |
| "grad_norm": 0.521918773651123, | |
| "learning_rate": 9.955469911747511e-05, | |
| "loss": 0.0314, | |
| "step": 28200 | |
| }, | |
| { | |
| "epoch": 1.3875269660717788, | |
| "grad_norm": 0.17727439105510712, | |
| "learning_rate": 9.954323139999871e-05, | |
| "loss": 0.0276, | |
| "step": 28300 | |
| }, | |
| { | |
| "epoch": 1.3924298882133752, | |
| "grad_norm": 0.3565279245376587, | |
| "learning_rate": 9.95316185686459e-05, | |
| "loss": 0.0259, | |
| "step": 28400 | |
| }, | |
| { | |
| "epoch": 1.3973328103549716, | |
| "grad_norm": 0.17430347204208374, | |
| "learning_rate": 9.951986065743111e-05, | |
| "loss": 0.03, | |
| "step": 28500 | |
| }, | |
| { | |
| "epoch": 1.402235732496568, | |
| "grad_norm": 0.13704228401184082, | |
| "learning_rate": 9.950795770079368e-05, | |
| "loss": 0.0274, | |
| "step": 28600 | |
| }, | |
| { | |
| "epoch": 1.4071386546381643, | |
| "grad_norm": 0.23465342819690704, | |
| "learning_rate": 9.949590973359778e-05, | |
| "loss": 0.0266, | |
| "step": 28700 | |
| }, | |
| { | |
| "epoch": 1.4120415767797607, | |
| "grad_norm": 0.18756377696990967, | |
| "learning_rate": 9.948371679113229e-05, | |
| "loss": 0.0274, | |
| "step": 28800 | |
| }, | |
| { | |
| "epoch": 1.416944498921357, | |
| "grad_norm": 0.17243963479995728, | |
| "learning_rate": 9.947137890911084e-05, | |
| "loss": 0.0254, | |
| "step": 28900 | |
| }, | |
| { | |
| "epoch": 1.4218474210629535, | |
| "grad_norm": 0.14038296043872833, | |
| "learning_rate": 9.945889612367149e-05, | |
| "loss": 0.0252, | |
| "step": 29000 | |
| }, | |
| { | |
| "epoch": 1.4267503432045499, | |
| "grad_norm": 0.25634902715682983, | |
| "learning_rate": 9.944626847137677e-05, | |
| "loss": 0.0288, | |
| "step": 29100 | |
| }, | |
| { | |
| "epoch": 1.4316532653461462, | |
| "grad_norm": 0.4887700080871582, | |
| "learning_rate": 9.94334959892135e-05, | |
| "loss": 0.0288, | |
| "step": 29200 | |
| }, | |
| { | |
| "epoch": 1.4365561874877426, | |
| "grad_norm": 0.33462828397750854, | |
| "learning_rate": 9.942057871459275e-05, | |
| "loss": 0.0295, | |
| "step": 29300 | |
| }, | |
| { | |
| "epoch": 1.441459109629339, | |
| "grad_norm": 0.3429461419582367, | |
| "learning_rate": 9.940751668534966e-05, | |
| "loss": 0.0272, | |
| "step": 29400 | |
| }, | |
| { | |
| "epoch": 1.4463620317709354, | |
| "grad_norm": 0.3318476676940918, | |
| "learning_rate": 9.939430993974339e-05, | |
| "loss": 0.0288, | |
| "step": 29500 | |
| }, | |
| { | |
| "epoch": 1.4512649539125317, | |
| "grad_norm": 0.3497527241706848, | |
| "learning_rate": 9.938095851645696e-05, | |
| "loss": 0.0314, | |
| "step": 29600 | |
| }, | |
| { | |
| "epoch": 1.4561678760541283, | |
| "grad_norm": 0.38459303975105286, | |
| "learning_rate": 9.936746245459716e-05, | |
| "loss": 0.0284, | |
| "step": 29700 | |
| }, | |
| { | |
| "epoch": 1.4610707981957247, | |
| "grad_norm": 0.34776678681373596, | |
| "learning_rate": 9.935382179369442e-05, | |
| "loss": 0.0254, | |
| "step": 29800 | |
| }, | |
| { | |
| "epoch": 1.465973720337321, | |
| "grad_norm": 0.13150684535503387, | |
| "learning_rate": 9.934003657370275e-05, | |
| "loss": 0.0274, | |
| "step": 29900 | |
| }, | |
| { | |
| "epoch": 1.4708766424789175, | |
| "grad_norm": 0.3473518192768097, | |
| "learning_rate": 9.932610683499951e-05, | |
| "loss": 0.0287, | |
| "step": 30000 | |
| }, | |
| { | |
| "epoch": 1.4757795646205139, | |
| "grad_norm": 0.10194269567728043, | |
| "learning_rate": 9.931203261838542e-05, | |
| "loss": 0.0265, | |
| "step": 30100 | |
| }, | |
| { | |
| "epoch": 1.4806824867621102, | |
| "grad_norm": 0.2663326859474182, | |
| "learning_rate": 9.929781396508435e-05, | |
| "loss": 0.0256, | |
| "step": 30200 | |
| }, | |
| { | |
| "epoch": 1.4855854089037066, | |
| "grad_norm": 0.37899428606033325, | |
| "learning_rate": 9.928345091674324e-05, | |
| "loss": 0.0287, | |
| "step": 30300 | |
| }, | |
| { | |
| "epoch": 1.490488331045303, | |
| "grad_norm": 0.16388994455337524, | |
| "learning_rate": 9.926894351543196e-05, | |
| "loss": 0.0281, | |
| "step": 30400 | |
| }, | |
| { | |
| "epoch": 1.4953912531868994, | |
| "grad_norm": 0.1827380210161209, | |
| "learning_rate": 9.92542918036432e-05, | |
| "loss": 0.0279, | |
| "step": 30500 | |
| }, | |
| { | |
| "epoch": 1.5002941753284957, | |
| "grad_norm": 0.3441345691680908, | |
| "learning_rate": 9.923949582429237e-05, | |
| "loss": 0.0299, | |
| "step": 30600 | |
| }, | |
| { | |
| "epoch": 1.5051970974700923, | |
| "grad_norm": 0.22477801144123077, | |
| "learning_rate": 9.92245556207174e-05, | |
| "loss": 0.0263, | |
| "step": 30700 | |
| }, | |
| { | |
| "epoch": 1.5101000196116887, | |
| "grad_norm": 0.2690620422363281, | |
| "learning_rate": 9.920947123667866e-05, | |
| "loss": 0.0281, | |
| "step": 30800 | |
| }, | |
| { | |
| "epoch": 1.515002941753285, | |
| "grad_norm": 0.3723180592060089, | |
| "learning_rate": 9.919424271635884e-05, | |
| "loss": 0.0275, | |
| "step": 30900 | |
| }, | |
| { | |
| "epoch": 1.5199058638948815, | |
| "grad_norm": 0.4014623165130615, | |
| "learning_rate": 9.917887010436285e-05, | |
| "loss": 0.0259, | |
| "step": 31000 | |
| }, | |
| { | |
| "epoch": 1.5248087860364778, | |
| "grad_norm": 0.29620611667633057, | |
| "learning_rate": 9.916335344571758e-05, | |
| "loss": 0.0281, | |
| "step": 31100 | |
| }, | |
| { | |
| "epoch": 1.5297117081780742, | |
| "grad_norm": 0.3609066605567932, | |
| "learning_rate": 9.914769278587189e-05, | |
| "loss": 0.0249, | |
| "step": 31200 | |
| }, | |
| { | |
| "epoch": 1.5346146303196706, | |
| "grad_norm": 0.20832450687885284, | |
| "learning_rate": 9.913188817069641e-05, | |
| "loss": 0.0306, | |
| "step": 31300 | |
| }, | |
| { | |
| "epoch": 1.539517552461267, | |
| "grad_norm": 0.14223641157150269, | |
| "learning_rate": 9.91159396464834e-05, | |
| "loss": 0.0261, | |
| "step": 31400 | |
| }, | |
| { | |
| "epoch": 1.5444204746028634, | |
| "grad_norm": 0.2740347981452942, | |
| "learning_rate": 9.909984725994666e-05, | |
| "loss": 0.0306, | |
| "step": 31500 | |
| }, | |
| { | |
| "epoch": 1.5493233967444597, | |
| "grad_norm": 0.2499512881040573, | |
| "learning_rate": 9.908361105822134e-05, | |
| "loss": 0.0271, | |
| "step": 31600 | |
| }, | |
| { | |
| "epoch": 1.5542263188860561, | |
| "grad_norm": 0.32326480746269226, | |
| "learning_rate": 9.90672310888639e-05, | |
| "loss": 0.0273, | |
| "step": 31700 | |
| }, | |
| { | |
| "epoch": 1.5591292410276525, | |
| "grad_norm": 0.3028701841831207, | |
| "learning_rate": 9.90507073998518e-05, | |
| "loss": 0.0277, | |
| "step": 31800 | |
| }, | |
| { | |
| "epoch": 1.5640321631692489, | |
| "grad_norm": 0.3670022487640381, | |
| "learning_rate": 9.903404003958352e-05, | |
| "loss": 0.026, | |
| "step": 31900 | |
| }, | |
| { | |
| "epoch": 1.5689350853108452, | |
| "grad_norm": 0.19097840785980225, | |
| "learning_rate": 9.901722905687835e-05, | |
| "loss": 0.0245, | |
| "step": 32000 | |
| }, | |
| { | |
| "epoch": 1.5738380074524416, | |
| "grad_norm": 0.3168416917324066, | |
| "learning_rate": 9.900027450097629e-05, | |
| "loss": 0.0273, | |
| "step": 32100 | |
| }, | |
| { | |
| "epoch": 1.578740929594038, | |
| "grad_norm": 0.3735883831977844, | |
| "learning_rate": 9.898317642153776e-05, | |
| "loss": 0.0285, | |
| "step": 32200 | |
| }, | |
| { | |
| "epoch": 1.5836438517356344, | |
| "grad_norm": 0.17389927804470062, | |
| "learning_rate": 9.896593486864369e-05, | |
| "loss": 0.0279, | |
| "step": 32300 | |
| }, | |
| { | |
| "epoch": 1.5885467738772308, | |
| "grad_norm": 0.402938574552536, | |
| "learning_rate": 9.89485498927952e-05, | |
| "loss": 0.0271, | |
| "step": 32400 | |
| }, | |
| { | |
| "epoch": 1.5934496960188271, | |
| "grad_norm": 0.11623658984899521, | |
| "learning_rate": 9.893102154491348e-05, | |
| "loss": 0.0282, | |
| "step": 32500 | |
| }, | |
| { | |
| "epoch": 1.5983526181604235, | |
| "grad_norm": 0.4419660270214081, | |
| "learning_rate": 9.891334987633969e-05, | |
| "loss": 0.0274, | |
| "step": 32600 | |
| }, | |
| { | |
| "epoch": 1.6032555403020199, | |
| "grad_norm": 0.19593150913715363, | |
| "learning_rate": 9.889553493883476e-05, | |
| "loss": 0.0276, | |
| "step": 32700 | |
| }, | |
| { | |
| "epoch": 1.6081584624436163, | |
| "grad_norm": 0.313765287399292, | |
| "learning_rate": 9.887757678457927e-05, | |
| "loss": 0.0302, | |
| "step": 32800 | |
| }, | |
| { | |
| "epoch": 1.6130613845852126, | |
| "grad_norm": 0.15113359689712524, | |
| "learning_rate": 9.885947546617332e-05, | |
| "loss": 0.0289, | |
| "step": 32900 | |
| }, | |
| { | |
| "epoch": 1.617964306726809, | |
| "grad_norm": 0.16863977909088135, | |
| "learning_rate": 9.884123103663629e-05, | |
| "loss": 0.0249, | |
| "step": 33000 | |
| }, | |
| { | |
| "epoch": 1.6228672288684056, | |
| "grad_norm": 0.4099023640155792, | |
| "learning_rate": 9.882284354940674e-05, | |
| "loss": 0.0272, | |
| "step": 33100 | |
| }, | |
| { | |
| "epoch": 1.627770151010002, | |
| "grad_norm": 0.18381527066230774, | |
| "learning_rate": 9.88043130583423e-05, | |
| "loss": 0.0256, | |
| "step": 33200 | |
| }, | |
| { | |
| "epoch": 1.6326730731515984, | |
| "grad_norm": 0.21517693996429443, | |
| "learning_rate": 9.878563961771942e-05, | |
| "loss": 0.0276, | |
| "step": 33300 | |
| }, | |
| { | |
| "epoch": 1.6375759952931948, | |
| "grad_norm": 0.23215888440608978, | |
| "learning_rate": 9.876682328223329e-05, | |
| "loss": 0.0231, | |
| "step": 33400 | |
| }, | |
| { | |
| "epoch": 1.6424789174347911, | |
| "grad_norm": 0.31246307492256165, | |
| "learning_rate": 9.87478641069976e-05, | |
| "loss": 0.0248, | |
| "step": 33500 | |
| }, | |
| { | |
| "epoch": 1.6473818395763875, | |
| "grad_norm": 0.2109220176935196, | |
| "learning_rate": 9.872876214754442e-05, | |
| "loss": 0.026, | |
| "step": 33600 | |
| }, | |
| { | |
| "epoch": 1.6522847617179839, | |
| "grad_norm": 0.17266134917736053, | |
| "learning_rate": 9.870951745982411e-05, | |
| "loss": 0.0305, | |
| "step": 33700 | |
| }, | |
| { | |
| "epoch": 1.6571876838595803, | |
| "grad_norm": 0.24855448305606842, | |
| "learning_rate": 9.869013010020504e-05, | |
| "loss": 0.0268, | |
| "step": 33800 | |
| }, | |
| { | |
| "epoch": 1.6620906060011769, | |
| "grad_norm": 0.4860318899154663, | |
| "learning_rate": 9.867060012547343e-05, | |
| "loss": 0.0279, | |
| "step": 33900 | |
| }, | |
| { | |
| "epoch": 1.6669935281427732, | |
| "grad_norm": 0.17674511671066284, | |
| "learning_rate": 9.86509275928333e-05, | |
| "loss": 0.0284, | |
| "step": 34000 | |
| }, | |
| { | |
| "epoch": 1.6718964502843696, | |
| "grad_norm": 0.25708097219467163, | |
| "learning_rate": 9.863111255990618e-05, | |
| "loss": 0.028, | |
| "step": 34100 | |
| }, | |
| { | |
| "epoch": 1.676799372425966, | |
| "grad_norm": 0.18099327385425568, | |
| "learning_rate": 9.861115508473102e-05, | |
| "loss": 0.0277, | |
| "step": 34200 | |
| }, | |
| { | |
| "epoch": 1.6817022945675624, | |
| "grad_norm": 0.6304388046264648, | |
| "learning_rate": 9.859105522576395e-05, | |
| "loss": 0.0275, | |
| "step": 34300 | |
| }, | |
| { | |
| "epoch": 1.6866052167091587, | |
| "grad_norm": 0.3071998655796051, | |
| "learning_rate": 9.857081304187817e-05, | |
| "loss": 0.03, | |
| "step": 34400 | |
| }, | |
| { | |
| "epoch": 1.6915081388507551, | |
| "grad_norm": 0.2949463725090027, | |
| "learning_rate": 9.855042859236376e-05, | |
| "loss": 0.0256, | |
| "step": 34500 | |
| }, | |
| { | |
| "epoch": 1.6964110609923515, | |
| "grad_norm": 0.22837311029434204, | |
| "learning_rate": 9.852990193692748e-05, | |
| "loss": 0.0249, | |
| "step": 34600 | |
| }, | |
| { | |
| "epoch": 1.7013139831339479, | |
| "grad_norm": 0.21440072357654572, | |
| "learning_rate": 9.850923313569266e-05, | |
| "loss": 0.0261, | |
| "step": 34700 | |
| }, | |
| { | |
| "epoch": 1.7062169052755443, | |
| "grad_norm": 0.2580870985984802, | |
| "learning_rate": 9.848842224919891e-05, | |
| "loss": 0.0257, | |
| "step": 34800 | |
| }, | |
| { | |
| "epoch": 1.7111198274171406, | |
| "grad_norm": 0.23171992599964142, | |
| "learning_rate": 9.846746933840209e-05, | |
| "loss": 0.0284, | |
| "step": 34900 | |
| }, | |
| { | |
| "epoch": 1.716022749558737, | |
| "grad_norm": 0.18120083212852478, | |
| "learning_rate": 9.844637446467399e-05, | |
| "loss": 0.0262, | |
| "step": 35000 | |
| }, | |
| { | |
| "epoch": 1.7209256717003334, | |
| "grad_norm": 0.2012002021074295, | |
| "learning_rate": 9.842513768980226e-05, | |
| "loss": 0.0255, | |
| "step": 35100 | |
| }, | |
| { | |
| "epoch": 1.7258285938419298, | |
| "grad_norm": 0.2966020405292511, | |
| "learning_rate": 9.840375907599013e-05, | |
| "loss": 0.025, | |
| "step": 35200 | |
| }, | |
| { | |
| "epoch": 1.7307315159835261, | |
| "grad_norm": 0.1654832810163498, | |
| "learning_rate": 9.838223868585637e-05, | |
| "loss": 0.0265, | |
| "step": 35300 | |
| }, | |
| { | |
| "epoch": 1.7356344381251225, | |
| "grad_norm": 0.21913449466228485, | |
| "learning_rate": 9.836057658243491e-05, | |
| "loss": 0.0282, | |
| "step": 35400 | |
| }, | |
| { | |
| "epoch": 1.740537360266719, | |
| "grad_norm": 0.21052640676498413, | |
| "learning_rate": 9.833877282917485e-05, | |
| "loss": 0.0247, | |
| "step": 35500 | |
| }, | |
| { | |
| "epoch": 1.7454402824083153, | |
| "grad_norm": 0.27234575152397156, | |
| "learning_rate": 9.831682748994013e-05, | |
| "loss": 0.0279, | |
| "step": 35600 | |
| }, | |
| { | |
| "epoch": 1.7503432045499117, | |
| "grad_norm": 0.30390360951423645, | |
| "learning_rate": 9.829474062900945e-05, | |
| "loss": 0.0281, | |
| "step": 35700 | |
| }, | |
| { | |
| "epoch": 1.755246126691508, | |
| "grad_norm": 0.3458337187767029, | |
| "learning_rate": 9.827251231107599e-05, | |
| "loss": 0.024, | |
| "step": 35800 | |
| }, | |
| { | |
| "epoch": 1.7601490488331044, | |
| "grad_norm": 0.26846686005592346, | |
| "learning_rate": 9.82501426012473e-05, | |
| "loss": 0.0316, | |
| "step": 35900 | |
| }, | |
| { | |
| "epoch": 1.7650519709747008, | |
| "grad_norm": 0.21699777245521545, | |
| "learning_rate": 9.822763156504502e-05, | |
| "loss": 0.0274, | |
| "step": 36000 | |
| }, | |
| { | |
| "epoch": 1.7699548931162972, | |
| "grad_norm": 0.25235772132873535, | |
| "learning_rate": 9.82049792684048e-05, | |
| "loss": 0.0273, | |
| "step": 36100 | |
| }, | |
| { | |
| "epoch": 1.7748578152578935, | |
| "grad_norm": 0.06357789784669876, | |
| "learning_rate": 9.818218577767599e-05, | |
| "loss": 0.0233, | |
| "step": 36200 | |
| }, | |
| { | |
| "epoch": 1.7797607373994901, | |
| "grad_norm": 0.25367268919944763, | |
| "learning_rate": 9.815925115962155e-05, | |
| "loss": 0.0249, | |
| "step": 36300 | |
| }, | |
| { | |
| "epoch": 1.7846636595410865, | |
| "grad_norm": 0.23693925142288208, | |
| "learning_rate": 9.813617548141776e-05, | |
| "loss": 0.0248, | |
| "step": 36400 | |
| }, | |
| { | |
| "epoch": 1.789566581682683, | |
| "grad_norm": 0.08688795566558838, | |
| "learning_rate": 9.811295881065415e-05, | |
| "loss": 0.0251, | |
| "step": 36500 | |
| }, | |
| { | |
| "epoch": 1.7944695038242793, | |
| "grad_norm": 0.3065630793571472, | |
| "learning_rate": 9.808960121533312e-05, | |
| "loss": 0.0284, | |
| "step": 36600 | |
| }, | |
| { | |
| "epoch": 1.7993724259658757, | |
| "grad_norm": 0.20038124918937683, | |
| "learning_rate": 9.806610276386992e-05, | |
| "loss": 0.028, | |
| "step": 36700 | |
| }, | |
| { | |
| "epoch": 1.804275348107472, | |
| "grad_norm": 0.25850433111190796, | |
| "learning_rate": 9.804246352509233e-05, | |
| "loss": 0.0274, | |
| "step": 36800 | |
| }, | |
| { | |
| "epoch": 1.8091782702490684, | |
| "grad_norm": 0.21738532185554504, | |
| "learning_rate": 9.801868356824051e-05, | |
| "loss": 0.0309, | |
| "step": 36900 | |
| }, | |
| { | |
| "epoch": 1.8140811923906648, | |
| "grad_norm": 0.19677628576755524, | |
| "learning_rate": 9.799476296296682e-05, | |
| "loss": 0.0273, | |
| "step": 37000 | |
| }, | |
| { | |
| "epoch": 1.8189841145322614, | |
| "grad_norm": 0.5938333868980408, | |
| "learning_rate": 9.797070177933554e-05, | |
| "loss": 0.025, | |
| "step": 37100 | |
| }, | |
| { | |
| "epoch": 1.8238870366738578, | |
| "grad_norm": 0.38806554675102234, | |
| "learning_rate": 9.794650008782273e-05, | |
| "loss": 0.0281, | |
| "step": 37200 | |
| }, | |
| { | |
| "epoch": 1.8287899588154541, | |
| "grad_norm": 0.3352005183696747, | |
| "learning_rate": 9.792215795931601e-05, | |
| "loss": 0.0265, | |
| "step": 37300 | |
| }, | |
| { | |
| "epoch": 1.8336928809570505, | |
| "grad_norm": 0.281361848115921, | |
| "learning_rate": 9.789767546511433e-05, | |
| "loss": 0.0301, | |
| "step": 37400 | |
| }, | |
| { | |
| "epoch": 1.838595803098647, | |
| "grad_norm": 0.265520304441452, | |
| "learning_rate": 9.787305267692778e-05, | |
| "loss": 0.0257, | |
| "step": 37500 | |
| }, | |
| { | |
| "epoch": 1.8434987252402433, | |
| "grad_norm": 0.1471584439277649, | |
| "learning_rate": 9.784828966687736e-05, | |
| "loss": 0.0245, | |
| "step": 37600 | |
| }, | |
| { | |
| "epoch": 1.8484016473818397, | |
| "grad_norm": 0.12242023646831512, | |
| "learning_rate": 9.782338650749483e-05, | |
| "loss": 0.0254, | |
| "step": 37700 | |
| }, | |
| { | |
| "epoch": 1.853304569523436, | |
| "grad_norm": 0.22480016946792603, | |
| "learning_rate": 9.779834327172241e-05, | |
| "loss": 0.0254, | |
| "step": 37800 | |
| }, | |
| { | |
| "epoch": 1.8582074916650324, | |
| "grad_norm": 0.4627992808818817, | |
| "learning_rate": 9.777316003291262e-05, | |
| "loss": 0.0251, | |
| "step": 37900 | |
| }, | |
| { | |
| "epoch": 1.8631104138066288, | |
| "grad_norm": 0.3330112397670746, | |
| "learning_rate": 9.774783686482807e-05, | |
| "loss": 0.025, | |
| "step": 38000 | |
| }, | |
| { | |
| "epoch": 1.8680133359482252, | |
| "grad_norm": 0.46279048919677734, | |
| "learning_rate": 9.772237384164121e-05, | |
| "loss": 0.0283, | |
| "step": 38100 | |
| }, | |
| { | |
| "epoch": 1.8729162580898215, | |
| "grad_norm": 0.39556947350502014, | |
| "learning_rate": 9.769677103793411e-05, | |
| "loss": 0.0245, | |
| "step": 38200 | |
| }, | |
| { | |
| "epoch": 1.877819180231418, | |
| "grad_norm": 0.27834552526474, | |
| "learning_rate": 9.767102852869832e-05, | |
| "loss": 0.0271, | |
| "step": 38300 | |
| }, | |
| { | |
| "epoch": 1.8827221023730143, | |
| "grad_norm": 0.16157115995883942, | |
| "learning_rate": 9.764514638933455e-05, | |
| "loss": 0.0292, | |
| "step": 38400 | |
| }, | |
| { | |
| "epoch": 1.8876250245146107, | |
| "grad_norm": 0.1439429223537445, | |
| "learning_rate": 9.761912469565251e-05, | |
| "loss": 0.0282, | |
| "step": 38500 | |
| }, | |
| { | |
| "epoch": 1.892527946656207, | |
| "grad_norm": 0.3553975224494934, | |
| "learning_rate": 9.759296352387063e-05, | |
| "loss": 0.0265, | |
| "step": 38600 | |
| }, | |
| { | |
| "epoch": 1.8974308687978034, | |
| "grad_norm": 0.37897151708602905, | |
| "learning_rate": 9.756666295061593e-05, | |
| "loss": 0.0272, | |
| "step": 38700 | |
| }, | |
| { | |
| "epoch": 1.9023337909393998, | |
| "grad_norm": 0.2812596261501312, | |
| "learning_rate": 9.754022305292372e-05, | |
| "loss": 0.0289, | |
| "step": 38800 | |
| }, | |
| { | |
| "epoch": 1.9072367130809962, | |
| "grad_norm": 0.39995622634887695, | |
| "learning_rate": 9.751364390823741e-05, | |
| "loss": 0.025, | |
| "step": 38900 | |
| }, | |
| { | |
| "epoch": 1.9121396352225926, | |
| "grad_norm": 0.4930385649204254, | |
| "learning_rate": 9.748692559440822e-05, | |
| "loss": 0.0287, | |
| "step": 39000 | |
| }, | |
| { | |
| "epoch": 1.917042557364189, | |
| "grad_norm": 0.2677227854728699, | |
| "learning_rate": 9.746006818969504e-05, | |
| "loss": 0.025, | |
| "step": 39100 | |
| }, | |
| { | |
| "epoch": 1.9219454795057853, | |
| "grad_norm": 0.23274803161621094, | |
| "learning_rate": 9.743307177276419e-05, | |
| "loss": 0.0277, | |
| "step": 39200 | |
| }, | |
| { | |
| "epoch": 1.9268484016473817, | |
| "grad_norm": 0.6990579962730408, | |
| "learning_rate": 9.74059364226891e-05, | |
| "loss": 0.0259, | |
| "step": 39300 | |
| }, | |
| { | |
| "epoch": 1.931751323788978, | |
| "grad_norm": 0.19638150930404663, | |
| "learning_rate": 9.737866221895018e-05, | |
| "loss": 0.0249, | |
| "step": 39400 | |
| }, | |
| { | |
| "epoch": 1.9366542459305747, | |
| "grad_norm": 0.22524696588516235, | |
| "learning_rate": 9.735124924143455e-05, | |
| "loss": 0.0239, | |
| "step": 39500 | |
| }, | |
| { | |
| "epoch": 1.941557168072171, | |
| "grad_norm": 0.2928408682346344, | |
| "learning_rate": 9.732369757043576e-05, | |
| "loss": 0.0275, | |
| "step": 39600 | |
| }, | |
| { | |
| "epoch": 1.9464600902137674, | |
| "grad_norm": 0.26414749026298523, | |
| "learning_rate": 9.729600728665365e-05, | |
| "loss": 0.0264, | |
| "step": 39700 | |
| }, | |
| { | |
| "epoch": 1.9513630123553638, | |
| "grad_norm": 0.36181434988975525, | |
| "learning_rate": 9.726817847119403e-05, | |
| "loss": 0.0284, | |
| "step": 39800 | |
| }, | |
| { | |
| "epoch": 1.9562659344969602, | |
| "grad_norm": 0.18077948689460754, | |
| "learning_rate": 9.72402112055685e-05, | |
| "loss": 0.028, | |
| "step": 39900 | |
| }, | |
| { | |
| "epoch": 1.9611688566385566, | |
| "grad_norm": 0.20345261693000793, | |
| "learning_rate": 9.721210557169416e-05, | |
| "loss": 0.0256, | |
| "step": 40000 | |
| }, | |
| { | |
| "epoch": 1.966071778780153, | |
| "grad_norm": 0.1142786517739296, | |
| "learning_rate": 9.71838616518934e-05, | |
| "loss": 0.0258, | |
| "step": 40100 | |
| }, | |
| { | |
| "epoch": 1.9709747009217493, | |
| "grad_norm": 0.18993543088436127, | |
| "learning_rate": 9.715547952889366e-05, | |
| "loss": 0.0276, | |
| "step": 40200 | |
| }, | |
| { | |
| "epoch": 1.975877623063346, | |
| "grad_norm": 0.21492281556129456, | |
| "learning_rate": 9.712695928582718e-05, | |
| "loss": 0.026, | |
| "step": 40300 | |
| }, | |
| { | |
| "epoch": 1.9807805452049423, | |
| "grad_norm": 0.3433995246887207, | |
| "learning_rate": 9.709830100623078e-05, | |
| "loss": 0.0261, | |
| "step": 40400 | |
| }, | |
| { | |
| "epoch": 1.9856834673465387, | |
| "grad_norm": 0.2835942506790161, | |
| "learning_rate": 9.706950477404555e-05, | |
| "loss": 0.0278, | |
| "step": 40500 | |
| }, | |
| { | |
| "epoch": 1.990586389488135, | |
| "grad_norm": 0.1534029096364975, | |
| "learning_rate": 9.70405706736167e-05, | |
| "loss": 0.0299, | |
| "step": 40600 | |
| }, | |
| { | |
| "epoch": 1.9954893116297314, | |
| "grad_norm": 0.24719446897506714, | |
| "learning_rate": 9.70114987896932e-05, | |
| "loss": 0.0272, | |
| "step": 40700 | |
| }, | |
| { | |
| "epoch": 2.0, | |
| "eval_loss": 0.029068127274513245, | |
| "eval_runtime": 233.9259, | |
| "eval_samples_per_second": 21.374, | |
| "eval_steps_per_second": 21.374, | |
| "step": 40792 | |
| }, | |
| { | |
| "epoch": 2.000392233771328, | |
| "grad_norm": 0.14692184329032898, | |
| "learning_rate": 9.698228920742765e-05, | |
| "loss": 0.0282, | |
| "step": 40800 | |
| }, | |
| { | |
| "epoch": 2.005295155912924, | |
| "grad_norm": 0.14744846522808075, | |
| "learning_rate": 9.695294201237593e-05, | |
| "loss": 0.0088, | |
| "step": 40900 | |
| }, | |
| { | |
| "epoch": 2.0101980780545206, | |
| "grad_norm": 0.03153422847390175, | |
| "learning_rate": 9.692345729049703e-05, | |
| "loss": 0.0099, | |
| "step": 41000 | |
| }, | |
| { | |
| "epoch": 2.015101000196117, | |
| "grad_norm": 0.17189723253250122, | |
| "learning_rate": 9.689383512815273e-05, | |
| "loss": 0.0102, | |
| "step": 41100 | |
| }, | |
| { | |
| "epoch": 2.0200039223377133, | |
| "grad_norm": 0.056081388145685196, | |
| "learning_rate": 9.686407561210741e-05, | |
| "loss": 0.01, | |
| "step": 41200 | |
| }, | |
| { | |
| "epoch": 2.0249068444793097, | |
| "grad_norm": 0.18959540128707886, | |
| "learning_rate": 9.683417882952773e-05, | |
| "loss": 0.0118, | |
| "step": 41300 | |
| }, | |
| { | |
| "epoch": 2.029809766620906, | |
| "grad_norm": 0.1800815314054489, | |
| "learning_rate": 9.68041448679824e-05, | |
| "loss": 0.0108, | |
| "step": 41400 | |
| }, | |
| { | |
| "epoch": 2.0347126887625024, | |
| "grad_norm": 0.024594638496637344, | |
| "learning_rate": 9.6773973815442e-05, | |
| "loss": 0.0089, | |
| "step": 41500 | |
| }, | |
| { | |
| "epoch": 2.039615610904099, | |
| "grad_norm": 0.12218187749385834, | |
| "learning_rate": 9.674366576027859e-05, | |
| "loss": 0.0101, | |
| "step": 41600 | |
| }, | |
| { | |
| "epoch": 2.044518533045695, | |
| "grad_norm": 0.009762736968696117, | |
| "learning_rate": 9.671322079126553e-05, | |
| "loss": 0.0087, | |
| "step": 41700 | |
| }, | |
| { | |
| "epoch": 2.0494214551872916, | |
| "grad_norm": 0.23141342401504517, | |
| "learning_rate": 9.668263899757722e-05, | |
| "loss": 0.0103, | |
| "step": 41800 | |
| }, | |
| { | |
| "epoch": 2.054324377328888, | |
| "grad_norm": 0.22465147078037262, | |
| "learning_rate": 9.665192046878879e-05, | |
| "loss": 0.011, | |
| "step": 41900 | |
| }, | |
| { | |
| "epoch": 2.0592272994704843, | |
| "grad_norm": 0.2304316610097885, | |
| "learning_rate": 9.662106529487593e-05, | |
| "loss": 0.01, | |
| "step": 42000 | |
| }, | |
| { | |
| "epoch": 2.0641302216120807, | |
| "grad_norm": 0.253606915473938, | |
| "learning_rate": 9.659007356621453e-05, | |
| "loss": 0.01, | |
| "step": 42100 | |
| }, | |
| { | |
| "epoch": 2.069033143753677, | |
| "grad_norm": 0.09148263186216354, | |
| "learning_rate": 9.655894537358045e-05, | |
| "loss": 0.0104, | |
| "step": 42200 | |
| }, | |
| { | |
| "epoch": 2.0739360658952735, | |
| "grad_norm": 0.20658960938453674, | |
| "learning_rate": 9.652768080814927e-05, | |
| "loss": 0.0103, | |
| "step": 42300 | |
| }, | |
| { | |
| "epoch": 2.07883898803687, | |
| "grad_norm": 0.0633637011051178, | |
| "learning_rate": 9.6496279961496e-05, | |
| "loss": 0.0112, | |
| "step": 42400 | |
| }, | |
| { | |
| "epoch": 2.083741910178466, | |
| "grad_norm": 0.20948578417301178, | |
| "learning_rate": 9.646474292559483e-05, | |
| "loss": 0.0097, | |
| "step": 42500 | |
| }, | |
| { | |
| "epoch": 2.0886448323200626, | |
| "grad_norm": 0.24738560616970062, | |
| "learning_rate": 9.643306979281885e-05, | |
| "loss": 0.0093, | |
| "step": 42600 | |
| }, | |
| { | |
| "epoch": 2.093547754461659, | |
| "grad_norm": 0.11444352567195892, | |
| "learning_rate": 9.640126065593981e-05, | |
| "loss": 0.0093, | |
| "step": 42700 | |
| }, | |
| { | |
| "epoch": 2.0984506766032553, | |
| "grad_norm": 0.20413364470005035, | |
| "learning_rate": 9.636931560812777e-05, | |
| "loss": 0.0116, | |
| "step": 42800 | |
| }, | |
| { | |
| "epoch": 2.1033535987448517, | |
| "grad_norm": 0.051684655249118805, | |
| "learning_rate": 9.633723474295092e-05, | |
| "loss": 0.0082, | |
| "step": 42900 | |
| }, | |
| { | |
| "epoch": 2.1082565208864485, | |
| "grad_norm": 0.1661156564950943, | |
| "learning_rate": 9.630501815437525e-05, | |
| "loss": 0.0112, | |
| "step": 43000 | |
| }, | |
| { | |
| "epoch": 2.113159443028045, | |
| "grad_norm": 0.11993784457445145, | |
| "learning_rate": 9.627266593676427e-05, | |
| "loss": 0.0089, | |
| "step": 43100 | |
| }, | |
| { | |
| "epoch": 2.1180623651696413, | |
| "grad_norm": 0.21582569181919098, | |
| "learning_rate": 9.624017818487879e-05, | |
| "loss": 0.0103, | |
| "step": 43200 | |
| }, | |
| { | |
| "epoch": 2.1229652873112377, | |
| "grad_norm": 0.3271758556365967, | |
| "learning_rate": 9.620755499387658e-05, | |
| "loss": 0.0104, | |
| "step": 43300 | |
| }, | |
| { | |
| "epoch": 2.127868209452834, | |
| "grad_norm": 0.08778655529022217, | |
| "learning_rate": 9.61747964593121e-05, | |
| "loss": 0.012, | |
| "step": 43400 | |
| }, | |
| { | |
| "epoch": 2.1327711315944304, | |
| "grad_norm": 0.32174956798553467, | |
| "learning_rate": 9.614190267713626e-05, | |
| "loss": 0.0108, | |
| "step": 43500 | |
| }, | |
| { | |
| "epoch": 2.137674053736027, | |
| "grad_norm": 0.03531063720583916, | |
| "learning_rate": 9.610887374369612e-05, | |
| "loss": 0.0113, | |
| "step": 43600 | |
| }, | |
| { | |
| "epoch": 2.142576975877623, | |
| "grad_norm": 0.31483060121536255, | |
| "learning_rate": 9.607570975573462e-05, | |
| "loss": 0.0098, | |
| "step": 43700 | |
| }, | |
| { | |
| "epoch": 2.1474798980192196, | |
| "grad_norm": 0.0525469034910202, | |
| "learning_rate": 9.604241081039021e-05, | |
| "loss": 0.0095, | |
| "step": 43800 | |
| }, | |
| { | |
| "epoch": 2.152382820160816, | |
| "grad_norm": 0.11407173424959183, | |
| "learning_rate": 9.600897700519668e-05, | |
| "loss": 0.0104, | |
| "step": 43900 | |
| }, | |
| { | |
| "epoch": 2.1572857423024123, | |
| "grad_norm": 0.14029061794281006, | |
| "learning_rate": 9.597540843808286e-05, | |
| "loss": 0.0108, | |
| "step": 44000 | |
| }, | |
| { | |
| "epoch": 2.1621886644440087, | |
| "grad_norm": 0.21073657274246216, | |
| "learning_rate": 9.594170520737227e-05, | |
| "loss": 0.0112, | |
| "step": 44100 | |
| }, | |
| { | |
| "epoch": 2.167091586585605, | |
| "grad_norm": 0.05304975062608719, | |
| "learning_rate": 9.590786741178281e-05, | |
| "loss": 0.0109, | |
| "step": 44200 | |
| }, | |
| { | |
| "epoch": 2.1719945087272015, | |
| "grad_norm": 0.04336349293589592, | |
| "learning_rate": 9.587389515042664e-05, | |
| "loss": 0.0107, | |
| "step": 44300 | |
| }, | |
| { | |
| "epoch": 2.176897430868798, | |
| "grad_norm": 0.1172260120511055, | |
| "learning_rate": 9.58397885228097e-05, | |
| "loss": 0.0111, | |
| "step": 44400 | |
| }, | |
| { | |
| "epoch": 2.181800353010394, | |
| "grad_norm": 0.18478664755821228, | |
| "learning_rate": 9.580554762883148e-05, | |
| "loss": 0.0091, | |
| "step": 44500 | |
| }, | |
| { | |
| "epoch": 2.1867032751519906, | |
| "grad_norm": 0.09167050570249557, | |
| "learning_rate": 9.57711725687848e-05, | |
| "loss": 0.0114, | |
| "step": 44600 | |
| }, | |
| { | |
| "epoch": 2.191606197293587, | |
| "grad_norm": 0.01435750164091587, | |
| "learning_rate": 9.573666344335541e-05, | |
| "loss": 0.0101, | |
| "step": 44700 | |
| }, | |
| { | |
| "epoch": 2.1965091194351833, | |
| "grad_norm": 0.01749005913734436, | |
| "learning_rate": 9.570202035362177e-05, | |
| "loss": 0.0107, | |
| "step": 44800 | |
| }, | |
| { | |
| "epoch": 2.2014120415767797, | |
| "grad_norm": 0.27203837037086487, | |
| "learning_rate": 9.56672434010547e-05, | |
| "loss": 0.0118, | |
| "step": 44900 | |
| }, | |
| { | |
| "epoch": 2.206314963718376, | |
| "grad_norm": 0.29238438606262207, | |
| "learning_rate": 9.563233268751712e-05, | |
| "loss": 0.0103, | |
| "step": 45000 | |
| }, | |
| { | |
| "epoch": 2.2112178858599725, | |
| "grad_norm": 0.026590676978230476, | |
| "learning_rate": 9.559728831526378e-05, | |
| "loss": 0.0128, | |
| "step": 45100 | |
| }, | |
| { | |
| "epoch": 2.216120808001569, | |
| "grad_norm": 0.2632576525211334, | |
| "learning_rate": 9.556211038694084e-05, | |
| "loss": 0.0103, | |
| "step": 45200 | |
| }, | |
| { | |
| "epoch": 2.2210237301431652, | |
| "grad_norm": 0.096591517329216, | |
| "learning_rate": 9.552679900558574e-05, | |
| "loss": 0.0082, | |
| "step": 45300 | |
| }, | |
| { | |
| "epoch": 2.2259266522847616, | |
| "grad_norm": 0.10889133810997009, | |
| "learning_rate": 9.549135427462672e-05, | |
| "loss": 0.0105, | |
| "step": 45400 | |
| }, | |
| { | |
| "epoch": 2.230829574426358, | |
| "grad_norm": 0.44789671897888184, | |
| "learning_rate": 9.545577629788269e-05, | |
| "loss": 0.0114, | |
| "step": 45500 | |
| }, | |
| { | |
| "epoch": 2.2357324965679544, | |
| "grad_norm": 0.46069878339767456, | |
| "learning_rate": 9.542006517956277e-05, | |
| "loss": 0.0119, | |
| "step": 45600 | |
| }, | |
| { | |
| "epoch": 2.2406354187095507, | |
| "grad_norm": 0.18736761808395386, | |
| "learning_rate": 9.53842210242661e-05, | |
| "loss": 0.0106, | |
| "step": 45700 | |
| }, | |
| { | |
| "epoch": 2.245538340851147, | |
| "grad_norm": 0.1695462465286255, | |
| "learning_rate": 9.534824393698149e-05, | |
| "loss": 0.0124, | |
| "step": 45800 | |
| }, | |
| { | |
| "epoch": 2.2504412629927435, | |
| "grad_norm": 0.22079171240329742, | |
| "learning_rate": 9.531213402308709e-05, | |
| "loss": 0.01, | |
| "step": 45900 | |
| }, | |
| { | |
| "epoch": 2.25534418513434, | |
| "grad_norm": 0.18455089628696442, | |
| "learning_rate": 9.527589138835011e-05, | |
| "loss": 0.0118, | |
| "step": 46000 | |
| }, | |
| { | |
| "epoch": 2.2602471072759363, | |
| "grad_norm": 0.08626440167427063, | |
| "learning_rate": 9.52395161389265e-05, | |
| "loss": 0.0098, | |
| "step": 46100 | |
| }, | |
| { | |
| "epoch": 2.2651500294175326, | |
| "grad_norm": 0.15146349370479584, | |
| "learning_rate": 9.520300838136069e-05, | |
| "loss": 0.0122, | |
| "step": 46200 | |
| }, | |
| { | |
| "epoch": 2.270052951559129, | |
| "grad_norm": 0.06669634580612183, | |
| "learning_rate": 9.516636822258514e-05, | |
| "loss": 0.011, | |
| "step": 46300 | |
| }, | |
| { | |
| "epoch": 2.2749558737007254, | |
| "grad_norm": 0.121702179312706, | |
| "learning_rate": 9.512959576992022e-05, | |
| "loss": 0.0125, | |
| "step": 46400 | |
| }, | |
| { | |
| "epoch": 2.279858795842322, | |
| "grad_norm": 0.10746506601572037, | |
| "learning_rate": 9.509269113107371e-05, | |
| "loss": 0.0111, | |
| "step": 46500 | |
| }, | |
| { | |
| "epoch": 2.2847617179839186, | |
| "grad_norm": 0.30326399207115173, | |
| "learning_rate": 9.505565441414061e-05, | |
| "loss": 0.0105, | |
| "step": 46600 | |
| }, | |
| { | |
| "epoch": 2.289664640125515, | |
| "grad_norm": 0.12522533535957336, | |
| "learning_rate": 9.50184857276028e-05, | |
| "loss": 0.013, | |
| "step": 46700 | |
| }, | |
| { | |
| "epoch": 2.2945675622671113, | |
| "grad_norm": 0.19467896223068237, | |
| "learning_rate": 9.498118518032864e-05, | |
| "loss": 0.0099, | |
| "step": 46800 | |
| }, | |
| { | |
| "epoch": 2.2994704844087077, | |
| "grad_norm": 0.22523294389247894, | |
| "learning_rate": 9.494375288157276e-05, | |
| "loss": 0.0094, | |
| "step": 46900 | |
| }, | |
| { | |
| "epoch": 2.304373406550304, | |
| "grad_norm": 0.07772246748209, | |
| "learning_rate": 9.490618894097569e-05, | |
| "loss": 0.0098, | |
| "step": 47000 | |
| }, | |
| { | |
| "epoch": 2.3092763286919005, | |
| "grad_norm": 0.40480080246925354, | |
| "learning_rate": 9.486849346856355e-05, | |
| "loss": 0.0102, | |
| "step": 47100 | |
| }, | |
| { | |
| "epoch": 2.314179250833497, | |
| "grad_norm": 0.15440988540649414, | |
| "learning_rate": 9.483066657474771e-05, | |
| "loss": 0.0113, | |
| "step": 47200 | |
| }, | |
| { | |
| "epoch": 2.3190821729750932, | |
| "grad_norm": 0.24852220714092255, | |
| "learning_rate": 9.479270837032446e-05, | |
| "loss": 0.011, | |
| "step": 47300 | |
| }, | |
| { | |
| "epoch": 2.3239850951166896, | |
| "grad_norm": 0.04837888851761818, | |
| "learning_rate": 9.475461896647475e-05, | |
| "loss": 0.0099, | |
| "step": 47400 | |
| }, | |
| { | |
| "epoch": 2.328888017258286, | |
| "grad_norm": 0.10291013866662979, | |
| "learning_rate": 9.471639847476377e-05, | |
| "loss": 0.01, | |
| "step": 47500 | |
| }, | |
| { | |
| "epoch": 2.3337909393998824, | |
| "grad_norm": 0.1631251573562622, | |
| "learning_rate": 9.46780470071407e-05, | |
| "loss": 0.0114, | |
| "step": 47600 | |
| }, | |
| { | |
| "epoch": 2.3386938615414787, | |
| "grad_norm": 0.10871821641921997, | |
| "learning_rate": 9.463956467593833e-05, | |
| "loss": 0.0109, | |
| "step": 47700 | |
| }, | |
| { | |
| "epoch": 2.343596783683075, | |
| "grad_norm": 0.2547306716442108, | |
| "learning_rate": 9.460095159387279e-05, | |
| "loss": 0.0121, | |
| "step": 47800 | |
| }, | |
| { | |
| "epoch": 2.3484997058246715, | |
| "grad_norm": 0.33700889348983765, | |
| "learning_rate": 9.456220787404315e-05, | |
| "loss": 0.0118, | |
| "step": 47900 | |
| }, | |
| { | |
| "epoch": 2.353402627966268, | |
| "grad_norm": 0.21864359080791473, | |
| "learning_rate": 9.452333362993115e-05, | |
| "loss": 0.0113, | |
| "step": 48000 | |
| }, | |
| { | |
| "epoch": 2.3583055501078642, | |
| "grad_norm": 0.30565133690834045, | |
| "learning_rate": 9.44843289754008e-05, | |
| "loss": 0.0119, | |
| "step": 48100 | |
| }, | |
| { | |
| "epoch": 2.3632084722494606, | |
| "grad_norm": 0.3426356911659241, | |
| "learning_rate": 9.444519402469813e-05, | |
| "loss": 0.013, | |
| "step": 48200 | |
| }, | |
| { | |
| "epoch": 2.368111394391057, | |
| "grad_norm": 0.260628342628479, | |
| "learning_rate": 9.44059288924508e-05, | |
| "loss": 0.0119, | |
| "step": 48300 | |
| }, | |
| { | |
| "epoch": 2.3730143165326534, | |
| "grad_norm": 0.12320883572101593, | |
| "learning_rate": 9.436653369366777e-05, | |
| "loss": 0.0117, | |
| "step": 48400 | |
| }, | |
| { | |
| "epoch": 2.3779172386742498, | |
| "grad_norm": 0.3849811851978302, | |
| "learning_rate": 9.432700854373894e-05, | |
| "loss": 0.0118, | |
| "step": 48500 | |
| }, | |
| { | |
| "epoch": 2.382820160815846, | |
| "grad_norm": 0.27962300181388855, | |
| "learning_rate": 9.42873535584349e-05, | |
| "loss": 0.0126, | |
| "step": 48600 | |
| }, | |
| { | |
| "epoch": 2.3877230829574425, | |
| "grad_norm": 0.08513534814119339, | |
| "learning_rate": 9.424756885390649e-05, | |
| "loss": 0.0116, | |
| "step": 48700 | |
| }, | |
| { | |
| "epoch": 2.392626005099039, | |
| "grad_norm": 0.0896671712398529, | |
| "learning_rate": 9.420765454668454e-05, | |
| "loss": 0.0108, | |
| "step": 48800 | |
| }, | |
| { | |
| "epoch": 2.3975289272406353, | |
| "grad_norm": 0.2822389304637909, | |
| "learning_rate": 9.416761075367944e-05, | |
| "loss": 0.0104, | |
| "step": 48900 | |
| }, | |
| { | |
| "epoch": 2.4024318493822316, | |
| "grad_norm": 0.15120819211006165, | |
| "learning_rate": 9.412743759218089e-05, | |
| "loss": 0.0112, | |
| "step": 49000 | |
| }, | |
| { | |
| "epoch": 2.407334771523828, | |
| "grad_norm": 0.09544061869382858, | |
| "learning_rate": 9.408713517985748e-05, | |
| "loss": 0.0109, | |
| "step": 49100 | |
| }, | |
| { | |
| "epoch": 2.412237693665425, | |
| "grad_norm": 0.2666862905025482, | |
| "learning_rate": 9.404670363475645e-05, | |
| "loss": 0.0105, | |
| "step": 49200 | |
| }, | |
| { | |
| "epoch": 2.417140615807021, | |
| "grad_norm": 0.2836065888404846, | |
| "learning_rate": 9.400614307530316e-05, | |
| "loss": 0.0105, | |
| "step": 49300 | |
| }, | |
| { | |
| "epoch": 2.4220435379486176, | |
| "grad_norm": 0.110966756939888, | |
| "learning_rate": 9.396545362030096e-05, | |
| "loss": 0.0119, | |
| "step": 49400 | |
| }, | |
| { | |
| "epoch": 2.426946460090214, | |
| "grad_norm": 0.09849441796541214, | |
| "learning_rate": 9.392463538893071e-05, | |
| "loss": 0.0106, | |
| "step": 49500 | |
| }, | |
| { | |
| "epoch": 2.4318493822318104, | |
| "grad_norm": 0.21830490231513977, | |
| "learning_rate": 9.388368850075043e-05, | |
| "loss": 0.0101, | |
| "step": 49600 | |
| }, | |
| { | |
| "epoch": 2.4367523043734067, | |
| "grad_norm": 0.07172509282827377, | |
| "learning_rate": 9.3842613075695e-05, | |
| "loss": 0.01, | |
| "step": 49700 | |
| }, | |
| { | |
| "epoch": 2.441655226515003, | |
| "grad_norm": 0.21159067749977112, | |
| "learning_rate": 9.380140923407583e-05, | |
| "loss": 0.0117, | |
| "step": 49800 | |
| }, | |
| { | |
| "epoch": 2.4465581486565995, | |
| "grad_norm": 0.046259116381406784, | |
| "learning_rate": 9.376007709658039e-05, | |
| "loss": 0.0113, | |
| "step": 49900 | |
| }, | |
| { | |
| "epoch": 2.451461070798196, | |
| "grad_norm": 0.2556672692298889, | |
| "learning_rate": 9.371861678427198e-05, | |
| "loss": 0.0111, | |
| "step": 50000 | |
| }, | |
| { | |
| "epoch": 2.4563639929397922, | |
| "grad_norm": 0.1805562973022461, | |
| "learning_rate": 9.367702841858935e-05, | |
| "loss": 0.0117, | |
| "step": 50100 | |
| }, | |
| { | |
| "epoch": 2.4612669150813886, | |
| "grad_norm": 0.05842866376042366, | |
| "learning_rate": 9.363531212134625e-05, | |
| "loss": 0.0109, | |
| "step": 50200 | |
| }, | |
| { | |
| "epoch": 2.466169837222985, | |
| "grad_norm": 0.20916195213794708, | |
| "learning_rate": 9.359346801473121e-05, | |
| "loss": 0.0108, | |
| "step": 50300 | |
| }, | |
| { | |
| "epoch": 2.4710727593645814, | |
| "grad_norm": 0.17823529243469238, | |
| "learning_rate": 9.35514962213071e-05, | |
| "loss": 0.0118, | |
| "step": 50400 | |
| }, | |
| { | |
| "epoch": 2.4759756815061778, | |
| "grad_norm": 0.14857697486877441, | |
| "learning_rate": 9.350939686401077e-05, | |
| "loss": 0.0126, | |
| "step": 50500 | |
| }, | |
| { | |
| "epoch": 2.480878603647774, | |
| "grad_norm": 0.13901618123054504, | |
| "learning_rate": 9.346717006615276e-05, | |
| "loss": 0.013, | |
| "step": 50600 | |
| }, | |
| { | |
| "epoch": 2.4857815257893705, | |
| "grad_norm": 0.21642690896987915, | |
| "learning_rate": 9.342481595141683e-05, | |
| "loss": 0.0094, | |
| "step": 50700 | |
| }, | |
| { | |
| "epoch": 2.490684447930967, | |
| "grad_norm": 0.0339026153087616, | |
| "learning_rate": 9.338233464385968e-05, | |
| "loss": 0.0106, | |
| "step": 50800 | |
| }, | |
| { | |
| "epoch": 2.4955873700725633, | |
| "grad_norm": 0.06651703268289566, | |
| "learning_rate": 9.333972626791057e-05, | |
| "loss": 0.0118, | |
| "step": 50900 | |
| }, | |
| { | |
| "epoch": 2.5004902922141596, | |
| "grad_norm": 0.22575177252292633, | |
| "learning_rate": 9.329699094837092e-05, | |
| "loss": 0.0097, | |
| "step": 51000 | |
| }, | |
| { | |
| "epoch": 2.505393214355756, | |
| "grad_norm": 0.16404572129249573, | |
| "learning_rate": 9.325412881041402e-05, | |
| "loss": 0.0122, | |
| "step": 51100 | |
| }, | |
| { | |
| "epoch": 2.5102961364973524, | |
| "grad_norm": 0.14393766224384308, | |
| "learning_rate": 9.321113997958458e-05, | |
| "loss": 0.013, | |
| "step": 51200 | |
| }, | |
| { | |
| "epoch": 2.5151990586389488, | |
| "grad_norm": 0.26015493273735046, | |
| "learning_rate": 9.316802458179839e-05, | |
| "loss": 0.011, | |
| "step": 51300 | |
| }, | |
| { | |
| "epoch": 2.520101980780545, | |
| "grad_norm": 0.21623893082141876, | |
| "learning_rate": 9.312478274334197e-05, | |
| "loss": 0.012, | |
| "step": 51400 | |
| }, | |
| { | |
| "epoch": 2.5250049029221415, | |
| "grad_norm": 0.035357438027858734, | |
| "learning_rate": 9.308141459087222e-05, | |
| "loss": 0.0111, | |
| "step": 51500 | |
| }, | |
| { | |
| "epoch": 2.529907825063738, | |
| "grad_norm": 0.19075345993041992, | |
| "learning_rate": 9.303792025141597e-05, | |
| "loss": 0.0119, | |
| "step": 51600 | |
| }, | |
| { | |
| "epoch": 2.5348107472053343, | |
| "grad_norm": 0.13520535826683044, | |
| "learning_rate": 9.29942998523697e-05, | |
| "loss": 0.0105, | |
| "step": 51700 | |
| }, | |
| { | |
| "epoch": 2.5397136693469307, | |
| "grad_norm": 0.07859195023775101, | |
| "learning_rate": 9.295055352149909e-05, | |
| "loss": 0.0125, | |
| "step": 51800 | |
| }, | |
| { | |
| "epoch": 2.544616591488527, | |
| "grad_norm": 0.23639413714408875, | |
| "learning_rate": 9.290668138693867e-05, | |
| "loss": 0.0111, | |
| "step": 51900 | |
| }, | |
| { | |
| "epoch": 2.5495195136301234, | |
| "grad_norm": 0.2909085750579834, | |
| "learning_rate": 9.286268357719152e-05, | |
| "loss": 0.0123, | |
| "step": 52000 | |
| }, | |
| { | |
| "epoch": 2.55442243577172, | |
| "grad_norm": 0.0753180980682373, | |
| "learning_rate": 9.281856022112874e-05, | |
| "loss": 0.0132, | |
| "step": 52100 | |
| }, | |
| { | |
| "epoch": 2.559325357913316, | |
| "grad_norm": 0.21286705136299133, | |
| "learning_rate": 9.27743114479892e-05, | |
| "loss": 0.0115, | |
| "step": 52200 | |
| }, | |
| { | |
| "epoch": 2.5642282800549125, | |
| "grad_norm": 0.1657351553440094, | |
| "learning_rate": 9.272993738737914e-05, | |
| "loss": 0.0118, | |
| "step": 52300 | |
| }, | |
| { | |
| "epoch": 2.569131202196509, | |
| "grad_norm": 0.18879827857017517, | |
| "learning_rate": 9.268543816927174e-05, | |
| "loss": 0.013, | |
| "step": 52400 | |
| }, | |
| { | |
| "epoch": 2.5740341243381053, | |
| "grad_norm": 0.1545262485742569, | |
| "learning_rate": 9.264081392400679e-05, | |
| "loss": 0.0107, | |
| "step": 52500 | |
| }, | |
| { | |
| "epoch": 2.5789370464797017, | |
| "grad_norm": 0.1411045491695404, | |
| "learning_rate": 9.259606478229028e-05, | |
| "loss": 0.0121, | |
| "step": 52600 | |
| }, | |
| { | |
| "epoch": 2.583839968621298, | |
| "grad_norm": 0.15814360976219177, | |
| "learning_rate": 9.255119087519404e-05, | |
| "loss": 0.0114, | |
| "step": 52700 | |
| }, | |
| { | |
| "epoch": 2.5887428907628944, | |
| "grad_norm": 0.10805512964725494, | |
| "learning_rate": 9.25061923341553e-05, | |
| "loss": 0.0103, | |
| "step": 52800 | |
| }, | |
| { | |
| "epoch": 2.593645812904491, | |
| "grad_norm": 0.10512258112430573, | |
| "learning_rate": 9.24610692909764e-05, | |
| "loss": 0.0106, | |
| "step": 52900 | |
| }, | |
| { | |
| "epoch": 2.5985487350460876, | |
| "grad_norm": 0.17147375643253326, | |
| "learning_rate": 9.241582187782434e-05, | |
| "loss": 0.0114, | |
| "step": 53000 | |
| }, | |
| { | |
| "epoch": 2.603451657187684, | |
| "grad_norm": 0.23781518638134003, | |
| "learning_rate": 9.237045022723036e-05, | |
| "loss": 0.0101, | |
| "step": 53100 | |
| }, | |
| { | |
| "epoch": 2.6083545793292804, | |
| "grad_norm": 0.2328936904668808, | |
| "learning_rate": 9.232495447208967e-05, | |
| "loss": 0.013, | |
| "step": 53200 | |
| }, | |
| { | |
| "epoch": 2.6132575014708768, | |
| "grad_norm": 0.39143115282058716, | |
| "learning_rate": 9.22793347456609e-05, | |
| "loss": 0.0117, | |
| "step": 53300 | |
| }, | |
| { | |
| "epoch": 2.618160423612473, | |
| "grad_norm": 0.19932937622070312, | |
| "learning_rate": 9.223359118156587e-05, | |
| "loss": 0.0125, | |
| "step": 53400 | |
| }, | |
| { | |
| "epoch": 2.6230633457540695, | |
| "grad_norm": 0.31379079818725586, | |
| "learning_rate": 9.218772391378908e-05, | |
| "loss": 0.0099, | |
| "step": 53500 | |
| }, | |
| { | |
| "epoch": 2.627966267895666, | |
| "grad_norm": 0.3086315095424652, | |
| "learning_rate": 9.214173307667739e-05, | |
| "loss": 0.0122, | |
| "step": 53600 | |
| }, | |
| { | |
| "epoch": 2.6328691900372623, | |
| "grad_norm": 0.1653462052345276, | |
| "learning_rate": 9.209561880493961e-05, | |
| "loss": 0.0157, | |
| "step": 53700 | |
| }, | |
| { | |
| "epoch": 2.6377721121788587, | |
| "grad_norm": 0.2647700309753418, | |
| "learning_rate": 9.204938123364604e-05, | |
| "loss": 0.0118, | |
| "step": 53800 | |
| }, | |
| { | |
| "epoch": 2.642675034320455, | |
| "grad_norm": 0.042278312146663666, | |
| "learning_rate": 9.200302049822815e-05, | |
| "loss": 0.0143, | |
| "step": 53900 | |
| }, | |
| { | |
| "epoch": 2.6475779564620514, | |
| "grad_norm": 0.162938192486763, | |
| "learning_rate": 9.195653673447822e-05, | |
| "loss": 0.013, | |
| "step": 54000 | |
| }, | |
| { | |
| "epoch": 2.652480878603648, | |
| "grad_norm": 0.06934090703725815, | |
| "learning_rate": 9.19099300785488e-05, | |
| "loss": 0.0109, | |
| "step": 54100 | |
| }, | |
| { | |
| "epoch": 2.657383800745244, | |
| "grad_norm": 0.32144516706466675, | |
| "learning_rate": 9.186320066695244e-05, | |
| "loss": 0.0122, | |
| "step": 54200 | |
| }, | |
| { | |
| "epoch": 2.6622867228868405, | |
| "grad_norm": 0.1582040935754776, | |
| "learning_rate": 9.181634863656125e-05, | |
| "loss": 0.0108, | |
| "step": 54300 | |
| }, | |
| { | |
| "epoch": 2.667189645028437, | |
| "grad_norm": 0.05845705047249794, | |
| "learning_rate": 9.176937412460647e-05, | |
| "loss": 0.0109, | |
| "step": 54400 | |
| }, | |
| { | |
| "epoch": 2.6720925671700333, | |
| "grad_norm": 0.32942378520965576, | |
| "learning_rate": 9.172227726867811e-05, | |
| "loss": 0.0131, | |
| "step": 54500 | |
| }, | |
| { | |
| "epoch": 2.6769954893116297, | |
| "grad_norm": 0.1683381050825119, | |
| "learning_rate": 9.167505820672453e-05, | |
| "loss": 0.0114, | |
| "step": 54600 | |
| }, | |
| { | |
| "epoch": 2.681898411453226, | |
| "grad_norm": 0.061455611139535904, | |
| "learning_rate": 9.162771707705204e-05, | |
| "loss": 0.0114, | |
| "step": 54700 | |
| }, | |
| { | |
| "epoch": 2.6868013335948224, | |
| "grad_norm": 0.25896042585372925, | |
| "learning_rate": 9.158025401832447e-05, | |
| "loss": 0.0117, | |
| "step": 54800 | |
| }, | |
| { | |
| "epoch": 2.691704255736419, | |
| "grad_norm": 0.18199358880519867, | |
| "learning_rate": 9.15326691695628e-05, | |
| "loss": 0.0118, | |
| "step": 54900 | |
| }, | |
| { | |
| "epoch": 2.696607177878015, | |
| "grad_norm": 0.44476819038391113, | |
| "learning_rate": 9.148496267014472e-05, | |
| "loss": 0.0106, | |
| "step": 55000 | |
| }, | |
| { | |
| "epoch": 2.7015101000196116, | |
| "grad_norm": 0.16368268430233002, | |
| "learning_rate": 9.143713465980428e-05, | |
| "loss": 0.0112, | |
| "step": 55100 | |
| }, | |
| { | |
| "epoch": 2.706413022161208, | |
| "grad_norm": 0.23616598546504974, | |
| "learning_rate": 9.13891852786314e-05, | |
| "loss": 0.0126, | |
| "step": 55200 | |
| }, | |
| { | |
| "epoch": 2.7113159443028043, | |
| "grad_norm": 0.2906864285469055, | |
| "learning_rate": 9.134111466707151e-05, | |
| "loss": 0.0127, | |
| "step": 55300 | |
| }, | |
| { | |
| "epoch": 2.716218866444401, | |
| "grad_norm": 0.14189335703849792, | |
| "learning_rate": 9.129292296592514e-05, | |
| "loss": 0.0111, | |
| "step": 55400 | |
| }, | |
| { | |
| "epoch": 2.7211217885859975, | |
| "grad_norm": 0.08627419173717499, | |
| "learning_rate": 9.124461031634748e-05, | |
| "loss": 0.0108, | |
| "step": 55500 | |
| }, | |
| { | |
| "epoch": 2.726024710727594, | |
| "grad_norm": 0.13704799115657806, | |
| "learning_rate": 9.1196176859848e-05, | |
| "loss": 0.0129, | |
| "step": 55600 | |
| }, | |
| { | |
| "epoch": 2.7309276328691903, | |
| "grad_norm": 0.05801895260810852, | |
| "learning_rate": 9.114762273828998e-05, | |
| "loss": 0.011, | |
| "step": 55700 | |
| }, | |
| { | |
| "epoch": 2.7358305550107866, | |
| "grad_norm": 0.31558868288993835, | |
| "learning_rate": 9.10989480938902e-05, | |
| "loss": 0.0125, | |
| "step": 55800 | |
| }, | |
| { | |
| "epoch": 2.740733477152383, | |
| "grad_norm": 0.06790540367364883, | |
| "learning_rate": 9.105015306921836e-05, | |
| "loss": 0.0111, | |
| "step": 55900 | |
| }, | |
| { | |
| "epoch": 2.7456363992939794, | |
| "grad_norm": 0.3000267744064331, | |
| "learning_rate": 9.100123780719685e-05, | |
| "loss": 0.0121, | |
| "step": 56000 | |
| }, | |
| { | |
| "epoch": 2.750539321435576, | |
| "grad_norm": 0.22068212926387787, | |
| "learning_rate": 9.095220245110019e-05, | |
| "loss": 0.0106, | |
| "step": 56100 | |
| }, | |
| { | |
| "epoch": 2.755442243577172, | |
| "grad_norm": 0.18099111318588257, | |
| "learning_rate": 9.090304714455469e-05, | |
| "loss": 0.0122, | |
| "step": 56200 | |
| }, | |
| { | |
| "epoch": 2.7603451657187685, | |
| "grad_norm": 0.2899315059185028, | |
| "learning_rate": 9.085377203153795e-05, | |
| "loss": 0.0124, | |
| "step": 56300 | |
| }, | |
| { | |
| "epoch": 2.765248087860365, | |
| "grad_norm": 0.3247213363647461, | |
| "learning_rate": 9.080437725637854e-05, | |
| "loss": 0.0122, | |
| "step": 56400 | |
| }, | |
| { | |
| "epoch": 2.7701510100019613, | |
| "grad_norm": 0.07788939774036407, | |
| "learning_rate": 9.07548629637555e-05, | |
| "loss": 0.0106, | |
| "step": 56500 | |
| }, | |
| { | |
| "epoch": 2.7750539321435577, | |
| "grad_norm": 0.13057078421115875, | |
| "learning_rate": 9.070522929869796e-05, | |
| "loss": 0.0111, | |
| "step": 56600 | |
| }, | |
| { | |
| "epoch": 2.779956854285154, | |
| "grad_norm": 0.0548156201839447, | |
| "learning_rate": 9.065547640658462e-05, | |
| "loss": 0.0118, | |
| "step": 56700 | |
| }, | |
| { | |
| "epoch": 2.7848597764267504, | |
| "grad_norm": 0.24302823841571808, | |
| "learning_rate": 9.060560443314353e-05, | |
| "loss": 0.0116, | |
| "step": 56800 | |
| }, | |
| { | |
| "epoch": 2.789762698568347, | |
| "grad_norm": 0.27404043078422546, | |
| "learning_rate": 9.055561352445143e-05, | |
| "loss": 0.013, | |
| "step": 56900 | |
| }, | |
| { | |
| "epoch": 2.794665620709943, | |
| "grad_norm": 0.5266895890235901, | |
| "learning_rate": 9.050550382693346e-05, | |
| "loss": 0.01, | |
| "step": 57000 | |
| }, | |
| { | |
| "epoch": 2.7995685428515396, | |
| "grad_norm": 0.1247442364692688, | |
| "learning_rate": 9.04552754873627e-05, | |
| "loss": 0.011, | |
| "step": 57100 | |
| }, | |
| { | |
| "epoch": 2.804471464993136, | |
| "grad_norm": 0.24913443624973297, | |
| "learning_rate": 9.040492865285973e-05, | |
| "loss": 0.0119, | |
| "step": 57200 | |
| }, | |
| { | |
| "epoch": 2.8093743871347323, | |
| "grad_norm": 0.11120960861444473, | |
| "learning_rate": 9.035446347089221e-05, | |
| "loss": 0.012, | |
| "step": 57300 | |
| }, | |
| { | |
| "epoch": 2.8142773092763287, | |
| "grad_norm": 0.2342406064271927, | |
| "learning_rate": 9.030388008927444e-05, | |
| "loss": 0.0128, | |
| "step": 57400 | |
| }, | |
| { | |
| "epoch": 2.819180231417925, | |
| "grad_norm": 0.19678667187690735, | |
| "learning_rate": 9.025317865616693e-05, | |
| "loss": 0.0142, | |
| "step": 57500 | |
| }, | |
| { | |
| "epoch": 2.8240831535595214, | |
| "grad_norm": 0.342568576335907, | |
| "learning_rate": 9.020235932007597e-05, | |
| "loss": 0.0104, | |
| "step": 57600 | |
| }, | |
| { | |
| "epoch": 2.828986075701118, | |
| "grad_norm": 0.27162832021713257, | |
| "learning_rate": 9.015142222985322e-05, | |
| "loss": 0.0105, | |
| "step": 57700 | |
| }, | |
| { | |
| "epoch": 2.833888997842714, | |
| "grad_norm": 0.3186455965042114, | |
| "learning_rate": 9.010036753469518e-05, | |
| "loss": 0.0121, | |
| "step": 57800 | |
| }, | |
| { | |
| "epoch": 2.8387919199843106, | |
| "grad_norm": 0.3207502067089081, | |
| "learning_rate": 9.004919538414287e-05, | |
| "loss": 0.0124, | |
| "step": 57900 | |
| }, | |
| { | |
| "epoch": 2.843694842125907, | |
| "grad_norm": 0.42707347869873047, | |
| "learning_rate": 8.999790592808133e-05, | |
| "loss": 0.0126, | |
| "step": 58000 | |
| }, | |
| { | |
| "epoch": 2.8485977642675033, | |
| "grad_norm": 0.24902117252349854, | |
| "learning_rate": 8.99464993167392e-05, | |
| "loss": 0.0113, | |
| "step": 58100 | |
| }, | |
| { | |
| "epoch": 2.8535006864090997, | |
| "grad_norm": 0.16693954169750214, | |
| "learning_rate": 8.989497570068824e-05, | |
| "loss": 0.0126, | |
| "step": 58200 | |
| }, | |
| { | |
| "epoch": 2.858403608550696, | |
| "grad_norm": 0.41618144512176514, | |
| "learning_rate": 8.984333523084298e-05, | |
| "loss": 0.0132, | |
| "step": 58300 | |
| }, | |
| { | |
| "epoch": 2.8633065306922925, | |
| "grad_norm": 0.36843499541282654, | |
| "learning_rate": 8.979157805846013e-05, | |
| "loss": 0.0119, | |
| "step": 58400 | |
| }, | |
| { | |
| "epoch": 2.868209452833889, | |
| "grad_norm": 0.3611968755722046, | |
| "learning_rate": 8.973970433513832e-05, | |
| "loss": 0.0104, | |
| "step": 58500 | |
| }, | |
| { | |
| "epoch": 2.873112374975485, | |
| "grad_norm": 0.18182240426540375, | |
| "learning_rate": 8.968771421281749e-05, | |
| "loss": 0.0106, | |
| "step": 58600 | |
| }, | |
| { | |
| "epoch": 2.8780152971170816, | |
| "grad_norm": 0.21805822849273682, | |
| "learning_rate": 8.963560784377858e-05, | |
| "loss": 0.0117, | |
| "step": 58700 | |
| }, | |
| { | |
| "epoch": 2.882918219258678, | |
| "grad_norm": 0.1938275694847107, | |
| "learning_rate": 8.958338538064297e-05, | |
| "loss": 0.0122, | |
| "step": 58800 | |
| }, | |
| { | |
| "epoch": 2.8878211414002743, | |
| "grad_norm": 0.27438417077064514, | |
| "learning_rate": 8.953104697637209e-05, | |
| "loss": 0.0131, | |
| "step": 58900 | |
| }, | |
| { | |
| "epoch": 2.8927240635418707, | |
| "grad_norm": 0.21847118437290192, | |
| "learning_rate": 8.9478592784267e-05, | |
| "loss": 0.011, | |
| "step": 59000 | |
| }, | |
| { | |
| "epoch": 2.897626985683467, | |
| "grad_norm": 0.3260138928890228, | |
| "learning_rate": 8.942602295796785e-05, | |
| "loss": 0.0135, | |
| "step": 59100 | |
| }, | |
| { | |
| "epoch": 2.9025299078250635, | |
| "grad_norm": 0.2678786814212799, | |
| "learning_rate": 8.937333765145355e-05, | |
| "loss": 0.013, | |
| "step": 59200 | |
| }, | |
| { | |
| "epoch": 2.90743282996666, | |
| "grad_norm": 0.352421373128891, | |
| "learning_rate": 8.932053701904122e-05, | |
| "loss": 0.0131, | |
| "step": 59300 | |
| }, | |
| { | |
| "epoch": 2.9123357521082567, | |
| "grad_norm": 0.25367945432662964, | |
| "learning_rate": 8.926762121538577e-05, | |
| "loss": 0.0125, | |
| "step": 59400 | |
| }, | |
| { | |
| "epoch": 2.917238674249853, | |
| "grad_norm": 0.07467200607061386, | |
| "learning_rate": 8.921459039547946e-05, | |
| "loss": 0.0105, | |
| "step": 59500 | |
| }, | |
| { | |
| "epoch": 2.9221415963914494, | |
| "grad_norm": 0.42252108454704285, | |
| "learning_rate": 8.916144471465144e-05, | |
| "loss": 0.012, | |
| "step": 59600 | |
| }, | |
| { | |
| "epoch": 2.927044518533046, | |
| "grad_norm": 0.14323803782463074, | |
| "learning_rate": 8.910818432856728e-05, | |
| "loss": 0.0109, | |
| "step": 59700 | |
| }, | |
| { | |
| "epoch": 2.931947440674642, | |
| "grad_norm": 0.18532708287239075, | |
| "learning_rate": 8.905480939322855e-05, | |
| "loss": 0.0128, | |
| "step": 59800 | |
| }, | |
| { | |
| "epoch": 2.9368503628162386, | |
| "grad_norm": 0.30199986696243286, | |
| "learning_rate": 8.900132006497231e-05, | |
| "loss": 0.0123, | |
| "step": 59900 | |
| }, | |
| { | |
| "epoch": 2.941753284957835, | |
| "grad_norm": 0.0945155918598175, | |
| "learning_rate": 8.894771650047072e-05, | |
| "loss": 0.0135, | |
| "step": 60000 | |
| }, | |
| { | |
| "epoch": 2.9466562070994313, | |
| "grad_norm": 0.1907569169998169, | |
| "learning_rate": 8.889399885673046e-05, | |
| "loss": 0.0133, | |
| "step": 60100 | |
| }, | |
| { | |
| "epoch": 2.9515591292410277, | |
| "grad_norm": 0.048369161784648895, | |
| "learning_rate": 8.884016729109246e-05, | |
| "loss": 0.0121, | |
| "step": 60200 | |
| }, | |
| { | |
| "epoch": 2.956462051382624, | |
| "grad_norm": 0.28431326150894165, | |
| "learning_rate": 8.878622196123127e-05, | |
| "loss": 0.0141, | |
| "step": 60300 | |
| }, | |
| { | |
| "epoch": 2.9613649735242205, | |
| "grad_norm": 0.3524870276451111, | |
| "learning_rate": 8.873216302515463e-05, | |
| "loss": 0.0121, | |
| "step": 60400 | |
| }, | |
| { | |
| "epoch": 2.966267895665817, | |
| "grad_norm": 0.10910056531429291, | |
| "learning_rate": 8.867799064120313e-05, | |
| "loss": 0.011, | |
| "step": 60500 | |
| }, | |
| { | |
| "epoch": 2.971170817807413, | |
| "grad_norm": 0.5855691432952881, | |
| "learning_rate": 8.862370496804958e-05, | |
| "loss": 0.0112, | |
| "step": 60600 | |
| }, | |
| { | |
| "epoch": 2.9760737399490096, | |
| "grad_norm": 0.2968946099281311, | |
| "learning_rate": 8.856930616469861e-05, | |
| "loss": 0.0095, | |
| "step": 60700 | |
| }, | |
| { | |
| "epoch": 2.980976662090606, | |
| "grad_norm": 0.07704410701990128, | |
| "learning_rate": 8.851479439048624e-05, | |
| "loss": 0.0109, | |
| "step": 60800 | |
| }, | |
| { | |
| "epoch": 2.9858795842322023, | |
| "grad_norm": 0.11921142786741257, | |
| "learning_rate": 8.846016980507942e-05, | |
| "loss": 0.0131, | |
| "step": 60900 | |
| }, | |
| { | |
| "epoch": 2.9907825063737987, | |
| "grad_norm": 0.1547405868768692, | |
| "learning_rate": 8.840543256847548e-05, | |
| "loss": 0.0131, | |
| "step": 61000 | |
| }, | |
| { | |
| "epoch": 2.995685428515395, | |
| "grad_norm": 0.1697314977645874, | |
| "learning_rate": 8.83505828410017e-05, | |
| "loss": 0.0127, | |
| "step": 61100 | |
| }, | |
| { | |
| "epoch": 3.0, | |
| "eval_loss": 0.030867407098412514, | |
| "eval_runtime": 223.9786, | |
| "eval_samples_per_second": 22.324, | |
| "eval_steps_per_second": 22.324, | |
| "step": 61188 | |
| } | |
| ], | |
| "logging_steps": 100, | |
| "max_steps": 203960, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 10, | |
| "save_steps": 500, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": false | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 6.448056809124618e+18, | |
| "train_batch_size": 16, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |