| { |
| "best_global_step": null, |
| "best_metric": null, |
| "best_model_checkpoint": null, |
| "epoch": 2.253636549887318, |
| "eval_steps": 500, |
| "global_step": 11000, |
| "is_hyper_param_search": false, |
| "is_local_process_zero": true, |
| "is_world_process_zero": true, |
| "log_history": [ |
| { |
| "epoch": 0.002048760499897562, |
| "grad_norm": 7.843175411224365, |
| "learning_rate": 1.8e-07, |
| "loss": 1.1173, |
| "step": 10 |
| }, |
| { |
| "epoch": 0.004097520999795124, |
| "grad_norm": 9.780545234680176, |
| "learning_rate": 3.8e-07, |
| "loss": 1.1759, |
| "step": 20 |
| }, |
| { |
| "epoch": 0.006146281499692686, |
| "grad_norm": 10.942689895629883, |
| "learning_rate": 5.800000000000001e-07, |
| "loss": 1.1623, |
| "step": 30 |
| }, |
| { |
| "epoch": 0.008195041999590248, |
| "grad_norm": 7.134339332580566, |
| "learning_rate": 7.799999999999999e-07, |
| "loss": 1.0603, |
| "step": 40 |
| }, |
| { |
| "epoch": 0.01024380249948781, |
| "grad_norm": 8.591020584106445, |
| "learning_rate": 9.8e-07, |
| "loss": 0.9755, |
| "step": 50 |
| }, |
| { |
| "epoch": 0.012292562999385371, |
| "grad_norm": 5.172506332397461, |
| "learning_rate": 1.18e-06, |
| "loss": 1.009, |
| "step": 60 |
| }, |
| { |
| "epoch": 0.014341323499282934, |
| "grad_norm": 7.914861679077148, |
| "learning_rate": 1.38e-06, |
| "loss": 0.897, |
| "step": 70 |
| }, |
| { |
| "epoch": 0.016390083999180495, |
| "grad_norm": 5.01186466217041, |
| "learning_rate": 1.5800000000000001e-06, |
| "loss": 0.7337, |
| "step": 80 |
| }, |
| { |
| "epoch": 0.01843884449907806, |
| "grad_norm": 2.484205722808838, |
| "learning_rate": 1.7800000000000001e-06, |
| "loss": 0.4849, |
| "step": 90 |
| }, |
| { |
| "epoch": 0.02048760499897562, |
| "grad_norm": 2.200183391571045, |
| "learning_rate": 1.98e-06, |
| "loss": 0.3821, |
| "step": 100 |
| }, |
| { |
| "epoch": 0.02253636549887318, |
| "grad_norm": 1.9665900468826294, |
| "learning_rate": 2.1800000000000003e-06, |
| "loss": 0.2953, |
| "step": 110 |
| }, |
| { |
| "epoch": 0.024585125998770743, |
| "grad_norm": 1.3625904321670532, |
| "learning_rate": 2.38e-06, |
| "loss": 0.2106, |
| "step": 120 |
| }, |
| { |
| "epoch": 0.026633886498668307, |
| "grad_norm": 0.7679103016853333, |
| "learning_rate": 2.58e-06, |
| "loss": 0.1343, |
| "step": 130 |
| }, |
| { |
| "epoch": 0.028682646998565868, |
| "grad_norm": 0.5932427048683167, |
| "learning_rate": 2.78e-06, |
| "loss": 0.0913, |
| "step": 140 |
| }, |
| { |
| "epoch": 0.03073140749846343, |
| "grad_norm": 0.6499311923980713, |
| "learning_rate": 2.98e-06, |
| "loss": 0.084, |
| "step": 150 |
| }, |
| { |
| "epoch": 0.03278016799836099, |
| "grad_norm": 0.4982171058654785, |
| "learning_rate": 3.18e-06, |
| "loss": 0.0769, |
| "step": 160 |
| }, |
| { |
| "epoch": 0.034828928498258555, |
| "grad_norm": 0.41531991958618164, |
| "learning_rate": 3.38e-06, |
| "loss": 0.0623, |
| "step": 170 |
| }, |
| { |
| "epoch": 0.03687768899815612, |
| "grad_norm": 0.33730119466781616, |
| "learning_rate": 3.58e-06, |
| "loss": 0.053, |
| "step": 180 |
| }, |
| { |
| "epoch": 0.03892644949805368, |
| "grad_norm": 0.34425678849220276, |
| "learning_rate": 3.7800000000000002e-06, |
| "loss": 0.0449, |
| "step": 190 |
| }, |
| { |
| "epoch": 0.04097520999795124, |
| "grad_norm": 0.2820582389831543, |
| "learning_rate": 3.98e-06, |
| "loss": 0.0503, |
| "step": 200 |
| }, |
| { |
| "epoch": 0.0430239704978488, |
| "grad_norm": 0.47338512539863586, |
| "learning_rate": 4.18e-06, |
| "loss": 0.0445, |
| "step": 210 |
| }, |
| { |
| "epoch": 0.04507273099774636, |
| "grad_norm": 0.42368489503860474, |
| "learning_rate": 4.3799999999999996e-06, |
| "loss": 0.0458, |
| "step": 220 |
| }, |
| { |
| "epoch": 0.04712149149764393, |
| "grad_norm": 0.3917086720466614, |
| "learning_rate": 4.58e-06, |
| "loss": 0.042, |
| "step": 230 |
| }, |
| { |
| "epoch": 0.049170251997541485, |
| "grad_norm": 0.2868223190307617, |
| "learning_rate": 4.78e-06, |
| "loss": 0.0386, |
| "step": 240 |
| }, |
| { |
| "epoch": 0.05121901249743905, |
| "grad_norm": 0.31068718433380127, |
| "learning_rate": 4.980000000000001e-06, |
| "loss": 0.0444, |
| "step": 250 |
| }, |
| { |
| "epoch": 0.053267772997336614, |
| "grad_norm": 0.5668361186981201, |
| "learning_rate": 5.18e-06, |
| "loss": 0.0347, |
| "step": 260 |
| }, |
| { |
| "epoch": 0.05531653349723417, |
| "grad_norm": 0.29416772723197937, |
| "learning_rate": 5.38e-06, |
| "loss": 0.0372, |
| "step": 270 |
| }, |
| { |
| "epoch": 0.057365293997131736, |
| "grad_norm": 0.30019980669021606, |
| "learning_rate": 5.58e-06, |
| "loss": 0.032, |
| "step": 280 |
| }, |
| { |
| "epoch": 0.0594140544970293, |
| "grad_norm": 0.24995483458042145, |
| "learning_rate": 5.7800000000000006e-06, |
| "loss": 0.0332, |
| "step": 290 |
| }, |
| { |
| "epoch": 0.06146281499692686, |
| "grad_norm": 0.22572636604309082, |
| "learning_rate": 5.98e-06, |
| "loss": 0.0305, |
| "step": 300 |
| }, |
| { |
| "epoch": 0.06351157549682442, |
| "grad_norm": 0.38446372747421265, |
| "learning_rate": 6.18e-06, |
| "loss": 0.0276, |
| "step": 310 |
| }, |
| { |
| "epoch": 0.06556033599672198, |
| "grad_norm": 0.35160934925079346, |
| "learning_rate": 6.38e-06, |
| "loss": 0.0319, |
| "step": 320 |
| }, |
| { |
| "epoch": 0.06760909649661954, |
| "grad_norm": 0.3805447518825531, |
| "learning_rate": 6.58e-06, |
| "loss": 0.0313, |
| "step": 330 |
| }, |
| { |
| "epoch": 0.06965785699651711, |
| "grad_norm": 0.43175026774406433, |
| "learning_rate": 6.78e-06, |
| "loss": 0.0268, |
| "step": 340 |
| }, |
| { |
| "epoch": 0.07170661749641467, |
| "grad_norm": 0.4864197373390198, |
| "learning_rate": 6.98e-06, |
| "loss": 0.0252, |
| "step": 350 |
| }, |
| { |
| "epoch": 0.07375537799631224, |
| "grad_norm": 0.34749090671539307, |
| "learning_rate": 7.180000000000001e-06, |
| "loss": 0.0325, |
| "step": 360 |
| }, |
| { |
| "epoch": 0.07580413849620979, |
| "grad_norm": 0.3528898358345032, |
| "learning_rate": 7.3800000000000005e-06, |
| "loss": 0.029, |
| "step": 370 |
| }, |
| { |
| "epoch": 0.07785289899610735, |
| "grad_norm": 0.3244384229183197, |
| "learning_rate": 7.5799999999999994e-06, |
| "loss": 0.0317, |
| "step": 380 |
| }, |
| { |
| "epoch": 0.07990165949600492, |
| "grad_norm": 0.2572486102581024, |
| "learning_rate": 7.780000000000002e-06, |
| "loss": 0.024, |
| "step": 390 |
| }, |
| { |
| "epoch": 0.08195041999590248, |
| "grad_norm": 0.43141183257102966, |
| "learning_rate": 7.98e-06, |
| "loss": 0.0248, |
| "step": 400 |
| }, |
| { |
| "epoch": 0.08399918049580005, |
| "grad_norm": 0.46238183975219727, |
| "learning_rate": 8.18e-06, |
| "loss": 0.0219, |
| "step": 410 |
| }, |
| { |
| "epoch": 0.0860479409956976, |
| "grad_norm": 1.017019271850586, |
| "learning_rate": 8.38e-06, |
| "loss": 0.0252, |
| "step": 420 |
| }, |
| { |
| "epoch": 0.08809670149559516, |
| "grad_norm": 0.2638542950153351, |
| "learning_rate": 8.58e-06, |
| "loss": 0.0273, |
| "step": 430 |
| }, |
| { |
| "epoch": 0.09014546199549273, |
| "grad_norm": 0.3530164659023285, |
| "learning_rate": 8.78e-06, |
| "loss": 0.0229, |
| "step": 440 |
| }, |
| { |
| "epoch": 0.09219422249539029, |
| "grad_norm": 0.18311157822608948, |
| "learning_rate": 8.98e-06, |
| "loss": 0.0266, |
| "step": 450 |
| }, |
| { |
| "epoch": 0.09424298299528786, |
| "grad_norm": 0.3219856321811676, |
| "learning_rate": 9.18e-06, |
| "loss": 0.0256, |
| "step": 460 |
| }, |
| { |
| "epoch": 0.09629174349518542, |
| "grad_norm": 0.252760648727417, |
| "learning_rate": 9.38e-06, |
| "loss": 0.0223, |
| "step": 470 |
| }, |
| { |
| "epoch": 0.09834050399508297, |
| "grad_norm": 0.21674038469791412, |
| "learning_rate": 9.580000000000001e-06, |
| "loss": 0.024, |
| "step": 480 |
| }, |
| { |
| "epoch": 0.10038926449498053, |
| "grad_norm": 0.24689117074012756, |
| "learning_rate": 9.780000000000001e-06, |
| "loss": 0.0194, |
| "step": 490 |
| }, |
| { |
| "epoch": 0.1024380249948781, |
| "grad_norm": 0.2953890860080719, |
| "learning_rate": 9.980000000000001e-06, |
| "loss": 0.0192, |
| "step": 500 |
| }, |
| { |
| "epoch": 0.10448678549477566, |
| "grad_norm": 0.28237539529800415, |
| "learning_rate": 1.0179999999999999e-05, |
| "loss": 0.0214, |
| "step": 510 |
| }, |
| { |
| "epoch": 0.10653554599467323, |
| "grad_norm": 0.31394603848457336, |
| "learning_rate": 1.0379999999999999e-05, |
| "loss": 0.0182, |
| "step": 520 |
| }, |
| { |
| "epoch": 0.10858430649457078, |
| "grad_norm": 0.28832370042800903, |
| "learning_rate": 1.058e-05, |
| "loss": 0.0236, |
| "step": 530 |
| }, |
| { |
| "epoch": 0.11063306699446834, |
| "grad_norm": 0.21490629017353058, |
| "learning_rate": 1.078e-05, |
| "loss": 0.0195, |
| "step": 540 |
| }, |
| { |
| "epoch": 0.11268182749436591, |
| "grad_norm": 0.37322184443473816, |
| "learning_rate": 1.098e-05, |
| "loss": 0.0191, |
| "step": 550 |
| }, |
| { |
| "epoch": 0.11473058799426347, |
| "grad_norm": 0.33188939094543457, |
| "learning_rate": 1.118e-05, |
| "loss": 0.0188, |
| "step": 560 |
| }, |
| { |
| "epoch": 0.11677934849416104, |
| "grad_norm": 0.25061842799186707, |
| "learning_rate": 1.1380000000000001e-05, |
| "loss": 0.0254, |
| "step": 570 |
| }, |
| { |
| "epoch": 0.1188281089940586, |
| "grad_norm": 0.273547500371933, |
| "learning_rate": 1.1580000000000001e-05, |
| "loss": 0.0183, |
| "step": 580 |
| }, |
| { |
| "epoch": 0.12087686949395615, |
| "grad_norm": 0.4790613353252411, |
| "learning_rate": 1.178e-05, |
| "loss": 0.0206, |
| "step": 590 |
| }, |
| { |
| "epoch": 0.12292562999385372, |
| "grad_norm": 0.6439072489738464, |
| "learning_rate": 1.198e-05, |
| "loss": 0.0225, |
| "step": 600 |
| }, |
| { |
| "epoch": 0.12497439049375128, |
| "grad_norm": 0.23024041950702667, |
| "learning_rate": 1.2180000000000002e-05, |
| "loss": 0.0254, |
| "step": 610 |
| }, |
| { |
| "epoch": 0.12702315099364883, |
| "grad_norm": 0.29340267181396484, |
| "learning_rate": 1.238e-05, |
| "loss": 0.021, |
| "step": 620 |
| }, |
| { |
| "epoch": 0.1290719114935464, |
| "grad_norm": 0.34328433871269226, |
| "learning_rate": 1.258e-05, |
| "loss": 0.0218, |
| "step": 630 |
| }, |
| { |
| "epoch": 0.13112067199344396, |
| "grad_norm": 0.25611504912376404, |
| "learning_rate": 1.278e-05, |
| "loss": 0.0235, |
| "step": 640 |
| }, |
| { |
| "epoch": 0.13316943249334154, |
| "grad_norm": 0.4698771834373474, |
| "learning_rate": 1.298e-05, |
| "loss": 0.0205, |
| "step": 650 |
| }, |
| { |
| "epoch": 0.1352181929932391, |
| "grad_norm": 0.39816421270370483, |
| "learning_rate": 1.3180000000000001e-05, |
| "loss": 0.023, |
| "step": 660 |
| }, |
| { |
| "epoch": 0.13726695349313664, |
| "grad_norm": 0.4990415573120117, |
| "learning_rate": 1.338e-05, |
| "loss": 0.023, |
| "step": 670 |
| }, |
| { |
| "epoch": 0.13931571399303422, |
| "grad_norm": 0.25273269414901733, |
| "learning_rate": 1.358e-05, |
| "loss": 0.0199, |
| "step": 680 |
| }, |
| { |
| "epoch": 0.14136447449293177, |
| "grad_norm": 0.36715972423553467, |
| "learning_rate": 1.378e-05, |
| "loss": 0.0235, |
| "step": 690 |
| }, |
| { |
| "epoch": 0.14341323499282935, |
| "grad_norm": 0.2521001696586609, |
| "learning_rate": 1.3980000000000002e-05, |
| "loss": 0.0207, |
| "step": 700 |
| }, |
| { |
| "epoch": 0.1454619954927269, |
| "grad_norm": 0.39710763096809387, |
| "learning_rate": 1.4180000000000001e-05, |
| "loss": 0.0196, |
| "step": 710 |
| }, |
| { |
| "epoch": 0.14751075599262448, |
| "grad_norm": 0.236027792096138, |
| "learning_rate": 1.438e-05, |
| "loss": 0.0172, |
| "step": 720 |
| }, |
| { |
| "epoch": 0.14955951649252203, |
| "grad_norm": 0.22290261089801788, |
| "learning_rate": 1.458e-05, |
| "loss": 0.0155, |
| "step": 730 |
| }, |
| { |
| "epoch": 0.15160827699241958, |
| "grad_norm": 0.22860530018806458, |
| "learning_rate": 1.4779999999999999e-05, |
| "loss": 0.0175, |
| "step": 740 |
| }, |
| { |
| "epoch": 0.15365703749231716, |
| "grad_norm": 0.2443128079175949, |
| "learning_rate": 1.498e-05, |
| "loss": 0.0176, |
| "step": 750 |
| }, |
| { |
| "epoch": 0.1557057979922147, |
| "grad_norm": 0.35519939661026, |
| "learning_rate": 1.518e-05, |
| "loss": 0.0194, |
| "step": 760 |
| }, |
| { |
| "epoch": 0.15775455849211228, |
| "grad_norm": 0.3162888288497925, |
| "learning_rate": 1.5380000000000002e-05, |
| "loss": 0.0197, |
| "step": 770 |
| }, |
| { |
| "epoch": 0.15980331899200984, |
| "grad_norm": 0.5799744129180908, |
| "learning_rate": 1.558e-05, |
| "loss": 0.0186, |
| "step": 780 |
| }, |
| { |
| "epoch": 0.16185207949190739, |
| "grad_norm": 0.44717493653297424, |
| "learning_rate": 1.578e-05, |
| "loss": 0.0223, |
| "step": 790 |
| }, |
| { |
| "epoch": 0.16390083999180496, |
| "grad_norm": 0.3271142542362213, |
| "learning_rate": 1.598e-05, |
| "loss": 0.0213, |
| "step": 800 |
| }, |
| { |
| "epoch": 0.16594960049170251, |
| "grad_norm": 0.3523205518722534, |
| "learning_rate": 1.618e-05, |
| "loss": 0.0174, |
| "step": 810 |
| }, |
| { |
| "epoch": 0.1679983609916001, |
| "grad_norm": 0.285109281539917, |
| "learning_rate": 1.6380000000000002e-05, |
| "loss": 0.02, |
| "step": 820 |
| }, |
| { |
| "epoch": 0.17004712149149764, |
| "grad_norm": 0.24754098057746887, |
| "learning_rate": 1.658e-05, |
| "loss": 0.0217, |
| "step": 830 |
| }, |
| { |
| "epoch": 0.1720958819913952, |
| "grad_norm": 0.89817875623703, |
| "learning_rate": 1.6780000000000002e-05, |
| "loss": 0.0194, |
| "step": 840 |
| }, |
| { |
| "epoch": 0.17414464249129277, |
| "grad_norm": 0.32891687750816345, |
| "learning_rate": 1.698e-05, |
| "loss": 0.0173, |
| "step": 850 |
| }, |
| { |
| "epoch": 0.17619340299119032, |
| "grad_norm": 0.2777133584022522, |
| "learning_rate": 1.718e-05, |
| "loss": 0.0192, |
| "step": 860 |
| }, |
| { |
| "epoch": 0.1782421634910879, |
| "grad_norm": 0.340331494808197, |
| "learning_rate": 1.7380000000000003e-05, |
| "loss": 0.0161, |
| "step": 870 |
| }, |
| { |
| "epoch": 0.18029092399098545, |
| "grad_norm": 0.3982824683189392, |
| "learning_rate": 1.758e-05, |
| "loss": 0.0208, |
| "step": 880 |
| }, |
| { |
| "epoch": 0.182339684490883, |
| "grad_norm": 0.2832046449184418, |
| "learning_rate": 1.778e-05, |
| "loss": 0.0147, |
| "step": 890 |
| }, |
| { |
| "epoch": 0.18438844499078058, |
| "grad_norm": 0.20551380515098572, |
| "learning_rate": 1.798e-05, |
| "loss": 0.0208, |
| "step": 900 |
| }, |
| { |
| "epoch": 0.18643720549067813, |
| "grad_norm": 0.2500019967556, |
| "learning_rate": 1.818e-05, |
| "loss": 0.0158, |
| "step": 910 |
| }, |
| { |
| "epoch": 0.1884859659905757, |
| "grad_norm": 0.25372886657714844, |
| "learning_rate": 1.838e-05, |
| "loss": 0.0201, |
| "step": 920 |
| }, |
| { |
| "epoch": 0.19053472649047326, |
| "grad_norm": 0.22380007803440094, |
| "learning_rate": 1.8579999999999998e-05, |
| "loss": 0.0159, |
| "step": 930 |
| }, |
| { |
| "epoch": 0.19258348699037084, |
| "grad_norm": 0.17833854258060455, |
| "learning_rate": 1.878e-05, |
| "loss": 0.0195, |
| "step": 940 |
| }, |
| { |
| "epoch": 0.1946322474902684, |
| "grad_norm": 0.42063650488853455, |
| "learning_rate": 1.898e-05, |
| "loss": 0.0183, |
| "step": 950 |
| }, |
| { |
| "epoch": 0.19668100799016594, |
| "grad_norm": 0.4654707908630371, |
| "learning_rate": 1.918e-05, |
| "loss": 0.0215, |
| "step": 960 |
| }, |
| { |
| "epoch": 0.19872976849006352, |
| "grad_norm": 0.19289351999759674, |
| "learning_rate": 1.938e-05, |
| "loss": 0.0165, |
| "step": 970 |
| }, |
| { |
| "epoch": 0.20077852898996107, |
| "grad_norm": 0.29176753759384155, |
| "learning_rate": 1.958e-05, |
| "loss": 0.0157, |
| "step": 980 |
| }, |
| { |
| "epoch": 0.20282728948985865, |
| "grad_norm": 0.2714230716228485, |
| "learning_rate": 1.978e-05, |
| "loss": 0.0175, |
| "step": 990 |
| }, |
| { |
| "epoch": 0.2048760499897562, |
| "grad_norm": 0.31601327657699585, |
| "learning_rate": 1.9980000000000002e-05, |
| "loss": 0.0196, |
| "step": 1000 |
| }, |
| { |
| "epoch": 0.20692481048965375, |
| "grad_norm": 0.20912113785743713, |
| "learning_rate": 2.018e-05, |
| "loss": 0.0163, |
| "step": 1010 |
| }, |
| { |
| "epoch": 0.20897357098955133, |
| "grad_norm": 0.3189891278743744, |
| "learning_rate": 2.038e-05, |
| "loss": 0.0211, |
| "step": 1020 |
| }, |
| { |
| "epoch": 0.21102233148944888, |
| "grad_norm": 0.3637097179889679, |
| "learning_rate": 2.0580000000000003e-05, |
| "loss": 0.0179, |
| "step": 1030 |
| }, |
| { |
| "epoch": 0.21307109198934646, |
| "grad_norm": 0.2773891091346741, |
| "learning_rate": 2.078e-05, |
| "loss": 0.0176, |
| "step": 1040 |
| }, |
| { |
| "epoch": 0.215119852489244, |
| "grad_norm": 0.3385528028011322, |
| "learning_rate": 2.0980000000000002e-05, |
| "loss": 0.0183, |
| "step": 1050 |
| }, |
| { |
| "epoch": 0.21716861298914156, |
| "grad_norm": 0.3089531362056732, |
| "learning_rate": 2.118e-05, |
| "loss": 0.0201, |
| "step": 1060 |
| }, |
| { |
| "epoch": 0.21921737348903914, |
| "grad_norm": 0.3494015336036682, |
| "learning_rate": 2.1380000000000002e-05, |
| "loss": 0.0141, |
| "step": 1070 |
| }, |
| { |
| "epoch": 0.2212661339889367, |
| "grad_norm": 0.27114707231521606, |
| "learning_rate": 2.1580000000000003e-05, |
| "loss": 0.0183, |
| "step": 1080 |
| }, |
| { |
| "epoch": 0.22331489448883426, |
| "grad_norm": 0.5669891834259033, |
| "learning_rate": 2.178e-05, |
| "loss": 0.021, |
| "step": 1090 |
| }, |
| { |
| "epoch": 0.22536365498873182, |
| "grad_norm": 0.31582093238830566, |
| "learning_rate": 2.198e-05, |
| "loss": 0.0167, |
| "step": 1100 |
| }, |
| { |
| "epoch": 0.22741241548862937, |
| "grad_norm": 0.36996686458587646, |
| "learning_rate": 2.2179999999999998e-05, |
| "loss": 0.0147, |
| "step": 1110 |
| }, |
| { |
| "epoch": 0.22946117598852694, |
| "grad_norm": 0.4034775495529175, |
| "learning_rate": 2.238e-05, |
| "loss": 0.0161, |
| "step": 1120 |
| }, |
| { |
| "epoch": 0.2315099364884245, |
| "grad_norm": 0.15472246706485748, |
| "learning_rate": 2.258e-05, |
| "loss": 0.0146, |
| "step": 1130 |
| }, |
| { |
| "epoch": 0.23355869698832207, |
| "grad_norm": 0.17472679913043976, |
| "learning_rate": 2.278e-05, |
| "loss": 0.0153, |
| "step": 1140 |
| }, |
| { |
| "epoch": 0.23560745748821962, |
| "grad_norm": 0.4465899169445038, |
| "learning_rate": 2.298e-05, |
| "loss": 0.0198, |
| "step": 1150 |
| }, |
| { |
| "epoch": 0.2376562179881172, |
| "grad_norm": 0.24651303887367249, |
| "learning_rate": 2.3179999999999998e-05, |
| "loss": 0.0164, |
| "step": 1160 |
| }, |
| { |
| "epoch": 0.23970497848801475, |
| "grad_norm": 0.25848498940467834, |
| "learning_rate": 2.338e-05, |
| "loss": 0.0172, |
| "step": 1170 |
| }, |
| { |
| "epoch": 0.2417537389879123, |
| "grad_norm": 0.4422129988670349, |
| "learning_rate": 2.358e-05, |
| "loss": 0.0167, |
| "step": 1180 |
| }, |
| { |
| "epoch": 0.24380249948780988, |
| "grad_norm": 0.21565169095993042, |
| "learning_rate": 2.378e-05, |
| "loss": 0.0173, |
| "step": 1190 |
| }, |
| { |
| "epoch": 0.24585125998770743, |
| "grad_norm": 0.3105074167251587, |
| "learning_rate": 2.398e-05, |
| "loss": 0.0179, |
| "step": 1200 |
| }, |
| { |
| "epoch": 0.247900020487605, |
| "grad_norm": 0.33732542395591736, |
| "learning_rate": 2.4180000000000002e-05, |
| "loss": 0.0171, |
| "step": 1210 |
| }, |
| { |
| "epoch": 0.24994878098750256, |
| "grad_norm": 0.3114495873451233, |
| "learning_rate": 2.438e-05, |
| "loss": 0.0196, |
| "step": 1220 |
| }, |
| { |
| "epoch": 0.2519975414874001, |
| "grad_norm": 0.2519095838069916, |
| "learning_rate": 2.4580000000000002e-05, |
| "loss": 0.0157, |
| "step": 1230 |
| }, |
| { |
| "epoch": 0.25404630198729766, |
| "grad_norm": 0.22751782834529877, |
| "learning_rate": 2.478e-05, |
| "loss": 0.0191, |
| "step": 1240 |
| }, |
| { |
| "epoch": 0.25609506248719527, |
| "grad_norm": 0.24779312312602997, |
| "learning_rate": 2.498e-05, |
| "loss": 0.0178, |
| "step": 1250 |
| }, |
| { |
| "epoch": 0.2581438229870928, |
| "grad_norm": 0.3079783022403717, |
| "learning_rate": 2.5180000000000003e-05, |
| "loss": 0.0152, |
| "step": 1260 |
| }, |
| { |
| "epoch": 0.26019258348699037, |
| "grad_norm": 0.3924476206302643, |
| "learning_rate": 2.538e-05, |
| "loss": 0.0171, |
| "step": 1270 |
| }, |
| { |
| "epoch": 0.2622413439868879, |
| "grad_norm": 0.17969749867916107, |
| "learning_rate": 2.5580000000000002e-05, |
| "loss": 0.0178, |
| "step": 1280 |
| }, |
| { |
| "epoch": 0.26429010448678547, |
| "grad_norm": 0.35174599289894104, |
| "learning_rate": 2.578e-05, |
| "loss": 0.017, |
| "step": 1290 |
| }, |
| { |
| "epoch": 0.2663388649866831, |
| "grad_norm": 0.5065320730209351, |
| "learning_rate": 2.5980000000000002e-05, |
| "loss": 0.0152, |
| "step": 1300 |
| }, |
| { |
| "epoch": 0.2683876254865806, |
| "grad_norm": 0.29134535789489746, |
| "learning_rate": 2.618e-05, |
| "loss": 0.016, |
| "step": 1310 |
| }, |
| { |
| "epoch": 0.2704363859864782, |
| "grad_norm": 0.3602963089942932, |
| "learning_rate": 2.6379999999999998e-05, |
| "loss": 0.0177, |
| "step": 1320 |
| }, |
| { |
| "epoch": 0.27248514648637573, |
| "grad_norm": 0.330760657787323, |
| "learning_rate": 2.658e-05, |
| "loss": 0.0159, |
| "step": 1330 |
| }, |
| { |
| "epoch": 0.2745339069862733, |
| "grad_norm": 0.4138154983520508, |
| "learning_rate": 2.678e-05, |
| "loss": 0.0179, |
| "step": 1340 |
| }, |
| { |
| "epoch": 0.2765826674861709, |
| "grad_norm": 0.26150503754615784, |
| "learning_rate": 2.698e-05, |
| "loss": 0.0153, |
| "step": 1350 |
| }, |
| { |
| "epoch": 0.27863142798606844, |
| "grad_norm": 0.22318997979164124, |
| "learning_rate": 2.718e-05, |
| "loss": 0.0197, |
| "step": 1360 |
| }, |
| { |
| "epoch": 0.280680188485966, |
| "grad_norm": 0.19625113904476166, |
| "learning_rate": 2.738e-05, |
| "loss": 0.0169, |
| "step": 1370 |
| }, |
| { |
| "epoch": 0.28272894898586354, |
| "grad_norm": 0.3631628453731537, |
| "learning_rate": 2.758e-05, |
| "loss": 0.0177, |
| "step": 1380 |
| }, |
| { |
| "epoch": 0.2847777094857611, |
| "grad_norm": 0.328561395406723, |
| "learning_rate": 2.778e-05, |
| "loss": 0.0156, |
| "step": 1390 |
| }, |
| { |
| "epoch": 0.2868264699856587, |
| "grad_norm": 0.3231017291545868, |
| "learning_rate": 2.798e-05, |
| "loss": 0.0141, |
| "step": 1400 |
| }, |
| { |
| "epoch": 0.28887523048555624, |
| "grad_norm": 0.46342310309410095, |
| "learning_rate": 2.818e-05, |
| "loss": 0.0143, |
| "step": 1410 |
| }, |
| { |
| "epoch": 0.2909239909854538, |
| "grad_norm": 0.22275222837924957, |
| "learning_rate": 2.838e-05, |
| "loss": 0.022, |
| "step": 1420 |
| }, |
| { |
| "epoch": 0.29297275148535135, |
| "grad_norm": 0.40549004077911377, |
| "learning_rate": 2.858e-05, |
| "loss": 0.0187, |
| "step": 1430 |
| }, |
| { |
| "epoch": 0.29502151198524895, |
| "grad_norm": 0.19027435779571533, |
| "learning_rate": 2.8780000000000002e-05, |
| "loss": 0.0161, |
| "step": 1440 |
| }, |
| { |
| "epoch": 0.2970702724851465, |
| "grad_norm": 0.21578077971935272, |
| "learning_rate": 2.898e-05, |
| "loss": 0.0172, |
| "step": 1450 |
| }, |
| { |
| "epoch": 0.29911903298504405, |
| "grad_norm": 0.1968519687652588, |
| "learning_rate": 2.9180000000000002e-05, |
| "loss": 0.0165, |
| "step": 1460 |
| }, |
| { |
| "epoch": 0.3011677934849416, |
| "grad_norm": 0.19503623247146606, |
| "learning_rate": 2.938e-05, |
| "loss": 0.014, |
| "step": 1470 |
| }, |
| { |
| "epoch": 0.30321655398483915, |
| "grad_norm": 0.36131182312965393, |
| "learning_rate": 2.958e-05, |
| "loss": 0.0153, |
| "step": 1480 |
| }, |
| { |
| "epoch": 0.30526531448473676, |
| "grad_norm": 0.17971989512443542, |
| "learning_rate": 2.9780000000000003e-05, |
| "loss": 0.0145, |
| "step": 1490 |
| }, |
| { |
| "epoch": 0.3073140749846343, |
| "grad_norm": 0.1925322711467743, |
| "learning_rate": 2.998e-05, |
| "loss": 0.0146, |
| "step": 1500 |
| }, |
| { |
| "epoch": 0.30936283548453186, |
| "grad_norm": 0.3785518705844879, |
| "learning_rate": 2.9999992618302022e-05, |
| "loss": 0.0135, |
| "step": 1510 |
| }, |
| { |
| "epoch": 0.3114115959844294, |
| "grad_norm": 0.21011051535606384, |
| "learning_rate": 2.999996710133069e-05, |
| "loss": 0.0159, |
| "step": 1520 |
| }, |
| { |
| "epoch": 0.31346035648432696, |
| "grad_norm": 0.2054048627614975, |
| "learning_rate": 2.9999923357984923e-05, |
| "loss": 0.0147, |
| "step": 1530 |
| }, |
| { |
| "epoch": 0.31550911698422457, |
| "grad_norm": 0.41478946805000305, |
| "learning_rate": 2.9999861388317874e-05, |
| "loss": 0.0126, |
| "step": 1540 |
| }, |
| { |
| "epoch": 0.3175578774841221, |
| "grad_norm": 0.46991223096847534, |
| "learning_rate": 2.9999781192404847e-05, |
| "loss": 0.0175, |
| "step": 1550 |
| }, |
| { |
| "epoch": 0.31960663798401967, |
| "grad_norm": 0.2961818277835846, |
| "learning_rate": 2.9999682770343283e-05, |
| "loss": 0.0161, |
| "step": 1560 |
| }, |
| { |
| "epoch": 0.3216553984839172, |
| "grad_norm": 0.4989997446537018, |
| "learning_rate": 2.9999566122252777e-05, |
| "loss": 0.0154, |
| "step": 1570 |
| }, |
| { |
| "epoch": 0.32370415898381477, |
| "grad_norm": 0.21468670666217804, |
| "learning_rate": 2.9999431248275065e-05, |
| "loss": 0.0167, |
| "step": 1580 |
| }, |
| { |
| "epoch": 0.3257529194837124, |
| "grad_norm": 0.26594579219818115, |
| "learning_rate": 2.9999278148574027e-05, |
| "loss": 0.016, |
| "step": 1590 |
| }, |
| { |
| "epoch": 0.32780167998360993, |
| "grad_norm": 0.3335718810558319, |
| "learning_rate": 2.9999106823335705e-05, |
| "loss": 0.0161, |
| "step": 1600 |
| }, |
| { |
| "epoch": 0.3298504404835075, |
| "grad_norm": 0.22297169268131256, |
| "learning_rate": 2.999891727276827e-05, |
| "loss": 0.0177, |
| "step": 1610 |
| }, |
| { |
| "epoch": 0.33189920098340503, |
| "grad_norm": 0.3370715379714966, |
| "learning_rate": 2.9998709497102037e-05, |
| "loss": 0.0143, |
| "step": 1620 |
| }, |
| { |
| "epoch": 0.3339479614833026, |
| "grad_norm": 0.3054150640964508, |
| "learning_rate": 2.9998483496589485e-05, |
| "loss": 0.0182, |
| "step": 1630 |
| }, |
| { |
| "epoch": 0.3359967219832002, |
| "grad_norm": 0.22999905049800873, |
| "learning_rate": 2.9998239271505217e-05, |
| "loss": 0.0164, |
| "step": 1640 |
| }, |
| { |
| "epoch": 0.33804548248309774, |
| "grad_norm": 0.27013978362083435, |
| "learning_rate": 2.9997976822145995e-05, |
| "loss": 0.0153, |
| "step": 1650 |
| }, |
| { |
| "epoch": 0.3400942429829953, |
| "grad_norm": 0.3299371898174286, |
| "learning_rate": 2.9997696148830714e-05, |
| "loss": 0.0181, |
| "step": 1660 |
| }, |
| { |
| "epoch": 0.34214300348289284, |
| "grad_norm": 0.24037165939807892, |
| "learning_rate": 2.9997397251900423e-05, |
| "loss": 0.0165, |
| "step": 1670 |
| }, |
| { |
| "epoch": 0.3441917639827904, |
| "grad_norm": 0.3235500752925873, |
| "learning_rate": 2.999708013171831e-05, |
| "loss": 0.0167, |
| "step": 1680 |
| }, |
| { |
| "epoch": 0.346240524482688, |
| "grad_norm": 0.46328964829444885, |
| "learning_rate": 2.9996744788669705e-05, |
| "loss": 0.015, |
| "step": 1690 |
| }, |
| { |
| "epoch": 0.34828928498258555, |
| "grad_norm": 0.22368527948856354, |
| "learning_rate": 2.999639122316208e-05, |
| "loss": 0.0154, |
| "step": 1700 |
| }, |
| { |
| "epoch": 0.3503380454824831, |
| "grad_norm": 0.2481120079755783, |
| "learning_rate": 2.999601943562505e-05, |
| "loss": 0.0193, |
| "step": 1710 |
| }, |
| { |
| "epoch": 0.35238680598238065, |
| "grad_norm": 0.22972320020198822, |
| "learning_rate": 2.9995629426510377e-05, |
| "loss": 0.0151, |
| "step": 1720 |
| }, |
| { |
| "epoch": 0.3544355664822782, |
| "grad_norm": 0.40488821268081665, |
| "learning_rate": 2.9995221196291952e-05, |
| "loss": 0.0112, |
| "step": 1730 |
| }, |
| { |
| "epoch": 0.3564843269821758, |
| "grad_norm": 0.6761015057563782, |
| "learning_rate": 2.999479474546582e-05, |
| "loss": 0.0186, |
| "step": 1740 |
| }, |
| { |
| "epoch": 0.35853308748207335, |
| "grad_norm": 0.49233734607696533, |
| "learning_rate": 2.9994350074550153e-05, |
| "loss": 0.0169, |
| "step": 1750 |
| }, |
| { |
| "epoch": 0.3605818479819709, |
| "grad_norm": 0.34043994545936584, |
| "learning_rate": 2.9993887184085273e-05, |
| "loss": 0.0151, |
| "step": 1760 |
| }, |
| { |
| "epoch": 0.36263060848186845, |
| "grad_norm": 0.19139300286769867, |
| "learning_rate": 2.9993406074633628e-05, |
| "loss": 0.0137, |
| "step": 1770 |
| }, |
| { |
| "epoch": 0.364679368981766, |
| "grad_norm": 0.36463215947151184, |
| "learning_rate": 2.9992906746779824e-05, |
| "loss": 0.0154, |
| "step": 1780 |
| }, |
| { |
| "epoch": 0.3667281294816636, |
| "grad_norm": 0.6888066530227661, |
| "learning_rate": 2.9992389201130576e-05, |
| "loss": 0.0174, |
| "step": 1790 |
| }, |
| { |
| "epoch": 0.36877688998156116, |
| "grad_norm": 0.19548574090003967, |
| "learning_rate": 2.999185343831476e-05, |
| "loss": 0.0195, |
| "step": 1800 |
| }, |
| { |
| "epoch": 0.3708256504814587, |
| "grad_norm": 0.3563873767852783, |
| "learning_rate": 2.999129945898338e-05, |
| "loss": 0.0175, |
| "step": 1810 |
| }, |
| { |
| "epoch": 0.37287441098135626, |
| "grad_norm": 0.19386987388134003, |
| "learning_rate": 2.999072726380957e-05, |
| "loss": 0.0192, |
| "step": 1820 |
| }, |
| { |
| "epoch": 0.3749231714812538, |
| "grad_norm": 0.2008875459432602, |
| "learning_rate": 2.9990136853488603e-05, |
| "loss": 0.0138, |
| "step": 1830 |
| }, |
| { |
| "epoch": 0.3769719319811514, |
| "grad_norm": 0.27989453077316284, |
| "learning_rate": 2.9989528228737878e-05, |
| "loss": 0.0148, |
| "step": 1840 |
| }, |
| { |
| "epoch": 0.37902069248104897, |
| "grad_norm": 0.1453440636396408, |
| "learning_rate": 2.9988901390296935e-05, |
| "loss": 0.0154, |
| "step": 1850 |
| }, |
| { |
| "epoch": 0.3810694529809465, |
| "grad_norm": 0.29804226756095886, |
| "learning_rate": 2.9988256338927442e-05, |
| "loss": 0.0135, |
| "step": 1860 |
| }, |
| { |
| "epoch": 0.38311821348084407, |
| "grad_norm": 0.3325411379337311, |
| "learning_rate": 2.9987593075413198e-05, |
| "loss": 0.0179, |
| "step": 1870 |
| }, |
| { |
| "epoch": 0.3851669739807417, |
| "grad_norm": 0.3169262707233429, |
| "learning_rate": 2.9986911600560128e-05, |
| "loss": 0.0162, |
| "step": 1880 |
| }, |
| { |
| "epoch": 0.38721573448063923, |
| "grad_norm": 0.25286865234375, |
| "learning_rate": 2.998621191519629e-05, |
| "loss": 0.0167, |
| "step": 1890 |
| }, |
| { |
| "epoch": 0.3892644949805368, |
| "grad_norm": 0.4364844560623169, |
| "learning_rate": 2.9985494020171873e-05, |
| "loss": 0.016, |
| "step": 1900 |
| }, |
| { |
| "epoch": 0.39131325548043433, |
| "grad_norm": 0.27653005719184875, |
| "learning_rate": 2.9984757916359176e-05, |
| "loss": 0.0169, |
| "step": 1910 |
| }, |
| { |
| "epoch": 0.3933620159803319, |
| "grad_norm": 0.21330104768276215, |
| "learning_rate": 2.998400360465264e-05, |
| "loss": 0.0151, |
| "step": 1920 |
| }, |
| { |
| "epoch": 0.3954107764802295, |
| "grad_norm": 0.19991207122802734, |
| "learning_rate": 2.998323108596883e-05, |
| "loss": 0.0124, |
| "step": 1930 |
| }, |
| { |
| "epoch": 0.39745953698012704, |
| "grad_norm": 0.2834044098854065, |
| "learning_rate": 2.9982440361246422e-05, |
| "loss": 0.0152, |
| "step": 1940 |
| }, |
| { |
| "epoch": 0.3995082974800246, |
| "grad_norm": 0.46298515796661377, |
| "learning_rate": 2.998163143144622e-05, |
| "loss": 0.0164, |
| "step": 1950 |
| }, |
| { |
| "epoch": 0.40155705797992214, |
| "grad_norm": 0.7101344466209412, |
| "learning_rate": 2.9980804297551155e-05, |
| "loss": 0.0165, |
| "step": 1960 |
| }, |
| { |
| "epoch": 0.4036058184798197, |
| "grad_norm": 0.23057594895362854, |
| "learning_rate": 2.997995896056627e-05, |
| "loss": 0.0177, |
| "step": 1970 |
| }, |
| { |
| "epoch": 0.4056545789797173, |
| "grad_norm": 0.23996594548225403, |
| "learning_rate": 2.997909542151873e-05, |
| "loss": 0.0189, |
| "step": 1980 |
| }, |
| { |
| "epoch": 0.40770333947961485, |
| "grad_norm": 0.2810029685497284, |
| "learning_rate": 2.9978213681457823e-05, |
| "loss": 0.0123, |
| "step": 1990 |
| }, |
| { |
| "epoch": 0.4097520999795124, |
| "grad_norm": 0.48422539234161377, |
| "learning_rate": 2.9977313741454933e-05, |
| "loss": 0.0178, |
| "step": 2000 |
| }, |
| { |
| "epoch": 0.41180086047940995, |
| "grad_norm": 0.4375326335430145, |
| "learning_rate": 2.997639560260358e-05, |
| "loss": 0.0175, |
| "step": 2010 |
| }, |
| { |
| "epoch": 0.4138496209793075, |
| "grad_norm": 0.2832667827606201, |
| "learning_rate": 2.9975459266019387e-05, |
| "loss": 0.0144, |
| "step": 2020 |
| }, |
| { |
| "epoch": 0.4158983814792051, |
| "grad_norm": 0.20203521847724915, |
| "learning_rate": 2.9974504732840092e-05, |
| "loss": 0.0151, |
| "step": 2030 |
| }, |
| { |
| "epoch": 0.41794714197910265, |
| "grad_norm": 0.3503471314907074, |
| "learning_rate": 2.997353200422555e-05, |
| "loss": 0.0166, |
| "step": 2040 |
| }, |
| { |
| "epoch": 0.4199959024790002, |
| "grad_norm": 0.22971761226654053, |
| "learning_rate": 2.99725410813577e-05, |
| "loss": 0.015, |
| "step": 2050 |
| }, |
| { |
| "epoch": 0.42204466297889776, |
| "grad_norm": 0.23142488300800323, |
| "learning_rate": 2.997153196544063e-05, |
| "loss": 0.0158, |
| "step": 2060 |
| }, |
| { |
| "epoch": 0.4240934234787953, |
| "grad_norm": 0.30666008591651917, |
| "learning_rate": 2.997050465770049e-05, |
| "loss": 0.0155, |
| "step": 2070 |
| }, |
| { |
| "epoch": 0.4261421839786929, |
| "grad_norm": 0.19076187908649445, |
| "learning_rate": 2.9969459159385562e-05, |
| "loss": 0.0154, |
| "step": 2080 |
| }, |
| { |
| "epoch": 0.42819094447859046, |
| "grad_norm": 0.38849496841430664, |
| "learning_rate": 2.9968395471766237e-05, |
| "loss": 0.0149, |
| "step": 2090 |
| }, |
| { |
| "epoch": 0.430239704978488, |
| "grad_norm": 0.2731882929801941, |
| "learning_rate": 2.996731359613498e-05, |
| "loss": 0.0154, |
| "step": 2100 |
| }, |
| { |
| "epoch": 0.43228846547838556, |
| "grad_norm": 0.19518789649009705, |
| "learning_rate": 2.996621353380638e-05, |
| "loss": 0.0168, |
| "step": 2110 |
| }, |
| { |
| "epoch": 0.4343372259782831, |
| "grad_norm": 0.21378087997436523, |
| "learning_rate": 2.9965095286117113e-05, |
| "loss": 0.0155, |
| "step": 2120 |
| }, |
| { |
| "epoch": 0.4363859864781807, |
| "grad_norm": 0.31314411759376526, |
| "learning_rate": 2.996395885442596e-05, |
| "loss": 0.0145, |
| "step": 2130 |
| }, |
| { |
| "epoch": 0.43843474697807827, |
| "grad_norm": 0.22869321703910828, |
| "learning_rate": 2.996280424011379e-05, |
| "loss": 0.0154, |
| "step": 2140 |
| }, |
| { |
| "epoch": 0.4404835074779758, |
| "grad_norm": 0.2786426842212677, |
| "learning_rate": 2.9961631444583565e-05, |
| "loss": 0.0146, |
| "step": 2150 |
| }, |
| { |
| "epoch": 0.4425322679778734, |
| "grad_norm": 0.3453676998615265, |
| "learning_rate": 2.9960440469260347e-05, |
| "loss": 0.0144, |
| "step": 2160 |
| }, |
| { |
| "epoch": 0.4445810284777709, |
| "grad_norm": 0.28185728192329407, |
| "learning_rate": 2.995923131559128e-05, |
| "loss": 0.0171, |
| "step": 2170 |
| }, |
| { |
| "epoch": 0.44662978897766853, |
| "grad_norm": 0.2775164842605591, |
| "learning_rate": 2.99580039850456e-05, |
| "loss": 0.0122, |
| "step": 2180 |
| }, |
| { |
| "epoch": 0.4486785494775661, |
| "grad_norm": 0.22074560821056366, |
| "learning_rate": 2.9956758479114635e-05, |
| "loss": 0.0161, |
| "step": 2190 |
| }, |
| { |
| "epoch": 0.45072730997746363, |
| "grad_norm": 0.16912080347537994, |
| "learning_rate": 2.9955494799311787e-05, |
| "loss": 0.0144, |
| "step": 2200 |
| }, |
| { |
| "epoch": 0.4527760704773612, |
| "grad_norm": 0.32740625739097595, |
| "learning_rate": 2.995421294717254e-05, |
| "loss": 0.0156, |
| "step": 2210 |
| }, |
| { |
| "epoch": 0.45482483097725873, |
| "grad_norm": 0.4528239965438843, |
| "learning_rate": 2.995291292425448e-05, |
| "loss": 0.0117, |
| "step": 2220 |
| }, |
| { |
| "epoch": 0.45687359147715634, |
| "grad_norm": 0.26946398615837097, |
| "learning_rate": 2.9951594732137248e-05, |
| "loss": 0.0154, |
| "step": 2230 |
| }, |
| { |
| "epoch": 0.4589223519770539, |
| "grad_norm": 0.2879970073699951, |
| "learning_rate": 2.995025837242257e-05, |
| "loss": 0.0151, |
| "step": 2240 |
| }, |
| { |
| "epoch": 0.46097111247695144, |
| "grad_norm": 0.16226619482040405, |
| "learning_rate": 2.994890384673426e-05, |
| "loss": 0.0124, |
| "step": 2250 |
| }, |
| { |
| "epoch": 0.463019872976849, |
| "grad_norm": 0.23087091743946075, |
| "learning_rate": 2.9947531156718174e-05, |
| "loss": 0.0148, |
| "step": 2260 |
| }, |
| { |
| "epoch": 0.4650686334767466, |
| "grad_norm": 0.342555969953537, |
| "learning_rate": 2.994614030404228e-05, |
| "loss": 0.0188, |
| "step": 2270 |
| }, |
| { |
| "epoch": 0.46711739397664415, |
| "grad_norm": 0.2865619361400604, |
| "learning_rate": 2.994473129039659e-05, |
| "loss": 0.0139, |
| "step": 2280 |
| }, |
| { |
| "epoch": 0.4691661544765417, |
| "grad_norm": 0.3614512085914612, |
| "learning_rate": 2.9943304117493183e-05, |
| "loss": 0.0136, |
| "step": 2290 |
| }, |
| { |
| "epoch": 0.47121491497643925, |
| "grad_norm": 0.3581768572330475, |
| "learning_rate": 2.9941858787066207e-05, |
| "loss": 0.014, |
| "step": 2300 |
| }, |
| { |
| "epoch": 0.4732636754763368, |
| "grad_norm": 0.19030360877513885, |
| "learning_rate": 2.9940395300871888e-05, |
| "loss": 0.011, |
| "step": 2310 |
| }, |
| { |
| "epoch": 0.4753124359762344, |
| "grad_norm": 0.18200714886188507, |
| "learning_rate": 2.993891366068849e-05, |
| "loss": 0.014, |
| "step": 2320 |
| }, |
| { |
| "epoch": 0.47736119647613195, |
| "grad_norm": 0.28966355323791504, |
| "learning_rate": 2.993741386831634e-05, |
| "loss": 0.0144, |
| "step": 2330 |
| }, |
| { |
| "epoch": 0.4794099569760295, |
| "grad_norm": 0.28325843811035156, |
| "learning_rate": 2.9935895925577837e-05, |
| "loss": 0.0132, |
| "step": 2340 |
| }, |
| { |
| "epoch": 0.48145871747592706, |
| "grad_norm": 0.1583881676197052, |
| "learning_rate": 2.9934359834317425e-05, |
| "loss": 0.018, |
| "step": 2350 |
| }, |
| { |
| "epoch": 0.4835074779758246, |
| "grad_norm": 0.22520609200000763, |
| "learning_rate": 2.993280559640159e-05, |
| "loss": 0.0166, |
| "step": 2360 |
| }, |
| { |
| "epoch": 0.4855562384757222, |
| "grad_norm": 0.19799301028251648, |
| "learning_rate": 2.9931233213718887e-05, |
| "loss": 0.0176, |
| "step": 2370 |
| }, |
| { |
| "epoch": 0.48760499897561976, |
| "grad_norm": 0.20375941693782806, |
| "learning_rate": 2.992964268817991e-05, |
| "loss": 0.0142, |
| "step": 2380 |
| }, |
| { |
| "epoch": 0.4896537594755173, |
| "grad_norm": 0.2075621485710144, |
| "learning_rate": 2.9928034021717286e-05, |
| "loss": 0.0135, |
| "step": 2390 |
| }, |
| { |
| "epoch": 0.49170251997541486, |
| "grad_norm": 0.49743375182151794, |
| "learning_rate": 2.9926407216285707e-05, |
| "loss": 0.0162, |
| "step": 2400 |
| }, |
| { |
| "epoch": 0.4937512804753124, |
| "grad_norm": 0.17372027039527893, |
| "learning_rate": 2.9924762273861894e-05, |
| "loss": 0.0166, |
| "step": 2410 |
| }, |
| { |
| "epoch": 0.49580004097521, |
| "grad_norm": 0.34063801169395447, |
| "learning_rate": 2.992309919644461e-05, |
| "loss": 0.0141, |
| "step": 2420 |
| }, |
| { |
| "epoch": 0.49784880147510757, |
| "grad_norm": 0.38418591022491455, |
| "learning_rate": 2.9921417986054633e-05, |
| "loss": 0.0139, |
| "step": 2430 |
| }, |
| { |
| "epoch": 0.4998975619750051, |
| "grad_norm": 0.7703158259391785, |
| "learning_rate": 2.9919718644734815e-05, |
| "loss": 0.0132, |
| "step": 2440 |
| }, |
| { |
| "epoch": 0.5019463224749027, |
| "grad_norm": 0.27611294388771057, |
| "learning_rate": 2.991800117455e-05, |
| "loss": 0.0166, |
| "step": 2450 |
| }, |
| { |
| "epoch": 0.5039950829748002, |
| "grad_norm": 0.5423811674118042, |
| "learning_rate": 2.9916265577587085e-05, |
| "loss": 0.0164, |
| "step": 2460 |
| }, |
| { |
| "epoch": 0.5060438434746978, |
| "grad_norm": 0.309264600276947, |
| "learning_rate": 2.9914511855954986e-05, |
| "loss": 0.0165, |
| "step": 2470 |
| }, |
| { |
| "epoch": 0.5080926039745953, |
| "grad_norm": 0.3045426309108734, |
| "learning_rate": 2.9912740011784626e-05, |
| "loss": 0.0161, |
| "step": 2480 |
| }, |
| { |
| "epoch": 0.510141364474493, |
| "grad_norm": 0.5828838348388672, |
| "learning_rate": 2.991095004722898e-05, |
| "loss": 0.0178, |
| "step": 2490 |
| }, |
| { |
| "epoch": 0.5121901249743905, |
| "grad_norm": 0.29532894492149353, |
| "learning_rate": 2.9909141964463017e-05, |
| "loss": 0.0154, |
| "step": 2500 |
| }, |
| { |
| "epoch": 0.5142388854742881, |
| "grad_norm": 0.2532103359699249, |
| "learning_rate": 2.990731576568372e-05, |
| "loss": 0.0161, |
| "step": 2510 |
| }, |
| { |
| "epoch": 0.5162876459741856, |
| "grad_norm": 0.25809621810913086, |
| "learning_rate": 2.99054714531101e-05, |
| "loss": 0.0151, |
| "step": 2520 |
| }, |
| { |
| "epoch": 0.5183364064740832, |
| "grad_norm": 0.30241936445236206, |
| "learning_rate": 2.9903609028983178e-05, |
| "loss": 0.0145, |
| "step": 2530 |
| }, |
| { |
| "epoch": 0.5203851669739807, |
| "grad_norm": 0.35125237703323364, |
| "learning_rate": 2.990172849556596e-05, |
| "loss": 0.0158, |
| "step": 2540 |
| }, |
| { |
| "epoch": 0.5224339274738783, |
| "grad_norm": 0.25003692507743835, |
| "learning_rate": 2.989982985514348e-05, |
| "loss": 0.0169, |
| "step": 2550 |
| }, |
| { |
| "epoch": 0.5244826879737758, |
| "grad_norm": 0.2216779887676239, |
| "learning_rate": 2.9897913110022766e-05, |
| "loss": 0.0158, |
| "step": 2560 |
| }, |
| { |
| "epoch": 0.5265314484736734, |
| "grad_norm": 0.2502020299434662, |
| "learning_rate": 2.989597826253284e-05, |
| "loss": 0.0138, |
| "step": 2570 |
| }, |
| { |
| "epoch": 0.5285802089735709, |
| "grad_norm": 0.24631421267986298, |
| "learning_rate": 2.989402531502473e-05, |
| "loss": 0.0214, |
| "step": 2580 |
| }, |
| { |
| "epoch": 0.5306289694734686, |
| "grad_norm": 0.19882503151893616, |
| "learning_rate": 2.989205426987144e-05, |
| "loss": 0.0131, |
| "step": 2590 |
| }, |
| { |
| "epoch": 0.5326777299733662, |
| "grad_norm": 0.2288060188293457, |
| "learning_rate": 2.9890065129467988e-05, |
| "loss": 0.0157, |
| "step": 2600 |
| }, |
| { |
| "epoch": 0.5347264904732637, |
| "grad_norm": 0.22312426567077637, |
| "learning_rate": 2.9888057896231364e-05, |
| "loss": 0.0142, |
| "step": 2610 |
| }, |
| { |
| "epoch": 0.5367752509731613, |
| "grad_norm": 0.2598424553871155, |
| "learning_rate": 2.988603257260054e-05, |
| "loss": 0.0132, |
| "step": 2620 |
| }, |
| { |
| "epoch": 0.5388240114730588, |
| "grad_norm": 0.5017476677894592, |
| "learning_rate": 2.988398916103648e-05, |
| "loss": 0.0136, |
| "step": 2630 |
| }, |
| { |
| "epoch": 0.5408727719729564, |
| "grad_norm": 0.3976192772388458, |
| "learning_rate": 2.988192766402212e-05, |
| "loss": 0.0148, |
| "step": 2640 |
| }, |
| { |
| "epoch": 0.5429215324728539, |
| "grad_norm": 0.25762036442756653, |
| "learning_rate": 2.9879848084062378e-05, |
| "loss": 0.0155, |
| "step": 2650 |
| }, |
| { |
| "epoch": 0.5449702929727515, |
| "grad_norm": 0.16714882850646973, |
| "learning_rate": 2.9877750423684132e-05, |
| "loss": 0.0152, |
| "step": 2660 |
| }, |
| { |
| "epoch": 0.547019053472649, |
| "grad_norm": 0.16751503944396973, |
| "learning_rate": 2.987563468543624e-05, |
| "loss": 0.0142, |
| "step": 2670 |
| }, |
| { |
| "epoch": 0.5490678139725466, |
| "grad_norm": 0.4317106008529663, |
| "learning_rate": 2.9873500871889533e-05, |
| "loss": 0.0171, |
| "step": 2680 |
| }, |
| { |
| "epoch": 0.5511165744724442, |
| "grad_norm": 0.3907589614391327, |
| "learning_rate": 2.9871348985636778e-05, |
| "loss": 0.0152, |
| "step": 2690 |
| }, |
| { |
| "epoch": 0.5531653349723418, |
| "grad_norm": 0.3212253451347351, |
| "learning_rate": 2.986917902929273e-05, |
| "loss": 0.0148, |
| "step": 2700 |
| }, |
| { |
| "epoch": 0.5552140954722393, |
| "grad_norm": 0.21875205636024475, |
| "learning_rate": 2.9866991005494095e-05, |
| "loss": 0.0149, |
| "step": 2710 |
| }, |
| { |
| "epoch": 0.5572628559721369, |
| "grad_norm": 0.3636499047279358, |
| "learning_rate": 2.986478491689952e-05, |
| "loss": 0.0162, |
| "step": 2720 |
| }, |
| { |
| "epoch": 0.5593116164720344, |
| "grad_norm": 0.2761183977127075, |
| "learning_rate": 2.986256076618961e-05, |
| "loss": 0.0175, |
| "step": 2730 |
| }, |
| { |
| "epoch": 0.561360376971932, |
| "grad_norm": 0.17025238275527954, |
| "learning_rate": 2.986031855606692e-05, |
| "loss": 0.0137, |
| "step": 2740 |
| }, |
| { |
| "epoch": 0.5634091374718295, |
| "grad_norm": 0.17907533049583435, |
| "learning_rate": 2.9858058289255954e-05, |
| "loss": 0.0149, |
| "step": 2750 |
| }, |
| { |
| "epoch": 0.5654578979717271, |
| "grad_norm": 0.24703305959701538, |
| "learning_rate": 2.985577996850314e-05, |
| "loss": 0.0119, |
| "step": 2760 |
| }, |
| { |
| "epoch": 0.5675066584716246, |
| "grad_norm": 0.3591955304145813, |
| "learning_rate": 2.9853483596576856e-05, |
| "loss": 0.016, |
| "step": 2770 |
| }, |
| { |
| "epoch": 0.5695554189715222, |
| "grad_norm": 0.3125201165676117, |
| "learning_rate": 2.9851169176267412e-05, |
| "loss": 0.0155, |
| "step": 2780 |
| }, |
| { |
| "epoch": 0.5716041794714198, |
| "grad_norm": 0.2338644415140152, |
| "learning_rate": 2.9848836710387044e-05, |
| "loss": 0.015, |
| "step": 2790 |
| }, |
| { |
| "epoch": 0.5736529399713174, |
| "grad_norm": 0.295673668384552, |
| "learning_rate": 2.9846486201769917e-05, |
| "loss": 0.0131, |
| "step": 2800 |
| }, |
| { |
| "epoch": 0.5757017004712149, |
| "grad_norm": 0.21388863027095795, |
| "learning_rate": 2.9844117653272122e-05, |
| "loss": 0.0164, |
| "step": 2810 |
| }, |
| { |
| "epoch": 0.5777504609711125, |
| "grad_norm": 0.2705565392971039, |
| "learning_rate": 2.9841731067771676e-05, |
| "loss": 0.0144, |
| "step": 2820 |
| }, |
| { |
| "epoch": 0.57979922147101, |
| "grad_norm": 0.23795472085475922, |
| "learning_rate": 2.9839326448168492e-05, |
| "loss": 0.0143, |
| "step": 2830 |
| }, |
| { |
| "epoch": 0.5818479819709076, |
| "grad_norm": 0.3617759346961975, |
| "learning_rate": 2.983690379738442e-05, |
| "loss": 0.0141, |
| "step": 2840 |
| }, |
| { |
| "epoch": 0.5838967424708051, |
| "grad_norm": 0.2801937460899353, |
| "learning_rate": 2.9834463118363204e-05, |
| "loss": 0.0143, |
| "step": 2850 |
| }, |
| { |
| "epoch": 0.5859455029707027, |
| "grad_norm": 0.2911805212497711, |
| "learning_rate": 2.98320044140705e-05, |
| "loss": 0.0155, |
| "step": 2860 |
| }, |
| { |
| "epoch": 0.5879942634706002, |
| "grad_norm": 0.19783718883991241, |
| "learning_rate": 2.9829527687493868e-05, |
| "loss": 0.0135, |
| "step": 2870 |
| }, |
| { |
| "epoch": 0.5900430239704979, |
| "grad_norm": 0.20388475060462952, |
| "learning_rate": 2.9827032941642765e-05, |
| "loss": 0.0126, |
| "step": 2880 |
| }, |
| { |
| "epoch": 0.5920917844703955, |
| "grad_norm": 0.18498742580413818, |
| "learning_rate": 2.9824520179548535e-05, |
| "loss": 0.0125, |
| "step": 2890 |
| }, |
| { |
| "epoch": 0.594140544970293, |
| "grad_norm": 0.1449342668056488, |
| "learning_rate": 2.9821989404264428e-05, |
| "loss": 0.0126, |
| "step": 2900 |
| }, |
| { |
| "epoch": 0.5961893054701906, |
| "grad_norm": 0.25185492634773254, |
| "learning_rate": 2.9819440618865572e-05, |
| "loss": 0.0156, |
| "step": 2910 |
| }, |
| { |
| "epoch": 0.5982380659700881, |
| "grad_norm": 0.20292209088802338, |
| "learning_rate": 2.981687382644898e-05, |
| "loss": 0.0135, |
| "step": 2920 |
| }, |
| { |
| "epoch": 0.6002868264699857, |
| "grad_norm": 0.2614067494869232, |
| "learning_rate": 2.9814289030133556e-05, |
| "loss": 0.0119, |
| "step": 2930 |
| }, |
| { |
| "epoch": 0.6023355869698832, |
| "grad_norm": 0.20472529530525208, |
| "learning_rate": 2.9811686233060058e-05, |
| "loss": 0.0142, |
| "step": 2940 |
| }, |
| { |
| "epoch": 0.6043843474697808, |
| "grad_norm": 0.24615521728992462, |
| "learning_rate": 2.9809065438391138e-05, |
| "loss": 0.0163, |
| "step": 2950 |
| }, |
| { |
| "epoch": 0.6064331079696783, |
| "grad_norm": 0.21681958436965942, |
| "learning_rate": 2.9806426649311304e-05, |
| "loss": 0.0121, |
| "step": 2960 |
| }, |
| { |
| "epoch": 0.6084818684695759, |
| "grad_norm": 0.222805917263031, |
| "learning_rate": 2.980376986902694e-05, |
| "loss": 0.0125, |
| "step": 2970 |
| }, |
| { |
| "epoch": 0.6105306289694735, |
| "grad_norm": 0.40812230110168457, |
| "learning_rate": 2.9801095100766276e-05, |
| "loss": 0.0149, |
| "step": 2980 |
| }, |
| { |
| "epoch": 0.6125793894693711, |
| "grad_norm": 0.25130754709243774, |
| "learning_rate": 2.979840234777942e-05, |
| "loss": 0.0161, |
| "step": 2990 |
| }, |
| { |
| "epoch": 0.6146281499692686, |
| "grad_norm": 0.6623209118843079, |
| "learning_rate": 2.979569161333831e-05, |
| "loss": 0.0139, |
| "step": 3000 |
| }, |
| { |
| "epoch": 0.6166769104691662, |
| "grad_norm": 0.30652934312820435, |
| "learning_rate": 2.9792962900736744e-05, |
| "loss": 0.0166, |
| "step": 3010 |
| }, |
| { |
| "epoch": 0.6187256709690637, |
| "grad_norm": 0.29974451661109924, |
| "learning_rate": 2.9790216213290372e-05, |
| "loss": 0.0145, |
| "step": 3020 |
| }, |
| { |
| "epoch": 0.6207744314689613, |
| "grad_norm": 0.23271848261356354, |
| "learning_rate": 2.9787451554336672e-05, |
| "loss": 0.014, |
| "step": 3030 |
| }, |
| { |
| "epoch": 0.6228231919688588, |
| "grad_norm": 0.16821375489234924, |
| "learning_rate": 2.9784668927234972e-05, |
| "loss": 0.0148, |
| "step": 3040 |
| }, |
| { |
| "epoch": 0.6248719524687564, |
| "grad_norm": 0.20723536610603333, |
| "learning_rate": 2.9781868335366418e-05, |
| "loss": 0.0182, |
| "step": 3050 |
| }, |
| { |
| "epoch": 0.6269207129686539, |
| "grad_norm": 0.2762635052204132, |
| "learning_rate": 2.9779049782134e-05, |
| "loss": 0.0172, |
| "step": 3060 |
| }, |
| { |
| "epoch": 0.6289694734685515, |
| "grad_norm": 0.2032575011253357, |
| "learning_rate": 2.9776213270962523e-05, |
| "loss": 0.0127, |
| "step": 3070 |
| }, |
| { |
| "epoch": 0.6310182339684491, |
| "grad_norm": 0.3603469431400299, |
| "learning_rate": 2.977335880529861e-05, |
| "loss": 0.0143, |
| "step": 3080 |
| }, |
| { |
| "epoch": 0.6330669944683467, |
| "grad_norm": 0.1752977967262268, |
| "learning_rate": 2.977048638861072e-05, |
| "loss": 0.0149, |
| "step": 3090 |
| }, |
| { |
| "epoch": 0.6351157549682442, |
| "grad_norm": 0.2731114327907562, |
| "learning_rate": 2.97675960243891e-05, |
| "loss": 0.0125, |
| "step": 3100 |
| }, |
| { |
| "epoch": 0.6371645154681418, |
| "grad_norm": 0.27266234159469604, |
| "learning_rate": 2.976468771614582e-05, |
| "loss": 0.0156, |
| "step": 3110 |
| }, |
| { |
| "epoch": 0.6392132759680393, |
| "grad_norm": 0.3300945460796356, |
| "learning_rate": 2.9761761467414736e-05, |
| "loss": 0.0177, |
| "step": 3120 |
| }, |
| { |
| "epoch": 0.6412620364679369, |
| "grad_norm": 0.3888791501522064, |
| "learning_rate": 2.9758817281751538e-05, |
| "loss": 0.0132, |
| "step": 3130 |
| }, |
| { |
| "epoch": 0.6433107969678344, |
| "grad_norm": 0.399882972240448, |
| "learning_rate": 2.975585516273367e-05, |
| "loss": 0.0175, |
| "step": 3140 |
| }, |
| { |
| "epoch": 0.645359557467732, |
| "grad_norm": 0.24711845815181732, |
| "learning_rate": 2.975287511396039e-05, |
| "loss": 0.0145, |
| "step": 3150 |
| }, |
| { |
| "epoch": 0.6474083179676295, |
| "grad_norm": 0.23811687529087067, |
| "learning_rate": 2.9749877139052745e-05, |
| "loss": 0.0159, |
| "step": 3160 |
| }, |
| { |
| "epoch": 0.6494570784675271, |
| "grad_norm": 0.33350464701652527, |
| "learning_rate": 2.9746861241653555e-05, |
| "loss": 0.0133, |
| "step": 3170 |
| }, |
| { |
| "epoch": 0.6515058389674248, |
| "grad_norm": 0.16725978255271912, |
| "learning_rate": 2.9743827425427416e-05, |
| "loss": 0.0128, |
| "step": 3180 |
| }, |
| { |
| "epoch": 0.6535545994673223, |
| "grad_norm": 0.22946153581142426, |
| "learning_rate": 2.9740775694060704e-05, |
| "loss": 0.0141, |
| "step": 3190 |
| }, |
| { |
| "epoch": 0.6556033599672199, |
| "grad_norm": 0.218031644821167, |
| "learning_rate": 2.973770605126156e-05, |
| "loss": 0.0138, |
| "step": 3200 |
| }, |
| { |
| "epoch": 0.6576521204671174, |
| "grad_norm": 0.19558770954608917, |
| "learning_rate": 2.9734618500759886e-05, |
| "loss": 0.0137, |
| "step": 3210 |
| }, |
| { |
| "epoch": 0.659700880967015, |
| "grad_norm": 0.235260009765625, |
| "learning_rate": 2.9731513046307356e-05, |
| "loss": 0.0137, |
| "step": 3220 |
| }, |
| { |
| "epoch": 0.6617496414669125, |
| "grad_norm": 0.5785341858863831, |
| "learning_rate": 2.972838969167738e-05, |
| "loss": 0.0117, |
| "step": 3230 |
| }, |
| { |
| "epoch": 0.6637984019668101, |
| "grad_norm": 0.27616119384765625, |
| "learning_rate": 2.972524844066513e-05, |
| "loss": 0.0149, |
| "step": 3240 |
| }, |
| { |
| "epoch": 0.6658471624667076, |
| "grad_norm": 0.25231319665908813, |
| "learning_rate": 2.9722089297087524e-05, |
| "loss": 0.0148, |
| "step": 3250 |
| }, |
| { |
| "epoch": 0.6678959229666052, |
| "grad_norm": 0.18168725073337555, |
| "learning_rate": 2.971891226478322e-05, |
| "loss": 0.0133, |
| "step": 3260 |
| }, |
| { |
| "epoch": 0.6699446834665027, |
| "grad_norm": 0.1884433478116989, |
| "learning_rate": 2.9715717347612615e-05, |
| "loss": 0.0142, |
| "step": 3270 |
| }, |
| { |
| "epoch": 0.6719934439664004, |
| "grad_norm": 0.3261781334877014, |
| "learning_rate": 2.971250454945783e-05, |
| "loss": 0.012, |
| "step": 3280 |
| }, |
| { |
| "epoch": 0.6740422044662979, |
| "grad_norm": 0.27586328983306885, |
| "learning_rate": 2.970927387422272e-05, |
| "loss": 0.0156, |
| "step": 3290 |
| }, |
| { |
| "epoch": 0.6760909649661955, |
| "grad_norm": 0.4391041398048401, |
| "learning_rate": 2.9706025325832858e-05, |
| "loss": 0.0157, |
| "step": 3300 |
| }, |
| { |
| "epoch": 0.678139725466093, |
| "grad_norm": 0.292903870344162, |
| "learning_rate": 2.9702758908235534e-05, |
| "loss": 0.0158, |
| "step": 3310 |
| }, |
| { |
| "epoch": 0.6801884859659906, |
| "grad_norm": 0.2626023292541504, |
| "learning_rate": 2.9699474625399765e-05, |
| "loss": 0.015, |
| "step": 3320 |
| }, |
| { |
| "epoch": 0.6822372464658881, |
| "grad_norm": 0.2589881122112274, |
| "learning_rate": 2.9696172481316246e-05, |
| "loss": 0.0136, |
| "step": 3330 |
| }, |
| { |
| "epoch": 0.6842860069657857, |
| "grad_norm": 0.25823044776916504, |
| "learning_rate": 2.9692852479997407e-05, |
| "loss": 0.0166, |
| "step": 3340 |
| }, |
| { |
| "epoch": 0.6863347674656832, |
| "grad_norm": 0.351821631193161, |
| "learning_rate": 2.9689514625477364e-05, |
| "loss": 0.016, |
| "step": 3350 |
| }, |
| { |
| "epoch": 0.6883835279655808, |
| "grad_norm": 0.2688700258731842, |
| "learning_rate": 2.968615892181192e-05, |
| "loss": 0.0123, |
| "step": 3360 |
| }, |
| { |
| "epoch": 0.6904322884654784, |
| "grad_norm": 0.579951286315918, |
| "learning_rate": 2.9682785373078567e-05, |
| "loss": 0.0131, |
| "step": 3370 |
| }, |
| { |
| "epoch": 0.692481048965376, |
| "grad_norm": 0.43358010053634644, |
| "learning_rate": 2.9679393983376493e-05, |
| "loss": 0.0133, |
| "step": 3380 |
| }, |
| { |
| "epoch": 0.6945298094652735, |
| "grad_norm": 0.16612811386585236, |
| "learning_rate": 2.9675984756826545e-05, |
| "loss": 0.0116, |
| "step": 3390 |
| }, |
| { |
| "epoch": 0.6965785699651711, |
| "grad_norm": 0.20509788393974304, |
| "learning_rate": 2.9672557697571273e-05, |
| "loss": 0.0152, |
| "step": 3400 |
| }, |
| { |
| "epoch": 0.6986273304650686, |
| "grad_norm": 0.1500205248594284, |
| "learning_rate": 2.966911280977486e-05, |
| "loss": 0.0135, |
| "step": 3410 |
| }, |
| { |
| "epoch": 0.7006760909649662, |
| "grad_norm": 0.18082526326179504, |
| "learning_rate": 2.966565009762318e-05, |
| "loss": 0.0146, |
| "step": 3420 |
| }, |
| { |
| "epoch": 0.7027248514648637, |
| "grad_norm": 0.25363779067993164, |
| "learning_rate": 2.966216956532375e-05, |
| "loss": 0.0151, |
| "step": 3430 |
| }, |
| { |
| "epoch": 0.7047736119647613, |
| "grad_norm": 0.18616405129432678, |
| "learning_rate": 2.9658671217105744e-05, |
| "loss": 0.0116, |
| "step": 3440 |
| }, |
| { |
| "epoch": 0.7068223724646588, |
| "grad_norm": 0.1986178755760193, |
| "learning_rate": 2.965515505721999e-05, |
| "loss": 0.0136, |
| "step": 3450 |
| }, |
| { |
| "epoch": 0.7088711329645564, |
| "grad_norm": 0.3392762839794159, |
| "learning_rate": 2.9651621089938948e-05, |
| "loss": 0.0118, |
| "step": 3460 |
| }, |
| { |
| "epoch": 0.7109198934644541, |
| "grad_norm": 0.1674731820821762, |
| "learning_rate": 2.9648069319556725e-05, |
| "loss": 0.0126, |
| "step": 3470 |
| }, |
| { |
| "epoch": 0.7129686539643516, |
| "grad_norm": 0.36203598976135254, |
| "learning_rate": 2.9644499750389057e-05, |
| "loss": 0.0168, |
| "step": 3480 |
| }, |
| { |
| "epoch": 0.7150174144642492, |
| "grad_norm": 0.4630207121372223, |
| "learning_rate": 2.9640912386773305e-05, |
| "loss": 0.013, |
| "step": 3490 |
| }, |
| { |
| "epoch": 0.7170661749641467, |
| "grad_norm": 0.4189489483833313, |
| "learning_rate": 2.9637307233068455e-05, |
| "loss": 0.0146, |
| "step": 3500 |
| }, |
| { |
| "epoch": 0.7191149354640443, |
| "grad_norm": 0.2173103541135788, |
| "learning_rate": 2.9633684293655108e-05, |
| "loss": 0.0194, |
| "step": 3510 |
| }, |
| { |
| "epoch": 0.7211636959639418, |
| "grad_norm": 0.3410756289958954, |
| "learning_rate": 2.963004357293548e-05, |
| "loss": 0.0132, |
| "step": 3520 |
| }, |
| { |
| "epoch": 0.7232124564638394, |
| "grad_norm": 0.24635352194309235, |
| "learning_rate": 2.962638507533339e-05, |
| "loss": 0.0141, |
| "step": 3530 |
| }, |
| { |
| "epoch": 0.7252612169637369, |
| "grad_norm": 0.3089825510978699, |
| "learning_rate": 2.962270880529425e-05, |
| "loss": 0.0159, |
| "step": 3540 |
| }, |
| { |
| "epoch": 0.7273099774636345, |
| "grad_norm": 0.2178957611322403, |
| "learning_rate": 2.9619014767285084e-05, |
| "loss": 0.0133, |
| "step": 3550 |
| }, |
| { |
| "epoch": 0.729358737963532, |
| "grad_norm": 0.32776984572410583, |
| "learning_rate": 2.9615302965794488e-05, |
| "loss": 0.0151, |
| "step": 3560 |
| }, |
| { |
| "epoch": 0.7314074984634297, |
| "grad_norm": 0.2422756850719452, |
| "learning_rate": 2.961157340533266e-05, |
| "loss": 0.0123, |
| "step": 3570 |
| }, |
| { |
| "epoch": 0.7334562589633272, |
| "grad_norm": 0.1641305536031723, |
| "learning_rate": 2.9607826090431364e-05, |
| "loss": 0.0131, |
| "step": 3580 |
| }, |
| { |
| "epoch": 0.7355050194632248, |
| "grad_norm": 0.16371628642082214, |
| "learning_rate": 2.9604061025643936e-05, |
| "loss": 0.0137, |
| "step": 3590 |
| }, |
| { |
| "epoch": 0.7375537799631223, |
| "grad_norm": 0.23734837770462036, |
| "learning_rate": 2.9600278215545292e-05, |
| "loss": 0.0151, |
| "step": 3600 |
| }, |
| { |
| "epoch": 0.7396025404630199, |
| "grad_norm": 0.17749202251434326, |
| "learning_rate": 2.95964776647319e-05, |
| "loss": 0.0114, |
| "step": 3610 |
| }, |
| { |
| "epoch": 0.7416513009629174, |
| "grad_norm": 0.3436014652252197, |
| "learning_rate": 2.9592659377821787e-05, |
| "loss": 0.0126, |
| "step": 3620 |
| }, |
| { |
| "epoch": 0.743700061462815, |
| "grad_norm": 0.15818221867084503, |
| "learning_rate": 2.9588823359454528e-05, |
| "loss": 0.0154, |
| "step": 3630 |
| }, |
| { |
| "epoch": 0.7457488219627125, |
| "grad_norm": 0.18272311985492706, |
| "learning_rate": 2.958496961429126e-05, |
| "loss": 0.0109, |
| "step": 3640 |
| }, |
| { |
| "epoch": 0.7477975824626101, |
| "grad_norm": 0.15854769945144653, |
| "learning_rate": 2.9581098147014628e-05, |
| "loss": 0.0136, |
| "step": 3650 |
| }, |
| { |
| "epoch": 0.7498463429625076, |
| "grad_norm": 0.17243251204490662, |
| "learning_rate": 2.9577208962328843e-05, |
| "loss": 0.0142, |
| "step": 3660 |
| }, |
| { |
| "epoch": 0.7518951034624053, |
| "grad_norm": 0.8100937008857727, |
| "learning_rate": 2.9573302064959633e-05, |
| "loss": 0.0182, |
| "step": 3670 |
| }, |
| { |
| "epoch": 0.7539438639623028, |
| "grad_norm": 0.2888268232345581, |
| "learning_rate": 2.9569377459654235e-05, |
| "loss": 0.0133, |
| "step": 3680 |
| }, |
| { |
| "epoch": 0.7559926244622004, |
| "grad_norm": 0.15330319106578827, |
| "learning_rate": 2.9565435151181424e-05, |
| "loss": 0.0167, |
| "step": 3690 |
| }, |
| { |
| "epoch": 0.7580413849620979, |
| "grad_norm": 0.3132888376712799, |
| "learning_rate": 2.9561475144331473e-05, |
| "loss": 0.0127, |
| "step": 3700 |
| }, |
| { |
| "epoch": 0.7600901454619955, |
| "grad_norm": 0.6971338391304016, |
| "learning_rate": 2.9557497443916162e-05, |
| "loss": 0.0154, |
| "step": 3710 |
| }, |
| { |
| "epoch": 0.762138905961893, |
| "grad_norm": 0.45506832003593445, |
| "learning_rate": 2.9553502054768774e-05, |
| "loss": 0.0145, |
| "step": 3720 |
| }, |
| { |
| "epoch": 0.7641876664617906, |
| "grad_norm": 0.23175564408302307, |
| "learning_rate": 2.954948898174409e-05, |
| "loss": 0.0158, |
| "step": 3730 |
| }, |
| { |
| "epoch": 0.7662364269616881, |
| "grad_norm": 0.17633609473705292, |
| "learning_rate": 2.954545822971836e-05, |
| "loss": 0.0136, |
| "step": 3740 |
| }, |
| { |
| "epoch": 0.7682851874615857, |
| "grad_norm": 0.22181996703147888, |
| "learning_rate": 2.9541409803589335e-05, |
| "loss": 0.0123, |
| "step": 3750 |
| }, |
| { |
| "epoch": 0.7703339479614834, |
| "grad_norm": 0.23021996021270752, |
| "learning_rate": 2.9537343708276235e-05, |
| "loss": 0.0175, |
| "step": 3760 |
| }, |
| { |
| "epoch": 0.7723827084613809, |
| "grad_norm": 0.3349889814853668, |
| "learning_rate": 2.953325994871975e-05, |
| "loss": 0.013, |
| "step": 3770 |
| }, |
| { |
| "epoch": 0.7744314689612785, |
| "grad_norm": 0.2657260298728943, |
| "learning_rate": 2.9529158529882023e-05, |
| "loss": 0.0108, |
| "step": 3780 |
| }, |
| { |
| "epoch": 0.776480229461176, |
| "grad_norm": 0.2826705276966095, |
| "learning_rate": 2.9525039456746677e-05, |
| "loss": 0.0129, |
| "step": 3790 |
| }, |
| { |
| "epoch": 0.7785289899610736, |
| "grad_norm": 0.18078115582466125, |
| "learning_rate": 2.952090273431877e-05, |
| "loss": 0.0121, |
| "step": 3800 |
| }, |
| { |
| "epoch": 0.7805777504609711, |
| "grad_norm": 0.20941874384880066, |
| "learning_rate": 2.9516748367624808e-05, |
| "loss": 0.0162, |
| "step": 3810 |
| }, |
| { |
| "epoch": 0.7826265109608687, |
| "grad_norm": 0.3546510934829712, |
| "learning_rate": 2.951257636171274e-05, |
| "loss": 0.0135, |
| "step": 3820 |
| }, |
| { |
| "epoch": 0.7846752714607662, |
| "grad_norm": 0.2289784550666809, |
| "learning_rate": 2.9508386721651944e-05, |
| "loss": 0.0141, |
| "step": 3830 |
| }, |
| { |
| "epoch": 0.7867240319606638, |
| "grad_norm": 0.4556593596935272, |
| "learning_rate": 2.9504179452533228e-05, |
| "loss": 0.0164, |
| "step": 3840 |
| }, |
| { |
| "epoch": 0.7887727924605613, |
| "grad_norm": 0.26459041237831116, |
| "learning_rate": 2.9499954559468826e-05, |
| "loss": 0.0143, |
| "step": 3850 |
| }, |
| { |
| "epoch": 0.790821552960459, |
| "grad_norm": 0.2827795147895813, |
| "learning_rate": 2.9495712047592378e-05, |
| "loss": 0.0131, |
| "step": 3860 |
| }, |
| { |
| "epoch": 0.7928703134603565, |
| "grad_norm": 0.2788614332675934, |
| "learning_rate": 2.949145192205893e-05, |
| "loss": 0.0133, |
| "step": 3870 |
| }, |
| { |
| "epoch": 0.7949190739602541, |
| "grad_norm": 0.1652802675962448, |
| "learning_rate": 2.9487174188044953e-05, |
| "loss": 0.0137, |
| "step": 3880 |
| }, |
| { |
| "epoch": 0.7969678344601516, |
| "grad_norm": 0.22082747519016266, |
| "learning_rate": 2.9482878850748276e-05, |
| "loss": 0.0144, |
| "step": 3890 |
| }, |
| { |
| "epoch": 0.7990165949600492, |
| "grad_norm": 0.19942745566368103, |
| "learning_rate": 2.9478565915388157e-05, |
| "loss": 0.0141, |
| "step": 3900 |
| }, |
| { |
| "epoch": 0.8010653554599467, |
| "grad_norm": 0.211129829287529, |
| "learning_rate": 2.9474235387205207e-05, |
| "loss": 0.0141, |
| "step": 3910 |
| }, |
| { |
| "epoch": 0.8031141159598443, |
| "grad_norm": 0.4587189257144928, |
| "learning_rate": 2.946988727146143e-05, |
| "loss": 0.0129, |
| "step": 3920 |
| }, |
| { |
| "epoch": 0.8051628764597418, |
| "grad_norm": 0.39450445771217346, |
| "learning_rate": 2.9465521573440196e-05, |
| "loss": 0.0129, |
| "step": 3930 |
| }, |
| { |
| "epoch": 0.8072116369596394, |
| "grad_norm": 0.2511252164840698, |
| "learning_rate": 2.946113829844624e-05, |
| "loss": 0.0151, |
| "step": 3940 |
| }, |
| { |
| "epoch": 0.8092603974595369, |
| "grad_norm": 0.2121558040380478, |
| "learning_rate": 2.945673745180566e-05, |
| "loss": 0.013, |
| "step": 3950 |
| }, |
| { |
| "epoch": 0.8113091579594346, |
| "grad_norm": 0.15654531121253967, |
| "learning_rate": 2.945231903886589e-05, |
| "loss": 0.0137, |
| "step": 3960 |
| }, |
| { |
| "epoch": 0.8133579184593321, |
| "grad_norm": 0.12897659838199615, |
| "learning_rate": 2.9447883064995728e-05, |
| "loss": 0.0147, |
| "step": 3970 |
| }, |
| { |
| "epoch": 0.8154066789592297, |
| "grad_norm": 0.2801649570465088, |
| "learning_rate": 2.9443429535585288e-05, |
| "loss": 0.0142, |
| "step": 3980 |
| }, |
| { |
| "epoch": 0.8174554394591272, |
| "grad_norm": 0.20357689261436462, |
| "learning_rate": 2.9438958456046044e-05, |
| "loss": 0.0118, |
| "step": 3990 |
| }, |
| { |
| "epoch": 0.8195041999590248, |
| "grad_norm": 0.2525339424610138, |
| "learning_rate": 2.9434469831810765e-05, |
| "loss": 0.0141, |
| "step": 4000 |
| }, |
| { |
| "epoch": 0.8215529604589223, |
| "grad_norm": 0.37886765599250793, |
| "learning_rate": 2.9429963668333566e-05, |
| "loss": 0.0137, |
| "step": 4010 |
| }, |
| { |
| "epoch": 0.8236017209588199, |
| "grad_norm": 0.29004308581352234, |
| "learning_rate": 2.9425439971089845e-05, |
| "loss": 0.0144, |
| "step": 4020 |
| }, |
| { |
| "epoch": 0.8256504814587174, |
| "grad_norm": 0.22543436288833618, |
| "learning_rate": 2.9420898745576336e-05, |
| "loss": 0.0115, |
| "step": 4030 |
| }, |
| { |
| "epoch": 0.827699241958615, |
| "grad_norm": 0.23177289962768555, |
| "learning_rate": 2.9416339997311044e-05, |
| "loss": 0.0139, |
| "step": 4040 |
| }, |
| { |
| "epoch": 0.8297480024585125, |
| "grad_norm": 0.36060988903045654, |
| "learning_rate": 2.941176373183329e-05, |
| "loss": 0.0164, |
| "step": 4050 |
| }, |
| { |
| "epoch": 0.8317967629584102, |
| "grad_norm": 0.24601984024047852, |
| "learning_rate": 2.9407169954703653e-05, |
| "loss": 0.0133, |
| "step": 4060 |
| }, |
| { |
| "epoch": 0.8338455234583078, |
| "grad_norm": 0.20672382414340973, |
| "learning_rate": 2.9402558671504018e-05, |
| "loss": 0.013, |
| "step": 4070 |
| }, |
| { |
| "epoch": 0.8358942839582053, |
| "grad_norm": 0.1864919811487198, |
| "learning_rate": 2.9397929887837524e-05, |
| "loss": 0.0135, |
| "step": 4080 |
| }, |
| { |
| "epoch": 0.8379430444581029, |
| "grad_norm": 0.3264613449573517, |
| "learning_rate": 2.939328360932858e-05, |
| "loss": 0.0138, |
| "step": 4090 |
| }, |
| { |
| "epoch": 0.8399918049580004, |
| "grad_norm": 0.39139190316200256, |
| "learning_rate": 2.938861984162285e-05, |
| "loss": 0.0123, |
| "step": 4100 |
| }, |
| { |
| "epoch": 0.842040565457898, |
| "grad_norm": 0.3254573941230774, |
| "learning_rate": 2.938393859038726e-05, |
| "loss": 0.0127, |
| "step": 4110 |
| }, |
| { |
| "epoch": 0.8440893259577955, |
| "grad_norm": 0.572283148765564, |
| "learning_rate": 2.9379239861309964e-05, |
| "loss": 0.0147, |
| "step": 4120 |
| }, |
| { |
| "epoch": 0.8461380864576931, |
| "grad_norm": 0.3349283039569855, |
| "learning_rate": 2.937452366010036e-05, |
| "loss": 0.0129, |
| "step": 4130 |
| }, |
| { |
| "epoch": 0.8481868469575906, |
| "grad_norm": 0.22565816342830658, |
| "learning_rate": 2.936978999248908e-05, |
| "loss": 0.012, |
| "step": 4140 |
| }, |
| { |
| "epoch": 0.8502356074574883, |
| "grad_norm": 0.33050665259361267, |
| "learning_rate": 2.936503886422798e-05, |
| "loss": 0.0125, |
| "step": 4150 |
| }, |
| { |
| "epoch": 0.8522843679573858, |
| "grad_norm": 0.17670375108718872, |
| "learning_rate": 2.9360270281090122e-05, |
| "loss": 0.0124, |
| "step": 4160 |
| }, |
| { |
| "epoch": 0.8543331284572834, |
| "grad_norm": 0.31105342507362366, |
| "learning_rate": 2.935548424886979e-05, |
| "loss": 0.0147, |
| "step": 4170 |
| }, |
| { |
| "epoch": 0.8563818889571809, |
| "grad_norm": 0.7212890386581421, |
| "learning_rate": 2.935068077338246e-05, |
| "loss": 0.0164, |
| "step": 4180 |
| }, |
| { |
| "epoch": 0.8584306494570785, |
| "grad_norm": 0.19543343782424927, |
| "learning_rate": 2.934585986046481e-05, |
| "loss": 0.0127, |
| "step": 4190 |
| }, |
| { |
| "epoch": 0.860479409956976, |
| "grad_norm": 0.177902951836586, |
| "learning_rate": 2.93410215159747e-05, |
| "loss": 0.0121, |
| "step": 4200 |
| }, |
| { |
| "epoch": 0.8625281704568736, |
| "grad_norm": 0.3001309931278229, |
| "learning_rate": 2.933616574579118e-05, |
| "loss": 0.0126, |
| "step": 4210 |
| }, |
| { |
| "epoch": 0.8645769309567711, |
| "grad_norm": 0.2270403355360031, |
| "learning_rate": 2.933129255581447e-05, |
| "loss": 0.0142, |
| "step": 4220 |
| }, |
| { |
| "epoch": 0.8666256914566687, |
| "grad_norm": 0.8695701956748962, |
| "learning_rate": 2.9326401951965954e-05, |
| "loss": 0.0137, |
| "step": 4230 |
| }, |
| { |
| "epoch": 0.8686744519565662, |
| "grad_norm": 0.3714264929294586, |
| "learning_rate": 2.9321493940188165e-05, |
| "loss": 0.016, |
| "step": 4240 |
| }, |
| { |
| "epoch": 0.8707232124564639, |
| "grad_norm": 0.273501455783844, |
| "learning_rate": 2.931656852644482e-05, |
| "loss": 0.0168, |
| "step": 4250 |
| }, |
| { |
| "epoch": 0.8727719729563614, |
| "grad_norm": 0.25349855422973633, |
| "learning_rate": 2.9311625716720748e-05, |
| "loss": 0.0156, |
| "step": 4260 |
| }, |
| { |
| "epoch": 0.874820733456259, |
| "grad_norm": 0.23911529779434204, |
| "learning_rate": 2.9306665517021927e-05, |
| "loss": 0.0148, |
| "step": 4270 |
| }, |
| { |
| "epoch": 0.8768694939561565, |
| "grad_norm": 0.25113216042518616, |
| "learning_rate": 2.9301687933375483e-05, |
| "loss": 0.0151, |
| "step": 4280 |
| }, |
| { |
| "epoch": 0.8789182544560541, |
| "grad_norm": 0.22121913731098175, |
| "learning_rate": 2.929669297182963e-05, |
| "loss": 0.012, |
| "step": 4290 |
| }, |
| { |
| "epoch": 0.8809670149559516, |
| "grad_norm": 0.44621706008911133, |
| "learning_rate": 2.929168063845373e-05, |
| "loss": 0.0135, |
| "step": 4300 |
| }, |
| { |
| "epoch": 0.8830157754558492, |
| "grad_norm": 0.32186633348464966, |
| "learning_rate": 2.9286650939338236e-05, |
| "loss": 0.0109, |
| "step": 4310 |
| }, |
| { |
| "epoch": 0.8850645359557467, |
| "grad_norm": 0.29777365922927856, |
| "learning_rate": 2.9281603880594706e-05, |
| "loss": 0.0115, |
| "step": 4320 |
| }, |
| { |
| "epoch": 0.8871132964556443, |
| "grad_norm": 0.4305902421474457, |
| "learning_rate": 2.9276539468355796e-05, |
| "loss": 0.0139, |
| "step": 4330 |
| }, |
| { |
| "epoch": 0.8891620569555418, |
| "grad_norm": 0.18245649337768555, |
| "learning_rate": 2.9271457708775244e-05, |
| "loss": 0.012, |
| "step": 4340 |
| }, |
| { |
| "epoch": 0.8912108174554395, |
| "grad_norm": 0.21634075045585632, |
| "learning_rate": 2.9266358608027863e-05, |
| "loss": 0.0143, |
| "step": 4350 |
| }, |
| { |
| "epoch": 0.8932595779553371, |
| "grad_norm": 0.183611661195755, |
| "learning_rate": 2.9261242172309543e-05, |
| "loss": 0.0131, |
| "step": 4360 |
| }, |
| { |
| "epoch": 0.8953083384552346, |
| "grad_norm": 0.23058868944644928, |
| "learning_rate": 2.925610840783723e-05, |
| "loss": 0.0138, |
| "step": 4370 |
| }, |
| { |
| "epoch": 0.8973570989551322, |
| "grad_norm": 0.27207133173942566, |
| "learning_rate": 2.9250957320848946e-05, |
| "loss": 0.0135, |
| "step": 4380 |
| }, |
| { |
| "epoch": 0.8994058594550297, |
| "grad_norm": 0.29274484515190125, |
| "learning_rate": 2.924578891760374e-05, |
| "loss": 0.0139, |
| "step": 4390 |
| }, |
| { |
| "epoch": 0.9014546199549273, |
| "grad_norm": 0.2229909896850586, |
| "learning_rate": 2.924060320438171e-05, |
| "loss": 0.0116, |
| "step": 4400 |
| }, |
| { |
| "epoch": 0.9035033804548248, |
| "grad_norm": 0.9632654190063477, |
| "learning_rate": 2.9235400187483978e-05, |
| "loss": 0.015, |
| "step": 4410 |
| }, |
| { |
| "epoch": 0.9055521409547224, |
| "grad_norm": 0.2625559866428375, |
| "learning_rate": 2.9230179873232716e-05, |
| "loss": 0.0152, |
| "step": 4420 |
| }, |
| { |
| "epoch": 0.9076009014546199, |
| "grad_norm": 0.26783913373947144, |
| "learning_rate": 2.922494226797108e-05, |
| "loss": 0.014, |
| "step": 4430 |
| }, |
| { |
| "epoch": 0.9096496619545175, |
| "grad_norm": 0.2952384948730469, |
| "learning_rate": 2.9219687378063275e-05, |
| "loss": 0.013, |
| "step": 4440 |
| }, |
| { |
| "epoch": 0.9116984224544151, |
| "grad_norm": 0.5369178056716919, |
| "learning_rate": 2.921441520989447e-05, |
| "loss": 0.0147, |
| "step": 4450 |
| }, |
| { |
| "epoch": 0.9137471829543127, |
| "grad_norm": 0.2858648896217346, |
| "learning_rate": 2.9209125769870854e-05, |
| "loss": 0.0121, |
| "step": 4460 |
| }, |
| { |
| "epoch": 0.9157959434542102, |
| "grad_norm": 0.13574151694774628, |
| "learning_rate": 2.92038190644196e-05, |
| "loss": 0.0121, |
| "step": 4470 |
| }, |
| { |
| "epoch": 0.9178447039541078, |
| "grad_norm": 0.22464297711849213, |
| "learning_rate": 2.9198495099988843e-05, |
| "loss": 0.0131, |
| "step": 4480 |
| }, |
| { |
| "epoch": 0.9198934644540053, |
| "grad_norm": 0.18517789244651794, |
| "learning_rate": 2.9193153883047712e-05, |
| "loss": 0.013, |
| "step": 4490 |
| }, |
| { |
| "epoch": 0.9219422249539029, |
| "grad_norm": 0.22209249436855316, |
| "learning_rate": 2.9187795420086287e-05, |
| "loss": 0.0111, |
| "step": 4500 |
| }, |
| { |
| "epoch": 0.9239909854538004, |
| "grad_norm": 0.13971543312072754, |
| "learning_rate": 2.9182419717615602e-05, |
| "loss": 0.0125, |
| "step": 4510 |
| }, |
| { |
| "epoch": 0.926039745953698, |
| "grad_norm": 0.18937425315380096, |
| "learning_rate": 2.917702678216765e-05, |
| "loss": 0.0119, |
| "step": 4520 |
| }, |
| { |
| "epoch": 0.9280885064535955, |
| "grad_norm": 0.5760051608085632, |
| "learning_rate": 2.9171616620295347e-05, |
| "loss": 0.0165, |
| "step": 4530 |
| }, |
| { |
| "epoch": 0.9301372669534932, |
| "grad_norm": 0.44574669003486633, |
| "learning_rate": 2.9166189238572554e-05, |
| "loss": 0.0146, |
| "step": 4540 |
| }, |
| { |
| "epoch": 0.9321860274533907, |
| "grad_norm": 0.213381826877594, |
| "learning_rate": 2.9160744643594052e-05, |
| "loss": 0.0124, |
| "step": 4550 |
| }, |
| { |
| "epoch": 0.9342347879532883, |
| "grad_norm": 0.17810748517513275, |
| "learning_rate": 2.915528284197554e-05, |
| "loss": 0.0143, |
| "step": 4560 |
| }, |
| { |
| "epoch": 0.9362835484531858, |
| "grad_norm": 0.367615282535553, |
| "learning_rate": 2.914980384035362e-05, |
| "loss": 0.0148, |
| "step": 4570 |
| }, |
| { |
| "epoch": 0.9383323089530834, |
| "grad_norm": 0.20603232085704803, |
| "learning_rate": 2.9144307645385787e-05, |
| "loss": 0.0125, |
| "step": 4580 |
| }, |
| { |
| "epoch": 0.9403810694529809, |
| "grad_norm": 0.2509411871433258, |
| "learning_rate": 2.9138794263750454e-05, |
| "loss": 0.0141, |
| "step": 4590 |
| }, |
| { |
| "epoch": 0.9424298299528785, |
| "grad_norm": 0.1960693895816803, |
| "learning_rate": 2.913326370214688e-05, |
| "loss": 0.0157, |
| "step": 4600 |
| }, |
| { |
| "epoch": 0.944478590452776, |
| "grad_norm": 0.8284222483634949, |
| "learning_rate": 2.9127715967295234e-05, |
| "loss": 0.0137, |
| "step": 4610 |
| }, |
| { |
| "epoch": 0.9465273509526736, |
| "grad_norm": 0.3766833543777466, |
| "learning_rate": 2.912215106593653e-05, |
| "loss": 0.0152, |
| "step": 4620 |
| }, |
| { |
| "epoch": 0.9485761114525711, |
| "grad_norm": 0.2750181555747986, |
| "learning_rate": 2.911656900483265e-05, |
| "loss": 0.0149, |
| "step": 4630 |
| }, |
| { |
| "epoch": 0.9506248719524688, |
| "grad_norm": 0.18956445157527924, |
| "learning_rate": 2.9110969790766322e-05, |
| "loss": 0.0131, |
| "step": 4640 |
| }, |
| { |
| "epoch": 0.9526736324523664, |
| "grad_norm": 0.12459101527929306, |
| "learning_rate": 2.910535343054112e-05, |
| "loss": 0.0114, |
| "step": 4650 |
| }, |
| { |
| "epoch": 0.9547223929522639, |
| "grad_norm": 0.2587648630142212, |
| "learning_rate": 2.909971993098145e-05, |
| "loss": 0.0124, |
| "step": 4660 |
| }, |
| { |
| "epoch": 0.9567711534521615, |
| "grad_norm": 0.14933425188064575, |
| "learning_rate": 2.9094069298932552e-05, |
| "loss": 0.0148, |
| "step": 4670 |
| }, |
| { |
| "epoch": 0.958819913952059, |
| "grad_norm": 0.2596816420555115, |
| "learning_rate": 2.908840154126047e-05, |
| "loss": 0.0137, |
| "step": 4680 |
| }, |
| { |
| "epoch": 0.9608686744519566, |
| "grad_norm": 0.6549279093742371, |
| "learning_rate": 2.9082716664852063e-05, |
| "loss": 0.0147, |
| "step": 4690 |
| }, |
| { |
| "epoch": 0.9629174349518541, |
| "grad_norm": 0.415600448846817, |
| "learning_rate": 2.9077014676615e-05, |
| "loss": 0.0149, |
| "step": 4700 |
| }, |
| { |
| "epoch": 0.9649661954517517, |
| "grad_norm": 0.2901279926300049, |
| "learning_rate": 2.9071295583477724e-05, |
| "loss": 0.0148, |
| "step": 4710 |
| }, |
| { |
| "epoch": 0.9670149559516492, |
| "grad_norm": 0.23200581967830658, |
| "learning_rate": 2.906555939238948e-05, |
| "loss": 0.0123, |
| "step": 4720 |
| }, |
| { |
| "epoch": 0.9690637164515468, |
| "grad_norm": 0.40415504574775696, |
| "learning_rate": 2.905980611032028e-05, |
| "loss": 0.0135, |
| "step": 4730 |
| }, |
| { |
| "epoch": 0.9711124769514444, |
| "grad_norm": 0.19962766766548157, |
| "learning_rate": 2.9054035744260904e-05, |
| "loss": 0.0133, |
| "step": 4740 |
| }, |
| { |
| "epoch": 0.973161237451342, |
| "grad_norm": 0.5521057844161987, |
| "learning_rate": 2.9048248301222885e-05, |
| "loss": 0.0137, |
| "step": 4750 |
| }, |
| { |
| "epoch": 0.9752099979512395, |
| "grad_norm": 0.1437607854604721, |
| "learning_rate": 2.904244378823852e-05, |
| "loss": 0.0118, |
| "step": 4760 |
| }, |
| { |
| "epoch": 0.9772587584511371, |
| "grad_norm": 0.23802931606769562, |
| "learning_rate": 2.9036622212360836e-05, |
| "loss": 0.0126, |
| "step": 4770 |
| }, |
| { |
| "epoch": 0.9793075189510346, |
| "grad_norm": 0.30804872512817383, |
| "learning_rate": 2.9030783580663596e-05, |
| "loss": 0.013, |
| "step": 4780 |
| }, |
| { |
| "epoch": 0.9813562794509322, |
| "grad_norm": 0.2821427881717682, |
| "learning_rate": 2.902492790024129e-05, |
| "loss": 0.0127, |
| "step": 4790 |
| }, |
| { |
| "epoch": 0.9834050399508297, |
| "grad_norm": 0.28503286838531494, |
| "learning_rate": 2.901905517820912e-05, |
| "loss": 0.015, |
| "step": 4800 |
| }, |
| { |
| "epoch": 0.9854538004507273, |
| "grad_norm": 0.3891165554523468, |
| "learning_rate": 2.9013165421703e-05, |
| "loss": 0.0109, |
| "step": 4810 |
| }, |
| { |
| "epoch": 0.9875025609506248, |
| "grad_norm": 0.3078186810016632, |
| "learning_rate": 2.9007258637879534e-05, |
| "loss": 0.0113, |
| "step": 4820 |
| }, |
| { |
| "epoch": 0.9895513214505224, |
| "grad_norm": 0.34799936413764954, |
| "learning_rate": 2.9001334833916025e-05, |
| "loss": 0.014, |
| "step": 4830 |
| }, |
| { |
| "epoch": 0.99160008195042, |
| "grad_norm": 0.45366302132606506, |
| "learning_rate": 2.899539401701046e-05, |
| "loss": 0.0133, |
| "step": 4840 |
| }, |
| { |
| "epoch": 0.9936488424503176, |
| "grad_norm": 0.2579045295715332, |
| "learning_rate": 2.8989436194381474e-05, |
| "loss": 0.0127, |
| "step": 4850 |
| }, |
| { |
| "epoch": 0.9956976029502151, |
| "grad_norm": 0.15534165501594543, |
| "learning_rate": 2.8983461373268404e-05, |
| "loss": 0.0147, |
| "step": 4860 |
| }, |
| { |
| "epoch": 0.9977463634501127, |
| "grad_norm": 0.37494421005249023, |
| "learning_rate": 2.8977469560931205e-05, |
| "loss": 0.0111, |
| "step": 4870 |
| }, |
| { |
| "epoch": 0.9997951239500102, |
| "grad_norm": 0.35213375091552734, |
| "learning_rate": 2.8971460764650502e-05, |
| "loss": 0.0157, |
| "step": 4880 |
| }, |
| { |
| "epoch": 1.0018438844499078, |
| "grad_norm": 0.2274678349494934, |
| "learning_rate": 2.8965434991727545e-05, |
| "loss": 0.0143, |
| "step": 4890 |
| }, |
| { |
| "epoch": 1.0038926449498053, |
| "grad_norm": 0.19185985624790192, |
| "learning_rate": 2.895939224948422e-05, |
| "loss": 0.0157, |
| "step": 4900 |
| }, |
| { |
| "epoch": 1.005941405449703, |
| "grad_norm": 0.366849809885025, |
| "learning_rate": 2.895333254526303e-05, |
| "loss": 0.0159, |
| "step": 4910 |
| }, |
| { |
| "epoch": 1.0079901659496004, |
| "grad_norm": 0.2487134486436844, |
| "learning_rate": 2.8947255886427084e-05, |
| "loss": 0.0155, |
| "step": 4920 |
| }, |
| { |
| "epoch": 1.010038926449498, |
| "grad_norm": 0.3345854580402374, |
| "learning_rate": 2.8941162280360095e-05, |
| "loss": 0.0142, |
| "step": 4930 |
| }, |
| { |
| "epoch": 1.0120876869493955, |
| "grad_norm": 0.2865547239780426, |
| "learning_rate": 2.8935051734466368e-05, |
| "loss": 0.0138, |
| "step": 4940 |
| }, |
| { |
| "epoch": 1.014136447449293, |
| "grad_norm": 0.2105558216571808, |
| "learning_rate": 2.8928924256170792e-05, |
| "loss": 0.0124, |
| "step": 4950 |
| }, |
| { |
| "epoch": 1.0161852079491906, |
| "grad_norm": 0.26346173882484436, |
| "learning_rate": 2.8922779852918836e-05, |
| "loss": 0.0132, |
| "step": 4960 |
| }, |
| { |
| "epoch": 1.0182339684490882, |
| "grad_norm": 0.20724141597747803, |
| "learning_rate": 2.8916618532176526e-05, |
| "loss": 0.013, |
| "step": 4970 |
| }, |
| { |
| "epoch": 1.020282728948986, |
| "grad_norm": 0.15110521018505096, |
| "learning_rate": 2.891044030143045e-05, |
| "loss": 0.0135, |
| "step": 4980 |
| }, |
| { |
| "epoch": 1.0223314894488835, |
| "grad_norm": 0.09837942570447922, |
| "learning_rate": 2.890424516818774e-05, |
| "loss": 0.0117, |
| "step": 4990 |
| }, |
| { |
| "epoch": 1.024380249948781, |
| "grad_norm": 0.28735998272895813, |
| "learning_rate": 2.8898033139976057e-05, |
| "loss": 0.0162, |
| "step": 5000 |
| }, |
| { |
| "epoch": 1.0264290104486786, |
| "grad_norm": 0.235224649310112, |
| "learning_rate": 2.889180422434362e-05, |
| "loss": 0.015, |
| "step": 5010 |
| }, |
| { |
| "epoch": 1.0284777709485762, |
| "grad_norm": 0.2208499163389206, |
| "learning_rate": 2.8885558428859136e-05, |
| "loss": 0.012, |
| "step": 5020 |
| }, |
| { |
| "epoch": 1.0305265314484737, |
| "grad_norm": 0.12630470097064972, |
| "learning_rate": 2.8879295761111835e-05, |
| "loss": 0.0133, |
| "step": 5030 |
| }, |
| { |
| "epoch": 1.0325752919483713, |
| "grad_norm": 0.12112489342689514, |
| "learning_rate": 2.8873016228711452e-05, |
| "loss": 0.0142, |
| "step": 5040 |
| }, |
| { |
| "epoch": 1.0346240524482688, |
| "grad_norm": 0.22429129481315613, |
| "learning_rate": 2.8866719839288215e-05, |
| "loss": 0.0146, |
| "step": 5050 |
| }, |
| { |
| "epoch": 1.0366728129481664, |
| "grad_norm": 0.2565072774887085, |
| "learning_rate": 2.886040660049283e-05, |
| "loss": 0.014, |
| "step": 5060 |
| }, |
| { |
| "epoch": 1.038721573448064, |
| "grad_norm": 0.14318262040615082, |
| "learning_rate": 2.8854076519996467e-05, |
| "loss": 0.0139, |
| "step": 5070 |
| }, |
| { |
| "epoch": 1.0407703339479615, |
| "grad_norm": 0.21522922813892365, |
| "learning_rate": 2.884772960549078e-05, |
| "loss": 0.0135, |
| "step": 5080 |
| }, |
| { |
| "epoch": 1.042819094447859, |
| "grad_norm": 0.3484971225261688, |
| "learning_rate": 2.8841365864687867e-05, |
| "loss": 0.0168, |
| "step": 5090 |
| }, |
| { |
| "epoch": 1.0448678549477566, |
| "grad_norm": 0.16776637732982635, |
| "learning_rate": 2.8834985305320276e-05, |
| "loss": 0.014, |
| "step": 5100 |
| }, |
| { |
| "epoch": 1.0469166154476541, |
| "grad_norm": 0.28611090779304504, |
| "learning_rate": 2.882858793514098e-05, |
| "loss": 0.0154, |
| "step": 5110 |
| }, |
| { |
| "epoch": 1.0489653759475517, |
| "grad_norm": 0.20541328191757202, |
| "learning_rate": 2.8822173761923398e-05, |
| "loss": 0.0122, |
| "step": 5120 |
| }, |
| { |
| "epoch": 1.0510141364474492, |
| "grad_norm": 0.31141290068626404, |
| "learning_rate": 2.8815742793461343e-05, |
| "loss": 0.0162, |
| "step": 5130 |
| }, |
| { |
| "epoch": 1.0530628969473468, |
| "grad_norm": 0.2866436243057251, |
| "learning_rate": 2.880929503756906e-05, |
| "loss": 0.0138, |
| "step": 5140 |
| }, |
| { |
| "epoch": 1.0551116574472443, |
| "grad_norm": 0.4340724050998688, |
| "learning_rate": 2.880283050208117e-05, |
| "loss": 0.0135, |
| "step": 5150 |
| }, |
| { |
| "epoch": 1.0571604179471419, |
| "grad_norm": 0.32904115319252014, |
| "learning_rate": 2.8796349194852704e-05, |
| "loss": 0.0128, |
| "step": 5160 |
| }, |
| { |
| "epoch": 1.0592091784470394, |
| "grad_norm": 0.20563213527202606, |
| "learning_rate": 2.8789851123759052e-05, |
| "loss": 0.0128, |
| "step": 5170 |
| }, |
| { |
| "epoch": 1.0612579389469372, |
| "grad_norm": 0.11401065438985825, |
| "learning_rate": 2.8783336296695985e-05, |
| "loss": 0.0139, |
| "step": 5180 |
| }, |
| { |
| "epoch": 1.0633066994468348, |
| "grad_norm": 0.2730553448200226, |
| "learning_rate": 2.877680472157964e-05, |
| "loss": 0.0133, |
| "step": 5190 |
| }, |
| { |
| "epoch": 1.0653554599467323, |
| "grad_norm": 0.24899455904960632, |
| "learning_rate": 2.8770256406346483e-05, |
| "loss": 0.012, |
| "step": 5200 |
| }, |
| { |
| "epoch": 1.0674042204466299, |
| "grad_norm": 0.3067129850387573, |
| "learning_rate": 2.8763691358953348e-05, |
| "loss": 0.0115, |
| "step": 5210 |
| }, |
| { |
| "epoch": 1.0694529809465274, |
| "grad_norm": 0.3275994062423706, |
| "learning_rate": 2.8757109587377377e-05, |
| "loss": 0.0145, |
| "step": 5220 |
| }, |
| { |
| "epoch": 1.071501741446425, |
| "grad_norm": 0.1287369430065155, |
| "learning_rate": 2.8750511099616047e-05, |
| "loss": 0.0119, |
| "step": 5230 |
| }, |
| { |
| "epoch": 1.0735505019463225, |
| "grad_norm": 0.18907569348812103, |
| "learning_rate": 2.8743895903687143e-05, |
| "loss": 0.0139, |
| "step": 5240 |
| }, |
| { |
| "epoch": 1.07559926244622, |
| "grad_norm": 0.18511714041233063, |
| "learning_rate": 2.8737264007628753e-05, |
| "loss": 0.0139, |
| "step": 5250 |
| }, |
| { |
| "epoch": 1.0776480229461176, |
| "grad_norm": 0.26233038306236267, |
| "learning_rate": 2.8730615419499247e-05, |
| "loss": 0.0153, |
| "step": 5260 |
| }, |
| { |
| "epoch": 1.0796967834460152, |
| "grad_norm": 0.18695448338985443, |
| "learning_rate": 2.8723950147377303e-05, |
| "loss": 0.0129, |
| "step": 5270 |
| }, |
| { |
| "epoch": 1.0817455439459127, |
| "grad_norm": 0.18551333248615265, |
| "learning_rate": 2.871726819936184e-05, |
| "loss": 0.0157, |
| "step": 5280 |
| }, |
| { |
| "epoch": 1.0837943044458103, |
| "grad_norm": 0.815523087978363, |
| "learning_rate": 2.8710569583572068e-05, |
| "loss": 0.0146, |
| "step": 5290 |
| }, |
| { |
| "epoch": 1.0858430649457078, |
| "grad_norm": 0.4258750379085541, |
| "learning_rate": 2.870385430814742e-05, |
| "loss": 0.0128, |
| "step": 5300 |
| }, |
| { |
| "epoch": 1.0878918254456054, |
| "grad_norm": 0.23373650014400482, |
| "learning_rate": 2.8697122381247605e-05, |
| "loss": 0.0151, |
| "step": 5310 |
| }, |
| { |
| "epoch": 1.089940585945503, |
| "grad_norm": 0.44140979647636414, |
| "learning_rate": 2.8690373811052545e-05, |
| "loss": 0.0119, |
| "step": 5320 |
| }, |
| { |
| "epoch": 1.0919893464454005, |
| "grad_norm": 0.35255903005599976, |
| "learning_rate": 2.8683608605762378e-05, |
| "loss": 0.0157, |
| "step": 5330 |
| }, |
| { |
| "epoch": 1.094038106945298, |
| "grad_norm": 0.28442883491516113, |
| "learning_rate": 2.8676826773597485e-05, |
| "loss": 0.0142, |
| "step": 5340 |
| }, |
| { |
| "epoch": 1.0960868674451956, |
| "grad_norm": 0.19205980002880096, |
| "learning_rate": 2.8670028322798413e-05, |
| "loss": 0.0137, |
| "step": 5350 |
| }, |
| { |
| "epoch": 1.0981356279450931, |
| "grad_norm": 0.1406937539577484, |
| "learning_rate": 2.8663213261625932e-05, |
| "loss": 0.0155, |
| "step": 5360 |
| }, |
| { |
| "epoch": 1.100184388444991, |
| "grad_norm": 0.33962225914001465, |
| "learning_rate": 2.865638159836098e-05, |
| "loss": 0.0145, |
| "step": 5370 |
| }, |
| { |
| "epoch": 1.1022331489448884, |
| "grad_norm": 0.4572281837463379, |
| "learning_rate": 2.8649533341304677e-05, |
| "loss": 0.0126, |
| "step": 5380 |
| }, |
| { |
| "epoch": 1.104281909444786, |
| "grad_norm": 0.17225702106952667, |
| "learning_rate": 2.8642668498778297e-05, |
| "loss": 0.0132, |
| "step": 5390 |
| }, |
| { |
| "epoch": 1.1063306699446835, |
| "grad_norm": 0.14350320398807526, |
| "learning_rate": 2.8635787079123268e-05, |
| "loss": 0.0116, |
| "step": 5400 |
| }, |
| { |
| "epoch": 1.108379430444581, |
| "grad_norm": 0.21226346492767334, |
| "learning_rate": 2.8628889090701175e-05, |
| "loss": 0.0116, |
| "step": 5410 |
| }, |
| { |
| "epoch": 1.1104281909444786, |
| "grad_norm": 0.16102077066898346, |
| "learning_rate": 2.862197454189371e-05, |
| "loss": 0.0116, |
| "step": 5420 |
| }, |
| { |
| "epoch": 1.1124769514443762, |
| "grad_norm": 0.3439859449863434, |
| "learning_rate": 2.8615043441102708e-05, |
| "loss": 0.0135, |
| "step": 5430 |
| }, |
| { |
| "epoch": 1.1145257119442737, |
| "grad_norm": 0.2246524840593338, |
| "learning_rate": 2.860809579675011e-05, |
| "loss": 0.0128, |
| "step": 5440 |
| }, |
| { |
| "epoch": 1.1165744724441713, |
| "grad_norm": 0.2989099621772766, |
| "learning_rate": 2.860113161727796e-05, |
| "loss": 0.0138, |
| "step": 5450 |
| }, |
| { |
| "epoch": 1.1186232329440688, |
| "grad_norm": 0.38425007462501526, |
| "learning_rate": 2.8594150911148393e-05, |
| "loss": 0.0141, |
| "step": 5460 |
| }, |
| { |
| "epoch": 1.1206719934439664, |
| "grad_norm": 0.2721428871154785, |
| "learning_rate": 2.8587153686843626e-05, |
| "loss": 0.0117, |
| "step": 5470 |
| }, |
| { |
| "epoch": 1.122720753943864, |
| "grad_norm": 0.20072659850120544, |
| "learning_rate": 2.858013995286594e-05, |
| "loss": 0.0181, |
| "step": 5480 |
| }, |
| { |
| "epoch": 1.1247695144437615, |
| "grad_norm": 0.23932218551635742, |
| "learning_rate": 2.857310971773769e-05, |
| "loss": 0.0158, |
| "step": 5490 |
| }, |
| { |
| "epoch": 1.126818274943659, |
| "grad_norm": 0.20797573029994965, |
| "learning_rate": 2.8566062990001274e-05, |
| "loss": 0.0123, |
| "step": 5500 |
| }, |
| { |
| "epoch": 1.1288670354435566, |
| "grad_norm": 0.2794961631298065, |
| "learning_rate": 2.8558999778219126e-05, |
| "loss": 0.0128, |
| "step": 5510 |
| }, |
| { |
| "epoch": 1.1309157959434541, |
| "grad_norm": 0.1914968192577362, |
| "learning_rate": 2.8551920090973722e-05, |
| "loss": 0.0122, |
| "step": 5520 |
| }, |
| { |
| "epoch": 1.1329645564433517, |
| "grad_norm": 0.1378844678401947, |
| "learning_rate": 2.854482393686754e-05, |
| "loss": 0.0104, |
| "step": 5530 |
| }, |
| { |
| "epoch": 1.1350133169432493, |
| "grad_norm": 0.34908342361450195, |
| "learning_rate": 2.8537711324523086e-05, |
| "loss": 0.0138, |
| "step": 5540 |
| }, |
| { |
| "epoch": 1.1370620774431468, |
| "grad_norm": 0.8649958968162537, |
| "learning_rate": 2.8530582262582845e-05, |
| "loss": 0.0143, |
| "step": 5550 |
| }, |
| { |
| "epoch": 1.1391108379430444, |
| "grad_norm": 0.2970515787601471, |
| "learning_rate": 2.8523436759709306e-05, |
| "loss": 0.0113, |
| "step": 5560 |
| }, |
| { |
| "epoch": 1.141159598442942, |
| "grad_norm": 0.17652669548988342, |
| "learning_rate": 2.8516274824584926e-05, |
| "loss": 0.0136, |
| "step": 5570 |
| }, |
| { |
| "epoch": 1.1432083589428397, |
| "grad_norm": 0.17546609044075012, |
| "learning_rate": 2.8509096465912136e-05, |
| "loss": 0.0128, |
| "step": 5580 |
| }, |
| { |
| "epoch": 1.1452571194427372, |
| "grad_norm": 0.2901144325733185, |
| "learning_rate": 2.8501901692413315e-05, |
| "loss": 0.0131, |
| "step": 5590 |
| }, |
| { |
| "epoch": 1.1473058799426348, |
| "grad_norm": 0.41484323143959045, |
| "learning_rate": 2.8494690512830788e-05, |
| "loss": 0.013, |
| "step": 5600 |
| }, |
| { |
| "epoch": 1.1493546404425323, |
| "grad_norm": 0.19848094880580902, |
| "learning_rate": 2.8487462935926823e-05, |
| "loss": 0.0148, |
| "step": 5610 |
| }, |
| { |
| "epoch": 1.1514034009424299, |
| "grad_norm": 0.22155901789665222, |
| "learning_rate": 2.8480218970483606e-05, |
| "loss": 0.0149, |
| "step": 5620 |
| }, |
| { |
| "epoch": 1.1534521614423274, |
| "grad_norm": 0.37396571040153503, |
| "learning_rate": 2.847295862530324e-05, |
| "loss": 0.0113, |
| "step": 5630 |
| }, |
| { |
| "epoch": 1.155500921942225, |
| "grad_norm": 0.3009156584739685, |
| "learning_rate": 2.846568190920773e-05, |
| "loss": 0.0119, |
| "step": 5640 |
| }, |
| { |
| "epoch": 1.1575496824421225, |
| "grad_norm": 0.3635924160480499, |
| "learning_rate": 2.8458388831038965e-05, |
| "loss": 0.012, |
| "step": 5650 |
| }, |
| { |
| "epoch": 1.15959844294202, |
| "grad_norm": 0.44420886039733887, |
| "learning_rate": 2.8451079399658725e-05, |
| "loss": 0.0121, |
| "step": 5660 |
| }, |
| { |
| "epoch": 1.1616472034419176, |
| "grad_norm": 0.19199533760547638, |
| "learning_rate": 2.8443753623948665e-05, |
| "loss": 0.0125, |
| "step": 5670 |
| }, |
| { |
| "epoch": 1.1636959639418152, |
| "grad_norm": 0.4054226875305176, |
| "learning_rate": 2.8436411512810288e-05, |
| "loss": 0.0119, |
| "step": 5680 |
| }, |
| { |
| "epoch": 1.1657447244417127, |
| "grad_norm": 0.16558250784873962, |
| "learning_rate": 2.842905307516495e-05, |
| "loss": 0.0128, |
| "step": 5690 |
| }, |
| { |
| "epoch": 1.1677934849416103, |
| "grad_norm": 0.1912321299314499, |
| "learning_rate": 2.8421678319953854e-05, |
| "loss": 0.0126, |
| "step": 5700 |
| }, |
| { |
| "epoch": 1.1698422454415078, |
| "grad_norm": 0.2967148423194885, |
| "learning_rate": 2.841428725613801e-05, |
| "loss": 0.0115, |
| "step": 5710 |
| }, |
| { |
| "epoch": 1.1718910059414054, |
| "grad_norm": 0.2288387566804886, |
| "learning_rate": 2.8406879892698272e-05, |
| "loss": 0.0135, |
| "step": 5720 |
| }, |
| { |
| "epoch": 1.173939766441303, |
| "grad_norm": 0.25127553939819336, |
| "learning_rate": 2.8399456238635274e-05, |
| "loss": 0.0119, |
| "step": 5730 |
| }, |
| { |
| "epoch": 1.1759885269412005, |
| "grad_norm": 0.247034952044487, |
| "learning_rate": 2.8392016302969457e-05, |
| "loss": 0.0129, |
| "step": 5740 |
| }, |
| { |
| "epoch": 1.1780372874410983, |
| "grad_norm": 0.4078435003757477, |
| "learning_rate": 2.8384560094741055e-05, |
| "loss": 0.0133, |
| "step": 5750 |
| }, |
| { |
| "epoch": 1.1800860479409958, |
| "grad_norm": 0.3641800582408905, |
| "learning_rate": 2.8377087623010054e-05, |
| "loss": 0.0139, |
| "step": 5760 |
| }, |
| { |
| "epoch": 1.1821348084408934, |
| "grad_norm": 0.287808895111084, |
| "learning_rate": 2.8369598896856213e-05, |
| "loss": 0.0133, |
| "step": 5770 |
| }, |
| { |
| "epoch": 1.184183568940791, |
| "grad_norm": 0.337286114692688, |
| "learning_rate": 2.836209392537904e-05, |
| "loss": 0.014, |
| "step": 5780 |
| }, |
| { |
| "epoch": 1.1862323294406885, |
| "grad_norm": 0.5720916986465454, |
| "learning_rate": 2.8354572717697787e-05, |
| "loss": 0.0121, |
| "step": 5790 |
| }, |
| { |
| "epoch": 1.188281089940586, |
| "grad_norm": 0.27736276388168335, |
| "learning_rate": 2.8347035282951428e-05, |
| "loss": 0.0135, |
| "step": 5800 |
| }, |
| { |
| "epoch": 1.1903298504404836, |
| "grad_norm": 0.19946762919425964, |
| "learning_rate": 2.8339481630298653e-05, |
| "loss": 0.0129, |
| "step": 5810 |
| }, |
| { |
| "epoch": 1.1923786109403811, |
| "grad_norm": 0.11657186597585678, |
| "learning_rate": 2.833191176891787e-05, |
| "loss": 0.0097, |
| "step": 5820 |
| }, |
| { |
| "epoch": 1.1944273714402787, |
| "grad_norm": 0.2431401163339615, |
| "learning_rate": 2.832432570800717e-05, |
| "loss": 0.0116, |
| "step": 5830 |
| }, |
| { |
| "epoch": 1.1964761319401762, |
| "grad_norm": 0.1273585706949234, |
| "learning_rate": 2.8316723456784332e-05, |
| "loss": 0.0119, |
| "step": 5840 |
| }, |
| { |
| "epoch": 1.1985248924400738, |
| "grad_norm": 0.41356801986694336, |
| "learning_rate": 2.8309105024486808e-05, |
| "loss": 0.0113, |
| "step": 5850 |
| }, |
| { |
| "epoch": 1.2005736529399713, |
| "grad_norm": 0.29914039373397827, |
| "learning_rate": 2.830147042037172e-05, |
| "loss": 0.014, |
| "step": 5860 |
| }, |
| { |
| "epoch": 1.2026224134398689, |
| "grad_norm": 0.5134176015853882, |
| "learning_rate": 2.8293819653715816e-05, |
| "loss": 0.0149, |
| "step": 5870 |
| }, |
| { |
| "epoch": 1.2046711739397664, |
| "grad_norm": 0.36912184953689575, |
| "learning_rate": 2.828615273381551e-05, |
| "loss": 0.0135, |
| "step": 5880 |
| }, |
| { |
| "epoch": 1.206719934439664, |
| "grad_norm": 0.34139683842658997, |
| "learning_rate": 2.827846966998683e-05, |
| "loss": 0.0109, |
| "step": 5890 |
| }, |
| { |
| "epoch": 1.2087686949395615, |
| "grad_norm": 0.18890626728534698, |
| "learning_rate": 2.8270770471565422e-05, |
| "loss": 0.0127, |
| "step": 5900 |
| }, |
| { |
| "epoch": 1.210817455439459, |
| "grad_norm": 0.19830946624279022, |
| "learning_rate": 2.826305514790654e-05, |
| "loss": 0.013, |
| "step": 5910 |
| }, |
| { |
| "epoch": 1.2128662159393566, |
| "grad_norm": 0.33469995856285095, |
| "learning_rate": 2.825532370838503e-05, |
| "loss": 0.013, |
| "step": 5920 |
| }, |
| { |
| "epoch": 1.2149149764392542, |
| "grad_norm": 0.1715225726366043, |
| "learning_rate": 2.824757616239532e-05, |
| "loss": 0.015, |
| "step": 5930 |
| }, |
| { |
| "epoch": 1.2169637369391517, |
| "grad_norm": 0.22815775871276855, |
| "learning_rate": 2.8239812519351404e-05, |
| "loss": 0.0144, |
| "step": 5940 |
| }, |
| { |
| "epoch": 1.2190124974390493, |
| "grad_norm": 0.1659063696861267, |
| "learning_rate": 2.823203278868685e-05, |
| "loss": 0.0159, |
| "step": 5950 |
| }, |
| { |
| "epoch": 1.2210612579389468, |
| "grad_norm": 0.22280147671699524, |
| "learning_rate": 2.8224236979854752e-05, |
| "loss": 0.0167, |
| "step": 5960 |
| }, |
| { |
| "epoch": 1.2231100184388446, |
| "grad_norm": 0.1894383728504181, |
| "learning_rate": 2.821642510232776e-05, |
| "loss": 0.0161, |
| "step": 5970 |
| }, |
| { |
| "epoch": 1.2251587789387421, |
| "grad_norm": 0.3990509808063507, |
| "learning_rate": 2.8208597165598046e-05, |
| "loss": 0.0164, |
| "step": 5980 |
| }, |
| { |
| "epoch": 1.2272075394386397, |
| "grad_norm": 0.2119562178850174, |
| "learning_rate": 2.8200753179177285e-05, |
| "loss": 0.0149, |
| "step": 5990 |
| }, |
| { |
| "epoch": 1.2292562999385372, |
| "grad_norm": 0.1994301676750183, |
| "learning_rate": 2.819289315259666e-05, |
| "loss": 0.0152, |
| "step": 6000 |
| }, |
| { |
| "epoch": 1.2313050604384348, |
| "grad_norm": 0.2767426669597626, |
| "learning_rate": 2.8185017095406845e-05, |
| "loss": 0.0122, |
| "step": 6010 |
| }, |
| { |
| "epoch": 1.2333538209383323, |
| "grad_norm": 0.34808582067489624, |
| "learning_rate": 2.8177125017177998e-05, |
| "loss": 0.0116, |
| "step": 6020 |
| }, |
| { |
| "epoch": 1.23540258143823, |
| "grad_norm": 0.5330958366394043, |
| "learning_rate": 2.8169216927499727e-05, |
| "loss": 0.0152, |
| "step": 6030 |
| }, |
| { |
| "epoch": 1.2374513419381274, |
| "grad_norm": 0.23534716665744781, |
| "learning_rate": 2.816129283598112e-05, |
| "loss": 0.0155, |
| "step": 6040 |
| }, |
| { |
| "epoch": 1.239500102438025, |
| "grad_norm": 0.21551455557346344, |
| "learning_rate": 2.815335275225068e-05, |
| "loss": 0.0145, |
| "step": 6050 |
| }, |
| { |
| "epoch": 1.2415488629379225, |
| "grad_norm": 0.2130284458398819, |
| "learning_rate": 2.8145396685956374e-05, |
| "loss": 0.0133, |
| "step": 6060 |
| }, |
| { |
| "epoch": 1.24359762343782, |
| "grad_norm": 0.49103668332099915, |
| "learning_rate": 2.813742464676556e-05, |
| "loss": 0.0138, |
| "step": 6070 |
| }, |
| { |
| "epoch": 1.2456463839377176, |
| "grad_norm": 0.1830034852027893, |
| "learning_rate": 2.8129436644365025e-05, |
| "loss": 0.0117, |
| "step": 6080 |
| }, |
| { |
| "epoch": 1.2476951444376152, |
| "grad_norm": 0.20279261469841003, |
| "learning_rate": 2.8121432688460938e-05, |
| "loss": 0.0135, |
| "step": 6090 |
| }, |
| { |
| "epoch": 1.2497439049375128, |
| "grad_norm": 0.2792716324329376, |
| "learning_rate": 2.8113412788778867e-05, |
| "loss": 0.013, |
| "step": 6100 |
| }, |
| { |
| "epoch": 1.2517926654374103, |
| "grad_norm": 0.32436633110046387, |
| "learning_rate": 2.8105376955063744e-05, |
| "loss": 0.0137, |
| "step": 6110 |
| }, |
| { |
| "epoch": 1.2538414259373079, |
| "grad_norm": 0.14203034341335297, |
| "learning_rate": 2.8097325197079862e-05, |
| "loss": 0.0164, |
| "step": 6120 |
| }, |
| { |
| "epoch": 1.2558901864372056, |
| "grad_norm": 0.43020227551460266, |
| "learning_rate": 2.808925752461087e-05, |
| "loss": 0.0098, |
| "step": 6130 |
| }, |
| { |
| "epoch": 1.2579389469371032, |
| "grad_norm": 0.3624671995639801, |
| "learning_rate": 2.808117394745975e-05, |
| "loss": 0.0127, |
| "step": 6140 |
| }, |
| { |
| "epoch": 1.2599877074370007, |
| "grad_norm": 0.30588075518608093, |
| "learning_rate": 2.807307447544881e-05, |
| "loss": 0.0128, |
| "step": 6150 |
| }, |
| { |
| "epoch": 1.2620364679368983, |
| "grad_norm": 0.28108155727386475, |
| "learning_rate": 2.806495911841967e-05, |
| "loss": 0.0126, |
| "step": 6160 |
| }, |
| { |
| "epoch": 1.2640852284367958, |
| "grad_norm": 0.2909795641899109, |
| "learning_rate": 2.8056827886233255e-05, |
| "loss": 0.0131, |
| "step": 6170 |
| }, |
| { |
| "epoch": 1.2661339889366934, |
| "grad_norm": 0.4865269064903259, |
| "learning_rate": 2.8048680788769783e-05, |
| "loss": 0.0128, |
| "step": 6180 |
| }, |
| { |
| "epoch": 1.268182749436591, |
| "grad_norm": 0.19630564749240875, |
| "learning_rate": 2.8040517835928742e-05, |
| "loss": 0.0126, |
| "step": 6190 |
| }, |
| { |
| "epoch": 1.2702315099364885, |
| "grad_norm": 0.318464994430542, |
| "learning_rate": 2.803233903762889e-05, |
| "loss": 0.0121, |
| "step": 6200 |
| }, |
| { |
| "epoch": 1.272280270436386, |
| "grad_norm": 0.3491669297218323, |
| "learning_rate": 2.8024144403808235e-05, |
| "loss": 0.0137, |
| "step": 6210 |
| }, |
| { |
| "epoch": 1.2743290309362836, |
| "grad_norm": 0.23283307254314423, |
| "learning_rate": 2.8015933944424037e-05, |
| "loss": 0.0104, |
| "step": 6220 |
| }, |
| { |
| "epoch": 1.2763777914361811, |
| "grad_norm": 0.13112273812294006, |
| "learning_rate": 2.800770766945277e-05, |
| "loss": 0.0121, |
| "step": 6230 |
| }, |
| { |
| "epoch": 1.2784265519360787, |
| "grad_norm": 0.2623870074748993, |
| "learning_rate": 2.7999465588890135e-05, |
| "loss": 0.0138, |
| "step": 6240 |
| }, |
| { |
| "epoch": 1.2804753124359762, |
| "grad_norm": 0.10492726415395737, |
| "learning_rate": 2.799120771275105e-05, |
| "loss": 0.0134, |
| "step": 6250 |
| }, |
| { |
| "epoch": 1.2825240729358738, |
| "grad_norm": 0.2797611653804779, |
| "learning_rate": 2.7982934051069597e-05, |
| "loss": 0.0122, |
| "step": 6260 |
| }, |
| { |
| "epoch": 1.2845728334357713, |
| "grad_norm": 0.2502429783344269, |
| "learning_rate": 2.7974644613899063e-05, |
| "loss": 0.0117, |
| "step": 6270 |
| }, |
| { |
| "epoch": 1.2866215939356689, |
| "grad_norm": 0.13542234897613525, |
| "learning_rate": 2.796633941131189e-05, |
| "loss": 0.0114, |
| "step": 6280 |
| }, |
| { |
| "epoch": 1.2886703544355664, |
| "grad_norm": 0.16352857649326324, |
| "learning_rate": 2.795801845339969e-05, |
| "loss": 0.0132, |
| "step": 6290 |
| }, |
| { |
| "epoch": 1.290719114935464, |
| "grad_norm": 1.1394368410110474, |
| "learning_rate": 2.7949681750273208e-05, |
| "loss": 0.0131, |
| "step": 6300 |
| }, |
| { |
| "epoch": 1.2927678754353615, |
| "grad_norm": 0.16163109242916107, |
| "learning_rate": 2.7941329312062322e-05, |
| "loss": 0.0139, |
| "step": 6310 |
| }, |
| { |
| "epoch": 1.294816635935259, |
| "grad_norm": 0.18337693810462952, |
| "learning_rate": 2.7932961148916037e-05, |
| "loss": 0.0139, |
| "step": 6320 |
| }, |
| { |
| "epoch": 1.2968653964351566, |
| "grad_norm": 0.2580595016479492, |
| "learning_rate": 2.7924577271002456e-05, |
| "loss": 0.0166, |
| "step": 6330 |
| }, |
| { |
| "epoch": 1.2989141569350542, |
| "grad_norm": 0.28770142793655396, |
| "learning_rate": 2.7916177688508782e-05, |
| "loss": 0.0126, |
| "step": 6340 |
| }, |
| { |
| "epoch": 1.3009629174349517, |
| "grad_norm": 0.2400563806295395, |
| "learning_rate": 2.7907762411641302e-05, |
| "loss": 0.0158, |
| "step": 6350 |
| }, |
| { |
| "epoch": 1.3030116779348493, |
| "grad_norm": 0.2640220522880554, |
| "learning_rate": 2.7899331450625367e-05, |
| "loss": 0.0113, |
| "step": 6360 |
| }, |
| { |
| "epoch": 1.305060438434747, |
| "grad_norm": 0.15067867934703827, |
| "learning_rate": 2.7890884815705397e-05, |
| "loss": 0.0122, |
| "step": 6370 |
| }, |
| { |
| "epoch": 1.3071091989346446, |
| "grad_norm": 0.4740571677684784, |
| "learning_rate": 2.788242251714484e-05, |
| "loss": 0.0158, |
| "step": 6380 |
| }, |
| { |
| "epoch": 1.3091579594345422, |
| "grad_norm": 0.28013890981674194, |
| "learning_rate": 2.787394456522619e-05, |
| "loss": 0.011, |
| "step": 6390 |
| }, |
| { |
| "epoch": 1.3112067199344397, |
| "grad_norm": 0.12139856070280075, |
| "learning_rate": 2.7865450970250965e-05, |
| "loss": 0.0107, |
| "step": 6400 |
| }, |
| { |
| "epoch": 1.3132554804343373, |
| "grad_norm": 0.24238674342632294, |
| "learning_rate": 2.7856941742539676e-05, |
| "loss": 0.0143, |
| "step": 6410 |
| }, |
| { |
| "epoch": 1.3153042409342348, |
| "grad_norm": 0.43777018785476685, |
| "learning_rate": 2.784841689243184e-05, |
| "loss": 0.013, |
| "step": 6420 |
| }, |
| { |
| "epoch": 1.3173530014341324, |
| "grad_norm": 0.30293965339660645, |
| "learning_rate": 2.783987643028595e-05, |
| "loss": 0.0126, |
| "step": 6430 |
| }, |
| { |
| "epoch": 1.31940176193403, |
| "grad_norm": 0.17805764079093933, |
| "learning_rate": 2.783132036647948e-05, |
| "loss": 0.0144, |
| "step": 6440 |
| }, |
| { |
| "epoch": 1.3214505224339275, |
| "grad_norm": 0.19755655527114868, |
| "learning_rate": 2.7822748711408852e-05, |
| "loss": 0.0127, |
| "step": 6450 |
| }, |
| { |
| "epoch": 1.323499282933825, |
| "grad_norm": 0.12483599781990051, |
| "learning_rate": 2.7814161475489428e-05, |
| "loss": 0.0131, |
| "step": 6460 |
| }, |
| { |
| "epoch": 1.3255480434337226, |
| "grad_norm": 0.14574919641017914, |
| "learning_rate": 2.780555866915552e-05, |
| "loss": 0.0135, |
| "step": 6470 |
| }, |
| { |
| "epoch": 1.3275968039336201, |
| "grad_norm": 0.25004565715789795, |
| "learning_rate": 2.7796940302860338e-05, |
| "loss": 0.0137, |
| "step": 6480 |
| }, |
| { |
| "epoch": 1.3296455644335177, |
| "grad_norm": 0.3793397843837738, |
| "learning_rate": 2.7788306387076018e-05, |
| "loss": 0.0151, |
| "step": 6490 |
| }, |
| { |
| "epoch": 1.3316943249334152, |
| "grad_norm": 0.2987273335456848, |
| "learning_rate": 2.7779656932293584e-05, |
| "loss": 0.0149, |
| "step": 6500 |
| }, |
| { |
| "epoch": 1.3337430854333128, |
| "grad_norm": 0.3328090012073517, |
| "learning_rate": 2.777099194902293e-05, |
| "loss": 0.0121, |
| "step": 6510 |
| }, |
| { |
| "epoch": 1.3357918459332105, |
| "grad_norm": 0.2181820124387741, |
| "learning_rate": 2.776231144779283e-05, |
| "loss": 0.0146, |
| "step": 6520 |
| }, |
| { |
| "epoch": 1.337840606433108, |
| "grad_norm": 0.2091411054134369, |
| "learning_rate": 2.775361543915091e-05, |
| "loss": 0.013, |
| "step": 6530 |
| }, |
| { |
| "epoch": 1.3398893669330056, |
| "grad_norm": 0.24733686447143555, |
| "learning_rate": 2.774490393366365e-05, |
| "loss": 0.0137, |
| "step": 6540 |
| }, |
| { |
| "epoch": 1.3419381274329032, |
| "grad_norm": 0.20312221348285675, |
| "learning_rate": 2.7736176941916346e-05, |
| "loss": 0.0141, |
| "step": 6550 |
| }, |
| { |
| "epoch": 1.3439868879328007, |
| "grad_norm": 0.5528634190559387, |
| "learning_rate": 2.772743447451311e-05, |
| "loss": 0.0116, |
| "step": 6560 |
| }, |
| { |
| "epoch": 1.3460356484326983, |
| "grad_norm": 0.21864120662212372, |
| "learning_rate": 2.7718676542076872e-05, |
| "loss": 0.0125, |
| "step": 6570 |
| }, |
| { |
| "epoch": 1.3480844089325958, |
| "grad_norm": 0.313936710357666, |
| "learning_rate": 2.770990315524934e-05, |
| "loss": 0.0113, |
| "step": 6580 |
| }, |
| { |
| "epoch": 1.3501331694324934, |
| "grad_norm": 0.15080596506595612, |
| "learning_rate": 2.7701114324691015e-05, |
| "loss": 0.011, |
| "step": 6590 |
| }, |
| { |
| "epoch": 1.352181929932391, |
| "grad_norm": 0.2524586319923401, |
| "learning_rate": 2.7692310061081145e-05, |
| "loss": 0.0119, |
| "step": 6600 |
| }, |
| { |
| "epoch": 1.3542306904322885, |
| "grad_norm": 0.3850145936012268, |
| "learning_rate": 2.7683490375117742e-05, |
| "loss": 0.013, |
| "step": 6610 |
| }, |
| { |
| "epoch": 1.356279450932186, |
| "grad_norm": 0.6408969759941101, |
| "learning_rate": 2.7674655277517556e-05, |
| "loss": 0.0138, |
| "step": 6620 |
| }, |
| { |
| "epoch": 1.3583282114320836, |
| "grad_norm": 0.25372737646102905, |
| "learning_rate": 2.7665804779016072e-05, |
| "loss": 0.0151, |
| "step": 6630 |
| }, |
| { |
| "epoch": 1.3603769719319811, |
| "grad_norm": 0.329515665769577, |
| "learning_rate": 2.765693889036747e-05, |
| "loss": 0.0103, |
| "step": 6640 |
| }, |
| { |
| "epoch": 1.3624257324318787, |
| "grad_norm": 0.21986433863639832, |
| "learning_rate": 2.764805762234465e-05, |
| "loss": 0.0132, |
| "step": 6650 |
| }, |
| { |
| "epoch": 1.3644744929317763, |
| "grad_norm": 0.23815444111824036, |
| "learning_rate": 2.7639160985739182e-05, |
| "loss": 0.0145, |
| "step": 6660 |
| }, |
| { |
| "epoch": 1.3665232534316738, |
| "grad_norm": 0.40145987272262573, |
| "learning_rate": 2.7630248991361317e-05, |
| "loss": 0.0118, |
| "step": 6670 |
| }, |
| { |
| "epoch": 1.3685720139315714, |
| "grad_norm": 0.41701415181159973, |
| "learning_rate": 2.762132165003998e-05, |
| "loss": 0.0125, |
| "step": 6680 |
| }, |
| { |
| "epoch": 1.370620774431469, |
| "grad_norm": 0.3597390949726105, |
| "learning_rate": 2.7612378972622722e-05, |
| "loss": 0.0111, |
| "step": 6690 |
| }, |
| { |
| "epoch": 1.3726695349313665, |
| "grad_norm": 0.13960440456867218, |
| "learning_rate": 2.7603420969975743e-05, |
| "loss": 0.0112, |
| "step": 6700 |
| }, |
| { |
| "epoch": 1.374718295431264, |
| "grad_norm": 0.302381306886673, |
| "learning_rate": 2.7594447652983862e-05, |
| "loss": 0.0143, |
| "step": 6710 |
| }, |
| { |
| "epoch": 1.3767670559311616, |
| "grad_norm": 0.25259408354759216, |
| "learning_rate": 2.7585459032550497e-05, |
| "loss": 0.0121, |
| "step": 6720 |
| }, |
| { |
| "epoch": 1.378815816431059, |
| "grad_norm": 0.6331823468208313, |
| "learning_rate": 2.7576455119597684e-05, |
| "loss": 0.0122, |
| "step": 6730 |
| }, |
| { |
| "epoch": 1.3808645769309567, |
| "grad_norm": 0.22952502965927124, |
| "learning_rate": 2.7567435925066012e-05, |
| "loss": 0.0117, |
| "step": 6740 |
| }, |
| { |
| "epoch": 1.3829133374308542, |
| "grad_norm": 0.22592857480049133, |
| "learning_rate": 2.7558401459914665e-05, |
| "loss": 0.0115, |
| "step": 6750 |
| }, |
| { |
| "epoch": 1.3849620979307518, |
| "grad_norm": 0.17219476401805878, |
| "learning_rate": 2.7549351735121362e-05, |
| "loss": 0.014, |
| "step": 6760 |
| }, |
| { |
| "epoch": 1.3870108584306495, |
| "grad_norm": 0.3418867588043213, |
| "learning_rate": 2.7540286761682377e-05, |
| "loss": 0.0141, |
| "step": 6770 |
| }, |
| { |
| "epoch": 1.389059618930547, |
| "grad_norm": 0.512632429599762, |
| "learning_rate": 2.75312065506125e-05, |
| "loss": 0.0126, |
| "step": 6780 |
| }, |
| { |
| "epoch": 1.3911083794304446, |
| "grad_norm": 0.244014173746109, |
| "learning_rate": 2.7522111112945052e-05, |
| "loss": 0.0122, |
| "step": 6790 |
| }, |
| { |
| "epoch": 1.3931571399303422, |
| "grad_norm": 0.213824063539505, |
| "learning_rate": 2.751300045973184e-05, |
| "loss": 0.0122, |
| "step": 6800 |
| }, |
| { |
| "epoch": 1.3952059004302397, |
| "grad_norm": 0.11551392823457718, |
| "learning_rate": 2.750387460204317e-05, |
| "loss": 0.0117, |
| "step": 6810 |
| }, |
| { |
| "epoch": 1.3972546609301373, |
| "grad_norm": 0.8220825791358948, |
| "learning_rate": 2.7494733550967827e-05, |
| "loss": 0.0159, |
| "step": 6820 |
| }, |
| { |
| "epoch": 1.3993034214300348, |
| "grad_norm": 0.4638221561908722, |
| "learning_rate": 2.748557731761304e-05, |
| "loss": 0.0165, |
| "step": 6830 |
| }, |
| { |
| "epoch": 1.4013521819299324, |
| "grad_norm": 0.17546994984149933, |
| "learning_rate": 2.74764059131045e-05, |
| "loss": 0.0134, |
| "step": 6840 |
| }, |
| { |
| "epoch": 1.40340094242983, |
| "grad_norm": 0.5194903612136841, |
| "learning_rate": 2.7467219348586322e-05, |
| "loss": 0.0128, |
| "step": 6850 |
| }, |
| { |
| "epoch": 1.4054497029297275, |
| "grad_norm": 0.22370308637619019, |
| "learning_rate": 2.7458017635221056e-05, |
| "loss": 0.0114, |
| "step": 6860 |
| }, |
| { |
| "epoch": 1.407498463429625, |
| "grad_norm": 0.2080584466457367, |
| "learning_rate": 2.7448800784189652e-05, |
| "loss": 0.0148, |
| "step": 6870 |
| }, |
| { |
| "epoch": 1.4095472239295226, |
| "grad_norm": 0.24876311421394348, |
| "learning_rate": 2.7439568806691448e-05, |
| "loss": 0.0111, |
| "step": 6880 |
| }, |
| { |
| "epoch": 1.4115959844294201, |
| "grad_norm": 0.20740638673305511, |
| "learning_rate": 2.7430321713944167e-05, |
| "loss": 0.0129, |
| "step": 6890 |
| }, |
| { |
| "epoch": 1.4136447449293177, |
| "grad_norm": 0.12317488342523575, |
| "learning_rate": 2.7421059517183906e-05, |
| "loss": 0.0106, |
| "step": 6900 |
| }, |
| { |
| "epoch": 1.4156935054292155, |
| "grad_norm": 0.17415420711040497, |
| "learning_rate": 2.7411782227665097e-05, |
| "loss": 0.0117, |
| "step": 6910 |
| }, |
| { |
| "epoch": 1.417742265929113, |
| "grad_norm": 0.4563613533973694, |
| "learning_rate": 2.7402489856660528e-05, |
| "loss": 0.0136, |
| "step": 6920 |
| }, |
| { |
| "epoch": 1.4197910264290106, |
| "grad_norm": 0.1932782679796219, |
| "learning_rate": 2.7393182415461307e-05, |
| "loss": 0.0135, |
| "step": 6930 |
| }, |
| { |
| "epoch": 1.4218397869289081, |
| "grad_norm": 0.1388777494430542, |
| "learning_rate": 2.7383859915376853e-05, |
| "loss": 0.0139, |
| "step": 6940 |
| }, |
| { |
| "epoch": 1.4238885474288057, |
| "grad_norm": 0.17383243143558502, |
| "learning_rate": 2.7374522367734875e-05, |
| "loss": 0.0126, |
| "step": 6950 |
| }, |
| { |
| "epoch": 1.4259373079287032, |
| "grad_norm": 0.288992315530777, |
| "learning_rate": 2.7365169783881378e-05, |
| "loss": 0.0124, |
| "step": 6960 |
| }, |
| { |
| "epoch": 1.4279860684286008, |
| "grad_norm": 0.19477839767932892, |
| "learning_rate": 2.735580217518064e-05, |
| "loss": 0.0114, |
| "step": 6970 |
| }, |
| { |
| "epoch": 1.4300348289284983, |
| "grad_norm": 0.48894810676574707, |
| "learning_rate": 2.7346419553015175e-05, |
| "loss": 0.0121, |
| "step": 6980 |
| }, |
| { |
| "epoch": 1.4320835894283959, |
| "grad_norm": 0.25800785422325134, |
| "learning_rate": 2.733702192878577e-05, |
| "loss": 0.0126, |
| "step": 6990 |
| }, |
| { |
| "epoch": 1.4341323499282934, |
| "grad_norm": 0.2877299189567566, |
| "learning_rate": 2.732760931391141e-05, |
| "loss": 0.0146, |
| "step": 7000 |
| }, |
| { |
| "epoch": 1.436181110428191, |
| "grad_norm": 0.3342173993587494, |
| "learning_rate": 2.731818171982932e-05, |
| "loss": 0.0136, |
| "step": 7010 |
| }, |
| { |
| "epoch": 1.4382298709280885, |
| "grad_norm": 0.17168273031711578, |
| "learning_rate": 2.730873915799491e-05, |
| "loss": 0.0105, |
| "step": 7020 |
| }, |
| { |
| "epoch": 1.440278631427986, |
| "grad_norm": 0.2379096895456314, |
| "learning_rate": 2.7299281639881783e-05, |
| "loss": 0.0127, |
| "step": 7030 |
| }, |
| { |
| "epoch": 1.4423273919278836, |
| "grad_norm": 0.22283445298671722, |
| "learning_rate": 2.728980917698172e-05, |
| "loss": 0.0139, |
| "step": 7040 |
| }, |
| { |
| "epoch": 1.4443761524277812, |
| "grad_norm": 0.20537123084068298, |
| "learning_rate": 2.7280321780804655e-05, |
| "loss": 0.0127, |
| "step": 7050 |
| }, |
| { |
| "epoch": 1.4464249129276787, |
| "grad_norm": 0.3638201057910919, |
| "learning_rate": 2.7270819462878666e-05, |
| "loss": 0.0165, |
| "step": 7060 |
| }, |
| { |
| "epoch": 1.4484736734275763, |
| "grad_norm": 0.38690879940986633, |
| "learning_rate": 2.7261302234749965e-05, |
| "loss": 0.0138, |
| "step": 7070 |
| }, |
| { |
| "epoch": 1.4505224339274738, |
| "grad_norm": 0.2533414661884308, |
| "learning_rate": 2.7251770107982888e-05, |
| "loss": 0.017, |
| "step": 7080 |
| }, |
| { |
| "epoch": 1.4525711944273714, |
| "grad_norm": 0.22235548496246338, |
| "learning_rate": 2.7242223094159856e-05, |
| "loss": 0.015, |
| "step": 7090 |
| }, |
| { |
| "epoch": 1.454619954927269, |
| "grad_norm": 0.3074291944503784, |
| "learning_rate": 2.7232661204881408e-05, |
| "loss": 0.0118, |
| "step": 7100 |
| }, |
| { |
| "epoch": 1.4566687154271665, |
| "grad_norm": 0.31307297945022583, |
| "learning_rate": 2.7223084451766123e-05, |
| "loss": 0.0138, |
| "step": 7110 |
| }, |
| { |
| "epoch": 1.458717475927064, |
| "grad_norm": 0.5192930102348328, |
| "learning_rate": 2.721349284645067e-05, |
| "loss": 0.0111, |
| "step": 7120 |
| }, |
| { |
| "epoch": 1.4607662364269616, |
| "grad_norm": 0.6757801175117493, |
| "learning_rate": 2.7203886400589752e-05, |
| "loss": 0.0124, |
| "step": 7130 |
| }, |
| { |
| "epoch": 1.4628149969268591, |
| "grad_norm": 0.40396848320961, |
| "learning_rate": 2.7194265125856105e-05, |
| "loss": 0.0145, |
| "step": 7140 |
| }, |
| { |
| "epoch": 1.4648637574267567, |
| "grad_norm": 0.32780084013938904, |
| "learning_rate": 2.718462903394049e-05, |
| "loss": 0.0119, |
| "step": 7150 |
| }, |
| { |
| "epoch": 1.4669125179266544, |
| "grad_norm": 0.31194281578063965, |
| "learning_rate": 2.7174978136551662e-05, |
| "loss": 0.0137, |
| "step": 7160 |
| }, |
| { |
| "epoch": 1.468961278426552, |
| "grad_norm": 0.1790454387664795, |
| "learning_rate": 2.716531244541637e-05, |
| "loss": 0.0145, |
| "step": 7170 |
| }, |
| { |
| "epoch": 1.4710100389264495, |
| "grad_norm": 0.40029376745224, |
| "learning_rate": 2.715563197227934e-05, |
| "loss": 0.0138, |
| "step": 7180 |
| }, |
| { |
| "epoch": 1.473058799426347, |
| "grad_norm": 0.23436768352985382, |
| "learning_rate": 2.714593672890327e-05, |
| "loss": 0.0133, |
| "step": 7190 |
| }, |
| { |
| "epoch": 1.4751075599262446, |
| "grad_norm": 0.11921199411153793, |
| "learning_rate": 2.7136226727068783e-05, |
| "loss": 0.0164, |
| "step": 7200 |
| }, |
| { |
| "epoch": 1.4771563204261422, |
| "grad_norm": 0.18620969355106354, |
| "learning_rate": 2.712650197857445e-05, |
| "loss": 0.0135, |
| "step": 7210 |
| }, |
| { |
| "epoch": 1.4792050809260398, |
| "grad_norm": 0.46304601430892944, |
| "learning_rate": 2.7116762495236763e-05, |
| "loss": 0.0131, |
| "step": 7220 |
| }, |
| { |
| "epoch": 1.4812538414259373, |
| "grad_norm": 0.2806619107723236, |
| "learning_rate": 2.7107008288890106e-05, |
| "loss": 0.0148, |
| "step": 7230 |
| }, |
| { |
| "epoch": 1.4833026019258349, |
| "grad_norm": 0.16201840341091156, |
| "learning_rate": 2.709723937138676e-05, |
| "loss": 0.013, |
| "step": 7240 |
| }, |
| { |
| "epoch": 1.4853513624257324, |
| "grad_norm": 0.27622106671333313, |
| "learning_rate": 2.7087455754596883e-05, |
| "loss": 0.0133, |
| "step": 7250 |
| }, |
| { |
| "epoch": 1.48740012292563, |
| "grad_norm": 0.14993157982826233, |
| "learning_rate": 2.7077657450408498e-05, |
| "loss": 0.0116, |
| "step": 7260 |
| }, |
| { |
| "epoch": 1.4894488834255275, |
| "grad_norm": 0.19468294084072113, |
| "learning_rate": 2.7067844470727456e-05, |
| "loss": 0.0127, |
| "step": 7270 |
| }, |
| { |
| "epoch": 1.491497643925425, |
| "grad_norm": 0.3145378530025482, |
| "learning_rate": 2.7058016827477455e-05, |
| "loss": 0.0127, |
| "step": 7280 |
| }, |
| { |
| "epoch": 1.4935464044253226, |
| "grad_norm": 0.24917137622833252, |
| "learning_rate": 2.7048174532600018e-05, |
| "loss": 0.0126, |
| "step": 7290 |
| }, |
| { |
| "epoch": 1.4955951649252204, |
| "grad_norm": 0.14209450781345367, |
| "learning_rate": 2.7038317598054455e-05, |
| "loss": 0.0141, |
| "step": 7300 |
| }, |
| { |
| "epoch": 1.497643925425118, |
| "grad_norm": 0.14404326677322388, |
| "learning_rate": 2.702844603581787e-05, |
| "loss": 0.0122, |
| "step": 7310 |
| }, |
| { |
| "epoch": 1.4996926859250155, |
| "grad_norm": 0.41484013199806213, |
| "learning_rate": 2.701855985788514e-05, |
| "loss": 0.0126, |
| "step": 7320 |
| }, |
| { |
| "epoch": 1.501741446424913, |
| "grad_norm": 0.4072243869304657, |
| "learning_rate": 2.7008659076268912e-05, |
| "loss": 0.014, |
| "step": 7330 |
| }, |
| { |
| "epoch": 1.5037902069248106, |
| "grad_norm": 0.3767257630825043, |
| "learning_rate": 2.6998743702999562e-05, |
| "loss": 0.0138, |
| "step": 7340 |
| }, |
| { |
| "epoch": 1.5058389674247081, |
| "grad_norm": 0.25109589099884033, |
| "learning_rate": 2.6988813750125202e-05, |
| "loss": 0.0124, |
| "step": 7350 |
| }, |
| { |
| "epoch": 1.5078877279246057, |
| "grad_norm": 0.21523475646972656, |
| "learning_rate": 2.6978869229711668e-05, |
| "loss": 0.0126, |
| "step": 7360 |
| }, |
| { |
| "epoch": 1.5099364884245032, |
| "grad_norm": 0.6786904335021973, |
| "learning_rate": 2.6968910153842484e-05, |
| "loss": 0.0143, |
| "step": 7370 |
| }, |
| { |
| "epoch": 1.5119852489244008, |
| "grad_norm": 0.18138009309768677, |
| "learning_rate": 2.6958936534618874e-05, |
| "loss": 0.0135, |
| "step": 7380 |
| }, |
| { |
| "epoch": 1.5140340094242983, |
| "grad_norm": 0.2903938293457031, |
| "learning_rate": 2.694894838415971e-05, |
| "loss": 0.0136, |
| "step": 7390 |
| }, |
| { |
| "epoch": 1.5160827699241959, |
| "grad_norm": 0.30754369497299194, |
| "learning_rate": 2.6938945714601555e-05, |
| "loss": 0.013, |
| "step": 7400 |
| }, |
| { |
| "epoch": 1.5181315304240934, |
| "grad_norm": 0.3371429443359375, |
| "learning_rate": 2.6928928538098587e-05, |
| "loss": 0.012, |
| "step": 7410 |
| }, |
| { |
| "epoch": 1.520180290923991, |
| "grad_norm": 0.17103278636932373, |
| "learning_rate": 2.6918896866822615e-05, |
| "loss": 0.0126, |
| "step": 7420 |
| }, |
| { |
| "epoch": 1.5222290514238885, |
| "grad_norm": 0.20670472085475922, |
| "learning_rate": 2.690885071296308e-05, |
| "loss": 0.0125, |
| "step": 7430 |
| }, |
| { |
| "epoch": 1.524277811923786, |
| "grad_norm": 0.3495783507823944, |
| "learning_rate": 2.6898790088726993e-05, |
| "loss": 0.0145, |
| "step": 7440 |
| }, |
| { |
| "epoch": 1.5263265724236836, |
| "grad_norm": 0.12448907643556595, |
| "learning_rate": 2.6888715006338968e-05, |
| "loss": 0.0123, |
| "step": 7450 |
| }, |
| { |
| "epoch": 1.5283753329235812, |
| "grad_norm": 0.38223981857299805, |
| "learning_rate": 2.6878625478041172e-05, |
| "loss": 0.0139, |
| "step": 7460 |
| }, |
| { |
| "epoch": 1.5304240934234787, |
| "grad_norm": 0.3152601420879364, |
| "learning_rate": 2.6868521516093342e-05, |
| "loss": 0.0125, |
| "step": 7470 |
| }, |
| { |
| "epoch": 1.5324728539233763, |
| "grad_norm": 0.2686951458454132, |
| "learning_rate": 2.6858403132772745e-05, |
| "loss": 0.0142, |
| "step": 7480 |
| }, |
| { |
| "epoch": 1.5345216144232738, |
| "grad_norm": 0.1875125914812088, |
| "learning_rate": 2.6848270340374162e-05, |
| "loss": 0.0114, |
| "step": 7490 |
| }, |
| { |
| "epoch": 1.5365703749231714, |
| "grad_norm": 0.2119808793067932, |
| "learning_rate": 2.68381231512099e-05, |
| "loss": 0.0111, |
| "step": 7500 |
| }, |
| { |
| "epoch": 1.538619135423069, |
| "grad_norm": 0.3255583345890045, |
| "learning_rate": 2.6827961577609745e-05, |
| "loss": 0.013, |
| "step": 7510 |
| }, |
| { |
| "epoch": 1.5406678959229665, |
| "grad_norm": 0.24628223478794098, |
| "learning_rate": 2.681778563192097e-05, |
| "loss": 0.0132, |
| "step": 7520 |
| }, |
| { |
| "epoch": 1.542716656422864, |
| "grad_norm": 0.17762424051761627, |
| "learning_rate": 2.6807595326508308e-05, |
| "loss": 0.011, |
| "step": 7530 |
| }, |
| { |
| "epoch": 1.5447654169227616, |
| "grad_norm": 0.14972083270549774, |
| "learning_rate": 2.679739067375394e-05, |
| "loss": 0.0143, |
| "step": 7540 |
| }, |
| { |
| "epoch": 1.5468141774226591, |
| "grad_norm": 0.16703253984451294, |
| "learning_rate": 2.6787171686057487e-05, |
| "loss": 0.0122, |
| "step": 7550 |
| }, |
| { |
| "epoch": 1.5488629379225567, |
| "grad_norm": 0.2700400650501251, |
| "learning_rate": 2.677693837583597e-05, |
| "loss": 0.0139, |
| "step": 7560 |
| }, |
| { |
| "epoch": 1.5509116984224545, |
| "grad_norm": 0.33240482211112976, |
| "learning_rate": 2.6766690755523836e-05, |
| "loss": 0.0119, |
| "step": 7570 |
| }, |
| { |
| "epoch": 1.552960458922352, |
| "grad_norm": 0.25089046359062195, |
| "learning_rate": 2.675642883757291e-05, |
| "loss": 0.0124, |
| "step": 7580 |
| }, |
| { |
| "epoch": 1.5550092194222496, |
| "grad_norm": 0.17166078090667725, |
| "learning_rate": 2.674615263445239e-05, |
| "loss": 0.0132, |
| "step": 7590 |
| }, |
| { |
| "epoch": 1.5570579799221471, |
| "grad_norm": 0.29602622985839844, |
| "learning_rate": 2.673586215864882e-05, |
| "loss": 0.0117, |
| "step": 7600 |
| }, |
| { |
| "epoch": 1.5591067404220447, |
| "grad_norm": 0.1616889387369156, |
| "learning_rate": 2.6725557422666113e-05, |
| "loss": 0.0118, |
| "step": 7610 |
| }, |
| { |
| "epoch": 1.5611555009219422, |
| "grad_norm": 0.6027488112449646, |
| "learning_rate": 2.6715238439025493e-05, |
| "loss": 0.0118, |
| "step": 7620 |
| }, |
| { |
| "epoch": 1.5632042614218398, |
| "grad_norm": 0.1984269767999649, |
| "learning_rate": 2.6704905220265493e-05, |
| "loss": 0.0109, |
| "step": 7630 |
| }, |
| { |
| "epoch": 1.5652530219217373, |
| "grad_norm": 0.30951443314552307, |
| "learning_rate": 2.6694557778941947e-05, |
| "loss": 0.0161, |
| "step": 7640 |
| }, |
| { |
| "epoch": 1.5673017824216349, |
| "grad_norm": 0.14493227005004883, |
| "learning_rate": 2.6684196127627977e-05, |
| "loss": 0.013, |
| "step": 7650 |
| }, |
| { |
| "epoch": 1.5693505429215324, |
| "grad_norm": 0.3415921628475189, |
| "learning_rate": 2.6673820278913963e-05, |
| "loss": 0.0143, |
| "step": 7660 |
| }, |
| { |
| "epoch": 1.5713993034214302, |
| "grad_norm": 0.32538485527038574, |
| "learning_rate": 2.6663430245407544e-05, |
| "loss": 0.0135, |
| "step": 7670 |
| }, |
| { |
| "epoch": 1.5734480639213277, |
| "grad_norm": 0.31267184019088745, |
| "learning_rate": 2.6653026039733585e-05, |
| "loss": 0.0134, |
| "step": 7680 |
| }, |
| { |
| "epoch": 1.5754968244212253, |
| "grad_norm": 0.4506918787956238, |
| "learning_rate": 2.6642607674534176e-05, |
| "loss": 0.0128, |
| "step": 7690 |
| }, |
| { |
| "epoch": 1.5775455849211228, |
| "grad_norm": 0.2968555688858032, |
| "learning_rate": 2.663217516246862e-05, |
| "loss": 0.011, |
| "step": 7700 |
| }, |
| { |
| "epoch": 1.5795943454210204, |
| "grad_norm": 0.16254118084907532, |
| "learning_rate": 2.66217285162134e-05, |
| "loss": 0.0153, |
| "step": 7710 |
| }, |
| { |
| "epoch": 1.581643105920918, |
| "grad_norm": 0.34746769070625305, |
| "learning_rate": 2.6611267748462175e-05, |
| "loss": 0.0139, |
| "step": 7720 |
| }, |
| { |
| "epoch": 1.5836918664208155, |
| "grad_norm": 0.2644629180431366, |
| "learning_rate": 2.6600792871925764e-05, |
| "loss": 0.0134, |
| "step": 7730 |
| }, |
| { |
| "epoch": 1.585740626920713, |
| "grad_norm": 0.15760578215122223, |
| "learning_rate": 2.659030389933213e-05, |
| "loss": 0.0102, |
| "step": 7740 |
| }, |
| { |
| "epoch": 1.5877893874206106, |
| "grad_norm": 0.21237245202064514, |
| "learning_rate": 2.6579800843426358e-05, |
| "loss": 0.0121, |
| "step": 7750 |
| }, |
| { |
| "epoch": 1.5898381479205081, |
| "grad_norm": 0.6923648118972778, |
| "learning_rate": 2.656928371697067e-05, |
| "loss": 0.0139, |
| "step": 7760 |
| }, |
| { |
| "epoch": 1.5918869084204057, |
| "grad_norm": 0.5918242931365967, |
| "learning_rate": 2.6558752532744346e-05, |
| "loss": 0.0132, |
| "step": 7770 |
| }, |
| { |
| "epoch": 1.5939356689203033, |
| "grad_norm": 0.40372413396835327, |
| "learning_rate": 2.654820730354377e-05, |
| "loss": 0.0129, |
| "step": 7780 |
| }, |
| { |
| "epoch": 1.5959844294202008, |
| "grad_norm": 0.39598509669303894, |
| "learning_rate": 2.6537648042182406e-05, |
| "loss": 0.0125, |
| "step": 7790 |
| }, |
| { |
| "epoch": 1.5980331899200984, |
| "grad_norm": 0.24353650212287903, |
| "learning_rate": 2.6527074761490735e-05, |
| "loss": 0.0145, |
| "step": 7800 |
| }, |
| { |
| "epoch": 1.600081950419996, |
| "grad_norm": 0.18843866884708405, |
| "learning_rate": 2.6516487474316303e-05, |
| "loss": 0.0138, |
| "step": 7810 |
| }, |
| { |
| "epoch": 1.6021307109198935, |
| "grad_norm": 0.18718990683555603, |
| "learning_rate": 2.650588619352365e-05, |
| "loss": 0.011, |
| "step": 7820 |
| }, |
| { |
| "epoch": 1.604179471419791, |
| "grad_norm": 0.24120080471038818, |
| "learning_rate": 2.649527093199434e-05, |
| "loss": 0.0118, |
| "step": 7830 |
| }, |
| { |
| "epoch": 1.6062282319196886, |
| "grad_norm": 0.23180128633975983, |
| "learning_rate": 2.6484641702626914e-05, |
| "loss": 0.0119, |
| "step": 7840 |
| }, |
| { |
| "epoch": 1.608276992419586, |
| "grad_norm": 0.3271956145763397, |
| "learning_rate": 2.647399851833689e-05, |
| "loss": 0.0183, |
| "step": 7850 |
| }, |
| { |
| "epoch": 1.6103257529194837, |
| "grad_norm": 0.20692825317382812, |
| "learning_rate": 2.6463341392056735e-05, |
| "loss": 0.0167, |
| "step": 7860 |
| }, |
| { |
| "epoch": 1.6123745134193812, |
| "grad_norm": 0.20381584763526917, |
| "learning_rate": 2.6452670336735867e-05, |
| "loss": 0.0142, |
| "step": 7870 |
| }, |
| { |
| "epoch": 1.6144232739192788, |
| "grad_norm": 0.3706916570663452, |
| "learning_rate": 2.644198536534063e-05, |
| "loss": 0.0155, |
| "step": 7880 |
| }, |
| { |
| "epoch": 1.6164720344191763, |
| "grad_norm": 0.15305541455745697, |
| "learning_rate": 2.6431286490854255e-05, |
| "loss": 0.0124, |
| "step": 7890 |
| }, |
| { |
| "epoch": 1.6185207949190739, |
| "grad_norm": 0.3070746958255768, |
| "learning_rate": 2.6420573726276903e-05, |
| "loss": 0.0109, |
| "step": 7900 |
| }, |
| { |
| "epoch": 1.6205695554189714, |
| "grad_norm": 0.17869031429290771, |
| "learning_rate": 2.6409847084625575e-05, |
| "loss": 0.0112, |
| "step": 7910 |
| }, |
| { |
| "epoch": 1.622618315918869, |
| "grad_norm": 0.37803635001182556, |
| "learning_rate": 2.6399106578934167e-05, |
| "loss": 0.0105, |
| "step": 7920 |
| }, |
| { |
| "epoch": 1.6246670764187665, |
| "grad_norm": 0.3635694980621338, |
| "learning_rate": 2.63883522222534e-05, |
| "loss": 0.0123, |
| "step": 7930 |
| }, |
| { |
| "epoch": 1.626715836918664, |
| "grad_norm": 0.32251060009002686, |
| "learning_rate": 2.6377584027650827e-05, |
| "loss": 0.0116, |
| "step": 7940 |
| }, |
| { |
| "epoch": 1.6287645974185616, |
| "grad_norm": 0.2724003195762634, |
| "learning_rate": 2.636680200821083e-05, |
| "loss": 0.0109, |
| "step": 7950 |
| }, |
| { |
| "epoch": 1.6308133579184594, |
| "grad_norm": 0.1469365805387497, |
| "learning_rate": 2.635600617703457e-05, |
| "loss": 0.0145, |
| "step": 7960 |
| }, |
| { |
| "epoch": 1.632862118418357, |
| "grad_norm": 0.199873685836792, |
| "learning_rate": 2.6345196547240007e-05, |
| "loss": 0.0122, |
| "step": 7970 |
| }, |
| { |
| "epoch": 1.6349108789182545, |
| "grad_norm": 0.17220766842365265, |
| "learning_rate": 2.6334373131961854e-05, |
| "loss": 0.0105, |
| "step": 7980 |
| }, |
| { |
| "epoch": 1.636959639418152, |
| "grad_norm": 0.3674171566963196, |
| "learning_rate": 2.6323535944351583e-05, |
| "loss": 0.0129, |
| "step": 7990 |
| }, |
| { |
| "epoch": 1.6390083999180496, |
| "grad_norm": 0.17417308688163757, |
| "learning_rate": 2.63126849975774e-05, |
| "loss": 0.0118, |
| "step": 8000 |
| }, |
| { |
| "epoch": 1.6410571604179471, |
| "grad_norm": 0.20215755701065063, |
| "learning_rate": 2.6301820304824234e-05, |
| "loss": 0.0112, |
| "step": 8010 |
| }, |
| { |
| "epoch": 1.6431059209178447, |
| "grad_norm": 0.2049969583749771, |
| "learning_rate": 2.6290941879293703e-05, |
| "loss": 0.0128, |
| "step": 8020 |
| }, |
| { |
| "epoch": 1.6451546814177422, |
| "grad_norm": 0.20151877403259277, |
| "learning_rate": 2.628004973420412e-05, |
| "loss": 0.013, |
| "step": 8030 |
| }, |
| { |
| "epoch": 1.6472034419176398, |
| "grad_norm": 0.43364572525024414, |
| "learning_rate": 2.626914388279047e-05, |
| "loss": 0.0126, |
| "step": 8040 |
| }, |
| { |
| "epoch": 1.6492522024175373, |
| "grad_norm": 0.28903061151504517, |
| "learning_rate": 2.6258224338304394e-05, |
| "loss": 0.012, |
| "step": 8050 |
| }, |
| { |
| "epoch": 1.6513009629174351, |
| "grad_norm": 0.14658614993095398, |
| "learning_rate": 2.6247291114014166e-05, |
| "loss": 0.014, |
| "step": 8060 |
| }, |
| { |
| "epoch": 1.6533497234173327, |
| "grad_norm": 0.5910778641700745, |
| "learning_rate": 2.6236344223204683e-05, |
| "loss": 0.0133, |
| "step": 8070 |
| }, |
| { |
| "epoch": 1.6553984839172302, |
| "grad_norm": 0.33621618151664734, |
| "learning_rate": 2.6225383679177453e-05, |
| "loss": 0.0111, |
| "step": 8080 |
| }, |
| { |
| "epoch": 1.6574472444171278, |
| "grad_norm": 0.2710290849208832, |
| "learning_rate": 2.6214409495250568e-05, |
| "loss": 0.0113, |
| "step": 8090 |
| }, |
| { |
| "epoch": 1.6594960049170253, |
| "grad_norm": 0.2991192042827606, |
| "learning_rate": 2.620342168475869e-05, |
| "loss": 0.0119, |
| "step": 8100 |
| }, |
| { |
| "epoch": 1.6615447654169229, |
| "grad_norm": 0.228810653090477, |
| "learning_rate": 2.619242026105306e-05, |
| "loss": 0.0123, |
| "step": 8110 |
| }, |
| { |
| "epoch": 1.6635935259168204, |
| "grad_norm": 0.1651410013437271, |
| "learning_rate": 2.618140523750143e-05, |
| "loss": 0.0109, |
| "step": 8120 |
| }, |
| { |
| "epoch": 1.665642286416718, |
| "grad_norm": 0.30259543657302856, |
| "learning_rate": 2.6170376627488103e-05, |
| "loss": 0.0147, |
| "step": 8130 |
| }, |
| { |
| "epoch": 1.6676910469166155, |
| "grad_norm": 0.3596413731575012, |
| "learning_rate": 2.6159334444413873e-05, |
| "loss": 0.0122, |
| "step": 8140 |
| }, |
| { |
| "epoch": 1.669739807416513, |
| "grad_norm": 0.3656299412250519, |
| "learning_rate": 2.614827870169604e-05, |
| "loss": 0.0128, |
| "step": 8150 |
| }, |
| { |
| "epoch": 1.6717885679164106, |
| "grad_norm": 0.6322624087333679, |
| "learning_rate": 2.6137209412768367e-05, |
| "loss": 0.0126, |
| "step": 8160 |
| }, |
| { |
| "epoch": 1.6738373284163082, |
| "grad_norm": 0.18824884295463562, |
| "learning_rate": 2.612612659108109e-05, |
| "loss": 0.0147, |
| "step": 8170 |
| }, |
| { |
| "epoch": 1.6758860889162057, |
| "grad_norm": 0.2941751480102539, |
| "learning_rate": 2.611503025010088e-05, |
| "loss": 0.0136, |
| "step": 8180 |
| }, |
| { |
| "epoch": 1.6779348494161033, |
| "grad_norm": 0.2813405990600586, |
| "learning_rate": 2.6103920403310832e-05, |
| "loss": 0.0113, |
| "step": 8190 |
| }, |
| { |
| "epoch": 1.6799836099160008, |
| "grad_norm": 0.20631378889083862, |
| "learning_rate": 2.6092797064210465e-05, |
| "loss": 0.0109, |
| "step": 8200 |
| }, |
| { |
| "epoch": 1.6820323704158984, |
| "grad_norm": 0.16025294363498688, |
| "learning_rate": 2.6081660246315685e-05, |
| "loss": 0.0152, |
| "step": 8210 |
| }, |
| { |
| "epoch": 1.684081130915796, |
| "grad_norm": 0.19440360367298126, |
| "learning_rate": 2.6070509963158774e-05, |
| "loss": 0.0149, |
| "step": 8220 |
| }, |
| { |
| "epoch": 1.6861298914156935, |
| "grad_norm": 0.6030908226966858, |
| "learning_rate": 2.6059346228288375e-05, |
| "loss": 0.0117, |
| "step": 8230 |
| }, |
| { |
| "epoch": 1.688178651915591, |
| "grad_norm": 0.4236409366130829, |
| "learning_rate": 2.604816905526948e-05, |
| "loss": 0.0127, |
| "step": 8240 |
| }, |
| { |
| "epoch": 1.6902274124154886, |
| "grad_norm": 0.2380395531654358, |
| "learning_rate": 2.603697845768341e-05, |
| "loss": 0.0118, |
| "step": 8250 |
| }, |
| { |
| "epoch": 1.6922761729153861, |
| "grad_norm": 0.18295350670814514, |
| "learning_rate": 2.6025774449127792e-05, |
| "loss": 0.0129, |
| "step": 8260 |
| }, |
| { |
| "epoch": 1.6943249334152837, |
| "grad_norm": 0.21567794680595398, |
| "learning_rate": 2.6014557043216562e-05, |
| "loss": 0.0109, |
| "step": 8270 |
| }, |
| { |
| "epoch": 1.6963736939151812, |
| "grad_norm": 0.23002085089683533, |
| "learning_rate": 2.6003326253579913e-05, |
| "loss": 0.015, |
| "step": 8280 |
| }, |
| { |
| "epoch": 1.6984224544150788, |
| "grad_norm": 0.22134102880954742, |
| "learning_rate": 2.599208209386432e-05, |
| "loss": 0.0131, |
| "step": 8290 |
| }, |
| { |
| "epoch": 1.7004712149149763, |
| "grad_norm": 0.2828630805015564, |
| "learning_rate": 2.5980824577732502e-05, |
| "loss": 0.0116, |
| "step": 8300 |
| }, |
| { |
| "epoch": 1.7025199754148739, |
| "grad_norm": 0.36930739879608154, |
| "learning_rate": 2.5969553718863395e-05, |
| "loss": 0.0124, |
| "step": 8310 |
| }, |
| { |
| "epoch": 1.7045687359147714, |
| "grad_norm": 0.5323567986488342, |
| "learning_rate": 2.595826953095216e-05, |
| "loss": 0.0128, |
| "step": 8320 |
| }, |
| { |
| "epoch": 1.706617496414669, |
| "grad_norm": 0.21462911367416382, |
| "learning_rate": 2.594697202771014e-05, |
| "loss": 0.0119, |
| "step": 8330 |
| }, |
| { |
| "epoch": 1.7086662569145665, |
| "grad_norm": 0.16089050471782684, |
| "learning_rate": 2.5935661222864888e-05, |
| "loss": 0.01, |
| "step": 8340 |
| }, |
| { |
| "epoch": 1.7107150174144643, |
| "grad_norm": 0.12649470567703247, |
| "learning_rate": 2.592433713016008e-05, |
| "loss": 0.0128, |
| "step": 8350 |
| }, |
| { |
| "epoch": 1.7127637779143619, |
| "grad_norm": 0.15089091658592224, |
| "learning_rate": 2.5912999763355553e-05, |
| "loss": 0.0124, |
| "step": 8360 |
| }, |
| { |
| "epoch": 1.7148125384142594, |
| "grad_norm": 0.29391559958457947, |
| "learning_rate": 2.590164913622729e-05, |
| "loss": 0.013, |
| "step": 8370 |
| }, |
| { |
| "epoch": 1.716861298914157, |
| "grad_norm": 0.1641857922077179, |
| "learning_rate": 2.589028526256737e-05, |
| "loss": 0.0125, |
| "step": 8380 |
| }, |
| { |
| "epoch": 1.7189100594140545, |
| "grad_norm": 0.4624747037887573, |
| "learning_rate": 2.587890815618397e-05, |
| "loss": 0.0114, |
| "step": 8390 |
| }, |
| { |
| "epoch": 1.720958819913952, |
| "grad_norm": 0.16161824762821198, |
| "learning_rate": 2.5867517830901343e-05, |
| "loss": 0.0122, |
| "step": 8400 |
| }, |
| { |
| "epoch": 1.7230075804138496, |
| "grad_norm": 0.12558385729789734, |
| "learning_rate": 2.5856114300559812e-05, |
| "loss": 0.0127, |
| "step": 8410 |
| }, |
| { |
| "epoch": 1.7250563409137472, |
| "grad_norm": 0.2348787486553192, |
| "learning_rate": 2.584469757901574e-05, |
| "loss": 0.0137, |
| "step": 8420 |
| }, |
| { |
| "epoch": 1.7271051014136447, |
| "grad_norm": 0.4634511470794678, |
| "learning_rate": 2.5833267680141526e-05, |
| "loss": 0.0131, |
| "step": 8430 |
| }, |
| { |
| "epoch": 1.7291538619135423, |
| "grad_norm": 0.3090566396713257, |
| "learning_rate": 2.5821824617825564e-05, |
| "loss": 0.0141, |
| "step": 8440 |
| }, |
| { |
| "epoch": 1.73120262241344, |
| "grad_norm": 0.26835501194000244, |
| "learning_rate": 2.5810368405972265e-05, |
| "loss": 0.0128, |
| "step": 8450 |
| }, |
| { |
| "epoch": 1.7332513829133376, |
| "grad_norm": 0.13316889107227325, |
| "learning_rate": 2.5798899058501996e-05, |
| "loss": 0.0099, |
| "step": 8460 |
| }, |
| { |
| "epoch": 1.7353001434132351, |
| "grad_norm": 0.32415223121643066, |
| "learning_rate": 2.578741658935111e-05, |
| "loss": 0.014, |
| "step": 8470 |
| }, |
| { |
| "epoch": 1.7373489039131327, |
| "grad_norm": 0.20000790059566498, |
| "learning_rate": 2.5775921012471875e-05, |
| "loss": 0.0132, |
| "step": 8480 |
| }, |
| { |
| "epoch": 1.7393976644130302, |
| "grad_norm": 0.37306877970695496, |
| "learning_rate": 2.5764412341832505e-05, |
| "loss": 0.0136, |
| "step": 8490 |
| }, |
| { |
| "epoch": 1.7414464249129278, |
| "grad_norm": 0.19551271200180054, |
| "learning_rate": 2.5752890591417125e-05, |
| "loss": 0.0147, |
| "step": 8500 |
| }, |
| { |
| "epoch": 1.7434951854128253, |
| "grad_norm": 0.19909143447875977, |
| "learning_rate": 2.5741355775225744e-05, |
| "loss": 0.0158, |
| "step": 8510 |
| }, |
| { |
| "epoch": 1.7455439459127229, |
| "grad_norm": 0.20295865833759308, |
| "learning_rate": 2.5729807907274254e-05, |
| "loss": 0.0119, |
| "step": 8520 |
| }, |
| { |
| "epoch": 1.7475927064126204, |
| "grad_norm": 0.16480283439159393, |
| "learning_rate": 2.5718247001594403e-05, |
| "loss": 0.0125, |
| "step": 8530 |
| }, |
| { |
| "epoch": 1.749641466912518, |
| "grad_norm": 0.13466161489486694, |
| "learning_rate": 2.5706673072233776e-05, |
| "loss": 0.0103, |
| "step": 8540 |
| }, |
| { |
| "epoch": 1.7516902274124155, |
| "grad_norm": 0.17151710391044617, |
| "learning_rate": 2.5695086133255795e-05, |
| "loss": 0.0108, |
| "step": 8550 |
| }, |
| { |
| "epoch": 1.753738987912313, |
| "grad_norm": 0.16230301558971405, |
| "learning_rate": 2.5683486198739687e-05, |
| "loss": 0.0119, |
| "step": 8560 |
| }, |
| { |
| "epoch": 1.7557877484122106, |
| "grad_norm": 0.226009801030159, |
| "learning_rate": 2.5671873282780452e-05, |
| "loss": 0.0109, |
| "step": 8570 |
| }, |
| { |
| "epoch": 1.7578365089121082, |
| "grad_norm": 0.23014304041862488, |
| "learning_rate": 2.5660247399488887e-05, |
| "loss": 0.0142, |
| "step": 8580 |
| }, |
| { |
| "epoch": 1.7598852694120057, |
| "grad_norm": 0.13345304131507874, |
| "learning_rate": 2.564860856299154e-05, |
| "loss": 0.012, |
| "step": 8590 |
| }, |
| { |
| "epoch": 1.7619340299119033, |
| "grad_norm": 0.30926987528800964, |
| "learning_rate": 2.5636956787430693e-05, |
| "loss": 0.0101, |
| "step": 8600 |
| }, |
| { |
| "epoch": 1.7639827904118008, |
| "grad_norm": 0.3453812599182129, |
| "learning_rate": 2.5625292086964346e-05, |
| "loss": 0.0124, |
| "step": 8610 |
| }, |
| { |
| "epoch": 1.7660315509116984, |
| "grad_norm": 0.37684300541877747, |
| "learning_rate": 2.5613614475766212e-05, |
| "loss": 0.0124, |
| "step": 8620 |
| }, |
| { |
| "epoch": 1.768080311411596, |
| "grad_norm": 0.3058297634124756, |
| "learning_rate": 2.5601923968025694e-05, |
| "loss": 0.0133, |
| "step": 8630 |
| }, |
| { |
| "epoch": 1.7701290719114935, |
| "grad_norm": 0.3661828935146332, |
| "learning_rate": 2.5590220577947862e-05, |
| "loss": 0.014, |
| "step": 8640 |
| }, |
| { |
| "epoch": 1.772177832411391, |
| "grad_norm": 0.1543501615524292, |
| "learning_rate": 2.5578504319753427e-05, |
| "loss": 0.0116, |
| "step": 8650 |
| }, |
| { |
| "epoch": 1.7742265929112886, |
| "grad_norm": 0.1778232753276825, |
| "learning_rate": 2.556677520767876e-05, |
| "loss": 0.0126, |
| "step": 8660 |
| }, |
| { |
| "epoch": 1.7762753534111861, |
| "grad_norm": 0.22795453667640686, |
| "learning_rate": 2.5555033255975835e-05, |
| "loss": 0.0123, |
| "step": 8670 |
| }, |
| { |
| "epoch": 1.7783241139110837, |
| "grad_norm": 0.18020279705524445, |
| "learning_rate": 2.5543278478912228e-05, |
| "loss": 0.0115, |
| "step": 8680 |
| }, |
| { |
| "epoch": 1.7803728744109812, |
| "grad_norm": 0.3434680700302124, |
| "learning_rate": 2.55315108907711e-05, |
| "loss": 0.014, |
| "step": 8690 |
| }, |
| { |
| "epoch": 1.7824216349108788, |
| "grad_norm": 0.33293616771698, |
| "learning_rate": 2.551973050585119e-05, |
| "loss": 0.0124, |
| "step": 8700 |
| }, |
| { |
| "epoch": 1.7844703954107763, |
| "grad_norm": 0.26626941561698914, |
| "learning_rate": 2.5507937338466768e-05, |
| "loss": 0.0133, |
| "step": 8710 |
| }, |
| { |
| "epoch": 1.786519155910674, |
| "grad_norm": 0.2310871183872223, |
| "learning_rate": 2.549613140294764e-05, |
| "loss": 0.0108, |
| "step": 8720 |
| }, |
| { |
| "epoch": 1.7885679164105714, |
| "grad_norm": 0.15284205973148346, |
| "learning_rate": 2.548431271363914e-05, |
| "loss": 0.0139, |
| "step": 8730 |
| }, |
| { |
| "epoch": 1.7906166769104692, |
| "grad_norm": 0.4070996642112732, |
| "learning_rate": 2.5472481284902092e-05, |
| "loss": 0.0136, |
| "step": 8740 |
| }, |
| { |
| "epoch": 1.7926654374103668, |
| "grad_norm": 0.43006783723831177, |
| "learning_rate": 2.5460637131112788e-05, |
| "loss": 0.0132, |
| "step": 8750 |
| }, |
| { |
| "epoch": 1.7947141979102643, |
| "grad_norm": 0.31472328305244446, |
| "learning_rate": 2.5448780266662995e-05, |
| "loss": 0.0109, |
| "step": 8760 |
| }, |
| { |
| "epoch": 1.7967629584101619, |
| "grad_norm": 0.20050498843193054, |
| "learning_rate": 2.5436910705959926e-05, |
| "loss": 0.0105, |
| "step": 8770 |
| }, |
| { |
| "epoch": 1.7988117189100594, |
| "grad_norm": 0.30332106351852417, |
| "learning_rate": 2.542502846342621e-05, |
| "loss": 0.0121, |
| "step": 8780 |
| }, |
| { |
| "epoch": 1.800860479409957, |
| "grad_norm": 0.34002912044525146, |
| "learning_rate": 2.5413133553499892e-05, |
| "loss": 0.0141, |
| "step": 8790 |
| }, |
| { |
| "epoch": 1.8029092399098545, |
| "grad_norm": 0.16000936925411224, |
| "learning_rate": 2.540122599063442e-05, |
| "loss": 0.0136, |
| "step": 8800 |
| }, |
| { |
| "epoch": 1.804958000409752, |
| "grad_norm": 0.23254422843456268, |
| "learning_rate": 2.538930578929859e-05, |
| "loss": 0.0119, |
| "step": 8810 |
| }, |
| { |
| "epoch": 1.8070067609096496, |
| "grad_norm": 0.28054049611091614, |
| "learning_rate": 2.537737296397658e-05, |
| "loss": 0.0122, |
| "step": 8820 |
| }, |
| { |
| "epoch": 1.8090555214095472, |
| "grad_norm": 0.27173829078674316, |
| "learning_rate": 2.5365427529167903e-05, |
| "loss": 0.0124, |
| "step": 8830 |
| }, |
| { |
| "epoch": 1.811104281909445, |
| "grad_norm": 0.2194901704788208, |
| "learning_rate": 2.5353469499387375e-05, |
| "loss": 0.0125, |
| "step": 8840 |
| }, |
| { |
| "epoch": 1.8131530424093425, |
| "grad_norm": 0.20670931041240692, |
| "learning_rate": 2.534149888916514e-05, |
| "loss": 0.0116, |
| "step": 8850 |
| }, |
| { |
| "epoch": 1.81520180290924, |
| "grad_norm": 0.31378597021102905, |
| "learning_rate": 2.5329515713046616e-05, |
| "loss": 0.0131, |
| "step": 8860 |
| }, |
| { |
| "epoch": 1.8172505634091376, |
| "grad_norm": 0.2812603712081909, |
| "learning_rate": 2.531751998559249e-05, |
| "loss": 0.0131, |
| "step": 8870 |
| }, |
| { |
| "epoch": 1.8192993239090351, |
| "grad_norm": 0.1855655014514923, |
| "learning_rate": 2.5305511721378703e-05, |
| "loss": 0.0105, |
| "step": 8880 |
| }, |
| { |
| "epoch": 1.8213480844089327, |
| "grad_norm": 0.15127666294574738, |
| "learning_rate": 2.529349093499642e-05, |
| "loss": 0.0113, |
| "step": 8890 |
| }, |
| { |
| "epoch": 1.8233968449088302, |
| "grad_norm": 0.376337468624115, |
| "learning_rate": 2.528145764105205e-05, |
| "loss": 0.0113, |
| "step": 8900 |
| }, |
| { |
| "epoch": 1.8254456054087278, |
| "grad_norm": 0.2767731249332428, |
| "learning_rate": 2.5269411854167166e-05, |
| "loss": 0.0136, |
| "step": 8910 |
| }, |
| { |
| "epoch": 1.8274943659086254, |
| "grad_norm": 0.6750398874282837, |
| "learning_rate": 2.5257353588978534e-05, |
| "loss": 0.0129, |
| "step": 8920 |
| }, |
| { |
| "epoch": 1.829543126408523, |
| "grad_norm": 0.3173629939556122, |
| "learning_rate": 2.5245282860138083e-05, |
| "loss": 0.0109, |
| "step": 8930 |
| }, |
| { |
| "epoch": 1.8315918869084205, |
| "grad_norm": 0.22351804375648499, |
| "learning_rate": 2.5233199682312894e-05, |
| "loss": 0.0133, |
| "step": 8940 |
| }, |
| { |
| "epoch": 1.833640647408318, |
| "grad_norm": 0.17554111778736115, |
| "learning_rate": 2.5221104070185163e-05, |
| "loss": 0.0117, |
| "step": 8950 |
| }, |
| { |
| "epoch": 1.8356894079082156, |
| "grad_norm": 0.16689617931842804, |
| "learning_rate": 2.5208996038452204e-05, |
| "loss": 0.0131, |
| "step": 8960 |
| }, |
| { |
| "epoch": 1.837738168408113, |
| "grad_norm": 0.21691209077835083, |
| "learning_rate": 2.5196875601826403e-05, |
| "loss": 0.0133, |
| "step": 8970 |
| }, |
| { |
| "epoch": 1.8397869289080107, |
| "grad_norm": 0.2426990270614624, |
| "learning_rate": 2.518474277503525e-05, |
| "loss": 0.0113, |
| "step": 8980 |
| }, |
| { |
| "epoch": 1.8418356894079082, |
| "grad_norm": 0.1473204344511032, |
| "learning_rate": 2.5172597572821266e-05, |
| "loss": 0.0111, |
| "step": 8990 |
| }, |
| { |
| "epoch": 1.8438844499078058, |
| "grad_norm": 0.2578393518924713, |
| "learning_rate": 2.5160440009942015e-05, |
| "loss": 0.0109, |
| "step": 9000 |
| }, |
| { |
| "epoch": 1.8459332104077033, |
| "grad_norm": 0.5229294896125793, |
| "learning_rate": 2.514827010117009e-05, |
| "loss": 0.0155, |
| "step": 9010 |
| }, |
| { |
| "epoch": 1.8479819709076009, |
| "grad_norm": 0.35728609561920166, |
| "learning_rate": 2.5136087861293068e-05, |
| "loss": 0.0127, |
| "step": 9020 |
| }, |
| { |
| "epoch": 1.8500307314074984, |
| "grad_norm": 0.20786592364311218, |
| "learning_rate": 2.5123893305113525e-05, |
| "loss": 0.0123, |
| "step": 9030 |
| }, |
| { |
| "epoch": 1.852079491907396, |
| "grad_norm": 0.2943900525569916, |
| "learning_rate": 2.5111686447448997e-05, |
| "loss": 0.0118, |
| "step": 9040 |
| }, |
| { |
| "epoch": 1.8541282524072935, |
| "grad_norm": 0.266979843378067, |
| "learning_rate": 2.5099467303131966e-05, |
| "loss": 0.0152, |
| "step": 9050 |
| }, |
| { |
| "epoch": 1.856177012907191, |
| "grad_norm": 0.45919889211654663, |
| "learning_rate": 2.508723588700984e-05, |
| "loss": 0.0128, |
| "step": 9060 |
| }, |
| { |
| "epoch": 1.8582257734070886, |
| "grad_norm": 0.6927161812782288, |
| "learning_rate": 2.5074992213944952e-05, |
| "loss": 0.0144, |
| "step": 9070 |
| }, |
| { |
| "epoch": 1.8602745339069862, |
| "grad_norm": 0.30469781160354614, |
| "learning_rate": 2.506273629881451e-05, |
| "loss": 0.0138, |
| "step": 9080 |
| }, |
| { |
| "epoch": 1.8623232944068837, |
| "grad_norm": 0.1494228094816208, |
| "learning_rate": 2.5050468156510617e-05, |
| "loss": 0.0109, |
| "step": 9090 |
| }, |
| { |
| "epoch": 1.8643720549067813, |
| "grad_norm": 0.2666446566581726, |
| "learning_rate": 2.5038187801940215e-05, |
| "loss": 0.0113, |
| "step": 9100 |
| }, |
| { |
| "epoch": 1.8664208154066788, |
| "grad_norm": 0.16080419719219208, |
| "learning_rate": 2.5025895250025088e-05, |
| "loss": 0.0137, |
| "step": 9110 |
| }, |
| { |
| "epoch": 1.8684695759065764, |
| "grad_norm": 0.20594751834869385, |
| "learning_rate": 2.501359051570186e-05, |
| "loss": 0.0101, |
| "step": 9120 |
| }, |
| { |
| "epoch": 1.8705183364064741, |
| "grad_norm": 0.17402420938014984, |
| "learning_rate": 2.500127361392193e-05, |
| "loss": 0.0127, |
| "step": 9130 |
| }, |
| { |
| "epoch": 1.8725670969063717, |
| "grad_norm": 0.16892752051353455, |
| "learning_rate": 2.49889445596515e-05, |
| "loss": 0.0173, |
| "step": 9140 |
| }, |
| { |
| "epoch": 1.8746158574062692, |
| "grad_norm": 0.19873949885368347, |
| "learning_rate": 2.497660336787154e-05, |
| "loss": 0.0105, |
| "step": 9150 |
| }, |
| { |
| "epoch": 1.8766646179061668, |
| "grad_norm": 0.27695828676223755, |
| "learning_rate": 2.496425005357775e-05, |
| "loss": 0.0117, |
| "step": 9160 |
| }, |
| { |
| "epoch": 1.8787133784060643, |
| "grad_norm": 0.3204931914806366, |
| "learning_rate": 2.4951884631780578e-05, |
| "loss": 0.0116, |
| "step": 9170 |
| }, |
| { |
| "epoch": 1.8807621389059619, |
| "grad_norm": 0.16970880329608917, |
| "learning_rate": 2.493950711750518e-05, |
| "loss": 0.0137, |
| "step": 9180 |
| }, |
| { |
| "epoch": 1.8828108994058594, |
| "grad_norm": 0.3381524980068207, |
| "learning_rate": 2.49271175257914e-05, |
| "loss": 0.0107, |
| "step": 9190 |
| }, |
| { |
| "epoch": 1.884859659905757, |
| "grad_norm": 0.7062802314758301, |
| "learning_rate": 2.4914715871693757e-05, |
| "loss": 0.0125, |
| "step": 9200 |
| }, |
| { |
| "epoch": 1.8869084204056545, |
| "grad_norm": 0.24380052089691162, |
| "learning_rate": 2.4902302170281438e-05, |
| "loss": 0.0114, |
| "step": 9210 |
| }, |
| { |
| "epoch": 1.888957180905552, |
| "grad_norm": 0.39946651458740234, |
| "learning_rate": 2.4889876436638252e-05, |
| "loss": 0.0131, |
| "step": 9220 |
| }, |
| { |
| "epoch": 1.8910059414054499, |
| "grad_norm": 0.38546937704086304, |
| "learning_rate": 2.4877438685862643e-05, |
| "loss": 0.0138, |
| "step": 9230 |
| }, |
| { |
| "epoch": 1.8930547019053474, |
| "grad_norm": 0.23303157091140747, |
| "learning_rate": 2.486498893306765e-05, |
| "loss": 0.0108, |
| "step": 9240 |
| }, |
| { |
| "epoch": 1.895103462405245, |
| "grad_norm": 0.24158629775047302, |
| "learning_rate": 2.485252719338089e-05, |
| "loss": 0.0148, |
| "step": 9250 |
| }, |
| { |
| "epoch": 1.8971522229051425, |
| "grad_norm": 0.32543596625328064, |
| "learning_rate": 2.4840053481944563e-05, |
| "loss": 0.0124, |
| "step": 9260 |
| }, |
| { |
| "epoch": 1.89920098340504, |
| "grad_norm": 0.18101559579372406, |
| "learning_rate": 2.4827567813915398e-05, |
| "loss": 0.0109, |
| "step": 9270 |
| }, |
| { |
| "epoch": 1.9012497439049376, |
| "grad_norm": 0.5253106951713562, |
| "learning_rate": 2.481507020446466e-05, |
| "loss": 0.01, |
| "step": 9280 |
| }, |
| { |
| "epoch": 1.9032985044048352, |
| "grad_norm": 0.33761027455329895, |
| "learning_rate": 2.4802560668778122e-05, |
| "loss": 0.0132, |
| "step": 9290 |
| }, |
| { |
| "epoch": 1.9053472649047327, |
| "grad_norm": 0.18968777358531952, |
| "learning_rate": 2.479003922205605e-05, |
| "loss": 0.0126, |
| "step": 9300 |
| }, |
| { |
| "epoch": 1.9073960254046303, |
| "grad_norm": 0.16011153161525726, |
| "learning_rate": 2.4777505879513183e-05, |
| "loss": 0.01, |
| "step": 9310 |
| }, |
| { |
| "epoch": 1.9094447859045278, |
| "grad_norm": 0.22958874702453613, |
| "learning_rate": 2.4764960656378714e-05, |
| "loss": 0.0133, |
| "step": 9320 |
| }, |
| { |
| "epoch": 1.9114935464044254, |
| "grad_norm": 0.3954741358757019, |
| "learning_rate": 2.4752403567896274e-05, |
| "loss": 0.0115, |
| "step": 9330 |
| }, |
| { |
| "epoch": 1.913542306904323, |
| "grad_norm": 0.21476221084594727, |
| "learning_rate": 2.47398346293239e-05, |
| "loss": 0.0108, |
| "step": 9340 |
| }, |
| { |
| "epoch": 1.9155910674042205, |
| "grad_norm": 0.5305155515670776, |
| "learning_rate": 2.4727253855934055e-05, |
| "loss": 0.012, |
| "step": 9350 |
| }, |
| { |
| "epoch": 1.917639827904118, |
| "grad_norm": 0.1582081913948059, |
| "learning_rate": 2.4714661263013548e-05, |
| "loss": 0.0145, |
| "step": 9360 |
| }, |
| { |
| "epoch": 1.9196885884040156, |
| "grad_norm": 0.25267645716667175, |
| "learning_rate": 2.4702056865863575e-05, |
| "loss": 0.0123, |
| "step": 9370 |
| }, |
| { |
| "epoch": 1.9217373489039131, |
| "grad_norm": 0.23755228519439697, |
| "learning_rate": 2.4689440679799672e-05, |
| "loss": 0.0129, |
| "step": 9380 |
| }, |
| { |
| "epoch": 1.9237861094038107, |
| "grad_norm": 0.3205947279930115, |
| "learning_rate": 2.4676812720151678e-05, |
| "loss": 0.0084, |
| "step": 9390 |
| }, |
| { |
| "epoch": 1.9258348699037082, |
| "grad_norm": 0.13462692499160767, |
| "learning_rate": 2.466417300226377e-05, |
| "loss": 0.0121, |
| "step": 9400 |
| }, |
| { |
| "epoch": 1.9278836304036058, |
| "grad_norm": 0.1858198195695877, |
| "learning_rate": 2.4651521541494394e-05, |
| "loss": 0.012, |
| "step": 9410 |
| }, |
| { |
| "epoch": 1.9299323909035033, |
| "grad_norm": 0.4120533764362335, |
| "learning_rate": 2.4638858353216267e-05, |
| "loss": 0.0125, |
| "step": 9420 |
| }, |
| { |
| "epoch": 1.9319811514034009, |
| "grad_norm": 0.16376164555549622, |
| "learning_rate": 2.4626183452816355e-05, |
| "loss": 0.0132, |
| "step": 9430 |
| }, |
| { |
| "epoch": 1.9340299119032984, |
| "grad_norm": 0.29123422503471375, |
| "learning_rate": 2.4613496855695855e-05, |
| "loss": 0.0142, |
| "step": 9440 |
| }, |
| { |
| "epoch": 1.936078672403196, |
| "grad_norm": 0.19462132453918457, |
| "learning_rate": 2.4600798577270183e-05, |
| "loss": 0.0123, |
| "step": 9450 |
| }, |
| { |
| "epoch": 1.9381274329030935, |
| "grad_norm": 0.8115758895874023, |
| "learning_rate": 2.4588088632968944e-05, |
| "loss": 0.0147, |
| "step": 9460 |
| }, |
| { |
| "epoch": 1.940176193402991, |
| "grad_norm": 0.47200286388397217, |
| "learning_rate": 2.4575367038235916e-05, |
| "loss": 0.0122, |
| "step": 9470 |
| }, |
| { |
| "epoch": 1.9422249539028886, |
| "grad_norm": 0.1522137075662613, |
| "learning_rate": 2.4562633808529043e-05, |
| "loss": 0.0146, |
| "step": 9480 |
| }, |
| { |
| "epoch": 1.9442737144027862, |
| "grad_norm": 0.8687599301338196, |
| "learning_rate": 2.4549888959320396e-05, |
| "loss": 0.0132, |
| "step": 9490 |
| }, |
| { |
| "epoch": 1.9463224749026837, |
| "grad_norm": 0.19402065873146057, |
| "learning_rate": 2.4537132506096162e-05, |
| "loss": 0.0147, |
| "step": 9500 |
| }, |
| { |
| "epoch": 1.9483712354025813, |
| "grad_norm": 0.29472532868385315, |
| "learning_rate": 2.452436446435664e-05, |
| "loss": 0.0126, |
| "step": 9510 |
| }, |
| { |
| "epoch": 1.950419995902479, |
| "grad_norm": 0.34857723116874695, |
| "learning_rate": 2.451158484961621e-05, |
| "loss": 0.018, |
| "step": 9520 |
| }, |
| { |
| "epoch": 1.9524687564023766, |
| "grad_norm": 0.2275514304637909, |
| "learning_rate": 2.4498793677403304e-05, |
| "loss": 0.0109, |
| "step": 9530 |
| }, |
| { |
| "epoch": 1.9545175169022742, |
| "grad_norm": 0.1869262009859085, |
| "learning_rate": 2.4485990963260398e-05, |
| "loss": 0.0124, |
| "step": 9540 |
| }, |
| { |
| "epoch": 1.9565662774021717, |
| "grad_norm": 0.3066518008708954, |
| "learning_rate": 2.4473176722744e-05, |
| "loss": 0.012, |
| "step": 9550 |
| }, |
| { |
| "epoch": 1.9586150379020693, |
| "grad_norm": 0.184419646859169, |
| "learning_rate": 2.446035097142463e-05, |
| "loss": 0.0151, |
| "step": 9560 |
| }, |
| { |
| "epoch": 1.9606637984019668, |
| "grad_norm": 0.1928374320268631, |
| "learning_rate": 2.444751372488677e-05, |
| "loss": 0.013, |
| "step": 9570 |
| }, |
| { |
| "epoch": 1.9627125589018644, |
| "grad_norm": 0.38961443305015564, |
| "learning_rate": 2.443466499872889e-05, |
| "loss": 0.0129, |
| "step": 9580 |
| }, |
| { |
| "epoch": 1.964761319401762, |
| "grad_norm": 0.19305481016635895, |
| "learning_rate": 2.4421804808563413e-05, |
| "loss": 0.0145, |
| "step": 9590 |
| }, |
| { |
| "epoch": 1.9668100799016595, |
| "grad_norm": 0.26148465275764465, |
| "learning_rate": 2.440893317001667e-05, |
| "loss": 0.0134, |
| "step": 9600 |
| }, |
| { |
| "epoch": 1.968858840401557, |
| "grad_norm": 0.4117526412010193, |
| "learning_rate": 2.4396050098728925e-05, |
| "loss": 0.0145, |
| "step": 9610 |
| }, |
| { |
| "epoch": 1.9709076009014548, |
| "grad_norm": 0.2972826659679413, |
| "learning_rate": 2.4383155610354314e-05, |
| "loss": 0.0142, |
| "step": 9620 |
| }, |
| { |
| "epoch": 1.9729563614013523, |
| "grad_norm": 0.39025503396987915, |
| "learning_rate": 2.437024972056087e-05, |
| "loss": 0.0163, |
| "step": 9630 |
| }, |
| { |
| "epoch": 1.9750051219012499, |
| "grad_norm": 0.1266939640045166, |
| "learning_rate": 2.4357332445030452e-05, |
| "loss": 0.0103, |
| "step": 9640 |
| }, |
| { |
| "epoch": 1.9770538824011474, |
| "grad_norm": 0.7581406235694885, |
| "learning_rate": 2.434440379945878e-05, |
| "loss": 0.0129, |
| "step": 9650 |
| }, |
| { |
| "epoch": 1.979102642901045, |
| "grad_norm": 0.18585029244422913, |
| "learning_rate": 2.4331463799555366e-05, |
| "loss": 0.0114, |
| "step": 9660 |
| }, |
| { |
| "epoch": 1.9811514034009425, |
| "grad_norm": 0.38119348883628845, |
| "learning_rate": 2.4318512461043544e-05, |
| "loss": 0.0114, |
| "step": 9670 |
| }, |
| { |
| "epoch": 1.98320016390084, |
| "grad_norm": 0.16554845869541168, |
| "learning_rate": 2.4305549799660397e-05, |
| "loss": 0.0126, |
| "step": 9680 |
| }, |
| { |
| "epoch": 1.9852489244007376, |
| "grad_norm": 0.3252345323562622, |
| "learning_rate": 2.429257583115679e-05, |
| "loss": 0.013, |
| "step": 9690 |
| }, |
| { |
| "epoch": 1.9872976849006352, |
| "grad_norm": 0.2885255813598633, |
| "learning_rate": 2.4279590571297316e-05, |
| "loss": 0.0134, |
| "step": 9700 |
| }, |
| { |
| "epoch": 1.9893464454005327, |
| "grad_norm": 0.3435351848602295, |
| "learning_rate": 2.4266594035860288e-05, |
| "loss": 0.0104, |
| "step": 9710 |
| }, |
| { |
| "epoch": 1.9913952059004303, |
| "grad_norm": 0.23861199617385864, |
| "learning_rate": 2.425358624063773e-05, |
| "loss": 0.0118, |
| "step": 9720 |
| }, |
| { |
| "epoch": 1.9934439664003278, |
| "grad_norm": 0.2492307424545288, |
| "learning_rate": 2.424056720143532e-05, |
| "loss": 0.0142, |
| "step": 9730 |
| }, |
| { |
| "epoch": 1.9954927269002254, |
| "grad_norm": 0.2493637651205063, |
| "learning_rate": 2.422753693407244e-05, |
| "loss": 0.0127, |
| "step": 9740 |
| }, |
| { |
| "epoch": 1.997541487400123, |
| "grad_norm": 0.20873196423053741, |
| "learning_rate": 2.421449545438209e-05, |
| "loss": 0.0108, |
| "step": 9750 |
| }, |
| { |
| "epoch": 1.9995902479000205, |
| "grad_norm": 0.2116081863641739, |
| "learning_rate": 2.4201442778210887e-05, |
| "loss": 0.0146, |
| "step": 9760 |
| }, |
| { |
| "epoch": 2.001639008399918, |
| "grad_norm": 0.13597454130649567, |
| "learning_rate": 2.418837892141907e-05, |
| "loss": 0.0104, |
| "step": 9770 |
| }, |
| { |
| "epoch": 2.0036877688998156, |
| "grad_norm": 0.18827858567237854, |
| "learning_rate": 2.4175303899880458e-05, |
| "loss": 0.0127, |
| "step": 9780 |
| }, |
| { |
| "epoch": 2.005736529399713, |
| "grad_norm": 0.4803963601589203, |
| "learning_rate": 2.4162217729482427e-05, |
| "loss": 0.0117, |
| "step": 9790 |
| }, |
| { |
| "epoch": 2.0077852898996107, |
| "grad_norm": 0.2918669879436493, |
| "learning_rate": 2.4149120426125918e-05, |
| "loss": 0.014, |
| "step": 9800 |
| }, |
| { |
| "epoch": 2.0098340503995082, |
| "grad_norm": 0.40370047092437744, |
| "learning_rate": 2.413601200572538e-05, |
| "loss": 0.0144, |
| "step": 9810 |
| }, |
| { |
| "epoch": 2.011882810899406, |
| "grad_norm": 0.6413276791572571, |
| "learning_rate": 2.4122892484208794e-05, |
| "loss": 0.0134, |
| "step": 9820 |
| }, |
| { |
| "epoch": 2.0139315713993033, |
| "grad_norm": 0.22806483507156372, |
| "learning_rate": 2.4109761877517603e-05, |
| "loss": 0.0123, |
| "step": 9830 |
| }, |
| { |
| "epoch": 2.015980331899201, |
| "grad_norm": 0.20831595361232758, |
| "learning_rate": 2.4096620201606737e-05, |
| "loss": 0.012, |
| "step": 9840 |
| }, |
| { |
| "epoch": 2.0180290923990984, |
| "grad_norm": 0.1265106350183487, |
| "learning_rate": 2.4083467472444573e-05, |
| "loss": 0.0124, |
| "step": 9850 |
| }, |
| { |
| "epoch": 2.020077852898996, |
| "grad_norm": 0.19140085577964783, |
| "learning_rate": 2.4070303706012912e-05, |
| "loss": 0.0113, |
| "step": 9860 |
| }, |
| { |
| "epoch": 2.0221266133988935, |
| "grad_norm": 0.1588415950536728, |
| "learning_rate": 2.4057128918306975e-05, |
| "loss": 0.0123, |
| "step": 9870 |
| }, |
| { |
| "epoch": 2.024175373898791, |
| "grad_norm": 0.1528339385986328, |
| "learning_rate": 2.4043943125335373e-05, |
| "loss": 0.0132, |
| "step": 9880 |
| }, |
| { |
| "epoch": 2.0262241343986886, |
| "grad_norm": 0.17474709451198578, |
| "learning_rate": 2.403074634312008e-05, |
| "loss": 0.0158, |
| "step": 9890 |
| }, |
| { |
| "epoch": 2.028272894898586, |
| "grad_norm": 0.1128094419836998, |
| "learning_rate": 2.4017538587696442e-05, |
| "loss": 0.0127, |
| "step": 9900 |
| }, |
| { |
| "epoch": 2.0303216553984837, |
| "grad_norm": 0.19044803082942963, |
| "learning_rate": 2.4004319875113118e-05, |
| "loss": 0.0121, |
| "step": 9910 |
| }, |
| { |
| "epoch": 2.0323704158983813, |
| "grad_norm": 0.18360526859760284, |
| "learning_rate": 2.3991090221432096e-05, |
| "loss": 0.0134, |
| "step": 9920 |
| }, |
| { |
| "epoch": 2.034419176398279, |
| "grad_norm": 0.3141976296901703, |
| "learning_rate": 2.397784964272865e-05, |
| "loss": 0.0122, |
| "step": 9930 |
| }, |
| { |
| "epoch": 2.0364679368981764, |
| "grad_norm": 0.26763567328453064, |
| "learning_rate": 2.3964598155091335e-05, |
| "loss": 0.0109, |
| "step": 9940 |
| }, |
| { |
| "epoch": 2.038516697398074, |
| "grad_norm": 0.5085217952728271, |
| "learning_rate": 2.3951335774621952e-05, |
| "loss": 0.0111, |
| "step": 9950 |
| }, |
| { |
| "epoch": 2.040565457897972, |
| "grad_norm": 0.29531678557395935, |
| "learning_rate": 2.3938062517435548e-05, |
| "loss": 0.0185, |
| "step": 9960 |
| }, |
| { |
| "epoch": 2.0426142183978695, |
| "grad_norm": 0.39689433574676514, |
| "learning_rate": 2.392477839966038e-05, |
| "loss": 0.0123, |
| "step": 9970 |
| }, |
| { |
| "epoch": 2.044662978897767, |
| "grad_norm": 0.34230825304985046, |
| "learning_rate": 2.3911483437437904e-05, |
| "loss": 0.0141, |
| "step": 9980 |
| }, |
| { |
| "epoch": 2.0467117393976646, |
| "grad_norm": 0.26099893450737, |
| "learning_rate": 2.3898177646922757e-05, |
| "loss": 0.0115, |
| "step": 9990 |
| }, |
| { |
| "epoch": 2.048760499897562, |
| "grad_norm": 0.3629753589630127, |
| "learning_rate": 2.3884861044282722e-05, |
| "loss": 0.0116, |
| "step": 10000 |
| }, |
| { |
| "epoch": 2.0508092603974597, |
| "grad_norm": 0.17743879556655884, |
| "learning_rate": 2.3871533645698732e-05, |
| "loss": 0.0112, |
| "step": 10010 |
| }, |
| { |
| "epoch": 2.0528580208973572, |
| "grad_norm": 0.17983455955982208, |
| "learning_rate": 2.3858195467364833e-05, |
| "loss": 0.0123, |
| "step": 10020 |
| }, |
| { |
| "epoch": 2.054906781397255, |
| "grad_norm": 0.17692384123802185, |
| "learning_rate": 2.3844846525488166e-05, |
| "loss": 0.0141, |
| "step": 10030 |
| }, |
| { |
| "epoch": 2.0569555418971524, |
| "grad_norm": 0.2711297273635864, |
| "learning_rate": 2.3831486836288957e-05, |
| "loss": 0.012, |
| "step": 10040 |
| }, |
| { |
| "epoch": 2.05900430239705, |
| "grad_norm": 0.37338387966156006, |
| "learning_rate": 2.381811641600048e-05, |
| "loss": 0.0112, |
| "step": 10050 |
| }, |
| { |
| "epoch": 2.0610530628969475, |
| "grad_norm": 0.23866449296474457, |
| "learning_rate": 2.3804735280869073e-05, |
| "loss": 0.0116, |
| "step": 10060 |
| }, |
| { |
| "epoch": 2.063101823396845, |
| "grad_norm": 0.25511065125465393, |
| "learning_rate": 2.3791343447154064e-05, |
| "loss": 0.0126, |
| "step": 10070 |
| }, |
| { |
| "epoch": 2.0651505838967426, |
| "grad_norm": 0.4390955865383148, |
| "learning_rate": 2.37779409311278e-05, |
| "loss": 0.0115, |
| "step": 10080 |
| }, |
| { |
| "epoch": 2.06719934439664, |
| "grad_norm": 0.16419632732868195, |
| "learning_rate": 2.3764527749075596e-05, |
| "loss": 0.0123, |
| "step": 10090 |
| }, |
| { |
| "epoch": 2.0692481048965377, |
| "grad_norm": 0.1460012048482895, |
| "learning_rate": 2.3751103917295735e-05, |
| "loss": 0.0127, |
| "step": 10100 |
| }, |
| { |
| "epoch": 2.071296865396435, |
| "grad_norm": 0.3381602466106415, |
| "learning_rate": 2.373766945209944e-05, |
| "loss": 0.0117, |
| "step": 10110 |
| }, |
| { |
| "epoch": 2.0733456258963328, |
| "grad_norm": 0.20190563797950745, |
| "learning_rate": 2.3724224369810856e-05, |
| "loss": 0.0108, |
| "step": 10120 |
| }, |
| { |
| "epoch": 2.0753943863962303, |
| "grad_norm": 0.1806822419166565, |
| "learning_rate": 2.3710768686767015e-05, |
| "loss": 0.0085, |
| "step": 10130 |
| }, |
| { |
| "epoch": 2.077443146896128, |
| "grad_norm": 0.3261871337890625, |
| "learning_rate": 2.3697302419317856e-05, |
| "loss": 0.0103, |
| "step": 10140 |
| }, |
| { |
| "epoch": 2.0794919073960254, |
| "grad_norm": 0.17560167610645294, |
| "learning_rate": 2.3683825583826146e-05, |
| "loss": 0.0141, |
| "step": 10150 |
| }, |
| { |
| "epoch": 2.081540667895923, |
| "grad_norm": 0.4527008831501007, |
| "learning_rate": 2.3670338196667528e-05, |
| "loss": 0.0099, |
| "step": 10160 |
| }, |
| { |
| "epoch": 2.0835894283958205, |
| "grad_norm": 0.4206450581550598, |
| "learning_rate": 2.3656840274230443e-05, |
| "loss": 0.0111, |
| "step": 10170 |
| }, |
| { |
| "epoch": 2.085638188895718, |
| "grad_norm": 0.4538799524307251, |
| "learning_rate": 2.3643331832916133e-05, |
| "loss": 0.0114, |
| "step": 10180 |
| }, |
| { |
| "epoch": 2.0876869493956156, |
| "grad_norm": 0.43454715609550476, |
| "learning_rate": 2.362981288913863e-05, |
| "loss": 0.0118, |
| "step": 10190 |
| }, |
| { |
| "epoch": 2.089735709895513, |
| "grad_norm": 0.5395990014076233, |
| "learning_rate": 2.3616283459324737e-05, |
| "loss": 0.0127, |
| "step": 10200 |
| }, |
| { |
| "epoch": 2.0917844703954107, |
| "grad_norm": 0.19712281227111816, |
| "learning_rate": 2.3602743559913973e-05, |
| "loss": 0.0133, |
| "step": 10210 |
| }, |
| { |
| "epoch": 2.0938332308953083, |
| "grad_norm": 0.34996548295021057, |
| "learning_rate": 2.35891932073586e-05, |
| "loss": 0.0136, |
| "step": 10220 |
| }, |
| { |
| "epoch": 2.095881991395206, |
| "grad_norm": 0.24226465821266174, |
| "learning_rate": 2.357563241812357e-05, |
| "loss": 0.013, |
| "step": 10230 |
| }, |
| { |
| "epoch": 2.0979307518951034, |
| "grad_norm": 0.5005622506141663, |
| "learning_rate": 2.3562061208686522e-05, |
| "loss": 0.0128, |
| "step": 10240 |
| }, |
| { |
| "epoch": 2.099979512395001, |
| "grad_norm": 0.2075444608926773, |
| "learning_rate": 2.354847959553776e-05, |
| "loss": 0.0095, |
| "step": 10250 |
| }, |
| { |
| "epoch": 2.1020282728948985, |
| "grad_norm": 0.6302710175514221, |
| "learning_rate": 2.353488759518022e-05, |
| "loss": 0.0118, |
| "step": 10260 |
| }, |
| { |
| "epoch": 2.104077033394796, |
| "grad_norm": 0.3023175299167633, |
| "learning_rate": 2.352128522412946e-05, |
| "loss": 0.0136, |
| "step": 10270 |
| }, |
| { |
| "epoch": 2.1061257938946936, |
| "grad_norm": 0.17477557063102722, |
| "learning_rate": 2.3507672498913653e-05, |
| "loss": 0.0113, |
| "step": 10280 |
| }, |
| { |
| "epoch": 2.108174554394591, |
| "grad_norm": 0.35146307945251465, |
| "learning_rate": 2.3494049436073537e-05, |
| "loss": 0.0125, |
| "step": 10290 |
| }, |
| { |
| "epoch": 2.1102233148944887, |
| "grad_norm": 0.15435707569122314, |
| "learning_rate": 2.348041605216242e-05, |
| "loss": 0.0109, |
| "step": 10300 |
| }, |
| { |
| "epoch": 2.112272075394386, |
| "grad_norm": 0.14636899530887604, |
| "learning_rate": 2.3466772363746146e-05, |
| "loss": 0.011, |
| "step": 10310 |
| }, |
| { |
| "epoch": 2.1143208358942838, |
| "grad_norm": 0.5210812091827393, |
| "learning_rate": 2.345311838740309e-05, |
| "loss": 0.0127, |
| "step": 10320 |
| }, |
| { |
| "epoch": 2.1163695963941813, |
| "grad_norm": 0.23452824354171753, |
| "learning_rate": 2.343945413972411e-05, |
| "loss": 0.0106, |
| "step": 10330 |
| }, |
| { |
| "epoch": 2.118418356894079, |
| "grad_norm": 0.7631458044052124, |
| "learning_rate": 2.3425779637312568e-05, |
| "loss": 0.012, |
| "step": 10340 |
| }, |
| { |
| "epoch": 2.120467117393977, |
| "grad_norm": 0.11999737471342087, |
| "learning_rate": 2.3412094896784264e-05, |
| "loss": 0.0106, |
| "step": 10350 |
| }, |
| { |
| "epoch": 2.1225158778938744, |
| "grad_norm": 0.24293369054794312, |
| "learning_rate": 2.339839993476745e-05, |
| "loss": 0.0119, |
| "step": 10360 |
| }, |
| { |
| "epoch": 2.124564638393772, |
| "grad_norm": 0.1800415962934494, |
| "learning_rate": 2.3384694767902804e-05, |
| "loss": 0.0119, |
| "step": 10370 |
| }, |
| { |
| "epoch": 2.1266133988936695, |
| "grad_norm": 0.7652474641799927, |
| "learning_rate": 2.3370979412843384e-05, |
| "loss": 0.0101, |
| "step": 10380 |
| }, |
| { |
| "epoch": 2.128662159393567, |
| "grad_norm": 0.14322306215763092, |
| "learning_rate": 2.3357253886254648e-05, |
| "loss": 0.0175, |
| "step": 10390 |
| }, |
| { |
| "epoch": 2.1307109198934646, |
| "grad_norm": 0.2438124716281891, |
| "learning_rate": 2.3343518204814407e-05, |
| "loss": 0.0131, |
| "step": 10400 |
| }, |
| { |
| "epoch": 2.132759680393362, |
| "grad_norm": 0.2166603058576584, |
| "learning_rate": 2.3329772385212798e-05, |
| "loss": 0.012, |
| "step": 10410 |
| }, |
| { |
| "epoch": 2.1348084408932597, |
| "grad_norm": 0.20191770792007446, |
| "learning_rate": 2.3316016444152296e-05, |
| "loss": 0.0117, |
| "step": 10420 |
| }, |
| { |
| "epoch": 2.1368572013931573, |
| "grad_norm": 0.15889151394367218, |
| "learning_rate": 2.3302250398347668e-05, |
| "loss": 0.0154, |
| "step": 10430 |
| }, |
| { |
| "epoch": 2.138905961893055, |
| "grad_norm": 0.2019120305776596, |
| "learning_rate": 2.328847426452595e-05, |
| "loss": 0.0137, |
| "step": 10440 |
| }, |
| { |
| "epoch": 2.1409547223929524, |
| "grad_norm": 0.2322073131799698, |
| "learning_rate": 2.3274688059426456e-05, |
| "loss": 0.0105, |
| "step": 10450 |
| }, |
| { |
| "epoch": 2.14300348289285, |
| "grad_norm": 0.19652584195137024, |
| "learning_rate": 2.3260891799800715e-05, |
| "loss": 0.0115, |
| "step": 10460 |
| }, |
| { |
| "epoch": 2.1450522433927475, |
| "grad_norm": 0.132473886013031, |
| "learning_rate": 2.3247085502412482e-05, |
| "loss": 0.0112, |
| "step": 10470 |
| }, |
| { |
| "epoch": 2.147101003892645, |
| "grad_norm": 0.23023509979248047, |
| "learning_rate": 2.3233269184037717e-05, |
| "loss": 0.0095, |
| "step": 10480 |
| }, |
| { |
| "epoch": 2.1491497643925426, |
| "grad_norm": 0.4320821166038513, |
| "learning_rate": 2.3219442861464547e-05, |
| "loss": 0.0141, |
| "step": 10490 |
| }, |
| { |
| "epoch": 2.15119852489244, |
| "grad_norm": 0.28234395384788513, |
| "learning_rate": 2.3205606551493255e-05, |
| "loss": 0.0101, |
| "step": 10500 |
| }, |
| { |
| "epoch": 2.1532472853923377, |
| "grad_norm": 0.19151149690151215, |
| "learning_rate": 2.319176027093627e-05, |
| "loss": 0.0134, |
| "step": 10510 |
| }, |
| { |
| "epoch": 2.1552960458922352, |
| "grad_norm": 0.22751007974147797, |
| "learning_rate": 2.3177904036618117e-05, |
| "loss": 0.013, |
| "step": 10520 |
| }, |
| { |
| "epoch": 2.1573448063921328, |
| "grad_norm": 0.2100066840648651, |
| "learning_rate": 2.316403786537544e-05, |
| "loss": 0.0113, |
| "step": 10530 |
| }, |
| { |
| "epoch": 2.1593935668920303, |
| "grad_norm": 0.30193421244621277, |
| "learning_rate": 2.3150161774056933e-05, |
| "loss": 0.0127, |
| "step": 10540 |
| }, |
| { |
| "epoch": 2.161442327391928, |
| "grad_norm": 0.17684316635131836, |
| "learning_rate": 2.313627577952336e-05, |
| "loss": 0.0126, |
| "step": 10550 |
| }, |
| { |
| "epoch": 2.1634910878918254, |
| "grad_norm": 0.19200824201107025, |
| "learning_rate": 2.312237989864752e-05, |
| "loss": 0.0124, |
| "step": 10560 |
| }, |
| { |
| "epoch": 2.165539848391723, |
| "grad_norm": 0.22590887546539307, |
| "learning_rate": 2.3108474148314205e-05, |
| "loss": 0.011, |
| "step": 10570 |
| }, |
| { |
| "epoch": 2.1675886088916205, |
| "grad_norm": 0.18505719304084778, |
| "learning_rate": 2.309455854542023e-05, |
| "loss": 0.0159, |
| "step": 10580 |
| }, |
| { |
| "epoch": 2.169637369391518, |
| "grad_norm": 0.31873178482055664, |
| "learning_rate": 2.308063310687435e-05, |
| "loss": 0.0149, |
| "step": 10590 |
| }, |
| { |
| "epoch": 2.1716861298914156, |
| "grad_norm": 0.2074051946401596, |
| "learning_rate": 2.306669784959729e-05, |
| "loss": 0.0123, |
| "step": 10600 |
| }, |
| { |
| "epoch": 2.173734890391313, |
| "grad_norm": 0.2341642826795578, |
| "learning_rate": 2.3052752790521703e-05, |
| "loss": 0.0118, |
| "step": 10610 |
| }, |
| { |
| "epoch": 2.1757836508912107, |
| "grad_norm": 0.33994895219802856, |
| "learning_rate": 2.3038797946592152e-05, |
| "loss": 0.0107, |
| "step": 10620 |
| }, |
| { |
| "epoch": 2.1778324113911083, |
| "grad_norm": 0.28025034070014954, |
| "learning_rate": 2.3024833334765087e-05, |
| "loss": 0.0106, |
| "step": 10630 |
| }, |
| { |
| "epoch": 2.179881171891006, |
| "grad_norm": 0.22036705911159515, |
| "learning_rate": 2.301085897200883e-05, |
| "loss": 0.0127, |
| "step": 10640 |
| }, |
| { |
| "epoch": 2.1819299323909034, |
| "grad_norm": 0.46761807799339294, |
| "learning_rate": 2.299687487530354e-05, |
| "loss": 0.0122, |
| "step": 10650 |
| }, |
| { |
| "epoch": 2.183978692890801, |
| "grad_norm": 0.1661396473646164, |
| "learning_rate": 2.2982881061641228e-05, |
| "loss": 0.011, |
| "step": 10660 |
| }, |
| { |
| "epoch": 2.1860274533906985, |
| "grad_norm": 0.23082272708415985, |
| "learning_rate": 2.296887754802569e-05, |
| "loss": 0.0112, |
| "step": 10670 |
| }, |
| { |
| "epoch": 2.188076213890596, |
| "grad_norm": 0.1763562262058258, |
| "learning_rate": 2.295486435147251e-05, |
| "loss": 0.0122, |
| "step": 10680 |
| }, |
| { |
| "epoch": 2.1901249743904936, |
| "grad_norm": 0.11439327150583267, |
| "learning_rate": 2.294084148900905e-05, |
| "loss": 0.0111, |
| "step": 10690 |
| }, |
| { |
| "epoch": 2.192173734890391, |
| "grad_norm": 0.20816867053508759, |
| "learning_rate": 2.292680897767441e-05, |
| "loss": 0.0121, |
| "step": 10700 |
| }, |
| { |
| "epoch": 2.1942224953902887, |
| "grad_norm": 0.16031624376773834, |
| "learning_rate": 2.291276683451941e-05, |
| "loss": 0.0095, |
| "step": 10710 |
| }, |
| { |
| "epoch": 2.1962712558901862, |
| "grad_norm": 0.10238825529813766, |
| "learning_rate": 2.2898715076606584e-05, |
| "loss": 0.0113, |
| "step": 10720 |
| }, |
| { |
| "epoch": 2.198320016390084, |
| "grad_norm": 0.47116655111312866, |
| "learning_rate": 2.2884653721010135e-05, |
| "loss": 0.0106, |
| "step": 10730 |
| }, |
| { |
| "epoch": 2.200368776889982, |
| "grad_norm": 0.2346353679895401, |
| "learning_rate": 2.2870582784815946e-05, |
| "loss": 0.0101, |
| "step": 10740 |
| }, |
| { |
| "epoch": 2.202417537389879, |
| "grad_norm": 0.18897022306919098, |
| "learning_rate": 2.2856502285121525e-05, |
| "loss": 0.0142, |
| "step": 10750 |
| }, |
| { |
| "epoch": 2.204466297889777, |
| "grad_norm": 0.44136667251586914, |
| "learning_rate": 2.2842412239036004e-05, |
| "loss": 0.0119, |
| "step": 10760 |
| }, |
| { |
| "epoch": 2.2065150583896744, |
| "grad_norm": 0.27752813696861267, |
| "learning_rate": 2.2828312663680125e-05, |
| "loss": 0.012, |
| "step": 10770 |
| }, |
| { |
| "epoch": 2.208563818889572, |
| "grad_norm": 0.20734326541423798, |
| "learning_rate": 2.2814203576186194e-05, |
| "loss": 0.0154, |
| "step": 10780 |
| }, |
| { |
| "epoch": 2.2106125793894695, |
| "grad_norm": 0.22143232822418213, |
| "learning_rate": 2.2800084993698093e-05, |
| "loss": 0.0106, |
| "step": 10790 |
| }, |
| { |
| "epoch": 2.212661339889367, |
| "grad_norm": 0.14383463561534882, |
| "learning_rate": 2.2785956933371224e-05, |
| "loss": 0.0103, |
| "step": 10800 |
| }, |
| { |
| "epoch": 2.2147101003892646, |
| "grad_norm": 0.23359259963035583, |
| "learning_rate": 2.2771819412372505e-05, |
| "loss": 0.0124, |
| "step": 10810 |
| }, |
| { |
| "epoch": 2.216758860889162, |
| "grad_norm": 0.23750093579292297, |
| "learning_rate": 2.2757672447880374e-05, |
| "loss": 0.0093, |
| "step": 10820 |
| }, |
| { |
| "epoch": 2.2188076213890597, |
| "grad_norm": 0.14557082951068878, |
| "learning_rate": 2.2743516057084712e-05, |
| "loss": 0.0097, |
| "step": 10830 |
| }, |
| { |
| "epoch": 2.2208563818889573, |
| "grad_norm": 0.2520831227302551, |
| "learning_rate": 2.2729350257186877e-05, |
| "loss": 0.0108, |
| "step": 10840 |
| }, |
| { |
| "epoch": 2.222905142388855, |
| "grad_norm": 0.23964238166809082, |
| "learning_rate": 2.2715175065399655e-05, |
| "loss": 0.0139, |
| "step": 10850 |
| }, |
| { |
| "epoch": 2.2249539028887524, |
| "grad_norm": 0.21454685926437378, |
| "learning_rate": 2.270099049894723e-05, |
| "loss": 0.0106, |
| "step": 10860 |
| }, |
| { |
| "epoch": 2.22700266338865, |
| "grad_norm": 0.2823438048362732, |
| "learning_rate": 2.268679657506519e-05, |
| "loss": 0.0091, |
| "step": 10870 |
| }, |
| { |
| "epoch": 2.2290514238885475, |
| "grad_norm": 0.2413579225540161, |
| "learning_rate": 2.2672593311000498e-05, |
| "loss": 0.0105, |
| "step": 10880 |
| }, |
| { |
| "epoch": 2.231100184388445, |
| "grad_norm": 0.10397704690694809, |
| "learning_rate": 2.2658380724011452e-05, |
| "loss": 0.0113, |
| "step": 10890 |
| }, |
| { |
| "epoch": 2.2331489448883426, |
| "grad_norm": 0.28332334756851196, |
| "learning_rate": 2.264415883136769e-05, |
| "loss": 0.0121, |
| "step": 10900 |
| }, |
| { |
| "epoch": 2.23519770538824, |
| "grad_norm": 0.2038186490535736, |
| "learning_rate": 2.262992765035015e-05, |
| "loss": 0.011, |
| "step": 10910 |
| }, |
| { |
| "epoch": 2.2372464658881377, |
| "grad_norm": 0.3021252155303955, |
| "learning_rate": 2.2615687198251062e-05, |
| "loss": 0.0122, |
| "step": 10920 |
| }, |
| { |
| "epoch": 2.2392952263880352, |
| "grad_norm": 0.16428732872009277, |
| "learning_rate": 2.2601437492373914e-05, |
| "loss": 0.0097, |
| "step": 10930 |
| }, |
| { |
| "epoch": 2.241343986887933, |
| "grad_norm": 0.30075010657310486, |
| "learning_rate": 2.2587178550033444e-05, |
| "loss": 0.0117, |
| "step": 10940 |
| }, |
| { |
| "epoch": 2.2433927473878303, |
| "grad_norm": 0.1838994175195694, |
| "learning_rate": 2.2572910388555615e-05, |
| "loss": 0.0111, |
| "step": 10950 |
| }, |
| { |
| "epoch": 2.245441507887728, |
| "grad_norm": 0.20049159228801727, |
| "learning_rate": 2.255863302527758e-05, |
| "loss": 0.0118, |
| "step": 10960 |
| }, |
| { |
| "epoch": 2.2474902683876254, |
| "grad_norm": 0.33610478043556213, |
| "learning_rate": 2.2544346477547685e-05, |
| "loss": 0.0138, |
| "step": 10970 |
| }, |
| { |
| "epoch": 2.249539028887523, |
| "grad_norm": 0.1395016312599182, |
| "learning_rate": 2.253005076272544e-05, |
| "loss": 0.0129, |
| "step": 10980 |
| }, |
| { |
| "epoch": 2.2515877893874205, |
| "grad_norm": 0.21686992049217224, |
| "learning_rate": 2.2515745898181477e-05, |
| "loss": 0.0126, |
| "step": 10990 |
| }, |
| { |
| "epoch": 2.253636549887318, |
| "grad_norm": 0.354717880487442, |
| "learning_rate": 2.250143190129756e-05, |
| "loss": 0.0107, |
| "step": 11000 |
| } |
| ], |
| "logging_steps": 10, |
| "max_steps": 30000, |
| "num_input_tokens_seen": 0, |
| "num_train_epochs": 7, |
| "save_steps": 1000, |
| "stateful_callbacks": { |
| "TrainerControl": { |
| "args": { |
| "should_epoch_stop": false, |
| "should_evaluate": false, |
| "should_log": false, |
| "should_save": true, |
| "should_training_stop": false |
| }, |
| "attributes": {} |
| } |
| }, |
| "total_flos": 0.0, |
| "train_batch_size": 60, |
| "trial_name": null, |
| "trial_params": null |
| } |
|
|