{ "best_global_step": null, "best_metric": null, "best_model_checkpoint": null, "epoch": 23.28288707799767, "eval_steps": 500, "global_step": 40000, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.005820721769499418, "grad_norm": 2.4815659523010254, "learning_rate": 3.0000000000000004e-07, "loss": 2.7856, "step": 10 }, { "epoch": 0.011641443538998836, "grad_norm": 1.853691577911377, "learning_rate": 6.333333333333333e-07, "loss": 2.7586, "step": 20 }, { "epoch": 0.017462165308498253, "grad_norm": 2.5251176357269287, "learning_rate": 9.666666666666668e-07, "loss": 2.7366, "step": 30 }, { "epoch": 0.023282887077997673, "grad_norm": 1.90872061252594, "learning_rate": 1.3e-06, "loss": 2.7054, "step": 40 }, { "epoch": 0.02910360884749709, "grad_norm": 1.5593458414077759, "learning_rate": 1.6333333333333333e-06, "loss": 2.6814, "step": 50 }, { "epoch": 0.034924330616996506, "grad_norm": 1.4362269639968872, "learning_rate": 1.9666666666666668e-06, "loss": 2.6769, "step": 60 }, { "epoch": 0.04074505238649592, "grad_norm": 1.416746735572815, "learning_rate": 2.3e-06, "loss": 2.6765, "step": 70 }, { "epoch": 0.046565774155995346, "grad_norm": 2.1545817852020264, "learning_rate": 2.6333333333333337e-06, "loss": 2.6487, "step": 80 }, { "epoch": 0.05238649592549476, "grad_norm": 1.9410408735275269, "learning_rate": 2.966666666666667e-06, "loss": 2.6257, "step": 90 }, { "epoch": 0.05820721769499418, "grad_norm": 1.1579829454421997, "learning_rate": 3.3e-06, "loss": 2.6042, "step": 100 }, { "epoch": 0.0640279394644936, "grad_norm": 1.6039224863052368, "learning_rate": 3.633333333333334e-06, "loss": 2.5777, "step": 110 }, { "epoch": 0.06984866123399301, "grad_norm": 1.4700742959976196, "learning_rate": 3.966666666666667e-06, "loss": 2.5835, "step": 120 }, { "epoch": 0.07566938300349244, "grad_norm": 1.7858898639678955, "learning_rate": 4.2999999999999995e-06, "loss": 2.5432, "step": 130 }, { "epoch": 0.08149010477299184, "grad_norm": 1.9030888080596924, "learning_rate": 4.633333333333334e-06, "loss": 2.529, "step": 140 }, { "epoch": 0.08731082654249127, "grad_norm": 2.033205270767212, "learning_rate": 4.966666666666667e-06, "loss": 2.5034, "step": 150 }, { "epoch": 0.09313154831199069, "grad_norm": 2.061861276626587, "learning_rate": 5.3e-06, "loss": 2.477, "step": 160 }, { "epoch": 0.0989522700814901, "grad_norm": 1.8396316766738892, "learning_rate": 5.633333333333333e-06, "loss": 2.4301, "step": 170 }, { "epoch": 0.10477299185098952, "grad_norm": 4.38357400894165, "learning_rate": 5.9666666666666666e-06, "loss": 2.396, "step": 180 }, { "epoch": 0.11059371362048893, "grad_norm": 2.7974507808685303, "learning_rate": 6.300000000000001e-06, "loss": 2.3303, "step": 190 }, { "epoch": 0.11641443538998836, "grad_norm": 2.4046788215637207, "learning_rate": 6.633333333333333e-06, "loss": 2.297, "step": 200 }, { "epoch": 0.12223515715948778, "grad_norm": 3.4349629878997803, "learning_rate": 6.966666666666667e-06, "loss": 2.228, "step": 210 }, { "epoch": 0.1280558789289872, "grad_norm": 5.490354061126709, "learning_rate": 7.2999999999999996e-06, "loss": 2.1658, "step": 220 }, { "epoch": 0.13387660069848661, "grad_norm": 5.056091785430908, "learning_rate": 7.633333333333334e-06, "loss": 2.0951, "step": 230 }, { "epoch": 0.13969732246798602, "grad_norm": 4.961029052734375, "learning_rate": 7.966666666666666e-06, "loss": 2.0246, "step": 240 }, { "epoch": 0.14551804423748546, "grad_norm": 4.4147748947143555, "learning_rate": 8.3e-06, "loss": 1.9589, "step": 250 }, { "epoch": 0.15133876600698487, "grad_norm": 5.014352321624756, "learning_rate": 8.633333333333334e-06, "loss": 1.8529, "step": 260 }, { "epoch": 0.15715948777648428, "grad_norm": 5.476301670074463, "learning_rate": 8.966666666666668e-06, "loss": 1.8602, "step": 270 }, { "epoch": 0.1629802095459837, "grad_norm": 6.117679595947266, "learning_rate": 9.3e-06, "loss": 1.8203, "step": 280 }, { "epoch": 0.16880093131548313, "grad_norm": 7.895984172821045, "learning_rate": 9.633333333333335e-06, "loss": 1.771, "step": 290 }, { "epoch": 0.17462165308498254, "grad_norm": 4.653863906860352, "learning_rate": 9.966666666666667e-06, "loss": 1.7469, "step": 300 }, { "epoch": 0.18044237485448195, "grad_norm": 8.867400169372559, "learning_rate": 1.03e-05, "loss": 1.7651, "step": 310 }, { "epoch": 0.18626309662398138, "grad_norm": 4.071671485900879, "learning_rate": 1.0633333333333334e-05, "loss": 1.7401, "step": 320 }, { "epoch": 0.1920838183934808, "grad_norm": 3.9089016914367676, "learning_rate": 1.0966666666666666e-05, "loss": 1.7104, "step": 330 }, { "epoch": 0.1979045401629802, "grad_norm": 4.5396952629089355, "learning_rate": 1.13e-05, "loss": 1.6394, "step": 340 }, { "epoch": 0.20372526193247964, "grad_norm": 4.105408668518066, "learning_rate": 1.1633333333333334e-05, "loss": 1.6339, "step": 350 }, { "epoch": 0.20954598370197905, "grad_norm": 4.206538200378418, "learning_rate": 1.1966666666666668e-05, "loss": 1.6269, "step": 360 }, { "epoch": 0.21536670547147846, "grad_norm": 2.829434394836426, "learning_rate": 1.23e-05, "loss": 1.6194, "step": 370 }, { "epoch": 0.22118742724097787, "grad_norm": 2.9135076999664307, "learning_rate": 1.2633333333333333e-05, "loss": 1.5741, "step": 380 }, { "epoch": 0.2270081490104773, "grad_norm": 2.318284034729004, "learning_rate": 1.2966666666666669e-05, "loss": 1.5583, "step": 390 }, { "epoch": 0.23282887077997672, "grad_norm": 4.086777210235596, "learning_rate": 1.3300000000000001e-05, "loss": 1.5488, "step": 400 }, { "epoch": 0.23864959254947612, "grad_norm": 3.2144272327423096, "learning_rate": 1.3633333333333334e-05, "loss": 1.5464, "step": 410 }, { "epoch": 0.24447031431897556, "grad_norm": 2.889495849609375, "learning_rate": 1.3966666666666666e-05, "loss": 1.5396, "step": 420 }, { "epoch": 0.25029103608847497, "grad_norm": 2.6884067058563232, "learning_rate": 1.43e-05, "loss": 1.509, "step": 430 }, { "epoch": 0.2561117578579744, "grad_norm": 3.9395334720611572, "learning_rate": 1.4633333333333334e-05, "loss": 1.5089, "step": 440 }, { "epoch": 0.2619324796274738, "grad_norm": 3.8806025981903076, "learning_rate": 1.4966666666666668e-05, "loss": 1.4947, "step": 450 }, { "epoch": 0.26775320139697323, "grad_norm": 3.4137089252471924, "learning_rate": 1.53e-05, "loss": 1.5169, "step": 460 }, { "epoch": 0.27357392316647267, "grad_norm": 1.9820940494537354, "learning_rate": 1.563333333333333e-05, "loss": 1.4741, "step": 470 }, { "epoch": 0.27939464493597205, "grad_norm": 1.413993239402771, "learning_rate": 1.5966666666666667e-05, "loss": 1.4598, "step": 480 }, { "epoch": 0.2852153667054715, "grad_norm": 3.229191541671753, "learning_rate": 1.63e-05, "loss": 1.4436, "step": 490 }, { "epoch": 0.2910360884749709, "grad_norm": 2.5939457416534424, "learning_rate": 1.6633333333333336e-05, "loss": 1.4345, "step": 500 }, { "epoch": 0.2968568102444703, "grad_norm": 1.865046739578247, "learning_rate": 1.6966666666666668e-05, "loss": 1.4541, "step": 510 }, { "epoch": 0.30267753201396974, "grad_norm": 2.358537197113037, "learning_rate": 1.73e-05, "loss": 1.4125, "step": 520 }, { "epoch": 0.3084982537834691, "grad_norm": 2.1191585063934326, "learning_rate": 1.7633333333333336e-05, "loss": 1.417, "step": 530 }, { "epoch": 0.31431897555296856, "grad_norm": 1.3803356885910034, "learning_rate": 1.796666666666667e-05, "loss": 1.3735, "step": 540 }, { "epoch": 0.320139697322468, "grad_norm": 1.6117502450942993, "learning_rate": 1.83e-05, "loss": 1.3837, "step": 550 }, { "epoch": 0.3259604190919674, "grad_norm": 1.89707612991333, "learning_rate": 1.8633333333333333e-05, "loss": 1.3769, "step": 560 }, { "epoch": 0.3317811408614668, "grad_norm": 2.4454703330993652, "learning_rate": 1.896666666666667e-05, "loss": 1.3591, "step": 570 }, { "epoch": 0.33760186263096625, "grad_norm": 1.8364553451538086, "learning_rate": 1.93e-05, "loss": 1.3621, "step": 580 }, { "epoch": 0.34342258440046564, "grad_norm": 2.0097408294677734, "learning_rate": 1.9633333333333334e-05, "loss": 1.3657, "step": 590 }, { "epoch": 0.3492433061699651, "grad_norm": 1.7796767950057983, "learning_rate": 1.9966666666666666e-05, "loss": 1.3684, "step": 600 }, { "epoch": 0.3550640279394645, "grad_norm": 2.0909807682037354, "learning_rate": 2.0300000000000002e-05, "loss": 1.3868, "step": 610 }, { "epoch": 0.3608847497089639, "grad_norm": 1.6235121488571167, "learning_rate": 2.0633333333333335e-05, "loss": 1.331, "step": 620 }, { "epoch": 0.36670547147846333, "grad_norm": 1.6188032627105713, "learning_rate": 2.0966666666666667e-05, "loss": 1.3489, "step": 630 }, { "epoch": 0.37252619324796277, "grad_norm": 2.4604203701019287, "learning_rate": 2.13e-05, "loss": 1.3307, "step": 640 }, { "epoch": 0.37834691501746215, "grad_norm": 1.8102141618728638, "learning_rate": 2.1633333333333332e-05, "loss": 1.3258, "step": 650 }, { "epoch": 0.3841676367869616, "grad_norm": 2.669039487838745, "learning_rate": 2.1966666666666668e-05, "loss": 1.3274, "step": 660 }, { "epoch": 0.389988358556461, "grad_norm": 1.6969751119613647, "learning_rate": 2.23e-05, "loss": 1.3479, "step": 670 }, { "epoch": 0.3958090803259604, "grad_norm": 2.105466604232788, "learning_rate": 2.2633333333333336e-05, "loss": 1.3205, "step": 680 }, { "epoch": 0.40162980209545984, "grad_norm": 1.8735966682434082, "learning_rate": 2.2966666666666668e-05, "loss": 1.3131, "step": 690 }, { "epoch": 0.4074505238649593, "grad_norm": 1.5554078817367554, "learning_rate": 2.3300000000000004e-05, "loss": 1.3055, "step": 700 }, { "epoch": 0.41327124563445866, "grad_norm": 2.4699294567108154, "learning_rate": 2.3633333333333336e-05, "loss": 1.3409, "step": 710 }, { "epoch": 0.4190919674039581, "grad_norm": 1.7793229818344116, "learning_rate": 2.396666666666667e-05, "loss": 1.3094, "step": 720 }, { "epoch": 0.42491268917345754, "grad_norm": 1.9663199186325073, "learning_rate": 2.43e-05, "loss": 1.3264, "step": 730 }, { "epoch": 0.4307334109429569, "grad_norm": 1.5309123992919922, "learning_rate": 2.4633333333333334e-05, "loss": 1.2856, "step": 740 }, { "epoch": 0.43655413271245636, "grad_norm": 1.2400994300842285, "learning_rate": 2.496666666666667e-05, "loss": 1.2924, "step": 750 }, { "epoch": 0.44237485448195574, "grad_norm": 1.9461252689361572, "learning_rate": 2.5300000000000002e-05, "loss": 1.2889, "step": 760 }, { "epoch": 0.4481955762514552, "grad_norm": 1.7307933568954468, "learning_rate": 2.5633333333333338e-05, "loss": 1.2745, "step": 770 }, { "epoch": 0.4540162980209546, "grad_norm": 1.7702643871307373, "learning_rate": 2.5966666666666667e-05, "loss": 1.2873, "step": 780 }, { "epoch": 0.459837019790454, "grad_norm": 1.9930938482284546, "learning_rate": 2.6300000000000002e-05, "loss": 1.2722, "step": 790 }, { "epoch": 0.46565774155995343, "grad_norm": 1.7803294658660889, "learning_rate": 2.663333333333333e-05, "loss": 1.267, "step": 800 }, { "epoch": 0.47147846332945287, "grad_norm": 1.5366089344024658, "learning_rate": 2.6966666666666667e-05, "loss": 1.2823, "step": 810 }, { "epoch": 0.47729918509895225, "grad_norm": 1.6607978343963623, "learning_rate": 2.7300000000000003e-05, "loss": 1.2733, "step": 820 }, { "epoch": 0.4831199068684517, "grad_norm": 1.7741808891296387, "learning_rate": 2.7633333333333332e-05, "loss": 1.2683, "step": 830 }, { "epoch": 0.4889406286379511, "grad_norm": 1.7722482681274414, "learning_rate": 2.7966666666666668e-05, "loss": 1.27, "step": 840 }, { "epoch": 0.4947613504074505, "grad_norm": 1.9304368495941162, "learning_rate": 2.83e-05, "loss": 1.265, "step": 850 }, { "epoch": 0.5005820721769499, "grad_norm": 1.3524025678634644, "learning_rate": 2.8633333333333336e-05, "loss": 1.2618, "step": 860 }, { "epoch": 0.5064027939464494, "grad_norm": 1.3825901746749878, "learning_rate": 2.8966666666666668e-05, "loss": 1.2678, "step": 870 }, { "epoch": 0.5122235157159488, "grad_norm": 1.4791237115859985, "learning_rate": 2.93e-05, "loss": 1.265, "step": 880 }, { "epoch": 0.5180442374854481, "grad_norm": 0.9114635586738586, "learning_rate": 2.9633333333333336e-05, "loss": 1.26, "step": 890 }, { "epoch": 0.5238649592549476, "grad_norm": 1.118841290473938, "learning_rate": 2.9966666666666672e-05, "loss": 1.2445, "step": 900 }, { "epoch": 0.529685681024447, "grad_norm": 1.4991689920425415, "learning_rate": 3.03e-05, "loss": 1.2359, "step": 910 }, { "epoch": 0.5355064027939465, "grad_norm": 1.7330141067504883, "learning_rate": 3.063333333333334e-05, "loss": 1.2466, "step": 920 }, { "epoch": 0.5413271245634459, "grad_norm": 1.087327003479004, "learning_rate": 3.096666666666666e-05, "loss": 1.2319, "step": 930 }, { "epoch": 0.5471478463329453, "grad_norm": 1.3227134943008423, "learning_rate": 3.13e-05, "loss": 1.2356, "step": 940 }, { "epoch": 0.5529685681024447, "grad_norm": 1.2291042804718018, "learning_rate": 3.1633333333333334e-05, "loss": 1.2418, "step": 950 }, { "epoch": 0.5587892898719441, "grad_norm": 1.1533374786376953, "learning_rate": 3.196666666666667e-05, "loss": 1.215, "step": 960 }, { "epoch": 0.5646100116414435, "grad_norm": 1.5578526258468628, "learning_rate": 3.2300000000000006e-05, "loss": 1.22, "step": 970 }, { "epoch": 0.570430733410943, "grad_norm": 1.0608288049697876, "learning_rate": 3.263333333333333e-05, "loss": 1.2241, "step": 980 }, { "epoch": 0.5762514551804424, "grad_norm": 1.207740306854248, "learning_rate": 3.296666666666667e-05, "loss": 1.2262, "step": 990 }, { "epoch": 0.5820721769499418, "grad_norm": 1.648051381111145, "learning_rate": 3.33e-05, "loss": 1.2161, "step": 1000 }, { "epoch": 0.5878928987194412, "grad_norm": 1.2090243101119995, "learning_rate": 3.3633333333333335e-05, "loss": 1.2079, "step": 1010 }, { "epoch": 0.5937136204889406, "grad_norm": 1.0535739660263062, "learning_rate": 3.396666666666667e-05, "loss": 1.2116, "step": 1020 }, { "epoch": 0.59953434225844, "grad_norm": 1.4417856931686401, "learning_rate": 3.430000000000001e-05, "loss": 1.2094, "step": 1030 }, { "epoch": 0.6053550640279395, "grad_norm": 1.1288009881973267, "learning_rate": 3.463333333333333e-05, "loss": 1.2068, "step": 1040 }, { "epoch": 0.6111757857974389, "grad_norm": 1.2067853212356567, "learning_rate": 3.496666666666667e-05, "loss": 1.1992, "step": 1050 }, { "epoch": 0.6169965075669382, "grad_norm": 1.3694567680358887, "learning_rate": 3.53e-05, "loss": 1.2167, "step": 1060 }, { "epoch": 0.6228172293364377, "grad_norm": 1.3209940195083618, "learning_rate": 3.563333333333334e-05, "loss": 1.2072, "step": 1070 }, { "epoch": 0.6286379511059371, "grad_norm": 1.574528455734253, "learning_rate": 3.596666666666667e-05, "loss": 1.2104, "step": 1080 }, { "epoch": 0.6344586728754366, "grad_norm": 0.94853675365448, "learning_rate": 3.63e-05, "loss": 1.1989, "step": 1090 }, { "epoch": 0.640279394644936, "grad_norm": 0.9895154237747192, "learning_rate": 3.6633333333333334e-05, "loss": 1.1925, "step": 1100 }, { "epoch": 0.6461001164144354, "grad_norm": 1.1562230587005615, "learning_rate": 3.6966666666666666e-05, "loss": 1.1992, "step": 1110 }, { "epoch": 0.6519208381839348, "grad_norm": 1.487748146057129, "learning_rate": 3.73e-05, "loss": 1.2061, "step": 1120 }, { "epoch": 0.6577415599534342, "grad_norm": 1.1773114204406738, "learning_rate": 3.763333333333334e-05, "loss": 1.2123, "step": 1130 }, { "epoch": 0.6635622817229336, "grad_norm": 1.0233012437820435, "learning_rate": 3.796666666666667e-05, "loss": 1.1877, "step": 1140 }, { "epoch": 0.6693830034924331, "grad_norm": 1.2088332176208496, "learning_rate": 3.83e-05, "loss": 1.1991, "step": 1150 }, { "epoch": 0.6752037252619325, "grad_norm": 1.1784979104995728, "learning_rate": 3.8633333333333335e-05, "loss": 1.2043, "step": 1160 }, { "epoch": 0.681024447031432, "grad_norm": 1.274684190750122, "learning_rate": 3.896666666666667e-05, "loss": 1.2004, "step": 1170 }, { "epoch": 0.6868451688009313, "grad_norm": 1.0404913425445557, "learning_rate": 3.9300000000000007e-05, "loss": 1.198, "step": 1180 }, { "epoch": 0.6926658905704307, "grad_norm": 1.8056018352508545, "learning_rate": 3.963333333333333e-05, "loss": 1.2013, "step": 1190 }, { "epoch": 0.6984866123399301, "grad_norm": 1.635491967201233, "learning_rate": 3.996666666666667e-05, "loss": 1.1927, "step": 1200 }, { "epoch": 0.7043073341094296, "grad_norm": 1.3751031160354614, "learning_rate": 4.0300000000000004e-05, "loss": 1.1962, "step": 1210 }, { "epoch": 0.710128055878929, "grad_norm": 1.2732681035995483, "learning_rate": 4.0633333333333336e-05, "loss": 1.1757, "step": 1220 }, { "epoch": 0.7159487776484285, "grad_norm": 1.2477619647979736, "learning_rate": 4.096666666666667e-05, "loss": 1.1726, "step": 1230 }, { "epoch": 0.7217694994179278, "grad_norm": 0.7919170260429382, "learning_rate": 4.13e-05, "loss": 1.1697, "step": 1240 }, { "epoch": 0.7275902211874272, "grad_norm": 1.2502204179763794, "learning_rate": 4.1633333333333333e-05, "loss": 1.1714, "step": 1250 }, { "epoch": 0.7334109429569267, "grad_norm": 1.1982324123382568, "learning_rate": 4.196666666666667e-05, "loss": 1.1648, "step": 1260 }, { "epoch": 0.7392316647264261, "grad_norm": 1.118094801902771, "learning_rate": 4.23e-05, "loss": 1.1825, "step": 1270 }, { "epoch": 0.7450523864959255, "grad_norm": 1.0838721990585327, "learning_rate": 4.263333333333334e-05, "loss": 1.1682, "step": 1280 }, { "epoch": 0.7508731082654249, "grad_norm": 1.102858543395996, "learning_rate": 4.296666666666666e-05, "loss": 1.1635, "step": 1290 }, { "epoch": 0.7566938300349243, "grad_norm": 1.0336270332336426, "learning_rate": 4.33e-05, "loss": 1.16, "step": 1300 }, { "epoch": 0.7625145518044237, "grad_norm": 0.9466416835784912, "learning_rate": 4.3633333333333335e-05, "loss": 1.1533, "step": 1310 }, { "epoch": 0.7683352735739232, "grad_norm": 0.938991367816925, "learning_rate": 4.396666666666667e-05, "loss": 1.1432, "step": 1320 }, { "epoch": 0.7741559953434226, "grad_norm": 0.9698126316070557, "learning_rate": 4.43e-05, "loss": 1.1518, "step": 1330 }, { "epoch": 0.779976717112922, "grad_norm": 1.1571011543273926, "learning_rate": 4.463333333333334e-05, "loss": 1.154, "step": 1340 }, { "epoch": 0.7857974388824214, "grad_norm": 1.1788208484649658, "learning_rate": 4.496666666666667e-05, "loss": 1.1386, "step": 1350 }, { "epoch": 0.7916181606519208, "grad_norm": 1.057041049003601, "learning_rate": 4.53e-05, "loss": 1.1318, "step": 1360 }, { "epoch": 0.7974388824214202, "grad_norm": 1.1668750047683716, "learning_rate": 4.5633333333333336e-05, "loss": 1.1183, "step": 1370 }, { "epoch": 0.8032596041909197, "grad_norm": 1.0052945613861084, "learning_rate": 4.596666666666667e-05, "loss": 1.1301, "step": 1380 }, { "epoch": 0.8090803259604191, "grad_norm": 0.9670915007591248, "learning_rate": 4.630000000000001e-05, "loss": 1.1165, "step": 1390 }, { "epoch": 0.8149010477299186, "grad_norm": 1.0529544353485107, "learning_rate": 4.663333333333333e-05, "loss": 1.1204, "step": 1400 }, { "epoch": 0.8207217694994179, "grad_norm": 0.8462616205215454, "learning_rate": 4.696666666666667e-05, "loss": 1.1091, "step": 1410 }, { "epoch": 0.8265424912689173, "grad_norm": 0.9274290800094604, "learning_rate": 4.73e-05, "loss": 1.111, "step": 1420 }, { "epoch": 0.8323632130384168, "grad_norm": 0.9385411143302917, "learning_rate": 4.763333333333334e-05, "loss": 1.1056, "step": 1430 }, { "epoch": 0.8381839348079162, "grad_norm": 1.2521352767944336, "learning_rate": 4.796666666666667e-05, "loss": 1.1026, "step": 1440 }, { "epoch": 0.8440046565774156, "grad_norm": 1.1530576944351196, "learning_rate": 4.83e-05, "loss": 1.1086, "step": 1450 }, { "epoch": 0.8498253783469151, "grad_norm": 1.0335489511489868, "learning_rate": 4.8633333333333334e-05, "loss": 1.0962, "step": 1460 }, { "epoch": 0.8556461001164144, "grad_norm": 0.8587660193443298, "learning_rate": 4.8966666666666667e-05, "loss": 1.1009, "step": 1470 }, { "epoch": 0.8614668218859138, "grad_norm": 0.7225843667984009, "learning_rate": 4.93e-05, "loss": 1.1028, "step": 1480 }, { "epoch": 0.8672875436554133, "grad_norm": 0.8470284938812256, "learning_rate": 4.963333333333334e-05, "loss": 1.0864, "step": 1490 }, { "epoch": 0.8731082654249127, "grad_norm": 0.9672794342041016, "learning_rate": 4.996666666666667e-05, "loss": 1.1053, "step": 1500 }, { "epoch": 0.8789289871944121, "grad_norm": 0.8982253074645996, "learning_rate": 5.03e-05, "loss": 1.0839, "step": 1510 }, { "epoch": 0.8847497089639115, "grad_norm": 1.0546019077301025, "learning_rate": 5.0633333333333335e-05, "loss": 1.0761, "step": 1520 }, { "epoch": 0.8905704307334109, "grad_norm": 0.9504612684249878, "learning_rate": 5.0966666666666674e-05, "loss": 1.0699, "step": 1530 }, { "epoch": 0.8963911525029103, "grad_norm": 1.0018041133880615, "learning_rate": 5.130000000000001e-05, "loss": 1.087, "step": 1540 }, { "epoch": 0.9022118742724098, "grad_norm": 0.8834160566329956, "learning_rate": 5.163333333333333e-05, "loss": 1.0617, "step": 1550 }, { "epoch": 0.9080325960419092, "grad_norm": 0.7902634739875793, "learning_rate": 5.196666666666667e-05, "loss": 1.0511, "step": 1560 }, { "epoch": 0.9138533178114087, "grad_norm": 0.8926331400871277, "learning_rate": 5.2300000000000004e-05, "loss": 1.0543, "step": 1570 }, { "epoch": 0.919674039580908, "grad_norm": 1.3002363443374634, "learning_rate": 5.2633333333333336e-05, "loss": 1.0467, "step": 1580 }, { "epoch": 0.9254947613504074, "grad_norm": 1.1655949354171753, "learning_rate": 5.296666666666666e-05, "loss": 1.0563, "step": 1590 }, { "epoch": 0.9313154831199069, "grad_norm": 1.1790984869003296, "learning_rate": 5.330000000000001e-05, "loss": 1.0596, "step": 1600 }, { "epoch": 0.9371362048894063, "grad_norm": 1.2342787981033325, "learning_rate": 5.3633333333333334e-05, "loss": 1.046, "step": 1610 }, { "epoch": 0.9429569266589057, "grad_norm": 1.2507675886154175, "learning_rate": 5.3966666666666666e-05, "loss": 1.0439, "step": 1620 }, { "epoch": 0.9487776484284052, "grad_norm": 0.9425173401832581, "learning_rate": 5.4300000000000005e-05, "loss": 1.0364, "step": 1630 }, { "epoch": 0.9545983701979045, "grad_norm": 1.042507290840149, "learning_rate": 5.463333333333334e-05, "loss": 1.04, "step": 1640 }, { "epoch": 0.9604190919674039, "grad_norm": 1.3278875350952148, "learning_rate": 5.496666666666666e-05, "loss": 1.04, "step": 1650 }, { "epoch": 0.9662398137369034, "grad_norm": 1.0992940664291382, "learning_rate": 5.530000000000001e-05, "loss": 1.0144, "step": 1660 }, { "epoch": 0.9720605355064028, "grad_norm": 0.9809670448303223, "learning_rate": 5.5633333333333335e-05, "loss": 1.022, "step": 1670 }, { "epoch": 0.9778812572759022, "grad_norm": 1.1678181886672974, "learning_rate": 5.596666666666667e-05, "loss": 1.0277, "step": 1680 }, { "epoch": 0.9837019790454016, "grad_norm": 1.109963059425354, "learning_rate": 5.63e-05, "loss": 1.0175, "step": 1690 }, { "epoch": 0.989522700814901, "grad_norm": 1.4591392278671265, "learning_rate": 5.663333333333334e-05, "loss": 1.0086, "step": 1700 }, { "epoch": 0.9953434225844005, "grad_norm": 1.1392290592193604, "learning_rate": 5.696666666666667e-05, "loss": 1.0058, "step": 1710 }, { "epoch": 1.0011641443538999, "grad_norm": 1.12555730342865, "learning_rate": 5.73e-05, "loss": 1.0102, "step": 1720 }, { "epoch": 1.0069848661233993, "grad_norm": 1.0687147378921509, "learning_rate": 5.7633333333333336e-05, "loss": 0.9963, "step": 1730 }, { "epoch": 1.0128055878928988, "grad_norm": 0.9165624976158142, "learning_rate": 5.796666666666667e-05, "loss": 0.9909, "step": 1740 }, { "epoch": 1.0186263096623982, "grad_norm": 1.1080509424209595, "learning_rate": 5.83e-05, "loss": 0.9953, "step": 1750 }, { "epoch": 1.0244470314318976, "grad_norm": 1.0263848304748535, "learning_rate": 5.863333333333334e-05, "loss": 0.9955, "step": 1760 }, { "epoch": 1.030267753201397, "grad_norm": 1.0806018114089966, "learning_rate": 5.896666666666667e-05, "loss": 0.9885, "step": 1770 }, { "epoch": 1.0360884749708963, "grad_norm": 1.037284255027771, "learning_rate": 5.93e-05, "loss": 0.9816, "step": 1780 }, { "epoch": 1.0419091967403957, "grad_norm": 0.9805722832679749, "learning_rate": 5.9633333333333344e-05, "loss": 0.9749, "step": 1790 }, { "epoch": 1.0477299185098952, "grad_norm": 0.9057533740997314, "learning_rate": 5.996666666666667e-05, "loss": 0.9796, "step": 1800 }, { "epoch": 1.0535506402793946, "grad_norm": 1.0553061962127686, "learning_rate": 6.03e-05, "loss": 0.9623, "step": 1810 }, { "epoch": 1.059371362048894, "grad_norm": 1.1339528560638428, "learning_rate": 6.063333333333333e-05, "loss": 0.9686, "step": 1820 }, { "epoch": 1.0651920838183935, "grad_norm": 1.008010745048523, "learning_rate": 6.0966666666666674e-05, "loss": 0.9616, "step": 1830 }, { "epoch": 1.071012805587893, "grad_norm": 1.019221544265747, "learning_rate": 6.13e-05, "loss": 0.9474, "step": 1840 }, { "epoch": 1.0768335273573924, "grad_norm": 1.1472612619400024, "learning_rate": 6.163333333333333e-05, "loss": 0.9574, "step": 1850 }, { "epoch": 1.0826542491268918, "grad_norm": 1.126043677330017, "learning_rate": 6.196666666666668e-05, "loss": 0.9743, "step": 1860 }, { "epoch": 1.0884749708963912, "grad_norm": 1.0530799627304077, "learning_rate": 6.23e-05, "loss": 0.9574, "step": 1870 }, { "epoch": 1.0942956926658907, "grad_norm": 1.1186445951461792, "learning_rate": 6.263333333333333e-05, "loss": 0.9435, "step": 1880 }, { "epoch": 1.1001164144353899, "grad_norm": 1.2041856050491333, "learning_rate": 6.296666666666667e-05, "loss": 0.9411, "step": 1890 }, { "epoch": 1.1059371362048893, "grad_norm": 1.115175724029541, "learning_rate": 6.330000000000001e-05, "loss": 0.9486, "step": 1900 }, { "epoch": 1.1117578579743888, "grad_norm": 1.301000952720642, "learning_rate": 6.363333333333334e-05, "loss": 0.9483, "step": 1910 }, { "epoch": 1.1175785797438882, "grad_norm": 1.3491665124893188, "learning_rate": 6.396666666666667e-05, "loss": 0.9444, "step": 1920 }, { "epoch": 1.1233993015133876, "grad_norm": 1.1432496309280396, "learning_rate": 6.43e-05, "loss": 0.9354, "step": 1930 }, { "epoch": 1.129220023282887, "grad_norm": 1.134177327156067, "learning_rate": 6.463333333333334e-05, "loss": 0.936, "step": 1940 }, { "epoch": 1.1350407450523865, "grad_norm": 1.0438448190689087, "learning_rate": 6.496666666666667e-05, "loss": 0.924, "step": 1950 }, { "epoch": 1.140861466821886, "grad_norm": 1.1216626167297363, "learning_rate": 6.53e-05, "loss": 0.9096, "step": 1960 }, { "epoch": 1.1466821885913854, "grad_norm": 1.1513433456420898, "learning_rate": 6.563333333333333e-05, "loss": 0.9374, "step": 1970 }, { "epoch": 1.1525029103608848, "grad_norm": 1.2888542413711548, "learning_rate": 6.596666666666667e-05, "loss": 0.9458, "step": 1980 }, { "epoch": 1.1583236321303843, "grad_norm": 1.1150907278060913, "learning_rate": 6.630000000000001e-05, "loss": 0.9098, "step": 1990 }, { "epoch": 1.1641443538998835, "grad_norm": 1.1481194496154785, "learning_rate": 6.663333333333333e-05, "loss": 0.9032, "step": 2000 }, { "epoch": 1.1699650756693831, "grad_norm": 1.2200440168380737, "learning_rate": 6.696666666666666e-05, "loss": 0.9034, "step": 2010 }, { "epoch": 1.1757857974388823, "grad_norm": 1.0101317167282104, "learning_rate": 6.730000000000001e-05, "loss": 0.9047, "step": 2020 }, { "epoch": 1.1816065192083818, "grad_norm": 1.3747870922088623, "learning_rate": 6.763333333333334e-05, "loss": 0.9059, "step": 2030 }, { "epoch": 1.1874272409778812, "grad_norm": 1.3034472465515137, "learning_rate": 6.796666666666666e-05, "loss": 0.9225, "step": 2040 }, { "epoch": 1.1932479627473807, "grad_norm": 1.1037561893463135, "learning_rate": 6.83e-05, "loss": 0.9072, "step": 2050 }, { "epoch": 1.19906868451688, "grad_norm": 1.2846330404281616, "learning_rate": 6.863333333333334e-05, "loss": 0.8841, "step": 2060 }, { "epoch": 1.2048894062863795, "grad_norm": 1.1986215114593506, "learning_rate": 6.896666666666667e-05, "loss": 0.8915, "step": 2070 }, { "epoch": 1.210710128055879, "grad_norm": 1.1313127279281616, "learning_rate": 6.93e-05, "loss": 0.8908, "step": 2080 }, { "epoch": 1.2165308498253784, "grad_norm": 1.23441743850708, "learning_rate": 6.963333333333334e-05, "loss": 0.8802, "step": 2090 }, { "epoch": 1.2223515715948778, "grad_norm": 1.095381498336792, "learning_rate": 6.996666666666667e-05, "loss": 0.8791, "step": 2100 }, { "epoch": 1.2281722933643773, "grad_norm": 1.0278679132461548, "learning_rate": 7.03e-05, "loss": 0.8715, "step": 2110 }, { "epoch": 1.2339930151338767, "grad_norm": 1.0416446924209595, "learning_rate": 7.063333333333333e-05, "loss": 0.8765, "step": 2120 }, { "epoch": 1.239813736903376, "grad_norm": 1.2162599563598633, "learning_rate": 7.096666666666667e-05, "loss": 0.8798, "step": 2130 }, { "epoch": 1.2456344586728754, "grad_norm": 1.0502229928970337, "learning_rate": 7.13e-05, "loss": 0.872, "step": 2140 }, { "epoch": 1.2514551804423748, "grad_norm": 1.2014297246932983, "learning_rate": 7.163333333333334e-05, "loss": 0.8783, "step": 2150 }, { "epoch": 1.2572759022118742, "grad_norm": 1.3122318983078003, "learning_rate": 7.196666666666668e-05, "loss": 0.8589, "step": 2160 }, { "epoch": 1.2630966239813737, "grad_norm": 1.4843741655349731, "learning_rate": 7.23e-05, "loss": 0.8529, "step": 2170 }, { "epoch": 1.2689173457508731, "grad_norm": 1.3992562294006348, "learning_rate": 7.263333333333334e-05, "loss": 0.8824, "step": 2180 }, { "epoch": 1.2747380675203726, "grad_norm": 1.5506564378738403, "learning_rate": 7.296666666666667e-05, "loss": 0.859, "step": 2190 }, { "epoch": 1.280558789289872, "grad_norm": 1.1957820653915405, "learning_rate": 7.33e-05, "loss": 0.8532, "step": 2200 }, { "epoch": 1.2863795110593714, "grad_norm": 1.2247660160064697, "learning_rate": 7.363333333333334e-05, "loss": 0.851, "step": 2210 }, { "epoch": 1.2922002328288706, "grad_norm": 1.197928786277771, "learning_rate": 7.396666666666667e-05, "loss": 0.8505, "step": 2220 }, { "epoch": 1.2980209545983703, "grad_norm": 1.1263822317123413, "learning_rate": 7.43e-05, "loss": 0.8605, "step": 2230 }, { "epoch": 1.3038416763678695, "grad_norm": 1.2196636199951172, "learning_rate": 7.463333333333334e-05, "loss": 0.8553, "step": 2240 }, { "epoch": 1.309662398137369, "grad_norm": 1.287122130393982, "learning_rate": 7.496666666666667e-05, "loss": 0.8686, "step": 2250 }, { "epoch": 1.3154831199068684, "grad_norm": 1.2078757286071777, "learning_rate": 7.53e-05, "loss": 0.8688, "step": 2260 }, { "epoch": 1.3213038416763678, "grad_norm": 1.5107771158218384, "learning_rate": 7.563333333333333e-05, "loss": 0.845, "step": 2270 }, { "epoch": 1.3271245634458673, "grad_norm": 1.2548962831497192, "learning_rate": 7.596666666666668e-05, "loss": 0.868, "step": 2280 }, { "epoch": 1.3329452852153667, "grad_norm": 1.3212847709655762, "learning_rate": 7.630000000000001e-05, "loss": 0.86, "step": 2290 }, { "epoch": 1.3387660069848661, "grad_norm": 1.2771493196487427, "learning_rate": 7.663333333333333e-05, "loss": 0.8313, "step": 2300 }, { "epoch": 1.3445867287543656, "grad_norm": 1.1689176559448242, "learning_rate": 7.696666666666668e-05, "loss": 0.8496, "step": 2310 }, { "epoch": 1.350407450523865, "grad_norm": 1.3042426109313965, "learning_rate": 7.730000000000001e-05, "loss": 0.8245, "step": 2320 }, { "epoch": 1.3562281722933645, "grad_norm": 1.2430737018585205, "learning_rate": 7.763333333333334e-05, "loss": 0.8443, "step": 2330 }, { "epoch": 1.362048894062864, "grad_norm": 1.3027112483978271, "learning_rate": 7.796666666666666e-05, "loss": 0.8265, "step": 2340 }, { "epoch": 1.367869615832363, "grad_norm": 1.4116811752319336, "learning_rate": 7.83e-05, "loss": 0.8332, "step": 2350 }, { "epoch": 1.3736903376018628, "grad_norm": 1.3568965196609497, "learning_rate": 7.863333333333334e-05, "loss": 0.8203, "step": 2360 }, { "epoch": 1.379511059371362, "grad_norm": 1.1511495113372803, "learning_rate": 7.896666666666667e-05, "loss": 0.8335, "step": 2370 }, { "epoch": 1.3853317811408614, "grad_norm": 1.3709437847137451, "learning_rate": 7.93e-05, "loss": 0.8141, "step": 2380 }, { "epoch": 1.3911525029103609, "grad_norm": 1.137596607208252, "learning_rate": 7.963333333333334e-05, "loss": 0.8333, "step": 2390 }, { "epoch": 1.3969732246798603, "grad_norm": 1.401602029800415, "learning_rate": 7.996666666666667e-05, "loss": 0.8265, "step": 2400 }, { "epoch": 1.4027939464493597, "grad_norm": 1.5351885557174683, "learning_rate": 8.030000000000001e-05, "loss": 0.8095, "step": 2410 }, { "epoch": 1.4086146682188592, "grad_norm": 1.1196638345718384, "learning_rate": 8.063333333333333e-05, "loss": 0.8095, "step": 2420 }, { "epoch": 1.4144353899883586, "grad_norm": 1.3064723014831543, "learning_rate": 8.096666666666667e-05, "loss": 0.8019, "step": 2430 }, { "epoch": 1.420256111757858, "grad_norm": 1.3565727472305298, "learning_rate": 8.13e-05, "loss": 0.8204, "step": 2440 }, { "epoch": 1.4260768335273575, "grad_norm": 1.284519910812378, "learning_rate": 8.163333333333334e-05, "loss": 0.8116, "step": 2450 }, { "epoch": 1.4318975552968567, "grad_norm": 1.4860012531280518, "learning_rate": 8.196666666666668e-05, "loss": 0.8223, "step": 2460 }, { "epoch": 1.4377182770663564, "grad_norm": 1.2199811935424805, "learning_rate": 8.23e-05, "loss": 0.8105, "step": 2470 }, { "epoch": 1.4435389988358556, "grad_norm": 1.2995200157165527, "learning_rate": 8.263333333333334e-05, "loss": 0.804, "step": 2480 }, { "epoch": 1.449359720605355, "grad_norm": 1.2752565145492554, "learning_rate": 8.296666666666667e-05, "loss": 0.783, "step": 2490 }, { "epoch": 1.4551804423748544, "grad_norm": 1.2391401529312134, "learning_rate": 8.33e-05, "loss": 0.8062, "step": 2500 }, { "epoch": 1.4610011641443539, "grad_norm": 1.1769263744354248, "learning_rate": 8.363333333333334e-05, "loss": 0.7922, "step": 2510 }, { "epoch": 1.4668218859138533, "grad_norm": 1.2160168886184692, "learning_rate": 8.396666666666667e-05, "loss": 0.7811, "step": 2520 }, { "epoch": 1.4726426076833528, "grad_norm": 1.3661754131317139, "learning_rate": 8.43e-05, "loss": 0.7734, "step": 2530 }, { "epoch": 1.4784633294528522, "grad_norm": 1.44180166721344, "learning_rate": 8.463333333333335e-05, "loss": 0.8035, "step": 2540 }, { "epoch": 1.4842840512223516, "grad_norm": 1.2356929779052734, "learning_rate": 8.496666666666667e-05, "loss": 0.7939, "step": 2550 }, { "epoch": 1.490104772991851, "grad_norm": 1.2838892936706543, "learning_rate": 8.53e-05, "loss": 0.7788, "step": 2560 }, { "epoch": 1.4959254947613503, "grad_norm": 1.2296321392059326, "learning_rate": 8.563333333333333e-05, "loss": 0.782, "step": 2570 }, { "epoch": 1.50174621653085, "grad_norm": 1.3221367597579956, "learning_rate": 8.596666666666668e-05, "loss": 0.7782, "step": 2580 }, { "epoch": 1.5075669383003492, "grad_norm": 1.215658187866211, "learning_rate": 8.63e-05, "loss": 0.7874, "step": 2590 }, { "epoch": 1.5133876600698488, "grad_norm": 1.2039812803268433, "learning_rate": 8.663333333333333e-05, "loss": 0.7801, "step": 2600 }, { "epoch": 1.519208381839348, "grad_norm": 1.1610695123672485, "learning_rate": 8.696666666666668e-05, "loss": 0.7663, "step": 2610 }, { "epoch": 1.5250291036088475, "grad_norm": 1.2785632610321045, "learning_rate": 8.730000000000001e-05, "loss": 0.7729, "step": 2620 }, { "epoch": 1.530849825378347, "grad_norm": 1.1557338237762451, "learning_rate": 8.763333333333334e-05, "loss": 0.7605, "step": 2630 }, { "epoch": 1.5366705471478463, "grad_norm": 1.242073655128479, "learning_rate": 8.796666666666667e-05, "loss": 0.7556, "step": 2640 }, { "epoch": 1.5424912689173458, "grad_norm": 1.3477911949157715, "learning_rate": 8.83e-05, "loss": 0.7711, "step": 2650 }, { "epoch": 1.5483119906868452, "grad_norm": 1.2573416233062744, "learning_rate": 8.863333333333334e-05, "loss": 0.7855, "step": 2660 }, { "epoch": 1.5541327124563447, "grad_norm": 1.143362283706665, "learning_rate": 8.896666666666667e-05, "loss": 0.7784, "step": 2670 }, { "epoch": 1.5599534342258439, "grad_norm": 1.2178300619125366, "learning_rate": 8.93e-05, "loss": 0.7716, "step": 2680 }, { "epoch": 1.5657741559953435, "grad_norm": 1.404107928276062, "learning_rate": 8.963333333333333e-05, "loss": 0.762, "step": 2690 }, { "epoch": 1.5715948777648427, "grad_norm": 1.344296932220459, "learning_rate": 8.996666666666667e-05, "loss": 0.7732, "step": 2700 }, { "epoch": 1.5774155995343424, "grad_norm": 1.2628093957901, "learning_rate": 9.030000000000001e-05, "loss": 0.7593, "step": 2710 }, { "epoch": 1.5832363213038416, "grad_norm": 1.3049347400665283, "learning_rate": 9.063333333333333e-05, "loss": 0.7536, "step": 2720 }, { "epoch": 1.589057043073341, "grad_norm": 1.54417085647583, "learning_rate": 9.096666666666666e-05, "loss": 0.7633, "step": 2730 }, { "epoch": 1.5948777648428405, "grad_norm": 1.3994975090026855, "learning_rate": 9.130000000000001e-05, "loss": 0.7543, "step": 2740 }, { "epoch": 1.60069848661234, "grad_norm": 1.3144352436065674, "learning_rate": 9.163333333333334e-05, "loss": 0.7438, "step": 2750 }, { "epoch": 1.6065192083818394, "grad_norm": 1.2317805290222168, "learning_rate": 9.196666666666666e-05, "loss": 0.7458, "step": 2760 }, { "epoch": 1.6123399301513388, "grad_norm": 1.5012415647506714, "learning_rate": 9.230000000000001e-05, "loss": 0.7434, "step": 2770 }, { "epoch": 1.6181606519208382, "grad_norm": 1.2595512866973877, "learning_rate": 9.263333333333334e-05, "loss": 0.7443, "step": 2780 }, { "epoch": 1.6239813736903375, "grad_norm": 1.57895827293396, "learning_rate": 9.296666666666667e-05, "loss": 0.7394, "step": 2790 }, { "epoch": 1.6298020954598371, "grad_norm": 1.3639267683029175, "learning_rate": 9.33e-05, "loss": 0.7482, "step": 2800 }, { "epoch": 1.6356228172293363, "grad_norm": 1.2522956132888794, "learning_rate": 9.363333333333334e-05, "loss": 0.7367, "step": 2810 }, { "epoch": 1.641443538998836, "grad_norm": 1.45496666431427, "learning_rate": 9.396666666666667e-05, "loss": 0.7276, "step": 2820 }, { "epoch": 1.6472642607683352, "grad_norm": 1.328822135925293, "learning_rate": 9.43e-05, "loss": 0.7364, "step": 2830 }, { "epoch": 1.6530849825378346, "grad_norm": 1.375406265258789, "learning_rate": 9.463333333333333e-05, "loss": 0.7379, "step": 2840 }, { "epoch": 1.658905704307334, "grad_norm": 1.2512245178222656, "learning_rate": 9.496666666666667e-05, "loss": 0.7365, "step": 2850 }, { "epoch": 1.6647264260768335, "grad_norm": 1.2965887784957886, "learning_rate": 9.53e-05, "loss": 0.7194, "step": 2860 }, { "epoch": 1.670547147846333, "grad_norm": 1.3652057647705078, "learning_rate": 9.563333333333334e-05, "loss": 0.7292, "step": 2870 }, { "epoch": 1.6763678696158324, "grad_norm": 1.4441159963607788, "learning_rate": 9.596666666666668e-05, "loss": 0.718, "step": 2880 }, { "epoch": 1.6821885913853318, "grad_norm": 1.1431790590286255, "learning_rate": 9.63e-05, "loss": 0.7193, "step": 2890 }, { "epoch": 1.688009313154831, "grad_norm": 1.5495961904525757, "learning_rate": 9.663333333333334e-05, "loss": 0.7315, "step": 2900 }, { "epoch": 1.6938300349243307, "grad_norm": 1.360645055770874, "learning_rate": 9.696666666666667e-05, "loss": 0.7488, "step": 2910 }, { "epoch": 1.69965075669383, "grad_norm": 1.3098105192184448, "learning_rate": 9.730000000000001e-05, "loss": 0.7254, "step": 2920 }, { "epoch": 1.7054714784633296, "grad_norm": 1.4514962434768677, "learning_rate": 9.763333333333334e-05, "loss": 0.7099, "step": 2930 }, { "epoch": 1.7112922002328288, "grad_norm": 1.4968982934951782, "learning_rate": 9.796666666666667e-05, "loss": 0.7114, "step": 2940 }, { "epoch": 1.7171129220023282, "grad_norm": 1.1940176486968994, "learning_rate": 9.83e-05, "loss": 0.7125, "step": 2950 }, { "epoch": 1.7229336437718277, "grad_norm": 1.3205413818359375, "learning_rate": 9.863333333333334e-05, "loss": 0.7395, "step": 2960 }, { "epoch": 1.728754365541327, "grad_norm": 1.4134654998779297, "learning_rate": 9.896666666666667e-05, "loss": 0.719, "step": 2970 }, { "epoch": 1.7345750873108265, "grad_norm": 1.3998537063598633, "learning_rate": 9.93e-05, "loss": 0.7087, "step": 2980 }, { "epoch": 1.740395809080326, "grad_norm": 1.2147654294967651, "learning_rate": 9.963333333333333e-05, "loss": 0.7083, "step": 2990 }, { "epoch": 1.7462165308498254, "grad_norm": 1.526343822479248, "learning_rate": 9.996666666666668e-05, "loss": 0.7236, "step": 3000 }, { "epoch": 1.7520372526193246, "grad_norm": 1.3985742330551147, "learning_rate": 9.999999384858465e-05, "loss": 0.7038, "step": 3010 }, { "epoch": 1.7578579743888243, "grad_norm": 1.3096967935562134, "learning_rate": 9.999997258443473e-05, "loss": 0.721, "step": 3020 }, { "epoch": 1.7636786961583235, "grad_norm": 1.268714189529419, "learning_rate": 9.999993613161331e-05, "loss": 0.7075, "step": 3030 }, { "epoch": 1.7694994179278232, "grad_norm": 1.2963391542434692, "learning_rate": 9.999988449013146e-05, "loss": 0.6937, "step": 3040 }, { "epoch": 1.7753201396973224, "grad_norm": 1.3826044797897339, "learning_rate": 9.99998176600049e-05, "loss": 0.7089, "step": 3050 }, { "epoch": 1.781140861466822, "grad_norm": 1.4136563539505005, "learning_rate": 9.999973564125389e-05, "loss": 0.7069, "step": 3060 }, { "epoch": 1.7869615832363213, "grad_norm": 1.4132513999938965, "learning_rate": 9.999963843390335e-05, "loss": 0.7157, "step": 3070 }, { "epoch": 1.7927823050058207, "grad_norm": 1.4122391939163208, "learning_rate": 9.999952603798282e-05, "loss": 0.7025, "step": 3080 }, { "epoch": 1.7986030267753201, "grad_norm": 1.565251350402832, "learning_rate": 9.999939845352646e-05, "loss": 0.7111, "step": 3090 }, { "epoch": 1.8044237485448196, "grad_norm": 1.3846335411071777, "learning_rate": 9.999925568057298e-05, "loss": 0.7265, "step": 3100 }, { "epoch": 1.810244470314319, "grad_norm": 1.4768803119659424, "learning_rate": 9.999909771916578e-05, "loss": 0.7316, "step": 3110 }, { "epoch": 1.8160651920838184, "grad_norm": 1.360030174255371, "learning_rate": 9.999892456935285e-05, "loss": 0.6923, "step": 3120 }, { "epoch": 1.8218859138533179, "grad_norm": 1.4551498889923096, "learning_rate": 9.999873623118679e-05, "loss": 0.7055, "step": 3130 }, { "epoch": 1.827706635622817, "grad_norm": 1.6587899923324585, "learning_rate": 9.999853270472479e-05, "loss": 0.7129, "step": 3140 }, { "epoch": 1.8335273573923168, "grad_norm": 1.3100780248641968, "learning_rate": 9.999831399002871e-05, "loss": 0.7006, "step": 3150 }, { "epoch": 1.839348079161816, "grad_norm": 1.7005400657653809, "learning_rate": 9.999808008716494e-05, "loss": 0.7042, "step": 3160 }, { "epoch": 1.8451688009313156, "grad_norm": 1.333143949508667, "learning_rate": 9.999783099620459e-05, "loss": 0.7148, "step": 3170 }, { "epoch": 1.8509895227008148, "grad_norm": 1.319303274154663, "learning_rate": 9.999756671722328e-05, "loss": 0.7184, "step": 3180 }, { "epoch": 1.8568102444703143, "grad_norm": 1.5092350244522095, "learning_rate": 9.99972872503013e-05, "loss": 0.7005, "step": 3190 }, { "epoch": 1.8626309662398137, "grad_norm": 1.4707974195480347, "learning_rate": 9.999699259552359e-05, "loss": 0.721, "step": 3200 }, { "epoch": 1.8684516880093132, "grad_norm": 1.2856364250183105, "learning_rate": 9.99966827529796e-05, "loss": 0.7046, "step": 3210 }, { "epoch": 1.8742724097788126, "grad_norm": 1.4254980087280273, "learning_rate": 9.999635772276348e-05, "loss": 0.7005, "step": 3220 }, { "epoch": 1.880093131548312, "grad_norm": 1.3656672239303589, "learning_rate": 9.999601750497396e-05, "loss": 0.6911, "step": 3230 }, { "epoch": 1.8859138533178115, "grad_norm": 1.5718532800674438, "learning_rate": 9.99956620997144e-05, "loss": 0.6772, "step": 3240 }, { "epoch": 1.8917345750873107, "grad_norm": 1.5857778787612915, "learning_rate": 9.999529150709275e-05, "loss": 0.6778, "step": 3250 }, { "epoch": 1.8975552968568103, "grad_norm": 1.5545802116394043, "learning_rate": 9.999490572722158e-05, "loss": 0.7119, "step": 3260 }, { "epoch": 1.9033760186263096, "grad_norm": 1.5440335273742676, "learning_rate": 9.99945047602181e-05, "loss": 0.691, "step": 3270 }, { "epoch": 1.9091967403958092, "grad_norm": 1.313011884689331, "learning_rate": 9.99940886062041e-05, "loss": 0.6785, "step": 3280 }, { "epoch": 1.9150174621653084, "grad_norm": 1.1551698446273804, "learning_rate": 9.999365726530599e-05, "loss": 0.6922, "step": 3290 }, { "epoch": 1.9208381839348079, "grad_norm": 1.4899983406066895, "learning_rate": 9.999321073765481e-05, "loss": 0.6788, "step": 3300 }, { "epoch": 1.9266589057043073, "grad_norm": 1.3196216821670532, "learning_rate": 9.99927490233862e-05, "loss": 0.6747, "step": 3310 }, { "epoch": 1.9324796274738067, "grad_norm": 1.4007161855697632, "learning_rate": 9.999227212264043e-05, "loss": 0.681, "step": 3320 }, { "epoch": 1.9383003492433062, "grad_norm": 1.4584583044052124, "learning_rate": 9.999178003556236e-05, "loss": 0.6859, "step": 3330 }, { "epoch": 1.9441210710128056, "grad_norm": 1.2124409675598145, "learning_rate": 9.999127276230146e-05, "loss": 0.6619, "step": 3340 }, { "epoch": 1.949941792782305, "grad_norm": 1.2438045740127563, "learning_rate": 9.999075030301184e-05, "loss": 0.6692, "step": 3350 }, { "epoch": 1.9557625145518043, "grad_norm": 1.4052081108093262, "learning_rate": 9.999021265785221e-05, "loss": 0.6635, "step": 3360 }, { "epoch": 1.961583236321304, "grad_norm": 1.345312237739563, "learning_rate": 9.998965982698589e-05, "loss": 0.6996, "step": 3370 }, { "epoch": 1.9674039580908032, "grad_norm": 1.3642547130584717, "learning_rate": 9.998909181058082e-05, "loss": 0.6679, "step": 3380 }, { "epoch": 1.9732246798603028, "grad_norm": 1.440560221672058, "learning_rate": 9.998850860880953e-05, "loss": 0.6653, "step": 3390 }, { "epoch": 1.979045401629802, "grad_norm": 1.6368601322174072, "learning_rate": 9.998791022184922e-05, "loss": 0.6764, "step": 3400 }, { "epoch": 1.9848661233993015, "grad_norm": 1.2400139570236206, "learning_rate": 9.99872966498816e-05, "loss": 0.668, "step": 3410 }, { "epoch": 1.990686845168801, "grad_norm": 1.4596027135849, "learning_rate": 9.998666789309313e-05, "loss": 0.6718, "step": 3420 }, { "epoch": 1.9965075669383003, "grad_norm": 1.2639851570129395, "learning_rate": 9.998602395167475e-05, "loss": 0.654, "step": 3430 }, { "epoch": 2.0023282887077998, "grad_norm": 1.4087293148040771, "learning_rate": 9.998536482582213e-05, "loss": 0.6609, "step": 3440 }, { "epoch": 2.008149010477299, "grad_norm": 1.5746890306472778, "learning_rate": 9.998469051573544e-05, "loss": 0.6554, "step": 3450 }, { "epoch": 2.0139697322467986, "grad_norm": 1.3763000965118408, "learning_rate": 9.998400102161954e-05, "loss": 0.6635, "step": 3460 }, { "epoch": 2.019790454016298, "grad_norm": 1.2402838468551636, "learning_rate": 9.998329634368388e-05, "loss": 0.6515, "step": 3470 }, { "epoch": 2.0256111757857975, "grad_norm": 1.4687516689300537, "learning_rate": 9.998257648214253e-05, "loss": 0.6555, "step": 3480 }, { "epoch": 2.0314318975552967, "grad_norm": 1.165253758430481, "learning_rate": 9.998184143721417e-05, "loss": 0.6473, "step": 3490 }, { "epoch": 2.0372526193247964, "grad_norm": 1.383156657218933, "learning_rate": 9.998109120912206e-05, "loss": 0.663, "step": 3500 }, { "epoch": 2.0430733410942956, "grad_norm": 1.1924165487289429, "learning_rate": 9.998032579809411e-05, "loss": 0.6494, "step": 3510 }, { "epoch": 2.0488940628637953, "grad_norm": 1.092405080795288, "learning_rate": 9.997954520436286e-05, "loss": 0.6407, "step": 3520 }, { "epoch": 2.0547147846332945, "grad_norm": 1.33918035030365, "learning_rate": 9.997874942816538e-05, "loss": 0.6469, "step": 3530 }, { "epoch": 2.060535506402794, "grad_norm": 1.2722212076187134, "learning_rate": 9.997793846974345e-05, "loss": 0.663, "step": 3540 }, { "epoch": 2.0663562281722934, "grad_norm": 1.3158340454101562, "learning_rate": 9.997711232934341e-05, "loss": 0.6615, "step": 3550 }, { "epoch": 2.0721769499417926, "grad_norm": 1.578774094581604, "learning_rate": 9.99762710072162e-05, "loss": 0.6474, "step": 3560 }, { "epoch": 2.0779976717112922, "grad_norm": 1.599096417427063, "learning_rate": 9.997541450361743e-05, "loss": 0.6727, "step": 3570 }, { "epoch": 2.0838183934807915, "grad_norm": 1.2816153764724731, "learning_rate": 9.997454281880723e-05, "loss": 0.6389, "step": 3580 }, { "epoch": 2.089639115250291, "grad_norm": 1.4457660913467407, "learning_rate": 9.997365595305044e-05, "loss": 0.6452, "step": 3590 }, { "epoch": 2.0954598370197903, "grad_norm": 1.2593332529067993, "learning_rate": 9.997275390661644e-05, "loss": 0.6529, "step": 3600 }, { "epoch": 2.10128055878929, "grad_norm": 1.167940616607666, "learning_rate": 9.997183667977926e-05, "loss": 0.6348, "step": 3610 }, { "epoch": 2.107101280558789, "grad_norm": 1.39714777469635, "learning_rate": 9.997090427281752e-05, "loss": 0.6448, "step": 3620 }, { "epoch": 2.112922002328289, "grad_norm": 1.2353298664093018, "learning_rate": 9.996995668601448e-05, "loss": 0.6364, "step": 3630 }, { "epoch": 2.118742724097788, "grad_norm": 1.4067635536193848, "learning_rate": 9.996899391965798e-05, "loss": 0.6383, "step": 3640 }, { "epoch": 2.1245634458672877, "grad_norm": 1.342576026916504, "learning_rate": 9.996801597404048e-05, "loss": 0.6466, "step": 3650 }, { "epoch": 2.130384167636787, "grad_norm": 1.4651687145233154, "learning_rate": 9.996702284945905e-05, "loss": 0.623, "step": 3660 }, { "epoch": 2.1362048894062866, "grad_norm": 1.3330045938491821, "learning_rate": 9.996601454621539e-05, "loss": 0.6373, "step": 3670 }, { "epoch": 2.142025611175786, "grad_norm": 1.2515190839767456, "learning_rate": 9.996499106461577e-05, "loss": 0.6369, "step": 3680 }, { "epoch": 2.147846332945285, "grad_norm": 1.4541981220245361, "learning_rate": 9.996395240497112e-05, "loss": 0.6541, "step": 3690 }, { "epoch": 2.1536670547147847, "grad_norm": 1.3340394496917725, "learning_rate": 9.996289856759696e-05, "loss": 0.6478, "step": 3700 }, { "epoch": 2.159487776484284, "grad_norm": 1.119783639907837, "learning_rate": 9.996182955281342e-05, "loss": 0.6359, "step": 3710 }, { "epoch": 2.1653084982537836, "grad_norm": 1.6080048084259033, "learning_rate": 9.996074536094519e-05, "loss": 0.6333, "step": 3720 }, { "epoch": 2.171129220023283, "grad_norm": 1.3285688161849976, "learning_rate": 9.995964599232168e-05, "loss": 0.6289, "step": 3730 }, { "epoch": 2.1769499417927825, "grad_norm": 1.1636394262313843, "learning_rate": 9.995853144727683e-05, "loss": 0.6345, "step": 3740 }, { "epoch": 2.1827706635622817, "grad_norm": 1.3392122983932495, "learning_rate": 9.99574017261492e-05, "loss": 0.6325, "step": 3750 }, { "epoch": 2.1885913853317813, "grad_norm": 1.2798062562942505, "learning_rate": 9.995625682928198e-05, "loss": 0.622, "step": 3760 }, { "epoch": 2.1944121071012805, "grad_norm": 1.2522615194320679, "learning_rate": 9.995509675702295e-05, "loss": 0.6397, "step": 3770 }, { "epoch": 2.2002328288707798, "grad_norm": 1.3435559272766113, "learning_rate": 9.995392150972451e-05, "loss": 0.6324, "step": 3780 }, { "epoch": 2.2060535506402794, "grad_norm": 1.2150763273239136, "learning_rate": 9.995273108774366e-05, "loss": 0.6137, "step": 3790 }, { "epoch": 2.2118742724097786, "grad_norm": 1.4472192525863647, "learning_rate": 9.995152549144205e-05, "loss": 0.6266, "step": 3800 }, { "epoch": 2.2176949941792783, "grad_norm": 1.2643681764602661, "learning_rate": 9.995030472118587e-05, "loss": 0.6131, "step": 3810 }, { "epoch": 2.2235157159487775, "grad_norm": 1.1018104553222656, "learning_rate": 9.9949068777346e-05, "loss": 0.6405, "step": 3820 }, { "epoch": 2.229336437718277, "grad_norm": 1.2990959882736206, "learning_rate": 9.994781766029786e-05, "loss": 0.6397, "step": 3830 }, { "epoch": 2.2351571594877764, "grad_norm": 1.3367003202438354, "learning_rate": 9.994655137042151e-05, "loss": 0.622, "step": 3840 }, { "epoch": 2.240977881257276, "grad_norm": 1.4297598600387573, "learning_rate": 9.99452699081016e-05, "loss": 0.6155, "step": 3850 }, { "epoch": 2.2467986030267753, "grad_norm": 1.3733325004577637, "learning_rate": 9.994397327372743e-05, "loss": 0.6141, "step": 3860 }, { "epoch": 2.252619324796275, "grad_norm": 1.4015048742294312, "learning_rate": 9.994266146769286e-05, "loss": 0.625, "step": 3870 }, { "epoch": 2.258440046565774, "grad_norm": 1.2799904346466064, "learning_rate": 9.994133449039642e-05, "loss": 0.6082, "step": 3880 }, { "epoch": 2.264260768335274, "grad_norm": 1.4037492275238037, "learning_rate": 9.993999234224118e-05, "loss": 0.6124, "step": 3890 }, { "epoch": 2.270081490104773, "grad_norm": 1.3427175283432007, "learning_rate": 9.993863502363485e-05, "loss": 0.6273, "step": 3900 }, { "epoch": 2.275902211874272, "grad_norm": 1.2601174116134644, "learning_rate": 9.993726253498976e-05, "loss": 0.6194, "step": 3910 }, { "epoch": 2.281722933643772, "grad_norm": 1.3284143209457397, "learning_rate": 9.993587487672282e-05, "loss": 0.618, "step": 3920 }, { "epoch": 2.287543655413271, "grad_norm": 1.5434640645980835, "learning_rate": 9.993447204925558e-05, "loss": 0.6136, "step": 3930 }, { "epoch": 2.2933643771827708, "grad_norm": 1.4554649591445923, "learning_rate": 9.993305405301416e-05, "loss": 0.6246, "step": 3940 }, { "epoch": 2.29918509895227, "grad_norm": 1.477989673614502, "learning_rate": 9.993162088842935e-05, "loss": 0.6188, "step": 3950 }, { "epoch": 2.3050058207217696, "grad_norm": 1.3366180658340454, "learning_rate": 9.993017255593646e-05, "loss": 0.6178, "step": 3960 }, { "epoch": 2.310826542491269, "grad_norm": 1.3944770097732544, "learning_rate": 9.992870905597548e-05, "loss": 0.6228, "step": 3970 }, { "epoch": 2.3166472642607685, "grad_norm": 1.2747650146484375, "learning_rate": 9.9927230388991e-05, "loss": 0.6065, "step": 3980 }, { "epoch": 2.3224679860302677, "grad_norm": 1.3399463891983032, "learning_rate": 9.992573655543215e-05, "loss": 0.5961, "step": 3990 }, { "epoch": 2.328288707799767, "grad_norm": 1.2757326364517212, "learning_rate": 9.992422755575277e-05, "loss": 0.626, "step": 4000 }, { "epoch": 2.3341094295692666, "grad_norm": 1.1210535764694214, "learning_rate": 9.992270339041123e-05, "loss": 0.6007, "step": 4010 }, { "epoch": 2.3399301513387663, "grad_norm": 1.4192754030227661, "learning_rate": 9.992116405987053e-05, "loss": 0.6151, "step": 4020 }, { "epoch": 2.3457508731082655, "grad_norm": 1.4590175151824951, "learning_rate": 9.991960956459828e-05, "loss": 0.6156, "step": 4030 }, { "epoch": 2.3515715948777647, "grad_norm": 1.2062103748321533, "learning_rate": 9.991803990506669e-05, "loss": 0.6068, "step": 4040 }, { "epoch": 2.3573923166472643, "grad_norm": 1.27995765209198, "learning_rate": 9.991645508175258e-05, "loss": 0.5906, "step": 4050 }, { "epoch": 2.3632130384167636, "grad_norm": 1.3409643173217773, "learning_rate": 9.99148550951374e-05, "loss": 0.615, "step": 4060 }, { "epoch": 2.369033760186263, "grad_norm": 1.3042147159576416, "learning_rate": 9.991323994570716e-05, "loss": 0.6053, "step": 4070 }, { "epoch": 2.3748544819557624, "grad_norm": 1.5942965745925903, "learning_rate": 9.99116096339525e-05, "loss": 0.5925, "step": 4080 }, { "epoch": 2.380675203725262, "grad_norm": 1.2858883142471313, "learning_rate": 9.990996416036869e-05, "loss": 0.6098, "step": 4090 }, { "epoch": 2.3864959254947613, "grad_norm": 1.4350899457931519, "learning_rate": 9.990830352545555e-05, "loss": 0.6242, "step": 4100 }, { "epoch": 2.392316647264261, "grad_norm": 1.396065354347229, "learning_rate": 9.990662772971756e-05, "loss": 0.6144, "step": 4110 }, { "epoch": 2.39813736903376, "grad_norm": 1.4168471097946167, "learning_rate": 9.990493677366376e-05, "loss": 0.5992, "step": 4120 }, { "epoch": 2.4039580908032594, "grad_norm": 1.2521262168884277, "learning_rate": 9.990323065780786e-05, "loss": 0.6017, "step": 4130 }, { "epoch": 2.409778812572759, "grad_norm": 1.4193687438964844, "learning_rate": 9.990150938266808e-05, "loss": 0.6013, "step": 4140 }, { "epoch": 2.4155995343422583, "grad_norm": 1.321452021598816, "learning_rate": 9.989977294876733e-05, "loss": 0.5987, "step": 4150 }, { "epoch": 2.421420256111758, "grad_norm": 1.2061758041381836, "learning_rate": 9.989802135663308e-05, "loss": 0.5778, "step": 4160 }, { "epoch": 2.427240977881257, "grad_norm": 1.2970432043075562, "learning_rate": 9.989625460679743e-05, "loss": 0.6062, "step": 4170 }, { "epoch": 2.433061699650757, "grad_norm": 1.3557347059249878, "learning_rate": 9.989447269979706e-05, "loss": 0.611, "step": 4180 }, { "epoch": 2.438882421420256, "grad_norm": 1.3134125471115112, "learning_rate": 9.989267563617328e-05, "loss": 0.5942, "step": 4190 }, { "epoch": 2.4447031431897557, "grad_norm": 1.5217485427856445, "learning_rate": 9.989086341647198e-05, "loss": 0.5903, "step": 4200 }, { "epoch": 2.450523864959255, "grad_norm": 1.3372539281845093, "learning_rate": 9.988903604124366e-05, "loss": 0.5941, "step": 4210 }, { "epoch": 2.4563445867287546, "grad_norm": 1.3015810251235962, "learning_rate": 9.988719351104343e-05, "loss": 0.5841, "step": 4220 }, { "epoch": 2.4621653084982538, "grad_norm": 1.341296672821045, "learning_rate": 9.9885335826431e-05, "loss": 0.5758, "step": 4230 }, { "epoch": 2.4679860302677534, "grad_norm": 1.3757985830307007, "learning_rate": 9.988346298797071e-05, "loss": 0.61, "step": 4240 }, { "epoch": 2.4738067520372526, "grad_norm": 1.3424694538116455, "learning_rate": 9.988157499623146e-05, "loss": 0.6109, "step": 4250 }, { "epoch": 2.479627473806752, "grad_norm": 1.190618872642517, "learning_rate": 9.987967185178677e-05, "loss": 0.5852, "step": 4260 }, { "epoch": 2.4854481955762515, "grad_norm": 1.2936655282974243, "learning_rate": 9.987775355521476e-05, "loss": 0.5778, "step": 4270 }, { "epoch": 2.4912689173457507, "grad_norm": 1.394820213317871, "learning_rate": 9.987582010709817e-05, "loss": 0.5879, "step": 4280 }, { "epoch": 2.4970896391152504, "grad_norm": 1.254529356956482, "learning_rate": 9.987387150802431e-05, "loss": 0.5788, "step": 4290 }, { "epoch": 2.5029103608847496, "grad_norm": 1.2970088720321655, "learning_rate": 9.987190775858517e-05, "loss": 0.5873, "step": 4300 }, { "epoch": 2.5087310826542493, "grad_norm": 1.6617860794067383, "learning_rate": 9.98699288593772e-05, "loss": 0.5796, "step": 4310 }, { "epoch": 2.5145518044237485, "grad_norm": 1.2412580251693726, "learning_rate": 9.986793481100161e-05, "loss": 0.5718, "step": 4320 }, { "epoch": 2.520372526193248, "grad_norm": 1.307213306427002, "learning_rate": 9.986592561406412e-05, "loss": 0.5767, "step": 4330 }, { "epoch": 2.5261932479627474, "grad_norm": 1.367361307144165, "learning_rate": 9.986390126917503e-05, "loss": 0.5799, "step": 4340 }, { "epoch": 2.5320139697322466, "grad_norm": 1.534752368927002, "learning_rate": 9.986186177694933e-05, "loss": 0.5774, "step": 4350 }, { "epoch": 2.5378346915017462, "grad_norm": 1.5030663013458252, "learning_rate": 9.985980713800656e-05, "loss": 0.5722, "step": 4360 }, { "epoch": 2.543655413271246, "grad_norm": 1.354374647140503, "learning_rate": 9.985773735297084e-05, "loss": 0.5886, "step": 4370 }, { "epoch": 2.549476135040745, "grad_norm": 1.2365741729736328, "learning_rate": 9.985565242247092e-05, "loss": 0.5733, "step": 4380 }, { "epoch": 2.5552968568102443, "grad_norm": 1.3295012712478638, "learning_rate": 9.985355234714016e-05, "loss": 0.5892, "step": 4390 }, { "epoch": 2.561117578579744, "grad_norm": 1.2309541702270508, "learning_rate": 9.985143712761652e-05, "loss": 0.6003, "step": 4400 }, { "epoch": 2.566938300349243, "grad_norm": 1.2548296451568604, "learning_rate": 9.984930676454252e-05, "loss": 0.5681, "step": 4410 }, { "epoch": 2.572759022118743, "grad_norm": 1.551564335823059, "learning_rate": 9.984716125856532e-05, "loss": 0.5882, "step": 4420 }, { "epoch": 2.578579743888242, "grad_norm": 1.4391086101531982, "learning_rate": 9.984500061033667e-05, "loss": 0.5733, "step": 4430 }, { "epoch": 2.5844004656577413, "grad_norm": 1.318411946296692, "learning_rate": 9.984282482051293e-05, "loss": 0.5918, "step": 4440 }, { "epoch": 2.590221187427241, "grad_norm": 1.2200992107391357, "learning_rate": 9.9840633889755e-05, "loss": 0.5654, "step": 4450 }, { "epoch": 2.5960419091967406, "grad_norm": 1.470212697982788, "learning_rate": 9.983842781872848e-05, "loss": 0.5794, "step": 4460 }, { "epoch": 2.60186263096624, "grad_norm": 1.360809564590454, "learning_rate": 9.98362066081035e-05, "loss": 0.5676, "step": 4470 }, { "epoch": 2.607683352735739, "grad_norm": 1.4357531070709229, "learning_rate": 9.983397025855479e-05, "loss": 0.5833, "step": 4480 }, { "epoch": 2.6135040745052387, "grad_norm": 1.3893333673477173, "learning_rate": 9.983171877076171e-05, "loss": 0.5868, "step": 4490 }, { "epoch": 2.619324796274738, "grad_norm": 1.3382655382156372, "learning_rate": 9.98294521454082e-05, "loss": 0.5752, "step": 4500 }, { "epoch": 2.6251455180442376, "grad_norm": 1.3002413511276245, "learning_rate": 9.98271703831828e-05, "loss": 0.5999, "step": 4510 }, { "epoch": 2.630966239813737, "grad_norm": 1.2967215776443481, "learning_rate": 9.982487348477865e-05, "loss": 0.5753, "step": 4520 }, { "epoch": 2.6367869615832364, "grad_norm": 1.4248729944229126, "learning_rate": 9.982256145089347e-05, "loss": 0.5832, "step": 4530 }, { "epoch": 2.6426076833527357, "grad_norm": 1.2417367696762085, "learning_rate": 9.982023428222962e-05, "loss": 0.5693, "step": 4540 }, { "epoch": 2.6484284051222353, "grad_norm": 1.3331282138824463, "learning_rate": 9.981789197949403e-05, "loss": 0.5675, "step": 4550 }, { "epoch": 2.6542491268917345, "grad_norm": 1.303347110748291, "learning_rate": 9.98155345433982e-05, "loss": 0.5604, "step": 4560 }, { "epoch": 2.6600698486612337, "grad_norm": 1.506927490234375, "learning_rate": 9.981316197465831e-05, "loss": 0.5651, "step": 4570 }, { "epoch": 2.6658905704307334, "grad_norm": 1.6198208332061768, "learning_rate": 9.981077427399504e-05, "loss": 0.5636, "step": 4580 }, { "epoch": 2.671711292200233, "grad_norm": 1.1931816339492798, "learning_rate": 9.980837144213371e-05, "loss": 0.576, "step": 4590 }, { "epoch": 2.6775320139697323, "grad_norm": 1.3799370527267456, "learning_rate": 9.980595347980426e-05, "loss": 0.5719, "step": 4600 }, { "epoch": 2.6833527357392315, "grad_norm": 1.4580270051956177, "learning_rate": 9.980352038774119e-05, "loss": 0.5739, "step": 4610 }, { "epoch": 2.689173457508731, "grad_norm": 1.3595774173736572, "learning_rate": 9.98010721666836e-05, "loss": 0.5821, "step": 4620 }, { "epoch": 2.6949941792782304, "grad_norm": 1.3344038724899292, "learning_rate": 9.979860881737523e-05, "loss": 0.5651, "step": 4630 }, { "epoch": 2.70081490104773, "grad_norm": 1.3082424402236938, "learning_rate": 9.979613034056434e-05, "loss": 0.5722, "step": 4640 }, { "epoch": 2.7066356228172292, "grad_norm": 1.2206871509552002, "learning_rate": 9.979363673700386e-05, "loss": 0.5627, "step": 4650 }, { "epoch": 2.712456344586729, "grad_norm": 1.1699495315551758, "learning_rate": 9.979112800745124e-05, "loss": 0.5827, "step": 4660 }, { "epoch": 2.718277066356228, "grad_norm": 1.3052709102630615, "learning_rate": 9.978860415266861e-05, "loss": 0.5718, "step": 4670 }, { "epoch": 2.724097788125728, "grad_norm": 1.3339544534683228, "learning_rate": 9.978606517342262e-05, "loss": 0.5408, "step": 4680 }, { "epoch": 2.729918509895227, "grad_norm": 1.3199431896209717, "learning_rate": 9.978351107048456e-05, "loss": 0.5658, "step": 4690 }, { "epoch": 2.735739231664726, "grad_norm": 1.5720640420913696, "learning_rate": 9.978094184463029e-05, "loss": 0.5689, "step": 4700 }, { "epoch": 2.741559953434226, "grad_norm": 1.2635021209716797, "learning_rate": 9.977835749664029e-05, "loss": 0.5696, "step": 4710 }, { "epoch": 2.7473806752037255, "grad_norm": 1.3324166536331177, "learning_rate": 9.97757580272996e-05, "loss": 0.5713, "step": 4720 }, { "epoch": 2.7532013969732247, "grad_norm": 1.388707160949707, "learning_rate": 9.977314343739786e-05, "loss": 0.5648, "step": 4730 }, { "epoch": 2.759022118742724, "grad_norm": 1.3717365264892578, "learning_rate": 9.977051372772934e-05, "loss": 0.5599, "step": 4740 }, { "epoch": 2.7648428405122236, "grad_norm": 1.2619632482528687, "learning_rate": 9.976786889909286e-05, "loss": 0.5799, "step": 4750 }, { "epoch": 2.770663562281723, "grad_norm": 1.341570496559143, "learning_rate": 9.976520895229185e-05, "loss": 0.5528, "step": 4760 }, { "epoch": 2.7764842840512225, "grad_norm": 1.4657649993896484, "learning_rate": 9.976253388813433e-05, "loss": 0.5638, "step": 4770 }, { "epoch": 2.7823050058207217, "grad_norm": 1.3337174654006958, "learning_rate": 9.975984370743293e-05, "loss": 0.5575, "step": 4780 }, { "epoch": 2.788125727590221, "grad_norm": 1.3726776838302612, "learning_rate": 9.975713841100485e-05, "loss": 0.5467, "step": 4790 }, { "epoch": 2.7939464493597206, "grad_norm": 1.2783139944076538, "learning_rate": 9.975441799967187e-05, "loss": 0.5584, "step": 4800 }, { "epoch": 2.7997671711292202, "grad_norm": 1.3473474979400635, "learning_rate": 9.975168247426039e-05, "loss": 0.5501, "step": 4810 }, { "epoch": 2.8055878928987195, "grad_norm": 1.206786870956421, "learning_rate": 9.974893183560139e-05, "loss": 0.5353, "step": 4820 }, { "epoch": 2.8114086146682187, "grad_norm": 1.3722354173660278, "learning_rate": 9.974616608453045e-05, "loss": 0.5726, "step": 4830 }, { "epoch": 2.8172293364377183, "grad_norm": 1.3391047716140747, "learning_rate": 9.974338522188772e-05, "loss": 0.573, "step": 4840 }, { "epoch": 2.8230500582072175, "grad_norm": 1.4288471937179565, "learning_rate": 9.974058924851797e-05, "loss": 0.5513, "step": 4850 }, { "epoch": 2.828870779976717, "grad_norm": 1.1524862051010132, "learning_rate": 9.973777816527051e-05, "loss": 0.553, "step": 4860 }, { "epoch": 2.8346915017462164, "grad_norm": 1.3390486240386963, "learning_rate": 9.973495197299931e-05, "loss": 0.5417, "step": 4870 }, { "epoch": 2.840512223515716, "grad_norm": 1.3148906230926514, "learning_rate": 9.973211067256287e-05, "loss": 0.5623, "step": 4880 }, { "epoch": 2.8463329452852153, "grad_norm": 1.1290092468261719, "learning_rate": 9.97292542648243e-05, "loss": 0.5592, "step": 4890 }, { "epoch": 2.852153667054715, "grad_norm": 1.393450379371643, "learning_rate": 9.972638275065131e-05, "loss": 0.5521, "step": 4900 }, { "epoch": 2.857974388824214, "grad_norm": 1.2780076265335083, "learning_rate": 9.972349613091621e-05, "loss": 0.5428, "step": 4910 }, { "epoch": 2.8637951105937134, "grad_norm": 1.3113853931427002, "learning_rate": 9.972059440649584e-05, "loss": 0.55, "step": 4920 }, { "epoch": 2.869615832363213, "grad_norm": 1.3131368160247803, "learning_rate": 9.971767757827168e-05, "loss": 0.5703, "step": 4930 }, { "epoch": 2.8754365541327127, "grad_norm": 1.4229904413223267, "learning_rate": 9.971474564712982e-05, "loss": 0.5504, "step": 4940 }, { "epoch": 2.881257275902212, "grad_norm": 1.3515968322753906, "learning_rate": 9.971179861396084e-05, "loss": 0.5538, "step": 4950 }, { "epoch": 2.887077997671711, "grad_norm": 1.4253073930740356, "learning_rate": 9.970883647966003e-05, "loss": 0.53, "step": 4960 }, { "epoch": 2.892898719441211, "grad_norm": 1.2481818199157715, "learning_rate": 9.970585924512717e-05, "loss": 0.5582, "step": 4970 }, { "epoch": 2.89871944121071, "grad_norm": 1.2361618280410767, "learning_rate": 9.970286691126669e-05, "loss": 0.5556, "step": 4980 }, { "epoch": 2.9045401629802097, "grad_norm": 1.3110727071762085, "learning_rate": 9.969985947898756e-05, "loss": 0.5475, "step": 4990 }, { "epoch": 2.910360884749709, "grad_norm": 1.5037683248519897, "learning_rate": 9.969683694920337e-05, "loss": 0.5718, "step": 5000 }, { "epoch": 2.9161816065192085, "grad_norm": 1.3044260740280151, "learning_rate": 9.969379932283228e-05, "loss": 0.5508, "step": 5010 }, { "epoch": 2.9220023282887078, "grad_norm": 1.234344482421875, "learning_rate": 9.969074660079704e-05, "loss": 0.5501, "step": 5020 }, { "epoch": 2.9278230500582074, "grad_norm": 1.3108245134353638, "learning_rate": 9.968767878402501e-05, "loss": 0.5477, "step": 5030 }, { "epoch": 2.9336437718277066, "grad_norm": 1.3185406923294067, "learning_rate": 9.968459587344808e-05, "loss": 0.544, "step": 5040 }, { "epoch": 2.939464493597206, "grad_norm": 1.4878636598587036, "learning_rate": 9.968149787000278e-05, "loss": 0.5528, "step": 5050 }, { "epoch": 2.9452852153667055, "grad_norm": 1.2969067096710205, "learning_rate": 9.967838477463018e-05, "loss": 0.5448, "step": 5060 }, { "epoch": 2.9511059371362047, "grad_norm": 1.5100456476211548, "learning_rate": 9.967525658827597e-05, "loss": 0.5297, "step": 5070 }, { "epoch": 2.9569266589057044, "grad_norm": 1.3019912242889404, "learning_rate": 9.967211331189042e-05, "loss": 0.5431, "step": 5080 }, { "epoch": 2.9627473806752036, "grad_norm": 1.2614737749099731, "learning_rate": 9.966895494642834e-05, "loss": 0.5551, "step": 5090 }, { "epoch": 2.9685681024447033, "grad_norm": 1.3661980628967285, "learning_rate": 9.96657814928492e-05, "loss": 0.5526, "step": 5100 }, { "epoch": 2.9743888242142025, "grad_norm": 1.4067302942276, "learning_rate": 9.966259295211697e-05, "loss": 0.5426, "step": 5110 }, { "epoch": 2.980209545983702, "grad_norm": 1.2591313123703003, "learning_rate": 9.965938932520028e-05, "loss": 0.5718, "step": 5120 }, { "epoch": 2.9860302677532014, "grad_norm": 1.4805954694747925, "learning_rate": 9.965617061307229e-05, "loss": 0.5312, "step": 5130 }, { "epoch": 2.9918509895227006, "grad_norm": 1.5289015769958496, "learning_rate": 9.965293681671077e-05, "loss": 0.5432, "step": 5140 }, { "epoch": 2.9976717112922002, "grad_norm": 1.2731318473815918, "learning_rate": 9.964968793709804e-05, "loss": 0.5423, "step": 5150 }, { "epoch": 3.0034924330616994, "grad_norm": 1.467517375946045, "learning_rate": 9.964642397522106e-05, "loss": 0.5468, "step": 5160 }, { "epoch": 3.009313154831199, "grad_norm": 1.3656615018844604, "learning_rate": 9.96431449320713e-05, "loss": 0.5153, "step": 5170 }, { "epoch": 3.0151338766006983, "grad_norm": 1.3679355382919312, "learning_rate": 9.963985080864486e-05, "loss": 0.5344, "step": 5180 }, { "epoch": 3.020954598370198, "grad_norm": 1.4017375707626343, "learning_rate": 9.96365416059424e-05, "loss": 0.5532, "step": 5190 }, { "epoch": 3.026775320139697, "grad_norm": 1.560525894165039, "learning_rate": 9.963321732496919e-05, "loss": 0.5557, "step": 5200 }, { "epoch": 3.032596041909197, "grad_norm": 1.5008668899536133, "learning_rate": 9.962987796673506e-05, "loss": 0.5487, "step": 5210 }, { "epoch": 3.038416763678696, "grad_norm": 1.4323164224624634, "learning_rate": 9.962652353225438e-05, "loss": 0.5423, "step": 5220 }, { "epoch": 3.0442374854481957, "grad_norm": 1.4053871631622314, "learning_rate": 9.962315402254619e-05, "loss": 0.5394, "step": 5230 }, { "epoch": 3.050058207217695, "grad_norm": 1.3439847230911255, "learning_rate": 9.9619769438634e-05, "loss": 0.5382, "step": 5240 }, { "epoch": 3.0558789289871946, "grad_norm": 1.423401117324829, "learning_rate": 9.9616369781546e-05, "loss": 0.5218, "step": 5250 }, { "epoch": 3.061699650756694, "grad_norm": 1.3146246671676636, "learning_rate": 9.961295505231491e-05, "loss": 0.5471, "step": 5260 }, { "epoch": 3.067520372526193, "grad_norm": 1.4071041345596313, "learning_rate": 9.960952525197804e-05, "loss": 0.546, "step": 5270 }, { "epoch": 3.0733410942956927, "grad_norm": 1.3692514896392822, "learning_rate": 9.960608038157724e-05, "loss": 0.5488, "step": 5280 }, { "epoch": 3.079161816065192, "grad_norm": 1.2678933143615723, "learning_rate": 9.960262044215901e-05, "loss": 0.5554, "step": 5290 }, { "epoch": 3.0849825378346916, "grad_norm": 1.3721624612808228, "learning_rate": 9.959914543477435e-05, "loss": 0.5475, "step": 5300 }, { "epoch": 3.090803259604191, "grad_norm": 1.3230259418487549, "learning_rate": 9.959565536047892e-05, "loss": 0.5416, "step": 5310 }, { "epoch": 3.0966239813736904, "grad_norm": 1.4343992471694946, "learning_rate": 9.959215022033288e-05, "loss": 0.5435, "step": 5320 }, { "epoch": 3.1024447031431897, "grad_norm": 1.2494285106658936, "learning_rate": 9.9588630015401e-05, "loss": 0.5218, "step": 5330 }, { "epoch": 3.1082654249126893, "grad_norm": 1.2102912664413452, "learning_rate": 9.958509474675264e-05, "loss": 0.5417, "step": 5340 }, { "epoch": 3.1140861466821885, "grad_norm": 1.3514965772628784, "learning_rate": 9.958154441546171e-05, "loss": 0.5288, "step": 5350 }, { "epoch": 3.119906868451688, "grad_norm": 1.2747243642807007, "learning_rate": 9.957797902260673e-05, "loss": 0.541, "step": 5360 }, { "epoch": 3.1257275902211874, "grad_norm": 1.3566619157791138, "learning_rate": 9.957439856927073e-05, "loss": 0.5453, "step": 5370 }, { "epoch": 3.131548311990687, "grad_norm": 1.3164036273956299, "learning_rate": 9.957080305654139e-05, "loss": 0.5236, "step": 5380 }, { "epoch": 3.1373690337601863, "grad_norm": 1.2035046815872192, "learning_rate": 9.956719248551092e-05, "loss": 0.5295, "step": 5390 }, { "epoch": 3.1431897555296855, "grad_norm": 1.2006549835205078, "learning_rate": 9.956356685727612e-05, "loss": 0.5348, "step": 5400 }, { "epoch": 3.149010477299185, "grad_norm": 1.354110836982727, "learning_rate": 9.955992617293836e-05, "loss": 0.5325, "step": 5410 }, { "epoch": 3.1548311990686844, "grad_norm": 1.2383904457092285, "learning_rate": 9.955627043360358e-05, "loss": 0.517, "step": 5420 }, { "epoch": 3.160651920838184, "grad_norm": 1.4119353294372559, "learning_rate": 9.955259964038231e-05, "loss": 0.5349, "step": 5430 }, { "epoch": 3.1664726426076832, "grad_norm": 1.4360922574996948, "learning_rate": 9.954891379438962e-05, "loss": 0.5346, "step": 5440 }, { "epoch": 3.172293364377183, "grad_norm": 1.5917502641677856, "learning_rate": 9.954521289674519e-05, "loss": 0.547, "step": 5450 }, { "epoch": 3.178114086146682, "grad_norm": 1.4314442873001099, "learning_rate": 9.954149694857325e-05, "loss": 0.5234, "step": 5460 }, { "epoch": 3.1839348079161818, "grad_norm": 1.3017076253890991, "learning_rate": 9.953776595100258e-05, "loss": 0.5254, "step": 5470 }, { "epoch": 3.189755529685681, "grad_norm": 1.371936559677124, "learning_rate": 9.95340199051666e-05, "loss": 0.5253, "step": 5480 }, { "epoch": 3.1955762514551806, "grad_norm": 1.3678311109542847, "learning_rate": 9.953025881220325e-05, "loss": 0.5147, "step": 5490 }, { "epoch": 3.20139697322468, "grad_norm": 1.3658735752105713, "learning_rate": 9.952648267325504e-05, "loss": 0.5238, "step": 5500 }, { "epoch": 3.207217694994179, "grad_norm": 1.3038593530654907, "learning_rate": 9.952269148946905e-05, "loss": 0.5311, "step": 5510 }, { "epoch": 3.2130384167636787, "grad_norm": 1.3068981170654297, "learning_rate": 9.951888526199697e-05, "loss": 0.5199, "step": 5520 }, { "epoch": 3.218859138533178, "grad_norm": 1.3260855674743652, "learning_rate": 9.951506399199501e-05, "loss": 0.5313, "step": 5530 }, { "epoch": 3.2246798603026776, "grad_norm": 1.282050609588623, "learning_rate": 9.951122768062399e-05, "loss": 0.5237, "step": 5540 }, { "epoch": 3.230500582072177, "grad_norm": 1.273202896118164, "learning_rate": 9.950737632904927e-05, "loss": 0.5162, "step": 5550 }, { "epoch": 3.2363213038416765, "grad_norm": 1.4833550453186035, "learning_rate": 9.950350993844077e-05, "loss": 0.535, "step": 5560 }, { "epoch": 3.2421420256111757, "grad_norm": 1.5928966999053955, "learning_rate": 9.949962850997303e-05, "loss": 0.513, "step": 5570 }, { "epoch": 3.2479627473806754, "grad_norm": 1.219640851020813, "learning_rate": 9.949573204482512e-05, "loss": 0.5269, "step": 5580 }, { "epoch": 3.2537834691501746, "grad_norm": 1.2813693284988403, "learning_rate": 9.949182054418064e-05, "loss": 0.5231, "step": 5590 }, { "epoch": 3.2596041909196742, "grad_norm": 1.2923225164413452, "learning_rate": 9.948789400922787e-05, "loss": 0.5241, "step": 5600 }, { "epoch": 3.2654249126891735, "grad_norm": 1.3776395320892334, "learning_rate": 9.948395244115953e-05, "loss": 0.5342, "step": 5610 }, { "epoch": 3.2712456344586727, "grad_norm": 1.4589473009109497, "learning_rate": 9.9479995841173e-05, "loss": 0.5347, "step": 5620 }, { "epoch": 3.2770663562281723, "grad_norm": 1.1041414737701416, "learning_rate": 9.947602421047017e-05, "loss": 0.5133, "step": 5630 }, { "epoch": 3.2828870779976715, "grad_norm": 1.2307438850402832, "learning_rate": 9.947203755025753e-05, "loss": 0.5331, "step": 5640 }, { "epoch": 3.288707799767171, "grad_norm": 1.3341834545135498, "learning_rate": 9.946803586174611e-05, "loss": 0.5111, "step": 5650 }, { "epoch": 3.2945285215366704, "grad_norm": 1.2455167770385742, "learning_rate": 9.946401914615151e-05, "loss": 0.5277, "step": 5660 }, { "epoch": 3.30034924330617, "grad_norm": 1.4423123598098755, "learning_rate": 9.945998740469394e-05, "loss": 0.5328, "step": 5670 }, { "epoch": 3.3061699650756693, "grad_norm": 1.4434070587158203, "learning_rate": 9.945594063859809e-05, "loss": 0.5258, "step": 5680 }, { "epoch": 3.311990686845169, "grad_norm": 1.1902482509613037, "learning_rate": 9.94518788490933e-05, "loss": 0.5307, "step": 5690 }, { "epoch": 3.317811408614668, "grad_norm": 1.1999151706695557, "learning_rate": 9.944780203741341e-05, "loss": 0.5031, "step": 5700 }, { "epoch": 3.323632130384168, "grad_norm": 1.4925495386123657, "learning_rate": 9.944371020479686e-05, "loss": 0.5298, "step": 5710 }, { "epoch": 3.329452852153667, "grad_norm": 1.1832563877105713, "learning_rate": 9.943960335248662e-05, "loss": 0.5275, "step": 5720 }, { "epoch": 3.3352735739231667, "grad_norm": 1.4086031913757324, "learning_rate": 9.943548148173027e-05, "loss": 0.532, "step": 5730 }, { "epoch": 3.341094295692666, "grad_norm": 1.3896619081497192, "learning_rate": 9.943134459377992e-05, "loss": 0.5188, "step": 5740 }, { "epoch": 3.346915017462165, "grad_norm": 1.1718300580978394, "learning_rate": 9.942719268989222e-05, "loss": 0.5204, "step": 5750 }, { "epoch": 3.352735739231665, "grad_norm": 1.334605097770691, "learning_rate": 9.942302577132844e-05, "loss": 0.5265, "step": 5760 }, { "epoch": 3.358556461001164, "grad_norm": 1.240610957145691, "learning_rate": 9.941884383935438e-05, "loss": 0.5211, "step": 5770 }, { "epoch": 3.3643771827706637, "grad_norm": 1.3100090026855469, "learning_rate": 9.941464689524039e-05, "loss": 0.5274, "step": 5780 }, { "epoch": 3.370197904540163, "grad_norm": 1.2016217708587646, "learning_rate": 9.941043494026139e-05, "loss": 0.5034, "step": 5790 }, { "epoch": 3.3760186263096625, "grad_norm": 1.374687671661377, "learning_rate": 9.940620797569685e-05, "loss": 0.516, "step": 5800 }, { "epoch": 3.3818393480791618, "grad_norm": 1.2523596286773682, "learning_rate": 9.940196600283082e-05, "loss": 0.5123, "step": 5810 }, { "epoch": 3.3876600698486614, "grad_norm": 1.4260362386703491, "learning_rate": 9.939770902295192e-05, "loss": 0.5088, "step": 5820 }, { "epoch": 3.3934807916181606, "grad_norm": 1.5519804954528809, "learning_rate": 9.939343703735329e-05, "loss": 0.5294, "step": 5830 }, { "epoch": 3.39930151338766, "grad_norm": 1.3274418115615845, "learning_rate": 9.938915004733264e-05, "loss": 0.5193, "step": 5840 }, { "epoch": 3.4051222351571595, "grad_norm": 1.2610960006713867, "learning_rate": 9.938484805419224e-05, "loss": 0.535, "step": 5850 }, { "epoch": 3.4109429569266587, "grad_norm": 1.4841417074203491, "learning_rate": 9.938053105923894e-05, "loss": 0.5023, "step": 5860 }, { "epoch": 3.4167636786961584, "grad_norm": 1.3334405422210693, "learning_rate": 9.937619906378413e-05, "loss": 0.5079, "step": 5870 }, { "epoch": 3.4225844004656576, "grad_norm": 1.4320570230484009, "learning_rate": 9.937185206914374e-05, "loss": 0.5522, "step": 5880 }, { "epoch": 3.4284051222351573, "grad_norm": 1.3657760620117188, "learning_rate": 9.936749007663829e-05, "loss": 0.5154, "step": 5890 }, { "epoch": 3.4342258440046565, "grad_norm": 1.4238898754119873, "learning_rate": 9.93631130875928e-05, "loss": 0.5142, "step": 5900 }, { "epoch": 3.440046565774156, "grad_norm": 1.3776757717132568, "learning_rate": 9.935872110333692e-05, "loss": 0.519, "step": 5910 }, { "epoch": 3.4458672875436553, "grad_norm": 1.3840113878250122, "learning_rate": 9.935431412520484e-05, "loss": 0.5214, "step": 5920 }, { "epoch": 3.451688009313155, "grad_norm": 1.3701118230819702, "learning_rate": 9.934989215453523e-05, "loss": 0.5135, "step": 5930 }, { "epoch": 3.457508731082654, "grad_norm": 1.2288986444473267, "learning_rate": 9.934545519267139e-05, "loss": 0.5041, "step": 5940 }, { "epoch": 3.463329452852154, "grad_norm": 1.175449013710022, "learning_rate": 9.934100324096117e-05, "loss": 0.5174, "step": 5950 }, { "epoch": 3.469150174621653, "grad_norm": 1.491661787033081, "learning_rate": 9.933653630075692e-05, "loss": 0.51, "step": 5960 }, { "epoch": 3.4749708963911523, "grad_norm": 1.36133873462677, "learning_rate": 9.93320543734156e-05, "loss": 0.4919, "step": 5970 }, { "epoch": 3.480791618160652, "grad_norm": 1.5339820384979248, "learning_rate": 9.932755746029871e-05, "loss": 0.5083, "step": 5980 }, { "epoch": 3.486612339930151, "grad_norm": 1.297743558883667, "learning_rate": 9.932304556277228e-05, "loss": 0.5048, "step": 5990 }, { "epoch": 3.492433061699651, "grad_norm": 1.314329981803894, "learning_rate": 9.93185186822069e-05, "loss": 0.5081, "step": 6000 }, { "epoch": 3.49825378346915, "grad_norm": 1.2274463176727295, "learning_rate": 9.931397681997773e-05, "loss": 0.4945, "step": 6010 }, { "epoch": 3.5040745052386497, "grad_norm": 1.4737827777862549, "learning_rate": 9.930941997746446e-05, "loss": 0.5169, "step": 6020 }, { "epoch": 3.509895227008149, "grad_norm": 1.3423832654953003, "learning_rate": 9.930484815605134e-05, "loss": 0.5044, "step": 6030 }, { "epoch": 3.5157159487776486, "grad_norm": 1.1909335851669312, "learning_rate": 9.930026135712717e-05, "loss": 0.5021, "step": 6040 }, { "epoch": 3.521536670547148, "grad_norm": 1.3570606708526611, "learning_rate": 9.92956595820853e-05, "loss": 0.4999, "step": 6050 }, { "epoch": 3.527357392316647, "grad_norm": 1.2635002136230469, "learning_rate": 9.929104283232362e-05, "loss": 0.5011, "step": 6060 }, { "epoch": 3.5331781140861467, "grad_norm": 1.4552651643753052, "learning_rate": 9.92864111092446e-05, "loss": 0.5056, "step": 6070 }, { "epoch": 3.5389988358556463, "grad_norm": 1.3079516887664795, "learning_rate": 9.92817644142552e-05, "loss": 0.5062, "step": 6080 }, { "epoch": 3.5448195576251456, "grad_norm": 1.1985629796981812, "learning_rate": 9.927710274876698e-05, "loss": 0.4997, "step": 6090 }, { "epoch": 3.5506402793946448, "grad_norm": 1.3855479955673218, "learning_rate": 9.927242611419603e-05, "loss": 0.4983, "step": 6100 }, { "epoch": 3.5564610011641444, "grad_norm": 1.1333374977111816, "learning_rate": 9.926773451196301e-05, "loss": 0.502, "step": 6110 }, { "epoch": 3.5622817229336436, "grad_norm": 1.533910870552063, "learning_rate": 9.926302794349306e-05, "loss": 0.5114, "step": 6120 }, { "epoch": 3.5681024447031433, "grad_norm": 1.4010874032974243, "learning_rate": 9.925830641021594e-05, "loss": 0.5248, "step": 6130 }, { "epoch": 3.5739231664726425, "grad_norm": 1.2345761060714722, "learning_rate": 9.925356991356593e-05, "loss": 0.4933, "step": 6140 }, { "epoch": 3.579743888242142, "grad_norm": 1.2987070083618164, "learning_rate": 9.924881845498184e-05, "loss": 0.4903, "step": 6150 }, { "epoch": 3.5855646100116414, "grad_norm": 1.2276204824447632, "learning_rate": 9.924405203590705e-05, "loss": 0.5111, "step": 6160 }, { "epoch": 3.591385331781141, "grad_norm": 1.323367714881897, "learning_rate": 9.923927065778946e-05, "loss": 0.5079, "step": 6170 }, { "epoch": 3.5972060535506403, "grad_norm": 1.3997156620025635, "learning_rate": 9.923447432208154e-05, "loss": 0.5197, "step": 6180 }, { "epoch": 3.6030267753201395, "grad_norm": 1.3657439947128296, "learning_rate": 9.922966303024027e-05, "loss": 0.5144, "step": 6190 }, { "epoch": 3.608847497089639, "grad_norm": 1.333682894706726, "learning_rate": 9.922483678372721e-05, "loss": 0.5084, "step": 6200 }, { "epoch": 3.614668218859139, "grad_norm": 1.3236114978790283, "learning_rate": 9.921999558400845e-05, "loss": 0.5022, "step": 6210 }, { "epoch": 3.620488940628638, "grad_norm": 1.3749001026153564, "learning_rate": 9.92151394325546e-05, "loss": 0.4901, "step": 6220 }, { "epoch": 3.6263096623981372, "grad_norm": 1.3203728199005127, "learning_rate": 9.921026833084084e-05, "loss": 0.511, "step": 6230 }, { "epoch": 3.632130384167637, "grad_norm": 1.2848238945007324, "learning_rate": 9.920538228034689e-05, "loss": 0.4991, "step": 6240 }, { "epoch": 3.637951105937136, "grad_norm": 1.267100214958191, "learning_rate": 9.920048128255699e-05, "loss": 0.5005, "step": 6250 }, { "epoch": 3.6437718277066358, "grad_norm": 1.4159588813781738, "learning_rate": 9.919556533895995e-05, "loss": 0.4935, "step": 6260 }, { "epoch": 3.649592549476135, "grad_norm": 1.2929341793060303, "learning_rate": 9.919063445104907e-05, "loss": 0.5087, "step": 6270 }, { "epoch": 3.655413271245634, "grad_norm": 1.1815531253814697, "learning_rate": 9.918568862032227e-05, "loss": 0.5085, "step": 6280 }, { "epoch": 3.661233993015134, "grad_norm": 1.2817507982254028, "learning_rate": 9.918072784828194e-05, "loss": 0.5053, "step": 6290 }, { "epoch": 3.6670547147846335, "grad_norm": 1.2423557043075562, "learning_rate": 9.917575213643501e-05, "loss": 0.4772, "step": 6300 }, { "epoch": 3.6728754365541327, "grad_norm": 1.4056448936462402, "learning_rate": 9.917076148629302e-05, "loss": 0.498, "step": 6310 }, { "epoch": 3.678696158323632, "grad_norm": 1.288513422012329, "learning_rate": 9.916575589937196e-05, "loss": 0.4848, "step": 6320 }, { "epoch": 3.6845168800931316, "grad_norm": 1.3175112009048462, "learning_rate": 9.916073537719239e-05, "loss": 0.498, "step": 6330 }, { "epoch": 3.690337601862631, "grad_norm": 1.4057283401489258, "learning_rate": 9.915569992127944e-05, "loss": 0.5021, "step": 6340 }, { "epoch": 3.6961583236321305, "grad_norm": 1.3772826194763184, "learning_rate": 9.915064953316273e-05, "loss": 0.4875, "step": 6350 }, { "epoch": 3.7019790454016297, "grad_norm": 1.2440996170043945, "learning_rate": 9.914558421437645e-05, "loss": 0.5031, "step": 6360 }, { "epoch": 3.7077997671711294, "grad_norm": 1.2028613090515137, "learning_rate": 9.914050396645929e-05, "loss": 0.4924, "step": 6370 }, { "epoch": 3.7136204889406286, "grad_norm": 1.4111618995666504, "learning_rate": 9.913540879095452e-05, "loss": 0.499, "step": 6380 }, { "epoch": 3.7194412107101282, "grad_norm": 1.1926265954971313, "learning_rate": 9.913029868940987e-05, "loss": 0.5134, "step": 6390 }, { "epoch": 3.7252619324796274, "grad_norm": 1.32298743724823, "learning_rate": 9.912517366337772e-05, "loss": 0.4935, "step": 6400 }, { "epoch": 3.7310826542491267, "grad_norm": 1.384425401687622, "learning_rate": 9.912003371441487e-05, "loss": 0.4998, "step": 6410 }, { "epoch": 3.7369033760186263, "grad_norm": 1.4014195203781128, "learning_rate": 9.911487884408271e-05, "loss": 0.5009, "step": 6420 }, { "epoch": 3.742724097788126, "grad_norm": 1.158780813217163, "learning_rate": 9.910970905394719e-05, "loss": 0.4867, "step": 6430 }, { "epoch": 3.748544819557625, "grad_norm": 1.2409323453903198, "learning_rate": 9.91045243455787e-05, "loss": 0.496, "step": 6440 }, { "epoch": 3.7543655413271244, "grad_norm": 1.3364369869232178, "learning_rate": 9.909932472055225e-05, "loss": 0.4867, "step": 6450 }, { "epoch": 3.760186263096624, "grad_norm": 1.3671120405197144, "learning_rate": 9.909411018044734e-05, "loss": 0.4895, "step": 6460 }, { "epoch": 3.7660069848661233, "grad_norm": 1.284104347229004, "learning_rate": 9.908888072684802e-05, "loss": 0.5079, "step": 6470 }, { "epoch": 3.771827706635623, "grad_norm": 1.3752398490905762, "learning_rate": 9.908363636134285e-05, "loss": 0.4865, "step": 6480 }, { "epoch": 3.777648428405122, "grad_norm": 1.1722854375839233, "learning_rate": 9.907837708552493e-05, "loss": 0.4885, "step": 6490 }, { "epoch": 3.7834691501746214, "grad_norm": 1.5468719005584717, "learning_rate": 9.90731029009919e-05, "loss": 0.4989, "step": 6500 }, { "epoch": 3.789289871944121, "grad_norm": 1.3475205898284912, "learning_rate": 9.906781380934589e-05, "loss": 0.4868, "step": 6510 }, { "epoch": 3.7951105937136207, "grad_norm": 1.3782157897949219, "learning_rate": 9.906250981219362e-05, "loss": 0.5035, "step": 6520 }, { "epoch": 3.80093131548312, "grad_norm": 1.262954831123352, "learning_rate": 9.905719091114628e-05, "loss": 0.5032, "step": 6530 }, { "epoch": 3.806752037252619, "grad_norm": 1.3369026184082031, "learning_rate": 9.905185710781964e-05, "loss": 0.4885, "step": 6540 }, { "epoch": 3.812572759022119, "grad_norm": 1.3380428552627563, "learning_rate": 9.904650840383392e-05, "loss": 0.4985, "step": 6550 }, { "epoch": 3.818393480791618, "grad_norm": 1.2748851776123047, "learning_rate": 9.904114480081397e-05, "loss": 0.5029, "step": 6560 }, { "epoch": 3.8242142025611177, "grad_norm": 1.2657837867736816, "learning_rate": 9.903576630038906e-05, "loss": 0.4984, "step": 6570 }, { "epoch": 3.830034924330617, "grad_norm": 1.3467230796813965, "learning_rate": 9.903037290419309e-05, "loss": 0.5329, "step": 6580 }, { "epoch": 3.8358556461001165, "grad_norm": 1.5105469226837158, "learning_rate": 9.902496461386439e-05, "loss": 0.4905, "step": 6590 }, { "epoch": 3.8416763678696157, "grad_norm": 1.3876874446868896, "learning_rate": 9.901954143104588e-05, "loss": 0.4986, "step": 6600 }, { "epoch": 3.8474970896391154, "grad_norm": 1.393911361694336, "learning_rate": 9.901410335738496e-05, "loss": 0.5008, "step": 6610 }, { "epoch": 3.8533178114086146, "grad_norm": 1.4654799699783325, "learning_rate": 9.900865039453358e-05, "loss": 0.4869, "step": 6620 }, { "epoch": 3.859138533178114, "grad_norm": 1.2269515991210938, "learning_rate": 9.900318254414821e-05, "loss": 0.5007, "step": 6630 }, { "epoch": 3.8649592549476135, "grad_norm": 1.472175121307373, "learning_rate": 9.899769980788985e-05, "loss": 0.5056, "step": 6640 }, { "epoch": 3.870779976717113, "grad_norm": 1.305052399635315, "learning_rate": 9.899220218742398e-05, "loss": 0.4834, "step": 6650 }, { "epoch": 3.8766006984866124, "grad_norm": 1.2856225967407227, "learning_rate": 9.898668968442066e-05, "loss": 0.4977, "step": 6660 }, { "epoch": 3.8824214202561116, "grad_norm": 1.2376476526260376, "learning_rate": 9.898116230055443e-05, "loss": 0.4856, "step": 6670 }, { "epoch": 3.8882421420256112, "grad_norm": 1.2468189001083374, "learning_rate": 9.897562003750437e-05, "loss": 0.4899, "step": 6680 }, { "epoch": 3.8940628637951105, "grad_norm": 1.4607261419296265, "learning_rate": 9.897006289695407e-05, "loss": 0.4972, "step": 6690 }, { "epoch": 3.89988358556461, "grad_norm": 1.447243332862854, "learning_rate": 9.896449088059164e-05, "loss": 0.4879, "step": 6700 }, { "epoch": 3.9057043073341093, "grad_norm": 1.256103277206421, "learning_rate": 9.89589039901097e-05, "loss": 0.4903, "step": 6710 }, { "epoch": 3.911525029103609, "grad_norm": 1.4003379344940186, "learning_rate": 9.895330222720542e-05, "loss": 0.4818, "step": 6720 }, { "epoch": 3.917345750873108, "grad_norm": 1.241776943206787, "learning_rate": 9.894768559358047e-05, "loss": 0.4937, "step": 6730 }, { "epoch": 3.923166472642608, "grad_norm": 1.3098689317703247, "learning_rate": 9.894205409094101e-05, "loss": 0.5042, "step": 6740 }, { "epoch": 3.928987194412107, "grad_norm": 1.3114290237426758, "learning_rate": 9.893640772099777e-05, "loss": 0.4782, "step": 6750 }, { "epoch": 3.9348079161816063, "grad_norm": 1.2668832540512085, "learning_rate": 9.893074648546595e-05, "loss": 0.488, "step": 6760 }, { "epoch": 3.940628637951106, "grad_norm": 1.319075107574463, "learning_rate": 9.892507038606528e-05, "loss": 0.4649, "step": 6770 }, { "epoch": 3.9464493597206056, "grad_norm": 1.4619354009628296, "learning_rate": 9.891937942452003e-05, "loss": 0.485, "step": 6780 }, { "epoch": 3.952270081490105, "grad_norm": 1.3690756559371948, "learning_rate": 9.891367360255895e-05, "loss": 0.4986, "step": 6790 }, { "epoch": 3.958090803259604, "grad_norm": 1.3349260091781616, "learning_rate": 9.890795292191532e-05, "loss": 0.5108, "step": 6800 }, { "epoch": 3.9639115250291037, "grad_norm": 1.2126715183258057, "learning_rate": 9.890221738432694e-05, "loss": 0.4768, "step": 6810 }, { "epoch": 3.969732246798603, "grad_norm": 1.455809235572815, "learning_rate": 9.88964669915361e-05, "loss": 0.4721, "step": 6820 }, { "epoch": 3.9755529685681026, "grad_norm": 1.4671248197555542, "learning_rate": 9.889070174528963e-05, "loss": 0.4768, "step": 6830 }, { "epoch": 3.981373690337602, "grad_norm": 1.3239575624465942, "learning_rate": 9.888492164733883e-05, "loss": 0.4838, "step": 6840 }, { "epoch": 3.987194412107101, "grad_norm": 1.2590062618255615, "learning_rate": 9.88791266994396e-05, "loss": 0.4766, "step": 6850 }, { "epoch": 3.9930151338766007, "grad_norm": 1.1893808841705322, "learning_rate": 9.887331690335223e-05, "loss": 0.4906, "step": 6860 }, { "epoch": 3.9988358556461003, "grad_norm": 1.1650422811508179, "learning_rate": 9.886749226084163e-05, "loss": 0.4951, "step": 6870 }, { "epoch": 4.0046565774155995, "grad_norm": 1.3535321950912476, "learning_rate": 9.886165277367714e-05, "loss": 0.4974, "step": 6880 }, { "epoch": 4.010477299185099, "grad_norm": 1.203468680381775, "learning_rate": 9.885579844363265e-05, "loss": 0.4899, "step": 6890 }, { "epoch": 4.016298020954598, "grad_norm": 1.2788859605789185, "learning_rate": 9.884992927248656e-05, "loss": 0.4784, "step": 6900 }, { "epoch": 4.022118742724098, "grad_norm": 1.425062894821167, "learning_rate": 9.884404526202178e-05, "loss": 0.4946, "step": 6910 }, { "epoch": 4.027939464493597, "grad_norm": 1.4433338642120361, "learning_rate": 9.883814641402568e-05, "loss": 0.4987, "step": 6920 }, { "epoch": 4.0337601862630965, "grad_norm": 1.1833258867263794, "learning_rate": 9.88322327302902e-05, "loss": 0.4839, "step": 6930 }, { "epoch": 4.039580908032596, "grad_norm": 1.252354621887207, "learning_rate": 9.882630421261176e-05, "loss": 0.471, "step": 6940 }, { "epoch": 4.045401629802096, "grad_norm": 1.3564316034317017, "learning_rate": 9.88203608627913e-05, "loss": 0.492, "step": 6950 }, { "epoch": 4.051222351571595, "grad_norm": 1.351953387260437, "learning_rate": 9.881440268263422e-05, "loss": 0.4994, "step": 6960 }, { "epoch": 4.057043073341094, "grad_norm": 1.1982601881027222, "learning_rate": 9.880842967395048e-05, "loss": 0.4738, "step": 6970 }, { "epoch": 4.0628637951105935, "grad_norm": 1.1731337308883667, "learning_rate": 9.880244183855452e-05, "loss": 0.4791, "step": 6980 }, { "epoch": 4.068684516880094, "grad_norm": 1.2485936880111694, "learning_rate": 9.879643917826527e-05, "loss": 0.4866, "step": 6990 }, { "epoch": 4.074505238649593, "grad_norm": 1.3642054796218872, "learning_rate": 9.87904216949062e-05, "loss": 0.4799, "step": 7000 }, { "epoch": 4.080325960419092, "grad_norm": 1.2799961566925049, "learning_rate": 9.878438939030526e-05, "loss": 0.4887, "step": 7010 }, { "epoch": 4.086146682188591, "grad_norm": 1.320765733718872, "learning_rate": 9.877834226629489e-05, "loss": 0.4668, "step": 7020 }, { "epoch": 4.09196740395809, "grad_norm": 1.420846939086914, "learning_rate": 9.877228032471206e-05, "loss": 0.4886, "step": 7030 }, { "epoch": 4.0977881257275905, "grad_norm": 1.3025811910629272, "learning_rate": 9.876620356739823e-05, "loss": 0.4967, "step": 7040 }, { "epoch": 4.10360884749709, "grad_norm": 1.4175794124603271, "learning_rate": 9.876011199619935e-05, "loss": 0.5057, "step": 7050 }, { "epoch": 4.109429569266589, "grad_norm": 1.279435396194458, "learning_rate": 9.875400561296589e-05, "loss": 0.4784, "step": 7060 }, { "epoch": 4.115250291036088, "grad_norm": 1.2974194288253784, "learning_rate": 9.874788441955278e-05, "loss": 0.4726, "step": 7070 }, { "epoch": 4.121071012805588, "grad_norm": 1.1871509552001953, "learning_rate": 9.874174841781951e-05, "loss": 0.4614, "step": 7080 }, { "epoch": 4.1268917345750875, "grad_norm": 1.2223505973815918, "learning_rate": 9.873559760963003e-05, "loss": 0.4807, "step": 7090 }, { "epoch": 4.132712456344587, "grad_norm": 1.2793678045272827, "learning_rate": 9.872943199685278e-05, "loss": 0.4733, "step": 7100 }, { "epoch": 4.138533178114086, "grad_norm": 1.3223806619644165, "learning_rate": 9.872325158136071e-05, "loss": 0.4837, "step": 7110 }, { "epoch": 4.144353899883585, "grad_norm": 1.2091032266616821, "learning_rate": 9.871705636503128e-05, "loss": 0.4972, "step": 7120 }, { "epoch": 4.150174621653085, "grad_norm": 1.2343100309371948, "learning_rate": 9.871084634974641e-05, "loss": 0.4726, "step": 7130 }, { "epoch": 4.1559953434225845, "grad_norm": 1.3967267274856567, "learning_rate": 9.870462153739257e-05, "loss": 0.4932, "step": 7140 }, { "epoch": 4.161816065192084, "grad_norm": 1.2172008752822876, "learning_rate": 9.869838192986067e-05, "loss": 0.4723, "step": 7150 }, { "epoch": 4.167636786961583, "grad_norm": 1.1934665441513062, "learning_rate": 9.869212752904616e-05, "loss": 0.4829, "step": 7160 }, { "epoch": 4.173457508731083, "grad_norm": 1.1700025796890259, "learning_rate": 9.868585833684894e-05, "loss": 0.4845, "step": 7170 }, { "epoch": 4.179278230500582, "grad_norm": 1.2237801551818848, "learning_rate": 9.867957435517342e-05, "loss": 0.4758, "step": 7180 }, { "epoch": 4.185098952270081, "grad_norm": 1.2563188076019287, "learning_rate": 9.867327558592854e-05, "loss": 0.4898, "step": 7190 }, { "epoch": 4.190919674039581, "grad_norm": 1.3118526935577393, "learning_rate": 9.866696203102766e-05, "loss": 0.476, "step": 7200 }, { "epoch": 4.19674039580908, "grad_norm": 1.2278954982757568, "learning_rate": 9.86606336923887e-05, "loss": 0.4624, "step": 7210 }, { "epoch": 4.20256111757858, "grad_norm": 1.2999484539031982, "learning_rate": 9.865429057193403e-05, "loss": 0.4903, "step": 7220 }, { "epoch": 4.208381839348079, "grad_norm": 1.2272859811782837, "learning_rate": 9.864793267159053e-05, "loss": 0.4758, "step": 7230 }, { "epoch": 4.214202561117578, "grad_norm": 1.0988699197769165, "learning_rate": 9.864155999328957e-05, "loss": 0.4741, "step": 7240 }, { "epoch": 4.220023282887078, "grad_norm": 1.5274351835250854, "learning_rate": 9.8635172538967e-05, "loss": 0.4681, "step": 7250 }, { "epoch": 4.225844004656578, "grad_norm": 1.1603261232376099, "learning_rate": 9.862877031056312e-05, "loss": 0.4842, "step": 7260 }, { "epoch": 4.231664726426077, "grad_norm": 1.3009566068649292, "learning_rate": 9.862235331002279e-05, "loss": 0.4764, "step": 7270 }, { "epoch": 4.237485448195576, "grad_norm": 1.268801212310791, "learning_rate": 9.861592153929533e-05, "loss": 0.469, "step": 7280 }, { "epoch": 4.243306169965075, "grad_norm": 1.3035651445388794, "learning_rate": 9.860947500033455e-05, "loss": 0.4751, "step": 7290 }, { "epoch": 4.2491268917345755, "grad_norm": 1.19972825050354, "learning_rate": 9.86030136950987e-05, "loss": 0.4875, "step": 7300 }, { "epoch": 4.254947613504075, "grad_norm": 1.2076289653778076, "learning_rate": 9.85965376255506e-05, "loss": 0.4623, "step": 7310 }, { "epoch": 4.260768335273574, "grad_norm": 1.175568699836731, "learning_rate": 9.859004679365747e-05, "loss": 0.4779, "step": 7320 }, { "epoch": 4.266589057043073, "grad_norm": 1.3565183877944946, "learning_rate": 9.858354120139108e-05, "loss": 0.4841, "step": 7330 }, { "epoch": 4.272409778812573, "grad_norm": 1.3250194787979126, "learning_rate": 9.857702085072764e-05, "loss": 0.4801, "step": 7340 }, { "epoch": 4.278230500582072, "grad_norm": 1.250728964805603, "learning_rate": 9.857048574364787e-05, "loss": 0.4694, "step": 7350 }, { "epoch": 4.284051222351572, "grad_norm": 1.488011360168457, "learning_rate": 9.856393588213698e-05, "loss": 0.4617, "step": 7360 }, { "epoch": 4.289871944121071, "grad_norm": 1.2864990234375, "learning_rate": 9.855737126818458e-05, "loss": 0.4639, "step": 7370 }, { "epoch": 4.29569266589057, "grad_norm": 1.2784711122512817, "learning_rate": 9.855079190378491e-05, "loss": 0.4781, "step": 7380 }, { "epoch": 4.30151338766007, "grad_norm": 1.1352078914642334, "learning_rate": 9.854419779093655e-05, "loss": 0.4757, "step": 7390 }, { "epoch": 4.307334109429569, "grad_norm": 1.2553211450576782, "learning_rate": 9.853758893164264e-05, "loss": 0.472, "step": 7400 }, { "epoch": 4.313154831199069, "grad_norm": 1.3039402961730957, "learning_rate": 9.853096532791078e-05, "loss": 0.4637, "step": 7410 }, { "epoch": 4.318975552968568, "grad_norm": 1.3409806489944458, "learning_rate": 9.852432698175304e-05, "loss": 0.4856, "step": 7420 }, { "epoch": 4.324796274738068, "grad_norm": 1.1921018362045288, "learning_rate": 9.851767389518597e-05, "loss": 0.4721, "step": 7430 }, { "epoch": 4.330616996507567, "grad_norm": 1.138107419013977, "learning_rate": 9.85110060702306e-05, "loss": 0.4614, "step": 7440 }, { "epoch": 4.336437718277066, "grad_norm": 1.3903526067733765, "learning_rate": 9.850432350891245e-05, "loss": 0.4623, "step": 7450 }, { "epoch": 4.342258440046566, "grad_norm": 1.2693548202514648, "learning_rate": 9.84976262132615e-05, "loss": 0.474, "step": 7460 }, { "epoch": 4.348079161816065, "grad_norm": 1.4243619441986084, "learning_rate": 9.849091418531222e-05, "loss": 0.4795, "step": 7470 }, { "epoch": 4.353899883585565, "grad_norm": 1.3414250612258911, "learning_rate": 9.848418742710353e-05, "loss": 0.473, "step": 7480 }, { "epoch": 4.359720605355064, "grad_norm": 1.2627335786819458, "learning_rate": 9.847744594067885e-05, "loss": 0.4755, "step": 7490 }, { "epoch": 4.365541327124563, "grad_norm": 1.2993561029434204, "learning_rate": 9.847068972808607e-05, "loss": 0.475, "step": 7500 }, { "epoch": 4.3713620488940625, "grad_norm": 1.5441068410873413, "learning_rate": 9.846391879137756e-05, "loss": 0.4518, "step": 7510 }, { "epoch": 4.377182770663563, "grad_norm": 1.3856806755065918, "learning_rate": 9.845713313261012e-05, "loss": 0.4868, "step": 7520 }, { "epoch": 4.383003492433062, "grad_norm": 1.2205406427383423, "learning_rate": 9.845033275384505e-05, "loss": 0.4899, "step": 7530 }, { "epoch": 4.388824214202561, "grad_norm": 1.2913535833358765, "learning_rate": 9.844351765714818e-05, "loss": 0.467, "step": 7540 }, { "epoch": 4.39464493597206, "grad_norm": 1.2896922826766968, "learning_rate": 9.843668784458971e-05, "loss": 0.4589, "step": 7550 }, { "epoch": 4.4004656577415595, "grad_norm": 1.3052088022232056, "learning_rate": 9.842984331824437e-05, "loss": 0.4705, "step": 7560 }, { "epoch": 4.40628637951106, "grad_norm": 1.3903422355651855, "learning_rate": 9.842298408019133e-05, "loss": 0.4638, "step": 7570 }, { "epoch": 4.412107101280559, "grad_norm": 1.424806833267212, "learning_rate": 9.841611013251429e-05, "loss": 0.4887, "step": 7580 }, { "epoch": 4.417927823050058, "grad_norm": 1.323277235031128, "learning_rate": 9.840922147730133e-05, "loss": 0.4671, "step": 7590 }, { "epoch": 4.423748544819557, "grad_norm": 1.2106677293777466, "learning_rate": 9.840231811664506e-05, "loss": 0.4776, "step": 7600 }, { "epoch": 4.429569266589057, "grad_norm": 1.3590097427368164, "learning_rate": 9.839540005264252e-05, "loss": 0.4624, "step": 7610 }, { "epoch": 4.435389988358557, "grad_norm": 1.0745830535888672, "learning_rate": 9.838846728739527e-05, "loss": 0.463, "step": 7620 }, { "epoch": 4.441210710128056, "grad_norm": 1.2718356847763062, "learning_rate": 9.838151982300927e-05, "loss": 0.4745, "step": 7630 }, { "epoch": 4.447031431897555, "grad_norm": 1.3798089027404785, "learning_rate": 9.8374557661595e-05, "loss": 0.4698, "step": 7640 }, { "epoch": 4.452852153667055, "grad_norm": 1.4077763557434082, "learning_rate": 9.836758080526735e-05, "loss": 0.4794, "step": 7650 }, { "epoch": 4.458672875436554, "grad_norm": 1.271268367767334, "learning_rate": 9.836058925614575e-05, "loss": 0.4728, "step": 7660 }, { "epoch": 4.4644935972060535, "grad_norm": 1.2982815504074097, "learning_rate": 9.8353583016354e-05, "loss": 0.4635, "step": 7670 }, { "epoch": 4.470314318975553, "grad_norm": 1.2887282371520996, "learning_rate": 9.834656208802044e-05, "loss": 0.4672, "step": 7680 }, { "epoch": 4.476135040745052, "grad_norm": 1.2812074422836304, "learning_rate": 9.833952647327784e-05, "loss": 0.4603, "step": 7690 }, { "epoch": 4.481955762514552, "grad_norm": 1.4211792945861816, "learning_rate": 9.833247617426342e-05, "loss": 0.4743, "step": 7700 }, { "epoch": 4.487776484284051, "grad_norm": 1.4753303527832031, "learning_rate": 9.832541119311889e-05, "loss": 0.4764, "step": 7710 }, { "epoch": 4.4935972060535505, "grad_norm": 1.2074272632598877, "learning_rate": 9.83183315319904e-05, "loss": 0.4535, "step": 7720 }, { "epoch": 4.49941792782305, "grad_norm": 1.2607215642929077, "learning_rate": 9.831123719302855e-05, "loss": 0.485, "step": 7730 }, { "epoch": 4.50523864959255, "grad_norm": 1.2705451250076294, "learning_rate": 9.830412817838842e-05, "loss": 0.4495, "step": 7740 }, { "epoch": 4.511059371362049, "grad_norm": 1.128655195236206, "learning_rate": 9.829700449022956e-05, "loss": 0.4605, "step": 7750 }, { "epoch": 4.516880093131548, "grad_norm": 1.2747613191604614, "learning_rate": 9.828986613071593e-05, "loss": 0.4591, "step": 7760 }, { "epoch": 4.5227008149010475, "grad_norm": 1.2967031002044678, "learning_rate": 9.828271310201601e-05, "loss": 0.4635, "step": 7770 }, { "epoch": 4.528521536670548, "grad_norm": 1.493586778640747, "learning_rate": 9.827554540630268e-05, "loss": 0.4783, "step": 7780 }, { "epoch": 4.534342258440047, "grad_norm": 1.1957406997680664, "learning_rate": 9.826836304575329e-05, "loss": 0.4543, "step": 7790 }, { "epoch": 4.540162980209546, "grad_norm": 1.2097604274749756, "learning_rate": 9.826116602254966e-05, "loss": 0.4605, "step": 7800 }, { "epoch": 4.545983701979045, "grad_norm": 1.3075367212295532, "learning_rate": 9.825395433887805e-05, "loss": 0.4558, "step": 7810 }, { "epoch": 4.551804423748544, "grad_norm": 1.2736986875534058, "learning_rate": 9.824672799692917e-05, "loss": 0.4715, "step": 7820 }, { "epoch": 4.5576251455180445, "grad_norm": 1.294832468032837, "learning_rate": 9.823948699889823e-05, "loss": 0.4877, "step": 7830 }, { "epoch": 4.563445867287544, "grad_norm": 1.0749993324279785, "learning_rate": 9.823223134698483e-05, "loss": 0.4659, "step": 7840 }, { "epoch": 4.569266589057043, "grad_norm": 1.1775898933410645, "learning_rate": 9.822496104339303e-05, "loss": 0.4626, "step": 7850 }, { "epoch": 4.575087310826542, "grad_norm": 1.2622172832489014, "learning_rate": 9.821767609033138e-05, "loss": 0.4565, "step": 7860 }, { "epoch": 4.580908032596042, "grad_norm": 1.2936962842941284, "learning_rate": 9.821037649001284e-05, "loss": 0.4487, "step": 7870 }, { "epoch": 4.5867287543655415, "grad_norm": 1.2991769313812256, "learning_rate": 9.820306224465486e-05, "loss": 0.4729, "step": 7880 }, { "epoch": 4.592549476135041, "grad_norm": 1.2297457456588745, "learning_rate": 9.819573335647928e-05, "loss": 0.4764, "step": 7890 }, { "epoch": 4.59837019790454, "grad_norm": 1.283860683441162, "learning_rate": 9.818838982771246e-05, "loss": 0.4566, "step": 7900 }, { "epoch": 4.604190919674039, "grad_norm": 1.2460992336273193, "learning_rate": 9.818103166058514e-05, "loss": 0.4902, "step": 7910 }, { "epoch": 4.610011641443539, "grad_norm": 1.2153984308242798, "learning_rate": 9.817365885733254e-05, "loss": 0.4544, "step": 7920 }, { "epoch": 4.6158323632130385, "grad_norm": 1.2643271684646606, "learning_rate": 9.816627142019434e-05, "loss": 0.4707, "step": 7930 }, { "epoch": 4.621653084982538, "grad_norm": 1.2082406282424927, "learning_rate": 9.815886935141463e-05, "loss": 0.472, "step": 7940 }, { "epoch": 4.627473806752037, "grad_norm": 1.1986031532287598, "learning_rate": 9.8151452653242e-05, "loss": 0.4888, "step": 7950 }, { "epoch": 4.633294528521537, "grad_norm": 1.2459275722503662, "learning_rate": 9.814402132792939e-05, "loss": 0.4982, "step": 7960 }, { "epoch": 4.639115250291036, "grad_norm": 1.508760929107666, "learning_rate": 9.813657537773428e-05, "loss": 0.4728, "step": 7970 }, { "epoch": 4.644935972060535, "grad_norm": 1.2922576665878296, "learning_rate": 9.812911480491854e-05, "loss": 0.479, "step": 7980 }, { "epoch": 4.650756693830035, "grad_norm": 1.271127462387085, "learning_rate": 9.81216396117485e-05, "loss": 0.4695, "step": 7990 }, { "epoch": 4.656577415599534, "grad_norm": 1.2605342864990234, "learning_rate": 9.811414980049491e-05, "loss": 0.4678, "step": 8000 }, { "epoch": 4.662398137369034, "grad_norm": 1.2637488842010498, "learning_rate": 9.810664537343301e-05, "loss": 0.47, "step": 8010 }, { "epoch": 4.668218859138533, "grad_norm": 1.1047930717468262, "learning_rate": 9.809912633284243e-05, "loss": 0.4705, "step": 8020 }, { "epoch": 4.674039580908032, "grad_norm": 1.197517991065979, "learning_rate": 9.809159268100725e-05, "loss": 0.4715, "step": 8030 }, { "epoch": 4.6798603026775325, "grad_norm": 1.195385217666626, "learning_rate": 9.808404442021599e-05, "loss": 0.457, "step": 8040 }, { "epoch": 4.685681024447032, "grad_norm": 1.2529963254928589, "learning_rate": 9.807648155276163e-05, "loss": 0.4539, "step": 8050 }, { "epoch": 4.691501746216531, "grad_norm": 1.3393511772155762, "learning_rate": 9.806890408094156e-05, "loss": 0.4678, "step": 8060 }, { "epoch": 4.69732246798603, "grad_norm": 1.2637048959732056, "learning_rate": 9.806131200705761e-05, "loss": 0.4569, "step": 8070 }, { "epoch": 4.703143189755529, "grad_norm": 1.2702840566635132, "learning_rate": 9.805370533341605e-05, "loss": 0.4713, "step": 8080 }, { "epoch": 4.7089639115250295, "grad_norm": 1.3869352340698242, "learning_rate": 9.804608406232762e-05, "loss": 0.4599, "step": 8090 }, { "epoch": 4.714784633294529, "grad_norm": 1.1356024742126465, "learning_rate": 9.803844819610741e-05, "loss": 0.4462, "step": 8100 }, { "epoch": 4.720605355064028, "grad_norm": 1.2935993671417236, "learning_rate": 9.803079773707504e-05, "loss": 0.4609, "step": 8110 }, { "epoch": 4.726426076833527, "grad_norm": 1.1321817636489868, "learning_rate": 9.802313268755447e-05, "loss": 0.4661, "step": 8120 }, { "epoch": 4.732246798603027, "grad_norm": 1.1032013893127441, "learning_rate": 9.801545304987419e-05, "loss": 0.4496, "step": 8130 }, { "epoch": 4.738067520372526, "grad_norm": 1.1836689710617065, "learning_rate": 9.800775882636704e-05, "loss": 0.4687, "step": 8140 }, { "epoch": 4.743888242142026, "grad_norm": 1.270978331565857, "learning_rate": 9.800005001937034e-05, "loss": 0.4653, "step": 8150 }, { "epoch": 4.749708963911525, "grad_norm": 1.1033276319503784, "learning_rate": 9.79923266312258e-05, "loss": 0.4526, "step": 8160 }, { "epoch": 4.755529685681024, "grad_norm": 1.2448137998580933, "learning_rate": 9.79845886642796e-05, "loss": 0.4514, "step": 8170 }, { "epoch": 4.761350407450524, "grad_norm": 1.2132056951522827, "learning_rate": 9.797683612088233e-05, "loss": 0.466, "step": 8180 }, { "epoch": 4.767171129220023, "grad_norm": 1.2340335845947266, "learning_rate": 9.796906900338898e-05, "loss": 0.463, "step": 8190 }, { "epoch": 4.772991850989523, "grad_norm": 1.3156864643096924, "learning_rate": 9.796128731415903e-05, "loss": 0.4562, "step": 8200 }, { "epoch": 4.778812572759022, "grad_norm": 1.0979681015014648, "learning_rate": 9.795349105555634e-05, "loss": 0.4501, "step": 8210 }, { "epoch": 4.784633294528522, "grad_norm": 1.1482429504394531, "learning_rate": 9.794568022994922e-05, "loss": 0.4625, "step": 8220 }, { "epoch": 4.790454016298021, "grad_norm": 1.4255115985870361, "learning_rate": 9.793785483971034e-05, "loss": 0.4663, "step": 8230 }, { "epoch": 4.79627473806752, "grad_norm": 1.1553003787994385, "learning_rate": 9.793001488721691e-05, "loss": 0.4543, "step": 8240 }, { "epoch": 4.80209545983702, "grad_norm": 1.2561644315719604, "learning_rate": 9.792216037485047e-05, "loss": 0.4544, "step": 8250 }, { "epoch": 4.807916181606519, "grad_norm": 1.3107410669326782, "learning_rate": 9.791429130499704e-05, "loss": 0.4586, "step": 8260 }, { "epoch": 4.813736903376019, "grad_norm": 1.2042505741119385, "learning_rate": 9.790640768004698e-05, "loss": 0.4883, "step": 8270 }, { "epoch": 4.819557625145518, "grad_norm": 1.1174209117889404, "learning_rate": 9.789850950239518e-05, "loss": 0.4617, "step": 8280 }, { "epoch": 4.825378346915017, "grad_norm": 1.265519380569458, "learning_rate": 9.789059677444089e-05, "loss": 0.4506, "step": 8290 }, { "epoch": 4.8311990686845165, "grad_norm": 1.1498188972473145, "learning_rate": 9.788266949858776e-05, "loss": 0.4487, "step": 8300 }, { "epoch": 4.837019790454017, "grad_norm": 1.2341748476028442, "learning_rate": 9.787472767724392e-05, "loss": 0.4634, "step": 8310 }, { "epoch": 4.842840512223516, "grad_norm": 1.4737939834594727, "learning_rate": 9.786677131282185e-05, "loss": 0.4679, "step": 8320 }, { "epoch": 4.848661233993015, "grad_norm": 1.217149257659912, "learning_rate": 9.785880040773853e-05, "loss": 0.4564, "step": 8330 }, { "epoch": 4.854481955762514, "grad_norm": 1.2193306684494019, "learning_rate": 9.785081496441527e-05, "loss": 0.4657, "step": 8340 }, { "epoch": 4.8603026775320135, "grad_norm": 1.1591646671295166, "learning_rate": 9.784281498527785e-05, "loss": 0.4715, "step": 8350 }, { "epoch": 4.866123399301514, "grad_norm": 1.2546614408493042, "learning_rate": 9.783480047275646e-05, "loss": 0.4617, "step": 8360 }, { "epoch": 4.871944121071013, "grad_norm": 1.3031560182571411, "learning_rate": 9.78267714292857e-05, "loss": 0.4455, "step": 8370 }, { "epoch": 4.877764842840512, "grad_norm": 1.147194743156433, "learning_rate": 9.781872785730454e-05, "loss": 0.4529, "step": 8380 }, { "epoch": 4.883585564610011, "grad_norm": 1.1950095891952515, "learning_rate": 9.781066975925646e-05, "loss": 0.4517, "step": 8390 }, { "epoch": 4.889406286379511, "grad_norm": 1.0820069313049316, "learning_rate": 9.780259713758928e-05, "loss": 0.4443, "step": 8400 }, { "epoch": 4.895227008149011, "grad_norm": 1.27780282497406, "learning_rate": 9.779450999475524e-05, "loss": 0.4723, "step": 8410 }, { "epoch": 4.90104772991851, "grad_norm": 1.1957966089248657, "learning_rate": 9.7786408333211e-05, "loss": 0.453, "step": 8420 }, { "epoch": 4.906868451688009, "grad_norm": 1.2056093215942383, "learning_rate": 9.777829215541764e-05, "loss": 0.4521, "step": 8430 }, { "epoch": 4.912689173457509, "grad_norm": 1.1954177618026733, "learning_rate": 9.777016146384064e-05, "loss": 0.4419, "step": 8440 }, { "epoch": 4.918509895227008, "grad_norm": 1.2177698612213135, "learning_rate": 9.776201626094988e-05, "loss": 0.4569, "step": 8450 }, { "epoch": 4.9243306169965075, "grad_norm": 1.218955636024475, "learning_rate": 9.775385654921965e-05, "loss": 0.4395, "step": 8460 }, { "epoch": 4.930151338766007, "grad_norm": 1.1752305030822754, "learning_rate": 9.774568233112868e-05, "loss": 0.4524, "step": 8470 }, { "epoch": 4.935972060535507, "grad_norm": 1.326972246170044, "learning_rate": 9.773749360916007e-05, "loss": 0.468, "step": 8480 }, { "epoch": 4.941792782305006, "grad_norm": 1.17852783203125, "learning_rate": 9.772929038580134e-05, "loss": 0.4582, "step": 8490 }, { "epoch": 4.947613504074505, "grad_norm": 1.2723263502120972, "learning_rate": 9.772107266354439e-05, "loss": 0.4428, "step": 8500 }, { "epoch": 4.9534342258440045, "grad_norm": 1.1110504865646362, "learning_rate": 9.77128404448856e-05, "loss": 0.4607, "step": 8510 }, { "epoch": 4.959254947613504, "grad_norm": 1.2682617902755737, "learning_rate": 9.770459373232565e-05, "loss": 0.4635, "step": 8520 }, { "epoch": 4.965075669383004, "grad_norm": 1.1968591213226318, "learning_rate": 9.769633252836969e-05, "loss": 0.4499, "step": 8530 }, { "epoch": 4.970896391152503, "grad_norm": 1.2225145101547241, "learning_rate": 9.768805683552724e-05, "loss": 0.451, "step": 8540 }, { "epoch": 4.976717112922002, "grad_norm": 1.2044776678085327, "learning_rate": 9.767976665631228e-05, "loss": 0.4479, "step": 8550 }, { "epoch": 4.9825378346915015, "grad_norm": 1.2465180158615112, "learning_rate": 9.767146199324311e-05, "loss": 0.4459, "step": 8560 }, { "epoch": 4.988358556461002, "grad_norm": 1.2140320539474487, "learning_rate": 9.766314284884249e-05, "loss": 0.4568, "step": 8570 }, { "epoch": 4.994179278230501, "grad_norm": 1.2004663944244385, "learning_rate": 9.765480922563752e-05, "loss": 0.4481, "step": 8580 }, { "epoch": 5.0, "grad_norm": 1.4215726852416992, "learning_rate": 9.764646112615978e-05, "loss": 0.4469, "step": 8590 }, { "epoch": 5.005820721769499, "grad_norm": 1.3227471113204956, "learning_rate": 9.763809855294517e-05, "loss": 0.4687, "step": 8600 }, { "epoch": 5.011641443538998, "grad_norm": 1.228171706199646, "learning_rate": 9.762972150853404e-05, "loss": 0.4645, "step": 8610 }, { "epoch": 5.0174621653084985, "grad_norm": 1.1037095785140991, "learning_rate": 9.762132999547111e-05, "loss": 0.4529, "step": 8620 }, { "epoch": 5.023282887077998, "grad_norm": 1.179108738899231, "learning_rate": 9.761292401630549e-05, "loss": 0.4551, "step": 8630 }, { "epoch": 5.029103608847497, "grad_norm": 1.1716400384902954, "learning_rate": 9.76045035735907e-05, "loss": 0.4479, "step": 8640 }, { "epoch": 5.034924330616996, "grad_norm": 1.19490385055542, "learning_rate": 9.759606866988464e-05, "loss": 0.4643, "step": 8650 }, { "epoch": 5.040745052386496, "grad_norm": 1.3011882305145264, "learning_rate": 9.758761930774963e-05, "loss": 0.4576, "step": 8660 }, { "epoch": 5.0465657741559955, "grad_norm": 1.3763166666030884, "learning_rate": 9.757915548975235e-05, "loss": 0.4515, "step": 8670 }, { "epoch": 5.052386495925495, "grad_norm": 1.1128971576690674, "learning_rate": 9.757067721846389e-05, "loss": 0.4593, "step": 8680 }, { "epoch": 5.058207217694994, "grad_norm": 1.2776720523834229, "learning_rate": 9.756218449645971e-05, "loss": 0.4551, "step": 8690 }, { "epoch": 5.064027939464494, "grad_norm": 1.3109244108200073, "learning_rate": 9.75536773263197e-05, "loss": 0.448, "step": 8700 }, { "epoch": 5.069848661233993, "grad_norm": 1.30994713306427, "learning_rate": 9.75451557106281e-05, "loss": 0.4332, "step": 8710 }, { "epoch": 5.0756693830034925, "grad_norm": 1.2049545049667358, "learning_rate": 9.753661965197354e-05, "loss": 0.4563, "step": 8720 }, { "epoch": 5.081490104772992, "grad_norm": 1.1810322999954224, "learning_rate": 9.752806915294908e-05, "loss": 0.4546, "step": 8730 }, { "epoch": 5.087310826542491, "grad_norm": 1.2136940956115723, "learning_rate": 9.75195042161521e-05, "loss": 0.4478, "step": 8740 }, { "epoch": 5.093131548311991, "grad_norm": 1.1704648733139038, "learning_rate": 9.751092484418442e-05, "loss": 0.4488, "step": 8750 }, { "epoch": 5.09895227008149, "grad_norm": 1.2521005868911743, "learning_rate": 9.750233103965224e-05, "loss": 0.467, "step": 8760 }, { "epoch": 5.104772991850989, "grad_norm": 1.275133728981018, "learning_rate": 9.749372280516611e-05, "loss": 0.4472, "step": 8770 }, { "epoch": 5.110593713620489, "grad_norm": 1.357383131980896, "learning_rate": 9.748510014334097e-05, "loss": 0.4608, "step": 8780 }, { "epoch": 5.116414435389989, "grad_norm": 1.1680039167404175, "learning_rate": 9.747646305679621e-05, "loss": 0.4621, "step": 8790 }, { "epoch": 5.122235157159488, "grad_norm": 1.2516281604766846, "learning_rate": 9.74678115481555e-05, "loss": 0.458, "step": 8800 }, { "epoch": 5.128055878928987, "grad_norm": 1.201210618019104, "learning_rate": 9.745914562004696e-05, "loss": 0.4574, "step": 8810 }, { "epoch": 5.133876600698486, "grad_norm": 1.2390812635421753, "learning_rate": 9.745046527510307e-05, "loss": 0.4541, "step": 8820 }, { "epoch": 5.139697322467986, "grad_norm": 1.263650894165039, "learning_rate": 9.744177051596068e-05, "loss": 0.468, "step": 8830 }, { "epoch": 5.145518044237486, "grad_norm": 1.1758009195327759, "learning_rate": 9.743306134526105e-05, "loss": 0.4424, "step": 8840 }, { "epoch": 5.151338766006985, "grad_norm": 1.2135963439941406, "learning_rate": 9.742433776564977e-05, "loss": 0.4609, "step": 8850 }, { "epoch": 5.157159487776484, "grad_norm": 1.1737381219863892, "learning_rate": 9.741559977977683e-05, "loss": 0.4551, "step": 8860 }, { "epoch": 5.162980209545983, "grad_norm": 1.1957175731658936, "learning_rate": 9.740684739029661e-05, "loss": 0.4376, "step": 8870 }, { "epoch": 5.1688009313154835, "grad_norm": 1.23824143409729, "learning_rate": 9.739808059986789e-05, "loss": 0.447, "step": 8880 }, { "epoch": 5.174621653084983, "grad_norm": 1.3649184703826904, "learning_rate": 9.738929941115373e-05, "loss": 0.448, "step": 8890 }, { "epoch": 5.180442374854482, "grad_norm": 1.1201858520507812, "learning_rate": 9.738050382682167e-05, "loss": 0.4656, "step": 8900 }, { "epoch": 5.186263096623981, "grad_norm": 1.2032488584518433, "learning_rate": 9.737169384954355e-05, "loss": 0.4383, "step": 8910 }, { "epoch": 5.192083818393481, "grad_norm": 1.2586114406585693, "learning_rate": 9.736286948199562e-05, "loss": 0.4316, "step": 8920 }, { "epoch": 5.19790454016298, "grad_norm": 1.1663180589675903, "learning_rate": 9.735403072685848e-05, "loss": 0.4416, "step": 8930 }, { "epoch": 5.20372526193248, "grad_norm": 1.6141871213912964, "learning_rate": 9.734517758681712e-05, "loss": 0.4529, "step": 8940 }, { "epoch": 5.209545983701979, "grad_norm": 1.2224981784820557, "learning_rate": 9.733631006456088e-05, "loss": 0.4603, "step": 8950 }, { "epoch": 5.215366705471478, "grad_norm": 1.1546244621276855, "learning_rate": 9.732742816278348e-05, "loss": 0.4577, "step": 8960 }, { "epoch": 5.221187427240978, "grad_norm": 1.2748117446899414, "learning_rate": 9.731853188418302e-05, "loss": 0.4405, "step": 8970 }, { "epoch": 5.227008149010477, "grad_norm": 1.1453349590301514, "learning_rate": 9.730962123146194e-05, "loss": 0.4525, "step": 8980 }, { "epoch": 5.232828870779977, "grad_norm": 1.2548115253448486, "learning_rate": 9.730069620732709e-05, "loss": 0.4602, "step": 8990 }, { "epoch": 5.238649592549476, "grad_norm": 1.2293144464492798, "learning_rate": 9.72917568144896e-05, "loss": 0.468, "step": 9000 }, { "epoch": 5.244470314318976, "grad_norm": 1.2596665620803833, "learning_rate": 9.728280305566509e-05, "loss": 0.4466, "step": 9010 }, { "epoch": 5.250291036088475, "grad_norm": 1.0937671661376953, "learning_rate": 9.727383493357343e-05, "loss": 0.4267, "step": 9020 }, { "epoch": 5.256111757857974, "grad_norm": 1.2252602577209473, "learning_rate": 9.726485245093891e-05, "loss": 0.4415, "step": 9030 }, { "epoch": 5.261932479627474, "grad_norm": 1.2319279909133911, "learning_rate": 9.725585561049018e-05, "loss": 0.4416, "step": 9040 }, { "epoch": 5.267753201396973, "grad_norm": 1.1982002258300781, "learning_rate": 9.724684441496022e-05, "loss": 0.4489, "step": 9050 }, { "epoch": 5.273573923166473, "grad_norm": 1.215409517288208, "learning_rate": 9.72378188670864e-05, "loss": 0.4401, "step": 9060 }, { "epoch": 5.279394644935972, "grad_norm": 1.4921222925186157, "learning_rate": 9.722877896961047e-05, "loss": 0.4467, "step": 9070 }, { "epoch": 5.285215366705471, "grad_norm": 1.1721526384353638, "learning_rate": 9.721972472527848e-05, "loss": 0.4302, "step": 9080 }, { "epoch": 5.2910360884749705, "grad_norm": 1.1213278770446777, "learning_rate": 9.721065613684089e-05, "loss": 0.4469, "step": 9090 }, { "epoch": 5.296856810244471, "grad_norm": 1.1082866191864014, "learning_rate": 9.72015732070525e-05, "loss": 0.4493, "step": 9100 }, { "epoch": 5.30267753201397, "grad_norm": 1.2003872394561768, "learning_rate": 9.719247593867244e-05, "loss": 0.4529, "step": 9110 }, { "epoch": 5.308498253783469, "grad_norm": 1.2206367254257202, "learning_rate": 9.718336433446423e-05, "loss": 0.4499, "step": 9120 }, { "epoch": 5.314318975552968, "grad_norm": 1.2403734922409058, "learning_rate": 9.717423839719574e-05, "loss": 0.4239, "step": 9130 }, { "epoch": 5.320139697322468, "grad_norm": 1.19962477684021, "learning_rate": 9.71650981296392e-05, "loss": 0.4534, "step": 9140 }, { "epoch": 5.325960419091968, "grad_norm": 1.1571992635726929, "learning_rate": 9.715594353457118e-05, "loss": 0.4576, "step": 9150 }, { "epoch": 5.331781140861467, "grad_norm": 1.2543100118637085, "learning_rate": 9.714677461477257e-05, "loss": 0.4475, "step": 9160 }, { "epoch": 5.337601862630966, "grad_norm": 1.3168036937713623, "learning_rate": 9.713759137302869e-05, "loss": 0.4566, "step": 9170 }, { "epoch": 5.343422584400465, "grad_norm": 1.2853105068206787, "learning_rate": 9.712839381212914e-05, "loss": 0.4543, "step": 9180 }, { "epoch": 5.349243306169965, "grad_norm": 1.1657466888427734, "learning_rate": 9.71191819348679e-05, "loss": 0.4501, "step": 9190 }, { "epoch": 5.355064027939465, "grad_norm": 1.2479323148727417, "learning_rate": 9.710995574404331e-05, "loss": 0.4527, "step": 9200 }, { "epoch": 5.360884749708964, "grad_norm": 1.2229150533676147, "learning_rate": 9.710071524245802e-05, "loss": 0.4512, "step": 9210 }, { "epoch": 5.366705471478463, "grad_norm": 1.0888222455978394, "learning_rate": 9.709146043291906e-05, "loss": 0.4628, "step": 9220 }, { "epoch": 5.372526193247963, "grad_norm": 1.29571533203125, "learning_rate": 9.70821913182378e-05, "loss": 0.441, "step": 9230 }, { "epoch": 5.378346915017462, "grad_norm": 1.224395751953125, "learning_rate": 9.707290790122995e-05, "loss": 0.4464, "step": 9240 }, { "epoch": 5.3841676367869615, "grad_norm": 1.3054018020629883, "learning_rate": 9.706361018471557e-05, "loss": 0.462, "step": 9250 }, { "epoch": 5.389988358556461, "grad_norm": 1.1657228469848633, "learning_rate": 9.705429817151906e-05, "loss": 0.4383, "step": 9260 }, { "epoch": 5.395809080325961, "grad_norm": 1.1609225273132324, "learning_rate": 9.704497186446917e-05, "loss": 0.4435, "step": 9270 }, { "epoch": 5.40162980209546, "grad_norm": 1.1216706037521362, "learning_rate": 9.703563126639896e-05, "loss": 0.4372, "step": 9280 }, { "epoch": 5.407450523864959, "grad_norm": 1.1904447078704834, "learning_rate": 9.70262763801459e-05, "loss": 0.4324, "step": 9290 }, { "epoch": 5.4132712456344585, "grad_norm": 1.2905863523483276, "learning_rate": 9.701690720855171e-05, "loss": 0.4486, "step": 9300 }, { "epoch": 5.419091967403958, "grad_norm": 1.1330195665359497, "learning_rate": 9.700752375446253e-05, "loss": 0.4471, "step": 9310 }, { "epoch": 5.424912689173458, "grad_norm": 1.160679578781128, "learning_rate": 9.69981260207288e-05, "loss": 0.4398, "step": 9320 }, { "epoch": 5.430733410942957, "grad_norm": 1.2331434488296509, "learning_rate": 9.698871401020529e-05, "loss": 0.4595, "step": 9330 }, { "epoch": 5.436554132712456, "grad_norm": 1.2500107288360596, "learning_rate": 9.697928772575112e-05, "loss": 0.4655, "step": 9340 }, { "epoch": 5.4423748544819555, "grad_norm": 1.0772045850753784, "learning_rate": 9.696984717022976e-05, "loss": 0.4319, "step": 9350 }, { "epoch": 5.448195576251456, "grad_norm": 1.2648571729660034, "learning_rate": 9.6960392346509e-05, "loss": 0.4507, "step": 9360 }, { "epoch": 5.454016298020955, "grad_norm": 1.2652721405029297, "learning_rate": 9.695092325746097e-05, "loss": 0.4524, "step": 9370 }, { "epoch": 5.459837019790454, "grad_norm": 1.2282251119613647, "learning_rate": 9.694143990596211e-05, "loss": 0.4634, "step": 9380 }, { "epoch": 5.465657741559953, "grad_norm": 1.0741585493087769, "learning_rate": 9.693194229489325e-05, "loss": 0.4598, "step": 9390 }, { "epoch": 5.471478463329452, "grad_norm": 1.2216109037399292, "learning_rate": 9.692243042713944e-05, "loss": 0.4484, "step": 9400 }, { "epoch": 5.4772991850989525, "grad_norm": 1.401711344718933, "learning_rate": 9.691290430559022e-05, "loss": 0.4507, "step": 9410 }, { "epoch": 5.483119906868452, "grad_norm": 1.243691086769104, "learning_rate": 9.690336393313932e-05, "loss": 0.4572, "step": 9420 }, { "epoch": 5.488940628637951, "grad_norm": 1.2016209363937378, "learning_rate": 9.689380931268487e-05, "loss": 0.4376, "step": 9430 }, { "epoch": 5.49476135040745, "grad_norm": 1.1911054849624634, "learning_rate": 9.688424044712932e-05, "loss": 0.4479, "step": 9440 }, { "epoch": 5.50058207217695, "grad_norm": 1.1341463327407837, "learning_rate": 9.687465733937942e-05, "loss": 0.4405, "step": 9450 }, { "epoch": 5.5064027939464495, "grad_norm": 1.3516442775726318, "learning_rate": 9.686505999234627e-05, "loss": 0.4559, "step": 9460 }, { "epoch": 5.512223515715949, "grad_norm": 1.1728814840316772, "learning_rate": 9.685544840894529e-05, "loss": 0.4558, "step": 9470 }, { "epoch": 5.518044237485448, "grad_norm": 1.2213845252990723, "learning_rate": 9.684582259209624e-05, "loss": 0.4434, "step": 9480 }, { "epoch": 5.523864959254947, "grad_norm": 1.1362087726593018, "learning_rate": 9.683618254472317e-05, "loss": 0.4499, "step": 9490 }, { "epoch": 5.529685681024447, "grad_norm": 1.2430719137191772, "learning_rate": 9.682652826975449e-05, "loss": 0.4459, "step": 9500 }, { "epoch": 5.5355064027939465, "grad_norm": 1.0758908987045288, "learning_rate": 9.681685977012291e-05, "loss": 0.4363, "step": 9510 }, { "epoch": 5.541327124563446, "grad_norm": 1.1709176301956177, "learning_rate": 9.680717704876546e-05, "loss": 0.4439, "step": 9520 }, { "epoch": 5.547147846332946, "grad_norm": 1.281245470046997, "learning_rate": 9.679748010862349e-05, "loss": 0.4508, "step": 9530 }, { "epoch": 5.552968568102445, "grad_norm": 1.2025375366210938, "learning_rate": 9.678776895264267e-05, "loss": 0.4458, "step": 9540 }, { "epoch": 5.558789289871944, "grad_norm": 1.2330130338668823, "learning_rate": 9.6778043583773e-05, "loss": 0.4571, "step": 9550 }, { "epoch": 5.564610011641443, "grad_norm": 1.1634575128555298, "learning_rate": 9.67683040049688e-05, "loss": 0.4529, "step": 9560 }, { "epoch": 5.570430733410943, "grad_norm": 1.2935731410980225, "learning_rate": 9.675855021918869e-05, "loss": 0.4433, "step": 9570 }, { "epoch": 5.576251455180443, "grad_norm": 1.1737756729125977, "learning_rate": 9.674878222939561e-05, "loss": 0.4505, "step": 9580 }, { "epoch": 5.582072176949942, "grad_norm": 1.2380452156066895, "learning_rate": 9.673900003855681e-05, "loss": 0.434, "step": 9590 }, { "epoch": 5.587892898719441, "grad_norm": 1.090334177017212, "learning_rate": 9.672920364964389e-05, "loss": 0.4467, "step": 9600 }, { "epoch": 5.59371362048894, "grad_norm": 1.196605920791626, "learning_rate": 9.671939306563269e-05, "loss": 0.46, "step": 9610 }, { "epoch": 5.5995343422584405, "grad_norm": 1.2136880159378052, "learning_rate": 9.670956828950345e-05, "loss": 0.4664, "step": 9620 }, { "epoch": 5.60535506402794, "grad_norm": 1.2051433324813843, "learning_rate": 9.669972932424065e-05, "loss": 0.4503, "step": 9630 }, { "epoch": 5.611175785797439, "grad_norm": 1.11421799659729, "learning_rate": 9.668987617283312e-05, "loss": 0.4405, "step": 9640 }, { "epoch": 5.616996507566938, "grad_norm": 1.0728652477264404, "learning_rate": 9.668000883827397e-05, "loss": 0.4455, "step": 9650 }, { "epoch": 5.622817229336437, "grad_norm": 1.0755887031555176, "learning_rate": 9.667012732356067e-05, "loss": 0.4402, "step": 9660 }, { "epoch": 5.6286379511059375, "grad_norm": 1.1643118858337402, "learning_rate": 9.666023163169493e-05, "loss": 0.4452, "step": 9670 }, { "epoch": 5.634458672875437, "grad_norm": 1.359443187713623, "learning_rate": 9.665032176568281e-05, "loss": 0.4569, "step": 9680 }, { "epoch": 5.640279394644936, "grad_norm": 1.1849114894866943, "learning_rate": 9.664039772853469e-05, "loss": 0.4401, "step": 9690 }, { "epoch": 5.646100116414435, "grad_norm": 1.1891063451766968, "learning_rate": 9.663045952326518e-05, "loss": 0.4388, "step": 9700 }, { "epoch": 5.651920838183935, "grad_norm": 1.077213168144226, "learning_rate": 9.662050715289328e-05, "loss": 0.4531, "step": 9710 }, { "epoch": 5.657741559953434, "grad_norm": 1.0870544910430908, "learning_rate": 9.661054062044226e-05, "loss": 0.4487, "step": 9720 }, { "epoch": 5.663562281722934, "grad_norm": 1.177158236503601, "learning_rate": 9.660055992893968e-05, "loss": 0.4484, "step": 9730 }, { "epoch": 5.669383003492433, "grad_norm": 1.0926679372787476, "learning_rate": 9.659056508141739e-05, "loss": 0.4383, "step": 9740 }, { "epoch": 5.675203725261932, "grad_norm": 1.0781042575836182, "learning_rate": 9.658055608091161e-05, "loss": 0.4371, "step": 9750 }, { "epoch": 5.681024447031432, "grad_norm": 1.127973198890686, "learning_rate": 9.657053293046276e-05, "loss": 0.4568, "step": 9760 }, { "epoch": 5.686845168800931, "grad_norm": 1.1510965824127197, "learning_rate": 9.656049563311564e-05, "loss": 0.4407, "step": 9770 }, { "epoch": 5.692665890570431, "grad_norm": 1.160111904144287, "learning_rate": 9.655044419191929e-05, "loss": 0.4312, "step": 9780 }, { "epoch": 5.69848661233993, "grad_norm": 1.1596847772598267, "learning_rate": 9.654037860992711e-05, "loss": 0.4383, "step": 9790 }, { "epoch": 5.70430733410943, "grad_norm": 1.0604851245880127, "learning_rate": 9.653029889019672e-05, "loss": 0.4329, "step": 9800 }, { "epoch": 5.710128055878929, "grad_norm": 1.323602557182312, "learning_rate": 9.65202050357901e-05, "loss": 0.4326, "step": 9810 }, { "epoch": 5.715948777648428, "grad_norm": 1.3075898885726929, "learning_rate": 9.651009704977347e-05, "loss": 0.4313, "step": 9820 }, { "epoch": 5.721769499417928, "grad_norm": 1.1967765092849731, "learning_rate": 9.649997493521738e-05, "loss": 0.4494, "step": 9830 }, { "epoch": 5.727590221187427, "grad_norm": 1.3972444534301758, "learning_rate": 9.64898386951967e-05, "loss": 0.4516, "step": 9840 }, { "epoch": 5.733410942956927, "grad_norm": 1.1762176752090454, "learning_rate": 9.647968833279049e-05, "loss": 0.4573, "step": 9850 }, { "epoch": 5.739231664726426, "grad_norm": 1.2071372270584106, "learning_rate": 9.646952385108218e-05, "loss": 0.4425, "step": 9860 }, { "epoch": 5.745052386495925, "grad_norm": 1.16315758228302, "learning_rate": 9.645934525315951e-05, "loss": 0.4284, "step": 9870 }, { "epoch": 5.7508731082654245, "grad_norm": 1.1568176746368408, "learning_rate": 9.644915254211442e-05, "loss": 0.4314, "step": 9880 }, { "epoch": 5.756693830034925, "grad_norm": 1.2904456853866577, "learning_rate": 9.643894572104321e-05, "loss": 0.4551, "step": 9890 }, { "epoch": 5.762514551804424, "grad_norm": 1.0880452394485474, "learning_rate": 9.642872479304644e-05, "loss": 0.4484, "step": 9900 }, { "epoch": 5.768335273573923, "grad_norm": 1.3711472749710083, "learning_rate": 9.641848976122895e-05, "loss": 0.4361, "step": 9910 }, { "epoch": 5.774155995343422, "grad_norm": 1.1714434623718262, "learning_rate": 9.64082406286999e-05, "loss": 0.4414, "step": 9920 }, { "epoch": 5.779976717112922, "grad_norm": 1.1895512342453003, "learning_rate": 9.639797739857269e-05, "loss": 0.4251, "step": 9930 }, { "epoch": 5.785797438882422, "grad_norm": 1.1834602355957031, "learning_rate": 9.638770007396498e-05, "loss": 0.4633, "step": 9940 }, { "epoch": 5.791618160651921, "grad_norm": 1.1963309049606323, "learning_rate": 9.63774086579988e-05, "loss": 0.4609, "step": 9950 }, { "epoch": 5.79743888242142, "grad_norm": 1.149266242980957, "learning_rate": 9.63671031538004e-05, "loss": 0.4291, "step": 9960 }, { "epoch": 5.80325960419092, "grad_norm": 1.1031285524368286, "learning_rate": 9.635678356450031e-05, "loss": 0.4474, "step": 9970 }, { "epoch": 5.809080325960419, "grad_norm": 1.2793986797332764, "learning_rate": 9.634644989323336e-05, "loss": 0.44, "step": 9980 }, { "epoch": 5.814901047729919, "grad_norm": 1.2074823379516602, "learning_rate": 9.633610214313861e-05, "loss": 0.4444, "step": 9990 }, { "epoch": 5.820721769499418, "grad_norm": 1.1358811855316162, "learning_rate": 9.632574031735951e-05, "loss": 0.4432, "step": 10000 }, { "epoch": 5.826542491268917, "grad_norm": 1.1019166707992554, "learning_rate": 9.631536441904364e-05, "loss": 0.4385, "step": 10010 }, { "epoch": 5.832363213038417, "grad_norm": 1.050550103187561, "learning_rate": 9.630497445134293e-05, "loss": 0.437, "step": 10020 }, { "epoch": 5.838183934807916, "grad_norm": 1.0588078498840332, "learning_rate": 9.62945704174136e-05, "loss": 0.4436, "step": 10030 }, { "epoch": 5.8440046565774155, "grad_norm": 1.102324366569519, "learning_rate": 9.628415232041612e-05, "loss": 0.464, "step": 10040 }, { "epoch": 5.849825378346915, "grad_norm": 1.0650187730789185, "learning_rate": 9.627372016351524e-05, "loss": 0.4367, "step": 10050 }, { "epoch": 5.855646100116415, "grad_norm": 1.1911996603012085, "learning_rate": 9.626327394987995e-05, "loss": 0.4562, "step": 10060 }, { "epoch": 5.861466821885914, "grad_norm": 1.1704671382904053, "learning_rate": 9.625281368268355e-05, "loss": 0.4455, "step": 10070 }, { "epoch": 5.867287543655413, "grad_norm": 1.0666558742523193, "learning_rate": 9.624233936510357e-05, "loss": 0.4339, "step": 10080 }, { "epoch": 5.8731082654249125, "grad_norm": 1.1353360414505005, "learning_rate": 9.623185100032187e-05, "loss": 0.4287, "step": 10090 }, { "epoch": 5.878928987194412, "grad_norm": 1.154634714126587, "learning_rate": 9.62213485915245e-05, "loss": 0.4411, "step": 10100 }, { "epoch": 5.884749708963912, "grad_norm": 1.0689419507980347, "learning_rate": 9.621083214190186e-05, "loss": 0.4366, "step": 10110 }, { "epoch": 5.890570430733411, "grad_norm": 1.1403982639312744, "learning_rate": 9.62003016546485e-05, "loss": 0.4355, "step": 10120 }, { "epoch": 5.89639115250291, "grad_norm": 1.183569312095642, "learning_rate": 9.618975713296339e-05, "loss": 0.4345, "step": 10130 }, { "epoch": 5.9022118742724095, "grad_norm": 1.2546907663345337, "learning_rate": 9.61791985800496e-05, "loss": 0.4248, "step": 10140 }, { "epoch": 5.90803259604191, "grad_norm": 1.203068733215332, "learning_rate": 9.616862599911458e-05, "loss": 0.4299, "step": 10150 }, { "epoch": 5.913853317811409, "grad_norm": 1.1775448322296143, "learning_rate": 9.615803939337e-05, "loss": 0.4392, "step": 10160 }, { "epoch": 5.919674039580908, "grad_norm": 1.0826560258865356, "learning_rate": 9.614743876603178e-05, "loss": 0.4395, "step": 10170 }, { "epoch": 5.925494761350407, "grad_norm": 1.1860491037368774, "learning_rate": 9.613682412032013e-05, "loss": 0.4367, "step": 10180 }, { "epoch": 5.931315483119906, "grad_norm": 1.1262001991271973, "learning_rate": 9.612619545945947e-05, "loss": 0.4397, "step": 10190 }, { "epoch": 5.9371362048894065, "grad_norm": 1.1638795137405396, "learning_rate": 9.611555278667852e-05, "loss": 0.4456, "step": 10200 }, { "epoch": 5.942956926658906, "grad_norm": 1.1187233924865723, "learning_rate": 9.610489610521024e-05, "loss": 0.4244, "step": 10210 }, { "epoch": 5.948777648428405, "grad_norm": 1.2408429384231567, "learning_rate": 9.609422541829187e-05, "loss": 0.4324, "step": 10220 }, { "epoch": 5.954598370197904, "grad_norm": 1.0988081693649292, "learning_rate": 9.608354072916486e-05, "loss": 0.4407, "step": 10230 }, { "epoch": 5.960419091967404, "grad_norm": 1.1182656288146973, "learning_rate": 9.607284204107493e-05, "loss": 0.4345, "step": 10240 }, { "epoch": 5.9662398137369035, "grad_norm": 1.2350497245788574, "learning_rate": 9.606212935727208e-05, "loss": 0.4412, "step": 10250 }, { "epoch": 5.972060535506403, "grad_norm": 1.1989935636520386, "learning_rate": 9.605140268101052e-05, "loss": 0.4417, "step": 10260 }, { "epoch": 5.977881257275902, "grad_norm": 1.065822958946228, "learning_rate": 9.604066201554875e-05, "loss": 0.4425, "step": 10270 }, { "epoch": 5.983701979045401, "grad_norm": 1.0371700525283813, "learning_rate": 9.60299073641495e-05, "loss": 0.4628, "step": 10280 }, { "epoch": 5.989522700814901, "grad_norm": 1.217310905456543, "learning_rate": 9.601913873007974e-05, "loss": 0.4581, "step": 10290 }, { "epoch": 5.9953434225844005, "grad_norm": 1.0645116567611694, "learning_rate": 9.60083561166107e-05, "loss": 0.4285, "step": 10300 }, { "epoch": 6.0011641443539, "grad_norm": 1.1252890825271606, "learning_rate": 9.599755952701783e-05, "loss": 0.4384, "step": 10310 }, { "epoch": 6.006984866123399, "grad_norm": 1.153885006904602, "learning_rate": 9.598674896458089e-05, "loss": 0.4352, "step": 10320 }, { "epoch": 6.012805587892899, "grad_norm": 1.2031645774841309, "learning_rate": 9.597592443258383e-05, "loss": 0.4329, "step": 10330 }, { "epoch": 6.018626309662398, "grad_norm": 1.1451680660247803, "learning_rate": 9.596508593431483e-05, "loss": 0.4229, "step": 10340 }, { "epoch": 6.024447031431897, "grad_norm": 1.2061463594436646, "learning_rate": 9.59542334730664e-05, "loss": 0.4349, "step": 10350 }, { "epoch": 6.030267753201397, "grad_norm": 1.1202276945114136, "learning_rate": 9.594336705213516e-05, "loss": 0.4217, "step": 10360 }, { "epoch": 6.036088474970897, "grad_norm": 1.0572315454483032, "learning_rate": 9.593248667482208e-05, "loss": 0.4438, "step": 10370 }, { "epoch": 6.041909196740396, "grad_norm": 1.5315200090408325, "learning_rate": 9.592159234443233e-05, "loss": 0.4416, "step": 10380 }, { "epoch": 6.047729918509895, "grad_norm": 1.1754790544509888, "learning_rate": 9.59106840642753e-05, "loss": 0.4265, "step": 10390 }, { "epoch": 6.053550640279394, "grad_norm": 1.0414848327636719, "learning_rate": 9.589976183766467e-05, "loss": 0.4306, "step": 10400 }, { "epoch": 6.0593713620488945, "grad_norm": 1.1338282823562622, "learning_rate": 9.58888256679183e-05, "loss": 0.4309, "step": 10410 }, { "epoch": 6.065192083818394, "grad_norm": 1.114861249923706, "learning_rate": 9.587787555835832e-05, "loss": 0.4337, "step": 10420 }, { "epoch": 6.071012805587893, "grad_norm": 1.239292860031128, "learning_rate": 9.586691151231107e-05, "loss": 0.4268, "step": 10430 }, { "epoch": 6.076833527357392, "grad_norm": 1.103458046913147, "learning_rate": 9.585593353310715e-05, "loss": 0.4376, "step": 10440 }, { "epoch": 6.082654249126891, "grad_norm": 1.1896706819534302, "learning_rate": 9.58449416240814e-05, "loss": 0.441, "step": 10450 }, { "epoch": 6.0884749708963914, "grad_norm": 1.1596651077270508, "learning_rate": 9.583393578857283e-05, "loss": 0.4369, "step": 10460 }, { "epoch": 6.094295692665891, "grad_norm": 1.0901907682418823, "learning_rate": 9.582291602992474e-05, "loss": 0.432, "step": 10470 }, { "epoch": 6.10011641443539, "grad_norm": 1.1093502044677734, "learning_rate": 9.581188235148466e-05, "loss": 0.4336, "step": 10480 }, { "epoch": 6.105937136204889, "grad_norm": 1.1595104932785034, "learning_rate": 9.58008347566043e-05, "loss": 0.4305, "step": 10490 }, { "epoch": 6.111757857974389, "grad_norm": 1.1431313753128052, "learning_rate": 9.578977324863965e-05, "loss": 0.4291, "step": 10500 }, { "epoch": 6.117578579743888, "grad_norm": 1.0276639461517334, "learning_rate": 9.577869783095089e-05, "loss": 0.4372, "step": 10510 }, { "epoch": 6.123399301513388, "grad_norm": 1.2332006692886353, "learning_rate": 9.576760850690245e-05, "loss": 0.4486, "step": 10520 }, { "epoch": 6.129220023282887, "grad_norm": 1.210195779800415, "learning_rate": 9.575650527986298e-05, "loss": 0.4501, "step": 10530 }, { "epoch": 6.135040745052386, "grad_norm": 1.0663834810256958, "learning_rate": 9.574538815320531e-05, "loss": 0.4341, "step": 10540 }, { "epoch": 6.140861466821886, "grad_norm": 1.1294327974319458, "learning_rate": 9.573425713030656e-05, "loss": 0.4337, "step": 10550 }, { "epoch": 6.146682188591385, "grad_norm": 1.261742115020752, "learning_rate": 9.572311221454806e-05, "loss": 0.4409, "step": 10560 }, { "epoch": 6.152502910360885, "grad_norm": 1.1514836549758911, "learning_rate": 9.57119534093153e-05, "loss": 0.4374, "step": 10570 }, { "epoch": 6.158323632130384, "grad_norm": 1.2600032091140747, "learning_rate": 9.570078071799806e-05, "loss": 0.4316, "step": 10580 }, { "epoch": 6.164144353899884, "grad_norm": 1.146855354309082, "learning_rate": 9.568959414399028e-05, "loss": 0.4406, "step": 10590 }, { "epoch": 6.169965075669383, "grad_norm": 1.094628095626831, "learning_rate": 9.567839369069018e-05, "loss": 0.4367, "step": 10600 }, { "epoch": 6.175785797438882, "grad_norm": 1.1379281282424927, "learning_rate": 9.566717936150013e-05, "loss": 0.4481, "step": 10610 }, { "epoch": 6.181606519208382, "grad_norm": 1.0662263631820679, "learning_rate": 9.565595115982678e-05, "loss": 0.4329, "step": 10620 }, { "epoch": 6.187427240977882, "grad_norm": 1.1526529788970947, "learning_rate": 9.564470908908094e-05, "loss": 0.4428, "step": 10630 }, { "epoch": 6.193247962747381, "grad_norm": 1.0656840801239014, "learning_rate": 9.563345315267764e-05, "loss": 0.4247, "step": 10640 }, { "epoch": 6.19906868451688, "grad_norm": 1.1878753900527954, "learning_rate": 9.562218335403616e-05, "loss": 0.4198, "step": 10650 }, { "epoch": 6.204889406286379, "grad_norm": 1.1607093811035156, "learning_rate": 9.561089969657999e-05, "loss": 0.4373, "step": 10660 }, { "epoch": 6.2107101280558785, "grad_norm": 1.238834023475647, "learning_rate": 9.559960218373673e-05, "loss": 0.4314, "step": 10670 }, { "epoch": 6.216530849825379, "grad_norm": 1.1636608839035034, "learning_rate": 9.558829081893836e-05, "loss": 0.4298, "step": 10680 }, { "epoch": 6.222351571594878, "grad_norm": 1.1652792692184448, "learning_rate": 9.55769656056209e-05, "loss": 0.4252, "step": 10690 }, { "epoch": 6.228172293364377, "grad_norm": 1.1140106916427612, "learning_rate": 9.556562654722469e-05, "loss": 0.4445, "step": 10700 }, { "epoch": 6.233993015133876, "grad_norm": 1.124434232711792, "learning_rate": 9.555427364719422e-05, "loss": 0.4408, "step": 10710 }, { "epoch": 6.239813736903376, "grad_norm": 1.0622814893722534, "learning_rate": 9.55429069089782e-05, "loss": 0.4242, "step": 10720 }, { "epoch": 6.245634458672876, "grad_norm": 1.1699867248535156, "learning_rate": 9.553152633602956e-05, "loss": 0.4238, "step": 10730 }, { "epoch": 6.251455180442375, "grad_norm": 1.1542474031448364, "learning_rate": 9.552013193180543e-05, "loss": 0.4335, "step": 10740 }, { "epoch": 6.257275902211874, "grad_norm": 1.16769540309906, "learning_rate": 9.550872369976707e-05, "loss": 0.4355, "step": 10750 }, { "epoch": 6.263096623981374, "grad_norm": 1.2809594869613647, "learning_rate": 9.549730164338007e-05, "loss": 0.4433, "step": 10760 }, { "epoch": 6.268917345750873, "grad_norm": 1.0805333852767944, "learning_rate": 9.548586576611408e-05, "loss": 0.4397, "step": 10770 }, { "epoch": 6.2747380675203726, "grad_norm": 0.9927615523338318, "learning_rate": 9.54744160714431e-05, "loss": 0.4328, "step": 10780 }, { "epoch": 6.280558789289872, "grad_norm": 1.168687105178833, "learning_rate": 9.546295256284516e-05, "loss": 0.4196, "step": 10790 }, { "epoch": 6.286379511059371, "grad_norm": 1.1638096570968628, "learning_rate": 9.545147524380265e-05, "loss": 0.4299, "step": 10800 }, { "epoch": 6.292200232828871, "grad_norm": 1.0638295412063599, "learning_rate": 9.543998411780201e-05, "loss": 0.425, "step": 10810 }, { "epoch": 6.29802095459837, "grad_norm": 1.0953001976013184, "learning_rate": 9.542847918833397e-05, "loss": 0.4322, "step": 10820 }, { "epoch": 6.3038416763678695, "grad_norm": 1.1604491472244263, "learning_rate": 9.541696045889343e-05, "loss": 0.4484, "step": 10830 }, { "epoch": 6.309662398137369, "grad_norm": 1.1640875339508057, "learning_rate": 9.540542793297947e-05, "loss": 0.4333, "step": 10840 }, { "epoch": 6.315483119906869, "grad_norm": 1.1085528135299683, "learning_rate": 9.539388161409537e-05, "loss": 0.4285, "step": 10850 }, { "epoch": 6.321303841676368, "grad_norm": 1.1325092315673828, "learning_rate": 9.538232150574857e-05, "loss": 0.4297, "step": 10860 }, { "epoch": 6.327124563445867, "grad_norm": 0.9773824214935303, "learning_rate": 9.537074761145076e-05, "loss": 0.4217, "step": 10870 }, { "epoch": 6.3329452852153665, "grad_norm": 1.163080096244812, "learning_rate": 9.535915993471778e-05, "loss": 0.4288, "step": 10880 }, { "epoch": 6.338766006984866, "grad_norm": 1.2986643314361572, "learning_rate": 9.534755847906964e-05, "loss": 0.425, "step": 10890 }, { "epoch": 6.344586728754366, "grad_norm": 1.1157046556472778, "learning_rate": 9.533594324803057e-05, "loss": 0.4257, "step": 10900 }, { "epoch": 6.350407450523865, "grad_norm": 1.2195558547973633, "learning_rate": 9.532431424512895e-05, "loss": 0.4348, "step": 10910 }, { "epoch": 6.356228172293364, "grad_norm": 1.1103503704071045, "learning_rate": 9.531267147389741e-05, "loss": 0.4163, "step": 10920 }, { "epoch": 6.3620488940628634, "grad_norm": 1.0580779314041138, "learning_rate": 9.530101493787266e-05, "loss": 0.4376, "step": 10930 }, { "epoch": 6.3678696158323636, "grad_norm": 1.3015564680099487, "learning_rate": 9.528934464059571e-05, "loss": 0.4154, "step": 10940 }, { "epoch": 6.373690337601863, "grad_norm": 1.1535801887512207, "learning_rate": 9.527766058561163e-05, "loss": 0.4247, "step": 10950 }, { "epoch": 6.379511059371362, "grad_norm": 1.194312334060669, "learning_rate": 9.526596277646976e-05, "loss": 0.4242, "step": 10960 }, { "epoch": 6.385331781140861, "grad_norm": 1.169007658958435, "learning_rate": 9.525425121672358e-05, "loss": 0.4329, "step": 10970 }, { "epoch": 6.391152502910361, "grad_norm": 1.2944763898849487, "learning_rate": 9.524252590993074e-05, "loss": 0.4393, "step": 10980 }, { "epoch": 6.3969732246798605, "grad_norm": 1.1932092905044556, "learning_rate": 9.523078685965309e-05, "loss": 0.4436, "step": 10990 }, { "epoch": 6.40279394644936, "grad_norm": 1.0643428564071655, "learning_rate": 9.521903406945664e-05, "loss": 0.4567, "step": 11000 }, { "epoch": 6.408614668218859, "grad_norm": 1.091320276260376, "learning_rate": 9.520726754291158e-05, "loss": 0.4281, "step": 11010 }, { "epoch": 6.414435389988358, "grad_norm": 1.2120248079299927, "learning_rate": 9.519548728359227e-05, "loss": 0.4321, "step": 11020 }, { "epoch": 6.420256111757858, "grad_norm": 1.1102219820022583, "learning_rate": 9.518369329507726e-05, "loss": 0.466, "step": 11030 }, { "epoch": 6.4260768335273575, "grad_norm": 1.0905125141143799, "learning_rate": 9.51718855809492e-05, "loss": 0.4188, "step": 11040 }, { "epoch": 6.431897555296857, "grad_norm": 1.057079792022705, "learning_rate": 9.516006414479502e-05, "loss": 0.4327, "step": 11050 }, { "epoch": 6.437718277066356, "grad_norm": 1.1189899444580078, "learning_rate": 9.514822899020572e-05, "loss": 0.4147, "step": 11060 }, { "epoch": 6.443538998835856, "grad_norm": 1.17814302444458, "learning_rate": 9.513638012077654e-05, "loss": 0.4421, "step": 11070 }, { "epoch": 6.449359720605355, "grad_norm": 1.1818537712097168, "learning_rate": 9.512451754010683e-05, "loss": 0.4296, "step": 11080 }, { "epoch": 6.455180442374854, "grad_norm": 1.1778303384780884, "learning_rate": 9.511264125180013e-05, "loss": 0.4256, "step": 11090 }, { "epoch": 6.461001164144354, "grad_norm": 1.1640859842300415, "learning_rate": 9.510075125946414e-05, "loss": 0.4414, "step": 11100 }, { "epoch": 6.466821885913854, "grad_norm": 1.0255920886993408, "learning_rate": 9.508884756671075e-05, "loss": 0.4431, "step": 11110 }, { "epoch": 6.472642607683353, "grad_norm": 1.0453333854675293, "learning_rate": 9.507693017715596e-05, "loss": 0.4419, "step": 11120 }, { "epoch": 6.478463329452852, "grad_norm": 1.0942779779434204, "learning_rate": 9.506499909441997e-05, "loss": 0.4238, "step": 11130 }, { "epoch": 6.484284051222351, "grad_norm": 1.0920370817184448, "learning_rate": 9.505305432212713e-05, "loss": 0.4417, "step": 11140 }, { "epoch": 6.490104772991851, "grad_norm": 1.1869456768035889, "learning_rate": 9.504109586390595e-05, "loss": 0.4374, "step": 11150 }, { "epoch": 6.495925494761351, "grad_norm": 1.0890856981277466, "learning_rate": 9.502912372338908e-05, "loss": 0.4382, "step": 11160 }, { "epoch": 6.50174621653085, "grad_norm": 1.133954644203186, "learning_rate": 9.501713790421335e-05, "loss": 0.4409, "step": 11170 }, { "epoch": 6.507566938300349, "grad_norm": 1.055872917175293, "learning_rate": 9.500513841001974e-05, "loss": 0.4161, "step": 11180 }, { "epoch": 6.513387660069848, "grad_norm": 1.0772318840026855, "learning_rate": 9.499312524445336e-05, "loss": 0.4407, "step": 11190 }, { "epoch": 6.5192083818393485, "grad_norm": 1.0758205652236938, "learning_rate": 9.498109841116351e-05, "loss": 0.4364, "step": 11200 }, { "epoch": 6.525029103608848, "grad_norm": 1.1146408319473267, "learning_rate": 9.496905791380363e-05, "loss": 0.4373, "step": 11210 }, { "epoch": 6.530849825378347, "grad_norm": 1.1203211545944214, "learning_rate": 9.495700375603129e-05, "loss": 0.4212, "step": 11220 }, { "epoch": 6.536670547147846, "grad_norm": 1.070094347000122, "learning_rate": 9.494493594150822e-05, "loss": 0.4388, "step": 11230 }, { "epoch": 6.542491268917345, "grad_norm": 1.1066251993179321, "learning_rate": 9.493285447390032e-05, "loss": 0.4196, "step": 11240 }, { "epoch": 6.548311990686845, "grad_norm": 1.0776093006134033, "learning_rate": 9.492075935687761e-05, "loss": 0.4439, "step": 11250 }, { "epoch": 6.554132712456345, "grad_norm": 1.1758688688278198, "learning_rate": 9.490865059411427e-05, "loss": 0.4153, "step": 11260 }, { "epoch": 6.559953434225844, "grad_norm": 1.207234263420105, "learning_rate": 9.489652818928863e-05, "loss": 0.4322, "step": 11270 }, { "epoch": 6.565774155995343, "grad_norm": 1.0448273420333862, "learning_rate": 9.488439214608315e-05, "loss": 0.4292, "step": 11280 }, { "epoch": 6.571594877764843, "grad_norm": 1.1425508260726929, "learning_rate": 9.487224246818444e-05, "loss": 0.4296, "step": 11290 }, { "epoch": 6.577415599534342, "grad_norm": 1.1052415370941162, "learning_rate": 9.486007915928325e-05, "loss": 0.4282, "step": 11300 }, { "epoch": 6.583236321303842, "grad_norm": 1.1361849308013916, "learning_rate": 9.484790222307448e-05, "loss": 0.4193, "step": 11310 }, { "epoch": 6.589057043073341, "grad_norm": 1.0850998163223267, "learning_rate": 9.483571166325716e-05, "loss": 0.4217, "step": 11320 }, { "epoch": 6.59487776484284, "grad_norm": 1.1107968091964722, "learning_rate": 9.482350748353444e-05, "loss": 0.4294, "step": 11330 }, { "epoch": 6.60069848661234, "grad_norm": 1.106458306312561, "learning_rate": 9.481128968761363e-05, "loss": 0.442, "step": 11340 }, { "epoch": 6.606519208381839, "grad_norm": 1.0412203073501587, "learning_rate": 9.479905827920621e-05, "loss": 0.4276, "step": 11350 }, { "epoch": 6.612339930151339, "grad_norm": 1.1850414276123047, "learning_rate": 9.478681326202773e-05, "loss": 0.42, "step": 11360 }, { "epoch": 6.618160651920838, "grad_norm": 1.0146780014038086, "learning_rate": 9.477455463979791e-05, "loss": 0.4422, "step": 11370 }, { "epoch": 6.623981373690338, "grad_norm": 1.0450239181518555, "learning_rate": 9.476228241624059e-05, "loss": 0.4095, "step": 11380 }, { "epoch": 6.629802095459837, "grad_norm": 1.0603525638580322, "learning_rate": 9.474999659508374e-05, "loss": 0.4149, "step": 11390 }, { "epoch": 6.635622817229336, "grad_norm": 1.1770741939544678, "learning_rate": 9.47376971800595e-05, "loss": 0.4231, "step": 11400 }, { "epoch": 6.6414435389988355, "grad_norm": 1.097055196762085, "learning_rate": 9.472538417490409e-05, "loss": 0.4245, "step": 11410 }, { "epoch": 6.647264260768336, "grad_norm": 1.0512722730636597, "learning_rate": 9.471305758335784e-05, "loss": 0.4326, "step": 11420 }, { "epoch": 6.653084982537835, "grad_norm": 1.1340248584747314, "learning_rate": 9.47007174091653e-05, "loss": 0.4329, "step": 11430 }, { "epoch": 6.658905704307334, "grad_norm": 1.1124579906463623, "learning_rate": 9.468836365607507e-05, "loss": 0.4232, "step": 11440 }, { "epoch": 6.664726426076833, "grad_norm": 1.0351412296295166, "learning_rate": 9.467599632783988e-05, "loss": 0.4275, "step": 11450 }, { "epoch": 6.670547147846333, "grad_norm": 1.0541598796844482, "learning_rate": 9.466361542821662e-05, "loss": 0.4476, "step": 11460 }, { "epoch": 6.676367869615833, "grad_norm": 1.1260101795196533, "learning_rate": 9.465122096096625e-05, "loss": 0.4324, "step": 11470 }, { "epoch": 6.682188591385332, "grad_norm": 1.0781068801879883, "learning_rate": 9.463881292985391e-05, "loss": 0.4288, "step": 11480 }, { "epoch": 6.688009313154831, "grad_norm": 1.0325977802276611, "learning_rate": 9.462639133864881e-05, "loss": 0.4186, "step": 11490 }, { "epoch": 6.69383003492433, "grad_norm": 1.1719515323638916, "learning_rate": 9.461395619112432e-05, "loss": 0.4326, "step": 11500 }, { "epoch": 6.69965075669383, "grad_norm": 1.1341193914413452, "learning_rate": 9.460150749105791e-05, "loss": 0.416, "step": 11510 }, { "epoch": 6.70547147846333, "grad_norm": 1.0681689977645874, "learning_rate": 9.458904524223116e-05, "loss": 0.4149, "step": 11520 }, { "epoch": 6.711292200232829, "grad_norm": 1.0980799198150635, "learning_rate": 9.457656944842976e-05, "loss": 0.4327, "step": 11530 }, { "epoch": 6.717112922002328, "grad_norm": 1.0611377954483032, "learning_rate": 9.456408011344353e-05, "loss": 0.4271, "step": 11540 }, { "epoch": 6.722933643771828, "grad_norm": 1.196511149406433, "learning_rate": 9.455157724106643e-05, "loss": 0.4329, "step": 11550 }, { "epoch": 6.728754365541327, "grad_norm": 1.036826252937317, "learning_rate": 9.453906083509647e-05, "loss": 0.4229, "step": 11560 }, { "epoch": 6.7345750873108265, "grad_norm": 1.0754517316818237, "learning_rate": 9.45265308993358e-05, "loss": 0.4272, "step": 11570 }, { "epoch": 6.740395809080326, "grad_norm": 1.2046021223068237, "learning_rate": 9.451398743759071e-05, "loss": 0.4108, "step": 11580 }, { "epoch": 6.746216530849825, "grad_norm": 1.107710838317871, "learning_rate": 9.450143045367156e-05, "loss": 0.4399, "step": 11590 }, { "epoch": 6.752037252619325, "grad_norm": 1.062930941581726, "learning_rate": 9.448885995139283e-05, "loss": 0.4229, "step": 11600 }, { "epoch": 6.757857974388824, "grad_norm": 1.0242741107940674, "learning_rate": 9.44762759345731e-05, "loss": 0.4239, "step": 11610 }, { "epoch": 6.7636786961583235, "grad_norm": 1.1527130603790283, "learning_rate": 9.446367840703509e-05, "loss": 0.424, "step": 11620 }, { "epoch": 6.769499417927823, "grad_norm": 1.0369648933410645, "learning_rate": 9.445106737260556e-05, "loss": 0.4118, "step": 11630 }, { "epoch": 6.775320139697323, "grad_norm": 1.0816560983657837, "learning_rate": 9.443844283511543e-05, "loss": 0.4243, "step": 11640 }, { "epoch": 6.781140861466822, "grad_norm": 1.1626890897750854, "learning_rate": 9.442580479839968e-05, "loss": 0.4195, "step": 11650 }, { "epoch": 6.786961583236321, "grad_norm": 1.1271620988845825, "learning_rate": 9.441315326629745e-05, "loss": 0.4161, "step": 11660 }, { "epoch": 6.7927823050058205, "grad_norm": 1.2551257610321045, "learning_rate": 9.44004882426519e-05, "loss": 0.4202, "step": 11670 }, { "epoch": 6.79860302677532, "grad_norm": 1.0512614250183105, "learning_rate": 9.438780973131037e-05, "loss": 0.4234, "step": 11680 }, { "epoch": 6.80442374854482, "grad_norm": 1.1225522756576538, "learning_rate": 9.437511773612423e-05, "loss": 0.4309, "step": 11690 }, { "epoch": 6.810244470314319, "grad_norm": 1.1074063777923584, "learning_rate": 9.436241226094896e-05, "loss": 0.4233, "step": 11700 }, { "epoch": 6.816065192083818, "grad_norm": 1.0148998498916626, "learning_rate": 9.434969330964418e-05, "loss": 0.439, "step": 11710 }, { "epoch": 6.821885913853317, "grad_norm": 1.1011461019515991, "learning_rate": 9.433696088607356e-05, "loss": 0.4155, "step": 11720 }, { "epoch": 6.8277066356228175, "grad_norm": 1.0215378999710083, "learning_rate": 9.432421499410486e-05, "loss": 0.4237, "step": 11730 }, { "epoch": 6.833527357392317, "grad_norm": 1.0764968395233154, "learning_rate": 9.431145563760998e-05, "loss": 0.4165, "step": 11740 }, { "epoch": 6.839348079161816, "grad_norm": 1.0311015844345093, "learning_rate": 9.429868282046484e-05, "loss": 0.4248, "step": 11750 }, { "epoch": 6.845168800931315, "grad_norm": 1.245847463607788, "learning_rate": 9.428589654654951e-05, "loss": 0.4184, "step": 11760 }, { "epoch": 6.850989522700814, "grad_norm": 1.0638270378112793, "learning_rate": 9.42730968197481e-05, "loss": 0.4136, "step": 11770 }, { "epoch": 6.8568102444703145, "grad_norm": 1.0847885608673096, "learning_rate": 9.426028364394883e-05, "loss": 0.4129, "step": 11780 }, { "epoch": 6.862630966239814, "grad_norm": 1.1459589004516602, "learning_rate": 9.424745702304402e-05, "loss": 0.4282, "step": 11790 }, { "epoch": 6.868451688009313, "grad_norm": 1.1197868585586548, "learning_rate": 9.423461696093006e-05, "loss": 0.4284, "step": 11800 }, { "epoch": 6.874272409778813, "grad_norm": 1.1056216955184937, "learning_rate": 9.422176346150741e-05, "loss": 0.4185, "step": 11810 }, { "epoch": 6.880093131548312, "grad_norm": 1.1117873191833496, "learning_rate": 9.420889652868063e-05, "loss": 0.4307, "step": 11820 }, { "epoch": 6.8859138533178115, "grad_norm": 1.1013410091400146, "learning_rate": 9.419601616635836e-05, "loss": 0.4354, "step": 11830 }, { "epoch": 6.891734575087311, "grad_norm": 1.2381882667541504, "learning_rate": 9.418312237845331e-05, "loss": 0.424, "step": 11840 }, { "epoch": 6.89755529685681, "grad_norm": 1.1059656143188477, "learning_rate": 9.417021516888225e-05, "loss": 0.4221, "step": 11850 }, { "epoch": 6.90337601862631, "grad_norm": 1.1584208011627197, "learning_rate": 9.415729454156608e-05, "loss": 0.4335, "step": 11860 }, { "epoch": 6.909196740395809, "grad_norm": 1.1902217864990234, "learning_rate": 9.414436050042973e-05, "loss": 0.4315, "step": 11870 }, { "epoch": 6.915017462165308, "grad_norm": 1.1598560810089111, "learning_rate": 9.413141304940223e-05, "loss": 0.4047, "step": 11880 }, { "epoch": 6.920838183934808, "grad_norm": 1.2127048969268799, "learning_rate": 9.411845219241666e-05, "loss": 0.4182, "step": 11890 }, { "epoch": 6.926658905704308, "grad_norm": 1.0939815044403076, "learning_rate": 9.410547793341021e-05, "loss": 0.432, "step": 11900 }, { "epoch": 6.932479627473807, "grad_norm": 1.1422399282455444, "learning_rate": 9.409249027632408e-05, "loss": 0.4242, "step": 11910 }, { "epoch": 6.938300349243306, "grad_norm": 1.0820708274841309, "learning_rate": 9.407948922510362e-05, "loss": 0.4171, "step": 11920 }, { "epoch": 6.944121071012805, "grad_norm": 1.1027655601501465, "learning_rate": 9.406647478369817e-05, "loss": 0.4142, "step": 11930 }, { "epoch": 6.949941792782305, "grad_norm": 1.0510139465332031, "learning_rate": 9.405344695606118e-05, "loss": 0.4303, "step": 11940 }, { "epoch": 6.955762514551805, "grad_norm": 1.0570377111434937, "learning_rate": 9.404040574615018e-05, "loss": 0.4207, "step": 11950 }, { "epoch": 6.961583236321304, "grad_norm": 0.9969397783279419, "learning_rate": 9.402735115792674e-05, "loss": 0.4153, "step": 11960 }, { "epoch": 6.967403958090803, "grad_norm": 1.2601842880249023, "learning_rate": 9.401428319535649e-05, "loss": 0.4195, "step": 11970 }, { "epoch": 6.973224679860302, "grad_norm": 1.2812349796295166, "learning_rate": 9.400120186240912e-05, "loss": 0.4177, "step": 11980 }, { "epoch": 6.9790454016298025, "grad_norm": 1.0860190391540527, "learning_rate": 9.398810716305844e-05, "loss": 0.4195, "step": 11990 }, { "epoch": 6.984866123399302, "grad_norm": 0.9889526963233948, "learning_rate": 9.397499910128222e-05, "loss": 0.421, "step": 12000 }, { "epoch": 6.990686845168801, "grad_norm": 1.0831211805343628, "learning_rate": 9.396187768106237e-05, "loss": 0.4244, "step": 12010 }, { "epoch": 6.9965075669383, "grad_norm": 1.0833789110183716, "learning_rate": 9.394874290638482e-05, "loss": 0.4232, "step": 12020 }, { "epoch": 7.002328288707799, "grad_norm": 1.0702120065689087, "learning_rate": 9.393559478123959e-05, "loss": 0.4308, "step": 12030 }, { "epoch": 7.008149010477299, "grad_norm": 1.1631996631622314, "learning_rate": 9.39224333096207e-05, "loss": 0.4215, "step": 12040 }, { "epoch": 7.013969732246799, "grad_norm": 1.2984942197799683, "learning_rate": 9.390925849552629e-05, "loss": 0.4368, "step": 12050 }, { "epoch": 7.019790454016298, "grad_norm": 1.1032174825668335, "learning_rate": 9.389607034295849e-05, "loss": 0.433, "step": 12060 }, { "epoch": 7.025611175785797, "grad_norm": 1.0518343448638916, "learning_rate": 9.388286885592355e-05, "loss": 0.4171, "step": 12070 }, { "epoch": 7.031431897555297, "grad_norm": 1.113012433052063, "learning_rate": 9.386965403843168e-05, "loss": 0.415, "step": 12080 }, { "epoch": 7.037252619324796, "grad_norm": 1.0390342473983765, "learning_rate": 9.385642589449726e-05, "loss": 0.41, "step": 12090 }, { "epoch": 7.043073341094296, "grad_norm": 1.0417174100875854, "learning_rate": 9.38431844281386e-05, "loss": 0.4287, "step": 12100 }, { "epoch": 7.048894062863795, "grad_norm": 1.1154087781906128, "learning_rate": 9.38299296433781e-05, "loss": 0.4173, "step": 12110 }, { "epoch": 7.054714784633295, "grad_norm": 1.0812678337097168, "learning_rate": 9.381666154424226e-05, "loss": 0.4276, "step": 12120 }, { "epoch": 7.060535506402794, "grad_norm": 1.0318371057510376, "learning_rate": 9.380338013476157e-05, "loss": 0.4176, "step": 12130 }, { "epoch": 7.066356228172293, "grad_norm": 0.9921733140945435, "learning_rate": 9.379008541897054e-05, "loss": 0.4189, "step": 12140 }, { "epoch": 7.072176949941793, "grad_norm": 1.172006368637085, "learning_rate": 9.377677740090777e-05, "loss": 0.4191, "step": 12150 }, { "epoch": 7.077997671711292, "grad_norm": 1.0760810375213623, "learning_rate": 9.376345608461588e-05, "loss": 0.4332, "step": 12160 }, { "epoch": 7.083818393480792, "grad_norm": 1.081378698348999, "learning_rate": 9.375012147414155e-05, "loss": 0.417, "step": 12170 }, { "epoch": 7.089639115250291, "grad_norm": 0.9981160163879395, "learning_rate": 9.373677357353545e-05, "loss": 0.4219, "step": 12180 }, { "epoch": 7.09545983701979, "grad_norm": 1.1465191841125488, "learning_rate": 9.372341238685237e-05, "loss": 0.446, "step": 12190 }, { "epoch": 7.1012805587892895, "grad_norm": 0.9542228579521179, "learning_rate": 9.371003791815102e-05, "loss": 0.4093, "step": 12200 }, { "epoch": 7.10710128055879, "grad_norm": 1.087906002998352, "learning_rate": 9.369665017149429e-05, "loss": 0.4252, "step": 12210 }, { "epoch": 7.112922002328289, "grad_norm": 1.0203044414520264, "learning_rate": 9.368324915094895e-05, "loss": 0.4243, "step": 12220 }, { "epoch": 7.118742724097788, "grad_norm": 1.1408214569091797, "learning_rate": 9.366983486058591e-05, "loss": 0.4379, "step": 12230 }, { "epoch": 7.124563445867287, "grad_norm": 1.0312377214431763, "learning_rate": 9.365640730448009e-05, "loss": 0.4204, "step": 12240 }, { "epoch": 7.130384167636787, "grad_norm": 1.0310145616531372, "learning_rate": 9.36429664867104e-05, "loss": 0.4206, "step": 12250 }, { "epoch": 7.136204889406287, "grad_norm": 1.2142966985702515, "learning_rate": 9.362951241135982e-05, "loss": 0.4275, "step": 12260 }, { "epoch": 7.142025611175786, "grad_norm": 1.1600873470306396, "learning_rate": 9.361604508251534e-05, "loss": 0.4269, "step": 12270 }, { "epoch": 7.147846332945285, "grad_norm": 0.9538432955741882, "learning_rate": 9.360256450426799e-05, "loss": 0.4311, "step": 12280 }, { "epoch": 7.153667054714784, "grad_norm": 1.123543620109558, "learning_rate": 9.358907068071279e-05, "loss": 0.4174, "step": 12290 }, { "epoch": 7.159487776484284, "grad_norm": 1.0078386068344116, "learning_rate": 9.357556361594882e-05, "loss": 0.4072, "step": 12300 }, { "epoch": 7.165308498253784, "grad_norm": 1.0555713176727295, "learning_rate": 9.356204331407917e-05, "loss": 0.4173, "step": 12310 }, { "epoch": 7.171129220023283, "grad_norm": 1.0537853240966797, "learning_rate": 9.354850977921094e-05, "loss": 0.4215, "step": 12320 }, { "epoch": 7.176949941792782, "grad_norm": 1.0913580656051636, "learning_rate": 9.353496301545529e-05, "loss": 0.4211, "step": 12330 }, { "epoch": 7.182770663562282, "grad_norm": 1.1140040159225464, "learning_rate": 9.352140302692733e-05, "loss": 0.4186, "step": 12340 }, { "epoch": 7.188591385331781, "grad_norm": 1.1183130741119385, "learning_rate": 9.350782981774627e-05, "loss": 0.4282, "step": 12350 }, { "epoch": 7.1944121071012805, "grad_norm": 1.1373964548110962, "learning_rate": 9.349424339203526e-05, "loss": 0.4089, "step": 12360 }, { "epoch": 7.20023282887078, "grad_norm": 1.122900366783142, "learning_rate": 9.34806437539215e-05, "loss": 0.4145, "step": 12370 }, { "epoch": 7.206053550640279, "grad_norm": 1.100482702255249, "learning_rate": 9.346703090753622e-05, "loss": 0.4331, "step": 12380 }, { "epoch": 7.211874272409779, "grad_norm": 0.9747170805931091, "learning_rate": 9.345340485701461e-05, "loss": 0.4103, "step": 12390 }, { "epoch": 7.217694994179278, "grad_norm": 1.0283538103103638, "learning_rate": 9.343976560649595e-05, "loss": 0.4333, "step": 12400 }, { "epoch": 7.2235157159487775, "grad_norm": 1.2306547164916992, "learning_rate": 9.342611316012344e-05, "loss": 0.4277, "step": 12410 }, { "epoch": 7.229336437718277, "grad_norm": 1.0849560499191284, "learning_rate": 9.341244752204437e-05, "loss": 0.4369, "step": 12420 }, { "epoch": 7.235157159487777, "grad_norm": 0.9939927458763123, "learning_rate": 9.339876869640995e-05, "loss": 0.4274, "step": 12430 }, { "epoch": 7.240977881257276, "grad_norm": 1.082054853439331, "learning_rate": 9.33850766873755e-05, "loss": 0.4146, "step": 12440 }, { "epoch": 7.246798603026775, "grad_norm": 1.0823357105255127, "learning_rate": 9.337137149910028e-05, "loss": 0.3999, "step": 12450 }, { "epoch": 7.2526193247962745, "grad_norm": 1.026820182800293, "learning_rate": 9.335765313574753e-05, "loss": 0.42, "step": 12460 }, { "epoch": 7.258440046565774, "grad_norm": 1.1621155738830566, "learning_rate": 9.334392160148457e-05, "loss": 0.4226, "step": 12470 }, { "epoch": 7.264260768335274, "grad_norm": 1.1778234243392944, "learning_rate": 9.333017690048264e-05, "loss": 0.4251, "step": 12480 }, { "epoch": 7.270081490104773, "grad_norm": 1.1120575666427612, "learning_rate": 9.331641903691706e-05, "loss": 0.418, "step": 12490 }, { "epoch": 7.275902211874272, "grad_norm": 1.036484956741333, "learning_rate": 9.330264801496707e-05, "loss": 0.4212, "step": 12500 }, { "epoch": 7.281722933643771, "grad_norm": 1.18160080909729, "learning_rate": 9.328886383881594e-05, "loss": 0.409, "step": 12510 }, { "epoch": 7.2875436554132715, "grad_norm": 1.0187621116638184, "learning_rate": 9.327506651265095e-05, "loss": 0.4115, "step": 12520 }, { "epoch": 7.293364377182771, "grad_norm": 1.062299132347107, "learning_rate": 9.326125604066338e-05, "loss": 0.4084, "step": 12530 }, { "epoch": 7.29918509895227, "grad_norm": 1.1724785566329956, "learning_rate": 9.324743242704847e-05, "loss": 0.4228, "step": 12540 }, { "epoch": 7.305005820721769, "grad_norm": 1.1290628910064697, "learning_rate": 9.323359567600546e-05, "loss": 0.4126, "step": 12550 }, { "epoch": 7.310826542491269, "grad_norm": 1.0068773031234741, "learning_rate": 9.321974579173761e-05, "loss": 0.4154, "step": 12560 }, { "epoch": 7.3166472642607685, "grad_norm": 1.0603946447372437, "learning_rate": 9.320588277845213e-05, "loss": 0.428, "step": 12570 }, { "epoch": 7.322467986030268, "grad_norm": 1.0753757953643799, "learning_rate": 9.319200664036026e-05, "loss": 0.4159, "step": 12580 }, { "epoch": 7.328288707799767, "grad_norm": 1.0627237558364868, "learning_rate": 9.31781173816772e-05, "loss": 0.4044, "step": 12590 }, { "epoch": 7.334109429569267, "grad_norm": 1.0124423503875732, "learning_rate": 9.316421500662212e-05, "loss": 0.4301, "step": 12600 }, { "epoch": 7.339930151338766, "grad_norm": 0.9962136745452881, "learning_rate": 9.31502995194182e-05, "loss": 0.4077, "step": 12610 }, { "epoch": 7.3457508731082655, "grad_norm": 1.0032124519348145, "learning_rate": 9.31363709242926e-05, "loss": 0.419, "step": 12620 }, { "epoch": 7.351571594877765, "grad_norm": 1.050986647605896, "learning_rate": 9.312242922547647e-05, "loss": 0.4163, "step": 12630 }, { "epoch": 7.357392316647264, "grad_norm": 1.0610908269882202, "learning_rate": 9.310847442720492e-05, "loss": 0.4188, "step": 12640 }, { "epoch": 7.363213038416764, "grad_norm": 0.9651063084602356, "learning_rate": 9.309450653371706e-05, "loss": 0.4068, "step": 12650 }, { "epoch": 7.369033760186263, "grad_norm": 1.1005926132202148, "learning_rate": 9.308052554925595e-05, "loss": 0.4206, "step": 12660 }, { "epoch": 7.374854481955762, "grad_norm": 1.1583558320999146, "learning_rate": 9.306653147806867e-05, "loss": 0.416, "step": 12670 }, { "epoch": 7.380675203725262, "grad_norm": 1.138871431350708, "learning_rate": 9.305252432440622e-05, "loss": 0.4066, "step": 12680 }, { "epoch": 7.386495925494762, "grad_norm": 1.151634931564331, "learning_rate": 9.303850409252361e-05, "loss": 0.4308, "step": 12690 }, { "epoch": 7.392316647264261, "grad_norm": 1.1589715480804443, "learning_rate": 9.302447078667985e-05, "loss": 0.4275, "step": 12700 }, { "epoch": 7.39813736903376, "grad_norm": 1.0658639669418335, "learning_rate": 9.301042441113783e-05, "loss": 0.4012, "step": 12710 }, { "epoch": 7.403958090803259, "grad_norm": 1.196771502494812, "learning_rate": 9.299636497016451e-05, "loss": 0.4225, "step": 12720 }, { "epoch": 7.409778812572759, "grad_norm": 1.0127888917922974, "learning_rate": 9.298229246803076e-05, "loss": 0.4188, "step": 12730 }, { "epoch": 7.415599534342259, "grad_norm": 1.197018027305603, "learning_rate": 9.296820690901144e-05, "loss": 0.4131, "step": 12740 }, { "epoch": 7.421420256111758, "grad_norm": 1.0120923519134521, "learning_rate": 9.295410829738539e-05, "loss": 0.4119, "step": 12750 }, { "epoch": 7.427240977881257, "grad_norm": 1.0240423679351807, "learning_rate": 9.293999663743535e-05, "loss": 0.4217, "step": 12760 }, { "epoch": 7.433061699650756, "grad_norm": 1.0795915126800537, "learning_rate": 9.292587193344813e-05, "loss": 0.4083, "step": 12770 }, { "epoch": 7.4388824214202565, "grad_norm": 1.0149883031845093, "learning_rate": 9.291173418971437e-05, "loss": 0.4267, "step": 12780 }, { "epoch": 7.444703143189756, "grad_norm": 1.1535934209823608, "learning_rate": 9.28975834105288e-05, "loss": 0.4191, "step": 12790 }, { "epoch": 7.450523864959255, "grad_norm": 1.0547106266021729, "learning_rate": 9.288341960019004e-05, "loss": 0.4193, "step": 12800 }, { "epoch": 7.456344586728754, "grad_norm": 1.2035844326019287, "learning_rate": 9.286924276300067e-05, "loss": 0.4341, "step": 12810 }, { "epoch": 7.462165308498253, "grad_norm": 1.1143449544906616, "learning_rate": 9.285505290326726e-05, "loss": 0.4266, "step": 12820 }, { "epoch": 7.467986030267753, "grad_norm": 1.085013747215271, "learning_rate": 9.284085002530027e-05, "loss": 0.415, "step": 12830 }, { "epoch": 7.473806752037253, "grad_norm": 1.1223331689834595, "learning_rate": 9.282663413341422e-05, "loss": 0.4185, "step": 12840 }, { "epoch": 7.479627473806752, "grad_norm": 1.130003571510315, "learning_rate": 9.281240523192747e-05, "loss": 0.4191, "step": 12850 }, { "epoch": 7.485448195576251, "grad_norm": 1.1886863708496094, "learning_rate": 9.279816332516242e-05, "loss": 0.4174, "step": 12860 }, { "epoch": 7.491268917345751, "grad_norm": 1.0865352153778076, "learning_rate": 9.278390841744536e-05, "loss": 0.4259, "step": 12870 }, { "epoch": 7.49708963911525, "grad_norm": 1.0709068775177002, "learning_rate": 9.276964051310658e-05, "loss": 0.423, "step": 12880 }, { "epoch": 7.50291036088475, "grad_norm": 1.050933599472046, "learning_rate": 9.275535961648027e-05, "loss": 0.4228, "step": 12890 }, { "epoch": 7.508731082654249, "grad_norm": 1.0036391019821167, "learning_rate": 9.274106573190459e-05, "loss": 0.4128, "step": 12900 }, { "epoch": 7.514551804423749, "grad_norm": 1.1013675928115845, "learning_rate": 9.272675886372168e-05, "loss": 0.4156, "step": 12910 }, { "epoch": 7.520372526193248, "grad_norm": 1.141252875328064, "learning_rate": 9.271243901627754e-05, "loss": 0.4318, "step": 12920 }, { "epoch": 7.526193247962747, "grad_norm": 1.0778552293777466, "learning_rate": 9.269810619392219e-05, "loss": 0.4156, "step": 12930 }, { "epoch": 7.532013969732247, "grad_norm": 1.004145860671997, "learning_rate": 9.268376040100955e-05, "loss": 0.3926, "step": 12940 }, { "epoch": 7.537834691501747, "grad_norm": 1.034703254699707, "learning_rate": 9.266940164189752e-05, "loss": 0.4119, "step": 12950 }, { "epoch": 7.543655413271246, "grad_norm": 1.011061429977417, "learning_rate": 9.265502992094787e-05, "loss": 0.405, "step": 12960 }, { "epoch": 7.549476135040745, "grad_norm": 1.0176446437835693, "learning_rate": 9.264064524252638e-05, "loss": 0.4173, "step": 12970 }, { "epoch": 7.555296856810244, "grad_norm": 1.120569109916687, "learning_rate": 9.262624761100271e-05, "loss": 0.4152, "step": 12980 }, { "epoch": 7.5611175785797435, "grad_norm": 1.086146593093872, "learning_rate": 9.261183703075051e-05, "loss": 0.4201, "step": 12990 }, { "epoch": 7.566938300349244, "grad_norm": 1.11162531375885, "learning_rate": 9.259741350614733e-05, "loss": 0.4183, "step": 13000 }, { "epoch": 7.572759022118743, "grad_norm": 1.0800210237503052, "learning_rate": 9.258297704157464e-05, "loss": 0.418, "step": 13010 }, { "epoch": 7.578579743888242, "grad_norm": 1.173542857170105, "learning_rate": 9.256852764141786e-05, "loss": 0.4148, "step": 13020 }, { "epoch": 7.584400465657741, "grad_norm": 1.0032734870910645, "learning_rate": 9.255406531006634e-05, "loss": 0.4239, "step": 13030 }, { "epoch": 7.590221187427241, "grad_norm": 1.0073604583740234, "learning_rate": 9.253959005191335e-05, "loss": 0.403, "step": 13040 }, { "epoch": 7.596041909196741, "grad_norm": 1.0406454801559448, "learning_rate": 9.25251018713561e-05, "loss": 0.4177, "step": 13050 }, { "epoch": 7.60186263096624, "grad_norm": 1.0072952508926392, "learning_rate": 9.251060077279571e-05, "loss": 0.4155, "step": 13060 }, { "epoch": 7.607683352735739, "grad_norm": 0.9957343339920044, "learning_rate": 9.249608676063724e-05, "loss": 0.3949, "step": 13070 }, { "epoch": 7.613504074505238, "grad_norm": 1.1802197694778442, "learning_rate": 9.248155983928964e-05, "loss": 0.4186, "step": 13080 }, { "epoch": 7.619324796274738, "grad_norm": 1.0219165086746216, "learning_rate": 9.246702001316583e-05, "loss": 0.4229, "step": 13090 }, { "epoch": 7.625145518044238, "grad_norm": 1.112596869468689, "learning_rate": 9.245246728668262e-05, "loss": 0.4261, "step": 13100 }, { "epoch": 7.630966239813737, "grad_norm": 1.0253640413284302, "learning_rate": 9.243790166426073e-05, "loss": 0.4178, "step": 13110 }, { "epoch": 7.636786961583236, "grad_norm": 1.3009555339813232, "learning_rate": 9.242332315032484e-05, "loss": 0.4267, "step": 13120 }, { "epoch": 7.642607683352736, "grad_norm": 1.0278857946395874, "learning_rate": 9.240873174930349e-05, "loss": 0.4194, "step": 13130 }, { "epoch": 7.648428405122235, "grad_norm": 1.0696063041687012, "learning_rate": 9.239412746562917e-05, "loss": 0.4159, "step": 13140 }, { "epoch": 7.6542491268917345, "grad_norm": 1.1235365867614746, "learning_rate": 9.237951030373828e-05, "loss": 0.4166, "step": 13150 }, { "epoch": 7.660069848661234, "grad_norm": 1.0558385848999023, "learning_rate": 9.236488026807113e-05, "loss": 0.3936, "step": 13160 }, { "epoch": 7.665890570430733, "grad_norm": 1.0931476354599, "learning_rate": 9.235023736307193e-05, "loss": 0.4215, "step": 13170 }, { "epoch": 7.671711292200233, "grad_norm": 0.9833325147628784, "learning_rate": 9.233558159318881e-05, "loss": 0.4153, "step": 13180 }, { "epoch": 7.677532013969732, "grad_norm": 1.1677206754684448, "learning_rate": 9.232091296287382e-05, "loss": 0.405, "step": 13190 }, { "epoch": 7.6833527357392315, "grad_norm": 1.0126110315322876, "learning_rate": 9.230623147658288e-05, "loss": 0.4327, "step": 13200 }, { "epoch": 7.689173457508731, "grad_norm": 1.0461345911026, "learning_rate": 9.229153713877586e-05, "loss": 0.4156, "step": 13210 }, { "epoch": 7.694994179278231, "grad_norm": 0.9591967463493347, "learning_rate": 9.227682995391649e-05, "loss": 0.4251, "step": 13220 }, { "epoch": 7.70081490104773, "grad_norm": 1.2463735342025757, "learning_rate": 9.226210992647243e-05, "loss": 0.4216, "step": 13230 }, { "epoch": 7.706635622817229, "grad_norm": 0.9419711232185364, "learning_rate": 9.224737706091525e-05, "loss": 0.406, "step": 13240 }, { "epoch": 7.7124563445867285, "grad_norm": 1.0155892372131348, "learning_rate": 9.223263136172039e-05, "loss": 0.4105, "step": 13250 }, { "epoch": 7.718277066356228, "grad_norm": 0.9565831422805786, "learning_rate": 9.22178728333672e-05, "loss": 0.4212, "step": 13260 }, { "epoch": 7.724097788125728, "grad_norm": 1.149019479751587, "learning_rate": 9.220310148033897e-05, "loss": 0.4327, "step": 13270 }, { "epoch": 7.729918509895227, "grad_norm": 0.9870067834854126, "learning_rate": 9.21883173071228e-05, "loss": 0.4255, "step": 13280 }, { "epoch": 7.735739231664726, "grad_norm": 1.01222562789917, "learning_rate": 9.217352031820976e-05, "loss": 0.4221, "step": 13290 }, { "epoch": 7.741559953434226, "grad_norm": 1.0934587717056274, "learning_rate": 9.215871051809477e-05, "loss": 0.4299, "step": 13300 }, { "epoch": 7.7473806752037255, "grad_norm": 0.9718924164772034, "learning_rate": 9.214388791127666e-05, "loss": 0.3927, "step": 13310 }, { "epoch": 7.753201396973225, "grad_norm": 1.1482000350952148, "learning_rate": 9.212905250225814e-05, "loss": 0.4107, "step": 13320 }, { "epoch": 7.759022118742724, "grad_norm": 1.1636489629745483, "learning_rate": 9.211420429554583e-05, "loss": 0.4297, "step": 13330 }, { "epoch": 7.764842840512223, "grad_norm": 1.0584750175476074, "learning_rate": 9.209934329565022e-05, "loss": 0.4227, "step": 13340 }, { "epoch": 7.770663562281723, "grad_norm": 1.0846753120422363, "learning_rate": 9.208446950708568e-05, "loss": 0.4081, "step": 13350 }, { "epoch": 7.7764842840512225, "grad_norm": 0.9745173454284668, "learning_rate": 9.20695829343705e-05, "loss": 0.4155, "step": 13360 }, { "epoch": 7.782305005820722, "grad_norm": 0.9698967933654785, "learning_rate": 9.205468358202678e-05, "loss": 0.4083, "step": 13370 }, { "epoch": 7.788125727590221, "grad_norm": 0.9951441884040833, "learning_rate": 9.203977145458059e-05, "loss": 0.4121, "step": 13380 }, { "epoch": 7.793946449359721, "grad_norm": 0.9731894135475159, "learning_rate": 9.202484655656182e-05, "loss": 0.4176, "step": 13390 }, { "epoch": 7.79976717112922, "grad_norm": 0.9557101726531982, "learning_rate": 9.200990889250427e-05, "loss": 0.4055, "step": 13400 }, { "epoch": 7.8055878928987195, "grad_norm": 1.0695878267288208, "learning_rate": 9.19949584669456e-05, "loss": 0.4183, "step": 13410 }, { "epoch": 7.811408614668219, "grad_norm": 1.0623635053634644, "learning_rate": 9.197999528442738e-05, "loss": 0.4197, "step": 13420 }, { "epoch": 7.817229336437718, "grad_norm": 1.0792028903961182, "learning_rate": 9.196501934949499e-05, "loss": 0.4224, "step": 13430 }, { "epoch": 7.823050058207218, "grad_norm": 0.9815205931663513, "learning_rate": 9.195003066669776e-05, "loss": 0.4188, "step": 13440 }, { "epoch": 7.828870779976717, "grad_norm": 0.9977779388427734, "learning_rate": 9.193502924058884e-05, "loss": 0.423, "step": 13450 }, { "epoch": 7.834691501746216, "grad_norm": 0.9881324768066406, "learning_rate": 9.192001507572526e-05, "loss": 0.4115, "step": 13460 }, { "epoch": 7.840512223515716, "grad_norm": 0.9898665547370911, "learning_rate": 9.190498817666793e-05, "loss": 0.4194, "step": 13470 }, { "epoch": 7.846332945285216, "grad_norm": 0.9282224774360657, "learning_rate": 9.188994854798163e-05, "loss": 0.3988, "step": 13480 }, { "epoch": 7.852153667054715, "grad_norm": 1.097061038017273, "learning_rate": 9.187489619423499e-05, "loss": 0.4273, "step": 13490 }, { "epoch": 7.857974388824214, "grad_norm": 0.9966172575950623, "learning_rate": 9.185983112000056e-05, "loss": 0.4115, "step": 13500 }, { "epoch": 7.863795110593713, "grad_norm": 1.0639724731445312, "learning_rate": 9.184475332985464e-05, "loss": 0.4313, "step": 13510 }, { "epoch": 7.869615832363213, "grad_norm": 1.0703684091567993, "learning_rate": 9.182966282837754e-05, "loss": 0.4025, "step": 13520 }, { "epoch": 7.875436554132713, "grad_norm": 1.1001510620117188, "learning_rate": 9.18145596201533e-05, "loss": 0.4173, "step": 13530 }, { "epoch": 7.881257275902212, "grad_norm": 0.9160954356193542, "learning_rate": 9.179944370976991e-05, "loss": 0.4222, "step": 13540 }, { "epoch": 7.887077997671711, "grad_norm": 1.0540014505386353, "learning_rate": 9.178431510181918e-05, "loss": 0.416, "step": 13550 }, { "epoch": 7.89289871944121, "grad_norm": 0.9434552788734436, "learning_rate": 9.176917380089675e-05, "loss": 0.4118, "step": 13560 }, { "epoch": 7.8987194412107105, "grad_norm": 1.0100722312927246, "learning_rate": 9.175401981160219e-05, "loss": 0.3998, "step": 13570 }, { "epoch": 7.90454016298021, "grad_norm": 1.064465880393982, "learning_rate": 9.173885313853885e-05, "loss": 0.4157, "step": 13580 }, { "epoch": 7.910360884749709, "grad_norm": 1.024595856666565, "learning_rate": 9.172367378631398e-05, "loss": 0.409, "step": 13590 }, { "epoch": 7.916181606519208, "grad_norm": 1.0334914922714233, "learning_rate": 9.170848175953866e-05, "loss": 0.4134, "step": 13600 }, { "epoch": 7.922002328288707, "grad_norm": 1.0912902355194092, "learning_rate": 9.169327706282784e-05, "loss": 0.4159, "step": 13610 }, { "epoch": 7.927823050058207, "grad_norm": 1.135477900505066, "learning_rate": 9.167805970080029e-05, "loss": 0.4246, "step": 13620 }, { "epoch": 7.933643771827707, "grad_norm": 1.1723566055297852, "learning_rate": 9.166282967807864e-05, "loss": 0.4169, "step": 13630 }, { "epoch": 7.939464493597206, "grad_norm": 1.0386918783187866, "learning_rate": 9.16475869992894e-05, "loss": 0.4138, "step": 13640 }, { "epoch": 7.945285215366706, "grad_norm": 0.9688805341720581, "learning_rate": 9.163233166906284e-05, "loss": 0.4109, "step": 13650 }, { "epoch": 7.951105937136205, "grad_norm": 1.0964045524597168, "learning_rate": 9.161706369203317e-05, "loss": 0.4166, "step": 13660 }, { "epoch": 7.956926658905704, "grad_norm": 1.0753111839294434, "learning_rate": 9.16017830728384e-05, "loss": 0.4011, "step": 13670 }, { "epoch": 7.962747380675204, "grad_norm": 1.0757765769958496, "learning_rate": 9.158648981612035e-05, "loss": 0.4191, "step": 13680 }, { "epoch": 7.968568102444703, "grad_norm": 1.1689661741256714, "learning_rate": 9.157118392652472e-05, "loss": 0.4091, "step": 13690 }, { "epoch": 7.974388824214203, "grad_norm": 0.9531298279762268, "learning_rate": 9.155586540870104e-05, "loss": 0.4172, "step": 13700 }, { "epoch": 7.980209545983702, "grad_norm": 1.0935940742492676, "learning_rate": 9.154053426730267e-05, "loss": 0.4338, "step": 13710 }, { "epoch": 7.986030267753201, "grad_norm": 1.175133466720581, "learning_rate": 9.15251905069868e-05, "loss": 0.4094, "step": 13720 }, { "epoch": 7.991850989522701, "grad_norm": 1.0658917427062988, "learning_rate": 9.150983413241446e-05, "loss": 0.4216, "step": 13730 }, { "epoch": 7.997671711292201, "grad_norm": 1.0328277349472046, "learning_rate": 9.149446514825051e-05, "loss": 0.4136, "step": 13740 }, { "epoch": 8.0034924330617, "grad_norm": 1.1568903923034668, "learning_rate": 9.147908355916365e-05, "loss": 0.4169, "step": 13750 }, { "epoch": 8.009313154831199, "grad_norm": 1.0283167362213135, "learning_rate": 9.146368936982642e-05, "loss": 0.4065, "step": 13760 }, { "epoch": 8.015133876600698, "grad_norm": 1.1364847421646118, "learning_rate": 9.144828258491511e-05, "loss": 0.4137, "step": 13770 }, { "epoch": 8.020954598370198, "grad_norm": 1.2190624475479126, "learning_rate": 9.143286320910996e-05, "loss": 0.4064, "step": 13780 }, { "epoch": 8.026775320139697, "grad_norm": 1.1609857082366943, "learning_rate": 9.141743124709491e-05, "loss": 0.4186, "step": 13790 }, { "epoch": 8.032596041909196, "grad_norm": 0.9524860978126526, "learning_rate": 9.140198670355784e-05, "loss": 0.4212, "step": 13800 }, { "epoch": 8.038416763678697, "grad_norm": 0.9807899594306946, "learning_rate": 9.138652958319034e-05, "loss": 0.4332, "step": 13810 }, { "epoch": 8.044237485448196, "grad_norm": 1.0859240293502808, "learning_rate": 9.137105989068791e-05, "loss": 0.4139, "step": 13820 }, { "epoch": 8.050058207217695, "grad_norm": 1.0412572622299194, "learning_rate": 9.135557763074983e-05, "loss": 0.4128, "step": 13830 }, { "epoch": 8.055878928987195, "grad_norm": 0.990929365158081, "learning_rate": 9.13400828080792e-05, "loss": 0.4167, "step": 13840 }, { "epoch": 8.061699650756694, "grad_norm": 1.043331265449524, "learning_rate": 9.132457542738292e-05, "loss": 0.4029, "step": 13850 }, { "epoch": 8.067520372526193, "grad_norm": 1.0212920904159546, "learning_rate": 9.130905549337174e-05, "loss": 0.4166, "step": 13860 }, { "epoch": 8.073341094295692, "grad_norm": 1.1034218072891235, "learning_rate": 9.129352301076021e-05, "loss": 0.4023, "step": 13870 }, { "epoch": 8.079161816065191, "grad_norm": 1.1403546333312988, "learning_rate": 9.127797798426668e-05, "loss": 0.4049, "step": 13880 }, { "epoch": 8.08498253783469, "grad_norm": 1.0253198146820068, "learning_rate": 9.126242041861333e-05, "loss": 0.4272, "step": 13890 }, { "epoch": 8.090803259604192, "grad_norm": 0.9663296341896057, "learning_rate": 9.124685031852611e-05, "loss": 0.4177, "step": 13900 }, { "epoch": 8.09662398137369, "grad_norm": 1.0601532459259033, "learning_rate": 9.123126768873482e-05, "loss": 0.4027, "step": 13910 }, { "epoch": 8.10244470314319, "grad_norm": 1.129341721534729, "learning_rate": 9.121567253397308e-05, "loss": 0.4246, "step": 13920 }, { "epoch": 8.10826542491269, "grad_norm": 0.9593374133110046, "learning_rate": 9.120006485897824e-05, "loss": 0.3998, "step": 13930 }, { "epoch": 8.114086146682189, "grad_norm": 1.195024847984314, "learning_rate": 9.118444466849152e-05, "loss": 0.4123, "step": 13940 }, { "epoch": 8.119906868451688, "grad_norm": 1.1759412288665771, "learning_rate": 9.116881196725793e-05, "loss": 0.4274, "step": 13950 }, { "epoch": 8.125727590221187, "grad_norm": 0.996015727519989, "learning_rate": 9.115316676002627e-05, "loss": 0.4138, "step": 13960 }, { "epoch": 8.131548311990686, "grad_norm": 1.0961450338363647, "learning_rate": 9.113750905154911e-05, "loss": 0.4074, "step": 13970 }, { "epoch": 8.137369033760187, "grad_norm": 0.9858422875404358, "learning_rate": 9.112183884658289e-05, "loss": 0.4096, "step": 13980 }, { "epoch": 8.143189755529686, "grad_norm": 1.0299017429351807, "learning_rate": 9.11061561498878e-05, "loss": 0.4129, "step": 13990 }, { "epoch": 8.149010477299186, "grad_norm": 1.0018177032470703, "learning_rate": 9.109046096622779e-05, "loss": 0.3967, "step": 14000 }, { "epoch": 8.154831199068685, "grad_norm": 1.0082489252090454, "learning_rate": 9.107475330037069e-05, "loss": 0.4128, "step": 14010 }, { "epoch": 8.160651920838184, "grad_norm": 0.9863461852073669, "learning_rate": 9.105903315708806e-05, "loss": 0.4155, "step": 14020 }, { "epoch": 8.166472642607683, "grad_norm": 1.0464304685592651, "learning_rate": 9.104330054115524e-05, "loss": 0.4161, "step": 14030 }, { "epoch": 8.172293364377182, "grad_norm": 0.9579399824142456, "learning_rate": 9.102755545735141e-05, "loss": 0.4128, "step": 14040 }, { "epoch": 8.178114086146682, "grad_norm": 1.0568180084228516, "learning_rate": 9.10117979104595e-05, "loss": 0.4019, "step": 14050 }, { "epoch": 8.18393480791618, "grad_norm": 1.1884466409683228, "learning_rate": 9.099602790526624e-05, "loss": 0.415, "step": 14060 }, { "epoch": 8.189755529685682, "grad_norm": 1.0618369579315186, "learning_rate": 9.098024544656212e-05, "loss": 0.4217, "step": 14070 }, { "epoch": 8.195576251455181, "grad_norm": 1.0375176668167114, "learning_rate": 9.096445053914148e-05, "loss": 0.4073, "step": 14080 }, { "epoch": 8.20139697322468, "grad_norm": 0.9826140999794006, "learning_rate": 9.094864318780236e-05, "loss": 0.4067, "step": 14090 }, { "epoch": 8.20721769499418, "grad_norm": 1.0318206548690796, "learning_rate": 9.093282339734663e-05, "loss": 0.4105, "step": 14100 }, { "epoch": 8.213038416763679, "grad_norm": 1.1102944612503052, "learning_rate": 9.091699117257992e-05, "loss": 0.3971, "step": 14110 }, { "epoch": 8.218859138533178, "grad_norm": 1.0034661293029785, "learning_rate": 9.090114651831163e-05, "loss": 0.4157, "step": 14120 }, { "epoch": 8.224679860302677, "grad_norm": 1.1608866453170776, "learning_rate": 9.088528943935497e-05, "loss": 0.4215, "step": 14130 }, { "epoch": 8.230500582072176, "grad_norm": 1.111953854560852, "learning_rate": 9.086941994052689e-05, "loss": 0.4246, "step": 14140 }, { "epoch": 8.236321303841676, "grad_norm": 1.049068570137024, "learning_rate": 9.085353802664813e-05, "loss": 0.4103, "step": 14150 }, { "epoch": 8.242142025611177, "grad_norm": 1.013900876045227, "learning_rate": 9.08376437025432e-05, "loss": 0.3935, "step": 14160 }, { "epoch": 8.247962747380676, "grad_norm": 1.0602614879608154, "learning_rate": 9.082173697304035e-05, "loss": 0.3971, "step": 14170 }, { "epoch": 8.253783469150175, "grad_norm": 1.0595523118972778, "learning_rate": 9.080581784297166e-05, "loss": 0.4022, "step": 14180 }, { "epoch": 8.259604190919674, "grad_norm": 0.9380316734313965, "learning_rate": 9.078988631717291e-05, "loss": 0.4132, "step": 14190 }, { "epoch": 8.265424912689173, "grad_norm": 1.0101593732833862, "learning_rate": 9.077394240048369e-05, "loss": 0.4079, "step": 14200 }, { "epoch": 8.271245634458673, "grad_norm": 1.2302417755126953, "learning_rate": 9.075798609774736e-05, "loss": 0.4158, "step": 14210 }, { "epoch": 8.277066356228172, "grad_norm": 1.0067874193191528, "learning_rate": 9.0742017413811e-05, "loss": 0.4087, "step": 14220 }, { "epoch": 8.282887077997671, "grad_norm": 1.0359159708023071, "learning_rate": 9.072603635352548e-05, "loss": 0.4195, "step": 14230 }, { "epoch": 8.28870779976717, "grad_norm": 1.0380821228027344, "learning_rate": 9.071004292174541e-05, "loss": 0.4027, "step": 14240 }, { "epoch": 8.294528521536671, "grad_norm": 0.9736270308494568, "learning_rate": 9.06940371233292e-05, "loss": 0.4269, "step": 14250 }, { "epoch": 8.30034924330617, "grad_norm": 1.0730109214782715, "learning_rate": 9.067801896313898e-05, "loss": 0.4083, "step": 14260 }, { "epoch": 8.30616996507567, "grad_norm": 1.0114364624023438, "learning_rate": 9.066198844604064e-05, "loss": 0.4104, "step": 14270 }, { "epoch": 8.311990686845169, "grad_norm": 1.116198182106018, "learning_rate": 9.06459455769038e-05, "loss": 0.4014, "step": 14280 }, { "epoch": 8.317811408614668, "grad_norm": 1.0016919374465942, "learning_rate": 9.062989036060193e-05, "loss": 0.3834, "step": 14290 }, { "epoch": 8.323632130384167, "grad_norm": 1.0113623142242432, "learning_rate": 9.061382280201212e-05, "loss": 0.399, "step": 14300 }, { "epoch": 8.329452852153667, "grad_norm": 1.0704573392868042, "learning_rate": 9.059774290601528e-05, "loss": 0.3934, "step": 14310 }, { "epoch": 8.335273573923166, "grad_norm": 1.0522905588150024, "learning_rate": 9.058165067749606e-05, "loss": 0.4223, "step": 14320 }, { "epoch": 8.341094295692667, "grad_norm": 1.0716668367385864, "learning_rate": 9.056554612134288e-05, "loss": 0.4143, "step": 14330 }, { "epoch": 8.346915017462166, "grad_norm": 1.0469290018081665, "learning_rate": 9.054942924244785e-05, "loss": 0.4183, "step": 14340 }, { "epoch": 8.352735739231665, "grad_norm": 0.9919982552528381, "learning_rate": 9.053330004570686e-05, "loss": 0.4057, "step": 14350 }, { "epoch": 8.358556461001164, "grad_norm": 0.972710907459259, "learning_rate": 9.051715853601955e-05, "loss": 0.4132, "step": 14360 }, { "epoch": 8.364377182770664, "grad_norm": 1.0450752973556519, "learning_rate": 9.050100471828926e-05, "loss": 0.4055, "step": 14370 }, { "epoch": 8.370197904540163, "grad_norm": 0.9921730160713196, "learning_rate": 9.048483859742311e-05, "loss": 0.408, "step": 14380 }, { "epoch": 8.376018626309662, "grad_norm": 0.9929628968238831, "learning_rate": 9.046866017833193e-05, "loss": 0.4017, "step": 14390 }, { "epoch": 8.381839348079161, "grad_norm": 0.9519934058189392, "learning_rate": 9.045246946593029e-05, "loss": 0.409, "step": 14400 }, { "epoch": 8.38766006984866, "grad_norm": 1.025685429573059, "learning_rate": 9.043626646513652e-05, "loss": 0.4224, "step": 14410 }, { "epoch": 8.39348079161816, "grad_norm": 0.9756312370300293, "learning_rate": 9.042005118087267e-05, "loss": 0.4046, "step": 14420 }, { "epoch": 8.39930151338766, "grad_norm": 1.0838712453842163, "learning_rate": 9.040382361806448e-05, "loss": 0.3971, "step": 14430 }, { "epoch": 8.40512223515716, "grad_norm": 0.9824561476707458, "learning_rate": 9.038758378164148e-05, "loss": 0.4042, "step": 14440 }, { "epoch": 8.41094295692666, "grad_norm": 1.0384674072265625, "learning_rate": 9.037133167653691e-05, "loss": 0.413, "step": 14450 }, { "epoch": 8.416763678696158, "grad_norm": 1.0536165237426758, "learning_rate": 9.035506730768771e-05, "loss": 0.4138, "step": 14460 }, { "epoch": 8.422584400465658, "grad_norm": 1.0365841388702393, "learning_rate": 9.033879068003458e-05, "loss": 0.4156, "step": 14470 }, { "epoch": 8.428405122235157, "grad_norm": 1.0599855184555054, "learning_rate": 9.032250179852193e-05, "loss": 0.4165, "step": 14480 }, { "epoch": 8.434225844004656, "grad_norm": 1.1396762132644653, "learning_rate": 9.030620066809787e-05, "loss": 0.4144, "step": 14490 }, { "epoch": 8.440046565774155, "grad_norm": 1.0368608236312866, "learning_rate": 9.028988729371428e-05, "loss": 0.3909, "step": 14500 }, { "epoch": 8.445867287543656, "grad_norm": 1.052016019821167, "learning_rate": 9.027356168032673e-05, "loss": 0.4041, "step": 14510 }, { "epoch": 8.451688009313155, "grad_norm": 1.0533912181854248, "learning_rate": 9.02572238328945e-05, "loss": 0.397, "step": 14520 }, { "epoch": 8.457508731082655, "grad_norm": 0.9893972873687744, "learning_rate": 9.02408737563806e-05, "loss": 0.396, "step": 14530 }, { "epoch": 8.463329452852154, "grad_norm": 1.0210684537887573, "learning_rate": 9.022451145575174e-05, "loss": 0.4075, "step": 14540 }, { "epoch": 8.469150174621653, "grad_norm": 1.0764586925506592, "learning_rate": 9.02081369359784e-05, "loss": 0.4163, "step": 14550 }, { "epoch": 8.474970896391152, "grad_norm": 1.08277428150177, "learning_rate": 9.019175020203465e-05, "loss": 0.4209, "step": 14560 }, { "epoch": 8.480791618160652, "grad_norm": 1.178607702255249, "learning_rate": 9.017535125889842e-05, "loss": 0.4068, "step": 14570 }, { "epoch": 8.48661233993015, "grad_norm": 0.9546521306037903, "learning_rate": 9.015894011155124e-05, "loss": 0.409, "step": 14580 }, { "epoch": 8.49243306169965, "grad_norm": 1.0339173078536987, "learning_rate": 9.014251676497838e-05, "loss": 0.4002, "step": 14590 }, { "epoch": 8.498253783469151, "grad_norm": 0.9843594431877136, "learning_rate": 9.012608122416884e-05, "loss": 0.4068, "step": 14600 }, { "epoch": 8.50407450523865, "grad_norm": 0.9695240259170532, "learning_rate": 9.010963349411529e-05, "loss": 0.414, "step": 14610 }, { "epoch": 8.50989522700815, "grad_norm": 1.0095850229263306, "learning_rate": 9.00931735798141e-05, "loss": 0.4032, "step": 14620 }, { "epoch": 8.515715948777649, "grad_norm": 1.0794265270233154, "learning_rate": 9.00767014862654e-05, "loss": 0.4294, "step": 14630 }, { "epoch": 8.521536670547148, "grad_norm": 1.0555039644241333, "learning_rate": 9.006021721847295e-05, "loss": 0.4172, "step": 14640 }, { "epoch": 8.527357392316647, "grad_norm": 1.0006349086761475, "learning_rate": 9.004372078144423e-05, "loss": 0.4038, "step": 14650 }, { "epoch": 8.533178114086146, "grad_norm": 1.0117309093475342, "learning_rate": 9.002721218019043e-05, "loss": 0.4005, "step": 14660 }, { "epoch": 8.538998835855645, "grad_norm": 1.0060842037200928, "learning_rate": 9.001069141972642e-05, "loss": 0.3963, "step": 14670 }, { "epoch": 8.544819557625146, "grad_norm": 0.960910439491272, "learning_rate": 8.99941585050708e-05, "loss": 0.4068, "step": 14680 }, { "epoch": 8.550640279394646, "grad_norm": 1.025824785232544, "learning_rate": 8.997761344124578e-05, "loss": 0.3942, "step": 14690 }, { "epoch": 8.556461001164145, "grad_norm": 0.9592751264572144, "learning_rate": 8.996105623327737e-05, "loss": 0.4069, "step": 14700 }, { "epoch": 8.562281722933644, "grad_norm": 0.9722949266433716, "learning_rate": 8.994448688619517e-05, "loss": 0.4215, "step": 14710 }, { "epoch": 8.568102444703143, "grad_norm": 1.1254675388336182, "learning_rate": 8.992790540503253e-05, "loss": 0.4232, "step": 14720 }, { "epoch": 8.573923166472643, "grad_norm": 1.1505780220031738, "learning_rate": 8.991131179482648e-05, "loss": 0.4181, "step": 14730 }, { "epoch": 8.579743888242142, "grad_norm": 1.0071805715560913, "learning_rate": 8.989470606061768e-05, "loss": 0.4299, "step": 14740 }, { "epoch": 8.585564610011641, "grad_norm": 1.1228402853012085, "learning_rate": 8.987808820745056e-05, "loss": 0.4163, "step": 14750 }, { "epoch": 8.59138533178114, "grad_norm": 0.9309464693069458, "learning_rate": 8.986145824037315e-05, "loss": 0.4094, "step": 14760 }, { "epoch": 8.59720605355064, "grad_norm": 1.0718178749084473, "learning_rate": 8.984481616443721e-05, "loss": 0.426, "step": 14770 }, { "epoch": 8.60302677532014, "grad_norm": 1.0228396654129028, "learning_rate": 8.982816198469815e-05, "loss": 0.4075, "step": 14780 }, { "epoch": 8.60884749708964, "grad_norm": 0.9703723192214966, "learning_rate": 8.98114957062151e-05, "loss": 0.4091, "step": 14790 }, { "epoch": 8.614668218859139, "grad_norm": 1.0707064867019653, "learning_rate": 8.97948173340508e-05, "loss": 0.3842, "step": 14800 }, { "epoch": 8.620488940628638, "grad_norm": 1.0535907745361328, "learning_rate": 8.977812687327172e-05, "loss": 0.4022, "step": 14810 }, { "epoch": 8.626309662398137, "grad_norm": 1.0988256931304932, "learning_rate": 8.976142432894798e-05, "loss": 0.4077, "step": 14820 }, { "epoch": 8.632130384167636, "grad_norm": 1.1002925634384155, "learning_rate": 8.974470970615336e-05, "loss": 0.4175, "step": 14830 }, { "epoch": 8.637951105937136, "grad_norm": 1.051554799079895, "learning_rate": 8.972798300996534e-05, "loss": 0.3949, "step": 14840 }, { "epoch": 8.643771827706635, "grad_norm": 0.9977623820304871, "learning_rate": 8.971124424546504e-05, "loss": 0.3907, "step": 14850 }, { "epoch": 8.649592549476136, "grad_norm": 0.9544128179550171, "learning_rate": 8.969449341773724e-05, "loss": 0.4058, "step": 14860 }, { "epoch": 8.655413271245635, "grad_norm": 1.1533058881759644, "learning_rate": 8.967773053187042e-05, "loss": 0.4133, "step": 14870 }, { "epoch": 8.661233993015134, "grad_norm": 1.0667170286178589, "learning_rate": 8.966095559295668e-05, "loss": 0.4031, "step": 14880 }, { "epoch": 8.667054714784634, "grad_norm": 0.9156146049499512, "learning_rate": 8.964416860609184e-05, "loss": 0.3989, "step": 14890 }, { "epoch": 8.672875436554133, "grad_norm": 0.9319448471069336, "learning_rate": 8.962736957637532e-05, "loss": 0.3892, "step": 14900 }, { "epoch": 8.678696158323632, "grad_norm": 1.1412712335586548, "learning_rate": 8.96105585089102e-05, "loss": 0.4037, "step": 14910 }, { "epoch": 8.684516880093131, "grad_norm": 1.1103278398513794, "learning_rate": 8.959373540880329e-05, "loss": 0.4327, "step": 14920 }, { "epoch": 8.69033760186263, "grad_norm": 0.9844107031822205, "learning_rate": 8.957690028116495e-05, "loss": 0.3882, "step": 14930 }, { "epoch": 8.69615832363213, "grad_norm": 1.1058430671691895, "learning_rate": 8.956005313110928e-05, "loss": 0.4164, "step": 14940 }, { "epoch": 8.70197904540163, "grad_norm": 1.118375301361084, "learning_rate": 8.9543193963754e-05, "loss": 0.4034, "step": 14950 }, { "epoch": 8.70779976717113, "grad_norm": 1.0156724452972412, "learning_rate": 8.952632278422048e-05, "loss": 0.3905, "step": 14960 }, { "epoch": 8.713620488940629, "grad_norm": 1.0318684577941895, "learning_rate": 8.95094395976337e-05, "loss": 0.3956, "step": 14970 }, { "epoch": 8.719441210710128, "grad_norm": 0.9942728281021118, "learning_rate": 8.949254440912239e-05, "loss": 0.395, "step": 14980 }, { "epoch": 8.725261932479627, "grad_norm": 1.1245758533477783, "learning_rate": 8.94756372238188e-05, "loss": 0.4096, "step": 14990 }, { "epoch": 8.731082654249127, "grad_norm": 1.0414669513702393, "learning_rate": 8.945871804685892e-05, "loss": 0.4052, "step": 15000 }, { "epoch": 8.736903376018626, "grad_norm": 0.9166380763053894, "learning_rate": 8.944178688338236e-05, "loss": 0.3953, "step": 15010 }, { "epoch": 8.742724097788125, "grad_norm": 0.9768570065498352, "learning_rate": 8.942484373853233e-05, "loss": 0.3954, "step": 15020 }, { "epoch": 8.748544819557626, "grad_norm": 1.0154985189437866, "learning_rate": 8.940788861745572e-05, "loss": 0.4132, "step": 15030 }, { "epoch": 8.754365541327125, "grad_norm": 1.05706787109375, "learning_rate": 8.939092152530308e-05, "loss": 0.3927, "step": 15040 }, { "epoch": 8.760186263096625, "grad_norm": 1.0034314393997192, "learning_rate": 8.937394246722853e-05, "loss": 0.4095, "step": 15050 }, { "epoch": 8.766006984866124, "grad_norm": 0.9793813228607178, "learning_rate": 8.935695144838984e-05, "loss": 0.388, "step": 15060 }, { "epoch": 8.771827706635623, "grad_norm": 0.9863178730010986, "learning_rate": 8.933994847394849e-05, "loss": 0.4097, "step": 15070 }, { "epoch": 8.777648428405122, "grad_norm": 1.0555667877197266, "learning_rate": 8.932293354906949e-05, "loss": 0.3957, "step": 15080 }, { "epoch": 8.783469150174621, "grad_norm": 0.9283832311630249, "learning_rate": 8.930590667892153e-05, "loss": 0.3892, "step": 15090 }, { "epoch": 8.78928987194412, "grad_norm": 1.0447211265563965, "learning_rate": 8.928886786867696e-05, "loss": 0.4069, "step": 15100 }, { "epoch": 8.79511059371362, "grad_norm": 1.0646406412124634, "learning_rate": 8.927181712351168e-05, "loss": 0.4169, "step": 15110 }, { "epoch": 8.800931315483119, "grad_norm": 1.0350261926651, "learning_rate": 8.925475444860527e-05, "loss": 0.4066, "step": 15120 }, { "epoch": 8.80675203725262, "grad_norm": 1.0198256969451904, "learning_rate": 8.923767984914092e-05, "loss": 0.4028, "step": 15130 }, { "epoch": 8.81257275902212, "grad_norm": 0.9399476051330566, "learning_rate": 8.922059333030545e-05, "loss": 0.3882, "step": 15140 }, { "epoch": 8.818393480791618, "grad_norm": 0.9916517734527588, "learning_rate": 8.920349489728928e-05, "loss": 0.4015, "step": 15150 }, { "epoch": 8.824214202561118, "grad_norm": 1.0767563581466675, "learning_rate": 8.918638455528646e-05, "loss": 0.3986, "step": 15160 }, { "epoch": 8.830034924330617, "grad_norm": 1.1149097681045532, "learning_rate": 8.916926230949468e-05, "loss": 0.3875, "step": 15170 }, { "epoch": 8.835855646100116, "grad_norm": 1.054743766784668, "learning_rate": 8.915212816511522e-05, "loss": 0.3953, "step": 15180 }, { "epoch": 8.841676367869615, "grad_norm": 0.9338681697845459, "learning_rate": 8.913498212735296e-05, "loss": 0.409, "step": 15190 }, { "epoch": 8.847497089639115, "grad_norm": 0.9890562295913696, "learning_rate": 8.911782420141643e-05, "loss": 0.4046, "step": 15200 }, { "epoch": 8.853317811408616, "grad_norm": 0.9650630354881287, "learning_rate": 8.910065439251775e-05, "loss": 0.3905, "step": 15210 }, { "epoch": 8.859138533178115, "grad_norm": 0.9943445920944214, "learning_rate": 8.908347270587268e-05, "loss": 0.3969, "step": 15220 }, { "epoch": 8.864959254947614, "grad_norm": 1.054731011390686, "learning_rate": 8.906627914670054e-05, "loss": 0.4001, "step": 15230 }, { "epoch": 8.870779976717113, "grad_norm": 1.0063526630401611, "learning_rate": 8.904907372022427e-05, "loss": 0.4131, "step": 15240 }, { "epoch": 8.876600698486612, "grad_norm": 1.068113088607788, "learning_rate": 8.903185643167042e-05, "loss": 0.4088, "step": 15250 }, { "epoch": 8.882421420256112, "grad_norm": 1.0032548904418945, "learning_rate": 8.901462728626919e-05, "loss": 0.4042, "step": 15260 }, { "epoch": 8.88824214202561, "grad_norm": 1.0737358331680298, "learning_rate": 8.899738628925429e-05, "loss": 0.4139, "step": 15270 }, { "epoch": 8.89406286379511, "grad_norm": 1.0028564929962158, "learning_rate": 8.898013344586312e-05, "loss": 0.3922, "step": 15280 }, { "epoch": 8.89988358556461, "grad_norm": 1.0441405773162842, "learning_rate": 8.896286876133661e-05, "loss": 0.403, "step": 15290 }, { "epoch": 8.90570430733411, "grad_norm": 0.9643349647521973, "learning_rate": 8.894559224091933e-05, "loss": 0.3996, "step": 15300 }, { "epoch": 8.91152502910361, "grad_norm": 1.004678726196289, "learning_rate": 8.892830388985942e-05, "loss": 0.4052, "step": 15310 }, { "epoch": 8.917345750873109, "grad_norm": 1.1336777210235596, "learning_rate": 8.891100371340864e-05, "loss": 0.4208, "step": 15320 }, { "epoch": 8.923166472642608, "grad_norm": 1.0656025409698486, "learning_rate": 8.889369171682231e-05, "loss": 0.4061, "step": 15330 }, { "epoch": 8.928987194412107, "grad_norm": 1.010088562965393, "learning_rate": 8.887636790535936e-05, "loss": 0.4085, "step": 15340 }, { "epoch": 8.934807916181606, "grad_norm": 1.161237120628357, "learning_rate": 8.885903228428231e-05, "loss": 0.399, "step": 15350 }, { "epoch": 8.940628637951106, "grad_norm": 0.9955573678016663, "learning_rate": 8.884168485885727e-05, "loss": 0.4017, "step": 15360 }, { "epoch": 8.946449359720605, "grad_norm": 0.9828818440437317, "learning_rate": 8.882432563435393e-05, "loss": 0.3966, "step": 15370 }, { "epoch": 8.952270081490104, "grad_norm": 1.0250632762908936, "learning_rate": 8.880695461604556e-05, "loss": 0.392, "step": 15380 }, { "epoch": 8.958090803259605, "grad_norm": 0.97443026304245, "learning_rate": 8.878957180920901e-05, "loss": 0.4142, "step": 15390 }, { "epoch": 8.963911525029104, "grad_norm": 0.9944722652435303, "learning_rate": 8.877217721912473e-05, "loss": 0.4169, "step": 15400 }, { "epoch": 8.969732246798603, "grad_norm": 1.0169012546539307, "learning_rate": 8.875477085107673e-05, "loss": 0.3956, "step": 15410 }, { "epoch": 8.975552968568103, "grad_norm": 1.0751420259475708, "learning_rate": 8.87373527103526e-05, "loss": 0.4055, "step": 15420 }, { "epoch": 8.981373690337602, "grad_norm": 1.0018045902252197, "learning_rate": 8.871992280224353e-05, "loss": 0.3939, "step": 15430 }, { "epoch": 8.987194412107101, "grad_norm": 0.9731285572052002, "learning_rate": 8.870248113204422e-05, "loss": 0.3992, "step": 15440 }, { "epoch": 8.9930151338766, "grad_norm": 0.980826199054718, "learning_rate": 8.868502770505306e-05, "loss": 0.4087, "step": 15450 }, { "epoch": 8.9988358556461, "grad_norm": 1.1109097003936768, "learning_rate": 8.86675625265719e-05, "loss": 0.4052, "step": 15460 }, { "epoch": 9.004656577415599, "grad_norm": 1.180631160736084, "learning_rate": 8.865008560190618e-05, "loss": 0.4088, "step": 15470 }, { "epoch": 9.0104772991851, "grad_norm": 1.0816713571548462, "learning_rate": 8.863259693636496e-05, "loss": 0.3891, "step": 15480 }, { "epoch": 9.016298020954599, "grad_norm": 1.0288232564926147, "learning_rate": 8.861509653526083e-05, "loss": 0.4127, "step": 15490 }, { "epoch": 9.022118742724098, "grad_norm": 0.9954609274864197, "learning_rate": 8.859758440390993e-05, "loss": 0.3924, "step": 15500 }, { "epoch": 9.027939464493597, "grad_norm": 1.0634241104125977, "learning_rate": 8.858006054763202e-05, "loss": 0.3913, "step": 15510 }, { "epoch": 9.033760186263097, "grad_norm": 0.962258517742157, "learning_rate": 8.856252497175035e-05, "loss": 0.3992, "step": 15520 }, { "epoch": 9.039580908032596, "grad_norm": 1.1528639793395996, "learning_rate": 8.854497768159178e-05, "loss": 0.4106, "step": 15530 }, { "epoch": 9.045401629802095, "grad_norm": 1.0313599109649658, "learning_rate": 8.852741868248671e-05, "loss": 0.4177, "step": 15540 }, { "epoch": 9.051222351571594, "grad_norm": 0.9684638381004333, "learning_rate": 8.85098479797691e-05, "loss": 0.389, "step": 15550 }, { "epoch": 9.057043073341095, "grad_norm": 1.054498314857483, "learning_rate": 8.849226557877646e-05, "loss": 0.3966, "step": 15560 }, { "epoch": 9.062863795110594, "grad_norm": 0.9243377447128296, "learning_rate": 8.84746714848499e-05, "loss": 0.3932, "step": 15570 }, { "epoch": 9.068684516880094, "grad_norm": 0.9128921627998352, "learning_rate": 8.845706570333397e-05, "loss": 0.3986, "step": 15580 }, { "epoch": 9.074505238649593, "grad_norm": 1.1216557025909424, "learning_rate": 8.84394482395769e-05, "loss": 0.403, "step": 15590 }, { "epoch": 9.080325960419092, "grad_norm": 0.9462006092071533, "learning_rate": 8.842181909893038e-05, "loss": 0.3894, "step": 15600 }, { "epoch": 9.086146682188591, "grad_norm": 0.9625077247619629, "learning_rate": 8.840417828674969e-05, "loss": 0.3946, "step": 15610 }, { "epoch": 9.09196740395809, "grad_norm": 0.91734778881073, "learning_rate": 8.838652580839364e-05, "loss": 0.3947, "step": 15620 }, { "epoch": 9.09778812572759, "grad_norm": 0.9544431567192078, "learning_rate": 8.836886166922458e-05, "loss": 0.3831, "step": 15630 }, { "epoch": 9.103608847497089, "grad_norm": 1.0375651121139526, "learning_rate": 8.835118587460844e-05, "loss": 0.3991, "step": 15640 }, { "epoch": 9.10942956926659, "grad_norm": 0.9297574162483215, "learning_rate": 8.83334984299146e-05, "loss": 0.4069, "step": 15650 }, { "epoch": 9.115250291036089, "grad_norm": 1.0088951587677002, "learning_rate": 8.83157993405161e-05, "loss": 0.392, "step": 15660 }, { "epoch": 9.121071012805588, "grad_norm": 1.0672118663787842, "learning_rate": 8.829808861178943e-05, "loss": 0.3998, "step": 15670 }, { "epoch": 9.126891734575088, "grad_norm": 1.0888123512268066, "learning_rate": 8.828036624911464e-05, "loss": 0.3895, "step": 15680 }, { "epoch": 9.132712456344587, "grad_norm": 0.9593304395675659, "learning_rate": 8.826263225787532e-05, "loss": 0.3978, "step": 15690 }, { "epoch": 9.138533178114086, "grad_norm": 0.969878077507019, "learning_rate": 8.824488664345858e-05, "loss": 0.386, "step": 15700 }, { "epoch": 9.144353899883585, "grad_norm": 0.9153664112091064, "learning_rate": 8.822712941125508e-05, "loss": 0.4049, "step": 15710 }, { "epoch": 9.150174621653084, "grad_norm": 0.9127864837646484, "learning_rate": 8.820936056665898e-05, "loss": 0.4036, "step": 15720 }, { "epoch": 9.155995343422584, "grad_norm": 0.9258425235748291, "learning_rate": 8.819158011506801e-05, "loss": 0.3876, "step": 15730 }, { "epoch": 9.161816065192085, "grad_norm": 1.0277128219604492, "learning_rate": 8.81737880618834e-05, "loss": 0.4023, "step": 15740 }, { "epoch": 9.167636786961584, "grad_norm": 1.142960548400879, "learning_rate": 8.815598441250987e-05, "loss": 0.3911, "step": 15750 }, { "epoch": 9.173457508731083, "grad_norm": 0.9829190373420715, "learning_rate": 8.813816917235576e-05, "loss": 0.4071, "step": 15760 }, { "epoch": 9.179278230500582, "grad_norm": 1.179589867591858, "learning_rate": 8.812034234683282e-05, "loss": 0.4218, "step": 15770 }, { "epoch": 9.185098952270081, "grad_norm": 1.0653280019760132, "learning_rate": 8.810250394135637e-05, "loss": 0.4131, "step": 15780 }, { "epoch": 9.19091967403958, "grad_norm": 0.9940358400344849, "learning_rate": 8.808465396134529e-05, "loss": 0.4051, "step": 15790 }, { "epoch": 9.19674039580908, "grad_norm": 0.9872370362281799, "learning_rate": 8.806679241222189e-05, "loss": 0.4, "step": 15800 }, { "epoch": 9.202561117578579, "grad_norm": 1.0895814895629883, "learning_rate": 8.804891929941203e-05, "loss": 0.4056, "step": 15810 }, { "epoch": 9.208381839348078, "grad_norm": 0.9640239477157593, "learning_rate": 8.803103462834514e-05, "loss": 0.4243, "step": 15820 }, { "epoch": 9.21420256111758, "grad_norm": 1.051705241203308, "learning_rate": 8.801313840445408e-05, "loss": 0.4073, "step": 15830 }, { "epoch": 9.220023282887079, "grad_norm": 0.9793722033500671, "learning_rate": 8.799523063317524e-05, "loss": 0.3885, "step": 15840 }, { "epoch": 9.225844004656578, "grad_norm": 0.9495517015457153, "learning_rate": 8.797731131994854e-05, "loss": 0.4032, "step": 15850 }, { "epoch": 9.231664726426077, "grad_norm": 0.9357540607452393, "learning_rate": 8.795938047021739e-05, "loss": 0.3904, "step": 15860 }, { "epoch": 9.237485448195576, "grad_norm": 0.9782213568687439, "learning_rate": 8.794143808942872e-05, "loss": 0.3979, "step": 15870 }, { "epoch": 9.243306169965075, "grad_norm": 1.0884029865264893, "learning_rate": 8.792348418303296e-05, "loss": 0.4086, "step": 15880 }, { "epoch": 9.249126891734575, "grad_norm": 1.1364957094192505, "learning_rate": 8.790551875648398e-05, "loss": 0.3876, "step": 15890 }, { "epoch": 9.254947613504074, "grad_norm": 1.0641556978225708, "learning_rate": 8.788754181523926e-05, "loss": 0.402, "step": 15900 }, { "epoch": 9.260768335273575, "grad_norm": 0.9835860133171082, "learning_rate": 8.78695533647597e-05, "loss": 0.4072, "step": 15910 }, { "epoch": 9.266589057043074, "grad_norm": 0.8878558278083801, "learning_rate": 8.785155341050972e-05, "loss": 0.392, "step": 15920 }, { "epoch": 9.272409778812573, "grad_norm": 1.0202651023864746, "learning_rate": 8.783354195795721e-05, "loss": 0.4308, "step": 15930 }, { "epoch": 9.278230500582072, "grad_norm": 1.0167388916015625, "learning_rate": 8.78155190125736e-05, "loss": 0.3948, "step": 15940 }, { "epoch": 9.284051222351572, "grad_norm": 1.094007134437561, "learning_rate": 8.779748457983378e-05, "loss": 0.4052, "step": 15950 }, { "epoch": 9.28987194412107, "grad_norm": 1.0087147951126099, "learning_rate": 8.777943866521612e-05, "loss": 0.4037, "step": 15960 }, { "epoch": 9.29569266589057, "grad_norm": 1.018270492553711, "learning_rate": 8.77613812742025e-05, "loss": 0.4031, "step": 15970 }, { "epoch": 9.30151338766007, "grad_norm": 1.043968915939331, "learning_rate": 8.774331241227829e-05, "loss": 0.3986, "step": 15980 }, { "epoch": 9.307334109429569, "grad_norm": 1.0577888488769531, "learning_rate": 8.772523208493232e-05, "loss": 0.3951, "step": 15990 }, { "epoch": 9.31315483119907, "grad_norm": 0.9690160155296326, "learning_rate": 8.770714029765692e-05, "loss": 0.3979, "step": 16000 }, { "epoch": 9.318975552968569, "grad_norm": 1.0037480592727661, "learning_rate": 8.768903705594789e-05, "loss": 0.393, "step": 16010 }, { "epoch": 9.324796274738068, "grad_norm": 0.9506270885467529, "learning_rate": 8.767092236530453e-05, "loss": 0.3778, "step": 16020 }, { "epoch": 9.330616996507567, "grad_norm": 0.9943719506263733, "learning_rate": 8.76527962312296e-05, "loss": 0.3906, "step": 16030 }, { "epoch": 9.336437718277066, "grad_norm": 0.9879030585289001, "learning_rate": 8.763465865922934e-05, "loss": 0.3918, "step": 16040 }, { "epoch": 9.342258440046566, "grad_norm": 0.9589715003967285, "learning_rate": 8.761650965481347e-05, "loss": 0.4027, "step": 16050 }, { "epoch": 9.348079161816065, "grad_norm": 1.0049707889556885, "learning_rate": 8.759834922349516e-05, "loss": 0.3946, "step": 16060 }, { "epoch": 9.353899883585564, "grad_norm": 0.959449827671051, "learning_rate": 8.758017737079108e-05, "loss": 0.3975, "step": 16070 }, { "epoch": 9.359720605355063, "grad_norm": 0.983998715877533, "learning_rate": 8.756199410222137e-05, "loss": 0.4123, "step": 16080 }, { "epoch": 9.365541327124564, "grad_norm": 0.9339179992675781, "learning_rate": 8.754379942330963e-05, "loss": 0.4001, "step": 16090 }, { "epoch": 9.371362048894063, "grad_norm": 1.0160149335861206, "learning_rate": 8.75255933395829e-05, "loss": 0.3997, "step": 16100 }, { "epoch": 9.377182770663563, "grad_norm": 1.053941249847412, "learning_rate": 8.750737585657171e-05, "loss": 0.3936, "step": 16110 }, { "epoch": 9.383003492433062, "grad_norm": 0.9225413203239441, "learning_rate": 8.748914697981008e-05, "loss": 0.3978, "step": 16120 }, { "epoch": 9.388824214202561, "grad_norm": 1.0449206829071045, "learning_rate": 8.747090671483542e-05, "loss": 0.3854, "step": 16130 }, { "epoch": 9.39464493597206, "grad_norm": 0.9920462369918823, "learning_rate": 8.745265506718869e-05, "loss": 0.4051, "step": 16140 }, { "epoch": 9.40046565774156, "grad_norm": 0.9969406127929688, "learning_rate": 8.74343920424142e-05, "loss": 0.3934, "step": 16150 }, { "epoch": 9.406286379511059, "grad_norm": 0.9678584337234497, "learning_rate": 8.741611764605982e-05, "loss": 0.4101, "step": 16160 }, { "epoch": 9.412107101280558, "grad_norm": 1.0373449325561523, "learning_rate": 8.739783188367682e-05, "loss": 0.3827, "step": 16170 }, { "epoch": 9.417927823050059, "grad_norm": 1.0267305374145508, "learning_rate": 8.737953476081991e-05, "loss": 0.4137, "step": 16180 }, { "epoch": 9.423748544819558, "grad_norm": 1.0292011499404907, "learning_rate": 8.73612262830473e-05, "loss": 0.4003, "step": 16190 }, { "epoch": 9.429569266589057, "grad_norm": 1.0240932703018188, "learning_rate": 8.734290645592061e-05, "loss": 0.4077, "step": 16200 }, { "epoch": 9.435389988358557, "grad_norm": 0.9354557394981384, "learning_rate": 8.732457528500493e-05, "loss": 0.3755, "step": 16210 }, { "epoch": 9.441210710128056, "grad_norm": 1.007812738418579, "learning_rate": 8.730623277586875e-05, "loss": 0.388, "step": 16220 }, { "epoch": 9.447031431897555, "grad_norm": 1.002661943435669, "learning_rate": 8.72878789340841e-05, "loss": 0.4016, "step": 16230 }, { "epoch": 9.452852153667054, "grad_norm": 1.089888095855713, "learning_rate": 8.726951376522635e-05, "loss": 0.3964, "step": 16240 }, { "epoch": 9.458672875436553, "grad_norm": 0.9454418420791626, "learning_rate": 8.725113727487435e-05, "loss": 0.3891, "step": 16250 }, { "epoch": 9.464493597206054, "grad_norm": 0.9434506893157959, "learning_rate": 8.723274946861042e-05, "loss": 0.408, "step": 16260 }, { "epoch": 9.470314318975554, "grad_norm": 0.8769670724868774, "learning_rate": 8.721435035202026e-05, "loss": 0.3835, "step": 16270 }, { "epoch": 9.476135040745053, "grad_norm": 1.0076355934143066, "learning_rate": 8.719593993069306e-05, "loss": 0.3925, "step": 16280 }, { "epoch": 9.481955762514552, "grad_norm": 0.9992807507514954, "learning_rate": 8.717751821022139e-05, "loss": 0.4019, "step": 16290 }, { "epoch": 9.487776484284051, "grad_norm": 0.986914873123169, "learning_rate": 8.715908519620134e-05, "loss": 0.3875, "step": 16300 }, { "epoch": 9.49359720605355, "grad_norm": 0.984212338924408, "learning_rate": 8.71406408942323e-05, "loss": 0.4059, "step": 16310 }, { "epoch": 9.49941792782305, "grad_norm": 1.0375996828079224, "learning_rate": 8.712218530991723e-05, "loss": 0.3953, "step": 16320 }, { "epoch": 9.505238649592549, "grad_norm": 0.9502151012420654, "learning_rate": 8.710371844886241e-05, "loss": 0.3965, "step": 16330 }, { "epoch": 9.511059371362048, "grad_norm": 0.9855525493621826, "learning_rate": 8.708524031667758e-05, "loss": 0.3933, "step": 16340 }, { "epoch": 9.516880093131547, "grad_norm": 0.998002827167511, "learning_rate": 8.706675091897592e-05, "loss": 0.3879, "step": 16350 }, { "epoch": 9.522700814901048, "grad_norm": 0.9820283651351929, "learning_rate": 8.704825026137404e-05, "loss": 0.3905, "step": 16360 }, { "epoch": 9.528521536670548, "grad_norm": 0.9808084964752197, "learning_rate": 8.702973834949192e-05, "loss": 0.3902, "step": 16370 }, { "epoch": 9.534342258440047, "grad_norm": 1.1941365003585815, "learning_rate": 8.701121518895301e-05, "loss": 0.3919, "step": 16380 }, { "epoch": 9.540162980209546, "grad_norm": 0.9993866086006165, "learning_rate": 8.699268078538414e-05, "loss": 0.3923, "step": 16390 }, { "epoch": 9.545983701979045, "grad_norm": 1.001605749130249, "learning_rate": 8.69741351444156e-05, "loss": 0.3946, "step": 16400 }, { "epoch": 9.551804423748544, "grad_norm": 0.9920091032981873, "learning_rate": 8.695557827168101e-05, "loss": 0.408, "step": 16410 }, { "epoch": 9.557625145518044, "grad_norm": 0.9819961190223694, "learning_rate": 8.693701017281753e-05, "loss": 0.3842, "step": 16420 }, { "epoch": 9.563445867287543, "grad_norm": 0.8770875930786133, "learning_rate": 8.691843085346563e-05, "loss": 0.3817, "step": 16430 }, { "epoch": 9.569266589057044, "grad_norm": 0.9287124276161194, "learning_rate": 8.689984031926919e-05, "loss": 0.3919, "step": 16440 }, { "epoch": 9.575087310826543, "grad_norm": 0.9504148960113525, "learning_rate": 8.688123857587555e-05, "loss": 0.3794, "step": 16450 }, { "epoch": 9.580908032596042, "grad_norm": 0.999089777469635, "learning_rate": 8.686262562893544e-05, "loss": 0.3893, "step": 16460 }, { "epoch": 9.586728754365542, "grad_norm": 0.983268678188324, "learning_rate": 8.684400148410294e-05, "loss": 0.3934, "step": 16470 }, { "epoch": 9.59254947613504, "grad_norm": 1.0008933544158936, "learning_rate": 8.682536614703562e-05, "loss": 0.4039, "step": 16480 }, { "epoch": 9.59837019790454, "grad_norm": 0.9156473278999329, "learning_rate": 8.680671962339437e-05, "loss": 0.4012, "step": 16490 }, { "epoch": 9.60419091967404, "grad_norm": 1.0114648342132568, "learning_rate": 8.678806191884352e-05, "loss": 0.4029, "step": 16500 }, { "epoch": 9.610011641443538, "grad_norm": 1.041835904121399, "learning_rate": 8.67693930390508e-05, "loss": 0.385, "step": 16510 }, { "epoch": 9.615832363213038, "grad_norm": 1.0692673921585083, "learning_rate": 8.67507129896873e-05, "loss": 0.3895, "step": 16520 }, { "epoch": 9.621653084982539, "grad_norm": 0.9336790442466736, "learning_rate": 8.673202177642757e-05, "loss": 0.3923, "step": 16530 }, { "epoch": 9.627473806752038, "grad_norm": 0.849489688873291, "learning_rate": 8.671331940494945e-05, "loss": 0.3965, "step": 16540 }, { "epoch": 9.633294528521537, "grad_norm": 1.0159111022949219, "learning_rate": 8.669460588093427e-05, "loss": 0.3998, "step": 16550 }, { "epoch": 9.639115250291036, "grad_norm": 0.9951322674751282, "learning_rate": 8.667588121006667e-05, "loss": 0.3934, "step": 16560 }, { "epoch": 9.644935972060535, "grad_norm": 0.9095257520675659, "learning_rate": 8.665714539803475e-05, "loss": 0.3946, "step": 16570 }, { "epoch": 9.650756693830035, "grad_norm": 1.0537036657333374, "learning_rate": 8.663839845052993e-05, "loss": 0.4006, "step": 16580 }, { "epoch": 9.656577415599534, "grad_norm": 0.911099374294281, "learning_rate": 8.661964037324703e-05, "loss": 0.3908, "step": 16590 }, { "epoch": 9.662398137369033, "grad_norm": 0.9691621661186218, "learning_rate": 8.660087117188427e-05, "loss": 0.393, "step": 16600 }, { "epoch": 9.668218859138534, "grad_norm": 1.0195798873901367, "learning_rate": 8.658209085214325e-05, "loss": 0.4, "step": 16610 }, { "epoch": 9.674039580908033, "grad_norm": 0.9890906810760498, "learning_rate": 8.656329941972891e-05, "loss": 0.399, "step": 16620 }, { "epoch": 9.679860302677533, "grad_norm": 0.9746270775794983, "learning_rate": 8.654449688034963e-05, "loss": 0.3922, "step": 16630 }, { "epoch": 9.685681024447032, "grad_norm": 1.0125324726104736, "learning_rate": 8.652568323971706e-05, "loss": 0.3935, "step": 16640 }, { "epoch": 9.691501746216531, "grad_norm": 0.9662505984306335, "learning_rate": 8.650685850354636e-05, "loss": 0.3932, "step": 16650 }, { "epoch": 9.69732246798603, "grad_norm": 0.9122781157493591, "learning_rate": 8.648802267755593e-05, "loss": 0.3971, "step": 16660 }, { "epoch": 9.70314318975553, "grad_norm": 1.07173752784729, "learning_rate": 8.646917576746764e-05, "loss": 0.4033, "step": 16670 }, { "epoch": 9.708963911525029, "grad_norm": 0.9979486465454102, "learning_rate": 8.645031777900666e-05, "loss": 0.382, "step": 16680 }, { "epoch": 9.714784633294528, "grad_norm": 1.1028109788894653, "learning_rate": 8.643144871790154e-05, "loss": 0.3816, "step": 16690 }, { "epoch": 9.720605355064027, "grad_norm": 0.9720006585121155, "learning_rate": 8.641256858988424e-05, "loss": 0.4072, "step": 16700 }, { "epoch": 9.726426076833528, "grad_norm": 0.9981321096420288, "learning_rate": 8.639367740069e-05, "loss": 0.3823, "step": 16710 }, { "epoch": 9.732246798603027, "grad_norm": 1.0622848272323608, "learning_rate": 8.63747751560575e-05, "loss": 0.3963, "step": 16720 }, { "epoch": 9.738067520372526, "grad_norm": 1.0503687858581543, "learning_rate": 8.635586186172871e-05, "loss": 0.3836, "step": 16730 }, { "epoch": 9.743888242142026, "grad_norm": 0.9788289666175842, "learning_rate": 8.633693752344902e-05, "loss": 0.3896, "step": 16740 }, { "epoch": 9.749708963911525, "grad_norm": 1.0156762599945068, "learning_rate": 8.631800214696713e-05, "loss": 0.3873, "step": 16750 }, { "epoch": 9.755529685681024, "grad_norm": 1.0223188400268555, "learning_rate": 8.629905573803511e-05, "loss": 0.3983, "step": 16760 }, { "epoch": 9.761350407450523, "grad_norm": 1.0052690505981445, "learning_rate": 8.628009830240839e-05, "loss": 0.4045, "step": 16770 }, { "epoch": 9.767171129220023, "grad_norm": 0.9968734979629517, "learning_rate": 8.626112984584571e-05, "loss": 0.3892, "step": 16780 }, { "epoch": 9.772991850989523, "grad_norm": 1.0218911170959473, "learning_rate": 8.62421503741092e-05, "loss": 0.3939, "step": 16790 }, { "epoch": 9.778812572759023, "grad_norm": 0.9858812093734741, "learning_rate": 8.622315989296432e-05, "loss": 0.3934, "step": 16800 }, { "epoch": 9.784633294528522, "grad_norm": 1.0227458477020264, "learning_rate": 8.62041584081799e-05, "loss": 0.3996, "step": 16810 }, { "epoch": 9.790454016298021, "grad_norm": 0.9977705478668213, "learning_rate": 8.618514592552807e-05, "loss": 0.3985, "step": 16820 }, { "epoch": 9.79627473806752, "grad_norm": 1.1228009462356567, "learning_rate": 8.616612245078431e-05, "loss": 0.3916, "step": 16830 }, { "epoch": 9.80209545983702, "grad_norm": 1.037785291671753, "learning_rate": 8.614708798972746e-05, "loss": 0.3863, "step": 16840 }, { "epoch": 9.807916181606519, "grad_norm": 1.0333179235458374, "learning_rate": 8.61280425481397e-05, "loss": 0.3885, "step": 16850 }, { "epoch": 9.813736903376018, "grad_norm": 0.9782261252403259, "learning_rate": 8.61089861318065e-05, "loss": 0.3828, "step": 16860 }, { "epoch": 9.819557625145517, "grad_norm": 0.9899381995201111, "learning_rate": 8.608991874651673e-05, "loss": 0.3784, "step": 16870 }, { "epoch": 9.825378346915018, "grad_norm": 0.9337296485900879, "learning_rate": 8.607084039806255e-05, "loss": 0.4014, "step": 16880 }, { "epoch": 9.831199068684517, "grad_norm": 1.0337886810302734, "learning_rate": 8.605175109223944e-05, "loss": 0.3949, "step": 16890 }, { "epoch": 9.837019790454017, "grad_norm": 1.1782559156417847, "learning_rate": 8.603265083484624e-05, "loss": 0.4158, "step": 16900 }, { "epoch": 9.842840512223516, "grad_norm": 1.0882201194763184, "learning_rate": 8.60135396316851e-05, "loss": 0.3946, "step": 16910 }, { "epoch": 9.848661233993015, "grad_norm": 0.9627221822738647, "learning_rate": 8.599441748856152e-05, "loss": 0.3799, "step": 16920 }, { "epoch": 9.854481955762514, "grad_norm": 1.0332645177841187, "learning_rate": 8.597528441128427e-05, "loss": 0.3876, "step": 16930 }, { "epoch": 9.860302677532014, "grad_norm": 1.0918800830841064, "learning_rate": 8.595614040566549e-05, "loss": 0.3951, "step": 16940 }, { "epoch": 9.866123399301513, "grad_norm": 1.0045740604400635, "learning_rate": 8.593698547752063e-05, "loss": 0.3763, "step": 16950 }, { "epoch": 9.871944121071014, "grad_norm": 1.028402328491211, "learning_rate": 8.591781963266843e-05, "loss": 0.4034, "step": 16960 }, { "epoch": 9.877764842840513, "grad_norm": 0.9887080788612366, "learning_rate": 8.5898642876931e-05, "loss": 0.3898, "step": 16970 }, { "epoch": 9.883585564610012, "grad_norm": 0.9212883114814758, "learning_rate": 8.587945521613369e-05, "loss": 0.3847, "step": 16980 }, { "epoch": 9.889406286379511, "grad_norm": 0.9757636189460754, "learning_rate": 8.586025665610524e-05, "loss": 0.4015, "step": 16990 }, { "epoch": 9.89522700814901, "grad_norm": 1.022473931312561, "learning_rate": 8.584104720267765e-05, "loss": 0.4052, "step": 17000 }, { "epoch": 9.90104772991851, "grad_norm": 1.0031172037124634, "learning_rate": 8.582182686168625e-05, "loss": 0.385, "step": 17010 }, { "epoch": 9.906868451688009, "grad_norm": 1.0584319829940796, "learning_rate": 8.580259563896967e-05, "loss": 0.4032, "step": 17020 }, { "epoch": 9.912689173457508, "grad_norm": 1.054803490638733, "learning_rate": 8.578335354036983e-05, "loss": 0.3946, "step": 17030 }, { "epoch": 9.918509895227007, "grad_norm": 0.9575402140617371, "learning_rate": 8.576410057173201e-05, "loss": 0.3878, "step": 17040 }, { "epoch": 9.924330616996507, "grad_norm": 0.9389042258262634, "learning_rate": 8.574483673890474e-05, "loss": 0.3778, "step": 17050 }, { "epoch": 9.930151338766008, "grad_norm": 0.9930344820022583, "learning_rate": 8.572556204773983e-05, "loss": 0.3918, "step": 17060 }, { "epoch": 9.935972060535507, "grad_norm": 1.0026341676712036, "learning_rate": 8.570627650409246e-05, "loss": 0.4064, "step": 17070 }, { "epoch": 9.941792782305006, "grad_norm": 1.000767707824707, "learning_rate": 8.568698011382107e-05, "loss": 0.3992, "step": 17080 }, { "epoch": 9.947613504074505, "grad_norm": 0.9974197149276733, "learning_rate": 8.566767288278738e-05, "loss": 0.3778, "step": 17090 }, { "epoch": 9.953434225844005, "grad_norm": 0.9581155776977539, "learning_rate": 8.56483548168564e-05, "loss": 0.3932, "step": 17100 }, { "epoch": 9.959254947613504, "grad_norm": 1.0369043350219727, "learning_rate": 8.562902592189648e-05, "loss": 0.3987, "step": 17110 }, { "epoch": 9.965075669383003, "grad_norm": 0.9737750291824341, "learning_rate": 8.560968620377921e-05, "loss": 0.399, "step": 17120 }, { "epoch": 9.970896391152502, "grad_norm": 0.9955225586891174, "learning_rate": 8.559033566837951e-05, "loss": 0.4116, "step": 17130 }, { "epoch": 9.976717112922003, "grad_norm": 0.9365544319152832, "learning_rate": 8.557097432157551e-05, "loss": 0.3899, "step": 17140 }, { "epoch": 9.982537834691502, "grad_norm": 0.9970062971115112, "learning_rate": 8.555160216924872e-05, "loss": 0.4007, "step": 17150 }, { "epoch": 9.988358556461002, "grad_norm": 0.8926650285720825, "learning_rate": 8.55322192172839e-05, "loss": 0.3994, "step": 17160 }, { "epoch": 9.9941792782305, "grad_norm": 0.8537611961364746, "learning_rate": 8.551282547156902e-05, "loss": 0.3903, "step": 17170 }, { "epoch": 10.0, "grad_norm": 0.9665818810462952, "learning_rate": 8.549342093799544e-05, "loss": 0.3978, "step": 17180 }, { "epoch": 10.0058207217695, "grad_norm": 0.9007522463798523, "learning_rate": 8.547400562245773e-05, "loss": 0.3916, "step": 17190 }, { "epoch": 10.011641443538998, "grad_norm": 0.9699416756629944, "learning_rate": 8.545457953085374e-05, "loss": 0.3886, "step": 17200 }, { "epoch": 10.017462165308498, "grad_norm": 0.889772891998291, "learning_rate": 8.543514266908463e-05, "loss": 0.3849, "step": 17210 }, { "epoch": 10.023282887077997, "grad_norm": 0.8279592394828796, "learning_rate": 8.541569504305478e-05, "loss": 0.3891, "step": 17220 }, { "epoch": 10.029103608847498, "grad_norm": 0.9780682325363159, "learning_rate": 8.539623665867187e-05, "loss": 0.3921, "step": 17230 }, { "epoch": 10.034924330616997, "grad_norm": 0.9797926545143127, "learning_rate": 8.537676752184685e-05, "loss": 0.3876, "step": 17240 }, { "epoch": 10.040745052386496, "grad_norm": 1.0248407125473022, "learning_rate": 8.53572876384939e-05, "loss": 0.3915, "step": 17250 }, { "epoch": 10.046565774155995, "grad_norm": 1.0217307806015015, "learning_rate": 8.533779701453056e-05, "loss": 0.3859, "step": 17260 }, { "epoch": 10.052386495925495, "grad_norm": 1.1478126049041748, "learning_rate": 8.53182956558775e-05, "loss": 0.4048, "step": 17270 }, { "epoch": 10.058207217694994, "grad_norm": 1.030155062675476, "learning_rate": 8.529878356845877e-05, "loss": 0.4045, "step": 17280 }, { "epoch": 10.064027939464493, "grad_norm": 0.8832722902297974, "learning_rate": 8.527926075820158e-05, "loss": 0.3897, "step": 17290 }, { "epoch": 10.069848661233992, "grad_norm": 1.0355886220932007, "learning_rate": 8.525972723103648e-05, "loss": 0.4088, "step": 17300 }, { "epoch": 10.075669383003492, "grad_norm": 0.9063625931739807, "learning_rate": 8.524018299289722e-05, "loss": 0.3885, "step": 17310 }, { "epoch": 10.081490104772993, "grad_norm": 0.8771740198135376, "learning_rate": 8.522062804972083e-05, "loss": 0.3879, "step": 17320 }, { "epoch": 10.087310826542492, "grad_norm": 0.9558520913124084, "learning_rate": 8.520106240744759e-05, "loss": 0.4061, "step": 17330 }, { "epoch": 10.093131548311991, "grad_norm": 0.8802461624145508, "learning_rate": 8.518148607202102e-05, "loss": 0.3971, "step": 17340 }, { "epoch": 10.09895227008149, "grad_norm": 0.942169725894928, "learning_rate": 8.51618990493879e-05, "loss": 0.3961, "step": 17350 }, { "epoch": 10.10477299185099, "grad_norm": 0.9669355750083923, "learning_rate": 8.514230134549823e-05, "loss": 0.4008, "step": 17360 }, { "epoch": 10.110593713620489, "grad_norm": 0.9199807047843933, "learning_rate": 8.51226929663053e-05, "loss": 0.3787, "step": 17370 }, { "epoch": 10.116414435389988, "grad_norm": 0.9808560609817505, "learning_rate": 8.51030739177656e-05, "loss": 0.3879, "step": 17380 }, { "epoch": 10.122235157159487, "grad_norm": 0.9619638919830322, "learning_rate": 8.508344420583889e-05, "loss": 0.3917, "step": 17390 }, { "epoch": 10.128055878928988, "grad_norm": 0.9815690517425537, "learning_rate": 8.506380383648816e-05, "loss": 0.3879, "step": 17400 }, { "epoch": 10.133876600698487, "grad_norm": 1.052003026008606, "learning_rate": 8.504415281567963e-05, "loss": 0.3909, "step": 17410 }, { "epoch": 10.139697322467986, "grad_norm": 1.1139363050460815, "learning_rate": 8.502449114938275e-05, "loss": 0.4135, "step": 17420 }, { "epoch": 10.145518044237486, "grad_norm": 0.9614306688308716, "learning_rate": 8.500481884357025e-05, "loss": 0.3946, "step": 17430 }, { "epoch": 10.151338766006985, "grad_norm": 0.9198089838027954, "learning_rate": 8.498513590421801e-05, "loss": 0.3956, "step": 17440 }, { "epoch": 10.157159487776484, "grad_norm": 1.0611704587936401, "learning_rate": 8.496544233730522e-05, "loss": 0.3851, "step": 17450 }, { "epoch": 10.162980209545983, "grad_norm": 0.8887783885002136, "learning_rate": 8.494573814881426e-05, "loss": 0.389, "step": 17460 }, { "epoch": 10.168800931315483, "grad_norm": 1.028727650642395, "learning_rate": 8.492602334473074e-05, "loss": 0.3832, "step": 17470 }, { "epoch": 10.174621653084982, "grad_norm": 0.8881736397743225, "learning_rate": 8.49062979310435e-05, "loss": 0.4022, "step": 17480 }, { "epoch": 10.180442374854483, "grad_norm": 0.9344576597213745, "learning_rate": 8.488656191374458e-05, "loss": 0.387, "step": 17490 }, { "epoch": 10.186263096623982, "grad_norm": 0.991486668586731, "learning_rate": 8.48668152988293e-05, "loss": 0.3786, "step": 17500 }, { "epoch": 10.192083818393481, "grad_norm": 0.9163709878921509, "learning_rate": 8.484705809229612e-05, "loss": 0.3792, "step": 17510 }, { "epoch": 10.19790454016298, "grad_norm": 0.9571926593780518, "learning_rate": 8.482729030014677e-05, "loss": 0.3894, "step": 17520 }, { "epoch": 10.20372526193248, "grad_norm": 1.020944595336914, "learning_rate": 8.48075119283862e-05, "loss": 0.3903, "step": 17530 }, { "epoch": 10.209545983701979, "grad_norm": 1.0042294263839722, "learning_rate": 8.478772298302254e-05, "loss": 0.4095, "step": 17540 }, { "epoch": 10.215366705471478, "grad_norm": 1.0724114179611206, "learning_rate": 8.476792347006716e-05, "loss": 0.3979, "step": 17550 }, { "epoch": 10.221187427240977, "grad_norm": 0.9943491220474243, "learning_rate": 8.474811339553462e-05, "loss": 0.3961, "step": 17560 }, { "epoch": 10.227008149010477, "grad_norm": 0.9863641262054443, "learning_rate": 8.47282927654427e-05, "loss": 0.4001, "step": 17570 }, { "epoch": 10.232828870779977, "grad_norm": 0.9441685080528259, "learning_rate": 8.470846158581238e-05, "loss": 0.4102, "step": 17580 }, { "epoch": 10.238649592549477, "grad_norm": 0.9593627452850342, "learning_rate": 8.468861986266787e-05, "loss": 0.3901, "step": 17590 }, { "epoch": 10.244470314318976, "grad_norm": 0.9238294959068298, "learning_rate": 8.466876760203654e-05, "loss": 0.3717, "step": 17600 }, { "epoch": 10.250291036088475, "grad_norm": 0.9268264770507812, "learning_rate": 8.464890480994898e-05, "loss": 0.3654, "step": 17610 }, { "epoch": 10.256111757857974, "grad_norm": 0.9504760503768921, "learning_rate": 8.462903149243899e-05, "loss": 0.3762, "step": 17620 }, { "epoch": 10.261932479627474, "grad_norm": 0.9536232948303223, "learning_rate": 8.460914765554357e-05, "loss": 0.3943, "step": 17630 }, { "epoch": 10.267753201396973, "grad_norm": 0.9565178751945496, "learning_rate": 8.458925330530288e-05, "loss": 0.4054, "step": 17640 }, { "epoch": 10.273573923166472, "grad_norm": 0.9149278402328491, "learning_rate": 8.456934844776032e-05, "loss": 0.3877, "step": 17650 }, { "epoch": 10.279394644935971, "grad_norm": 0.9248025417327881, "learning_rate": 8.454943308896246e-05, "loss": 0.3936, "step": 17660 }, { "epoch": 10.285215366705472, "grad_norm": 0.9986094236373901, "learning_rate": 8.452950723495905e-05, "loss": 0.4038, "step": 17670 }, { "epoch": 10.291036088474971, "grad_norm": 0.9393812417984009, "learning_rate": 8.450957089180303e-05, "loss": 0.3796, "step": 17680 }, { "epoch": 10.29685681024447, "grad_norm": 0.9657251834869385, "learning_rate": 8.448962406555055e-05, "loss": 0.3786, "step": 17690 }, { "epoch": 10.30267753201397, "grad_norm": 0.9948515892028809, "learning_rate": 8.446966676226093e-05, "loss": 0.4138, "step": 17700 }, { "epoch": 10.308498253783469, "grad_norm": 0.9797689914703369, "learning_rate": 8.444969898799667e-05, "loss": 0.3979, "step": 17710 }, { "epoch": 10.314318975552968, "grad_norm": 1.096710443496704, "learning_rate": 8.442972074882343e-05, "loss": 0.395, "step": 17720 }, { "epoch": 10.320139697322467, "grad_norm": 0.9963417649269104, "learning_rate": 8.44097320508101e-05, "loss": 0.4037, "step": 17730 }, { "epoch": 10.325960419091967, "grad_norm": 1.0579969882965088, "learning_rate": 8.43897329000287e-05, "loss": 0.391, "step": 17740 }, { "epoch": 10.331781140861466, "grad_norm": 0.9009507298469543, "learning_rate": 8.436972330255448e-05, "loss": 0.3709, "step": 17750 }, { "epoch": 10.337601862630967, "grad_norm": 1.0239449739456177, "learning_rate": 8.434970326446579e-05, "loss": 0.381, "step": 17760 }, { "epoch": 10.343422584400466, "grad_norm": 0.92396080493927, "learning_rate": 8.432967279184418e-05, "loss": 0.4057, "step": 17770 }, { "epoch": 10.349243306169965, "grad_norm": 0.9746524691581726, "learning_rate": 8.430963189077441e-05, "loss": 0.3949, "step": 17780 }, { "epoch": 10.355064027939465, "grad_norm": 1.049760341644287, "learning_rate": 8.428958056734437e-05, "loss": 0.39, "step": 17790 }, { "epoch": 10.360884749708964, "grad_norm": 1.0604331493377686, "learning_rate": 8.426951882764513e-05, "loss": 0.3933, "step": 17800 }, { "epoch": 10.366705471478463, "grad_norm": 0.9911412596702576, "learning_rate": 8.424944667777089e-05, "loss": 0.3937, "step": 17810 }, { "epoch": 10.372526193247962, "grad_norm": 1.0019129514694214, "learning_rate": 8.422936412381905e-05, "loss": 0.3905, "step": 17820 }, { "epoch": 10.378346915017461, "grad_norm": 0.9231011271476746, "learning_rate": 8.420927117189017e-05, "loss": 0.3827, "step": 17830 }, { "epoch": 10.384167636786962, "grad_norm": 0.9384064674377441, "learning_rate": 8.418916782808795e-05, "loss": 0.4005, "step": 17840 }, { "epoch": 10.389988358556462, "grad_norm": 1.0310536623001099, "learning_rate": 8.416905409851926e-05, "loss": 0.4179, "step": 17850 }, { "epoch": 10.39580908032596, "grad_norm": 1.1868290901184082, "learning_rate": 8.41489299892941e-05, "loss": 0.3884, "step": 17860 }, { "epoch": 10.40162980209546, "grad_norm": 1.000658392906189, "learning_rate": 8.412879550652566e-05, "loss": 0.3859, "step": 17870 }, { "epoch": 10.40745052386496, "grad_norm": 0.969255805015564, "learning_rate": 8.410865065633029e-05, "loss": 0.403, "step": 17880 }, { "epoch": 10.413271245634458, "grad_norm": 0.8980079889297485, "learning_rate": 8.408849544482742e-05, "loss": 0.3837, "step": 17890 }, { "epoch": 10.419091967403958, "grad_norm": 0.9710978865623474, "learning_rate": 8.406832987813968e-05, "loss": 0.3891, "step": 17900 }, { "epoch": 10.424912689173457, "grad_norm": 0.9499503374099731, "learning_rate": 8.404815396239286e-05, "loss": 0.3874, "step": 17910 }, { "epoch": 10.430733410942956, "grad_norm": 0.9150850772857666, "learning_rate": 8.402796770371587e-05, "loss": 0.3974, "step": 17920 }, { "epoch": 10.436554132712457, "grad_norm": 0.9240706562995911, "learning_rate": 8.400777110824071e-05, "loss": 0.3928, "step": 17930 }, { "epoch": 10.442374854481956, "grad_norm": 0.934883713722229, "learning_rate": 8.398756418210263e-05, "loss": 0.3877, "step": 17940 }, { "epoch": 10.448195576251456, "grad_norm": 0.9045294523239136, "learning_rate": 8.396734693143993e-05, "loss": 0.3836, "step": 17950 }, { "epoch": 10.454016298020955, "grad_norm": 1.0833239555358887, "learning_rate": 8.39471193623941e-05, "loss": 0.3902, "step": 17960 }, { "epoch": 10.459837019790454, "grad_norm": 1.0409271717071533, "learning_rate": 8.392688148110974e-05, "loss": 0.3949, "step": 17970 }, { "epoch": 10.465657741559953, "grad_norm": 0.8864637613296509, "learning_rate": 8.390663329373456e-05, "loss": 0.3914, "step": 17980 }, { "epoch": 10.471478463329452, "grad_norm": 0.9357925057411194, "learning_rate": 8.388637480641944e-05, "loss": 0.3958, "step": 17990 }, { "epoch": 10.477299185098952, "grad_norm": 1.0160051584243774, "learning_rate": 8.386610602531837e-05, "loss": 0.4014, "step": 18000 }, { "epoch": 10.48311990686845, "grad_norm": 0.9866865277290344, "learning_rate": 8.384582695658847e-05, "loss": 0.3897, "step": 18010 }, { "epoch": 10.488940628637952, "grad_norm": 0.9702135920524597, "learning_rate": 8.382553760638999e-05, "loss": 0.3931, "step": 18020 }, { "epoch": 10.494761350407451, "grad_norm": 0.9299555420875549, "learning_rate": 8.380523798088631e-05, "loss": 0.3913, "step": 18030 }, { "epoch": 10.50058207217695, "grad_norm": 0.9349640607833862, "learning_rate": 8.378492808624389e-05, "loss": 0.3916, "step": 18040 }, { "epoch": 10.50640279394645, "grad_norm": 1.0012950897216797, "learning_rate": 8.376460792863237e-05, "loss": 0.3966, "step": 18050 }, { "epoch": 10.512223515715949, "grad_norm": 0.9553634524345398, "learning_rate": 8.374427751422444e-05, "loss": 0.3925, "step": 18060 }, { "epoch": 10.518044237485448, "grad_norm": 0.9426121115684509, "learning_rate": 8.3723936849196e-05, "loss": 0.3806, "step": 18070 }, { "epoch": 10.523864959254947, "grad_norm": 0.9251390099525452, "learning_rate": 8.370358593972595e-05, "loss": 0.3891, "step": 18080 }, { "epoch": 10.529685681024446, "grad_norm": 0.9848257303237915, "learning_rate": 8.36832247919964e-05, "loss": 0.3809, "step": 18090 }, { "epoch": 10.535506402793946, "grad_norm": 0.9541112184524536, "learning_rate": 8.36628534121925e-05, "loss": 0.3831, "step": 18100 }, { "epoch": 10.541327124563447, "grad_norm": 0.9072530269622803, "learning_rate": 8.364247180650254e-05, "loss": 0.3962, "step": 18110 }, { "epoch": 10.547147846332946, "grad_norm": 0.9500899314880371, "learning_rate": 8.362207998111794e-05, "loss": 0.3862, "step": 18120 }, { "epoch": 10.552968568102445, "grad_norm": 0.9345365166664124, "learning_rate": 8.360167794223318e-05, "loss": 0.3747, "step": 18130 }, { "epoch": 10.558789289871944, "grad_norm": 0.9664402008056641, "learning_rate": 8.358126569604586e-05, "loss": 0.3992, "step": 18140 }, { "epoch": 10.564610011641443, "grad_norm": 0.9770203232765198, "learning_rate": 8.356084324875668e-05, "loss": 0.3732, "step": 18150 }, { "epoch": 10.570430733410943, "grad_norm": 0.9495903849601746, "learning_rate": 8.354041060656945e-05, "loss": 0.3956, "step": 18160 }, { "epoch": 10.576251455180442, "grad_norm": 0.9336432814598083, "learning_rate": 8.351996777569106e-05, "loss": 0.3767, "step": 18170 }, { "epoch": 10.582072176949941, "grad_norm": 0.8032680153846741, "learning_rate": 8.349951476233148e-05, "loss": 0.3944, "step": 18180 }, { "epoch": 10.587892898719442, "grad_norm": 0.9940772652626038, "learning_rate": 8.347905157270386e-05, "loss": 0.3954, "step": 18190 }, { "epoch": 10.593713620488941, "grad_norm": 0.926501989364624, "learning_rate": 8.345857821302432e-05, "loss": 0.3796, "step": 18200 }, { "epoch": 10.59953434225844, "grad_norm": 0.9573755860328674, "learning_rate": 8.343809468951213e-05, "loss": 0.3723, "step": 18210 }, { "epoch": 10.60535506402794, "grad_norm": 0.9987868070602417, "learning_rate": 8.341760100838965e-05, "loss": 0.3855, "step": 18220 }, { "epoch": 10.611175785797439, "grad_norm": 0.9829151034355164, "learning_rate": 8.339709717588233e-05, "loss": 0.3997, "step": 18230 }, { "epoch": 10.616996507566938, "grad_norm": 1.0631672143936157, "learning_rate": 8.33765831982187e-05, "loss": 0.3936, "step": 18240 }, { "epoch": 10.622817229336437, "grad_norm": 1.0479860305786133, "learning_rate": 8.335605908163035e-05, "loss": 0.3997, "step": 18250 }, { "epoch": 10.628637951105937, "grad_norm": 0.917760968208313, "learning_rate": 8.333552483235196e-05, "loss": 0.387, "step": 18260 }, { "epoch": 10.634458672875436, "grad_norm": 0.9306721091270447, "learning_rate": 8.33149804566213e-05, "loss": 0.3816, "step": 18270 }, { "epoch": 10.640279394644937, "grad_norm": 0.9347231984138489, "learning_rate": 8.329442596067921e-05, "loss": 0.3714, "step": 18280 }, { "epoch": 10.646100116414436, "grad_norm": 0.8591166734695435, "learning_rate": 8.32738613507696e-05, "loss": 0.3913, "step": 18290 }, { "epoch": 10.651920838183935, "grad_norm": 1.2140979766845703, "learning_rate": 8.325328663313946e-05, "loss": 0.3885, "step": 18300 }, { "epoch": 10.657741559953434, "grad_norm": 1.096247673034668, "learning_rate": 8.323270181403884e-05, "loss": 0.3914, "step": 18310 }, { "epoch": 10.663562281722934, "grad_norm": 0.8836649060249329, "learning_rate": 8.321210689972086e-05, "loss": 0.3645, "step": 18320 }, { "epoch": 10.669383003492433, "grad_norm": 0.9551728963851929, "learning_rate": 8.319150189644174e-05, "loss": 0.3899, "step": 18330 }, { "epoch": 10.675203725261932, "grad_norm": 0.8546008467674255, "learning_rate": 8.31708868104607e-05, "loss": 0.3763, "step": 18340 }, { "epoch": 10.681024447031431, "grad_norm": 0.8562464714050293, "learning_rate": 8.315026164804007e-05, "loss": 0.3784, "step": 18350 }, { "epoch": 10.68684516880093, "grad_norm": 0.9812507033348083, "learning_rate": 8.312962641544524e-05, "loss": 0.3983, "step": 18360 }, { "epoch": 10.692665890570431, "grad_norm": 0.9956711530685425, "learning_rate": 8.310898111894465e-05, "loss": 0.3956, "step": 18370 }, { "epoch": 10.69848661233993, "grad_norm": 1.049071192741394, "learning_rate": 8.308832576480977e-05, "loss": 0.3957, "step": 18380 }, { "epoch": 10.70430733410943, "grad_norm": 0.8905249834060669, "learning_rate": 8.306766035931519e-05, "loss": 0.3885, "step": 18390 }, { "epoch": 10.71012805587893, "grad_norm": 0.870517909526825, "learning_rate": 8.304698490873847e-05, "loss": 0.396, "step": 18400 }, { "epoch": 10.715948777648428, "grad_norm": 0.8843123316764832, "learning_rate": 8.30262994193603e-05, "loss": 0.3991, "step": 18410 }, { "epoch": 10.721769499417928, "grad_norm": 0.9508868455886841, "learning_rate": 8.300560389746438e-05, "loss": 0.3885, "step": 18420 }, { "epoch": 10.727590221187427, "grad_norm": 0.9100558757781982, "learning_rate": 8.298489834933745e-05, "loss": 0.3855, "step": 18430 }, { "epoch": 10.733410942956926, "grad_norm": 0.885618269443512, "learning_rate": 8.296418278126934e-05, "loss": 0.3895, "step": 18440 }, { "epoch": 10.739231664726425, "grad_norm": 0.981644332408905, "learning_rate": 8.294345719955284e-05, "loss": 0.3861, "step": 18450 }, { "epoch": 10.745052386495926, "grad_norm": 1.071499228477478, "learning_rate": 8.29227216104839e-05, "loss": 0.3908, "step": 18460 }, { "epoch": 10.750873108265425, "grad_norm": 0.9536086916923523, "learning_rate": 8.290197602036137e-05, "loss": 0.3852, "step": 18470 }, { "epoch": 10.756693830034925, "grad_norm": 0.9378873705863953, "learning_rate": 8.288122043548725e-05, "loss": 0.3861, "step": 18480 }, { "epoch": 10.762514551804424, "grad_norm": 0.9300410747528076, "learning_rate": 8.286045486216657e-05, "loss": 0.3942, "step": 18490 }, { "epoch": 10.768335273573923, "grad_norm": 0.9408305287361145, "learning_rate": 8.283967930670733e-05, "loss": 0.3897, "step": 18500 }, { "epoch": 10.774155995343422, "grad_norm": 1.0752365589141846, "learning_rate": 8.281889377542058e-05, "loss": 0.3991, "step": 18510 }, { "epoch": 10.779976717112921, "grad_norm": 0.9809948205947876, "learning_rate": 8.279809827462045e-05, "loss": 0.3857, "step": 18520 }, { "epoch": 10.78579743888242, "grad_norm": 0.9074745774269104, "learning_rate": 8.277729281062402e-05, "loss": 0.3869, "step": 18530 }, { "epoch": 10.791618160651922, "grad_norm": 0.9109318256378174, "learning_rate": 8.27564773897515e-05, "loss": 0.4042, "step": 18540 }, { "epoch": 10.797438882421421, "grad_norm": 1.034658432006836, "learning_rate": 8.273565201832602e-05, "loss": 0.3866, "step": 18550 }, { "epoch": 10.80325960419092, "grad_norm": 0.9266527891159058, "learning_rate": 8.27148167026738e-05, "loss": 0.4051, "step": 18560 }, { "epoch": 10.80908032596042, "grad_norm": 0.871245801448822, "learning_rate": 8.269397144912405e-05, "loss": 0.392, "step": 18570 }, { "epoch": 10.814901047729919, "grad_norm": 0.9932366013526917, "learning_rate": 8.267311626400899e-05, "loss": 0.3731, "step": 18580 }, { "epoch": 10.820721769499418, "grad_norm": 1.0286118984222412, "learning_rate": 8.26522511536639e-05, "loss": 0.3859, "step": 18590 }, { "epoch": 10.826542491268917, "grad_norm": 0.9670930504798889, "learning_rate": 8.263137612442706e-05, "loss": 0.3983, "step": 18600 }, { "epoch": 10.832363213038416, "grad_norm": 1.014459252357483, "learning_rate": 8.261049118263971e-05, "loss": 0.3944, "step": 18610 }, { "epoch": 10.838183934807915, "grad_norm": 1.0700030326843262, "learning_rate": 8.258959633464619e-05, "loss": 0.3799, "step": 18620 }, { "epoch": 10.844004656577416, "grad_norm": 0.9720134735107422, "learning_rate": 8.256869158679377e-05, "loss": 0.382, "step": 18630 }, { "epoch": 10.849825378346916, "grad_norm": 0.880643904209137, "learning_rate": 8.254777694543278e-05, "loss": 0.3782, "step": 18640 }, { "epoch": 10.855646100116415, "grad_norm": 1.030015230178833, "learning_rate": 8.252685241691651e-05, "loss": 0.3818, "step": 18650 }, { "epoch": 10.861466821885914, "grad_norm": 1.0706971883773804, "learning_rate": 8.250591800760133e-05, "loss": 0.3924, "step": 18660 }, { "epoch": 10.867287543655413, "grad_norm": 1.1347997188568115, "learning_rate": 8.248497372384649e-05, "loss": 0.4151, "step": 18670 }, { "epoch": 10.873108265424912, "grad_norm": 0.9392638802528381, "learning_rate": 8.246401957201437e-05, "loss": 0.3836, "step": 18680 }, { "epoch": 10.878928987194412, "grad_norm": 0.9473598003387451, "learning_rate": 8.244305555847027e-05, "loss": 0.3949, "step": 18690 }, { "epoch": 10.884749708963911, "grad_norm": 0.9078813791275024, "learning_rate": 8.24220816895825e-05, "loss": 0.3903, "step": 18700 }, { "epoch": 10.89057043073341, "grad_norm": 0.9568220973014832, "learning_rate": 8.240109797172237e-05, "loss": 0.3879, "step": 18710 }, { "epoch": 10.896391152502911, "grad_norm": 0.887516438961029, "learning_rate": 8.238010441126416e-05, "loss": 0.3817, "step": 18720 }, { "epoch": 10.90221187427241, "grad_norm": 0.8363648056983948, "learning_rate": 8.23591010145852e-05, "loss": 0.37, "step": 18730 }, { "epoch": 10.90803259604191, "grad_norm": 1.037353515625, "learning_rate": 8.233808778806571e-05, "loss": 0.3878, "step": 18740 }, { "epoch": 10.913853317811409, "grad_norm": 0.9446734189987183, "learning_rate": 8.231706473808903e-05, "loss": 0.372, "step": 18750 }, { "epoch": 10.919674039580908, "grad_norm": 0.891288697719574, "learning_rate": 8.229603187104133e-05, "loss": 0.371, "step": 18760 }, { "epoch": 10.925494761350407, "grad_norm": 0.886999249458313, "learning_rate": 8.22749891933119e-05, "loss": 0.3813, "step": 18770 }, { "epoch": 10.931315483119906, "grad_norm": 0.9511243104934692, "learning_rate": 8.225393671129291e-05, "loss": 0.3867, "step": 18780 }, { "epoch": 10.937136204889406, "grad_norm": 0.9742743372917175, "learning_rate": 8.223287443137957e-05, "loss": 0.3869, "step": 18790 }, { "epoch": 10.942956926658905, "grad_norm": 1.052055835723877, "learning_rate": 8.221180235997004e-05, "loss": 0.3797, "step": 18800 }, { "epoch": 10.948777648428406, "grad_norm": 0.886188268661499, "learning_rate": 8.219072050346544e-05, "loss": 0.396, "step": 18810 }, { "epoch": 10.954598370197905, "grad_norm": 0.8932410478591919, "learning_rate": 8.216962886826992e-05, "loss": 0.3938, "step": 18820 }, { "epoch": 10.960419091967404, "grad_norm": 0.9877529144287109, "learning_rate": 8.214852746079054e-05, "loss": 0.38, "step": 18830 }, { "epoch": 10.966239813736903, "grad_norm": 0.8746454119682312, "learning_rate": 8.212741628743732e-05, "loss": 0.376, "step": 18840 }, { "epoch": 10.972060535506403, "grad_norm": 1.0313780307769775, "learning_rate": 8.210629535462333e-05, "loss": 0.3909, "step": 18850 }, { "epoch": 10.977881257275902, "grad_norm": 0.8965948820114136, "learning_rate": 8.208516466876453e-05, "loss": 0.3731, "step": 18860 }, { "epoch": 10.983701979045401, "grad_norm": 0.9084764719009399, "learning_rate": 8.206402423627986e-05, "loss": 0.3787, "step": 18870 }, { "epoch": 10.9895227008149, "grad_norm": 0.9327656626701355, "learning_rate": 8.204287406359124e-05, "loss": 0.3858, "step": 18880 }, { "epoch": 10.995343422584401, "grad_norm": 0.9394635558128357, "learning_rate": 8.20217141571235e-05, "loss": 0.3733, "step": 18890 }, { "epoch": 11.0011641443539, "grad_norm": 0.9663826823234558, "learning_rate": 8.200054452330449e-05, "loss": 0.3901, "step": 18900 }, { "epoch": 11.0069848661234, "grad_norm": 0.9967271089553833, "learning_rate": 8.197936516856499e-05, "loss": 0.375, "step": 18910 }, { "epoch": 11.012805587892899, "grad_norm": 1.0349831581115723, "learning_rate": 8.195817609933871e-05, "loss": 0.4, "step": 18920 }, { "epoch": 11.018626309662398, "grad_norm": 0.9117758274078369, "learning_rate": 8.193697732206233e-05, "loss": 0.3939, "step": 18930 }, { "epoch": 11.024447031431897, "grad_norm": 0.9306642413139343, "learning_rate": 8.19157688431755e-05, "loss": 0.3904, "step": 18940 }, { "epoch": 11.030267753201397, "grad_norm": 0.9806511402130127, "learning_rate": 8.189455066912077e-05, "loss": 0.3782, "step": 18950 }, { "epoch": 11.036088474970896, "grad_norm": 0.9662925004959106, "learning_rate": 8.187332280634369e-05, "loss": 0.3904, "step": 18960 }, { "epoch": 11.041909196740395, "grad_norm": 0.9084495902061462, "learning_rate": 8.18520852612927e-05, "loss": 0.3833, "step": 18970 }, { "epoch": 11.047729918509896, "grad_norm": 0.8785226941108704, "learning_rate": 8.183083804041921e-05, "loss": 0.3808, "step": 18980 }, { "epoch": 11.053550640279395, "grad_norm": 0.9635068774223328, "learning_rate": 8.180958115017757e-05, "loss": 0.372, "step": 18990 }, { "epoch": 11.059371362048894, "grad_norm": 0.9145845174789429, "learning_rate": 8.178831459702505e-05, "loss": 0.3729, "step": 19000 }, { "epoch": 11.065192083818394, "grad_norm": 1.061597466468811, "learning_rate": 8.17670383874219e-05, "loss": 0.3867, "step": 19010 }, { "epoch": 11.071012805587893, "grad_norm": 1.0363633632659912, "learning_rate": 8.174575252783124e-05, "loss": 0.3904, "step": 19020 }, { "epoch": 11.076833527357392, "grad_norm": 0.9618642330169678, "learning_rate": 8.172445702471914e-05, "loss": 0.3914, "step": 19030 }, { "epoch": 11.082654249126891, "grad_norm": 0.9475641846656799, "learning_rate": 8.170315188455466e-05, "loss": 0.3843, "step": 19040 }, { "epoch": 11.08847497089639, "grad_norm": 0.980542778968811, "learning_rate": 8.168183711380969e-05, "loss": 0.3779, "step": 19050 }, { "epoch": 11.09429569266589, "grad_norm": 1.0533289909362793, "learning_rate": 8.166051271895913e-05, "loss": 0.3716, "step": 19060 }, { "epoch": 11.10011641443539, "grad_norm": 0.9327096343040466, "learning_rate": 8.163917870648075e-05, "loss": 0.3834, "step": 19070 }, { "epoch": 11.10593713620489, "grad_norm": 0.9519631862640381, "learning_rate": 8.161783508285526e-05, "loss": 0.3824, "step": 19080 }, { "epoch": 11.11175785797439, "grad_norm": 1.011099100112915, "learning_rate": 8.159648185456628e-05, "loss": 0.3839, "step": 19090 }, { "epoch": 11.117578579743888, "grad_norm": 1.0122950077056885, "learning_rate": 8.157511902810038e-05, "loss": 0.3857, "step": 19100 }, { "epoch": 11.123399301513388, "grad_norm": 0.8921992182731628, "learning_rate": 8.155374660994701e-05, "loss": 0.3769, "step": 19110 }, { "epoch": 11.129220023282887, "grad_norm": 0.898792028427124, "learning_rate": 8.153236460659857e-05, "loss": 0.3785, "step": 19120 }, { "epoch": 11.135040745052386, "grad_norm": 1.0053539276123047, "learning_rate": 8.151097302455031e-05, "loss": 0.3859, "step": 19130 }, { "epoch": 11.140861466821885, "grad_norm": 0.8877506256103516, "learning_rate": 8.148957187030044e-05, "loss": 0.3818, "step": 19140 }, { "epoch": 11.146682188591384, "grad_norm": 1.0174238681793213, "learning_rate": 8.146816115035006e-05, "loss": 0.3971, "step": 19150 }, { "epoch": 11.152502910360885, "grad_norm": 1.01798677444458, "learning_rate": 8.14467408712032e-05, "loss": 0.393, "step": 19160 }, { "epoch": 11.158323632130385, "grad_norm": 0.9066681861877441, "learning_rate": 8.142531103936678e-05, "loss": 0.3857, "step": 19170 }, { "epoch": 11.164144353899884, "grad_norm": 1.0280799865722656, "learning_rate": 8.14038716613506e-05, "loss": 0.3802, "step": 19180 }, { "epoch": 11.169965075669383, "grad_norm": 0.8910545706748962, "learning_rate": 8.138242274366736e-05, "loss": 0.3827, "step": 19190 }, { "epoch": 11.175785797438882, "grad_norm": 0.918465256690979, "learning_rate": 8.136096429283271e-05, "loss": 0.3795, "step": 19200 }, { "epoch": 11.181606519208382, "grad_norm": 0.9243491291999817, "learning_rate": 8.133949631536515e-05, "loss": 0.3709, "step": 19210 }, { "epoch": 11.18742724097788, "grad_norm": 0.9887418150901794, "learning_rate": 8.131801881778607e-05, "loss": 0.3902, "step": 19220 }, { "epoch": 11.19324796274738, "grad_norm": 0.9314790368080139, "learning_rate": 8.129653180661978e-05, "loss": 0.3777, "step": 19230 }, { "epoch": 11.199068684516881, "grad_norm": 0.9157055616378784, "learning_rate": 8.127503528839346e-05, "loss": 0.3843, "step": 19240 }, { "epoch": 11.20488940628638, "grad_norm": 0.8590747117996216, "learning_rate": 8.125352926963721e-05, "loss": 0.3923, "step": 19250 }, { "epoch": 11.21071012805588, "grad_norm": 0.9029837250709534, "learning_rate": 8.123201375688395e-05, "loss": 0.3847, "step": 19260 }, { "epoch": 11.216530849825379, "grad_norm": 0.9490057229995728, "learning_rate": 8.121048875666954e-05, "loss": 0.3772, "step": 19270 }, { "epoch": 11.222351571594878, "grad_norm": 0.8583550453186035, "learning_rate": 8.118895427553274e-05, "loss": 0.3841, "step": 19280 }, { "epoch": 11.228172293364377, "grad_norm": 1.0221377611160278, "learning_rate": 8.116741032001511e-05, "loss": 0.3945, "step": 19290 }, { "epoch": 11.233993015133876, "grad_norm": 0.9121062755584717, "learning_rate": 8.114585689666114e-05, "loss": 0.3826, "step": 19300 }, { "epoch": 11.239813736903375, "grad_norm": 0.9380261898040771, "learning_rate": 8.112429401201821e-05, "loss": 0.3817, "step": 19310 }, { "epoch": 11.245634458672875, "grad_norm": 0.8931359052658081, "learning_rate": 8.110272167263656e-05, "loss": 0.3704, "step": 19320 }, { "epoch": 11.251455180442376, "grad_norm": 0.925885796546936, "learning_rate": 8.108113988506929e-05, "loss": 0.3838, "step": 19330 }, { "epoch": 11.257275902211875, "grad_norm": 0.8683913946151733, "learning_rate": 8.105954865587235e-05, "loss": 0.382, "step": 19340 }, { "epoch": 11.263096623981374, "grad_norm": 0.9885071516036987, "learning_rate": 8.103794799160463e-05, "loss": 0.3872, "step": 19350 }, { "epoch": 11.268917345750873, "grad_norm": 0.9206083416938782, "learning_rate": 8.101633789882781e-05, "loss": 0.3779, "step": 19360 }, { "epoch": 11.274738067520373, "grad_norm": 0.9173790216445923, "learning_rate": 8.099471838410648e-05, "loss": 0.3789, "step": 19370 }, { "epoch": 11.280558789289872, "grad_norm": 0.9504477977752686, "learning_rate": 8.097308945400806e-05, "loss": 0.3794, "step": 19380 }, { "epoch": 11.286379511059371, "grad_norm": 0.9435696005821228, "learning_rate": 8.095145111510288e-05, "loss": 0.3846, "step": 19390 }, { "epoch": 11.29220023282887, "grad_norm": 0.9466186761856079, "learning_rate": 8.092980337396406e-05, "loss": 0.3843, "step": 19400 }, { "epoch": 11.29802095459837, "grad_norm": 0.9327443242073059, "learning_rate": 8.090814623716763e-05, "loss": 0.3949, "step": 19410 }, { "epoch": 11.30384167636787, "grad_norm": 0.8897565603256226, "learning_rate": 8.088647971129246e-05, "loss": 0.3756, "step": 19420 }, { "epoch": 11.30966239813737, "grad_norm": 0.8757994174957275, "learning_rate": 8.086480380292026e-05, "loss": 0.3805, "step": 19430 }, { "epoch": 11.315483119906869, "grad_norm": 0.9435701966285706, "learning_rate": 8.084311851863562e-05, "loss": 0.3898, "step": 19440 }, { "epoch": 11.321303841676368, "grad_norm": 0.9447243213653564, "learning_rate": 8.082142386502591e-05, "loss": 0.3757, "step": 19450 }, { "epoch": 11.327124563445867, "grad_norm": 0.9003065228462219, "learning_rate": 8.079971984868145e-05, "loss": 0.3679, "step": 19460 }, { "epoch": 11.332945285215366, "grad_norm": 0.8638840317726135, "learning_rate": 8.077800647619532e-05, "loss": 0.3902, "step": 19470 }, { "epoch": 11.338766006984866, "grad_norm": 0.9713338017463684, "learning_rate": 8.075628375416345e-05, "loss": 0.3969, "step": 19480 }, { "epoch": 11.344586728754365, "grad_norm": 0.9276121854782104, "learning_rate": 8.073455168918464e-05, "loss": 0.39, "step": 19490 }, { "epoch": 11.350407450523864, "grad_norm": 0.8921542167663574, "learning_rate": 8.071281028786055e-05, "loss": 0.3769, "step": 19500 }, { "epoch": 11.356228172293365, "grad_norm": 0.923560619354248, "learning_rate": 8.069105955679562e-05, "loss": 0.3748, "step": 19510 }, { "epoch": 11.362048894062864, "grad_norm": 1.015700340270996, "learning_rate": 8.066929950259713e-05, "loss": 0.3823, "step": 19520 }, { "epoch": 11.367869615832364, "grad_norm": 1.024211049079895, "learning_rate": 8.064753013187522e-05, "loss": 0.3985, "step": 19530 }, { "epoch": 11.373690337601863, "grad_norm": 0.9669106006622314, "learning_rate": 8.062575145124289e-05, "loss": 0.3802, "step": 19540 }, { "epoch": 11.379511059371362, "grad_norm": 1.0071942806243896, "learning_rate": 8.060396346731587e-05, "loss": 0.393, "step": 19550 }, { "epoch": 11.385331781140861, "grad_norm": 0.8817033171653748, "learning_rate": 8.058216618671281e-05, "loss": 0.3818, "step": 19560 }, { "epoch": 11.39115250291036, "grad_norm": 0.8806056976318359, "learning_rate": 8.056035961605514e-05, "loss": 0.3931, "step": 19570 }, { "epoch": 11.39697322467986, "grad_norm": 0.9446294903755188, "learning_rate": 8.05385437619671e-05, "loss": 0.3906, "step": 19580 }, { "epoch": 11.40279394644936, "grad_norm": 0.8858463764190674, "learning_rate": 8.05167186310758e-05, "loss": 0.379, "step": 19590 }, { "epoch": 11.40861466821886, "grad_norm": 0.8987346887588501, "learning_rate": 8.049488423001113e-05, "loss": 0.3799, "step": 19600 }, { "epoch": 11.414435389988359, "grad_norm": 0.96622633934021, "learning_rate": 8.047304056540581e-05, "loss": 0.3654, "step": 19610 }, { "epoch": 11.420256111757858, "grad_norm": 0.9913033246994019, "learning_rate": 8.045118764389534e-05, "loss": 0.3719, "step": 19620 }, { "epoch": 11.426076833527357, "grad_norm": 0.9111454486846924, "learning_rate": 8.042932547211809e-05, "loss": 0.3913, "step": 19630 }, { "epoch": 11.431897555296857, "grad_norm": 0.9489163756370544, "learning_rate": 8.04074540567152e-05, "loss": 0.3634, "step": 19640 }, { "epoch": 11.437718277066356, "grad_norm": 0.957499086856842, "learning_rate": 8.038557340433063e-05, "loss": 0.378, "step": 19650 }, { "epoch": 11.443538998835855, "grad_norm": 0.9581336379051208, "learning_rate": 8.036368352161115e-05, "loss": 0.3942, "step": 19660 }, { "epoch": 11.449359720605354, "grad_norm": 0.9869627356529236, "learning_rate": 8.034178441520633e-05, "loss": 0.3982, "step": 19670 }, { "epoch": 11.455180442374855, "grad_norm": 0.9732761383056641, "learning_rate": 8.031987609176852e-05, "loss": 0.3658, "step": 19680 }, { "epoch": 11.461001164144355, "grad_norm": 1.0036364793777466, "learning_rate": 8.02979585579529e-05, "loss": 0.3795, "step": 19690 }, { "epoch": 11.466821885913854, "grad_norm": 0.9535634517669678, "learning_rate": 8.027603182041745e-05, "loss": 0.3889, "step": 19700 }, { "epoch": 11.472642607683353, "grad_norm": 0.9666116833686829, "learning_rate": 8.025409588582292e-05, "loss": 0.375, "step": 19710 }, { "epoch": 11.478463329452852, "grad_norm": 0.9186115264892578, "learning_rate": 8.023215076083288e-05, "loss": 0.39, "step": 19720 }, { "epoch": 11.484284051222351, "grad_norm": 0.9144423007965088, "learning_rate": 8.021019645211367e-05, "loss": 0.3833, "step": 19730 }, { "epoch": 11.49010477299185, "grad_norm": 0.957166314125061, "learning_rate": 8.018823296633441e-05, "loss": 0.3818, "step": 19740 }, { "epoch": 11.49592549476135, "grad_norm": 0.918369472026825, "learning_rate": 8.016626031016708e-05, "loss": 0.3778, "step": 19750 }, { "epoch": 11.501746216530849, "grad_norm": 0.9696377515792847, "learning_rate": 8.014427849028636e-05, "loss": 0.376, "step": 19760 }, { "epoch": 11.50756693830035, "grad_norm": 0.9965868592262268, "learning_rate": 8.012228751336974e-05, "loss": 0.3889, "step": 19770 }, { "epoch": 11.51338766006985, "grad_norm": 0.9389261603355408, "learning_rate": 8.01002873860975e-05, "loss": 0.3845, "step": 19780 }, { "epoch": 11.519208381839348, "grad_norm": 0.9288774132728577, "learning_rate": 8.00782781151527e-05, "loss": 0.3766, "step": 19790 }, { "epoch": 11.525029103608848, "grad_norm": 0.9013659358024597, "learning_rate": 8.005625970722119e-05, "loss": 0.3768, "step": 19800 }, { "epoch": 11.530849825378347, "grad_norm": 0.9434489607810974, "learning_rate": 8.003423216899158e-05, "loss": 0.3902, "step": 19810 }, { "epoch": 11.536670547147846, "grad_norm": 0.9885954260826111, "learning_rate": 8.001219550715522e-05, "loss": 0.3811, "step": 19820 }, { "epoch": 11.542491268917345, "grad_norm": 0.908875584602356, "learning_rate": 7.999014972840632e-05, "loss": 0.3873, "step": 19830 }, { "epoch": 11.548311990686845, "grad_norm": 0.9262266755104065, "learning_rate": 7.996809483944174e-05, "loss": 0.3885, "step": 19840 }, { "epoch": 11.554132712456344, "grad_norm": 0.9468299746513367, "learning_rate": 7.994603084696124e-05, "loss": 0.3953, "step": 19850 }, { "epoch": 11.559953434225845, "grad_norm": 0.9799118638038635, "learning_rate": 7.992395775766724e-05, "loss": 0.3936, "step": 19860 }, { "epoch": 11.565774155995344, "grad_norm": 0.8854632377624512, "learning_rate": 7.990187557826497e-05, "loss": 0.3873, "step": 19870 }, { "epoch": 11.571594877764843, "grad_norm": 0.856124997138977, "learning_rate": 7.987978431546242e-05, "loss": 0.3702, "step": 19880 }, { "epoch": 11.577415599534342, "grad_norm": 0.9848445653915405, "learning_rate": 7.985768397597031e-05, "loss": 0.3844, "step": 19890 }, { "epoch": 11.583236321303842, "grad_norm": 0.9193136096000671, "learning_rate": 7.983557456650216e-05, "loss": 0.381, "step": 19900 }, { "epoch": 11.58905704307334, "grad_norm": 0.951411247253418, "learning_rate": 7.981345609377422e-05, "loss": 0.3638, "step": 19910 }, { "epoch": 11.59487776484284, "grad_norm": 0.9582989811897278, "learning_rate": 7.97913285645055e-05, "loss": 0.3837, "step": 19920 }, { "epoch": 11.60069848661234, "grad_norm": 0.9096288084983826, "learning_rate": 7.976919198541776e-05, "loss": 0.3676, "step": 19930 }, { "epoch": 11.60651920838184, "grad_norm": 0.8624553084373474, "learning_rate": 7.974704636323548e-05, "loss": 0.3765, "step": 19940 }, { "epoch": 11.61233993015134, "grad_norm": 0.8865863680839539, "learning_rate": 7.972489170468597e-05, "loss": 0.3846, "step": 19950 }, { "epoch": 11.618160651920839, "grad_norm": 1.0035046339035034, "learning_rate": 7.970272801649918e-05, "loss": 0.381, "step": 19960 }, { "epoch": 11.623981373690338, "grad_norm": 0.9466930031776428, "learning_rate": 7.96805553054079e-05, "loss": 0.3829, "step": 19970 }, { "epoch": 11.629802095459837, "grad_norm": 0.9089265465736389, "learning_rate": 7.965837357814756e-05, "loss": 0.3685, "step": 19980 }, { "epoch": 11.635622817229336, "grad_norm": 0.9486823678016663, "learning_rate": 7.963618284145643e-05, "loss": 0.3911, "step": 19990 }, { "epoch": 11.641443538998836, "grad_norm": 0.9111713171005249, "learning_rate": 7.961398310207544e-05, "loss": 0.366, "step": 20000 }, { "epoch": 11.647264260768335, "grad_norm": 0.9083071351051331, "learning_rate": 7.95917743667483e-05, "loss": 0.3721, "step": 20010 }, { "epoch": 11.653084982537834, "grad_norm": 1.0182963609695435, "learning_rate": 7.956955664222144e-05, "loss": 0.3742, "step": 20020 }, { "epoch": 11.658905704307333, "grad_norm": 0.9165193438529968, "learning_rate": 7.954732993524399e-05, "loss": 0.3824, "step": 20030 }, { "epoch": 11.664726426076834, "grad_norm": 0.9419496655464172, "learning_rate": 7.952509425256786e-05, "loss": 0.3749, "step": 20040 }, { "epoch": 11.670547147846333, "grad_norm": 0.7483842968940735, "learning_rate": 7.950284960094767e-05, "loss": 0.3732, "step": 20050 }, { "epoch": 11.676367869615833, "grad_norm": 0.9486749172210693, "learning_rate": 7.948059598714076e-05, "loss": 0.391, "step": 20060 }, { "epoch": 11.682188591385332, "grad_norm": 1.104118824005127, "learning_rate": 7.945833341790717e-05, "loss": 0.3814, "step": 20070 }, { "epoch": 11.688009313154831, "grad_norm": 1.0248515605926514, "learning_rate": 7.94360619000097e-05, "loss": 0.3735, "step": 20080 }, { "epoch": 11.69383003492433, "grad_norm": 0.9341338872909546, "learning_rate": 7.941378144021381e-05, "loss": 0.3766, "step": 20090 }, { "epoch": 11.69965075669383, "grad_norm": 0.9124895334243774, "learning_rate": 7.939149204528777e-05, "loss": 0.3849, "step": 20100 }, { "epoch": 11.705471478463329, "grad_norm": 0.9156284928321838, "learning_rate": 7.936919372200246e-05, "loss": 0.3712, "step": 20110 }, { "epoch": 11.71129220023283, "grad_norm": 0.8602609634399414, "learning_rate": 7.934688647713158e-05, "loss": 0.3702, "step": 20120 }, { "epoch": 11.717112922002329, "grad_norm": 0.9604485034942627, "learning_rate": 7.932457031745143e-05, "loss": 0.3679, "step": 20130 }, { "epoch": 11.722933643771828, "grad_norm": 0.9936156868934631, "learning_rate": 7.930224524974108e-05, "loss": 0.3904, "step": 20140 }, { "epoch": 11.728754365541327, "grad_norm": 0.9715657234191895, "learning_rate": 7.927991128078232e-05, "loss": 0.3803, "step": 20150 }, { "epoch": 11.734575087310827, "grad_norm": 1.0216500759124756, "learning_rate": 7.925756841735958e-05, "loss": 0.378, "step": 20160 }, { "epoch": 11.740395809080326, "grad_norm": 0.9652760028839111, "learning_rate": 7.923521666626008e-05, "loss": 0.3741, "step": 20170 }, { "epoch": 11.746216530849825, "grad_norm": 0.9215492606163025, "learning_rate": 7.921285603427366e-05, "loss": 0.3703, "step": 20180 }, { "epoch": 11.752037252619324, "grad_norm": 0.9420367479324341, "learning_rate": 7.91904865281929e-05, "loss": 0.3858, "step": 20190 }, { "epoch": 11.757857974388823, "grad_norm": 0.9639178514480591, "learning_rate": 7.916810815481307e-05, "loss": 0.3886, "step": 20200 }, { "epoch": 11.763678696158324, "grad_norm": 1.0106834173202515, "learning_rate": 7.914572092093211e-05, "loss": 0.3877, "step": 20210 }, { "epoch": 11.769499417927824, "grad_norm": 0.931984543800354, "learning_rate": 7.912332483335068e-05, "loss": 0.3872, "step": 20220 }, { "epoch": 11.775320139697323, "grad_norm": 0.9370011687278748, "learning_rate": 7.910091989887213e-05, "loss": 0.3733, "step": 20230 }, { "epoch": 11.781140861466822, "grad_norm": 0.9416025280952454, "learning_rate": 7.907850612430248e-05, "loss": 0.3948, "step": 20240 }, { "epoch": 11.786961583236321, "grad_norm": 0.9113908410072327, "learning_rate": 7.905608351645044e-05, "loss": 0.3636, "step": 20250 }, { "epoch": 11.79278230500582, "grad_norm": 0.9033636450767517, "learning_rate": 7.90336520821274e-05, "loss": 0.3919, "step": 20260 }, { "epoch": 11.79860302677532, "grad_norm": 0.9146940112113953, "learning_rate": 7.901121182814746e-05, "loss": 0.3954, "step": 20270 }, { "epoch": 11.804423748544819, "grad_norm": 0.9198600649833679, "learning_rate": 7.898876276132736e-05, "loss": 0.3713, "step": 20280 }, { "epoch": 11.81024447031432, "grad_norm": 0.9694291949272156, "learning_rate": 7.896630488848654e-05, "loss": 0.3723, "step": 20290 }, { "epoch": 11.81606519208382, "grad_norm": 0.8769311904907227, "learning_rate": 7.89438382164471e-05, "loss": 0.3816, "step": 20300 }, { "epoch": 11.821885913853318, "grad_norm": 0.8313193917274475, "learning_rate": 7.892136275203383e-05, "loss": 0.3695, "step": 20310 }, { "epoch": 11.827706635622818, "grad_norm": 0.8415419459342957, "learning_rate": 7.889887850207418e-05, "loss": 0.3716, "step": 20320 }, { "epoch": 11.833527357392317, "grad_norm": 0.9024835824966431, "learning_rate": 7.887638547339827e-05, "loss": 0.3833, "step": 20330 }, { "epoch": 11.839348079161816, "grad_norm": 0.9867492318153381, "learning_rate": 7.885388367283891e-05, "loss": 0.398, "step": 20340 }, { "epoch": 11.845168800931315, "grad_norm": 0.981570303440094, "learning_rate": 7.88313731072315e-05, "loss": 0.3775, "step": 20350 }, { "epoch": 11.850989522700814, "grad_norm": 0.95677649974823, "learning_rate": 7.88088537834142e-05, "loss": 0.389, "step": 20360 }, { "epoch": 11.856810244470314, "grad_norm": 0.8787667155265808, "learning_rate": 7.878632570822778e-05, "loss": 0.3666, "step": 20370 }, { "epoch": 11.862630966239813, "grad_norm": 0.9012565016746521, "learning_rate": 7.876378888851567e-05, "loss": 0.3669, "step": 20380 }, { "epoch": 11.868451688009314, "grad_norm": 0.8726131916046143, "learning_rate": 7.874124333112396e-05, "loss": 0.3745, "step": 20390 }, { "epoch": 11.874272409778813, "grad_norm": 0.8699089884757996, "learning_rate": 7.871868904290138e-05, "loss": 0.3719, "step": 20400 }, { "epoch": 11.880093131548312, "grad_norm": 1.0195887088775635, "learning_rate": 7.869612603069935e-05, "loss": 0.381, "step": 20410 }, { "epoch": 11.885913853317811, "grad_norm": 0.918847382068634, "learning_rate": 7.867355430137192e-05, "loss": 0.3617, "step": 20420 }, { "epoch": 11.89173457508731, "grad_norm": 0.8652060627937317, "learning_rate": 7.865097386177577e-05, "loss": 0.3846, "step": 20430 }, { "epoch": 11.89755529685681, "grad_norm": 0.95042484998703, "learning_rate": 7.862838471877023e-05, "loss": 0.3804, "step": 20440 }, { "epoch": 11.90337601862631, "grad_norm": 0.8833785653114319, "learning_rate": 7.860578687921731e-05, "loss": 0.3657, "step": 20450 }, { "epoch": 11.909196740395808, "grad_norm": 0.8490141034126282, "learning_rate": 7.858318034998164e-05, "loss": 0.3635, "step": 20460 }, { "epoch": 11.91501746216531, "grad_norm": 1.0065689086914062, "learning_rate": 7.856056513793046e-05, "loss": 0.375, "step": 20470 }, { "epoch": 11.920838183934809, "grad_norm": 0.8571712970733643, "learning_rate": 7.85379412499337e-05, "loss": 0.3769, "step": 20480 }, { "epoch": 11.926658905704308, "grad_norm": 0.8776944279670715, "learning_rate": 7.851530869286389e-05, "loss": 0.3763, "step": 20490 }, { "epoch": 11.932479627473807, "grad_norm": 1.0383249521255493, "learning_rate": 7.849266747359619e-05, "loss": 0.3769, "step": 20500 }, { "epoch": 11.938300349243306, "grad_norm": 0.852577269077301, "learning_rate": 7.847001759900843e-05, "loss": 0.3784, "step": 20510 }, { "epoch": 11.944121071012805, "grad_norm": 0.9667934775352478, "learning_rate": 7.844735907598102e-05, "loss": 0.378, "step": 20520 }, { "epoch": 11.949941792782305, "grad_norm": 0.989570140838623, "learning_rate": 7.842469191139703e-05, "loss": 0.3689, "step": 20530 }, { "epoch": 11.955762514551804, "grad_norm": 0.9717575907707214, "learning_rate": 7.840201611214215e-05, "loss": 0.386, "step": 20540 }, { "epoch": 11.961583236321303, "grad_norm": 0.8907134532928467, "learning_rate": 7.837933168510469e-05, "loss": 0.3751, "step": 20550 }, { "epoch": 11.967403958090804, "grad_norm": 0.9433079361915588, "learning_rate": 7.835663863717559e-05, "loss": 0.3801, "step": 20560 }, { "epoch": 11.973224679860303, "grad_norm": 0.9404760003089905, "learning_rate": 7.833393697524838e-05, "loss": 0.3811, "step": 20570 }, { "epoch": 11.979045401629802, "grad_norm": 0.8385111093521118, "learning_rate": 7.831122670621922e-05, "loss": 0.3863, "step": 20580 }, { "epoch": 11.984866123399302, "grad_norm": 0.8646472692489624, "learning_rate": 7.82885078369869e-05, "loss": 0.3649, "step": 20590 }, { "epoch": 11.990686845168801, "grad_norm": 0.9308244585990906, "learning_rate": 7.826578037445283e-05, "loss": 0.3802, "step": 20600 }, { "epoch": 11.9965075669383, "grad_norm": 0.9162536263465881, "learning_rate": 7.824304432552097e-05, "loss": 0.3888, "step": 20610 }, { "epoch": 12.0023282887078, "grad_norm": 0.9341192841529846, "learning_rate": 7.822029969709798e-05, "loss": 0.366, "step": 20620 }, { "epoch": 12.008149010477299, "grad_norm": 0.9536611437797546, "learning_rate": 7.819754649609306e-05, "loss": 0.3669, "step": 20630 }, { "epoch": 12.013969732246798, "grad_norm": 0.8486018776893616, "learning_rate": 7.817478472941802e-05, "loss": 0.3707, "step": 20640 }, { "epoch": 12.019790454016299, "grad_norm": 0.9193456768989563, "learning_rate": 7.815201440398727e-05, "loss": 0.3785, "step": 20650 }, { "epoch": 12.025611175785798, "grad_norm": 1.0759291648864746, "learning_rate": 7.812923552671789e-05, "loss": 0.3863, "step": 20660 }, { "epoch": 12.031431897555297, "grad_norm": 1.0832123756408691, "learning_rate": 7.810644810452945e-05, "loss": 0.388, "step": 20670 }, { "epoch": 12.037252619324796, "grad_norm": 0.9369824528694153, "learning_rate": 7.808365214434417e-05, "loss": 0.3798, "step": 20680 }, { "epoch": 12.043073341094296, "grad_norm": 0.861233651638031, "learning_rate": 7.80608476530869e-05, "loss": 0.3792, "step": 20690 }, { "epoch": 12.048894062863795, "grad_norm": 0.9358461499214172, "learning_rate": 7.8038034637685e-05, "loss": 0.3754, "step": 20700 }, { "epoch": 12.054714784633294, "grad_norm": 0.9281898140907288, "learning_rate": 7.801521310506848e-05, "loss": 0.3758, "step": 20710 }, { "epoch": 12.060535506402793, "grad_norm": 0.9410876035690308, "learning_rate": 7.799238306216994e-05, "loss": 0.3845, "step": 20720 }, { "epoch": 12.066356228172292, "grad_norm": 0.9527221918106079, "learning_rate": 7.796954451592448e-05, "loss": 0.3752, "step": 20730 }, { "epoch": 12.072176949941793, "grad_norm": 0.8250777125358582, "learning_rate": 7.794669747326992e-05, "loss": 0.3695, "step": 20740 }, { "epoch": 12.077997671711293, "grad_norm": 0.9375665783882141, "learning_rate": 7.792384194114654e-05, "loss": 0.3613, "step": 20750 }, { "epoch": 12.083818393480792, "grad_norm": 0.9111078381538391, "learning_rate": 7.790097792649729e-05, "loss": 0.3691, "step": 20760 }, { "epoch": 12.089639115250291, "grad_norm": 0.9265091419219971, "learning_rate": 7.787810543626762e-05, "loss": 0.3785, "step": 20770 }, { "epoch": 12.09545983701979, "grad_norm": 0.8234249949455261, "learning_rate": 7.785522447740558e-05, "loss": 0.3936, "step": 20780 }, { "epoch": 12.10128055878929, "grad_norm": 0.9051051735877991, "learning_rate": 7.783233505686182e-05, "loss": 0.3906, "step": 20790 }, { "epoch": 12.107101280558789, "grad_norm": 0.8203218579292297, "learning_rate": 7.780943718158955e-05, "loss": 0.3722, "step": 20800 }, { "epoch": 12.112922002328288, "grad_norm": 0.9490086436271667, "learning_rate": 7.778653085854453e-05, "loss": 0.3723, "step": 20810 }, { "epoch": 12.118742724097789, "grad_norm": 0.9237866401672363, "learning_rate": 7.77636160946851e-05, "loss": 0.3734, "step": 20820 }, { "epoch": 12.124563445867288, "grad_norm": 0.9686608910560608, "learning_rate": 7.774069289697215e-05, "loss": 0.3783, "step": 20830 }, { "epoch": 12.130384167636787, "grad_norm": 1.0005509853363037, "learning_rate": 7.771776127236913e-05, "loss": 0.372, "step": 20840 }, { "epoch": 12.136204889406287, "grad_norm": 0.8720837831497192, "learning_rate": 7.769482122784212e-05, "loss": 0.377, "step": 20850 }, { "epoch": 12.142025611175786, "grad_norm": 0.856606662273407, "learning_rate": 7.767187277035963e-05, "loss": 0.3567, "step": 20860 }, { "epoch": 12.147846332945285, "grad_norm": 0.823923647403717, "learning_rate": 7.764891590689285e-05, "loss": 0.3567, "step": 20870 }, { "epoch": 12.153667054714784, "grad_norm": 0.8576198816299438, "learning_rate": 7.762595064441542e-05, "loss": 0.3682, "step": 20880 }, { "epoch": 12.159487776484283, "grad_norm": 0.8980448842048645, "learning_rate": 7.760297698990362e-05, "loss": 0.3668, "step": 20890 }, { "epoch": 12.165308498253783, "grad_norm": 0.9948905110359192, "learning_rate": 7.757999495033623e-05, "loss": 0.3805, "step": 20900 }, { "epoch": 12.171129220023284, "grad_norm": 0.9001995921134949, "learning_rate": 7.755700453269456e-05, "loss": 0.3763, "step": 20910 }, { "epoch": 12.176949941792783, "grad_norm": 0.932870626449585, "learning_rate": 7.753400574396254e-05, "loss": 0.3759, "step": 20920 }, { "epoch": 12.182770663562282, "grad_norm": 0.9351015090942383, "learning_rate": 7.751099859112655e-05, "loss": 0.3638, "step": 20930 }, { "epoch": 12.188591385331781, "grad_norm": 0.9210125207901001, "learning_rate": 7.748798308117557e-05, "loss": 0.3746, "step": 20940 }, { "epoch": 12.19441210710128, "grad_norm": 0.8551166653633118, "learning_rate": 7.746495922110112e-05, "loss": 0.3624, "step": 20950 }, { "epoch": 12.20023282887078, "grad_norm": 0.9035927057266235, "learning_rate": 7.744192701789723e-05, "loss": 0.3832, "step": 20960 }, { "epoch": 12.206053550640279, "grad_norm": 0.912771999835968, "learning_rate": 7.741888647856046e-05, "loss": 0.3772, "step": 20970 }, { "epoch": 12.211874272409778, "grad_norm": 0.8874492645263672, "learning_rate": 7.739583761008994e-05, "loss": 0.3876, "step": 20980 }, { "epoch": 12.217694994179277, "grad_norm": 0.9218620657920837, "learning_rate": 7.73727804194873e-05, "loss": 0.3877, "step": 20990 }, { "epoch": 12.223515715948778, "grad_norm": 0.9140023589134216, "learning_rate": 7.734971491375671e-05, "loss": 0.3821, "step": 21000 }, { "epoch": 12.229336437718278, "grad_norm": 0.9010871648788452, "learning_rate": 7.732664109990485e-05, "loss": 0.3645, "step": 21010 }, { "epoch": 12.235157159487777, "grad_norm": 0.8812329769134521, "learning_rate": 7.730355898494095e-05, "loss": 0.3698, "step": 21020 }, { "epoch": 12.240977881257276, "grad_norm": 0.8319091796875, "learning_rate": 7.728046857587673e-05, "loss": 0.3671, "step": 21030 }, { "epoch": 12.246798603026775, "grad_norm": 0.9173668622970581, "learning_rate": 7.725736987972647e-05, "loss": 0.3695, "step": 21040 }, { "epoch": 12.252619324796274, "grad_norm": 0.9318044781684875, "learning_rate": 7.723426290350691e-05, "loss": 0.3796, "step": 21050 }, { "epoch": 12.258440046565774, "grad_norm": 0.9237673878669739, "learning_rate": 7.721114765423736e-05, "loss": 0.372, "step": 21060 }, { "epoch": 12.264260768335273, "grad_norm": 0.9540320634841919, "learning_rate": 7.718802413893963e-05, "loss": 0.3758, "step": 21070 }, { "epoch": 12.270081490104772, "grad_norm": 0.9259331822395325, "learning_rate": 7.716489236463802e-05, "loss": 0.3873, "step": 21080 }, { "epoch": 12.275902211874273, "grad_norm": 0.898345410823822, "learning_rate": 7.714175233835936e-05, "loss": 0.3762, "step": 21090 }, { "epoch": 12.281722933643772, "grad_norm": 0.8397708535194397, "learning_rate": 7.711860406713299e-05, "loss": 0.3828, "step": 21100 }, { "epoch": 12.287543655413272, "grad_norm": 0.8805057406425476, "learning_rate": 7.70954475579907e-05, "loss": 0.3795, "step": 21110 }, { "epoch": 12.29336437718277, "grad_norm": 0.840254008769989, "learning_rate": 7.707228281796688e-05, "loss": 0.3719, "step": 21120 }, { "epoch": 12.29918509895227, "grad_norm": 0.8880007863044739, "learning_rate": 7.704910985409833e-05, "loss": 0.364, "step": 21130 }, { "epoch": 12.30500582072177, "grad_norm": 0.9736623167991638, "learning_rate": 7.702592867342439e-05, "loss": 0.3678, "step": 21140 }, { "epoch": 12.310826542491268, "grad_norm": 1.0823297500610352, "learning_rate": 7.700273928298691e-05, "loss": 0.3728, "step": 21150 }, { "epoch": 12.316647264260768, "grad_norm": 0.991179883480072, "learning_rate": 7.697954168983021e-05, "loss": 0.3666, "step": 21160 }, { "epoch": 12.322467986030269, "grad_norm": 1.051215648651123, "learning_rate": 7.695633590100109e-05, "loss": 0.3782, "step": 21170 }, { "epoch": 12.328288707799768, "grad_norm": 0.9286991953849792, "learning_rate": 7.693312192354886e-05, "loss": 0.3636, "step": 21180 }, { "epoch": 12.334109429569267, "grad_norm": 0.8849918246269226, "learning_rate": 7.690989976452532e-05, "loss": 0.3746, "step": 21190 }, { "epoch": 12.339930151338766, "grad_norm": 0.9892093539237976, "learning_rate": 7.688666943098475e-05, "loss": 0.3811, "step": 21200 }, { "epoch": 12.345750873108265, "grad_norm": 0.9606562852859497, "learning_rate": 7.686343092998389e-05, "loss": 0.3689, "step": 21210 }, { "epoch": 12.351571594877765, "grad_norm": 0.9262534379959106, "learning_rate": 7.684018426858202e-05, "loss": 0.3743, "step": 21220 }, { "epoch": 12.357392316647264, "grad_norm": 0.8906402587890625, "learning_rate": 7.681692945384084e-05, "loss": 0.3721, "step": 21230 }, { "epoch": 12.363213038416763, "grad_norm": 0.9850530028343201, "learning_rate": 7.679366649282456e-05, "loss": 0.3765, "step": 21240 }, { "epoch": 12.369033760186262, "grad_norm": 0.8414514660835266, "learning_rate": 7.677039539259983e-05, "loss": 0.3823, "step": 21250 }, { "epoch": 12.374854481955763, "grad_norm": 0.9963126182556152, "learning_rate": 7.674711616023581e-05, "loss": 0.3803, "step": 21260 }, { "epoch": 12.380675203725263, "grad_norm": 0.9133096933364868, "learning_rate": 7.672382880280413e-05, "loss": 0.3849, "step": 21270 }, { "epoch": 12.386495925494762, "grad_norm": 0.8474382162094116, "learning_rate": 7.670053332737885e-05, "loss": 0.382, "step": 21280 }, { "epoch": 12.392316647264261, "grad_norm": 0.9633595943450928, "learning_rate": 7.667722974103654e-05, "loss": 0.366, "step": 21290 }, { "epoch": 12.39813736903376, "grad_norm": 0.9064950942993164, "learning_rate": 7.66539180508562e-05, "loss": 0.3805, "step": 21300 }, { "epoch": 12.40395809080326, "grad_norm": 0.9336495995521545, "learning_rate": 7.663059826391932e-05, "loss": 0.3553, "step": 21310 }, { "epoch": 12.409778812572759, "grad_norm": 0.9385684728622437, "learning_rate": 7.660727038730981e-05, "loss": 0.3739, "step": 21320 }, { "epoch": 12.415599534342258, "grad_norm": 1.090997576713562, "learning_rate": 7.65839344281141e-05, "loss": 0.3718, "step": 21330 }, { "epoch": 12.421420256111757, "grad_norm": 0.9184780120849609, "learning_rate": 7.656059039342101e-05, "loss": 0.3794, "step": 21340 }, { "epoch": 12.427240977881258, "grad_norm": 0.8266398906707764, "learning_rate": 7.653723829032187e-05, "loss": 0.366, "step": 21350 }, { "epoch": 12.433061699650757, "grad_norm": 0.9776253700256348, "learning_rate": 7.65138781259104e-05, "loss": 0.3925, "step": 21360 }, { "epoch": 12.438882421420256, "grad_norm": 0.9026396870613098, "learning_rate": 7.649050990728279e-05, "loss": 0.3782, "step": 21370 }, { "epoch": 12.444703143189756, "grad_norm": 0.9742230772972107, "learning_rate": 7.646713364153774e-05, "loss": 0.3787, "step": 21380 }, { "epoch": 12.450523864959255, "grad_norm": 1.0216903686523438, "learning_rate": 7.64437493357763e-05, "loss": 0.3742, "step": 21390 }, { "epoch": 12.456344586728754, "grad_norm": 0.9246240258216858, "learning_rate": 7.642035699710202e-05, "loss": 0.385, "step": 21400 }, { "epoch": 12.462165308498253, "grad_norm": 1.0848970413208008, "learning_rate": 7.639695663262089e-05, "loss": 0.3746, "step": 21410 }, { "epoch": 12.467986030267753, "grad_norm": 0.8459434509277344, "learning_rate": 7.637354824944128e-05, "loss": 0.385, "step": 21420 }, { "epoch": 12.473806752037252, "grad_norm": 0.922215461730957, "learning_rate": 7.635013185467408e-05, "loss": 0.3796, "step": 21430 }, { "epoch": 12.479627473806753, "grad_norm": 0.9649181962013245, "learning_rate": 7.632670745543256e-05, "loss": 0.3792, "step": 21440 }, { "epoch": 12.485448195576252, "grad_norm": 0.9283530116081238, "learning_rate": 7.630327505883242e-05, "loss": 0.3777, "step": 21450 }, { "epoch": 12.491268917345751, "grad_norm": 0.8308735489845276, "learning_rate": 7.627983467199182e-05, "loss": 0.3677, "step": 21460 }, { "epoch": 12.49708963911525, "grad_norm": 0.9836928844451904, "learning_rate": 7.625638630203132e-05, "loss": 0.3706, "step": 21470 }, { "epoch": 12.50291036088475, "grad_norm": 0.9991427063941956, "learning_rate": 7.623292995607394e-05, "loss": 0.3677, "step": 21480 }, { "epoch": 12.508731082654249, "grad_norm": 0.9075595736503601, "learning_rate": 7.620946564124507e-05, "loss": 0.3851, "step": 21490 }, { "epoch": 12.514551804423748, "grad_norm": 0.9340375065803528, "learning_rate": 7.618599336467256e-05, "loss": 0.3751, "step": 21500 }, { "epoch": 12.520372526193247, "grad_norm": 0.9119265675544739, "learning_rate": 7.616251313348666e-05, "loss": 0.3801, "step": 21510 }, { "epoch": 12.526193247962748, "grad_norm": 0.8932326436042786, "learning_rate": 7.613902495482005e-05, "loss": 0.384, "step": 21520 }, { "epoch": 12.532013969732247, "grad_norm": 0.8134983777999878, "learning_rate": 7.611552883580784e-05, "loss": 0.3623, "step": 21530 }, { "epoch": 12.537834691501747, "grad_norm": 0.8545719981193542, "learning_rate": 7.609202478358748e-05, "loss": 0.3629, "step": 21540 }, { "epoch": 12.543655413271246, "grad_norm": 1.0076854228973389, "learning_rate": 7.606851280529895e-05, "loss": 0.3674, "step": 21550 }, { "epoch": 12.549476135040745, "grad_norm": 0.979397177696228, "learning_rate": 7.604499290808449e-05, "loss": 0.3609, "step": 21560 }, { "epoch": 12.555296856810244, "grad_norm": 0.9002043008804321, "learning_rate": 7.602146509908888e-05, "loss": 0.3666, "step": 21570 }, { "epoch": 12.561117578579744, "grad_norm": 0.8945687413215637, "learning_rate": 7.599792938545921e-05, "loss": 0.3699, "step": 21580 }, { "epoch": 12.566938300349243, "grad_norm": 0.9460280537605286, "learning_rate": 7.597438577434506e-05, "loss": 0.3855, "step": 21590 }, { "epoch": 12.572759022118742, "grad_norm": 0.8966083526611328, "learning_rate": 7.595083427289831e-05, "loss": 0.3625, "step": 21600 }, { "epoch": 12.578579743888243, "grad_norm": 0.9449682235717773, "learning_rate": 7.59272748882733e-05, "loss": 0.3858, "step": 21610 }, { "epoch": 12.584400465657742, "grad_norm": 0.8362062573432922, "learning_rate": 7.590370762762675e-05, "loss": 0.3683, "step": 21620 }, { "epoch": 12.590221187427241, "grad_norm": 0.9183404445648193, "learning_rate": 7.588013249811777e-05, "loss": 0.3695, "step": 21630 }, { "epoch": 12.59604190919674, "grad_norm": 0.8992018699645996, "learning_rate": 7.585654950690786e-05, "loss": 0.3796, "step": 21640 }, { "epoch": 12.60186263096624, "grad_norm": 0.9210944771766663, "learning_rate": 7.583295866116091e-05, "loss": 0.3649, "step": 21650 }, { "epoch": 12.607683352735739, "grad_norm": 0.9403299689292908, "learning_rate": 7.580935996804321e-05, "loss": 0.361, "step": 21660 }, { "epoch": 12.613504074505238, "grad_norm": 0.9280886054039001, "learning_rate": 7.57857534347234e-05, "loss": 0.377, "step": 21670 }, { "epoch": 12.619324796274737, "grad_norm": 0.8635410666465759, "learning_rate": 7.576213906837254e-05, "loss": 0.3754, "step": 21680 }, { "epoch": 12.625145518044237, "grad_norm": 0.8028340339660645, "learning_rate": 7.573851687616403e-05, "loss": 0.3732, "step": 21690 }, { "epoch": 12.630966239813738, "grad_norm": 0.9808146953582764, "learning_rate": 7.571488686527368e-05, "loss": 0.3822, "step": 21700 }, { "epoch": 12.636786961583237, "grad_norm": 0.8879321813583374, "learning_rate": 7.569124904287968e-05, "loss": 0.3766, "step": 21710 }, { "epoch": 12.642607683352736, "grad_norm": 0.9029263854026794, "learning_rate": 7.566760341616254e-05, "loss": 0.3761, "step": 21720 }, { "epoch": 12.648428405122235, "grad_norm": 0.8411194682121277, "learning_rate": 7.564394999230519e-05, "loss": 0.3872, "step": 21730 }, { "epoch": 12.654249126891735, "grad_norm": 0.7834049463272095, "learning_rate": 7.562028877849294e-05, "loss": 0.3614, "step": 21740 }, { "epoch": 12.660069848661234, "grad_norm": 0.9261874556541443, "learning_rate": 7.559661978191341e-05, "loss": 0.3581, "step": 21750 }, { "epoch": 12.665890570430733, "grad_norm": 0.8483310341835022, "learning_rate": 7.557294300975664e-05, "loss": 0.3644, "step": 21760 }, { "epoch": 12.671711292200232, "grad_norm": 0.8511804342269897, "learning_rate": 7.554925846921499e-05, "loss": 0.3709, "step": 21770 }, { "epoch": 12.677532013969731, "grad_norm": 0.8713361024856567, "learning_rate": 7.552556616748321e-05, "loss": 0.3714, "step": 21780 }, { "epoch": 12.683352735739232, "grad_norm": 0.8705841302871704, "learning_rate": 7.550186611175838e-05, "loss": 0.3833, "step": 21790 }, { "epoch": 12.689173457508732, "grad_norm": 0.9795161485671997, "learning_rate": 7.547815830923998e-05, "loss": 0.3703, "step": 21800 }, { "epoch": 12.69499417927823, "grad_norm": 0.914176881313324, "learning_rate": 7.54544427671298e-05, "loss": 0.3808, "step": 21810 }, { "epoch": 12.70081490104773, "grad_norm": 0.9185447096824646, "learning_rate": 7.543071949263198e-05, "loss": 0.3717, "step": 21820 }, { "epoch": 12.70663562281723, "grad_norm": 0.8247960209846497, "learning_rate": 7.540698849295305e-05, "loss": 0.39, "step": 21830 }, { "epoch": 12.712456344586728, "grad_norm": 0.8642902374267578, "learning_rate": 7.538324977530183e-05, "loss": 0.3756, "step": 21840 }, { "epoch": 12.718277066356228, "grad_norm": 0.9888301491737366, "learning_rate": 7.535950334688955e-05, "loss": 0.3872, "step": 21850 }, { "epoch": 12.724097788125727, "grad_norm": 0.8957217335700989, "learning_rate": 7.533574921492972e-05, "loss": 0.3737, "step": 21860 }, { "epoch": 12.729918509895228, "grad_norm": 0.9394975900650024, "learning_rate": 7.531198738663824e-05, "loss": 0.3732, "step": 21870 }, { "epoch": 12.735739231664727, "grad_norm": 0.921127200126648, "learning_rate": 7.528821786923333e-05, "loss": 0.373, "step": 21880 }, { "epoch": 12.741559953434226, "grad_norm": 0.8835353851318359, "learning_rate": 7.52644406699355e-05, "loss": 0.3663, "step": 21890 }, { "epoch": 12.747380675203726, "grad_norm": 0.9827650785446167, "learning_rate": 7.524065579596766e-05, "loss": 0.3667, "step": 21900 }, { "epoch": 12.753201396973225, "grad_norm": 0.8946456909179688, "learning_rate": 7.521686325455506e-05, "loss": 0.382, "step": 21910 }, { "epoch": 12.759022118742724, "grad_norm": 0.9777619242668152, "learning_rate": 7.51930630529252e-05, "loss": 0.364, "step": 21920 }, { "epoch": 12.764842840512223, "grad_norm": 0.9451289176940918, "learning_rate": 7.516925519830797e-05, "loss": 0.3876, "step": 21930 }, { "epoch": 12.770663562281722, "grad_norm": 0.983009397983551, "learning_rate": 7.514543969793557e-05, "loss": 0.3879, "step": 21940 }, { "epoch": 12.776484284051222, "grad_norm": 0.9401159882545471, "learning_rate": 7.512161655904251e-05, "loss": 0.3596, "step": 21950 }, { "epoch": 12.782305005820723, "grad_norm": 0.9166232347488403, "learning_rate": 7.509778578886563e-05, "loss": 0.3724, "step": 21960 }, { "epoch": 12.788125727590222, "grad_norm": 0.930877685546875, "learning_rate": 7.507394739464412e-05, "loss": 0.3587, "step": 21970 }, { "epoch": 12.793946449359721, "grad_norm": 0.9239055514335632, "learning_rate": 7.50501013836194e-05, "loss": 0.394, "step": 21980 }, { "epoch": 12.79976717112922, "grad_norm": 0.8730902075767517, "learning_rate": 7.50262477630353e-05, "loss": 0.3815, "step": 21990 }, { "epoch": 12.80558789289872, "grad_norm": 0.8904895186424255, "learning_rate": 7.500238654013794e-05, "loss": 0.3654, "step": 22000 }, { "epoch": 12.811408614668219, "grad_norm": 0.8957593441009521, "learning_rate": 7.497851772217566e-05, "loss": 0.3616, "step": 22010 }, { "epoch": 12.817229336437718, "grad_norm": 0.8962363004684448, "learning_rate": 7.495464131639924e-05, "loss": 0.3792, "step": 22020 }, { "epoch": 12.823050058207217, "grad_norm": 0.8621322512626648, "learning_rate": 7.493075733006166e-05, "loss": 0.3767, "step": 22030 }, { "epoch": 12.828870779976716, "grad_norm": 0.9712288975715637, "learning_rate": 7.490686577041828e-05, "loss": 0.3743, "step": 22040 }, { "epoch": 12.834691501746217, "grad_norm": 0.9499992728233337, "learning_rate": 7.488296664472668e-05, "loss": 0.3615, "step": 22050 }, { "epoch": 12.840512223515717, "grad_norm": 0.8443769216537476, "learning_rate": 7.485905996024682e-05, "loss": 0.3526, "step": 22060 }, { "epoch": 12.846332945285216, "grad_norm": 0.8765397667884827, "learning_rate": 7.483514572424093e-05, "loss": 0.3676, "step": 22070 }, { "epoch": 12.852153667054715, "grad_norm": 1.0246998071670532, "learning_rate": 7.481122394397349e-05, "loss": 0.3726, "step": 22080 }, { "epoch": 12.857974388824214, "grad_norm": 0.9399444460868835, "learning_rate": 7.478729462671131e-05, "loss": 0.3772, "step": 22090 }, { "epoch": 12.863795110593713, "grad_norm": 0.8318814635276794, "learning_rate": 7.47633577797235e-05, "loss": 0.3814, "step": 22100 }, { "epoch": 12.869615832363213, "grad_norm": 0.9059068560600281, "learning_rate": 7.473941341028144e-05, "loss": 0.3784, "step": 22110 }, { "epoch": 12.875436554132712, "grad_norm": 0.8209186792373657, "learning_rate": 7.471546152565879e-05, "loss": 0.361, "step": 22120 }, { "epoch": 12.881257275902211, "grad_norm": 0.8178325891494751, "learning_rate": 7.46915021331315e-05, "loss": 0.3629, "step": 22130 }, { "epoch": 12.887077997671712, "grad_norm": 0.8547857403755188, "learning_rate": 7.466753523997778e-05, "loss": 0.3676, "step": 22140 }, { "epoch": 12.892898719441211, "grad_norm": 0.8285050392150879, "learning_rate": 7.464356085347819e-05, "loss": 0.3721, "step": 22150 }, { "epoch": 12.89871944121071, "grad_norm": 0.9131461977958679, "learning_rate": 7.461957898091548e-05, "loss": 0.3698, "step": 22160 }, { "epoch": 12.90454016298021, "grad_norm": 0.9325342178344727, "learning_rate": 7.459558962957473e-05, "loss": 0.3609, "step": 22170 }, { "epoch": 12.910360884749709, "grad_norm": 0.8688215613365173, "learning_rate": 7.457159280674326e-05, "loss": 0.3702, "step": 22180 }, { "epoch": 12.916181606519208, "grad_norm": 0.7964382767677307, "learning_rate": 7.454758851971066e-05, "loss": 0.3549, "step": 22190 }, { "epoch": 12.922002328288707, "grad_norm": 0.9265076518058777, "learning_rate": 7.45235767757688e-05, "loss": 0.3712, "step": 22200 }, { "epoch": 12.927823050058207, "grad_norm": 0.9237794876098633, "learning_rate": 7.449955758221183e-05, "loss": 0.362, "step": 22210 }, { "epoch": 12.933643771827708, "grad_norm": 0.9003564119338989, "learning_rate": 7.447553094633615e-05, "loss": 0.3952, "step": 22220 }, { "epoch": 12.939464493597207, "grad_norm": 0.8454480767250061, "learning_rate": 7.445149687544039e-05, "loss": 0.382, "step": 22230 }, { "epoch": 12.945285215366706, "grad_norm": 0.8856659531593323, "learning_rate": 7.44274553768255e-05, "loss": 0.3669, "step": 22240 }, { "epoch": 12.951105937136205, "grad_norm": 0.8724136352539062, "learning_rate": 7.440340645779464e-05, "loss": 0.3733, "step": 22250 }, { "epoch": 12.956926658905704, "grad_norm": 0.8578822016716003, "learning_rate": 7.437935012565322e-05, "loss": 0.3687, "step": 22260 }, { "epoch": 12.962747380675204, "grad_norm": 0.8795007467269897, "learning_rate": 7.435528638770893e-05, "loss": 0.3767, "step": 22270 }, { "epoch": 12.968568102444703, "grad_norm": 0.8716334700584412, "learning_rate": 7.433121525127171e-05, "loss": 0.3667, "step": 22280 }, { "epoch": 12.974388824214202, "grad_norm": 0.8114678859710693, "learning_rate": 7.430713672365371e-05, "loss": 0.3601, "step": 22290 }, { "epoch": 12.980209545983701, "grad_norm": 0.9276398420333862, "learning_rate": 7.428305081216938e-05, "loss": 0.3674, "step": 22300 }, { "epoch": 12.9860302677532, "grad_norm": 0.8999015688896179, "learning_rate": 7.425895752413536e-05, "loss": 0.3683, "step": 22310 }, { "epoch": 12.991850989522701, "grad_norm": 0.8826491832733154, "learning_rate": 7.423485686687057e-05, "loss": 0.3809, "step": 22320 }, { "epoch": 12.9976717112922, "grad_norm": 0.9681684374809265, "learning_rate": 7.421074884769616e-05, "loss": 0.3797, "step": 22330 }, { "epoch": 13.0034924330617, "grad_norm": 1.0129884481430054, "learning_rate": 7.418663347393548e-05, "loss": 0.3825, "step": 22340 }, { "epoch": 13.009313154831199, "grad_norm": 0.8234730362892151, "learning_rate": 7.416251075291418e-05, "loss": 0.3746, "step": 22350 }, { "epoch": 13.015133876600698, "grad_norm": 0.8852059841156006, "learning_rate": 7.413838069196007e-05, "loss": 0.3585, "step": 22360 }, { "epoch": 13.020954598370198, "grad_norm": 0.8918631672859192, "learning_rate": 7.411424329840324e-05, "loss": 0.3689, "step": 22370 }, { "epoch": 13.026775320139697, "grad_norm": 0.9187667369842529, "learning_rate": 7.409009857957601e-05, "loss": 0.3701, "step": 22380 }, { "epoch": 13.032596041909196, "grad_norm": 0.8954730033874512, "learning_rate": 7.40659465428129e-05, "loss": 0.3617, "step": 22390 }, { "epoch": 13.038416763678697, "grad_norm": 0.900609016418457, "learning_rate": 7.404178719545063e-05, "loss": 0.3753, "step": 22400 }, { "epoch": 13.044237485448196, "grad_norm": 1.054720401763916, "learning_rate": 7.401762054482822e-05, "loss": 0.3926, "step": 22410 }, { "epoch": 13.050058207217695, "grad_norm": 0.8188084363937378, "learning_rate": 7.39934465982868e-05, "loss": 0.3704, "step": 22420 }, { "epoch": 13.055878928987195, "grad_norm": 0.8111404180526733, "learning_rate": 7.396926536316984e-05, "loss": 0.3672, "step": 22430 }, { "epoch": 13.061699650756694, "grad_norm": 0.9936233758926392, "learning_rate": 7.394507684682293e-05, "loss": 0.3518, "step": 22440 }, { "epoch": 13.067520372526193, "grad_norm": 0.9392058253288269, "learning_rate": 7.392088105659393e-05, "loss": 0.3832, "step": 22450 }, { "epoch": 13.073341094295692, "grad_norm": 0.7751203179359436, "learning_rate": 7.389667799983284e-05, "loss": 0.373, "step": 22460 }, { "epoch": 13.079161816065191, "grad_norm": 1.0100139379501343, "learning_rate": 7.387246768389193e-05, "loss": 0.3644, "step": 22470 }, { "epoch": 13.08498253783469, "grad_norm": 0.8842720985412598, "learning_rate": 7.384825011612563e-05, "loss": 0.3705, "step": 22480 }, { "epoch": 13.090803259604192, "grad_norm": 0.9453799724578857, "learning_rate": 7.382402530389066e-05, "loss": 0.3787, "step": 22490 }, { "epoch": 13.09662398137369, "grad_norm": 0.8821013569831848, "learning_rate": 7.379979325454582e-05, "loss": 0.3715, "step": 22500 }, { "epoch": 13.10244470314319, "grad_norm": 0.8689032196998596, "learning_rate": 7.37755539754522e-05, "loss": 0.3591, "step": 22510 }, { "epoch": 13.10826542491269, "grad_norm": 0.8747856616973877, "learning_rate": 7.375130747397302e-05, "loss": 0.3693, "step": 22520 }, { "epoch": 13.114086146682189, "grad_norm": 0.886725902557373, "learning_rate": 7.372705375747377e-05, "loss": 0.3603, "step": 22530 }, { "epoch": 13.119906868451688, "grad_norm": 0.8986651301383972, "learning_rate": 7.370279283332205e-05, "loss": 0.3629, "step": 22540 }, { "epoch": 13.125727590221187, "grad_norm": 0.8574289679527283, "learning_rate": 7.36785247088877e-05, "loss": 0.3604, "step": 22550 }, { "epoch": 13.131548311990686, "grad_norm": 0.8812527656555176, "learning_rate": 7.365424939154275e-05, "loss": 0.3706, "step": 22560 }, { "epoch": 13.137369033760187, "grad_norm": 0.927931010723114, "learning_rate": 7.362996688866138e-05, "loss": 0.3785, "step": 22570 }, { "epoch": 13.143189755529686, "grad_norm": 0.8739878535270691, "learning_rate": 7.360567720761999e-05, "loss": 0.3506, "step": 22580 }, { "epoch": 13.149010477299186, "grad_norm": 0.9395618438720703, "learning_rate": 7.358138035579711e-05, "loss": 0.3696, "step": 22590 }, { "epoch": 13.154831199068685, "grad_norm": 0.8288349509239197, "learning_rate": 7.355707634057354e-05, "loss": 0.3711, "step": 22600 }, { "epoch": 13.160651920838184, "grad_norm": 0.8514562845230103, "learning_rate": 7.353276516933215e-05, "loss": 0.3715, "step": 22610 }, { "epoch": 13.166472642607683, "grad_norm": 0.8491923809051514, "learning_rate": 7.350844684945806e-05, "loss": 0.3609, "step": 22620 }, { "epoch": 13.172293364377182, "grad_norm": 0.909537672996521, "learning_rate": 7.348412138833851e-05, "loss": 0.3523, "step": 22630 }, { "epoch": 13.178114086146682, "grad_norm": 0.938489556312561, "learning_rate": 7.345978879336295e-05, "loss": 0.3808, "step": 22640 }, { "epoch": 13.18393480791618, "grad_norm": 0.866823136806488, "learning_rate": 7.343544907192296e-05, "loss": 0.3685, "step": 22650 }, { "epoch": 13.189755529685682, "grad_norm": 0.9087203741073608, "learning_rate": 7.341110223141235e-05, "loss": 0.3695, "step": 22660 }, { "epoch": 13.195576251455181, "grad_norm": 0.9233509302139282, "learning_rate": 7.3386748279227e-05, "loss": 0.3703, "step": 22670 }, { "epoch": 13.20139697322468, "grad_norm": 0.8659322261810303, "learning_rate": 7.336238722276501e-05, "loss": 0.3758, "step": 22680 }, { "epoch": 13.20721769499418, "grad_norm": 0.9753334522247314, "learning_rate": 7.333801906942663e-05, "loss": 0.3939, "step": 22690 }, { "epoch": 13.213038416763679, "grad_norm": 0.8919965624809265, "learning_rate": 7.331364382661428e-05, "loss": 0.3683, "step": 22700 }, { "epoch": 13.218859138533178, "grad_norm": 0.8510006666183472, "learning_rate": 7.328926150173248e-05, "loss": 0.3669, "step": 22710 }, { "epoch": 13.224679860302677, "grad_norm": 0.915793776512146, "learning_rate": 7.326487210218795e-05, "loss": 0.3669, "step": 22720 }, { "epoch": 13.230500582072176, "grad_norm": 0.8834252953529358, "learning_rate": 7.324047563538955e-05, "loss": 0.3754, "step": 22730 }, { "epoch": 13.236321303841676, "grad_norm": 0.836448073387146, "learning_rate": 7.321607210874828e-05, "loss": 0.3718, "step": 22740 }, { "epoch": 13.242142025611177, "grad_norm": 0.8950358629226685, "learning_rate": 7.31916615296773e-05, "loss": 0.3718, "step": 22750 }, { "epoch": 13.247962747380676, "grad_norm": 0.9047633409500122, "learning_rate": 7.316724390559188e-05, "loss": 0.3787, "step": 22760 }, { "epoch": 13.253783469150175, "grad_norm": 0.9118289947509766, "learning_rate": 7.314281924390946e-05, "loss": 0.3665, "step": 22770 }, { "epoch": 13.259604190919674, "grad_norm": 0.9960862994194031, "learning_rate": 7.311838755204959e-05, "loss": 0.3699, "step": 22780 }, { "epoch": 13.265424912689173, "grad_norm": 0.8361077308654785, "learning_rate": 7.3093948837434e-05, "loss": 0.3744, "step": 22790 }, { "epoch": 13.271245634458673, "grad_norm": 0.8538800477981567, "learning_rate": 7.306950310748651e-05, "loss": 0.3843, "step": 22800 }, { "epoch": 13.277066356228172, "grad_norm": 0.9298845529556274, "learning_rate": 7.304505036963311e-05, "loss": 0.3688, "step": 22810 }, { "epoch": 13.282887077997671, "grad_norm": 0.927438497543335, "learning_rate": 7.302059063130186e-05, "loss": 0.3759, "step": 22820 }, { "epoch": 13.28870779976717, "grad_norm": 0.9113131165504456, "learning_rate": 7.2996123899923e-05, "loss": 0.372, "step": 22830 }, { "epoch": 13.294528521536671, "grad_norm": 0.9688734412193298, "learning_rate": 7.297165018292886e-05, "loss": 0.3734, "step": 22840 }, { "epoch": 13.30034924330617, "grad_norm": 0.8673002123832703, "learning_rate": 7.294716948775396e-05, "loss": 0.3561, "step": 22850 }, { "epoch": 13.30616996507567, "grad_norm": 0.9174924492835999, "learning_rate": 7.292268182183484e-05, "loss": 0.3661, "step": 22860 }, { "epoch": 13.311990686845169, "grad_norm": 0.8974834680557251, "learning_rate": 7.28981871926102e-05, "loss": 0.3532, "step": 22870 }, { "epoch": 13.317811408614668, "grad_norm": 0.8891708850860596, "learning_rate": 7.28736856075209e-05, "loss": 0.3672, "step": 22880 }, { "epoch": 13.323632130384167, "grad_norm": 0.9022500514984131, "learning_rate": 7.284917707400985e-05, "loss": 0.3749, "step": 22890 }, { "epoch": 13.329452852153667, "grad_norm": 0.9370237588882446, "learning_rate": 7.282466159952212e-05, "loss": 0.3646, "step": 22900 }, { "epoch": 13.335273573923166, "grad_norm": 0.8648034930229187, "learning_rate": 7.280013919150483e-05, "loss": 0.3777, "step": 22910 }, { "epoch": 13.341094295692667, "grad_norm": 0.8445641994476318, "learning_rate": 7.277560985740728e-05, "loss": 0.3712, "step": 22920 }, { "epoch": 13.346915017462166, "grad_norm": 0.8323317766189575, "learning_rate": 7.275107360468079e-05, "loss": 0.3719, "step": 22930 }, { "epoch": 13.352735739231665, "grad_norm": 0.852628767490387, "learning_rate": 7.272653044077885e-05, "loss": 0.3584, "step": 22940 }, { "epoch": 13.358556461001164, "grad_norm": 0.9539941549301147, "learning_rate": 7.270198037315703e-05, "loss": 0.3638, "step": 22950 }, { "epoch": 13.364377182770664, "grad_norm": 0.8217638731002808, "learning_rate": 7.267742340927297e-05, "loss": 0.3551, "step": 22960 }, { "epoch": 13.370197904540163, "grad_norm": 0.8922520279884338, "learning_rate": 7.265285955658645e-05, "loss": 0.3623, "step": 22970 }, { "epoch": 13.376018626309662, "grad_norm": 0.9186207056045532, "learning_rate": 7.26282888225593e-05, "loss": 0.3729, "step": 22980 }, { "epoch": 13.381839348079161, "grad_norm": 1.0959477424621582, "learning_rate": 7.260371121465548e-05, "loss": 0.3633, "step": 22990 }, { "epoch": 13.38766006984866, "grad_norm": 0.9776979088783264, "learning_rate": 7.2579126740341e-05, "loss": 0.3791, "step": 23000 }, { "epoch": 13.39348079161816, "grad_norm": 0.9533649682998657, "learning_rate": 7.2554535407084e-05, "loss": 0.3812, "step": 23010 }, { "epoch": 13.39930151338766, "grad_norm": 0.8154546618461609, "learning_rate": 7.252993722235464e-05, "loss": 0.382, "step": 23020 }, { "epoch": 13.40512223515716, "grad_norm": 0.8467314839363098, "learning_rate": 7.250533219362523e-05, "loss": 0.3796, "step": 23030 }, { "epoch": 13.41094295692666, "grad_norm": 0.9104031920433044, "learning_rate": 7.248072032837012e-05, "loss": 0.3719, "step": 23040 }, { "epoch": 13.416763678696158, "grad_norm": 0.9099923968315125, "learning_rate": 7.245610163406575e-05, "loss": 0.3886, "step": 23050 }, { "epoch": 13.422584400465658, "grad_norm": 0.8908551335334778, "learning_rate": 7.243147611819061e-05, "loss": 0.368, "step": 23060 }, { "epoch": 13.428405122235157, "grad_norm": 0.8628039956092834, "learning_rate": 7.240684378822531e-05, "loss": 0.3859, "step": 23070 }, { "epoch": 13.434225844004656, "grad_norm": 0.9439898133277893, "learning_rate": 7.238220465165248e-05, "loss": 0.371, "step": 23080 }, { "epoch": 13.440046565774155, "grad_norm": 1.1104811429977417, "learning_rate": 7.235755871595684e-05, "loss": 0.3768, "step": 23090 }, { "epoch": 13.445867287543656, "grad_norm": 0.8795697689056396, "learning_rate": 7.233290598862517e-05, "loss": 0.383, "step": 23100 }, { "epoch": 13.451688009313155, "grad_norm": 0.8345921635627747, "learning_rate": 7.230824647714635e-05, "loss": 0.3689, "step": 23110 }, { "epoch": 13.457508731082655, "grad_norm": 0.9146775603294373, "learning_rate": 7.228358018901124e-05, "loss": 0.3738, "step": 23120 }, { "epoch": 13.463329452852154, "grad_norm": 0.9424764513969421, "learning_rate": 7.225890713171286e-05, "loss": 0.3601, "step": 23130 }, { "epoch": 13.469150174621653, "grad_norm": 0.8913775086402893, "learning_rate": 7.223422731274618e-05, "loss": 0.3659, "step": 23140 }, { "epoch": 13.474970896391152, "grad_norm": 0.8539252877235413, "learning_rate": 7.220954073960832e-05, "loss": 0.3599, "step": 23150 }, { "epoch": 13.480791618160652, "grad_norm": 0.9190834760665894, "learning_rate": 7.218484741979838e-05, "loss": 0.3883, "step": 23160 }, { "epoch": 13.48661233993015, "grad_norm": 0.8655169010162354, "learning_rate": 7.216014736081756e-05, "loss": 0.381, "step": 23170 }, { "epoch": 13.49243306169965, "grad_norm": 0.9781033992767334, "learning_rate": 7.213544057016906e-05, "loss": 0.3622, "step": 23180 }, { "epoch": 13.498253783469151, "grad_norm": 0.8550151586532593, "learning_rate": 7.211072705535819e-05, "loss": 0.364, "step": 23190 }, { "epoch": 13.50407450523865, "grad_norm": 0.8867624402046204, "learning_rate": 7.208600682389224e-05, "loss": 0.3834, "step": 23200 }, { "epoch": 13.50989522700815, "grad_norm": 0.9311361908912659, "learning_rate": 7.206127988328055e-05, "loss": 0.3889, "step": 23210 }, { "epoch": 13.515715948777649, "grad_norm": 0.9450420141220093, "learning_rate": 7.203654624103453e-05, "loss": 0.3675, "step": 23220 }, { "epoch": 13.521536670547148, "grad_norm": 0.9166697859764099, "learning_rate": 7.201180590466761e-05, "loss": 0.3893, "step": 23230 }, { "epoch": 13.527357392316647, "grad_norm": 0.9188504219055176, "learning_rate": 7.198705888169523e-05, "loss": 0.3697, "step": 23240 }, { "epoch": 13.533178114086146, "grad_norm": 0.9466860294342041, "learning_rate": 7.196230517963491e-05, "loss": 0.385, "step": 23250 }, { "epoch": 13.538998835855645, "grad_norm": 0.8790097236633301, "learning_rate": 7.193754480600615e-05, "loss": 0.381, "step": 23260 }, { "epoch": 13.544819557625146, "grad_norm": 0.9275935888290405, "learning_rate": 7.19127777683305e-05, "loss": 0.373, "step": 23270 }, { "epoch": 13.550640279394646, "grad_norm": 0.7932361960411072, "learning_rate": 7.188800407413156e-05, "loss": 0.3667, "step": 23280 }, { "epoch": 13.556461001164145, "grad_norm": 0.9170732498168945, "learning_rate": 7.186322373093489e-05, "loss": 0.3735, "step": 23290 }, { "epoch": 13.562281722933644, "grad_norm": 0.832909345626831, "learning_rate": 7.18384367462681e-05, "loss": 0.3662, "step": 23300 }, { "epoch": 13.568102444703143, "grad_norm": 0.9035414457321167, "learning_rate": 7.181364312766085e-05, "loss": 0.3708, "step": 23310 }, { "epoch": 13.573923166472643, "grad_norm": 0.9432408213615417, "learning_rate": 7.178884288264477e-05, "loss": 0.3721, "step": 23320 }, { "epoch": 13.579743888242142, "grad_norm": 0.8216586112976074, "learning_rate": 7.176403601875353e-05, "loss": 0.3739, "step": 23330 }, { "epoch": 13.585564610011641, "grad_norm": 0.9649113416671753, "learning_rate": 7.173922254352279e-05, "loss": 0.3679, "step": 23340 }, { "epoch": 13.59138533178114, "grad_norm": 0.8361912965774536, "learning_rate": 7.171440246449024e-05, "loss": 0.3732, "step": 23350 }, { "epoch": 13.59720605355064, "grad_norm": 0.8422397375106812, "learning_rate": 7.168957578919555e-05, "loss": 0.3508, "step": 23360 }, { "epoch": 13.60302677532014, "grad_norm": 0.7808836102485657, "learning_rate": 7.16647425251804e-05, "loss": 0.3652, "step": 23370 }, { "epoch": 13.60884749708964, "grad_norm": 0.7774417996406555, "learning_rate": 7.163990267998852e-05, "loss": 0.3651, "step": 23380 }, { "epoch": 13.614668218859139, "grad_norm": 0.7561466097831726, "learning_rate": 7.161505626116556e-05, "loss": 0.3707, "step": 23390 }, { "epoch": 13.620488940628638, "grad_norm": 0.8417432308197021, "learning_rate": 7.159020327625923e-05, "loss": 0.3692, "step": 23400 }, { "epoch": 13.626309662398137, "grad_norm": 0.8548671007156372, "learning_rate": 7.15653437328192e-05, "loss": 0.3617, "step": 23410 }, { "epoch": 13.632130384167636, "grad_norm": 0.86459881067276, "learning_rate": 7.154047763839713e-05, "loss": 0.3756, "step": 23420 }, { "epoch": 13.637951105937136, "grad_norm": 0.7991750240325928, "learning_rate": 7.15156050005467e-05, "loss": 0.3606, "step": 23430 }, { "epoch": 13.643771827706635, "grad_norm": 0.9347641468048096, "learning_rate": 7.149072582682357e-05, "loss": 0.3721, "step": 23440 }, { "epoch": 13.649592549476136, "grad_norm": 0.8163397312164307, "learning_rate": 7.146584012478535e-05, "loss": 0.3655, "step": 23450 }, { "epoch": 13.655413271245635, "grad_norm": 0.8405846953392029, "learning_rate": 7.144094790199169e-05, "loss": 0.3671, "step": 23460 }, { "epoch": 13.661233993015134, "grad_norm": 0.8320028781890869, "learning_rate": 7.141604916600415e-05, "loss": 0.3717, "step": 23470 }, { "epoch": 13.667054714784634, "grad_norm": 0.8584665060043335, "learning_rate": 7.139114392438635e-05, "loss": 0.3693, "step": 23480 }, { "epoch": 13.672875436554133, "grad_norm": 0.9168875217437744, "learning_rate": 7.136623218470382e-05, "loss": 0.3754, "step": 23490 }, { "epoch": 13.678696158323632, "grad_norm": 0.8293243050575256, "learning_rate": 7.13413139545241e-05, "loss": 0.378, "step": 23500 }, { "epoch": 13.684516880093131, "grad_norm": 0.8496024012565613, "learning_rate": 7.131638924141668e-05, "loss": 0.3647, "step": 23510 }, { "epoch": 13.69033760186263, "grad_norm": 0.9114397168159485, "learning_rate": 7.129145805295304e-05, "loss": 0.3787, "step": 23520 }, { "epoch": 13.69615832363213, "grad_norm": 0.9111357927322388, "learning_rate": 7.126652039670661e-05, "loss": 0.3659, "step": 23530 }, { "epoch": 13.70197904540163, "grad_norm": 0.939799427986145, "learning_rate": 7.124157628025278e-05, "loss": 0.3713, "step": 23540 }, { "epoch": 13.70779976717113, "grad_norm": 0.8924121260643005, "learning_rate": 7.121662571116894e-05, "loss": 0.3534, "step": 23550 }, { "epoch": 13.713620488940629, "grad_norm": 0.8544423580169678, "learning_rate": 7.119166869703441e-05, "loss": 0.367, "step": 23560 }, { "epoch": 13.719441210710128, "grad_norm": 0.8870489597320557, "learning_rate": 7.116670524543044e-05, "loss": 0.3724, "step": 23570 }, { "epoch": 13.725261932479627, "grad_norm": 0.8803307414054871, "learning_rate": 7.114173536394032e-05, "loss": 0.3642, "step": 23580 }, { "epoch": 13.731082654249127, "grad_norm": 1.0116991996765137, "learning_rate": 7.111675906014917e-05, "loss": 0.3723, "step": 23590 }, { "epoch": 13.736903376018626, "grad_norm": 0.9802102446556091, "learning_rate": 7.109177634164421e-05, "loss": 0.3845, "step": 23600 }, { "epoch": 13.742724097788125, "grad_norm": 0.9900462627410889, "learning_rate": 7.106678721601449e-05, "loss": 0.3739, "step": 23610 }, { "epoch": 13.748544819557626, "grad_norm": 0.8208266496658325, "learning_rate": 7.104179169085103e-05, "loss": 0.3695, "step": 23620 }, { "epoch": 13.754365541327125, "grad_norm": 0.8230187892913818, "learning_rate": 7.101678977374683e-05, "loss": 0.3675, "step": 23630 }, { "epoch": 13.760186263096625, "grad_norm": 0.8889334201812744, "learning_rate": 7.099178147229685e-05, "loss": 0.3836, "step": 23640 }, { "epoch": 13.766006984866124, "grad_norm": 0.8857555985450745, "learning_rate": 7.096676679409789e-05, "loss": 0.3715, "step": 23650 }, { "epoch": 13.771827706635623, "grad_norm": 0.859822154045105, "learning_rate": 7.094174574674877e-05, "loss": 0.3799, "step": 23660 }, { "epoch": 13.777648428405122, "grad_norm": 0.8716594576835632, "learning_rate": 7.091671833785025e-05, "loss": 0.3695, "step": 23670 }, { "epoch": 13.783469150174621, "grad_norm": 1.0160565376281738, "learning_rate": 7.089168457500493e-05, "loss": 0.3769, "step": 23680 }, { "epoch": 13.78928987194412, "grad_norm": 0.8870261311531067, "learning_rate": 7.086664446581747e-05, "loss": 0.369, "step": 23690 }, { "epoch": 13.79511059371362, "grad_norm": 0.7947361469268799, "learning_rate": 7.084159801789438e-05, "loss": 0.3538, "step": 23700 }, { "epoch": 13.800931315483119, "grad_norm": 0.7842016220092773, "learning_rate": 7.081654523884411e-05, "loss": 0.3612, "step": 23710 }, { "epoch": 13.80675203725262, "grad_norm": 0.9271778464317322, "learning_rate": 7.0791486136277e-05, "loss": 0.371, "step": 23720 }, { "epoch": 13.81257275902212, "grad_norm": 0.8115108013153076, "learning_rate": 7.07664207178054e-05, "loss": 0.3672, "step": 23730 }, { "epoch": 13.818393480791618, "grad_norm": 0.7920642495155334, "learning_rate": 7.074134899104345e-05, "loss": 0.3622, "step": 23740 }, { "epoch": 13.824214202561118, "grad_norm": 0.870080292224884, "learning_rate": 7.071627096360735e-05, "loss": 0.3743, "step": 23750 }, { "epoch": 13.830034924330617, "grad_norm": 0.9487703442573547, "learning_rate": 7.069118664311511e-05, "loss": 0.3672, "step": 23760 }, { "epoch": 13.835855646100116, "grad_norm": 0.8792227506637573, "learning_rate": 7.06660960371867e-05, "loss": 0.3578, "step": 23770 }, { "epoch": 13.841676367869615, "grad_norm": 0.89492267370224, "learning_rate": 7.064099915344396e-05, "loss": 0.3724, "step": 23780 }, { "epoch": 13.847497089639115, "grad_norm": 1.0242363214492798, "learning_rate": 7.061589599951066e-05, "loss": 0.378, "step": 23790 }, { "epoch": 13.853317811408616, "grad_norm": 0.8039287328720093, "learning_rate": 7.05907865830125e-05, "loss": 0.3605, "step": 23800 }, { "epoch": 13.859138533178115, "grad_norm": 0.9403156638145447, "learning_rate": 7.056567091157703e-05, "loss": 0.3758, "step": 23810 }, { "epoch": 13.864959254947614, "grad_norm": 0.9517104029655457, "learning_rate": 7.054054899283375e-05, "loss": 0.3683, "step": 23820 }, { "epoch": 13.870779976717113, "grad_norm": 0.9098303914070129, "learning_rate": 7.051542083441403e-05, "loss": 0.3702, "step": 23830 }, { "epoch": 13.876600698486612, "grad_norm": 0.7919227480888367, "learning_rate": 7.049028644395113e-05, "loss": 0.3593, "step": 23840 }, { "epoch": 13.882421420256112, "grad_norm": 0.9336976408958435, "learning_rate": 7.046514582908024e-05, "loss": 0.396, "step": 23850 }, { "epoch": 13.88824214202561, "grad_norm": 0.9649502635002136, "learning_rate": 7.043999899743838e-05, "loss": 0.3727, "step": 23860 }, { "epoch": 13.89406286379511, "grad_norm": 0.9209492206573486, "learning_rate": 7.041484595666451e-05, "loss": 0.3645, "step": 23870 }, { "epoch": 13.89988358556461, "grad_norm": 0.9011490345001221, "learning_rate": 7.038968671439948e-05, "loss": 0.3778, "step": 23880 }, { "epoch": 13.90570430733411, "grad_norm": 0.9682804346084595, "learning_rate": 7.036452127828596e-05, "loss": 0.3655, "step": 23890 }, { "epoch": 13.91152502910361, "grad_norm": 0.8545828461647034, "learning_rate": 7.033934965596859e-05, "loss": 0.3651, "step": 23900 }, { "epoch": 13.917345750873109, "grad_norm": 0.9247050881385803, "learning_rate": 7.031417185509381e-05, "loss": 0.3762, "step": 23910 }, { "epoch": 13.923166472642608, "grad_norm": 0.9315139651298523, "learning_rate": 7.028898788331e-05, "loss": 0.3797, "step": 23920 }, { "epoch": 13.928987194412107, "grad_norm": 0.9091866612434387, "learning_rate": 7.026379774826736e-05, "loss": 0.3788, "step": 23930 }, { "epoch": 13.934807916181606, "grad_norm": 0.8792995810508728, "learning_rate": 7.0238601457618e-05, "loss": 0.3616, "step": 23940 }, { "epoch": 13.940628637951106, "grad_norm": 0.8240976929664612, "learning_rate": 7.02133990190159e-05, "loss": 0.367, "step": 23950 }, { "epoch": 13.946449359720605, "grad_norm": 0.8717221617698669, "learning_rate": 7.018819044011687e-05, "loss": 0.3616, "step": 23960 }, { "epoch": 13.952270081490104, "grad_norm": 0.8808758854866028, "learning_rate": 7.016297572857863e-05, "loss": 0.3554, "step": 23970 }, { "epoch": 13.958090803259605, "grad_norm": 0.7908383011817932, "learning_rate": 7.013775489206072e-05, "loss": 0.364, "step": 23980 }, { "epoch": 13.963911525029104, "grad_norm": 0.949173629283905, "learning_rate": 7.01125279382246e-05, "loss": 0.3662, "step": 23990 }, { "epoch": 13.969732246798603, "grad_norm": 0.8867692351341248, "learning_rate": 7.008729487473351e-05, "loss": 0.3713, "step": 24000 }, { "epoch": 13.975552968568103, "grad_norm": 0.879950225353241, "learning_rate": 7.006205570925263e-05, "loss": 0.3728, "step": 24010 }, { "epoch": 13.981373690337602, "grad_norm": 0.8576526641845703, "learning_rate": 7.003681044944892e-05, "loss": 0.3844, "step": 24020 }, { "epoch": 13.987194412107101, "grad_norm": 0.8256216049194336, "learning_rate": 7.001155910299126e-05, "loss": 0.3543, "step": 24030 }, { "epoch": 13.9930151338766, "grad_norm": 0.862705409526825, "learning_rate": 6.99863016775503e-05, "loss": 0.3586, "step": 24040 }, { "epoch": 13.9988358556461, "grad_norm": 0.8216114640235901, "learning_rate": 6.996103818079859e-05, "loss": 0.3655, "step": 24050 }, { "epoch": 14.004656577415599, "grad_norm": 0.8656325340270996, "learning_rate": 6.993576862041054e-05, "loss": 0.3693, "step": 24060 }, { "epoch": 14.0104772991851, "grad_norm": 0.9158079624176025, "learning_rate": 6.991049300406235e-05, "loss": 0.3662, "step": 24070 }, { "epoch": 14.016298020954599, "grad_norm": 0.8293671011924744, "learning_rate": 6.988521133943209e-05, "loss": 0.3735, "step": 24080 }, { "epoch": 14.022118742724098, "grad_norm": 0.8741906881332397, "learning_rate": 6.985992363419966e-05, "loss": 0.3742, "step": 24090 }, { "epoch": 14.027939464493597, "grad_norm": 0.8769538998603821, "learning_rate": 6.983462989604682e-05, "loss": 0.3597, "step": 24100 }, { "epoch": 14.033760186263097, "grad_norm": 0.8883635997772217, "learning_rate": 6.980933013265709e-05, "loss": 0.3739, "step": 24110 }, { "epoch": 14.039580908032596, "grad_norm": 0.8984344601631165, "learning_rate": 6.978402435171592e-05, "loss": 0.3693, "step": 24120 }, { "epoch": 14.045401629802095, "grad_norm": 0.912447988986969, "learning_rate": 6.975871256091052e-05, "loss": 0.3551, "step": 24130 }, { "epoch": 14.051222351571594, "grad_norm": 0.8040839433670044, "learning_rate": 6.973339476792995e-05, "loss": 0.3651, "step": 24140 }, { "epoch": 14.057043073341095, "grad_norm": 0.8497581481933594, "learning_rate": 6.970807098046505e-05, "loss": 0.3562, "step": 24150 }, { "epoch": 14.062863795110594, "grad_norm": 0.8163893222808838, "learning_rate": 6.968274120620858e-05, "loss": 0.3621, "step": 24160 }, { "epoch": 14.068684516880094, "grad_norm": 0.9406212568283081, "learning_rate": 6.965740545285499e-05, "loss": 0.3604, "step": 24170 }, { "epoch": 14.074505238649593, "grad_norm": 0.8638939261436462, "learning_rate": 6.963206372810068e-05, "loss": 0.3636, "step": 24180 }, { "epoch": 14.080325960419092, "grad_norm": 0.9338365793228149, "learning_rate": 6.960671603964375e-05, "loss": 0.3713, "step": 24190 }, { "epoch": 14.086146682188591, "grad_norm": 0.8919215798377991, "learning_rate": 6.958136239518418e-05, "loss": 0.3704, "step": 24200 }, { "epoch": 14.09196740395809, "grad_norm": 0.8231239318847656, "learning_rate": 6.955600280242371e-05, "loss": 0.3769, "step": 24210 }, { "epoch": 14.09778812572759, "grad_norm": 0.8998029232025146, "learning_rate": 6.953063726906596e-05, "loss": 0.373, "step": 24220 }, { "epoch": 14.103608847497089, "grad_norm": 0.8303180932998657, "learning_rate": 6.950526580281626e-05, "loss": 0.3664, "step": 24230 }, { "epoch": 14.10942956926659, "grad_norm": 0.9869544506072998, "learning_rate": 6.947988841138184e-05, "loss": 0.3673, "step": 24240 }, { "epoch": 14.115250291036089, "grad_norm": 0.8874381184577942, "learning_rate": 6.945450510247165e-05, "loss": 0.3698, "step": 24250 }, { "epoch": 14.121071012805588, "grad_norm": 0.9138217568397522, "learning_rate": 6.942911588379647e-05, "loss": 0.3731, "step": 24260 }, { "epoch": 14.126891734575088, "grad_norm": 0.9037309288978577, "learning_rate": 6.940372076306888e-05, "loss": 0.3774, "step": 24270 }, { "epoch": 14.132712456344587, "grad_norm": 0.8601304888725281, "learning_rate": 6.937831974800326e-05, "loss": 0.3561, "step": 24280 }, { "epoch": 14.138533178114086, "grad_norm": 0.8161399960517883, "learning_rate": 6.935291284631574e-05, "loss": 0.3661, "step": 24290 }, { "epoch": 14.144353899883585, "grad_norm": 0.9050516486167908, "learning_rate": 6.932750006572428e-05, "loss": 0.3605, "step": 24300 }, { "epoch": 14.150174621653084, "grad_norm": 0.8387322425842285, "learning_rate": 6.930208141394863e-05, "loss": 0.3657, "step": 24310 }, { "epoch": 14.155995343422584, "grad_norm": 0.8406665921211243, "learning_rate": 6.927665689871026e-05, "loss": 0.3787, "step": 24320 }, { "epoch": 14.161816065192085, "grad_norm": 0.7640193700790405, "learning_rate": 6.925122652773253e-05, "loss": 0.3555, "step": 24330 }, { "epoch": 14.167636786961584, "grad_norm": 0.9147052764892578, "learning_rate": 6.922579030874046e-05, "loss": 0.366, "step": 24340 }, { "epoch": 14.173457508731083, "grad_norm": 0.9227333664894104, "learning_rate": 6.920034824946093e-05, "loss": 0.3599, "step": 24350 }, { "epoch": 14.179278230500582, "grad_norm": 1.0102875232696533, "learning_rate": 6.917490035762255e-05, "loss": 0.3765, "step": 24360 }, { "epoch": 14.185098952270081, "grad_norm": 0.821762204170227, "learning_rate": 6.914944664095573e-05, "loss": 0.3439, "step": 24370 }, { "epoch": 14.19091967403958, "grad_norm": 1.0102052688598633, "learning_rate": 6.912398710719264e-05, "loss": 0.3617, "step": 24380 }, { "epoch": 14.19674039580908, "grad_norm": 0.8375685811042786, "learning_rate": 6.90985217640672e-05, "loss": 0.3778, "step": 24390 }, { "epoch": 14.202561117578579, "grad_norm": 0.903256893157959, "learning_rate": 6.90730506193151e-05, "loss": 0.3669, "step": 24400 }, { "epoch": 14.208381839348078, "grad_norm": 0.9171555638313293, "learning_rate": 6.904757368067384e-05, "loss": 0.3838, "step": 24410 }, { "epoch": 14.21420256111758, "grad_norm": 1.0091086626052856, "learning_rate": 6.90220909558826e-05, "loss": 0.3887, "step": 24420 }, { "epoch": 14.220023282887079, "grad_norm": 0.8879430294036865, "learning_rate": 6.899660245268237e-05, "loss": 0.3623, "step": 24430 }, { "epoch": 14.225844004656578, "grad_norm": 0.9147592186927795, "learning_rate": 6.897110817881592e-05, "loss": 0.3524, "step": 24440 }, { "epoch": 14.231664726426077, "grad_norm": 0.9122844338417053, "learning_rate": 6.894560814202769e-05, "loss": 0.3732, "step": 24450 }, { "epoch": 14.237485448195576, "grad_norm": 0.8178569078445435, "learning_rate": 6.892010235006394e-05, "loss": 0.3632, "step": 24460 }, { "epoch": 14.243306169965075, "grad_norm": 0.8668747544288635, "learning_rate": 6.889459081067264e-05, "loss": 0.3712, "step": 24470 }, { "epoch": 14.249126891734575, "grad_norm": 0.8927398920059204, "learning_rate": 6.886907353160356e-05, "loss": 0.3566, "step": 24480 }, { "epoch": 14.254947613504074, "grad_norm": 0.8403146266937256, "learning_rate": 6.884355052060814e-05, "loss": 0.3656, "step": 24490 }, { "epoch": 14.260768335273575, "grad_norm": 0.8151387572288513, "learning_rate": 6.88180217854396e-05, "loss": 0.3506, "step": 24500 }, { "epoch": 14.266589057043074, "grad_norm": 0.8362220525741577, "learning_rate": 6.87924873338529e-05, "loss": 0.3611, "step": 24510 }, { "epoch": 14.272409778812573, "grad_norm": 0.8322467803955078, "learning_rate": 6.876694717360475e-05, "loss": 0.3702, "step": 24520 }, { "epoch": 14.278230500582072, "grad_norm": 0.8977903723716736, "learning_rate": 6.874140131245355e-05, "loss": 0.3883, "step": 24530 }, { "epoch": 14.284051222351572, "grad_norm": 0.9062952995300293, "learning_rate": 6.871584975815948e-05, "loss": 0.3718, "step": 24540 }, { "epoch": 14.28987194412107, "grad_norm": 0.8714351654052734, "learning_rate": 6.86902925184844e-05, "loss": 0.3673, "step": 24550 }, { "epoch": 14.29569266589057, "grad_norm": 0.8113968968391418, "learning_rate": 6.866472960119195e-05, "loss": 0.3562, "step": 24560 }, { "epoch": 14.30151338766007, "grad_norm": 0.9047183990478516, "learning_rate": 6.863916101404748e-05, "loss": 0.3592, "step": 24570 }, { "epoch": 14.307334109429569, "grad_norm": 0.860538125038147, "learning_rate": 6.8613586764818e-05, "loss": 0.3496, "step": 24580 }, { "epoch": 14.31315483119907, "grad_norm": 0.8960812091827393, "learning_rate": 6.858800686127233e-05, "loss": 0.3885, "step": 24590 }, { "epoch": 14.318975552968569, "grad_norm": 0.938456654548645, "learning_rate": 6.856242131118097e-05, "loss": 0.3712, "step": 24600 }, { "epoch": 14.324796274738068, "grad_norm": 0.9463357329368591, "learning_rate": 6.853683012231614e-05, "loss": 0.3652, "step": 24610 }, { "epoch": 14.330616996507567, "grad_norm": 0.9288243651390076, "learning_rate": 6.851123330245173e-05, "loss": 0.3694, "step": 24620 }, { "epoch": 14.336437718277066, "grad_norm": 0.7941624522209167, "learning_rate": 6.848563085936343e-05, "loss": 0.3662, "step": 24630 }, { "epoch": 14.342258440046566, "grad_norm": 0.8936805129051208, "learning_rate": 6.846002280082853e-05, "loss": 0.3543, "step": 24640 }, { "epoch": 14.348079161816065, "grad_norm": 0.8898518681526184, "learning_rate": 6.843440913462614e-05, "loss": 0.3645, "step": 24650 }, { "epoch": 14.353899883585564, "grad_norm": 0.8298083543777466, "learning_rate": 6.840878986853698e-05, "loss": 0.3594, "step": 24660 }, { "epoch": 14.359720605355063, "grad_norm": 0.9087788462638855, "learning_rate": 6.838316501034352e-05, "loss": 0.365, "step": 24670 }, { "epoch": 14.365541327124564, "grad_norm": 0.8211778998374939, "learning_rate": 6.83575345678299e-05, "loss": 0.3596, "step": 24680 }, { "epoch": 14.371362048894063, "grad_norm": 0.9594720005989075, "learning_rate": 6.833189854878196e-05, "loss": 0.3666, "step": 24690 }, { "epoch": 14.377182770663563, "grad_norm": 0.8563826680183411, "learning_rate": 6.83062569609873e-05, "loss": 0.3636, "step": 24700 }, { "epoch": 14.383003492433062, "grad_norm": 0.9135967493057251, "learning_rate": 6.828060981223512e-05, "loss": 0.3514, "step": 24710 }, { "epoch": 14.388824214202561, "grad_norm": 0.9009190797805786, "learning_rate": 6.825495711031634e-05, "loss": 0.3696, "step": 24720 }, { "epoch": 14.39464493597206, "grad_norm": 0.8470876812934875, "learning_rate": 6.822929886302359e-05, "loss": 0.3593, "step": 24730 }, { "epoch": 14.40046565774156, "grad_norm": 0.8122824430465698, "learning_rate": 6.820363507815116e-05, "loss": 0.3536, "step": 24740 }, { "epoch": 14.406286379511059, "grad_norm": 0.9040880799293518, "learning_rate": 6.817796576349501e-05, "loss": 0.3584, "step": 24750 }, { "epoch": 14.412107101280558, "grad_norm": 0.8562291264533997, "learning_rate": 6.815229092685285e-05, "loss": 0.3494, "step": 24760 }, { "epoch": 14.417927823050059, "grad_norm": 0.8824843168258667, "learning_rate": 6.812661057602399e-05, "loss": 0.3653, "step": 24770 }, { "epoch": 14.423748544819558, "grad_norm": 0.8462730646133423, "learning_rate": 6.810092471880943e-05, "loss": 0.3665, "step": 24780 }, { "epoch": 14.429569266589057, "grad_norm": 0.9022014737129211, "learning_rate": 6.807523336301187e-05, "loss": 0.3616, "step": 24790 }, { "epoch": 14.435389988358557, "grad_norm": 0.8272255063056946, "learning_rate": 6.804953651643566e-05, "loss": 0.3704, "step": 24800 }, { "epoch": 14.441210710128056, "grad_norm": 0.8681949377059937, "learning_rate": 6.802383418688685e-05, "loss": 0.3524, "step": 24810 }, { "epoch": 14.447031431897555, "grad_norm": 0.7640222907066345, "learning_rate": 6.799812638217309e-05, "loss": 0.3501, "step": 24820 }, { "epoch": 14.452852153667054, "grad_norm": 0.8956701159477234, "learning_rate": 6.797241311010373e-05, "loss": 0.3524, "step": 24830 }, { "epoch": 14.458672875436553, "grad_norm": 0.9279035925865173, "learning_rate": 6.794669437848982e-05, "loss": 0.3607, "step": 24840 }, { "epoch": 14.464493597206054, "grad_norm": 0.8285792469978333, "learning_rate": 6.792097019514402e-05, "loss": 0.367, "step": 24850 }, { "epoch": 14.470314318975554, "grad_norm": 0.7854271531105042, "learning_rate": 6.789524056788064e-05, "loss": 0.3733, "step": 24860 }, { "epoch": 14.476135040745053, "grad_norm": 0.7717190980911255, "learning_rate": 6.786950550451567e-05, "loss": 0.3483, "step": 24870 }, { "epoch": 14.481955762514552, "grad_norm": 0.8563088774681091, "learning_rate": 6.784376501286676e-05, "loss": 0.3656, "step": 24880 }, { "epoch": 14.487776484284051, "grad_norm": 0.9349384307861328, "learning_rate": 6.781801910075316e-05, "loss": 0.3636, "step": 24890 }, { "epoch": 14.49359720605355, "grad_norm": 0.8152458071708679, "learning_rate": 6.779226777599581e-05, "loss": 0.3481, "step": 24900 }, { "epoch": 14.49941792782305, "grad_norm": 0.7959907054901123, "learning_rate": 6.776651104641729e-05, "loss": 0.35, "step": 24910 }, { "epoch": 14.505238649592549, "grad_norm": 0.8133774399757385, "learning_rate": 6.774074891984183e-05, "loss": 0.3513, "step": 24920 }, { "epoch": 14.511059371362048, "grad_norm": 0.9293308854103088, "learning_rate": 6.771498140409526e-05, "loss": 0.3681, "step": 24930 }, { "epoch": 14.516880093131547, "grad_norm": 0.8993431329727173, "learning_rate": 6.768920850700506e-05, "loss": 0.3701, "step": 24940 }, { "epoch": 14.522700814901048, "grad_norm": 0.8804574608802795, "learning_rate": 6.766343023640039e-05, "loss": 0.3551, "step": 24950 }, { "epoch": 14.528521536670548, "grad_norm": 0.8934838175773621, "learning_rate": 6.763764660011198e-05, "loss": 0.3595, "step": 24960 }, { "epoch": 14.534342258440047, "grad_norm": 0.7894715666770935, "learning_rate": 6.761185760597223e-05, "loss": 0.367, "step": 24970 }, { "epoch": 14.540162980209546, "grad_norm": 0.928125262260437, "learning_rate": 6.758606326181515e-05, "loss": 0.3693, "step": 24980 }, { "epoch": 14.545983701979045, "grad_norm": 0.9075110554695129, "learning_rate": 6.75602635754764e-05, "loss": 0.3638, "step": 24990 }, { "epoch": 14.551804423748544, "grad_norm": 0.8730425238609314, "learning_rate": 6.75344585547932e-05, "loss": 0.35, "step": 25000 }, { "epoch": 14.557625145518044, "grad_norm": 0.8084487318992615, "learning_rate": 6.750864820760449e-05, "loss": 0.3655, "step": 25010 }, { "epoch": 14.563445867287543, "grad_norm": 0.8254767060279846, "learning_rate": 6.748283254175072e-05, "loss": 0.3627, "step": 25020 }, { "epoch": 14.569266589057044, "grad_norm": 0.9460242390632629, "learning_rate": 6.745701156507404e-05, "loss": 0.3557, "step": 25030 }, { "epoch": 14.575087310826543, "grad_norm": 0.8331180214881897, "learning_rate": 6.743118528541818e-05, "loss": 0.3554, "step": 25040 }, { "epoch": 14.580908032596042, "grad_norm": 0.7955823540687561, "learning_rate": 6.740535371062846e-05, "loss": 0.3489, "step": 25050 }, { "epoch": 14.586728754365542, "grad_norm": 1.000473976135254, "learning_rate": 6.737951684855185e-05, "loss": 0.3666, "step": 25060 }, { "epoch": 14.59254947613504, "grad_norm": 0.8598228096961975, "learning_rate": 6.735367470703691e-05, "loss": 0.3499, "step": 25070 }, { "epoch": 14.59837019790454, "grad_norm": 0.8143960237503052, "learning_rate": 6.732782729393379e-05, "loss": 0.3574, "step": 25080 }, { "epoch": 14.60419091967404, "grad_norm": 0.8501980304718018, "learning_rate": 6.730197461709425e-05, "loss": 0.3583, "step": 25090 }, { "epoch": 14.610011641443538, "grad_norm": 0.885576069355011, "learning_rate": 6.727611668437164e-05, "loss": 0.3647, "step": 25100 }, { "epoch": 14.615832363213038, "grad_norm": 0.9724550843238831, "learning_rate": 6.725025350362094e-05, "loss": 0.3822, "step": 25110 }, { "epoch": 14.621653084982539, "grad_norm": 0.9274489283561707, "learning_rate": 6.72243850826987e-05, "loss": 0.3639, "step": 25120 }, { "epoch": 14.627473806752038, "grad_norm": 0.8872038125991821, "learning_rate": 6.719851142946305e-05, "loss": 0.3517, "step": 25130 }, { "epoch": 14.633294528521537, "grad_norm": 0.8169835209846497, "learning_rate": 6.717263255177372e-05, "loss": 0.3689, "step": 25140 }, { "epoch": 14.639115250291036, "grad_norm": 0.8366329073905945, "learning_rate": 6.714674845749205e-05, "loss": 0.3737, "step": 25150 }, { "epoch": 14.644935972060535, "grad_norm": 0.9830068945884705, "learning_rate": 6.712085915448092e-05, "loss": 0.3696, "step": 25160 }, { "epoch": 14.650756693830035, "grad_norm": 0.8020924925804138, "learning_rate": 6.709496465060486e-05, "loss": 0.3603, "step": 25170 }, { "epoch": 14.656577415599534, "grad_norm": 0.8826188445091248, "learning_rate": 6.706906495372987e-05, "loss": 0.3587, "step": 25180 }, { "epoch": 14.662398137369033, "grad_norm": 0.8326629400253296, "learning_rate": 6.704316007172365e-05, "loss": 0.355, "step": 25190 }, { "epoch": 14.668218859138534, "grad_norm": 0.8873414993286133, "learning_rate": 6.701725001245539e-05, "loss": 0.3645, "step": 25200 }, { "epoch": 14.674039580908033, "grad_norm": 0.9272871613502502, "learning_rate": 6.699133478379588e-05, "loss": 0.3501, "step": 25210 }, { "epoch": 14.679860302677533, "grad_norm": 0.8985930681228638, "learning_rate": 6.69654143936175e-05, "loss": 0.3576, "step": 25220 }, { "epoch": 14.685681024447032, "grad_norm": 0.8011118769645691, "learning_rate": 6.693948884979419e-05, "loss": 0.361, "step": 25230 }, { "epoch": 14.691501746216531, "grad_norm": 0.9545274376869202, "learning_rate": 6.691355816020142e-05, "loss": 0.3598, "step": 25240 }, { "epoch": 14.69732246798603, "grad_norm": 0.8299897313117981, "learning_rate": 6.688762233271624e-05, "loss": 0.3679, "step": 25250 }, { "epoch": 14.70314318975553, "grad_norm": 0.8447553515434265, "learning_rate": 6.68616813752173e-05, "loss": 0.363, "step": 25260 }, { "epoch": 14.708963911525029, "grad_norm": 1.0025029182434082, "learning_rate": 6.683573529558477e-05, "loss": 0.3782, "step": 25270 }, { "epoch": 14.714784633294528, "grad_norm": 0.8291882872581482, "learning_rate": 6.680978410170037e-05, "loss": 0.3588, "step": 25280 }, { "epoch": 14.720605355064027, "grad_norm": 0.8518157601356506, "learning_rate": 6.678382780144741e-05, "loss": 0.3552, "step": 25290 }, { "epoch": 14.726426076833528, "grad_norm": 0.8448993563652039, "learning_rate": 6.675786640271071e-05, "loss": 0.3517, "step": 25300 }, { "epoch": 14.732246798603027, "grad_norm": 0.802340567111969, "learning_rate": 6.673189991337665e-05, "loss": 0.368, "step": 25310 }, { "epoch": 14.738067520372526, "grad_norm": 0.8431757688522339, "learning_rate": 6.670592834133317e-05, "loss": 0.357, "step": 25320 }, { "epoch": 14.743888242142026, "grad_norm": 0.9485881328582764, "learning_rate": 6.667995169446979e-05, "loss": 0.3557, "step": 25330 }, { "epoch": 14.749708963911525, "grad_norm": 0.8475288152694702, "learning_rate": 6.665396998067747e-05, "loss": 0.3521, "step": 25340 }, { "epoch": 14.755529685681024, "grad_norm": 0.8160470128059387, "learning_rate": 6.66279832078488e-05, "loss": 0.3607, "step": 25350 }, { "epoch": 14.761350407450523, "grad_norm": 0.8066021800041199, "learning_rate": 6.660199138387786e-05, "loss": 0.3499, "step": 25360 }, { "epoch": 14.767171129220023, "grad_norm": 0.8549355268478394, "learning_rate": 6.65759945166603e-05, "loss": 0.3753, "step": 25370 }, { "epoch": 14.772991850989523, "grad_norm": 0.8518051505088806, "learning_rate": 6.654999261409326e-05, "loss": 0.3663, "step": 25380 }, { "epoch": 14.778812572759023, "grad_norm": 0.8104454874992371, "learning_rate": 6.652398568407544e-05, "loss": 0.3527, "step": 25390 }, { "epoch": 14.784633294528522, "grad_norm": 0.7404143810272217, "learning_rate": 6.649797373450707e-05, "loss": 0.3531, "step": 25400 }, { "epoch": 14.790454016298021, "grad_norm": 0.8082414269447327, "learning_rate": 6.647195677328988e-05, "loss": 0.3472, "step": 25410 }, { "epoch": 14.79627473806752, "grad_norm": 0.6992002725601196, "learning_rate": 6.644593480832712e-05, "loss": 0.3574, "step": 25420 }, { "epoch": 14.80209545983702, "grad_norm": 0.9123445749282837, "learning_rate": 6.641990784752363e-05, "loss": 0.3637, "step": 25430 }, { "epoch": 14.807916181606519, "grad_norm": 0.7956994771957397, "learning_rate": 6.639387589878566e-05, "loss": 0.3625, "step": 25440 }, { "epoch": 14.813736903376018, "grad_norm": 0.880879819393158, "learning_rate": 6.636783897002103e-05, "loss": 0.3679, "step": 25450 }, { "epoch": 14.819557625145517, "grad_norm": 0.9167891144752502, "learning_rate": 6.63417970691391e-05, "loss": 0.3733, "step": 25460 }, { "epoch": 14.825378346915018, "grad_norm": 0.8852885365486145, "learning_rate": 6.63157502040507e-05, "loss": 0.3504, "step": 25470 }, { "epoch": 14.831199068684517, "grad_norm": 0.922656774520874, "learning_rate": 6.628969838266819e-05, "loss": 0.3566, "step": 25480 }, { "epoch": 14.837019790454017, "grad_norm": 0.8533841967582703, "learning_rate": 6.626364161290541e-05, "loss": 0.3538, "step": 25490 }, { "epoch": 14.842840512223516, "grad_norm": 0.8182980418205261, "learning_rate": 6.623757990267774e-05, "loss": 0.3629, "step": 25500 }, { "epoch": 14.848661233993015, "grad_norm": 0.866116464138031, "learning_rate": 6.621151325990201e-05, "loss": 0.3765, "step": 25510 }, { "epoch": 14.854481955762514, "grad_norm": 0.7898160815238953, "learning_rate": 6.618544169249657e-05, "loss": 0.3543, "step": 25520 }, { "epoch": 14.860302677532014, "grad_norm": 0.8202651143074036, "learning_rate": 6.615936520838133e-05, "loss": 0.3715, "step": 25530 }, { "epoch": 14.866123399301513, "grad_norm": 1.0881128311157227, "learning_rate": 6.613328381547759e-05, "loss": 0.3669, "step": 25540 }, { "epoch": 14.871944121071014, "grad_norm": 0.924116849899292, "learning_rate": 6.610719752170821e-05, "loss": 0.3587, "step": 25550 }, { "epoch": 14.877764842840513, "grad_norm": 0.8833903670310974, "learning_rate": 6.60811063349975e-05, "loss": 0.3756, "step": 25560 }, { "epoch": 14.883585564610012, "grad_norm": 0.7994251251220703, "learning_rate": 6.605501026327127e-05, "loss": 0.377, "step": 25570 }, { "epoch": 14.889406286379511, "grad_norm": 0.8536614179611206, "learning_rate": 6.602890931445685e-05, "loss": 0.3609, "step": 25580 }, { "epoch": 14.89522700814901, "grad_norm": 0.8062093257904053, "learning_rate": 6.6002803496483e-05, "loss": 0.3583, "step": 25590 }, { "epoch": 14.90104772991851, "grad_norm": 0.8806256055831909, "learning_rate": 6.597669281727997e-05, "loss": 0.3665, "step": 25600 }, { "epoch": 14.906868451688009, "grad_norm": 0.8144710659980774, "learning_rate": 6.595057728477949e-05, "loss": 0.351, "step": 25610 }, { "epoch": 14.912689173457508, "grad_norm": 0.9302346110343933, "learning_rate": 6.59244569069148e-05, "loss": 0.3617, "step": 25620 }, { "epoch": 14.918509895227007, "grad_norm": 0.8503525853157043, "learning_rate": 6.589833169162054e-05, "loss": 0.3689, "step": 25630 }, { "epoch": 14.924330616996507, "grad_norm": 0.8578535318374634, "learning_rate": 6.587220164683291e-05, "loss": 0.3587, "step": 25640 }, { "epoch": 14.930151338766008, "grad_norm": 0.8513970375061035, "learning_rate": 6.58460667804895e-05, "loss": 0.3533, "step": 25650 }, { "epoch": 14.935972060535507, "grad_norm": 0.9219322800636292, "learning_rate": 6.581992710052938e-05, "loss": 0.3594, "step": 25660 }, { "epoch": 14.941792782305006, "grad_norm": 0.93956458568573, "learning_rate": 6.579378261489311e-05, "loss": 0.3623, "step": 25670 }, { "epoch": 14.947613504074505, "grad_norm": 0.7841658592224121, "learning_rate": 6.576763333152268e-05, "loss": 0.3528, "step": 25680 }, { "epoch": 14.953434225844005, "grad_norm": 0.7268421649932861, "learning_rate": 6.574147925836159e-05, "loss": 0.3631, "step": 25690 }, { "epoch": 14.959254947613504, "grad_norm": 0.9407269954681396, "learning_rate": 6.571532040335472e-05, "loss": 0.3697, "step": 25700 }, { "epoch": 14.965075669383003, "grad_norm": 0.9140018820762634, "learning_rate": 6.568915677444845e-05, "loss": 0.3552, "step": 25710 }, { "epoch": 14.970896391152502, "grad_norm": 0.8639675378799438, "learning_rate": 6.56629883795906e-05, "loss": 0.3672, "step": 25720 }, { "epoch": 14.976717112922003, "grad_norm": 0.8225834965705872, "learning_rate": 6.563681522673043e-05, "loss": 0.3719, "step": 25730 }, { "epoch": 14.982537834691502, "grad_norm": 0.7890604734420776, "learning_rate": 6.561063732381867e-05, "loss": 0.3569, "step": 25740 }, { "epoch": 14.988358556461002, "grad_norm": 0.8245360255241394, "learning_rate": 6.558445467880745e-05, "loss": 0.3421, "step": 25750 }, { "epoch": 14.9941792782305, "grad_norm": 0.8102015256881714, "learning_rate": 6.55582672996504e-05, "loss": 0.3606, "step": 25760 }, { "epoch": 15.0, "grad_norm": 0.79194575548172, "learning_rate": 6.553207519430253e-05, "loss": 0.3442, "step": 25770 }, { "epoch": 15.0058207217695, "grad_norm": 0.8890531659126282, "learning_rate": 6.550587837072032e-05, "loss": 0.3859, "step": 25780 }, { "epoch": 15.011641443538998, "grad_norm": 0.8478220701217651, "learning_rate": 6.547967683686166e-05, "loss": 0.3561, "step": 25790 }, { "epoch": 15.017462165308498, "grad_norm": 0.8849816918373108, "learning_rate": 6.545347060068591e-05, "loss": 0.3502, "step": 25800 }, { "epoch": 15.023282887077997, "grad_norm": 0.83918696641922, "learning_rate": 6.542725967015382e-05, "loss": 0.3557, "step": 25810 }, { "epoch": 15.029103608847498, "grad_norm": 0.827459990978241, "learning_rate": 6.540104405322757e-05, "loss": 0.3474, "step": 25820 }, { "epoch": 15.034924330616997, "grad_norm": 0.9590751528739929, "learning_rate": 6.537482375787077e-05, "loss": 0.3578, "step": 25830 }, { "epoch": 15.040745052386496, "grad_norm": 0.7314987182617188, "learning_rate": 6.534859879204845e-05, "loss": 0.3722, "step": 25840 }, { "epoch": 15.046565774155995, "grad_norm": 0.9036800265312195, "learning_rate": 6.532236916372709e-05, "loss": 0.3712, "step": 25850 }, { "epoch": 15.052386495925495, "grad_norm": 0.7526482343673706, "learning_rate": 6.529613488087454e-05, "loss": 0.3659, "step": 25860 }, { "epoch": 15.058207217694994, "grad_norm": 0.8313570022583008, "learning_rate": 6.526989595146009e-05, "loss": 0.3528, "step": 25870 }, { "epoch": 15.064027939464493, "grad_norm": 0.8581215143203735, "learning_rate": 6.524365238345441e-05, "loss": 0.3665, "step": 25880 }, { "epoch": 15.069848661233992, "grad_norm": 0.844317615032196, "learning_rate": 6.521740418482964e-05, "loss": 0.3642, "step": 25890 }, { "epoch": 15.075669383003492, "grad_norm": 0.8446950316429138, "learning_rate": 6.519115136355925e-05, "loss": 0.3589, "step": 25900 }, { "epoch": 15.081490104772993, "grad_norm": 0.8165686726570129, "learning_rate": 6.51648939276182e-05, "loss": 0.3596, "step": 25910 }, { "epoch": 15.087310826542492, "grad_norm": 0.8297081589698792, "learning_rate": 6.513863188498277e-05, "loss": 0.3467, "step": 25920 }, { "epoch": 15.093131548311991, "grad_norm": 0.7651437520980835, "learning_rate": 6.511236524363068e-05, "loss": 0.3539, "step": 25930 }, { "epoch": 15.09895227008149, "grad_norm": 0.8431874513626099, "learning_rate": 6.508609401154104e-05, "loss": 0.3469, "step": 25940 }, { "epoch": 15.10477299185099, "grad_norm": 0.98366379737854, "learning_rate": 6.505981819669439e-05, "loss": 0.3625, "step": 25950 }, { "epoch": 15.110593713620489, "grad_norm": 0.8645537495613098, "learning_rate": 6.503353780707258e-05, "loss": 0.3575, "step": 25960 }, { "epoch": 15.116414435389988, "grad_norm": 0.8917500972747803, "learning_rate": 6.500725285065895e-05, "loss": 0.3724, "step": 25970 }, { "epoch": 15.122235157159487, "grad_norm": 0.8587837219238281, "learning_rate": 6.498096333543813e-05, "loss": 0.3468, "step": 25980 }, { "epoch": 15.128055878928988, "grad_norm": 0.8514759540557861, "learning_rate": 6.49546692693962e-05, "loss": 0.3688, "step": 25990 }, { "epoch": 15.133876600698487, "grad_norm": 0.932528555393219, "learning_rate": 6.492837066052059e-05, "loss": 0.3747, "step": 26000 }, { "epoch": 15.139697322467986, "grad_norm": 0.8583822846412659, "learning_rate": 6.490206751680014e-05, "loss": 0.3488, "step": 26010 }, { "epoch": 15.145518044237486, "grad_norm": 1.02936851978302, "learning_rate": 6.487575984622505e-05, "loss": 0.362, "step": 26020 }, { "epoch": 15.151338766006985, "grad_norm": 0.8690603971481323, "learning_rate": 6.484944765678689e-05, "loss": 0.3725, "step": 26030 }, { "epoch": 15.157159487776484, "grad_norm": 0.7789080142974854, "learning_rate": 6.482313095647861e-05, "loss": 0.3506, "step": 26040 }, { "epoch": 15.162980209545983, "grad_norm": 0.7340533137321472, "learning_rate": 6.479680975329451e-05, "loss": 0.344, "step": 26050 }, { "epoch": 15.168800931315483, "grad_norm": 0.7558485865592957, "learning_rate": 6.477048405523031e-05, "loss": 0.3683, "step": 26060 }, { "epoch": 15.174621653084982, "grad_norm": 0.8031021356582642, "learning_rate": 6.474415387028304e-05, "loss": 0.3479, "step": 26070 }, { "epoch": 15.180442374854483, "grad_norm": 0.8569827079772949, "learning_rate": 6.471781920645114e-05, "loss": 0.3512, "step": 26080 }, { "epoch": 15.186263096623982, "grad_norm": 0.8196824789047241, "learning_rate": 6.469148007173434e-05, "loss": 0.3617, "step": 26090 }, { "epoch": 15.192083818393481, "grad_norm": 0.788165807723999, "learning_rate": 6.466513647413381e-05, "loss": 0.3659, "step": 26100 }, { "epoch": 15.19790454016298, "grad_norm": 0.8888673782348633, "learning_rate": 6.463878842165203e-05, "loss": 0.3613, "step": 26110 }, { "epoch": 15.20372526193248, "grad_norm": 0.8069729208946228, "learning_rate": 6.461243592229286e-05, "loss": 0.3684, "step": 26120 }, { "epoch": 15.209545983701979, "grad_norm": 0.8173063397407532, "learning_rate": 6.458607898406146e-05, "loss": 0.3542, "step": 26130 }, { "epoch": 15.215366705471478, "grad_norm": 0.8634331822395325, "learning_rate": 6.455971761496439e-05, "loss": 0.357, "step": 26140 }, { "epoch": 15.221187427240977, "grad_norm": 0.7633211016654968, "learning_rate": 6.453335182300953e-05, "loss": 0.3659, "step": 26150 }, { "epoch": 15.227008149010477, "grad_norm": 0.837256133556366, "learning_rate": 6.450698161620612e-05, "loss": 0.3667, "step": 26160 }, { "epoch": 15.232828870779977, "grad_norm": 0.8121291399002075, "learning_rate": 6.448060700256473e-05, "loss": 0.365, "step": 26170 }, { "epoch": 15.238649592549477, "grad_norm": 0.8867397308349609, "learning_rate": 6.445422799009726e-05, "loss": 0.3667, "step": 26180 }, { "epoch": 15.244470314318976, "grad_norm": 0.8844960927963257, "learning_rate": 6.442784458681699e-05, "loss": 0.3486, "step": 26190 }, { "epoch": 15.250291036088475, "grad_norm": 0.824310839176178, "learning_rate": 6.440145680073847e-05, "loss": 0.3614, "step": 26200 }, { "epoch": 15.256111757857974, "grad_norm": 0.8348096609115601, "learning_rate": 6.437506463987762e-05, "loss": 0.3661, "step": 26210 }, { "epoch": 15.261932479627474, "grad_norm": 0.8230703473091125, "learning_rate": 6.434866811225168e-05, "loss": 0.3694, "step": 26220 }, { "epoch": 15.267753201396973, "grad_norm": 0.8683792948722839, "learning_rate": 6.432226722587923e-05, "loss": 0.3661, "step": 26230 }, { "epoch": 15.273573923166472, "grad_norm": 0.8419182896614075, "learning_rate": 6.429586198878015e-05, "loss": 0.3641, "step": 26240 }, { "epoch": 15.279394644935971, "grad_norm": 0.8182449340820312, "learning_rate": 6.426945240897566e-05, "loss": 0.3643, "step": 26250 }, { "epoch": 15.285215366705472, "grad_norm": 0.8333450555801392, "learning_rate": 6.424303849448829e-05, "loss": 0.356, "step": 26260 }, { "epoch": 15.291036088474971, "grad_norm": 0.9059413075447083, "learning_rate": 6.42166202533419e-05, "loss": 0.3584, "step": 26270 }, { "epoch": 15.29685681024447, "grad_norm": 0.828937828540802, "learning_rate": 6.419019769356164e-05, "loss": 0.3457, "step": 26280 }, { "epoch": 15.30267753201397, "grad_norm": 0.7799766659736633, "learning_rate": 6.416377082317398e-05, "loss": 0.3735, "step": 26290 }, { "epoch": 15.308498253783469, "grad_norm": 0.8599786758422852, "learning_rate": 6.413733965020674e-05, "loss": 0.3559, "step": 26300 }, { "epoch": 15.314318975552968, "grad_norm": 0.7386131286621094, "learning_rate": 6.411090418268896e-05, "loss": 0.3538, "step": 26310 }, { "epoch": 15.320139697322467, "grad_norm": 0.8272466659545898, "learning_rate": 6.408446442865109e-05, "loss": 0.3513, "step": 26320 }, { "epoch": 15.325960419091967, "grad_norm": 0.8149201273918152, "learning_rate": 6.405802039612479e-05, "loss": 0.3589, "step": 26330 }, { "epoch": 15.331781140861466, "grad_norm": 0.9657898545265198, "learning_rate": 6.403157209314308e-05, "loss": 0.3585, "step": 26340 }, { "epoch": 15.337601862630967, "grad_norm": 0.8823891878128052, "learning_rate": 6.400511952774024e-05, "loss": 0.3664, "step": 26350 }, { "epoch": 15.343422584400466, "grad_norm": 0.8154417276382446, "learning_rate": 6.397866270795187e-05, "loss": 0.3553, "step": 26360 }, { "epoch": 15.349243306169965, "grad_norm": 0.947446346282959, "learning_rate": 6.395220164181489e-05, "loss": 0.3669, "step": 26370 }, { "epoch": 15.355064027939465, "grad_norm": 0.8857580423355103, "learning_rate": 6.39257363373674e-05, "loss": 0.3653, "step": 26380 }, { "epoch": 15.360884749708964, "grad_norm": 0.7945115566253662, "learning_rate": 6.389926680264892e-05, "loss": 0.3449, "step": 26390 }, { "epoch": 15.366705471478463, "grad_norm": 0.8686063289642334, "learning_rate": 6.387279304570017e-05, "loss": 0.3591, "step": 26400 }, { "epoch": 15.372526193247962, "grad_norm": 0.8673288226127625, "learning_rate": 6.384631507456319e-05, "loss": 0.3559, "step": 26410 }, { "epoch": 15.378346915017461, "grad_norm": 0.8106796741485596, "learning_rate": 6.381983289728126e-05, "loss": 0.3608, "step": 26420 }, { "epoch": 15.384167636786962, "grad_norm": 0.8582190871238708, "learning_rate": 6.3793346521899e-05, "loss": 0.3575, "step": 26430 }, { "epoch": 15.389988358556462, "grad_norm": 0.7970398664474487, "learning_rate": 6.376685595646226e-05, "loss": 0.3598, "step": 26440 }, { "epoch": 15.39580908032596, "grad_norm": 0.8340988159179688, "learning_rate": 6.374036120901816e-05, "loss": 0.3594, "step": 26450 }, { "epoch": 15.40162980209546, "grad_norm": 0.8276436924934387, "learning_rate": 6.371386228761514e-05, "loss": 0.3693, "step": 26460 }, { "epoch": 15.40745052386496, "grad_norm": 0.8086073398590088, "learning_rate": 6.368735920030283e-05, "loss": 0.3482, "step": 26470 }, { "epoch": 15.413271245634458, "grad_norm": 0.866829514503479, "learning_rate": 6.366085195513218e-05, "loss": 0.3531, "step": 26480 }, { "epoch": 15.419091967403958, "grad_norm": 0.7474760413169861, "learning_rate": 6.363434056015543e-05, "loss": 0.3421, "step": 26490 }, { "epoch": 15.424912689173457, "grad_norm": 0.7839873433113098, "learning_rate": 6.360782502342599e-05, "loss": 0.3584, "step": 26500 }, { "epoch": 15.430733410942956, "grad_norm": 0.8620923161506653, "learning_rate": 6.358130535299862e-05, "loss": 0.3577, "step": 26510 }, { "epoch": 15.436554132712457, "grad_norm": 0.8250285983085632, "learning_rate": 6.355478155692926e-05, "loss": 0.3755, "step": 26520 }, { "epoch": 15.442374854481956, "grad_norm": 0.7982230186462402, "learning_rate": 6.352825364327517e-05, "loss": 0.3414, "step": 26530 }, { "epoch": 15.448195576251456, "grad_norm": 0.8053463101387024, "learning_rate": 6.350172162009482e-05, "loss": 0.3733, "step": 26540 }, { "epoch": 15.454016298020955, "grad_norm": 0.8216500282287598, "learning_rate": 6.347518549544793e-05, "loss": 0.3629, "step": 26550 }, { "epoch": 15.459837019790454, "grad_norm": 0.7606666684150696, "learning_rate": 6.344864527739547e-05, "loss": 0.346, "step": 26560 }, { "epoch": 15.465657741559953, "grad_norm": 0.8795208930969238, "learning_rate": 6.342210097399966e-05, "loss": 0.3653, "step": 26570 }, { "epoch": 15.471478463329452, "grad_norm": 0.88360196352005, "learning_rate": 6.339555259332398e-05, "loss": 0.3648, "step": 26580 }, { "epoch": 15.477299185098952, "grad_norm": 0.7356986999511719, "learning_rate": 6.33690001434331e-05, "loss": 0.3481, "step": 26590 }, { "epoch": 15.48311990686845, "grad_norm": 0.8378396034240723, "learning_rate": 6.334244363239296e-05, "loss": 0.3637, "step": 26600 }, { "epoch": 15.488940628637952, "grad_norm": 0.8419702649116516, "learning_rate": 6.331588306827073e-05, "loss": 0.3535, "step": 26610 }, { "epoch": 15.494761350407451, "grad_norm": 0.8516611456871033, "learning_rate": 6.328931845913483e-05, "loss": 0.3416, "step": 26620 }, { "epoch": 15.50058207217695, "grad_norm": 0.8559938669204712, "learning_rate": 6.326274981305484e-05, "loss": 0.3715, "step": 26630 }, { "epoch": 15.50640279394645, "grad_norm": 0.8111240863800049, "learning_rate": 6.323617713810166e-05, "loss": 0.3725, "step": 26640 }, { "epoch": 15.512223515715949, "grad_norm": 0.8479250073432922, "learning_rate": 6.320960044234734e-05, "loss": 0.376, "step": 26650 }, { "epoch": 15.518044237485448, "grad_norm": 0.8218114972114563, "learning_rate": 6.318301973386518e-05, "loss": 0.3512, "step": 26660 }, { "epoch": 15.523864959254947, "grad_norm": 0.8793686628341675, "learning_rate": 6.315643502072971e-05, "loss": 0.3508, "step": 26670 }, { "epoch": 15.529685681024446, "grad_norm": 0.7312950491905212, "learning_rate": 6.312984631101667e-05, "loss": 0.3519, "step": 26680 }, { "epoch": 15.535506402793946, "grad_norm": 0.8220320343971252, "learning_rate": 6.310325361280297e-05, "loss": 0.3545, "step": 26690 }, { "epoch": 15.541327124563447, "grad_norm": 0.8440669775009155, "learning_rate": 6.30766569341668e-05, "loss": 0.3561, "step": 26700 }, { "epoch": 15.547147846332946, "grad_norm": 0.7695035338401794, "learning_rate": 6.305005628318753e-05, "loss": 0.3627, "step": 26710 }, { "epoch": 15.552968568102445, "grad_norm": 0.9124706387519836, "learning_rate": 6.302345166794572e-05, "loss": 0.3553, "step": 26720 }, { "epoch": 15.558789289871944, "grad_norm": 0.8581624627113342, "learning_rate": 6.299684309652316e-05, "loss": 0.3433, "step": 26730 }, { "epoch": 15.564610011641443, "grad_norm": 0.7871854901313782, "learning_rate": 6.297023057700283e-05, "loss": 0.3531, "step": 26740 }, { "epoch": 15.570430733410943, "grad_norm": 0.8074427247047424, "learning_rate": 6.294361411746891e-05, "loss": 0.3577, "step": 26750 }, { "epoch": 15.576251455180442, "grad_norm": 0.8499223589897156, "learning_rate": 6.291699372600677e-05, "loss": 0.3572, "step": 26760 }, { "epoch": 15.582072176949941, "grad_norm": 0.8971974849700928, "learning_rate": 6.2890369410703e-05, "loss": 0.3677, "step": 26770 }, { "epoch": 15.587892898719442, "grad_norm": 0.9249571561813354, "learning_rate": 6.286374117964534e-05, "loss": 0.3542, "step": 26780 }, { "epoch": 15.593713620488941, "grad_norm": 0.7961941361427307, "learning_rate": 6.283710904092277e-05, "loss": 0.3611, "step": 26790 }, { "epoch": 15.59953434225844, "grad_norm": 0.8245217800140381, "learning_rate": 6.281047300262542e-05, "loss": 0.3684, "step": 26800 }, { "epoch": 15.60535506402794, "grad_norm": 0.8921163082122803, "learning_rate": 6.278383307284461e-05, "loss": 0.3664, "step": 26810 }, { "epoch": 15.611175785797439, "grad_norm": 0.9463527202606201, "learning_rate": 6.275718925967284e-05, "loss": 0.3572, "step": 26820 }, { "epoch": 15.616996507566938, "grad_norm": 0.8397687077522278, "learning_rate": 6.273054157120382e-05, "loss": 0.3544, "step": 26830 }, { "epoch": 15.622817229336437, "grad_norm": 0.9004577994346619, "learning_rate": 6.270389001553238e-05, "loss": 0.3636, "step": 26840 }, { "epoch": 15.628637951105937, "grad_norm": 0.8625347018241882, "learning_rate": 6.26772346007546e-05, "loss": 0.3611, "step": 26850 }, { "epoch": 15.634458672875436, "grad_norm": 0.7825828790664673, "learning_rate": 6.265057533496767e-05, "loss": 0.3597, "step": 26860 }, { "epoch": 15.640279394644937, "grad_norm": 0.8272867798805237, "learning_rate": 6.262391222626997e-05, "loss": 0.3506, "step": 26870 }, { "epoch": 15.646100116414436, "grad_norm": 0.7491023540496826, "learning_rate": 6.259724528276106e-05, "loss": 0.3692, "step": 26880 }, { "epoch": 15.651920838183935, "grad_norm": 0.7978399991989136, "learning_rate": 6.257057451254162e-05, "loss": 0.3419, "step": 26890 }, { "epoch": 15.657741559953434, "grad_norm": 0.7905765175819397, "learning_rate": 6.254389992371357e-05, "loss": 0.3593, "step": 26900 }, { "epoch": 15.663562281722934, "grad_norm": 0.8977292776107788, "learning_rate": 6.25172215243799e-05, "loss": 0.3603, "step": 26910 }, { "epoch": 15.669383003492433, "grad_norm": 0.872273325920105, "learning_rate": 6.249053932264486e-05, "loss": 0.3438, "step": 26920 }, { "epoch": 15.675203725261932, "grad_norm": 0.8119707107543945, "learning_rate": 6.246385332661376e-05, "loss": 0.3451, "step": 26930 }, { "epoch": 15.681024447031431, "grad_norm": 0.8773928284645081, "learning_rate": 6.24371635443931e-05, "loss": 0.3708, "step": 26940 }, { "epoch": 15.68684516880093, "grad_norm": 0.8209101557731628, "learning_rate": 6.241046998409054e-05, "loss": 0.3563, "step": 26950 }, { "epoch": 15.692665890570431, "grad_norm": 0.8075182437896729, "learning_rate": 6.238377265381489e-05, "loss": 0.3652, "step": 26960 }, { "epoch": 15.69848661233993, "grad_norm": 0.8497335910797119, "learning_rate": 6.235707156167607e-05, "loss": 0.3557, "step": 26970 }, { "epoch": 15.70430733410943, "grad_norm": 0.7893561124801636, "learning_rate": 6.233036671578519e-05, "loss": 0.357, "step": 26980 }, { "epoch": 15.71012805587893, "grad_norm": 0.8800008893013, "learning_rate": 6.230365812425445e-05, "loss": 0.3651, "step": 26990 }, { "epoch": 15.715948777648428, "grad_norm": 0.8507137894630432, "learning_rate": 6.227694579519724e-05, "loss": 0.3701, "step": 27000 }, { "epoch": 15.721769499417928, "grad_norm": 0.9431421756744385, "learning_rate": 6.225022973672805e-05, "loss": 0.3788, "step": 27010 }, { "epoch": 15.727590221187427, "grad_norm": 0.8988215327262878, "learning_rate": 6.222350995696253e-05, "loss": 0.3609, "step": 27020 }, { "epoch": 15.733410942956926, "grad_norm": 0.978313148021698, "learning_rate": 6.21967864640174e-05, "loss": 0.3553, "step": 27030 }, { "epoch": 15.739231664726425, "grad_norm": 0.8811371922492981, "learning_rate": 6.217005926601059e-05, "loss": 0.3441, "step": 27040 }, { "epoch": 15.745052386495926, "grad_norm": 0.8417131900787354, "learning_rate": 6.214332837106111e-05, "loss": 0.3557, "step": 27050 }, { "epoch": 15.750873108265425, "grad_norm": 0.8079370260238647, "learning_rate": 6.21165937872891e-05, "loss": 0.3582, "step": 27060 }, { "epoch": 15.756693830034925, "grad_norm": 0.873866081237793, "learning_rate": 6.208985552281582e-05, "loss": 0.3529, "step": 27070 }, { "epoch": 15.762514551804424, "grad_norm": 0.819521427154541, "learning_rate": 6.206311358576364e-05, "loss": 0.3521, "step": 27080 }, { "epoch": 15.768335273573923, "grad_norm": 0.9079386591911316, "learning_rate": 6.203636798425608e-05, "loss": 0.3673, "step": 27090 }, { "epoch": 15.774155995343422, "grad_norm": 0.7882280945777893, "learning_rate": 6.20096187264177e-05, "loss": 0.3579, "step": 27100 }, { "epoch": 15.779976717112921, "grad_norm": 0.850636899471283, "learning_rate": 6.198286582037425e-05, "loss": 0.3549, "step": 27110 }, { "epoch": 15.78579743888242, "grad_norm": 0.8562852144241333, "learning_rate": 6.195610927425256e-05, "loss": 0.3571, "step": 27120 }, { "epoch": 15.791618160651922, "grad_norm": 0.776759147644043, "learning_rate": 6.192934909618056e-05, "loss": 0.3604, "step": 27130 }, { "epoch": 15.797438882421421, "grad_norm": 0.7809591293334961, "learning_rate": 6.190258529428728e-05, "loss": 0.3555, "step": 27140 }, { "epoch": 15.80325960419092, "grad_norm": 0.8698499798774719, "learning_rate": 6.187581787670285e-05, "loss": 0.3673, "step": 27150 }, { "epoch": 15.80908032596042, "grad_norm": 0.8490100502967834, "learning_rate": 6.184904685155852e-05, "loss": 0.3591, "step": 27160 }, { "epoch": 15.814901047729919, "grad_norm": 0.8671167492866516, "learning_rate": 6.18222722269866e-05, "loss": 0.3482, "step": 27170 }, { "epoch": 15.820721769499418, "grad_norm": 0.8000816106796265, "learning_rate": 6.179549401112053e-05, "loss": 0.3745, "step": 27180 }, { "epoch": 15.826542491268917, "grad_norm": 0.8356051445007324, "learning_rate": 6.176871221209482e-05, "loss": 0.3549, "step": 27190 }, { "epoch": 15.832363213038416, "grad_norm": 0.8911503553390503, "learning_rate": 6.174192683804508e-05, "loss": 0.3552, "step": 27200 }, { "epoch": 15.838183934807915, "grad_norm": 0.9509402513504028, "learning_rate": 6.1715137897108e-05, "loss": 0.3495, "step": 27210 }, { "epoch": 15.844004656577416, "grad_norm": 0.8156218528747559, "learning_rate": 6.168834539742134e-05, "loss": 0.3612, "step": 27220 }, { "epoch": 15.849825378346916, "grad_norm": 0.7960091829299927, "learning_rate": 6.166154934712397e-05, "loss": 0.3445, "step": 27230 }, { "epoch": 15.855646100116415, "grad_norm": 0.7989662289619446, "learning_rate": 6.163474975435581e-05, "loss": 0.3463, "step": 27240 }, { "epoch": 15.861466821885914, "grad_norm": 0.8087363243103027, "learning_rate": 6.160794662725787e-05, "loss": 0.352, "step": 27250 }, { "epoch": 15.867287543655413, "grad_norm": 0.7886871099472046, "learning_rate": 6.158113997397222e-05, "loss": 0.3591, "step": 27260 }, { "epoch": 15.873108265424912, "grad_norm": 0.8080302476882935, "learning_rate": 6.155432980264205e-05, "loss": 0.3438, "step": 27270 }, { "epoch": 15.878928987194412, "grad_norm": 0.7845944166183472, "learning_rate": 6.152751612141156e-05, "loss": 0.3521, "step": 27280 }, { "epoch": 15.884749708963911, "grad_norm": 0.8057480454444885, "learning_rate": 6.150069893842602e-05, "loss": 0.3617, "step": 27290 }, { "epoch": 15.89057043073341, "grad_norm": 0.8119341135025024, "learning_rate": 6.147387826183182e-05, "loss": 0.3669, "step": 27300 }, { "epoch": 15.896391152502911, "grad_norm": 0.7551400661468506, "learning_rate": 6.144705409977635e-05, "loss": 0.3518, "step": 27310 }, { "epoch": 15.90221187427241, "grad_norm": 0.7514864802360535, "learning_rate": 6.142022646040808e-05, "loss": 0.3463, "step": 27320 }, { "epoch": 15.90803259604191, "grad_norm": 0.7684576511383057, "learning_rate": 6.139339535187653e-05, "loss": 0.36, "step": 27330 }, { "epoch": 15.913853317811409, "grad_norm": 0.8893705606460571, "learning_rate": 6.136656078233232e-05, "loss": 0.3618, "step": 27340 }, { "epoch": 15.919674039580908, "grad_norm": 0.9159356355667114, "learning_rate": 6.133972275992707e-05, "loss": 0.3713, "step": 27350 }, { "epoch": 15.925494761350407, "grad_norm": 0.8188374042510986, "learning_rate": 6.131288129281342e-05, "loss": 0.3689, "step": 27360 }, { "epoch": 15.931315483119906, "grad_norm": 0.8410196900367737, "learning_rate": 6.128603638914516e-05, "loss": 0.361, "step": 27370 }, { "epoch": 15.937136204889406, "grad_norm": 0.7941893339157104, "learning_rate": 6.125918805707704e-05, "loss": 0.3533, "step": 27380 }, { "epoch": 15.942956926658905, "grad_norm": 0.8476890325546265, "learning_rate": 6.123233630476485e-05, "loss": 0.3622, "step": 27390 }, { "epoch": 15.948777648428406, "grad_norm": 0.83357834815979, "learning_rate": 6.120548114036547e-05, "loss": 0.3649, "step": 27400 }, { "epoch": 15.954598370197905, "grad_norm": 0.8665810823440552, "learning_rate": 6.117862257203679e-05, "loss": 0.3607, "step": 27410 }, { "epoch": 15.960419091967404, "grad_norm": 0.8509446382522583, "learning_rate": 6.115176060793771e-05, "loss": 0.3638, "step": 27420 }, { "epoch": 15.966239813736903, "grad_norm": 0.8449386954307556, "learning_rate": 6.112489525622822e-05, "loss": 0.3565, "step": 27430 }, { "epoch": 15.972060535506403, "grad_norm": 0.7431774735450745, "learning_rate": 6.109802652506928e-05, "loss": 0.3458, "step": 27440 }, { "epoch": 15.977881257275902, "grad_norm": 0.7034810781478882, "learning_rate": 6.107115442262291e-05, "loss": 0.3559, "step": 27450 }, { "epoch": 15.983701979045401, "grad_norm": 0.8041349649429321, "learning_rate": 6.104427895705214e-05, "loss": 0.3638, "step": 27460 }, { "epoch": 15.9895227008149, "grad_norm": 0.8368820548057556, "learning_rate": 6.101740013652103e-05, "loss": 0.3492, "step": 27470 }, { "epoch": 15.995343422584401, "grad_norm": 0.7631123661994934, "learning_rate": 6.099051796919465e-05, "loss": 0.3546, "step": 27480 }, { "epoch": 16.0011641443539, "grad_norm": 0.8535446524620056, "learning_rate": 6.096363246323911e-05, "loss": 0.366, "step": 27490 }, { "epoch": 16.0069848661234, "grad_norm": 0.8209224939346313, "learning_rate": 6.0936743626821504e-05, "loss": 0.3605, "step": 27500 }, { "epoch": 16.0128055878929, "grad_norm": 0.769600510597229, "learning_rate": 6.090985146810996e-05, "loss": 0.3543, "step": 27510 }, { "epoch": 16.018626309662398, "grad_norm": 0.8093259334564209, "learning_rate": 6.088295599527357e-05, "loss": 0.3747, "step": 27520 }, { "epoch": 16.024447031431897, "grad_norm": 0.8676464557647705, "learning_rate": 6.085605721648252e-05, "loss": 0.3468, "step": 27530 }, { "epoch": 16.030267753201397, "grad_norm": 0.8484489321708679, "learning_rate": 6.082915513990792e-05, "loss": 0.3624, "step": 27540 }, { "epoch": 16.036088474970896, "grad_norm": 0.7814247608184814, "learning_rate": 6.080224977372192e-05, "loss": 0.3507, "step": 27550 }, { "epoch": 16.041909196740395, "grad_norm": 0.7503460049629211, "learning_rate": 6.0775341126097666e-05, "loss": 0.3424, "step": 27560 }, { "epoch": 16.047729918509894, "grad_norm": 0.8474236726760864, "learning_rate": 6.074842920520926e-05, "loss": 0.348, "step": 27570 }, { "epoch": 16.053550640279393, "grad_norm": 0.8428366184234619, "learning_rate": 6.072151401923186e-05, "loss": 0.3573, "step": 27580 }, { "epoch": 16.059371362048893, "grad_norm": 0.8192843794822693, "learning_rate": 6.069459557634159e-05, "loss": 0.3374, "step": 27590 }, { "epoch": 16.065192083818392, "grad_norm": 0.8072295784950256, "learning_rate": 6.066767388471557e-05, "loss": 0.3625, "step": 27600 }, { "epoch": 16.07101280558789, "grad_norm": 0.8207958340644836, "learning_rate": 6.064074895253188e-05, "loss": 0.3593, "step": 27610 }, { "epoch": 16.076833527357394, "grad_norm": 0.8421236872673035, "learning_rate": 6.061382078796961e-05, "loss": 0.3502, "step": 27620 }, { "epoch": 16.082654249126893, "grad_norm": 0.8403932452201843, "learning_rate": 6.0586889399208814e-05, "loss": 0.3572, "step": 27630 }, { "epoch": 16.088474970896392, "grad_norm": 0.8726142048835754, "learning_rate": 6.0559954794430565e-05, "loss": 0.354, "step": 27640 }, { "epoch": 16.09429569266589, "grad_norm": 0.8089408278465271, "learning_rate": 6.053301698181687e-05, "loss": 0.3616, "step": 27650 }, { "epoch": 16.10011641443539, "grad_norm": 0.7932350039482117, "learning_rate": 6.0506075969550725e-05, "loss": 0.3547, "step": 27660 }, { "epoch": 16.10593713620489, "grad_norm": 0.9726024270057678, "learning_rate": 6.047913176581609e-05, "loss": 0.3695, "step": 27670 }, { "epoch": 16.11175785797439, "grad_norm": 0.8889328837394714, "learning_rate": 6.0452184378797904e-05, "loss": 0.3595, "step": 27680 }, { "epoch": 16.11757857974389, "grad_norm": 0.86598140001297, "learning_rate": 6.042523381668209e-05, "loss": 0.3526, "step": 27690 }, { "epoch": 16.123399301513388, "grad_norm": 0.8732963800430298, "learning_rate": 6.03982800876555e-05, "loss": 0.351, "step": 27700 }, { "epoch": 16.129220023282887, "grad_norm": 0.742963433265686, "learning_rate": 6.0371323199905975e-05, "loss": 0.3429, "step": 27710 }, { "epoch": 16.135040745052386, "grad_norm": 0.8058791160583496, "learning_rate": 6.03443631616223e-05, "loss": 0.3532, "step": 27720 }, { "epoch": 16.140861466821885, "grad_norm": 0.8026687502861023, "learning_rate": 6.031739998099421e-05, "loss": 0.3641, "step": 27730 }, { "epoch": 16.146682188591384, "grad_norm": 0.7769772410392761, "learning_rate": 6.029043366621243e-05, "loss": 0.3695, "step": 27740 }, { "epoch": 16.152502910360884, "grad_norm": 0.8922532796859741, "learning_rate": 6.0263464225468615e-05, "loss": 0.3615, "step": 27750 }, { "epoch": 16.158323632130383, "grad_norm": 0.8134825229644775, "learning_rate": 6.023649166695534e-05, "loss": 0.3552, "step": 27760 }, { "epoch": 16.164144353899882, "grad_norm": 0.8476783037185669, "learning_rate": 6.0209515998866186e-05, "loss": 0.3631, "step": 27770 }, { "epoch": 16.16996507566938, "grad_norm": 0.8164272308349609, "learning_rate": 6.018253722939563e-05, "loss": 0.3615, "step": 27780 }, { "epoch": 16.175785797438884, "grad_norm": 0.8596774339675903, "learning_rate": 6.015555536673914e-05, "loss": 0.3542, "step": 27790 }, { "epoch": 16.181606519208383, "grad_norm": 0.7570109963417053, "learning_rate": 6.0128570419093054e-05, "loss": 0.3502, "step": 27800 }, { "epoch": 16.187427240977883, "grad_norm": 0.7716780304908752, "learning_rate": 6.010158239465471e-05, "loss": 0.3533, "step": 27810 }, { "epoch": 16.19324796274738, "grad_norm": 0.876576840877533, "learning_rate": 6.007459130162235e-05, "loss": 0.3626, "step": 27820 }, { "epoch": 16.19906868451688, "grad_norm": 0.7927570343017578, "learning_rate": 6.004759714819516e-05, "loss": 0.3488, "step": 27830 }, { "epoch": 16.20488940628638, "grad_norm": 0.890956461429596, "learning_rate": 6.002059994257323e-05, "loss": 0.3609, "step": 27840 }, { "epoch": 16.21071012805588, "grad_norm": 0.8078141212463379, "learning_rate": 5.999359969295764e-05, "loss": 0.3641, "step": 27850 }, { "epoch": 16.21653084982538, "grad_norm": 0.7889083623886108, "learning_rate": 5.9966596407550314e-05, "loss": 0.3435, "step": 27860 }, { "epoch": 16.222351571594878, "grad_norm": 0.8036587834358215, "learning_rate": 5.993959009455416e-05, "loss": 0.3505, "step": 27870 }, { "epoch": 16.228172293364377, "grad_norm": 0.9406036734580994, "learning_rate": 5.991258076217298e-05, "loss": 0.3447, "step": 27880 }, { "epoch": 16.233993015133876, "grad_norm": 0.9369323253631592, "learning_rate": 5.988556841861147e-05, "loss": 0.3632, "step": 27890 }, { "epoch": 16.239813736903375, "grad_norm": 0.8358119130134583, "learning_rate": 5.985855307207531e-05, "loss": 0.3546, "step": 27900 }, { "epoch": 16.245634458672875, "grad_norm": 0.8518297076225281, "learning_rate": 5.9831534730771e-05, "loss": 0.3574, "step": 27910 }, { "epoch": 16.251455180442374, "grad_norm": 0.8356350064277649, "learning_rate": 5.980451340290605e-05, "loss": 0.3583, "step": 27920 }, { "epoch": 16.257275902211873, "grad_norm": 0.8452991247177124, "learning_rate": 5.97774890966888e-05, "loss": 0.3586, "step": 27930 }, { "epoch": 16.263096623981372, "grad_norm": 0.7291198372840881, "learning_rate": 5.975046182032851e-05, "loss": 0.3548, "step": 27940 }, { "epoch": 16.26891734575087, "grad_norm": 0.8008617162704468, "learning_rate": 5.972343158203537e-05, "loss": 0.3563, "step": 27950 }, { "epoch": 16.274738067520374, "grad_norm": 0.8157156109809875, "learning_rate": 5.969639839002045e-05, "loss": 0.3442, "step": 27960 }, { "epoch": 16.280558789289874, "grad_norm": 0.7771471738815308, "learning_rate": 5.966936225249572e-05, "loss": 0.3643, "step": 27970 }, { "epoch": 16.286379511059373, "grad_norm": 0.9235126972198486, "learning_rate": 5.9642323177674044e-05, "loss": 0.3516, "step": 27980 }, { "epoch": 16.292200232828872, "grad_norm": 0.7628045678138733, "learning_rate": 5.9615281173769154e-05, "loss": 0.344, "step": 27990 }, { "epoch": 16.29802095459837, "grad_norm": 0.7585437893867493, "learning_rate": 5.958823624899574e-05, "loss": 0.3579, "step": 28000 }, { "epoch": 16.30384167636787, "grad_norm": 0.8247695565223694, "learning_rate": 5.956118841156933e-05, "loss": 0.3514, "step": 28010 }, { "epoch": 16.30966239813737, "grad_norm": 0.9235171675682068, "learning_rate": 5.953413766970631e-05, "loss": 0.3703, "step": 28020 }, { "epoch": 16.31548311990687, "grad_norm": 0.7781187295913696, "learning_rate": 5.9507084031624e-05, "loss": 0.3593, "step": 28030 }, { "epoch": 16.321303841676368, "grad_norm": 0.8051145076751709, "learning_rate": 5.948002750554058e-05, "loss": 0.3666, "step": 28040 }, { "epoch": 16.327124563445867, "grad_norm": 0.842210054397583, "learning_rate": 5.9452968099675124e-05, "loss": 0.3454, "step": 28050 }, { "epoch": 16.332945285215366, "grad_norm": 0.8030927181243896, "learning_rate": 5.9425905822247527e-05, "loss": 0.3573, "step": 28060 }, { "epoch": 16.338766006984866, "grad_norm": 0.8120151162147522, "learning_rate": 5.939884068147864e-05, "loss": 0.3359, "step": 28070 }, { "epoch": 16.344586728754365, "grad_norm": 0.7600579857826233, "learning_rate": 5.937177268559011e-05, "loss": 0.3674, "step": 28080 }, { "epoch": 16.350407450523864, "grad_norm": 0.8106238842010498, "learning_rate": 5.934470184280448e-05, "loss": 0.3672, "step": 28090 }, { "epoch": 16.356228172293363, "grad_norm": 0.7792031168937683, "learning_rate": 5.931762816134516e-05, "loss": 0.3645, "step": 28100 }, { "epoch": 16.362048894062863, "grad_norm": 0.8433620929718018, "learning_rate": 5.9290551649436434e-05, "loss": 0.3533, "step": 28110 }, { "epoch": 16.36786961583236, "grad_norm": 0.7339833378791809, "learning_rate": 5.9263472315303416e-05, "loss": 0.3654, "step": 28120 }, { "epoch": 16.37369033760186, "grad_norm": 0.8230981826782227, "learning_rate": 5.9236390167172096e-05, "loss": 0.3484, "step": 28130 }, { "epoch": 16.379511059371364, "grad_norm": 0.7815616130828857, "learning_rate": 5.920930521326932e-05, "loss": 0.338, "step": 28140 }, { "epoch": 16.385331781140863, "grad_norm": 0.7505594491958618, "learning_rate": 5.918221746182276e-05, "loss": 0.3559, "step": 28150 }, { "epoch": 16.391152502910362, "grad_norm": 0.7830066084861755, "learning_rate": 5.9155126921061e-05, "loss": 0.3461, "step": 28160 }, { "epoch": 16.39697322467986, "grad_norm": 0.741695761680603, "learning_rate": 5.91280335992134e-05, "loss": 0.3537, "step": 28170 }, { "epoch": 16.40279394644936, "grad_norm": 0.766599714756012, "learning_rate": 5.91009375045102e-05, "loss": 0.345, "step": 28180 }, { "epoch": 16.40861466821886, "grad_norm": 0.7681957483291626, "learning_rate": 5.9073838645182476e-05, "loss": 0.3407, "step": 28190 }, { "epoch": 16.41443538998836, "grad_norm": 0.8440104722976685, "learning_rate": 5.904673702946217e-05, "loss": 0.3658, "step": 28200 }, { "epoch": 16.42025611175786, "grad_norm": 0.8585644364356995, "learning_rate": 5.9019632665582004e-05, "loss": 0.3569, "step": 28210 }, { "epoch": 16.426076833527357, "grad_norm": 0.8047969937324524, "learning_rate": 5.899252556177559e-05, "loss": 0.3572, "step": 28220 }, { "epoch": 16.431897555296857, "grad_norm": 0.8405523896217346, "learning_rate": 5.896541572627735e-05, "loss": 0.3671, "step": 28230 }, { "epoch": 16.437718277066356, "grad_norm": 0.8687349557876587, "learning_rate": 5.893830316732253e-05, "loss": 0.3534, "step": 28240 }, { "epoch": 16.443538998835855, "grad_norm": 0.8732990622520447, "learning_rate": 5.8911187893147214e-05, "loss": 0.3583, "step": 28250 }, { "epoch": 16.449359720605354, "grad_norm": 0.8825758099555969, "learning_rate": 5.888406991198828e-05, "loss": 0.3467, "step": 28260 }, { "epoch": 16.455180442374854, "grad_norm": 0.8065192699432373, "learning_rate": 5.885694923208349e-05, "loss": 0.3545, "step": 28270 }, { "epoch": 16.461001164144353, "grad_norm": 0.7693616151809692, "learning_rate": 5.882982586167138e-05, "loss": 0.3396, "step": 28280 }, { "epoch": 16.466821885913852, "grad_norm": 0.7759407758712769, "learning_rate": 5.880269980899131e-05, "loss": 0.34, "step": 28290 }, { "epoch": 16.47264260768335, "grad_norm": 0.7929189205169678, "learning_rate": 5.8775571082283465e-05, "loss": 0.3526, "step": 28300 }, { "epoch": 16.47846332945285, "grad_norm": 0.8144717216491699, "learning_rate": 5.8748439689788824e-05, "loss": 0.3522, "step": 28310 }, { "epoch": 16.484284051222353, "grad_norm": 0.8156712055206299, "learning_rate": 5.87213056397492e-05, "loss": 0.3638, "step": 28320 }, { "epoch": 16.490104772991852, "grad_norm": 0.8543603420257568, "learning_rate": 5.869416894040719e-05, "loss": 0.3573, "step": 28330 }, { "epoch": 16.49592549476135, "grad_norm": 0.9292714595794678, "learning_rate": 5.866702960000621e-05, "loss": 0.3454, "step": 28340 }, { "epoch": 16.50174621653085, "grad_norm": 0.8464938402175903, "learning_rate": 5.863988762679048e-05, "loss": 0.3545, "step": 28350 }, { "epoch": 16.50756693830035, "grad_norm": 0.8457704782485962, "learning_rate": 5.8612743029005e-05, "loss": 0.3523, "step": 28360 }, { "epoch": 16.51338766006985, "grad_norm": 0.7968124151229858, "learning_rate": 5.858559581489561e-05, "loss": 0.3531, "step": 28370 }, { "epoch": 16.51920838183935, "grad_norm": 0.8428890705108643, "learning_rate": 5.85584459927089e-05, "loss": 0.3487, "step": 28380 }, { "epoch": 16.525029103608848, "grad_norm": 0.8440232276916504, "learning_rate": 5.853129357069227e-05, "loss": 0.3465, "step": 28390 }, { "epoch": 16.530849825378347, "grad_norm": 0.7697357535362244, "learning_rate": 5.8504138557093913e-05, "loss": 0.3574, "step": 28400 }, { "epoch": 16.536670547147846, "grad_norm": 0.8513658046722412, "learning_rate": 5.8476980960162784e-05, "loss": 0.3587, "step": 28410 }, { "epoch": 16.542491268917345, "grad_norm": 0.8730418682098389, "learning_rate": 5.844982078814868e-05, "loss": 0.3563, "step": 28420 }, { "epoch": 16.548311990686845, "grad_norm": 0.7838836908340454, "learning_rate": 5.842265804930211e-05, "loss": 0.3513, "step": 28430 }, { "epoch": 16.554132712456344, "grad_norm": 0.8003295660018921, "learning_rate": 5.839549275187444e-05, "loss": 0.3644, "step": 28440 }, { "epoch": 16.559953434225843, "grad_norm": 0.8171358108520508, "learning_rate": 5.836832490411771e-05, "loss": 0.3432, "step": 28450 }, { "epoch": 16.565774155995342, "grad_norm": 0.7545079588890076, "learning_rate": 5.834115451428485e-05, "loss": 0.3655, "step": 28460 }, { "epoch": 16.57159487776484, "grad_norm": 0.9314609169960022, "learning_rate": 5.831398159062946e-05, "loss": 0.3452, "step": 28470 }, { "epoch": 16.57741559953434, "grad_norm": 0.7800509333610535, "learning_rate": 5.828680614140599e-05, "loss": 0.3703, "step": 28480 }, { "epoch": 16.583236321303843, "grad_norm": 0.8762701153755188, "learning_rate": 5.825962817486962e-05, "loss": 0.3632, "step": 28490 }, { "epoch": 16.589057043073343, "grad_norm": 0.9546051621437073, "learning_rate": 5.823244769927629e-05, "loss": 0.3592, "step": 28500 }, { "epoch": 16.594877764842842, "grad_norm": 0.8866776823997498, "learning_rate": 5.8205264722882716e-05, "loss": 0.3558, "step": 28510 }, { "epoch": 16.60069848661234, "grad_norm": 0.8378314971923828, "learning_rate": 5.817807925394636e-05, "loss": 0.3613, "step": 28520 }, { "epoch": 16.60651920838184, "grad_norm": 0.8429141044616699, "learning_rate": 5.815089130072546e-05, "loss": 0.3488, "step": 28530 }, { "epoch": 16.61233993015134, "grad_norm": 0.8703509569168091, "learning_rate": 5.8123700871479e-05, "loss": 0.3636, "step": 28540 }, { "epoch": 16.61816065192084, "grad_norm": 0.8749171495437622, "learning_rate": 5.809650797446671e-05, "loss": 0.3604, "step": 28550 }, { "epoch": 16.623981373690338, "grad_norm": 0.8289881944656372, "learning_rate": 5.806931261794907e-05, "loss": 0.3361, "step": 28560 }, { "epoch": 16.629802095459837, "grad_norm": 0.798164963722229, "learning_rate": 5.804211481018731e-05, "loss": 0.3466, "step": 28570 }, { "epoch": 16.635622817229336, "grad_norm": 0.7498195767402649, "learning_rate": 5.801491455944341e-05, "loss": 0.3548, "step": 28580 }, { "epoch": 16.641443538998836, "grad_norm": 0.7586014866828918, "learning_rate": 5.79877118739801e-05, "loss": 0.346, "step": 28590 }, { "epoch": 16.647264260768335, "grad_norm": 0.8847599029541016, "learning_rate": 5.7960506762060816e-05, "loss": 0.3653, "step": 28600 }, { "epoch": 16.653084982537834, "grad_norm": 0.8118208050727844, "learning_rate": 5.793329923194977e-05, "loss": 0.3531, "step": 28610 }, { "epoch": 16.658905704307333, "grad_norm": 0.8027520179748535, "learning_rate": 5.790608929191187e-05, "loss": 0.3713, "step": 28620 }, { "epoch": 16.664726426076832, "grad_norm": 0.8373742699623108, "learning_rate": 5.78788769502128e-05, "loss": 0.3414, "step": 28630 }, { "epoch": 16.67054714784633, "grad_norm": 0.8145909309387207, "learning_rate": 5.785166221511894e-05, "loss": 0.3574, "step": 28640 }, { "epoch": 16.67636786961583, "grad_norm": 0.805867612361908, "learning_rate": 5.7824445094897415e-05, "loss": 0.3541, "step": 28650 }, { "epoch": 16.682188591385334, "grad_norm": 0.8490023016929626, "learning_rate": 5.7797225597816065e-05, "loss": 0.3437, "step": 28660 }, { "epoch": 16.688009313154833, "grad_norm": 0.810542643070221, "learning_rate": 5.777000373214345e-05, "loss": 0.3674, "step": 28670 }, { "epoch": 16.693830034924332, "grad_norm": 0.7255892157554626, "learning_rate": 5.774277950614885e-05, "loss": 0.362, "step": 28680 }, { "epoch": 16.69965075669383, "grad_norm": 0.7354653477668762, "learning_rate": 5.771555292810227e-05, "loss": 0.3525, "step": 28690 }, { "epoch": 16.70547147846333, "grad_norm": 0.816547155380249, "learning_rate": 5.768832400627444e-05, "loss": 0.3619, "step": 28700 }, { "epoch": 16.71129220023283, "grad_norm": 0.7770837545394897, "learning_rate": 5.7661092748936775e-05, "loss": 0.3445, "step": 28710 }, { "epoch": 16.71711292200233, "grad_norm": 0.7640708088874817, "learning_rate": 5.76338591643614e-05, "loss": 0.3479, "step": 28720 }, { "epoch": 16.722933643771828, "grad_norm": 0.8059104681015015, "learning_rate": 5.760662326082118e-05, "loss": 0.3533, "step": 28730 }, { "epoch": 16.728754365541327, "grad_norm": 0.783650279045105, "learning_rate": 5.757938504658965e-05, "loss": 0.3406, "step": 28740 }, { "epoch": 16.734575087310827, "grad_norm": 0.8926831483840942, "learning_rate": 5.755214452994107e-05, "loss": 0.3477, "step": 28750 }, { "epoch": 16.740395809080326, "grad_norm": 0.7937654256820679, "learning_rate": 5.752490171915039e-05, "loss": 0.3624, "step": 28760 }, { "epoch": 16.746216530849825, "grad_norm": 0.8258665204048157, "learning_rate": 5.749765662249324e-05, "loss": 0.3514, "step": 28770 }, { "epoch": 16.752037252619324, "grad_norm": 0.8851333856582642, "learning_rate": 5.747040924824596e-05, "loss": 0.3607, "step": 28780 }, { "epoch": 16.757857974388823, "grad_norm": 0.7806821465492249, "learning_rate": 5.7443159604685613e-05, "loss": 0.3433, "step": 28790 }, { "epoch": 16.763678696158323, "grad_norm": 0.7829483151435852, "learning_rate": 5.74159077000899e-05, "loss": 0.3514, "step": 28800 }, { "epoch": 16.769499417927822, "grad_norm": 0.810696542263031, "learning_rate": 5.7388653542737235e-05, "loss": 0.3428, "step": 28810 }, { "epoch": 16.77532013969732, "grad_norm": 0.7866176962852478, "learning_rate": 5.736139714090672e-05, "loss": 0.3466, "step": 28820 }, { "epoch": 16.78114086146682, "grad_norm": 0.7822887897491455, "learning_rate": 5.73341385028781e-05, "loss": 0.3376, "step": 28830 }, { "epoch": 16.78696158323632, "grad_norm": 0.8181754350662231, "learning_rate": 5.7306877636931855e-05, "loss": 0.3453, "step": 28840 }, { "epoch": 16.792782305005822, "grad_norm": 0.8451094031333923, "learning_rate": 5.7279614551349125e-05, "loss": 0.3576, "step": 28850 }, { "epoch": 16.79860302677532, "grad_norm": 0.6861896514892578, "learning_rate": 5.725234925441169e-05, "loss": 0.3437, "step": 28860 }, { "epoch": 16.80442374854482, "grad_norm": 0.8074343800544739, "learning_rate": 5.7225081754402044e-05, "loss": 0.3613, "step": 28870 }, { "epoch": 16.81024447031432, "grad_norm": 0.8076528310775757, "learning_rate": 5.7197812059603326e-05, "loss": 0.3505, "step": 28880 }, { "epoch": 16.81606519208382, "grad_norm": 0.7930624485015869, "learning_rate": 5.717054017829934e-05, "loss": 0.3441, "step": 28890 }, { "epoch": 16.82188591385332, "grad_norm": 0.7868682146072388, "learning_rate": 5.7143266118774584e-05, "loss": 0.3578, "step": 28900 }, { "epoch": 16.827706635622818, "grad_norm": 0.6941824555397034, "learning_rate": 5.711598988931418e-05, "loss": 0.3497, "step": 28910 }, { "epoch": 16.833527357392317, "grad_norm": 0.7950880527496338, "learning_rate": 5.7088711498203954e-05, "loss": 0.3556, "step": 28920 }, { "epoch": 16.839348079161816, "grad_norm": 0.913692831993103, "learning_rate": 5.706143095373033e-05, "loss": 0.361, "step": 28930 }, { "epoch": 16.845168800931315, "grad_norm": 0.7534388303756714, "learning_rate": 5.703414826418042e-05, "loss": 0.3367, "step": 28940 }, { "epoch": 16.850989522700814, "grad_norm": 0.8206372857093811, "learning_rate": 5.7006863437842007e-05, "loss": 0.3424, "step": 28950 }, { "epoch": 16.856810244470314, "grad_norm": 0.8046417832374573, "learning_rate": 5.697957648300348e-05, "loss": 0.3428, "step": 28960 }, { "epoch": 16.862630966239813, "grad_norm": 0.7604328989982605, "learning_rate": 5.695228740795391e-05, "loss": 0.3533, "step": 28970 }, { "epoch": 16.868451688009312, "grad_norm": 0.7475972771644592, "learning_rate": 5.6924996220982985e-05, "loss": 0.3567, "step": 28980 }, { "epoch": 16.87427240977881, "grad_norm": 0.9893953204154968, "learning_rate": 5.6897702930381045e-05, "loss": 0.3637, "step": 28990 }, { "epoch": 16.88009313154831, "grad_norm": 0.7782140970230103, "learning_rate": 5.687040754443908e-05, "loss": 0.3427, "step": 29000 }, { "epoch": 16.88591385331781, "grad_norm": 0.781378984451294, "learning_rate": 5.6843110071448725e-05, "loss": 0.3686, "step": 29010 }, { "epoch": 16.891734575087312, "grad_norm": 0.8175413012504578, "learning_rate": 5.6815810519702194e-05, "loss": 0.351, "step": 29020 }, { "epoch": 16.89755529685681, "grad_norm": 0.8205431699752808, "learning_rate": 5.6788508897492396e-05, "loss": 0.3432, "step": 29030 }, { "epoch": 16.90337601862631, "grad_norm": 0.800632119178772, "learning_rate": 5.676120521311282e-05, "loss": 0.3574, "step": 29040 }, { "epoch": 16.90919674039581, "grad_norm": 0.8575453162193298, "learning_rate": 5.6733899474857634e-05, "loss": 0.3486, "step": 29050 }, { "epoch": 16.91501746216531, "grad_norm": 0.8550110459327698, "learning_rate": 5.670659169102157e-05, "loss": 0.3587, "step": 29060 }, { "epoch": 16.92083818393481, "grad_norm": 0.814070463180542, "learning_rate": 5.6679281869900044e-05, "loss": 0.3606, "step": 29070 }, { "epoch": 16.926658905704308, "grad_norm": 0.8580443859100342, "learning_rate": 5.6651970019789045e-05, "loss": 0.341, "step": 29080 }, { "epoch": 16.932479627473807, "grad_norm": 0.7540550827980042, "learning_rate": 5.662465614898519e-05, "loss": 0.3661, "step": 29090 }, { "epoch": 16.938300349243306, "grad_norm": 0.79789137840271, "learning_rate": 5.6597340265785695e-05, "loss": 0.3574, "step": 29100 }, { "epoch": 16.944121071012805, "grad_norm": 0.8177546262741089, "learning_rate": 5.657002237848843e-05, "loss": 0.3644, "step": 29110 }, { "epoch": 16.949941792782305, "grad_norm": 0.828069806098938, "learning_rate": 5.654270249539183e-05, "loss": 0.3416, "step": 29120 }, { "epoch": 16.955762514551804, "grad_norm": 0.763191282749176, "learning_rate": 5.651538062479498e-05, "loss": 0.3426, "step": 29130 }, { "epoch": 16.961583236321303, "grad_norm": 0.8449112176895142, "learning_rate": 5.648805677499751e-05, "loss": 0.3509, "step": 29140 }, { "epoch": 16.967403958090802, "grad_norm": 0.8393259048461914, "learning_rate": 5.646073095429969e-05, "loss": 0.3453, "step": 29150 }, { "epoch": 16.9732246798603, "grad_norm": 0.7534541487693787, "learning_rate": 5.643340317100241e-05, "loss": 0.3559, "step": 29160 }, { "epoch": 16.9790454016298, "grad_norm": 0.7556893229484558, "learning_rate": 5.64060734334071e-05, "loss": 0.3516, "step": 29170 }, { "epoch": 16.9848661233993, "grad_norm": 0.7626745104789734, "learning_rate": 5.637874174981583e-05, "loss": 0.351, "step": 29180 }, { "epoch": 16.990686845168803, "grad_norm": 0.7858979105949402, "learning_rate": 5.635140812853124e-05, "loss": 0.3377, "step": 29190 }, { "epoch": 16.996507566938302, "grad_norm": 0.6760004162788391, "learning_rate": 5.6324072577856544e-05, "loss": 0.3648, "step": 29200 }, { "epoch": 17.0023282887078, "grad_norm": 0.7889419198036194, "learning_rate": 5.629673510609559e-05, "loss": 0.3486, "step": 29210 }, { "epoch": 17.0081490104773, "grad_norm": 0.8262608051300049, "learning_rate": 5.626939572155276e-05, "loss": 0.365, "step": 29220 }, { "epoch": 17.0139697322468, "grad_norm": 0.8759471774101257, "learning_rate": 5.6242054432533054e-05, "loss": 0.3515, "step": 29230 }, { "epoch": 17.0197904540163, "grad_norm": 0.8052782416343689, "learning_rate": 5.621471124734201e-05, "loss": 0.353, "step": 29240 }, { "epoch": 17.025611175785798, "grad_norm": 0.7795354723930359, "learning_rate": 5.6187366174285794e-05, "loss": 0.3419, "step": 29250 }, { "epoch": 17.031431897555297, "grad_norm": 0.9155131578445435, "learning_rate": 5.616001922167109e-05, "loss": 0.3651, "step": 29260 }, { "epoch": 17.037252619324796, "grad_norm": 0.8257775902748108, "learning_rate": 5.61326703978052e-05, "loss": 0.372, "step": 29270 }, { "epoch": 17.043073341094296, "grad_norm": 0.7912125587463379, "learning_rate": 5.6105319710995964e-05, "loss": 0.3399, "step": 29280 }, { "epoch": 17.048894062863795, "grad_norm": 0.8820396065711975, "learning_rate": 5.60779671695518e-05, "loss": 0.3521, "step": 29290 }, { "epoch": 17.054714784633294, "grad_norm": 0.7451552748680115, "learning_rate": 5.6050612781781684e-05, "loss": 0.3419, "step": 29300 }, { "epoch": 17.060535506402793, "grad_norm": 0.7385024428367615, "learning_rate": 5.602325655599516e-05, "loss": 0.3495, "step": 29310 }, { "epoch": 17.066356228172292, "grad_norm": 0.7834216356277466, "learning_rate": 5.599589850050234e-05, "loss": 0.3517, "step": 29320 }, { "epoch": 17.07217694994179, "grad_norm": 0.850275993347168, "learning_rate": 5.5968538623613874e-05, "loss": 0.3595, "step": 29330 }, { "epoch": 17.07799767171129, "grad_norm": 0.9052079916000366, "learning_rate": 5.594117693364095e-05, "loss": 0.3504, "step": 29340 }, { "epoch": 17.08381839348079, "grad_norm": 0.8363013863563538, "learning_rate": 5.591381343889535e-05, "loss": 0.3644, "step": 29350 }, { "epoch": 17.08963911525029, "grad_norm": 0.743800699710846, "learning_rate": 5.5886448147689355e-05, "loss": 0.3443, "step": 29360 }, { "epoch": 17.095459837019792, "grad_norm": 0.7811647057533264, "learning_rate": 5.585908106833585e-05, "loss": 0.3366, "step": 29370 }, { "epoch": 17.10128055878929, "grad_norm": 0.7996224761009216, "learning_rate": 5.5831712209148226e-05, "loss": 0.3463, "step": 29380 }, { "epoch": 17.10710128055879, "grad_norm": 0.7637783885002136, "learning_rate": 5.58043415784404e-05, "loss": 0.3527, "step": 29390 }, { "epoch": 17.11292200232829, "grad_norm": 0.760132372379303, "learning_rate": 5.577696918452686e-05, "loss": 0.3493, "step": 29400 }, { "epoch": 17.11874272409779, "grad_norm": 0.7801423668861389, "learning_rate": 5.5749595035722604e-05, "loss": 0.3378, "step": 29410 }, { "epoch": 17.124563445867288, "grad_norm": 0.8867086172103882, "learning_rate": 5.5722219140343193e-05, "loss": 0.3573, "step": 29420 }, { "epoch": 17.130384167636787, "grad_norm": 0.922238826751709, "learning_rate": 5.56948415067047e-05, "loss": 0.3467, "step": 29430 }, { "epoch": 17.136204889406287, "grad_norm": 0.7105020880699158, "learning_rate": 5.5667462143123704e-05, "loss": 0.3374, "step": 29440 }, { "epoch": 17.142025611175786, "grad_norm": 0.766633927822113, "learning_rate": 5.564008105791737e-05, "loss": 0.3464, "step": 29450 }, { "epoch": 17.147846332945285, "grad_norm": 0.8239579200744629, "learning_rate": 5.5612698259403316e-05, "loss": 0.3627, "step": 29460 }, { "epoch": 17.153667054714784, "grad_norm": 0.8180544972419739, "learning_rate": 5.5585313755899724e-05, "loss": 0.3379, "step": 29470 }, { "epoch": 17.159487776484283, "grad_norm": 0.7822855114936829, "learning_rate": 5.5557927555725285e-05, "loss": 0.3542, "step": 29480 }, { "epoch": 17.165308498253783, "grad_norm": 0.8591811656951904, "learning_rate": 5.55305396671992e-05, "loss": 0.3457, "step": 29490 }, { "epoch": 17.171129220023282, "grad_norm": 0.8725078701972961, "learning_rate": 5.55031500986412e-05, "loss": 0.3434, "step": 29500 }, { "epoch": 17.17694994179278, "grad_norm": 0.8070479035377502, "learning_rate": 5.547575885837149e-05, "loss": 0.3565, "step": 29510 }, { "epoch": 17.18277066356228, "grad_norm": 0.7669195532798767, "learning_rate": 5.5448365954710825e-05, "loss": 0.3465, "step": 29520 }, { "epoch": 17.18859138533178, "grad_norm": 0.8357054591178894, "learning_rate": 5.5420971395980446e-05, "loss": 0.3561, "step": 29530 }, { "epoch": 17.194412107101282, "grad_norm": 0.8513782024383545, "learning_rate": 5.539357519050209e-05, "loss": 0.3428, "step": 29540 }, { "epoch": 17.20023282887078, "grad_norm": 0.7192535996437073, "learning_rate": 5.536617734659799e-05, "loss": 0.3513, "step": 29550 }, { "epoch": 17.20605355064028, "grad_norm": 0.799770176410675, "learning_rate": 5.533877787259091e-05, "loss": 0.3606, "step": 29560 }, { "epoch": 17.21187427240978, "grad_norm": 0.7469451427459717, "learning_rate": 5.5311376776804044e-05, "loss": 0.3472, "step": 29570 }, { "epoch": 17.21769499417928, "grad_norm": 0.8808695673942566, "learning_rate": 5.528397406756118e-05, "loss": 0.3476, "step": 29580 }, { "epoch": 17.22351571594878, "grad_norm": 0.7716600894927979, "learning_rate": 5.525656975318652e-05, "loss": 0.3439, "step": 29590 }, { "epoch": 17.229336437718278, "grad_norm": 0.7585768699645996, "learning_rate": 5.522916384200474e-05, "loss": 0.3503, "step": 29600 }, { "epoch": 17.235157159487777, "grad_norm": 0.8996527791023254, "learning_rate": 5.520175634234106e-05, "loss": 0.3552, "step": 29610 }, { "epoch": 17.240977881257276, "grad_norm": 0.7832033634185791, "learning_rate": 5.517434726252113e-05, "loss": 0.344, "step": 29620 }, { "epoch": 17.246798603026775, "grad_norm": 0.8300150632858276, "learning_rate": 5.514693661087113e-05, "loss": 0.3417, "step": 29630 }, { "epoch": 17.252619324796274, "grad_norm": 0.8066682815551758, "learning_rate": 5.511952439571769e-05, "loss": 0.3456, "step": 29640 }, { "epoch": 17.258440046565774, "grad_norm": 0.7938589453697205, "learning_rate": 5.509211062538791e-05, "loss": 0.3534, "step": 29650 }, { "epoch": 17.264260768335273, "grad_norm": 0.8709155321121216, "learning_rate": 5.506469530820939e-05, "loss": 0.3502, "step": 29660 }, { "epoch": 17.270081490104772, "grad_norm": 0.8109029531478882, "learning_rate": 5.503727845251014e-05, "loss": 0.3474, "step": 29670 }, { "epoch": 17.27590221187427, "grad_norm": 0.8080213069915771, "learning_rate": 5.50098600666187e-05, "loss": 0.3498, "step": 29680 }, { "epoch": 17.28172293364377, "grad_norm": 0.8627705574035645, "learning_rate": 5.498244015886406e-05, "loss": 0.3484, "step": 29690 }, { "epoch": 17.28754365541327, "grad_norm": 0.8218737840652466, "learning_rate": 5.495501873757565e-05, "loss": 0.3338, "step": 29700 }, { "epoch": 17.29336437718277, "grad_norm": 0.9444795846939087, "learning_rate": 5.492759581108336e-05, "loss": 0.3517, "step": 29710 }, { "epoch": 17.29918509895227, "grad_norm": 0.8266793489456177, "learning_rate": 5.490017138771759e-05, "loss": 0.3543, "step": 29720 }, { "epoch": 17.30500582072177, "grad_norm": 0.8940138220787048, "learning_rate": 5.487274547580912e-05, "loss": 0.3511, "step": 29730 }, { "epoch": 17.31082654249127, "grad_norm": 0.8400716185569763, "learning_rate": 5.484531808368923e-05, "loss": 0.3543, "step": 29740 }, { "epoch": 17.31664726426077, "grad_norm": 0.779987096786499, "learning_rate": 5.4817889219689656e-05, "loss": 0.3535, "step": 29750 }, { "epoch": 17.32246798603027, "grad_norm": 0.9400197863578796, "learning_rate": 5.4790458892142536e-05, "loss": 0.3513, "step": 29760 }, { "epoch": 17.328288707799768, "grad_norm": 0.8598145246505737, "learning_rate": 5.476302710938048e-05, "loss": 0.3468, "step": 29770 }, { "epoch": 17.334109429569267, "grad_norm": 0.7799885272979736, "learning_rate": 5.473559387973657e-05, "loss": 0.3607, "step": 29780 }, { "epoch": 17.339930151338766, "grad_norm": 0.8795459866523743, "learning_rate": 5.470815921154425e-05, "loss": 0.354, "step": 29790 }, { "epoch": 17.345750873108265, "grad_norm": 0.8709113597869873, "learning_rate": 5.468072311313749e-05, "loss": 0.3644, "step": 29800 }, { "epoch": 17.351571594877765, "grad_norm": 0.8056883811950684, "learning_rate": 5.465328559285063e-05, "loss": 0.3573, "step": 29810 }, { "epoch": 17.357392316647264, "grad_norm": 0.8263056874275208, "learning_rate": 5.462584665901849e-05, "loss": 0.3501, "step": 29820 }, { "epoch": 17.363213038416763, "grad_norm": 0.7946062088012695, "learning_rate": 5.4598406319976235e-05, "loss": 0.3412, "step": 29830 }, { "epoch": 17.369033760186262, "grad_norm": 0.8129686117172241, "learning_rate": 5.457096458405958e-05, "loss": 0.3455, "step": 29840 }, { "epoch": 17.37485448195576, "grad_norm": 0.9091408848762512, "learning_rate": 5.454352145960457e-05, "loss": 0.3533, "step": 29850 }, { "epoch": 17.38067520372526, "grad_norm": 0.8630486726760864, "learning_rate": 5.4516076954947715e-05, "loss": 0.374, "step": 29860 }, { "epoch": 17.38649592549476, "grad_norm": 0.8465051651000977, "learning_rate": 5.448863107842591e-05, "loss": 0.3659, "step": 29870 }, { "epoch": 17.39231664726426, "grad_norm": 0.8212530016899109, "learning_rate": 5.446118383837651e-05, "loss": 0.3422, "step": 29880 }, { "epoch": 17.398137369033762, "grad_norm": 0.7269655466079712, "learning_rate": 5.443373524313722e-05, "loss": 0.3323, "step": 29890 }, { "epoch": 17.40395809080326, "grad_norm": 0.7237100601196289, "learning_rate": 5.440628530104626e-05, "loss": 0.3427, "step": 29900 }, { "epoch": 17.40977881257276, "grad_norm": 0.7077711820602417, "learning_rate": 5.4378834020442146e-05, "loss": 0.3382, "step": 29910 }, { "epoch": 17.41559953434226, "grad_norm": 0.7395457625389099, "learning_rate": 5.4351381409663884e-05, "loss": 0.336, "step": 29920 }, { "epoch": 17.42142025611176, "grad_norm": 0.7666095495223999, "learning_rate": 5.432392747705084e-05, "loss": 0.3499, "step": 29930 }, { "epoch": 17.427240977881258, "grad_norm": 0.7887189388275146, "learning_rate": 5.429647223094278e-05, "loss": 0.3587, "step": 29940 }, { "epoch": 17.433061699650757, "grad_norm": 0.7696295380592346, "learning_rate": 5.4269015679679924e-05, "loss": 0.3487, "step": 29950 }, { "epoch": 17.438882421420256, "grad_norm": 0.7956116795539856, "learning_rate": 5.424155783160281e-05, "loss": 0.3609, "step": 29960 }, { "epoch": 17.444703143189756, "grad_norm": 0.8609016537666321, "learning_rate": 5.4214098695052415e-05, "loss": 0.3522, "step": 29970 }, { "epoch": 17.450523864959255, "grad_norm": 0.755420982837677, "learning_rate": 5.418663827837012e-05, "loss": 0.3515, "step": 29980 }, { "epoch": 17.456344586728754, "grad_norm": 0.747539758682251, "learning_rate": 5.415917658989763e-05, "loss": 0.3456, "step": 29990 }, { "epoch": 17.462165308498253, "grad_norm": 0.805091142654419, "learning_rate": 5.413171363797713e-05, "loss": 0.3426, "step": 30000 }, { "epoch": 17.467986030267753, "grad_norm": 0.7310168147087097, "learning_rate": 5.4104249430951116e-05, "loss": 0.3469, "step": 30010 }, { "epoch": 17.47380675203725, "grad_norm": 0.8720252513885498, "learning_rate": 5.4076783977162494e-05, "loss": 0.3513, "step": 30020 }, { "epoch": 17.47962747380675, "grad_norm": 0.731995701789856, "learning_rate": 5.4049317284954525e-05, "loss": 0.3556, "step": 30030 }, { "epoch": 17.48544819557625, "grad_norm": 0.8593712449073792, "learning_rate": 5.4021849362670884e-05, "loss": 0.3458, "step": 30040 }, { "epoch": 17.49126891734575, "grad_norm": 0.858826756477356, "learning_rate": 5.3994380218655604e-05, "loss": 0.3531, "step": 30050 }, { "epoch": 17.49708963911525, "grad_norm": 0.7970762848854065, "learning_rate": 5.396690986125309e-05, "loss": 0.3436, "step": 30060 }, { "epoch": 17.50291036088475, "grad_norm": 0.7746023535728455, "learning_rate": 5.3939438298808075e-05, "loss": 0.3482, "step": 30070 }, { "epoch": 17.50873108265425, "grad_norm": 0.7405223846435547, "learning_rate": 5.3911965539665744e-05, "loss": 0.3298, "step": 30080 }, { "epoch": 17.51455180442375, "grad_norm": 0.8057175278663635, "learning_rate": 5.388449159217156e-05, "loss": 0.3452, "step": 30090 }, { "epoch": 17.52037252619325, "grad_norm": 0.7597792148590088, "learning_rate": 5.3857016464671385e-05, "loss": 0.3541, "step": 30100 }, { "epoch": 17.52619324796275, "grad_norm": 0.7354208827018738, "learning_rate": 5.382954016551146e-05, "loss": 0.3539, "step": 30110 }, { "epoch": 17.532013969732247, "grad_norm": 0.7712478637695312, "learning_rate": 5.380206270303835e-05, "loss": 0.351, "step": 30120 }, { "epoch": 17.537834691501747, "grad_norm": 0.7719467282295227, "learning_rate": 5.377458408559897e-05, "loss": 0.3522, "step": 30130 }, { "epoch": 17.543655413271246, "grad_norm": 0.8558740019798279, "learning_rate": 5.374710432154061e-05, "loss": 0.3535, "step": 30140 }, { "epoch": 17.549476135040745, "grad_norm": 0.797627866268158, "learning_rate": 5.3719623419210886e-05, "loss": 0.3659, "step": 30150 }, { "epoch": 17.555296856810244, "grad_norm": 0.8634498119354248, "learning_rate": 5.3692141386957786e-05, "loss": 0.3464, "step": 30160 }, { "epoch": 17.561117578579744, "grad_norm": 0.7816588878631592, "learning_rate": 5.3664658233129616e-05, "loss": 0.3496, "step": 30170 }, { "epoch": 17.566938300349243, "grad_norm": 0.7019344568252563, "learning_rate": 5.363717396607504e-05, "loss": 0.3542, "step": 30180 }, { "epoch": 17.572759022118742, "grad_norm": 0.8423434495925903, "learning_rate": 5.360968859414305e-05, "loss": 0.3658, "step": 30190 }, { "epoch": 17.57857974388824, "grad_norm": 0.7975437641143799, "learning_rate": 5.358220212568295e-05, "loss": 0.3425, "step": 30200 }, { "epoch": 17.58440046565774, "grad_norm": 0.9544422030448914, "learning_rate": 5.355471456904444e-05, "loss": 0.3387, "step": 30210 }, { "epoch": 17.59022118742724, "grad_norm": 0.8711733818054199, "learning_rate": 5.3527225932577495e-05, "loss": 0.3598, "step": 30220 }, { "epoch": 17.59604190919674, "grad_norm": 0.7305257320404053, "learning_rate": 5.349973622463246e-05, "loss": 0.361, "step": 30230 }, { "epoch": 17.601862630966238, "grad_norm": 0.7723055481910706, "learning_rate": 5.3472245453559956e-05, "loss": 0.3445, "step": 30240 }, { "epoch": 17.60768335273574, "grad_norm": 0.7553175091743469, "learning_rate": 5.3444753627710955e-05, "loss": 0.3361, "step": 30250 }, { "epoch": 17.61350407450524, "grad_norm": 0.7824915647506714, "learning_rate": 5.341726075543676e-05, "loss": 0.3447, "step": 30260 }, { "epoch": 17.61932479627474, "grad_norm": 0.7818645238876343, "learning_rate": 5.338976684508898e-05, "loss": 0.3412, "step": 30270 }, { "epoch": 17.62514551804424, "grad_norm": 0.8095875978469849, "learning_rate": 5.336227190501953e-05, "loss": 0.3526, "step": 30280 }, { "epoch": 17.630966239813738, "grad_norm": 1.130234956741333, "learning_rate": 5.3334775943580664e-05, "loss": 0.3502, "step": 30290 }, { "epoch": 17.636786961583237, "grad_norm": 0.9606425166130066, "learning_rate": 5.330727896912491e-05, "loss": 0.3498, "step": 30300 }, { "epoch": 17.642607683352736, "grad_norm": 0.8410587310791016, "learning_rate": 5.327978099000511e-05, "loss": 0.3369, "step": 30310 }, { "epoch": 17.648428405122235, "grad_norm": 0.7615038156509399, "learning_rate": 5.3252282014574465e-05, "loss": 0.3505, "step": 30320 }, { "epoch": 17.654249126891735, "grad_norm": 0.8383623957633972, "learning_rate": 5.322478205118641e-05, "loss": 0.3695, "step": 30330 }, { "epoch": 17.660069848661234, "grad_norm": 0.8004186153411865, "learning_rate": 5.3197281108194704e-05, "loss": 0.342, "step": 30340 }, { "epoch": 17.665890570430733, "grad_norm": 0.7164003849029541, "learning_rate": 5.316977919395342e-05, "loss": 0.3434, "step": 30350 }, { "epoch": 17.671711292200232, "grad_norm": 0.8234593868255615, "learning_rate": 5.314227631681691e-05, "loss": 0.3499, "step": 30360 }, { "epoch": 17.67753201396973, "grad_norm": 1.0784928798675537, "learning_rate": 5.311477248513982e-05, "loss": 0.3506, "step": 30370 }, { "epoch": 17.68335273573923, "grad_norm": 1.0883055925369263, "learning_rate": 5.30872677072771e-05, "loss": 0.3593, "step": 30380 }, { "epoch": 17.68917345750873, "grad_norm": 0.8163688778877258, "learning_rate": 5.3059761991583954e-05, "loss": 0.3528, "step": 30390 }, { "epoch": 17.69499417927823, "grad_norm": 0.8509805202484131, "learning_rate": 5.303225534641592e-05, "loss": 0.3563, "step": 30400 }, { "epoch": 17.70081490104773, "grad_norm": 0.8746610879898071, "learning_rate": 5.300474778012875e-05, "loss": 0.3442, "step": 30410 }, { "epoch": 17.70663562281723, "grad_norm": 0.87282794713974, "learning_rate": 5.297723930107855e-05, "loss": 0.3469, "step": 30420 }, { "epoch": 17.71245634458673, "grad_norm": 0.7796326875686646, "learning_rate": 5.294972991762167e-05, "loss": 0.3456, "step": 30430 }, { "epoch": 17.71827706635623, "grad_norm": 0.8242179155349731, "learning_rate": 5.292221963811472e-05, "loss": 0.3565, "step": 30440 }, { "epoch": 17.72409778812573, "grad_norm": 0.7836906313896179, "learning_rate": 5.28947084709146e-05, "loss": 0.3491, "step": 30450 }, { "epoch": 17.729918509895228, "grad_norm": 0.8535655736923218, "learning_rate": 5.2867196424378465e-05, "loss": 0.3474, "step": 30460 }, { "epoch": 17.735739231664727, "grad_norm": 0.8449023962020874, "learning_rate": 5.2839683506863765e-05, "loss": 0.3438, "step": 30470 }, { "epoch": 17.741559953434226, "grad_norm": 0.8102584481239319, "learning_rate": 5.281216972672821e-05, "loss": 0.3515, "step": 30480 }, { "epoch": 17.747380675203726, "grad_norm": 0.8400194644927979, "learning_rate": 5.278465509232973e-05, "loss": 0.3582, "step": 30490 }, { "epoch": 17.753201396973225, "grad_norm": 0.8500772714614868, "learning_rate": 5.275713961202655e-05, "loss": 0.3427, "step": 30500 }, { "epoch": 17.759022118742724, "grad_norm": 0.7850145101547241, "learning_rate": 5.2729623294177165e-05, "loss": 0.3395, "step": 30510 }, { "epoch": 17.764842840512223, "grad_norm": 0.7502774000167847, "learning_rate": 5.270210614714028e-05, "loss": 0.3601, "step": 30520 }, { "epoch": 17.770663562281722, "grad_norm": 0.8569589257240295, "learning_rate": 5.267458817927491e-05, "loss": 0.3593, "step": 30530 }, { "epoch": 17.77648428405122, "grad_norm": 0.7196128964424133, "learning_rate": 5.264706939894026e-05, "loss": 0.3443, "step": 30540 }, { "epoch": 17.78230500582072, "grad_norm": 0.8399613499641418, "learning_rate": 5.261954981449584e-05, "loss": 0.3505, "step": 30550 }, { "epoch": 17.78812572759022, "grad_norm": 0.7802991271018982, "learning_rate": 5.2592029434301324e-05, "loss": 0.3505, "step": 30560 }, { "epoch": 17.79394644935972, "grad_norm": 0.7330204248428345, "learning_rate": 5.256450826671672e-05, "loss": 0.3547, "step": 30570 }, { "epoch": 17.79976717112922, "grad_norm": 0.828274667263031, "learning_rate": 5.253698632010221e-05, "loss": 0.3518, "step": 30580 }, { "epoch": 17.80558789289872, "grad_norm": 0.7362048029899597, "learning_rate": 5.2509463602818246e-05, "loss": 0.3498, "step": 30590 }, { "epoch": 17.81140861466822, "grad_norm": 0.7092493772506714, "learning_rate": 5.248194012322549e-05, "loss": 0.3481, "step": 30600 }, { "epoch": 17.81722933643772, "grad_norm": 0.6925507187843323, "learning_rate": 5.245441588968486e-05, "loss": 0.3401, "step": 30610 }, { "epoch": 17.82305005820722, "grad_norm": 0.6853694319725037, "learning_rate": 5.242689091055748e-05, "loss": 0.339, "step": 30620 }, { "epoch": 17.828870779976718, "grad_norm": 0.8018394112586975, "learning_rate": 5.239936519420473e-05, "loss": 0.3371, "step": 30630 }, { "epoch": 17.834691501746217, "grad_norm": 0.7443342804908752, "learning_rate": 5.2371838748988175e-05, "loss": 0.3326, "step": 30640 }, { "epoch": 17.840512223515717, "grad_norm": 0.6761016249656677, "learning_rate": 5.234431158326965e-05, "loss": 0.3294, "step": 30650 }, { "epoch": 17.846332945285216, "grad_norm": 0.7830053567886353, "learning_rate": 5.231678370541115e-05, "loss": 0.3482, "step": 30660 }, { "epoch": 17.852153667054715, "grad_norm": 0.7351859211921692, "learning_rate": 5.228925512377495e-05, "loss": 0.3393, "step": 30670 }, { "epoch": 17.857974388824214, "grad_norm": 0.8110164999961853, "learning_rate": 5.2261725846723465e-05, "loss": 0.3451, "step": 30680 }, { "epoch": 17.863795110593713, "grad_norm": 0.7590148448944092, "learning_rate": 5.22341958826194e-05, "loss": 0.3442, "step": 30690 }, { "epoch": 17.869615832363213, "grad_norm": 0.7512082457542419, "learning_rate": 5.22066652398256e-05, "loss": 0.3499, "step": 30700 }, { "epoch": 17.875436554132712, "grad_norm": 0.8665536642074585, "learning_rate": 5.2179133926705185e-05, "loss": 0.348, "step": 30710 }, { "epoch": 17.88125727590221, "grad_norm": 0.7935031652450562, "learning_rate": 5.215160195162141e-05, "loss": 0.3453, "step": 30720 }, { "epoch": 17.88707799767171, "grad_norm": 0.7931444644927979, "learning_rate": 5.212406932293776e-05, "loss": 0.3577, "step": 30730 }, { "epoch": 17.89289871944121, "grad_norm": 0.7568791508674622, "learning_rate": 5.209653604901795e-05, "loss": 0.3441, "step": 30740 }, { "epoch": 17.89871944121071, "grad_norm": 0.7489747405052185, "learning_rate": 5.206900213822584e-05, "loss": 0.3449, "step": 30750 }, { "epoch": 17.904540162980208, "grad_norm": 0.7232306599617004, "learning_rate": 5.204146759892551e-05, "loss": 0.3336, "step": 30760 }, { "epoch": 17.91036088474971, "grad_norm": 0.8039589524269104, "learning_rate": 5.2013932439481216e-05, "loss": 0.3562, "step": 30770 }, { "epoch": 17.91618160651921, "grad_norm": 0.8052955269813538, "learning_rate": 5.198639666825743e-05, "loss": 0.3512, "step": 30780 }, { "epoch": 17.92200232828871, "grad_norm": 0.910214364528656, "learning_rate": 5.195886029361877e-05, "loss": 0.3524, "step": 30790 }, { "epoch": 17.92782305005821, "grad_norm": 0.7919678092002869, "learning_rate": 5.193132332393009e-05, "loss": 0.3377, "step": 30800 }, { "epoch": 17.933643771827708, "grad_norm": 0.7797943949699402, "learning_rate": 5.1903785767556376e-05, "loss": 0.3337, "step": 30810 }, { "epoch": 17.939464493597207, "grad_norm": 0.7239806652069092, "learning_rate": 5.187624763286282e-05, "loss": 0.3416, "step": 30820 }, { "epoch": 17.945285215366706, "grad_norm": 0.8105485439300537, "learning_rate": 5.184870892821475e-05, "loss": 0.3574, "step": 30830 }, { "epoch": 17.951105937136205, "grad_norm": 0.7286266684532166, "learning_rate": 5.182116966197773e-05, "loss": 0.3501, "step": 30840 }, { "epoch": 17.956926658905704, "grad_norm": 0.7989816069602966, "learning_rate": 5.1793629842517466e-05, "loss": 0.3515, "step": 30850 }, { "epoch": 17.962747380675204, "grad_norm": 0.8348557353019714, "learning_rate": 5.17660894781998e-05, "loss": 0.3358, "step": 30860 }, { "epoch": 17.968568102444703, "grad_norm": 0.8469457626342773, "learning_rate": 5.173854857739079e-05, "loss": 0.3585, "step": 30870 }, { "epoch": 17.974388824214202, "grad_norm": 0.7748771905899048, "learning_rate": 5.171100714845661e-05, "loss": 0.3581, "step": 30880 }, { "epoch": 17.9802095459837, "grad_norm": 0.7618610858917236, "learning_rate": 5.1683465199763646e-05, "loss": 0.3483, "step": 30890 }, { "epoch": 17.9860302677532, "grad_norm": 0.8354260921478271, "learning_rate": 5.16559227396784e-05, "loss": 0.3452, "step": 30900 }, { "epoch": 17.9918509895227, "grad_norm": 0.7577842473983765, "learning_rate": 5.1628379776567556e-05, "loss": 0.3428, "step": 30910 }, { "epoch": 17.9976717112922, "grad_norm": 0.7985525727272034, "learning_rate": 5.160083631879792e-05, "loss": 0.3388, "step": 30920 }, { "epoch": 18.003492433061698, "grad_norm": 0.7649548053741455, "learning_rate": 5.1573292374736484e-05, "loss": 0.3333, "step": 30930 }, { "epoch": 18.009313154831197, "grad_norm": 0.7190653681755066, "learning_rate": 5.1545747952750356e-05, "loss": 0.3426, "step": 30940 }, { "epoch": 18.0151338766007, "grad_norm": 0.8706685304641724, "learning_rate": 5.151820306120682e-05, "loss": 0.3556, "step": 30950 }, { "epoch": 18.0209545983702, "grad_norm": 0.8285284638404846, "learning_rate": 5.149065770847328e-05, "loss": 0.3475, "step": 30960 }, { "epoch": 18.0267753201397, "grad_norm": 0.7429248094558716, "learning_rate": 5.1463111902917297e-05, "loss": 0.347, "step": 30970 }, { "epoch": 18.032596041909198, "grad_norm": 0.7550114393234253, "learning_rate": 5.143556565290654e-05, "loss": 0.3489, "step": 30980 }, { "epoch": 18.038416763678697, "grad_norm": 0.7792723774909973, "learning_rate": 5.140801896680882e-05, "loss": 0.349, "step": 30990 }, { "epoch": 18.044237485448196, "grad_norm": 0.804306149482727, "learning_rate": 5.1380471852992144e-05, "loss": 0.3544, "step": 31000 }, { "epoch": 18.050058207217695, "grad_norm": 0.7554967999458313, "learning_rate": 5.135292431982457e-05, "loss": 0.3469, "step": 31010 }, { "epoch": 18.055878928987195, "grad_norm": 0.8046695590019226, "learning_rate": 5.1325376375674294e-05, "loss": 0.3493, "step": 31020 }, { "epoch": 18.061699650756694, "grad_norm": 0.8516713976860046, "learning_rate": 5.129782802890968e-05, "loss": 0.3657, "step": 31030 }, { "epoch": 18.067520372526193, "grad_norm": 0.8148659467697144, "learning_rate": 5.127027928789916e-05, "loss": 0.3469, "step": 31040 }, { "epoch": 18.073341094295692, "grad_norm": 0.7683861255645752, "learning_rate": 5.124273016101135e-05, "loss": 0.34, "step": 31050 }, { "epoch": 18.07916181606519, "grad_norm": 0.881624162197113, "learning_rate": 5.121518065661492e-05, "loss": 0.3456, "step": 31060 }, { "epoch": 18.08498253783469, "grad_norm": 0.8584743142127991, "learning_rate": 5.11876307830787e-05, "loss": 0.3543, "step": 31070 }, { "epoch": 18.09080325960419, "grad_norm": 0.819191038608551, "learning_rate": 5.1160080548771596e-05, "loss": 0.3461, "step": 31080 }, { "epoch": 18.09662398137369, "grad_norm": 0.7675065398216248, "learning_rate": 5.1132529962062656e-05, "loss": 0.3615, "step": 31090 }, { "epoch": 18.10244470314319, "grad_norm": 0.8253178596496582, "learning_rate": 5.110497903132101e-05, "loss": 0.3462, "step": 31100 }, { "epoch": 18.108265424912688, "grad_norm": 0.8388607501983643, "learning_rate": 5.107742776491592e-05, "loss": 0.3526, "step": 31110 }, { "epoch": 18.11408614668219, "grad_norm": 0.7118335366249084, "learning_rate": 5.104987617121673e-05, "loss": 0.3357, "step": 31120 }, { "epoch": 18.11990686845169, "grad_norm": 0.8011450171470642, "learning_rate": 5.102232425859287e-05, "loss": 0.3427, "step": 31130 }, { "epoch": 18.12572759022119, "grad_norm": 0.842621922492981, "learning_rate": 5.09947720354139e-05, "loss": 0.3538, "step": 31140 }, { "epoch": 18.131548311990688, "grad_norm": 0.8452784419059753, "learning_rate": 5.096721951004942e-05, "loss": 0.3519, "step": 31150 }, { "epoch": 18.137369033760187, "grad_norm": 0.7473915815353394, "learning_rate": 5.0939666690869227e-05, "loss": 0.3418, "step": 31160 }, { "epoch": 18.143189755529686, "grad_norm": 0.798251211643219, "learning_rate": 5.0912113586243096e-05, "loss": 0.3466, "step": 31170 }, { "epoch": 18.149010477299186, "grad_norm": 0.7611405849456787, "learning_rate": 5.0884560204540935e-05, "loss": 0.3411, "step": 31180 }, { "epoch": 18.154831199068685, "grad_norm": 0.7718116044998169, "learning_rate": 5.0857006554132736e-05, "loss": 0.3438, "step": 31190 }, { "epoch": 18.160651920838184, "grad_norm": 0.7396775484085083, "learning_rate": 5.0829452643388575e-05, "loss": 0.3481, "step": 31200 }, { "epoch": 18.166472642607683, "grad_norm": 0.721592903137207, "learning_rate": 5.08018984806786e-05, "loss": 0.3438, "step": 31210 }, { "epoch": 18.172293364377182, "grad_norm": 0.7327125668525696, "learning_rate": 5.0774344074373036e-05, "loss": 0.3536, "step": 31220 }, { "epoch": 18.17811408614668, "grad_norm": 0.7948656678199768, "learning_rate": 5.07467894328422e-05, "loss": 0.338, "step": 31230 }, { "epoch": 18.18393480791618, "grad_norm": 0.7179282903671265, "learning_rate": 5.0719234564456454e-05, "loss": 0.3472, "step": 31240 }, { "epoch": 18.18975552968568, "grad_norm": 0.8115698099136353, "learning_rate": 5.0691679477586216e-05, "loss": 0.3514, "step": 31250 }, { "epoch": 18.19557625145518, "grad_norm": 0.7324455976486206, "learning_rate": 5.0664124180602035e-05, "loss": 0.3318, "step": 31260 }, { "epoch": 18.20139697322468, "grad_norm": 0.7380615472793579, "learning_rate": 5.063656868187447e-05, "loss": 0.3386, "step": 31270 }, { "epoch": 18.207217694994178, "grad_norm": 0.7570245862007141, "learning_rate": 5.060901298977413e-05, "loss": 0.3552, "step": 31280 }, { "epoch": 18.213038416763677, "grad_norm": 0.6967106461524963, "learning_rate": 5.0581457112671725e-05, "loss": 0.3543, "step": 31290 }, { "epoch": 18.21885913853318, "grad_norm": 0.7857575416564941, "learning_rate": 5.0553901058938016e-05, "loss": 0.342, "step": 31300 }, { "epoch": 18.22467986030268, "grad_norm": 0.7169231176376343, "learning_rate": 5.052634483694377e-05, "loss": 0.3375, "step": 31310 }, { "epoch": 18.230500582072178, "grad_norm": 0.8260692954063416, "learning_rate": 5.049878845505988e-05, "loss": 0.352, "step": 31320 }, { "epoch": 18.236321303841677, "grad_norm": 0.7980726361274719, "learning_rate": 5.047123192165721e-05, "loss": 0.3415, "step": 31330 }, { "epoch": 18.242142025611177, "grad_norm": 0.7702462077140808, "learning_rate": 5.0443675245106735e-05, "loss": 0.3383, "step": 31340 }, { "epoch": 18.247962747380676, "grad_norm": 0.7840902805328369, "learning_rate": 5.0416118433779426e-05, "loss": 0.3231, "step": 31350 }, { "epoch": 18.253783469150175, "grad_norm": 0.7609191536903381, "learning_rate": 5.038856149604633e-05, "loss": 0.3412, "step": 31360 }, { "epoch": 18.259604190919674, "grad_norm": 0.7513986229896545, "learning_rate": 5.03610044402785e-05, "loss": 0.3335, "step": 31370 }, { "epoch": 18.265424912689173, "grad_norm": 0.7768571376800537, "learning_rate": 5.033344727484707e-05, "loss": 0.3369, "step": 31380 }, { "epoch": 18.271245634458673, "grad_norm": 0.6763558387756348, "learning_rate": 5.030589000812315e-05, "loss": 0.3399, "step": 31390 }, { "epoch": 18.277066356228172, "grad_norm": 0.6945101618766785, "learning_rate": 5.027833264847793e-05, "loss": 0.34, "step": 31400 }, { "epoch": 18.28288707799767, "grad_norm": 0.7109326124191284, "learning_rate": 5.025077520428258e-05, "loss": 0.3498, "step": 31410 }, { "epoch": 18.28870779976717, "grad_norm": 0.7222321629524231, "learning_rate": 5.022321768390837e-05, "loss": 0.3383, "step": 31420 }, { "epoch": 18.29452852153667, "grad_norm": 0.6854539513587952, "learning_rate": 5.0195660095726516e-05, "loss": 0.3416, "step": 31430 }, { "epoch": 18.30034924330617, "grad_norm": 0.7258181571960449, "learning_rate": 5.016810244810829e-05, "loss": 0.3417, "step": 31440 }, { "epoch": 18.306169965075668, "grad_norm": 0.7484527826309204, "learning_rate": 5.0140544749424976e-05, "loss": 0.3468, "step": 31450 }, { "epoch": 18.311990686845167, "grad_norm": 0.780432939529419, "learning_rate": 5.0112987008047874e-05, "loss": 0.3401, "step": 31460 }, { "epoch": 18.31781140861467, "grad_norm": 0.7603837251663208, "learning_rate": 5.008542923234831e-05, "loss": 0.3332, "step": 31470 }, { "epoch": 18.32363213038417, "grad_norm": 0.7459961771965027, "learning_rate": 5.00578714306976e-05, "loss": 0.3371, "step": 31480 }, { "epoch": 18.32945285215367, "grad_norm": 0.7956845164299011, "learning_rate": 5.0030313611467084e-05, "loss": 0.347, "step": 31490 }, { "epoch": 18.335273573923168, "grad_norm": 0.747351348400116, "learning_rate": 5.0002755783028074e-05, "loss": 0.3492, "step": 31500 }, { "epoch": 18.341094295692667, "grad_norm": 0.8316318988800049, "learning_rate": 4.997519795375194e-05, "loss": 0.3403, "step": 31510 }, { "epoch": 18.346915017462166, "grad_norm": 0.7322835922241211, "learning_rate": 4.9947640132010016e-05, "loss": 0.3321, "step": 31520 }, { "epoch": 18.352735739231665, "grad_norm": 0.6912382245063782, "learning_rate": 4.9920082326173625e-05, "loss": 0.3529, "step": 31530 }, { "epoch": 18.358556461001164, "grad_norm": 0.6833052039146423, "learning_rate": 4.9892524544614114e-05, "loss": 0.3551, "step": 31540 }, { "epoch": 18.364377182770664, "grad_norm": 0.8001623749732971, "learning_rate": 4.986496679570283e-05, "loss": 0.3377, "step": 31550 }, { "epoch": 18.370197904540163, "grad_norm": 0.8507075905799866, "learning_rate": 4.983740908781105e-05, "loss": 0.345, "step": 31560 }, { "epoch": 18.376018626309662, "grad_norm": 0.7046404480934143, "learning_rate": 4.9809851429310116e-05, "loss": 0.3473, "step": 31570 }, { "epoch": 18.38183934807916, "grad_norm": 0.8002596497535706, "learning_rate": 4.9782293828571275e-05, "loss": 0.3583, "step": 31580 }, { "epoch": 18.38766006984866, "grad_norm": 0.8776413202285767, "learning_rate": 4.9754736293965846e-05, "loss": 0.346, "step": 31590 }, { "epoch": 18.39348079161816, "grad_norm": 0.8492249846458435, "learning_rate": 4.972717883386502e-05, "loss": 0.3656, "step": 31600 }, { "epoch": 18.39930151338766, "grad_norm": 0.8189017176628113, "learning_rate": 4.9699621456640075e-05, "loss": 0.3505, "step": 31610 }, { "epoch": 18.405122235157158, "grad_norm": 0.691523015499115, "learning_rate": 4.9672064170662214e-05, "loss": 0.3368, "step": 31620 }, { "epoch": 18.410942956926657, "grad_norm": 0.7818205952644348, "learning_rate": 4.9644506984302583e-05, "loss": 0.3521, "step": 31630 }, { "epoch": 18.416763678696157, "grad_norm": 0.8971557021141052, "learning_rate": 4.9616949905932356e-05, "loss": 0.3548, "step": 31640 }, { "epoch": 18.42258440046566, "grad_norm": 0.7764553427696228, "learning_rate": 4.9589392943922615e-05, "loss": 0.3331, "step": 31650 }, { "epoch": 18.42840512223516, "grad_norm": 0.7676321268081665, "learning_rate": 4.956183610664447e-05, "loss": 0.3511, "step": 31660 }, { "epoch": 18.434225844004658, "grad_norm": 0.7803621292114258, "learning_rate": 4.9534279402468945e-05, "loss": 0.3516, "step": 31670 }, { "epoch": 18.440046565774157, "grad_norm": 0.7396121025085449, "learning_rate": 4.9506722839767036e-05, "loss": 0.3327, "step": 31680 }, { "epoch": 18.445867287543656, "grad_norm": 0.7655929327011108, "learning_rate": 4.947916642690972e-05, "loss": 0.3622, "step": 31690 }, { "epoch": 18.451688009313155, "grad_norm": 0.7586347460746765, "learning_rate": 4.9451610172267874e-05, "loss": 0.3432, "step": 31700 }, { "epoch": 18.457508731082655, "grad_norm": 0.6975140571594238, "learning_rate": 4.9424054084212376e-05, "loss": 0.3531, "step": 31710 }, { "epoch": 18.463329452852154, "grad_norm": 0.7207931280136108, "learning_rate": 4.939649817111407e-05, "loss": 0.3522, "step": 31720 }, { "epoch": 18.469150174621653, "grad_norm": 0.8620873689651489, "learning_rate": 4.936894244134365e-05, "loss": 0.3389, "step": 31730 }, { "epoch": 18.474970896391152, "grad_norm": 0.7594047784805298, "learning_rate": 4.9341386903271886e-05, "loss": 0.3398, "step": 31740 }, { "epoch": 18.48079161816065, "grad_norm": 0.7331551313400269, "learning_rate": 4.931383156526936e-05, "loss": 0.3498, "step": 31750 }, { "epoch": 18.48661233993015, "grad_norm": 0.7036387324333191, "learning_rate": 4.92862764357067e-05, "loss": 0.3501, "step": 31760 }, { "epoch": 18.49243306169965, "grad_norm": 0.7314353585243225, "learning_rate": 4.925872152295443e-05, "loss": 0.3521, "step": 31770 }, { "epoch": 18.49825378346915, "grad_norm": 0.829741895198822, "learning_rate": 4.923116683538296e-05, "loss": 0.3683, "step": 31780 }, { "epoch": 18.50407450523865, "grad_norm": 0.8373990654945374, "learning_rate": 4.920361238136273e-05, "loss": 0.3434, "step": 31790 }, { "epoch": 18.509895227008148, "grad_norm": 0.9142295122146606, "learning_rate": 4.9176058169264014e-05, "loss": 0.3494, "step": 31800 }, { "epoch": 18.515715948777647, "grad_norm": 0.7233287692070007, "learning_rate": 4.9148504207457074e-05, "loss": 0.3384, "step": 31810 }, { "epoch": 18.52153667054715, "grad_norm": 0.8350837826728821, "learning_rate": 4.912095050431208e-05, "loss": 0.3548, "step": 31820 }, { "epoch": 18.52735739231665, "grad_norm": 0.8769446015357971, "learning_rate": 4.909339706819911e-05, "loss": 0.3543, "step": 31830 }, { "epoch": 18.533178114086148, "grad_norm": 0.7669773697853088, "learning_rate": 4.906584390748819e-05, "loss": 0.348, "step": 31840 }, { "epoch": 18.538998835855647, "grad_norm": 0.7681154608726501, "learning_rate": 4.9038291030549195e-05, "loss": 0.3427, "step": 31850 }, { "epoch": 18.544819557625146, "grad_norm": 0.8245429992675781, "learning_rate": 4.9010738445751995e-05, "loss": 0.3456, "step": 31860 }, { "epoch": 18.550640279394646, "grad_norm": 0.7025579214096069, "learning_rate": 4.8983186161466364e-05, "loss": 0.3468, "step": 31870 }, { "epoch": 18.556461001164145, "grad_norm": 0.8179356455802917, "learning_rate": 4.89556341860619e-05, "loss": 0.3414, "step": 31880 }, { "epoch": 18.562281722933644, "grad_norm": 0.8177351951599121, "learning_rate": 4.892808252790822e-05, "loss": 0.3384, "step": 31890 }, { "epoch": 18.568102444703143, "grad_norm": 0.7941415309906006, "learning_rate": 4.890053119537475e-05, "loss": 0.3567, "step": 31900 }, { "epoch": 18.573923166472643, "grad_norm": 0.7530911564826965, "learning_rate": 4.887298019683087e-05, "loss": 0.3429, "step": 31910 }, { "epoch": 18.57974388824214, "grad_norm": 0.8562133312225342, "learning_rate": 4.884542954064587e-05, "loss": 0.3495, "step": 31920 }, { "epoch": 18.58556461001164, "grad_norm": 0.7902324795722961, "learning_rate": 4.881787923518887e-05, "loss": 0.3403, "step": 31930 }, { "epoch": 18.59138533178114, "grad_norm": 0.7777092456817627, "learning_rate": 4.879032928882896e-05, "loss": 0.3549, "step": 31940 }, { "epoch": 18.59720605355064, "grad_norm": 0.6859437823295593, "learning_rate": 4.876277970993505e-05, "loss": 0.3387, "step": 31950 }, { "epoch": 18.60302677532014, "grad_norm": 0.8399350047111511, "learning_rate": 4.873523050687602e-05, "loss": 0.3542, "step": 31960 }, { "epoch": 18.608847497089638, "grad_norm": 0.7415204048156738, "learning_rate": 4.870768168802056e-05, "loss": 0.341, "step": 31970 }, { "epoch": 18.614668218859137, "grad_norm": 1.2248846292495728, "learning_rate": 4.868013326173728e-05, "loss": 0.3572, "step": 31980 }, { "epoch": 18.620488940628636, "grad_norm": 0.7213358879089355, "learning_rate": 4.865258523639468e-05, "loss": 0.3425, "step": 31990 }, { "epoch": 18.62630966239814, "grad_norm": 0.7171351909637451, "learning_rate": 4.862503762036109e-05, "loss": 0.3418, "step": 32000 }, { "epoch": 18.63213038416764, "grad_norm": 0.8086410760879517, "learning_rate": 4.859749042200478e-05, "loss": 0.3391, "step": 32010 }, { "epoch": 18.637951105937137, "grad_norm": 0.7391202449798584, "learning_rate": 4.856994364969384e-05, "loss": 0.346, "step": 32020 }, { "epoch": 18.643771827706637, "grad_norm": 0.7739666104316711, "learning_rate": 4.854239731179625e-05, "loss": 0.3594, "step": 32030 }, { "epoch": 18.649592549476136, "grad_norm": 0.8583351969718933, "learning_rate": 4.85148514166799e-05, "loss": 0.3349, "step": 32040 }, { "epoch": 18.655413271245635, "grad_norm": 0.6843629479408264, "learning_rate": 4.8487305972712456e-05, "loss": 0.3502, "step": 32050 }, { "epoch": 18.661233993015134, "grad_norm": 0.6862215995788574, "learning_rate": 4.8459760988261526e-05, "loss": 0.3362, "step": 32060 }, { "epoch": 18.667054714784634, "grad_norm": 0.789715051651001, "learning_rate": 4.843221647169453e-05, "loss": 0.3337, "step": 32070 }, { "epoch": 18.672875436554133, "grad_norm": 0.744361400604248, "learning_rate": 4.840467243137878e-05, "loss": 0.3522, "step": 32080 }, { "epoch": 18.678696158323632, "grad_norm": 0.8183082938194275, "learning_rate": 4.837712887568143e-05, "loss": 0.3473, "step": 32090 }, { "epoch": 18.68451688009313, "grad_norm": 0.7788164615631104, "learning_rate": 4.8349585812969464e-05, "loss": 0.3368, "step": 32100 }, { "epoch": 18.69033760186263, "grad_norm": 0.7521683573722839, "learning_rate": 4.8322043251609775e-05, "loss": 0.3397, "step": 32110 }, { "epoch": 18.69615832363213, "grad_norm": 0.7466205358505249, "learning_rate": 4.8294501199969015e-05, "loss": 0.3496, "step": 32120 }, { "epoch": 18.70197904540163, "grad_norm": 0.6509891748428345, "learning_rate": 4.826695966641376e-05, "loss": 0.3331, "step": 32130 }, { "epoch": 18.707799767171128, "grad_norm": 0.7442918419837952, "learning_rate": 4.823941865931043e-05, "loss": 0.3516, "step": 32140 }, { "epoch": 18.713620488940627, "grad_norm": 0.8395334482192993, "learning_rate": 4.82118781870252e-05, "loss": 0.3436, "step": 32150 }, { "epoch": 18.719441210710126, "grad_norm": 0.6925535798072815, "learning_rate": 4.8184338257924185e-05, "loss": 0.35, "step": 32160 }, { "epoch": 18.725261932479626, "grad_norm": 0.724785566329956, "learning_rate": 4.815679888037324e-05, "loss": 0.3587, "step": 32170 }, { "epoch": 18.73108265424913, "grad_norm": 0.7350343465805054, "learning_rate": 4.8129260062738135e-05, "loss": 0.3549, "step": 32180 }, { "epoch": 18.736903376018628, "grad_norm": 0.8542187809944153, "learning_rate": 4.810172181338445e-05, "loss": 0.3366, "step": 32190 }, { "epoch": 18.742724097788127, "grad_norm": 0.7853947281837463, "learning_rate": 4.807418414067753e-05, "loss": 0.335, "step": 32200 }, { "epoch": 18.748544819557626, "grad_norm": 0.7951407432556152, "learning_rate": 4.804664705298264e-05, "loss": 0.341, "step": 32210 }, { "epoch": 18.754365541327125, "grad_norm": 0.735905647277832, "learning_rate": 4.80191105586648e-05, "loss": 0.3407, "step": 32220 }, { "epoch": 18.760186263096625, "grad_norm": 0.7373834252357483, "learning_rate": 4.799157466608886e-05, "loss": 0.3471, "step": 32230 }, { "epoch": 18.766006984866124, "grad_norm": 0.7328466176986694, "learning_rate": 4.796403938361951e-05, "loss": 0.3377, "step": 32240 }, { "epoch": 18.771827706635623, "grad_norm": 0.787578821182251, "learning_rate": 4.793650471962123e-05, "loss": 0.3421, "step": 32250 }, { "epoch": 18.777648428405122, "grad_norm": 0.8442027568817139, "learning_rate": 4.790897068245835e-05, "loss": 0.3354, "step": 32260 }, { "epoch": 18.78346915017462, "grad_norm": 0.785853922367096, "learning_rate": 4.7881437280494954e-05, "loss": 0.3385, "step": 32270 }, { "epoch": 18.78928987194412, "grad_norm": 0.7716873288154602, "learning_rate": 4.7853904522094965e-05, "loss": 0.352, "step": 32280 }, { "epoch": 18.79511059371362, "grad_norm": 0.6980934143066406, "learning_rate": 4.782637241562215e-05, "loss": 0.3453, "step": 32290 }, { "epoch": 18.80093131548312, "grad_norm": 0.688080370426178, "learning_rate": 4.779884096943997e-05, "loss": 0.3332, "step": 32300 }, { "epoch": 18.80675203725262, "grad_norm": 0.7395124435424805, "learning_rate": 4.777131019191182e-05, "loss": 0.3411, "step": 32310 }, { "epoch": 18.812572759022117, "grad_norm": 0.8450809121131897, "learning_rate": 4.774378009140076e-05, "loss": 0.338, "step": 32320 }, { "epoch": 18.818393480791617, "grad_norm": 0.7611128687858582, "learning_rate": 4.7716250676269735e-05, "loss": 0.3393, "step": 32330 }, { "epoch": 18.824214202561116, "grad_norm": 0.787765383720398, "learning_rate": 4.7688721954881485e-05, "loss": 0.3474, "step": 32340 }, { "epoch": 18.83003492433062, "grad_norm": 0.8255966305732727, "learning_rate": 4.7661193935598446e-05, "loss": 0.3375, "step": 32350 }, { "epoch": 18.835855646100118, "grad_norm": 0.723374605178833, "learning_rate": 4.763366662678296e-05, "loss": 0.3625, "step": 32360 }, { "epoch": 18.841676367869617, "grad_norm": 0.7190979719161987, "learning_rate": 4.7606140036797064e-05, "loss": 0.3449, "step": 32370 }, { "epoch": 18.847497089639116, "grad_norm": 0.7671477794647217, "learning_rate": 4.7578614174002614e-05, "loss": 0.3433, "step": 32380 }, { "epoch": 18.853317811408616, "grad_norm": 0.7784708738327026, "learning_rate": 4.755108904676125e-05, "loss": 0.3551, "step": 32390 }, { "epoch": 18.859138533178115, "grad_norm": 0.7466813325881958, "learning_rate": 4.752356466343436e-05, "loss": 0.3399, "step": 32400 }, { "epoch": 18.864959254947614, "grad_norm": 0.7384832501411438, "learning_rate": 4.7496041032383174e-05, "loss": 0.3453, "step": 32410 }, { "epoch": 18.870779976717113, "grad_norm": 0.7466758489608765, "learning_rate": 4.746851816196858e-05, "loss": 0.3518, "step": 32420 }, { "epoch": 18.876600698486612, "grad_norm": 0.7480140328407288, "learning_rate": 4.744099606055135e-05, "loss": 0.345, "step": 32430 }, { "epoch": 18.88242142025611, "grad_norm": 0.881687581539154, "learning_rate": 4.741347473649193e-05, "loss": 0.3562, "step": 32440 }, { "epoch": 18.88824214202561, "grad_norm": 0.7682483792304993, "learning_rate": 4.738595419815058e-05, "loss": 0.339, "step": 32450 }, { "epoch": 18.89406286379511, "grad_norm": 0.7316564321517944, "learning_rate": 4.7358434453887365e-05, "loss": 0.3383, "step": 32460 }, { "epoch": 18.89988358556461, "grad_norm": 0.7135927081108093, "learning_rate": 4.7330915512061976e-05, "loss": 0.3483, "step": 32470 }, { "epoch": 18.90570430733411, "grad_norm": 0.7053089737892151, "learning_rate": 4.730339738103402e-05, "loss": 0.3221, "step": 32480 }, { "epoch": 18.911525029103608, "grad_norm": 0.8181918263435364, "learning_rate": 4.727588006916271e-05, "loss": 0.3549, "step": 32490 }, { "epoch": 18.917345750873107, "grad_norm": 0.7738187909126282, "learning_rate": 4.724836358480711e-05, "loss": 0.3422, "step": 32500 }, { "epoch": 18.923166472642606, "grad_norm": 0.8612481355667114, "learning_rate": 4.722084793632601e-05, "loss": 0.3628, "step": 32510 }, { "epoch": 18.92898719441211, "grad_norm": 0.8409717082977295, "learning_rate": 4.719333313207792e-05, "loss": 0.351, "step": 32520 }, { "epoch": 18.934807916181608, "grad_norm": 0.763599693775177, "learning_rate": 4.716581918042114e-05, "loss": 0.3349, "step": 32530 }, { "epoch": 18.940628637951107, "grad_norm": 0.7488880753517151, "learning_rate": 4.7138306089713636e-05, "loss": 0.3595, "step": 32540 }, { "epoch": 18.946449359720607, "grad_norm": 0.8060206770896912, "learning_rate": 4.7110793868313183e-05, "loss": 0.3418, "step": 32550 }, { "epoch": 18.952270081490106, "grad_norm": 0.828363299369812, "learning_rate": 4.708328252457729e-05, "loss": 0.3506, "step": 32560 }, { "epoch": 18.958090803259605, "grad_norm": 0.8504939675331116, "learning_rate": 4.7055772066863135e-05, "loss": 0.352, "step": 32570 }, { "epoch": 18.963911525029104, "grad_norm": 0.7897292375564575, "learning_rate": 4.702826250352771e-05, "loss": 0.3686, "step": 32580 }, { "epoch": 18.969732246798603, "grad_norm": 0.877781867980957, "learning_rate": 4.7000753842927653e-05, "loss": 0.3478, "step": 32590 }, { "epoch": 18.975552968568103, "grad_norm": 0.7976166009902954, "learning_rate": 4.6973246093419384e-05, "loss": 0.3553, "step": 32600 }, { "epoch": 18.981373690337602, "grad_norm": 0.75192791223526, "learning_rate": 4.694573926335906e-05, "loss": 0.3335, "step": 32610 }, { "epoch": 18.9871944121071, "grad_norm": 0.7689452171325684, "learning_rate": 4.6918233361102476e-05, "loss": 0.3393, "step": 32620 }, { "epoch": 18.9930151338766, "grad_norm": 0.7403500080108643, "learning_rate": 4.689072839500525e-05, "loss": 0.3301, "step": 32630 }, { "epoch": 18.9988358556461, "grad_norm": 0.7380828857421875, "learning_rate": 4.6863224373422635e-05, "loss": 0.3493, "step": 32640 }, { "epoch": 19.0046565774156, "grad_norm": 0.6713395118713379, "learning_rate": 4.683572130470962e-05, "loss": 0.3576, "step": 32650 }, { "epoch": 19.010477299185098, "grad_norm": 0.7412142157554626, "learning_rate": 4.680821919722094e-05, "loss": 0.3279, "step": 32660 }, { "epoch": 19.016298020954597, "grad_norm": 0.8271126747131348, "learning_rate": 4.6780718059310975e-05, "loss": 0.3449, "step": 32670 }, { "epoch": 19.022118742724096, "grad_norm": 0.7360976338386536, "learning_rate": 4.675321789933389e-05, "loss": 0.3547, "step": 32680 }, { "epoch": 19.027939464493596, "grad_norm": 0.7112531661987305, "learning_rate": 4.6725718725643464e-05, "loss": 0.3335, "step": 32690 }, { "epoch": 19.0337601862631, "grad_norm": 0.8327864408493042, "learning_rate": 4.669822054659323e-05, "loss": 0.3378, "step": 32700 }, { "epoch": 19.039580908032598, "grad_norm": 0.7312436699867249, "learning_rate": 4.667072337053644e-05, "loss": 0.3598, "step": 32710 }, { "epoch": 19.045401629802097, "grad_norm": 0.7078747153282166, "learning_rate": 4.6643227205825965e-05, "loss": 0.3383, "step": 32720 }, { "epoch": 19.051222351571596, "grad_norm": 0.7283867001533508, "learning_rate": 4.6615732060814454e-05, "loss": 0.3361, "step": 32730 }, { "epoch": 19.057043073341095, "grad_norm": 0.6947492361068726, "learning_rate": 4.658823794385417e-05, "loss": 0.3428, "step": 32740 }, { "epoch": 19.062863795110594, "grad_norm": 0.7831544876098633, "learning_rate": 4.6560744863297115e-05, "loss": 0.3363, "step": 32750 }, { "epoch": 19.068684516880094, "grad_norm": 0.7656981945037842, "learning_rate": 4.653325282749498e-05, "loss": 0.3339, "step": 32760 }, { "epoch": 19.074505238649593, "grad_norm": 0.837428867816925, "learning_rate": 4.6505761844799075e-05, "loss": 0.3439, "step": 32770 }, { "epoch": 19.080325960419092, "grad_norm": 0.8031784296035767, "learning_rate": 4.647827192356048e-05, "loss": 0.3465, "step": 32780 }, { "epoch": 19.08614668218859, "grad_norm": 0.7633192539215088, "learning_rate": 4.645078307212989e-05, "loss": 0.3421, "step": 32790 }, { "epoch": 19.09196740395809, "grad_norm": 0.7817515730857849, "learning_rate": 4.642329529885768e-05, "loss": 0.3379, "step": 32800 }, { "epoch": 19.09778812572759, "grad_norm": 0.8793368935585022, "learning_rate": 4.639580861209393e-05, "loss": 0.3574, "step": 32810 }, { "epoch": 19.10360884749709, "grad_norm": 0.7740985155105591, "learning_rate": 4.636832302018835e-05, "loss": 0.3302, "step": 32820 }, { "epoch": 19.109429569266588, "grad_norm": 0.7543525099754333, "learning_rate": 4.6340838531490365e-05, "loss": 0.3363, "step": 32830 }, { "epoch": 19.115250291036087, "grad_norm": 0.7427484393119812, "learning_rate": 4.6313355154349e-05, "loss": 0.3394, "step": 32840 }, { "epoch": 19.121071012805587, "grad_norm": 0.7652061581611633, "learning_rate": 4.6285872897113025e-05, "loss": 0.3507, "step": 32850 }, { "epoch": 19.126891734575086, "grad_norm": 0.7003895044326782, "learning_rate": 4.625839176813077e-05, "loss": 0.3383, "step": 32860 }, { "epoch": 19.132712456344585, "grad_norm": 0.773963451385498, "learning_rate": 4.623091177575031e-05, "loss": 0.3453, "step": 32870 }, { "epoch": 19.138533178114088, "grad_norm": 0.7456396818161011, "learning_rate": 4.620343292831936e-05, "loss": 0.3389, "step": 32880 }, { "epoch": 19.144353899883587, "grad_norm": 0.7083576321601868, "learning_rate": 4.6175955234185206e-05, "loss": 0.3399, "step": 32890 }, { "epoch": 19.150174621653086, "grad_norm": 0.7872650623321533, "learning_rate": 4.614847870169492e-05, "loss": 0.3479, "step": 32900 }, { "epoch": 19.155995343422585, "grad_norm": 0.7452362775802612, "learning_rate": 4.612100333919509e-05, "loss": 0.3374, "step": 32910 }, { "epoch": 19.161816065192085, "grad_norm": 0.7564975023269653, "learning_rate": 4.609352915503202e-05, "loss": 0.3269, "step": 32920 }, { "epoch": 19.167636786961584, "grad_norm": 0.7357543706893921, "learning_rate": 4.606605615755166e-05, "loss": 0.3382, "step": 32930 }, { "epoch": 19.173457508731083, "grad_norm": 0.7436339855194092, "learning_rate": 4.6038584355099576e-05, "loss": 0.3414, "step": 32940 }, { "epoch": 19.179278230500582, "grad_norm": 0.7084823846817017, "learning_rate": 4.6011113756020964e-05, "loss": 0.3483, "step": 32950 }, { "epoch": 19.18509895227008, "grad_norm": 0.7018502354621887, "learning_rate": 4.598364436866066e-05, "loss": 0.3417, "step": 32960 }, { "epoch": 19.19091967403958, "grad_norm": 0.8343421220779419, "learning_rate": 4.595617620136316e-05, "loss": 0.3507, "step": 32970 }, { "epoch": 19.19674039580908, "grad_norm": 0.798288106918335, "learning_rate": 4.592870926247257e-05, "loss": 0.3417, "step": 32980 }, { "epoch": 19.20256111757858, "grad_norm": 0.767014741897583, "learning_rate": 4.5901243560332594e-05, "loss": 0.3358, "step": 32990 }, { "epoch": 19.20838183934808, "grad_norm": 0.8004962205886841, "learning_rate": 4.587377910328662e-05, "loss": 0.3363, "step": 33000 }, { "epoch": 19.214202561117578, "grad_norm": 0.7057169079780579, "learning_rate": 4.5846315899677586e-05, "loss": 0.3414, "step": 33010 }, { "epoch": 19.220023282887077, "grad_norm": 0.8351920247077942, "learning_rate": 4.5818853957848114e-05, "loss": 0.3395, "step": 33020 }, { "epoch": 19.225844004656576, "grad_norm": 0.9034773111343384, "learning_rate": 4.579139328614043e-05, "loss": 0.3431, "step": 33030 }, { "epoch": 19.231664726426075, "grad_norm": 0.6568132042884827, "learning_rate": 4.576393389289633e-05, "loss": 0.3395, "step": 33040 }, { "epoch": 19.237485448195578, "grad_norm": 0.7183129787445068, "learning_rate": 4.573647578645728e-05, "loss": 0.3503, "step": 33050 }, { "epoch": 19.243306169965077, "grad_norm": 0.8431748151779175, "learning_rate": 4.57090189751643e-05, "loss": 0.3366, "step": 33060 }, { "epoch": 19.249126891734576, "grad_norm": 0.808697521686554, "learning_rate": 4.568156346735806e-05, "loss": 0.3355, "step": 33070 }, { "epoch": 19.254947613504076, "grad_norm": 0.6851001381874084, "learning_rate": 4.565410927137882e-05, "loss": 0.3259, "step": 33080 }, { "epoch": 19.260768335273575, "grad_norm": 0.7239410281181335, "learning_rate": 4.562665639556644e-05, "loss": 0.3347, "step": 33090 }, { "epoch": 19.266589057043074, "grad_norm": 0.8064601421356201, "learning_rate": 4.559920484826037e-05, "loss": 0.3539, "step": 33100 }, { "epoch": 19.272409778812573, "grad_norm": 0.7697112560272217, "learning_rate": 4.5571754637799665e-05, "loss": 0.3488, "step": 33110 }, { "epoch": 19.278230500582072, "grad_norm": 0.7777717709541321, "learning_rate": 4.554430577252298e-05, "loss": 0.3322, "step": 33120 }, { "epoch": 19.28405122235157, "grad_norm": 0.7879396677017212, "learning_rate": 4.551685826076858e-05, "loss": 0.339, "step": 33130 }, { "epoch": 19.28987194412107, "grad_norm": 0.79716956615448, "learning_rate": 4.5489412110874246e-05, "loss": 0.3472, "step": 33140 }, { "epoch": 19.29569266589057, "grad_norm": 0.772157609462738, "learning_rate": 4.5461967331177444e-05, "loss": 0.3508, "step": 33150 }, { "epoch": 19.30151338766007, "grad_norm": 0.7485741972923279, "learning_rate": 4.5434523930015115e-05, "loss": 0.3329, "step": 33160 }, { "epoch": 19.30733410942957, "grad_norm": 0.6818084120750427, "learning_rate": 4.540708191572388e-05, "loss": 0.3562, "step": 33170 }, { "epoch": 19.313154831199068, "grad_norm": 0.7629798054695129, "learning_rate": 4.537964129663991e-05, "loss": 0.3406, "step": 33180 }, { "epoch": 19.318975552968567, "grad_norm": 0.8300849199295044, "learning_rate": 4.535220208109889e-05, "loss": 0.345, "step": 33190 }, { "epoch": 19.324796274738066, "grad_norm": 0.7350436449050903, "learning_rate": 4.5324764277436194e-05, "loss": 0.3533, "step": 33200 }, { "epoch": 19.330616996507565, "grad_norm": 0.7553167343139648, "learning_rate": 4.529732789398664e-05, "loss": 0.342, "step": 33210 }, { "epoch": 19.336437718277068, "grad_norm": 0.7740887403488159, "learning_rate": 4.526989293908472e-05, "loss": 0.3366, "step": 33220 }, { "epoch": 19.342258440046567, "grad_norm": 0.7641907334327698, "learning_rate": 4.524245942106442e-05, "loss": 0.3413, "step": 33230 }, { "epoch": 19.348079161816067, "grad_norm": 0.8209200501441956, "learning_rate": 4.5215027348259345e-05, "loss": 0.3381, "step": 33240 }, { "epoch": 19.353899883585566, "grad_norm": 0.8785572052001953, "learning_rate": 4.5187596729002616e-05, "loss": 0.3378, "step": 33250 }, { "epoch": 19.359720605355065, "grad_norm": 0.7954618334770203, "learning_rate": 4.516016757162693e-05, "loss": 0.3344, "step": 33260 }, { "epoch": 19.365541327124564, "grad_norm": 0.7850214838981628, "learning_rate": 4.513273988446457e-05, "loss": 0.3456, "step": 33270 }, { "epoch": 19.371362048894063, "grad_norm": 0.7909144163131714, "learning_rate": 4.5105313675847296e-05, "loss": 0.3426, "step": 33280 }, { "epoch": 19.377182770663563, "grad_norm": 0.6843576431274414, "learning_rate": 4.5077888954106495e-05, "loss": 0.3227, "step": 33290 }, { "epoch": 19.383003492433062, "grad_norm": 0.7487248182296753, "learning_rate": 4.505046572757309e-05, "loss": 0.3495, "step": 33300 }, { "epoch": 19.38882421420256, "grad_norm": 0.7395634651184082, "learning_rate": 4.502304400457749e-05, "loss": 0.3454, "step": 33310 }, { "epoch": 19.39464493597206, "grad_norm": 0.7380208969116211, "learning_rate": 4.499562379344973e-05, "loss": 0.3503, "step": 33320 }, { "epoch": 19.40046565774156, "grad_norm": 0.742085874080658, "learning_rate": 4.4968205102519306e-05, "loss": 0.3586, "step": 33330 }, { "epoch": 19.40628637951106, "grad_norm": 0.832878828048706, "learning_rate": 4.494078794011532e-05, "loss": 0.335, "step": 33340 }, { "epoch": 19.412107101280558, "grad_norm": 0.7669912576675415, "learning_rate": 4.491337231456639e-05, "loss": 0.3409, "step": 33350 }, { "epoch": 19.417927823050057, "grad_norm": 0.8586367964744568, "learning_rate": 4.4885958234200634e-05, "loss": 0.3505, "step": 33360 }, { "epoch": 19.423748544819556, "grad_norm": 0.7839811444282532, "learning_rate": 4.485854570734575e-05, "loss": 0.3595, "step": 33370 }, { "epoch": 19.429569266589056, "grad_norm": 0.7692723274230957, "learning_rate": 4.483113474232891e-05, "loss": 0.3447, "step": 33380 }, { "epoch": 19.435389988358555, "grad_norm": 0.7576361894607544, "learning_rate": 4.480372534747688e-05, "loss": 0.3333, "step": 33390 }, { "epoch": 19.441210710128058, "grad_norm": 0.7630125880241394, "learning_rate": 4.477631753111588e-05, "loss": 0.3408, "step": 33400 }, { "epoch": 19.447031431897557, "grad_norm": 0.7971280813217163, "learning_rate": 4.4748911301571686e-05, "loss": 0.3408, "step": 33410 }, { "epoch": 19.452852153667056, "grad_norm": 0.8367514610290527, "learning_rate": 4.472150666716961e-05, "loss": 0.338, "step": 33420 }, { "epoch": 19.458672875436555, "grad_norm": 0.8383968472480774, "learning_rate": 4.469410363623442e-05, "loss": 0.3348, "step": 33430 }, { "epoch": 19.464493597206054, "grad_norm": 0.6750313639640808, "learning_rate": 4.466670221709044e-05, "loss": 0.3306, "step": 33440 }, { "epoch": 19.470314318975554, "grad_norm": 0.6842035055160522, "learning_rate": 4.463930241806154e-05, "loss": 0.3314, "step": 33450 }, { "epoch": 19.476135040745053, "grad_norm": 0.7889201641082764, "learning_rate": 4.4611904247471006e-05, "loss": 0.3424, "step": 33460 }, { "epoch": 19.481955762514552, "grad_norm": 0.72495037317276, "learning_rate": 4.458450771364171e-05, "loss": 0.3387, "step": 33470 }, { "epoch": 19.48777648428405, "grad_norm": 0.7363576889038086, "learning_rate": 4.4557112824895965e-05, "loss": 0.3397, "step": 33480 }, { "epoch": 19.49359720605355, "grad_norm": 0.6914657950401306, "learning_rate": 4.452971958955563e-05, "loss": 0.3503, "step": 33490 }, { "epoch": 19.49941792782305, "grad_norm": 0.7652488946914673, "learning_rate": 4.450232801594208e-05, "loss": 0.3485, "step": 33500 }, { "epoch": 19.50523864959255, "grad_norm": 0.7597461938858032, "learning_rate": 4.447493811237609e-05, "loss": 0.3494, "step": 33510 }, { "epoch": 19.511059371362048, "grad_norm": 0.7060930132865906, "learning_rate": 4.444754988717804e-05, "loss": 0.3351, "step": 33520 }, { "epoch": 19.516880093131547, "grad_norm": 0.7017679214477539, "learning_rate": 4.442016334866771e-05, "loss": 0.3442, "step": 33530 }, { "epoch": 19.522700814901047, "grad_norm": 0.7248622179031372, "learning_rate": 4.4392778505164445e-05, "loss": 0.34, "step": 33540 }, { "epoch": 19.528521536670546, "grad_norm": 0.7606011629104614, "learning_rate": 4.436539536498702e-05, "loss": 0.3437, "step": 33550 }, { "epoch": 19.534342258440045, "grad_norm": 0.7212501168251038, "learning_rate": 4.433801393645369e-05, "loss": 0.3418, "step": 33560 }, { "epoch": 19.540162980209544, "grad_norm": 0.6886502504348755, "learning_rate": 4.431063422788226e-05, "loss": 0.3374, "step": 33570 }, { "epoch": 19.545983701979047, "grad_norm": 0.8165651559829712, "learning_rate": 4.428325624758991e-05, "loss": 0.3471, "step": 33580 }, { "epoch": 19.551804423748546, "grad_norm": 0.7683725357055664, "learning_rate": 4.4255880003893366e-05, "loss": 0.3481, "step": 33590 }, { "epoch": 19.557625145518045, "grad_norm": 0.7908358573913574, "learning_rate": 4.422850550510884e-05, "loss": 0.3503, "step": 33600 }, { "epoch": 19.563445867287545, "grad_norm": 0.7134390473365784, "learning_rate": 4.4201132759551934e-05, "loss": 0.3484, "step": 33610 }, { "epoch": 19.569266589057044, "grad_norm": 0.7284324765205383, "learning_rate": 4.4173761775537804e-05, "loss": 0.3439, "step": 33620 }, { "epoch": 19.575087310826543, "grad_norm": 0.7649917602539062, "learning_rate": 4.414639256138099e-05, "loss": 0.3418, "step": 33630 }, { "epoch": 19.580908032596042, "grad_norm": 0.7640083432197571, "learning_rate": 4.411902512539557e-05, "loss": 0.3406, "step": 33640 }, { "epoch": 19.58672875436554, "grad_norm": 0.7287074327468872, "learning_rate": 4.4091659475895044e-05, "loss": 0.3457, "step": 33650 }, { "epoch": 19.59254947613504, "grad_norm": 0.794236421585083, "learning_rate": 4.406429562119235e-05, "loss": 0.3506, "step": 33660 }, { "epoch": 19.59837019790454, "grad_norm": 0.667019784450531, "learning_rate": 4.4036933569599945e-05, "loss": 0.3417, "step": 33670 }, { "epoch": 19.60419091967404, "grad_norm": 0.7565139532089233, "learning_rate": 4.400957332942965e-05, "loss": 0.3404, "step": 33680 }, { "epoch": 19.61001164144354, "grad_norm": 0.6840291619300842, "learning_rate": 4.3982214908992844e-05, "loss": 0.3455, "step": 33690 }, { "epoch": 19.615832363213038, "grad_norm": 0.6802547574043274, "learning_rate": 4.3954858316600235e-05, "loss": 0.345, "step": 33700 }, { "epoch": 19.621653084982537, "grad_norm": 0.77154940366745, "learning_rate": 4.392750356056205e-05, "loss": 0.3424, "step": 33710 }, { "epoch": 19.627473806752036, "grad_norm": 1.0578277111053467, "learning_rate": 4.390015064918798e-05, "loss": 0.3404, "step": 33720 }, { "epoch": 19.633294528521535, "grad_norm": 0.8373546004295349, "learning_rate": 4.387279959078705e-05, "loss": 0.3341, "step": 33730 }, { "epoch": 19.639115250291034, "grad_norm": 0.8424969911575317, "learning_rate": 4.384545039366786e-05, "loss": 0.3487, "step": 33740 }, { "epoch": 19.644935972060537, "grad_norm": 0.666834831237793, "learning_rate": 4.381810306613831e-05, "loss": 0.3344, "step": 33750 }, { "epoch": 19.650756693830036, "grad_norm": 0.7717517018318176, "learning_rate": 4.3790757616505826e-05, "loss": 0.3498, "step": 33760 }, { "epoch": 19.656577415599536, "grad_norm": 0.7753255367279053, "learning_rate": 4.376341405307725e-05, "loss": 0.3446, "step": 33770 }, { "epoch": 19.662398137369035, "grad_norm": 0.7141785621643066, "learning_rate": 4.37360723841588e-05, "loss": 0.3292, "step": 33780 }, { "epoch": 19.668218859138534, "grad_norm": 0.7232694029808044, "learning_rate": 4.370873261805619e-05, "loss": 0.3418, "step": 33790 }, { "epoch": 19.674039580908033, "grad_norm": 0.7749433517456055, "learning_rate": 4.368139476307449e-05, "loss": 0.3442, "step": 33800 }, { "epoch": 19.679860302677533, "grad_norm": 0.7461733818054199, "learning_rate": 4.365405882751822e-05, "loss": 0.3309, "step": 33810 }, { "epoch": 19.68568102444703, "grad_norm": 0.7364381551742554, "learning_rate": 4.3626724819691326e-05, "loss": 0.3332, "step": 33820 }, { "epoch": 19.69150174621653, "grad_norm": 0.6850901246070862, "learning_rate": 4.359939274789715e-05, "loss": 0.3292, "step": 33830 }, { "epoch": 19.69732246798603, "grad_norm": 0.6812374591827393, "learning_rate": 4.357206262043848e-05, "loss": 0.3558, "step": 33840 }, { "epoch": 19.70314318975553, "grad_norm": 0.718180775642395, "learning_rate": 4.354473444561745e-05, "loss": 0.3365, "step": 33850 }, { "epoch": 19.70896391152503, "grad_norm": 0.7515060901641846, "learning_rate": 4.3517408231735644e-05, "loss": 0.335, "step": 33860 }, { "epoch": 19.714784633294528, "grad_norm": 0.6873344779014587, "learning_rate": 4.3490083987094086e-05, "loss": 0.329, "step": 33870 }, { "epoch": 19.720605355064027, "grad_norm": 0.8884050846099854, "learning_rate": 4.34627617199931e-05, "loss": 0.3371, "step": 33880 }, { "epoch": 19.726426076833526, "grad_norm": 0.7176953554153442, "learning_rate": 4.3435441438732526e-05, "loss": 0.3356, "step": 33890 }, { "epoch": 19.732246798603025, "grad_norm": 0.7973713278770447, "learning_rate": 4.340812315161149e-05, "loss": 0.3418, "step": 33900 }, { "epoch": 19.738067520372525, "grad_norm": 0.7413381338119507, "learning_rate": 4.338080686692859e-05, "loss": 0.3418, "step": 33910 }, { "epoch": 19.743888242142027, "grad_norm": 0.7312262654304504, "learning_rate": 4.3353492592981816e-05, "loss": 0.3374, "step": 33920 }, { "epoch": 19.749708963911527, "grad_norm": 0.7184549570083618, "learning_rate": 4.3326180338068485e-05, "loss": 0.3217, "step": 33930 }, { "epoch": 19.755529685681026, "grad_norm": 0.7438695430755615, "learning_rate": 4.3298870110485356e-05, "loss": 0.334, "step": 33940 }, { "epoch": 19.761350407450525, "grad_norm": 0.7380194664001465, "learning_rate": 4.3271561918528567e-05, "loss": 0.3448, "step": 33950 }, { "epoch": 19.767171129220024, "grad_norm": 0.7571950554847717, "learning_rate": 4.324425577049359e-05, "loss": 0.3432, "step": 33960 }, { "epoch": 19.772991850989523, "grad_norm": 0.7372188568115234, "learning_rate": 4.321695167467535e-05, "loss": 0.3571, "step": 33970 }, { "epoch": 19.778812572759023, "grad_norm": 0.7966983318328857, "learning_rate": 4.3189649639368093e-05, "loss": 0.3378, "step": 33980 }, { "epoch": 19.784633294528522, "grad_norm": 0.7272815704345703, "learning_rate": 4.316234967286547e-05, "loss": 0.3312, "step": 33990 }, { "epoch": 19.79045401629802, "grad_norm": 0.7704598307609558, "learning_rate": 4.313505178346046e-05, "loss": 0.3498, "step": 34000 }, { "epoch": 19.79627473806752, "grad_norm": 0.7544799447059631, "learning_rate": 4.3107755979445465e-05, "loss": 0.3428, "step": 34010 }, { "epoch": 19.80209545983702, "grad_norm": 0.8148561716079712, "learning_rate": 4.308046226911224e-05, "loss": 0.3436, "step": 34020 }, { "epoch": 19.80791618160652, "grad_norm": 0.7024657130241394, "learning_rate": 4.305317066075185e-05, "loss": 0.3508, "step": 34030 }, { "epoch": 19.813736903376018, "grad_norm": 0.7192490696907043, "learning_rate": 4.302588116265482e-05, "loss": 0.3424, "step": 34040 }, { "epoch": 19.819557625145517, "grad_norm": 0.7245880961418152, "learning_rate": 4.299859378311094e-05, "loss": 0.333, "step": 34050 }, { "epoch": 19.825378346915016, "grad_norm": 0.7604085803031921, "learning_rate": 4.2971308530409424e-05, "loss": 0.334, "step": 34060 }, { "epoch": 19.831199068684516, "grad_norm": 0.8304821252822876, "learning_rate": 4.2944025412838765e-05, "loss": 0.3464, "step": 34070 }, { "epoch": 19.837019790454015, "grad_norm": 0.7121114134788513, "learning_rate": 4.291674443868689e-05, "loss": 0.3275, "step": 34080 }, { "epoch": 19.842840512223514, "grad_norm": 0.6199884414672852, "learning_rate": 4.288946561624104e-05, "loss": 0.3292, "step": 34090 }, { "epoch": 19.848661233993017, "grad_norm": 0.6795803904533386, "learning_rate": 4.2862188953787794e-05, "loss": 0.3338, "step": 34100 }, { "epoch": 19.854481955762516, "grad_norm": 0.7955459356307983, "learning_rate": 4.283491445961308e-05, "loss": 0.3494, "step": 34110 }, { "epoch": 19.860302677532015, "grad_norm": 0.6851580142974854, "learning_rate": 4.2807642142002155e-05, "loss": 0.3459, "step": 34120 }, { "epoch": 19.866123399301514, "grad_norm": 0.7393444180488586, "learning_rate": 4.278037200923966e-05, "loss": 0.3325, "step": 34130 }, { "epoch": 19.871944121071014, "grad_norm": 0.7059219479560852, "learning_rate": 4.275310406960953e-05, "loss": 0.3313, "step": 34140 }, { "epoch": 19.877764842840513, "grad_norm": 0.7053835988044739, "learning_rate": 4.272583833139502e-05, "loss": 0.353, "step": 34150 }, { "epoch": 19.883585564610012, "grad_norm": 0.7062874436378479, "learning_rate": 4.2698574802878794e-05, "loss": 0.3383, "step": 34160 }, { "epoch": 19.88940628637951, "grad_norm": 0.6870304346084595, "learning_rate": 4.2671313492342734e-05, "loss": 0.3473, "step": 34170 }, { "epoch": 19.89522700814901, "grad_norm": 0.7226932644844055, "learning_rate": 4.264405440806813e-05, "loss": 0.3427, "step": 34180 }, { "epoch": 19.90104772991851, "grad_norm": 0.7560579180717468, "learning_rate": 4.26167975583356e-05, "loss": 0.3426, "step": 34190 }, { "epoch": 19.90686845168801, "grad_norm": 0.8021458387374878, "learning_rate": 4.2589542951425e-05, "loss": 0.3461, "step": 34200 }, { "epoch": 19.912689173457508, "grad_norm": 0.78035569190979, "learning_rate": 4.2562290595615615e-05, "loss": 0.3567, "step": 34210 }, { "epoch": 19.918509895227007, "grad_norm": 0.7851737141609192, "learning_rate": 4.2535040499185946e-05, "loss": 0.3383, "step": 34220 }, { "epoch": 19.924330616996507, "grad_norm": 0.7996530532836914, "learning_rate": 4.250779267041387e-05, "loss": 0.3365, "step": 34230 }, { "epoch": 19.930151338766006, "grad_norm": 0.7429865598678589, "learning_rate": 4.248054711757657e-05, "loss": 0.3307, "step": 34240 }, { "epoch": 19.935972060535505, "grad_norm": 0.7711613774299622, "learning_rate": 4.245330384895052e-05, "loss": 0.3428, "step": 34250 }, { "epoch": 19.941792782305004, "grad_norm": 0.7696585655212402, "learning_rate": 4.242606287281151e-05, "loss": 0.3332, "step": 34260 }, { "epoch": 19.947613504074504, "grad_norm": 1.2151157855987549, "learning_rate": 4.2398824197434595e-05, "loss": 0.3515, "step": 34270 }, { "epoch": 19.953434225844006, "grad_norm": 0.7175403237342834, "learning_rate": 4.23715878310942e-05, "loss": 0.3397, "step": 34280 }, { "epoch": 19.959254947613505, "grad_norm": 0.7058170437812805, "learning_rate": 4.234435378206402e-05, "loss": 0.3581, "step": 34290 }, { "epoch": 19.965075669383005, "grad_norm": 0.6595849990844727, "learning_rate": 4.2317122058617006e-05, "loss": 0.3282, "step": 34300 }, { "epoch": 19.970896391152504, "grad_norm": 0.758228063583374, "learning_rate": 4.2289892669025485e-05, "loss": 0.3436, "step": 34310 }, { "epoch": 19.976717112922003, "grad_norm": 0.7238243818283081, "learning_rate": 4.226266562156097e-05, "loss": 0.3363, "step": 34320 }, { "epoch": 19.982537834691502, "grad_norm": 0.8042359352111816, "learning_rate": 4.223544092449435e-05, "loss": 0.3449, "step": 34330 }, { "epoch": 19.988358556461, "grad_norm": 0.729564368724823, "learning_rate": 4.2208218586095784e-05, "loss": 0.3361, "step": 34340 }, { "epoch": 19.9941792782305, "grad_norm": 0.7643580436706543, "learning_rate": 4.218099861463466e-05, "loss": 0.3493, "step": 34350 }, { "epoch": 20.0, "grad_norm": 0.7492313385009766, "learning_rate": 4.215378101837972e-05, "loss": 0.359, "step": 34360 }, { "epoch": 20.0058207217695, "grad_norm": 0.7035341858863831, "learning_rate": 4.2126565805598937e-05, "loss": 0.3271, "step": 34370 }, { "epoch": 20.011641443539, "grad_norm": 0.657242476940155, "learning_rate": 4.209935298455957e-05, "loss": 0.3472, "step": 34380 }, { "epoch": 20.017462165308498, "grad_norm": 0.725962221622467, "learning_rate": 4.207214256352817e-05, "loss": 0.3369, "step": 34390 }, { "epoch": 20.023282887077997, "grad_norm": 0.8228978514671326, "learning_rate": 4.2044934550770524e-05, "loss": 0.3503, "step": 34400 }, { "epoch": 20.029103608847496, "grad_norm": 0.6943507194519043, "learning_rate": 4.201772895455174e-05, "loss": 0.3324, "step": 34410 }, { "epoch": 20.034924330616995, "grad_norm": 0.7264478206634521, "learning_rate": 4.199052578313613e-05, "loss": 0.3434, "step": 34420 }, { "epoch": 20.040745052386495, "grad_norm": 0.7429852485656738, "learning_rate": 4.1963325044787294e-05, "loss": 0.3468, "step": 34430 }, { "epoch": 20.046565774155994, "grad_norm": 0.848974883556366, "learning_rate": 4.193612674776814e-05, "loss": 0.3469, "step": 34440 }, { "epoch": 20.052386495925496, "grad_norm": 0.7689423561096191, "learning_rate": 4.1908930900340745e-05, "loss": 0.3486, "step": 34450 }, { "epoch": 20.058207217694996, "grad_norm": 0.7937766909599304, "learning_rate": 4.1881737510766536e-05, "loss": 0.348, "step": 34460 }, { "epoch": 20.064027939464495, "grad_norm": 0.6949288249015808, "learning_rate": 4.185454658730609e-05, "loss": 0.3465, "step": 34470 }, { "epoch": 20.069848661233994, "grad_norm": 0.8087432980537415, "learning_rate": 4.1827358138219355e-05, "loss": 0.3388, "step": 34480 }, { "epoch": 20.075669383003493, "grad_norm": 0.8497655987739563, "learning_rate": 4.1800172171765404e-05, "loss": 0.3574, "step": 34490 }, { "epoch": 20.081490104772993, "grad_norm": 0.722882866859436, "learning_rate": 4.177298869620264e-05, "loss": 0.3375, "step": 34500 }, { "epoch": 20.087310826542492, "grad_norm": 0.8632571697235107, "learning_rate": 4.1745807719788705e-05, "loss": 0.3513, "step": 34510 }, { "epoch": 20.09313154831199, "grad_norm": 0.752866268157959, "learning_rate": 4.1718629250780445e-05, "loss": 0.3446, "step": 34520 }, { "epoch": 20.09895227008149, "grad_norm": 0.8496826887130737, "learning_rate": 4.1691453297433956e-05, "loss": 0.335, "step": 34530 }, { "epoch": 20.10477299185099, "grad_norm": 0.7647242546081543, "learning_rate": 4.166427986800457e-05, "loss": 0.3305, "step": 34540 }, { "epoch": 20.11059371362049, "grad_norm": 0.7897390127182007, "learning_rate": 4.163710897074688e-05, "loss": 0.3417, "step": 34550 }, { "epoch": 20.116414435389988, "grad_norm": 0.7422825694084167, "learning_rate": 4.1609940613914686e-05, "loss": 0.3479, "step": 34560 }, { "epoch": 20.122235157159487, "grad_norm": 0.7164402008056641, "learning_rate": 4.1582774805760996e-05, "loss": 0.3252, "step": 34570 }, { "epoch": 20.128055878928986, "grad_norm": 0.6867008209228516, "learning_rate": 4.155561155453809e-05, "loss": 0.3397, "step": 34580 }, { "epoch": 20.133876600698486, "grad_norm": 0.7832051515579224, "learning_rate": 4.15284508684974e-05, "loss": 0.3528, "step": 34590 }, { "epoch": 20.139697322467985, "grad_norm": 0.7353395223617554, "learning_rate": 4.1501292755889675e-05, "loss": 0.3359, "step": 34600 }, { "epoch": 20.145518044237484, "grad_norm": 0.7369534969329834, "learning_rate": 4.1474137224964833e-05, "loss": 0.3332, "step": 34610 }, { "epoch": 20.151338766006983, "grad_norm": 0.7298052906990051, "learning_rate": 4.144698428397197e-05, "loss": 0.3405, "step": 34620 }, { "epoch": 20.157159487776486, "grad_norm": 0.7347239255905151, "learning_rate": 4.1419833941159466e-05, "loss": 0.3417, "step": 34630 }, { "epoch": 20.162980209545985, "grad_norm": 0.7693282961845398, "learning_rate": 4.1392686204774846e-05, "loss": 0.3494, "step": 34640 }, { "epoch": 20.168800931315484, "grad_norm": 0.8498129844665527, "learning_rate": 4.13655410830649e-05, "loss": 0.341, "step": 34650 }, { "epoch": 20.174621653084984, "grad_norm": 0.7338592410087585, "learning_rate": 4.1338398584275594e-05, "loss": 0.3396, "step": 34660 }, { "epoch": 20.180442374854483, "grad_norm": 0.7136600613594055, "learning_rate": 4.1311258716652104e-05, "loss": 0.3314, "step": 34670 }, { "epoch": 20.186263096623982, "grad_norm": 0.712332546710968, "learning_rate": 4.128412148843881e-05, "loss": 0.3354, "step": 34680 }, { "epoch": 20.19208381839348, "grad_norm": 0.7340875267982483, "learning_rate": 4.125698690787926e-05, "loss": 0.3425, "step": 34690 }, { "epoch": 20.19790454016298, "grad_norm": 0.7561855912208557, "learning_rate": 4.1229854983216245e-05, "loss": 0.3338, "step": 34700 }, { "epoch": 20.20372526193248, "grad_norm": 0.8030306696891785, "learning_rate": 4.120272572269175e-05, "loss": 0.3416, "step": 34710 }, { "epoch": 20.20954598370198, "grad_norm": 0.7260233163833618, "learning_rate": 4.117559913454687e-05, "loss": 0.3373, "step": 34720 }, { "epoch": 20.215366705471478, "grad_norm": 0.7310113310813904, "learning_rate": 4.114847522702201e-05, "loss": 0.3381, "step": 34730 }, { "epoch": 20.221187427240977, "grad_norm": 0.7368890643119812, "learning_rate": 4.112135400835664e-05, "loss": 0.3404, "step": 34740 }, { "epoch": 20.227008149010477, "grad_norm": 0.6835466027259827, "learning_rate": 4.109423548678949e-05, "loss": 0.3357, "step": 34750 }, { "epoch": 20.232828870779976, "grad_norm": 0.7424144744873047, "learning_rate": 4.106711967055848e-05, "loss": 0.341, "step": 34760 }, { "epoch": 20.238649592549475, "grad_norm": 0.748163104057312, "learning_rate": 4.1040006567900636e-05, "loss": 0.33, "step": 34770 }, { "epoch": 20.244470314318974, "grad_norm": 0.809276819229126, "learning_rate": 4.101289618705224e-05, "loss": 0.3381, "step": 34780 }, { "epoch": 20.250291036088473, "grad_norm": 0.8161876201629639, "learning_rate": 4.0985788536248675e-05, "loss": 0.3485, "step": 34790 }, { "epoch": 20.256111757857976, "grad_norm": 0.7947666049003601, "learning_rate": 4.095868362372454e-05, "loss": 0.3483, "step": 34800 }, { "epoch": 20.261932479627475, "grad_norm": 0.7506164908409119, "learning_rate": 4.0931581457713614e-05, "loss": 0.3339, "step": 34810 }, { "epoch": 20.267753201396975, "grad_norm": 0.8096303939819336, "learning_rate": 4.09044820464488e-05, "loss": 0.3535, "step": 34820 }, { "epoch": 20.273573923166474, "grad_norm": 0.7324687242507935, "learning_rate": 4.087738539816219e-05, "loss": 0.3304, "step": 34830 }, { "epoch": 20.279394644935973, "grad_norm": 0.7469944357872009, "learning_rate": 4.085029152108501e-05, "loss": 0.3502, "step": 34840 }, { "epoch": 20.285215366705472, "grad_norm": 0.6316313147544861, "learning_rate": 4.0823200423447714e-05, "loss": 0.3352, "step": 34850 }, { "epoch": 20.29103608847497, "grad_norm": 0.7292985320091248, "learning_rate": 4.079611211347981e-05, "loss": 0.3374, "step": 34860 }, { "epoch": 20.29685681024447, "grad_norm": 0.6351380944252014, "learning_rate": 4.076902659941002e-05, "loss": 0.3142, "step": 34870 }, { "epoch": 20.30267753201397, "grad_norm": 0.7251527905464172, "learning_rate": 4.074194388946624e-05, "loss": 0.3307, "step": 34880 }, { "epoch": 20.30849825378347, "grad_norm": 0.7062609791755676, "learning_rate": 4.071486399187545e-05, "loss": 0.3352, "step": 34890 }, { "epoch": 20.31431897555297, "grad_norm": 0.6925460696220398, "learning_rate": 4.0687786914863836e-05, "loss": 0.3324, "step": 34900 }, { "epoch": 20.320139697322467, "grad_norm": 0.753831684589386, "learning_rate": 4.0660712666656666e-05, "loss": 0.3487, "step": 34910 }, { "epoch": 20.325960419091967, "grad_norm": 0.8424142003059387, "learning_rate": 4.0633641255478394e-05, "loss": 0.3568, "step": 34920 }, { "epoch": 20.331781140861466, "grad_norm": 0.7585978507995605, "learning_rate": 4.0606572689552624e-05, "loss": 0.339, "step": 34930 }, { "epoch": 20.337601862630965, "grad_norm": 0.6995795369148254, "learning_rate": 4.0579506977102036e-05, "loss": 0.3376, "step": 34940 }, { "epoch": 20.343422584400464, "grad_norm": 0.7311017513275146, "learning_rate": 4.055244412634849e-05, "loss": 0.3344, "step": 34950 }, { "epoch": 20.349243306169964, "grad_norm": 0.8232046961784363, "learning_rate": 4.052538414551298e-05, "loss": 0.3416, "step": 34960 }, { "epoch": 20.355064027939463, "grad_norm": 0.6753091812133789, "learning_rate": 4.0498327042815596e-05, "loss": 0.346, "step": 34970 }, { "epoch": 20.360884749708966, "grad_norm": 0.7714649438858032, "learning_rate": 4.047127282647559e-05, "loss": 0.3465, "step": 34980 }, { "epoch": 20.366705471478465, "grad_norm": 0.6234942078590393, "learning_rate": 4.04442215047113e-05, "loss": 0.3175, "step": 34990 }, { "epoch": 20.372526193247964, "grad_norm": 0.7682901620864868, "learning_rate": 4.041717308574023e-05, "loss": 0.3492, "step": 35000 }, { "epoch": 20.378346915017463, "grad_norm": 0.8001797795295715, "learning_rate": 4.039012757777893e-05, "loss": 0.333, "step": 35010 }, { "epoch": 20.384167636786962, "grad_norm": 0.7527280449867249, "learning_rate": 4.036308498904314e-05, "loss": 0.3346, "step": 35020 }, { "epoch": 20.38998835855646, "grad_norm": 0.7917881011962891, "learning_rate": 4.033604532774771e-05, "loss": 0.3438, "step": 35030 }, { "epoch": 20.39580908032596, "grad_norm": 0.7669606804847717, "learning_rate": 4.030900860210652e-05, "loss": 0.3478, "step": 35040 }, { "epoch": 20.40162980209546, "grad_norm": 0.777649462223053, "learning_rate": 4.028197482033266e-05, "loss": 0.3687, "step": 35050 }, { "epoch": 20.40745052386496, "grad_norm": 0.8140813708305359, "learning_rate": 4.0254943990638246e-05, "loss": 0.3414, "step": 35060 }, { "epoch": 20.41327124563446, "grad_norm": 0.7021397352218628, "learning_rate": 4.022791612123454e-05, "loss": 0.3419, "step": 35070 }, { "epoch": 20.419091967403958, "grad_norm": 0.8021634221076965, "learning_rate": 4.020089122033192e-05, "loss": 0.3437, "step": 35080 }, { "epoch": 20.424912689173457, "grad_norm": 0.7159806489944458, "learning_rate": 4.01738692961398e-05, "loss": 0.337, "step": 35090 }, { "epoch": 20.430733410942956, "grad_norm": 0.7254035472869873, "learning_rate": 4.014685035686675e-05, "loss": 0.34, "step": 35100 }, { "epoch": 20.436554132712455, "grad_norm": 0.6485346555709839, "learning_rate": 4.011983441072039e-05, "loss": 0.342, "step": 35110 }, { "epoch": 20.442374854481955, "grad_norm": 0.752780556678772, "learning_rate": 4.0092821465907485e-05, "loss": 0.3549, "step": 35120 }, { "epoch": 20.448195576251454, "grad_norm": 0.68431556224823, "learning_rate": 4.006581153063383e-05, "loss": 0.3451, "step": 35130 }, { "epoch": 20.454016298020953, "grad_norm": 0.7042539715766907, "learning_rate": 4.003880461310432e-05, "loss": 0.3382, "step": 35140 }, { "epoch": 20.459837019790456, "grad_norm": 0.7448498606681824, "learning_rate": 4.001180072152298e-05, "loss": 0.3488, "step": 35150 }, { "epoch": 20.465657741559955, "grad_norm": 0.679750919342041, "learning_rate": 3.998479986409285e-05, "loss": 0.3434, "step": 35160 }, { "epoch": 20.471478463329454, "grad_norm": 0.852586030960083, "learning_rate": 3.995780204901607e-05, "loss": 0.3596, "step": 35170 }, { "epoch": 20.477299185098953, "grad_norm": 0.8126032948493958, "learning_rate": 3.993080728449391e-05, "loss": 0.3375, "step": 35180 }, { "epoch": 20.483119906868453, "grad_norm": 0.746708333492279, "learning_rate": 3.990381557872661e-05, "loss": 0.3423, "step": 35190 }, { "epoch": 20.488940628637952, "grad_norm": 0.7892195582389832, "learning_rate": 3.987682693991359e-05, "loss": 0.3555, "step": 35200 }, { "epoch": 20.49476135040745, "grad_norm": 0.7402273416519165, "learning_rate": 3.9849841376253226e-05, "loss": 0.3427, "step": 35210 }, { "epoch": 20.50058207217695, "grad_norm": 0.7187559008598328, "learning_rate": 3.982285889594306e-05, "loss": 0.3428, "step": 35220 }, { "epoch": 20.50640279394645, "grad_norm": 0.6295853853225708, "learning_rate": 3.9795879507179665e-05, "loss": 0.3256, "step": 35230 }, { "epoch": 20.51222351571595, "grad_norm": 0.6790810227394104, "learning_rate": 3.9768903218158634e-05, "loss": 0.3443, "step": 35240 }, { "epoch": 20.518044237485448, "grad_norm": 0.7991594672203064, "learning_rate": 3.974193003707468e-05, "loss": 0.3449, "step": 35250 }, { "epoch": 20.523864959254947, "grad_norm": 0.72548508644104, "learning_rate": 3.971495997212152e-05, "loss": 0.3503, "step": 35260 }, { "epoch": 20.529685681024446, "grad_norm": 0.7057595252990723, "learning_rate": 3.9687993031491985e-05, "loss": 0.3364, "step": 35270 }, { "epoch": 20.535506402793946, "grad_norm": 0.8201078772544861, "learning_rate": 3.966102922337787e-05, "loss": 0.3436, "step": 35280 }, { "epoch": 20.541327124563445, "grad_norm": 0.7237703800201416, "learning_rate": 3.963406855597009e-05, "loss": 0.3445, "step": 35290 }, { "epoch": 20.547147846332944, "grad_norm": 0.7639037370681763, "learning_rate": 3.960711103745861e-05, "loss": 0.3469, "step": 35300 }, { "epoch": 20.552968568102443, "grad_norm": 0.9144278168678284, "learning_rate": 3.958015667603237e-05, "loss": 0.3378, "step": 35310 }, { "epoch": 20.558789289871942, "grad_norm": 0.7301858067512512, "learning_rate": 3.955320547987943e-05, "loss": 0.3389, "step": 35320 }, { "epoch": 20.564610011641445, "grad_norm": 0.717623770236969, "learning_rate": 3.952625745718681e-05, "loss": 0.32, "step": 35330 }, { "epoch": 20.570430733410944, "grad_norm": 0.6454512476921082, "learning_rate": 3.949931261614064e-05, "loss": 0.3402, "step": 35340 }, { "epoch": 20.576251455180444, "grad_norm": 0.7154173851013184, "learning_rate": 3.947237096492605e-05, "loss": 0.3322, "step": 35350 }, { "epoch": 20.582072176949943, "grad_norm": 0.7641522288322449, "learning_rate": 3.944543251172719e-05, "loss": 0.3418, "step": 35360 }, { "epoch": 20.587892898719442, "grad_norm": 0.7956507205963135, "learning_rate": 3.941849726472725e-05, "loss": 0.3422, "step": 35370 }, { "epoch": 20.59371362048894, "grad_norm": 0.6622102856636047, "learning_rate": 3.939156523210846e-05, "loss": 0.3389, "step": 35380 }, { "epoch": 20.59953434225844, "grad_norm": 0.598272979259491, "learning_rate": 3.9364636422052046e-05, "loss": 0.3346, "step": 35390 }, { "epoch": 20.60535506402794, "grad_norm": 0.756969153881073, "learning_rate": 3.933771084273828e-05, "loss": 0.3374, "step": 35400 }, { "epoch": 20.61117578579744, "grad_norm": 0.7101261615753174, "learning_rate": 3.931078850234643e-05, "loss": 0.329, "step": 35410 }, { "epoch": 20.616996507566938, "grad_norm": 0.7236905694007874, "learning_rate": 3.928386940905483e-05, "loss": 0.3241, "step": 35420 }, { "epoch": 20.622817229336437, "grad_norm": 0.6530445218086243, "learning_rate": 3.925695357104073e-05, "loss": 0.3296, "step": 35430 }, { "epoch": 20.628637951105937, "grad_norm": 0.703557014465332, "learning_rate": 3.923004099648049e-05, "loss": 0.3367, "step": 35440 }, { "epoch": 20.634458672875436, "grad_norm": 0.7629899978637695, "learning_rate": 3.920313169354944e-05, "loss": 0.344, "step": 35450 }, { "epoch": 20.640279394644935, "grad_norm": 0.8275253772735596, "learning_rate": 3.9176225670421897e-05, "loss": 0.333, "step": 35460 }, { "epoch": 20.646100116414434, "grad_norm": 0.6992354393005371, "learning_rate": 3.9149322935271224e-05, "loss": 0.3427, "step": 35470 }, { "epoch": 20.651920838183933, "grad_norm": 0.7306963801383972, "learning_rate": 3.9122423496269725e-05, "loss": 0.3312, "step": 35480 }, { "epoch": 20.657741559953433, "grad_norm": 0.6485311388969421, "learning_rate": 3.909552736158877e-05, "loss": 0.3312, "step": 35490 }, { "epoch": 20.663562281722932, "grad_norm": 0.6495388150215149, "learning_rate": 3.90686345393987e-05, "loss": 0.3415, "step": 35500 }, { "epoch": 20.669383003492435, "grad_norm": 0.6983779072761536, "learning_rate": 3.9041745037868816e-05, "loss": 0.3282, "step": 35510 }, { "epoch": 20.675203725261934, "grad_norm": 0.6745324730873108, "learning_rate": 3.9014858865167465e-05, "loss": 0.3219, "step": 35520 }, { "epoch": 20.681024447031433, "grad_norm": 0.7530921697616577, "learning_rate": 3.8987976029461935e-05, "loss": 0.3365, "step": 35530 }, { "epoch": 20.686845168800932, "grad_norm": 0.6747531294822693, "learning_rate": 3.896109653891853e-05, "loss": 0.3412, "step": 35540 }, { "epoch": 20.69266589057043, "grad_norm": 0.7053173780441284, "learning_rate": 3.893422040170254e-05, "loss": 0.3435, "step": 35550 }, { "epoch": 20.69848661233993, "grad_norm": 0.6654060482978821, "learning_rate": 3.8907347625978207e-05, "loss": 0.3335, "step": 35560 }, { "epoch": 20.70430733410943, "grad_norm": 0.7690516710281372, "learning_rate": 3.88804782199088e-05, "loss": 0.3363, "step": 35570 }, { "epoch": 20.71012805587893, "grad_norm": 0.7181965708732605, "learning_rate": 3.8853612191656495e-05, "loss": 0.3409, "step": 35580 }, { "epoch": 20.71594877764843, "grad_norm": 0.6634014844894409, "learning_rate": 3.88267495493825e-05, "loss": 0.3262, "step": 35590 }, { "epoch": 20.721769499417928, "grad_norm": 0.6904743909835815, "learning_rate": 3.8799890301247004e-05, "loss": 0.3396, "step": 35600 }, { "epoch": 20.727590221187427, "grad_norm": 0.6765801906585693, "learning_rate": 3.8773034455409096e-05, "loss": 0.3305, "step": 35610 }, { "epoch": 20.733410942956926, "grad_norm": 0.6467931866645813, "learning_rate": 3.8746182020026904e-05, "loss": 0.3348, "step": 35620 }, { "epoch": 20.739231664726425, "grad_norm": 0.6828676462173462, "learning_rate": 3.871933300325745e-05, "loss": 0.3407, "step": 35630 }, { "epoch": 20.745052386495924, "grad_norm": 0.7830579280853271, "learning_rate": 3.869248741325679e-05, "loss": 0.3485, "step": 35640 }, { "epoch": 20.750873108265424, "grad_norm": 0.8389113545417786, "learning_rate": 3.866564525817992e-05, "loss": 0.3503, "step": 35650 }, { "epoch": 20.756693830034923, "grad_norm": 0.6869242787361145, "learning_rate": 3.8638806546180725e-05, "loss": 0.3395, "step": 35660 }, { "epoch": 20.762514551804422, "grad_norm": 0.6885790228843689, "learning_rate": 3.861197128541213e-05, "loss": 0.339, "step": 35670 }, { "epoch": 20.768335273573925, "grad_norm": 0.7009900808334351, "learning_rate": 3.858513948402599e-05, "loss": 0.3432, "step": 35680 }, { "epoch": 20.774155995343424, "grad_norm": 0.7759427428245544, "learning_rate": 3.8558311150173077e-05, "loss": 0.3391, "step": 35690 }, { "epoch": 20.779976717112923, "grad_norm": 0.720737874507904, "learning_rate": 3.853148629200312e-05, "loss": 0.3428, "step": 35700 }, { "epoch": 20.785797438882422, "grad_norm": 0.7062591910362244, "learning_rate": 3.850466491766482e-05, "loss": 0.3335, "step": 35710 }, { "epoch": 20.79161816065192, "grad_norm": 0.7146226167678833, "learning_rate": 3.847784703530583e-05, "loss": 0.3337, "step": 35720 }, { "epoch": 20.79743888242142, "grad_norm": 0.7646546363830566, "learning_rate": 3.845103265307266e-05, "loss": 0.3539, "step": 35730 }, { "epoch": 20.80325960419092, "grad_norm": 0.7527734041213989, "learning_rate": 3.842422177911086e-05, "loss": 0.3328, "step": 35740 }, { "epoch": 20.80908032596042, "grad_norm": 0.6917887926101685, "learning_rate": 3.8397414421564826e-05, "loss": 0.3239, "step": 35750 }, { "epoch": 20.81490104772992, "grad_norm": 0.7096051573753357, "learning_rate": 3.8370610588577935e-05, "loss": 0.3459, "step": 35760 }, { "epoch": 20.820721769499418, "grad_norm": 0.6435834765434265, "learning_rate": 3.834381028829251e-05, "loss": 0.3311, "step": 35770 }, { "epoch": 20.826542491268917, "grad_norm": 0.7264404296875, "learning_rate": 3.8317013528849745e-05, "loss": 0.3316, "step": 35780 }, { "epoch": 20.832363213038416, "grad_norm": 0.7097316980361938, "learning_rate": 3.8290220318389815e-05, "loss": 0.3279, "step": 35790 }, { "epoch": 20.838183934807915, "grad_norm": 0.7773378491401672, "learning_rate": 3.8263430665051746e-05, "loss": 0.3428, "step": 35800 }, { "epoch": 20.844004656577415, "grad_norm": 0.6812870502471924, "learning_rate": 3.8236644576973554e-05, "loss": 0.3368, "step": 35810 }, { "epoch": 20.849825378346914, "grad_norm": 0.6904159784317017, "learning_rate": 3.820986206229217e-05, "loss": 0.3357, "step": 35820 }, { "epoch": 20.855646100116413, "grad_norm": 0.6962639093399048, "learning_rate": 3.8183083129143384e-05, "loss": 0.3521, "step": 35830 }, { "epoch": 20.861466821885912, "grad_norm": 0.9988173246383667, "learning_rate": 3.815630778566193e-05, "loss": 0.3388, "step": 35840 }, { "epoch": 20.867287543655415, "grad_norm": 0.8270527720451355, "learning_rate": 3.812953603998145e-05, "loss": 0.3407, "step": 35850 }, { "epoch": 20.873108265424914, "grad_norm": 0.6863256692886353, "learning_rate": 3.8102767900234504e-05, "loss": 0.3308, "step": 35860 }, { "epoch": 20.878928987194413, "grad_norm": 0.7481356859207153, "learning_rate": 3.807600337455256e-05, "loss": 0.3373, "step": 35870 }, { "epoch": 20.884749708963913, "grad_norm": 0.7086185812950134, "learning_rate": 3.804924247106593e-05, "loss": 0.3333, "step": 35880 }, { "epoch": 20.890570430733412, "grad_norm": 0.6384220719337463, "learning_rate": 3.8022485197903925e-05, "loss": 0.3353, "step": 35890 }, { "epoch": 20.89639115250291, "grad_norm": 0.716096818447113, "learning_rate": 3.799573156319464e-05, "loss": 0.3451, "step": 35900 }, { "epoch": 20.90221187427241, "grad_norm": 0.7341987490653992, "learning_rate": 3.796898157506515e-05, "loss": 0.3273, "step": 35910 }, { "epoch": 20.90803259604191, "grad_norm": 0.657914936542511, "learning_rate": 3.794223524164143e-05, "loss": 0.3357, "step": 35920 }, { "epoch": 20.91385331781141, "grad_norm": 0.8117069005966187, "learning_rate": 3.7915492571048245e-05, "loss": 0.3432, "step": 35930 }, { "epoch": 20.919674039580908, "grad_norm": 0.6732069253921509, "learning_rate": 3.788875357140937e-05, "loss": 0.3408, "step": 35940 }, { "epoch": 20.925494761350407, "grad_norm": 0.7065292000770569, "learning_rate": 3.786201825084736e-05, "loss": 0.3373, "step": 35950 }, { "epoch": 20.931315483119906, "grad_norm": 0.6717864274978638, "learning_rate": 3.783528661748372e-05, "loss": 0.3277, "step": 35960 }, { "epoch": 20.937136204889406, "grad_norm": 0.7153463363647461, "learning_rate": 3.780855867943882e-05, "loss": 0.3388, "step": 35970 }, { "epoch": 20.942956926658905, "grad_norm": 0.7097715735435486, "learning_rate": 3.778183444483189e-05, "loss": 0.3413, "step": 35980 }, { "epoch": 20.948777648428404, "grad_norm": 0.7235670685768127, "learning_rate": 3.775511392178108e-05, "loss": 0.3329, "step": 35990 }, { "epoch": 20.954598370197903, "grad_norm": 0.7038096785545349, "learning_rate": 3.772839711840332e-05, "loss": 0.3409, "step": 36000 }, { "epoch": 20.960419091967402, "grad_norm": 0.6894518136978149, "learning_rate": 3.7701684042814515e-05, "loss": 0.3335, "step": 36010 }, { "epoch": 20.9662398137369, "grad_norm": 0.7342696189880371, "learning_rate": 3.76749747031294e-05, "loss": 0.3382, "step": 36020 }, { "epoch": 20.972060535506404, "grad_norm": 0.7034570574760437, "learning_rate": 3.764826910746152e-05, "loss": 0.3353, "step": 36030 }, { "epoch": 20.977881257275904, "grad_norm": 0.7351905107498169, "learning_rate": 3.762156726392338e-05, "loss": 0.3377, "step": 36040 }, { "epoch": 20.983701979045403, "grad_norm": 0.7013428807258606, "learning_rate": 3.759486918062625e-05, "loss": 0.3399, "step": 36050 }, { "epoch": 20.989522700814902, "grad_norm": 0.830778956413269, "learning_rate": 3.756817486568033e-05, "loss": 0.3374, "step": 36060 }, { "epoch": 20.9953434225844, "grad_norm": 0.717715859413147, "learning_rate": 3.7541484327194654e-05, "loss": 0.3319, "step": 36070 }, { "epoch": 21.0011641443539, "grad_norm": 0.6729212999343872, "learning_rate": 3.751479757327707e-05, "loss": 0.3285, "step": 36080 }, { "epoch": 21.0069848661234, "grad_norm": 0.7435158491134644, "learning_rate": 3.7488114612034345e-05, "loss": 0.3539, "step": 36090 }, { "epoch": 21.0128055878929, "grad_norm": 0.6963726282119751, "learning_rate": 3.7461435451572044e-05, "loss": 0.3289, "step": 36100 }, { "epoch": 21.018626309662398, "grad_norm": 0.6764479875564575, "learning_rate": 3.743476009999459e-05, "loss": 0.3475, "step": 36110 }, { "epoch": 21.024447031431897, "grad_norm": 0.6353332996368408, "learning_rate": 3.7408088565405245e-05, "loss": 0.3311, "step": 36120 }, { "epoch": 21.030267753201397, "grad_norm": 0.7238125205039978, "learning_rate": 3.738142085590612e-05, "loss": 0.3455, "step": 36130 }, { "epoch": 21.036088474970896, "grad_norm": 0.6817229986190796, "learning_rate": 3.7354756979598194e-05, "loss": 0.3377, "step": 36140 }, { "epoch": 21.041909196740395, "grad_norm": 0.7676522135734558, "learning_rate": 3.7328096944581187e-05, "loss": 0.3413, "step": 36150 }, { "epoch": 21.047729918509894, "grad_norm": 0.702826738357544, "learning_rate": 3.730144075895377e-05, "loss": 0.339, "step": 36160 }, { "epoch": 21.053550640279393, "grad_norm": 0.6991465091705322, "learning_rate": 3.727478843081335e-05, "loss": 0.3222, "step": 36170 }, { "epoch": 21.059371362048893, "grad_norm": 0.7830649018287659, "learning_rate": 3.72481399682562e-05, "loss": 0.3526, "step": 36180 }, { "epoch": 21.065192083818392, "grad_norm": 0.7136797904968262, "learning_rate": 3.722149537937747e-05, "loss": 0.3317, "step": 36190 }, { "epoch": 21.07101280558789, "grad_norm": 0.7212358117103577, "learning_rate": 3.7194854672271015e-05, "loss": 0.3263, "step": 36200 }, { "epoch": 21.076833527357394, "grad_norm": 0.6745208501815796, "learning_rate": 3.7168217855029644e-05, "loss": 0.3325, "step": 36210 }, { "epoch": 21.082654249126893, "grad_norm": 0.7285847067832947, "learning_rate": 3.7141584935744856e-05, "loss": 0.3219, "step": 36220 }, { "epoch": 21.088474970896392, "grad_norm": 0.7491427659988403, "learning_rate": 3.7114955922507055e-05, "loss": 0.3274, "step": 36230 }, { "epoch": 21.09429569266589, "grad_norm": 0.6166269183158875, "learning_rate": 3.708833082340545e-05, "loss": 0.3328, "step": 36240 }, { "epoch": 21.10011641443539, "grad_norm": 0.6450756192207336, "learning_rate": 3.7061709646528034e-05, "loss": 0.3237, "step": 36250 }, { "epoch": 21.10593713620489, "grad_norm": 0.8564639687538147, "learning_rate": 3.7035092399961604e-05, "loss": 0.3557, "step": 36260 }, { "epoch": 21.11175785797439, "grad_norm": 0.8188032507896423, "learning_rate": 3.700847909179177e-05, "loss": 0.3533, "step": 36270 }, { "epoch": 21.11757857974389, "grad_norm": 0.7216567397117615, "learning_rate": 3.698186973010297e-05, "loss": 0.3587, "step": 36280 }, { "epoch": 21.123399301513388, "grad_norm": 0.7605169415473938, "learning_rate": 3.695526432297844e-05, "loss": 0.3331, "step": 36290 }, { "epoch": 21.129220023282887, "grad_norm": 0.7688214182853699, "learning_rate": 3.692866287850017e-05, "loss": 0.3247, "step": 36300 }, { "epoch": 21.135040745052386, "grad_norm": 0.7863556742668152, "learning_rate": 3.6902065404749006e-05, "loss": 0.3305, "step": 36310 }, { "epoch": 21.140861466821885, "grad_norm": 0.7454606890678406, "learning_rate": 3.6875471909804516e-05, "loss": 0.3369, "step": 36320 }, { "epoch": 21.146682188591384, "grad_norm": 0.6164785027503967, "learning_rate": 3.6848882401745135e-05, "loss": 0.3487, "step": 36330 }, { "epoch": 21.152502910360884, "grad_norm": 0.6725181937217712, "learning_rate": 3.682229688864806e-05, "loss": 0.3403, "step": 36340 }, { "epoch": 21.158323632130383, "grad_norm": 0.6411550045013428, "learning_rate": 3.6795715378589235e-05, "loss": 0.3443, "step": 36350 }, { "epoch": 21.164144353899882, "grad_norm": 0.7443664073944092, "learning_rate": 3.676913787964345e-05, "loss": 0.3403, "step": 36360 }, { "epoch": 21.16996507566938, "grad_norm": 0.6662797331809998, "learning_rate": 3.674256439988423e-05, "loss": 0.3361, "step": 36370 }, { "epoch": 21.175785797438884, "grad_norm": 0.748651921749115, "learning_rate": 3.6715994947383904e-05, "loss": 0.3415, "step": 36380 }, { "epoch": 21.181606519208383, "grad_norm": 0.7558821439743042, "learning_rate": 3.668942953021357e-05, "loss": 0.3386, "step": 36390 }, { "epoch": 21.187427240977883, "grad_norm": 0.6591335535049438, "learning_rate": 3.66628681564431e-05, "loss": 0.3313, "step": 36400 }, { "epoch": 21.19324796274738, "grad_norm": 0.7725299596786499, "learning_rate": 3.663631083414114e-05, "loss": 0.334, "step": 36410 }, { "epoch": 21.19906868451688, "grad_norm": 0.6630872488021851, "learning_rate": 3.660975757137509e-05, "loss": 0.3305, "step": 36420 }, { "epoch": 21.20488940628638, "grad_norm": 0.789484977722168, "learning_rate": 3.658320837621114e-05, "loss": 0.359, "step": 36430 }, { "epoch": 21.21071012805588, "grad_norm": 0.8712199330329895, "learning_rate": 3.655666325671426e-05, "loss": 0.3317, "step": 36440 }, { "epoch": 21.21653084982538, "grad_norm": 0.7518192529678345, "learning_rate": 3.65301222209481e-05, "loss": 0.339, "step": 36450 }, { "epoch": 21.222351571594878, "grad_norm": 0.7395167350769043, "learning_rate": 3.650358527697519e-05, "loss": 0.3312, "step": 36460 }, { "epoch": 21.228172293364377, "grad_norm": 0.6968680024147034, "learning_rate": 3.64770524328567e-05, "loss": 0.3443, "step": 36470 }, { "epoch": 21.233993015133876, "grad_norm": 0.6698436141014099, "learning_rate": 3.645052369665265e-05, "loss": 0.3319, "step": 36480 }, { "epoch": 21.239813736903375, "grad_norm": 0.7237467169761658, "learning_rate": 3.6423999076421724e-05, "loss": 0.3489, "step": 36490 }, { "epoch": 21.245634458672875, "grad_norm": 0.7488775253295898, "learning_rate": 3.639747858022142e-05, "loss": 0.3327, "step": 36500 }, { "epoch": 21.251455180442374, "grad_norm": 0.6895870566368103, "learning_rate": 3.637096221610799e-05, "loss": 0.3505, "step": 36510 }, { "epoch": 21.257275902211873, "grad_norm": 0.7012002468109131, "learning_rate": 3.634444999213638e-05, "loss": 0.3293, "step": 36520 }, { "epoch": 21.263096623981372, "grad_norm": 0.6757987141609192, "learning_rate": 3.6317941916360296e-05, "loss": 0.3456, "step": 36530 }, { "epoch": 21.26891734575087, "grad_norm": 0.6966349482536316, "learning_rate": 3.629143799683221e-05, "loss": 0.346, "step": 36540 }, { "epoch": 21.274738067520374, "grad_norm": 0.781493604183197, "learning_rate": 3.626493824160331e-05, "loss": 0.3417, "step": 36550 }, { "epoch": 21.280558789289874, "grad_norm": 0.7341422438621521, "learning_rate": 3.623844265872352e-05, "loss": 0.33, "step": 36560 }, { "epoch": 21.286379511059373, "grad_norm": 0.7333785891532898, "learning_rate": 3.621195125624149e-05, "loss": 0.3375, "step": 36570 }, { "epoch": 21.292200232828872, "grad_norm": 0.6998869180679321, "learning_rate": 3.618546404220463e-05, "loss": 0.3379, "step": 36580 }, { "epoch": 21.29802095459837, "grad_norm": 0.6934114694595337, "learning_rate": 3.615898102465903e-05, "loss": 0.3484, "step": 36590 }, { "epoch": 21.30384167636787, "grad_norm": 0.706316351890564, "learning_rate": 3.6132502211649544e-05, "loss": 0.3226, "step": 36600 }, { "epoch": 21.30966239813737, "grad_norm": 0.7311387062072754, "learning_rate": 3.610602761121975e-05, "loss": 0.3359, "step": 36610 }, { "epoch": 21.31548311990687, "grad_norm": 0.8338699340820312, "learning_rate": 3.6079557231411897e-05, "loss": 0.3507, "step": 36620 }, { "epoch": 21.321303841676368, "grad_norm": 0.6485093235969543, "learning_rate": 3.6053091080267035e-05, "loss": 0.3237, "step": 36630 }, { "epoch": 21.327124563445867, "grad_norm": 0.6964545249938965, "learning_rate": 3.602662916582483e-05, "loss": 0.3347, "step": 36640 }, { "epoch": 21.332945285215366, "grad_norm": 0.655020534992218, "learning_rate": 3.600017149612375e-05, "loss": 0.331, "step": 36650 }, { "epoch": 21.338766006984866, "grad_norm": 0.7291299700737, "learning_rate": 3.5973718079200935e-05, "loss": 0.3512, "step": 36660 }, { "epoch": 21.344586728754365, "grad_norm": 0.7029775381088257, "learning_rate": 3.5947268923092216e-05, "loss": 0.3317, "step": 36670 }, { "epoch": 21.350407450523864, "grad_norm": 0.7071051001548767, "learning_rate": 3.592082403583216e-05, "loss": 0.3436, "step": 36680 }, { "epoch": 21.356228172293363, "grad_norm": 0.716827929019928, "learning_rate": 3.5894383425454004e-05, "loss": 0.3382, "step": 36690 }, { "epoch": 21.362048894062863, "grad_norm": 0.7631881833076477, "learning_rate": 3.586794709998975e-05, "loss": 0.3297, "step": 36700 }, { "epoch": 21.36786961583236, "grad_norm": 0.7699057459831238, "learning_rate": 3.584151506747002e-05, "loss": 0.3472, "step": 36710 }, { "epoch": 21.37369033760186, "grad_norm": 0.6941090822219849, "learning_rate": 3.581508733592418e-05, "loss": 0.3199, "step": 36720 }, { "epoch": 21.379511059371364, "grad_norm": 0.696228563785553, "learning_rate": 3.5788663913380297e-05, "loss": 0.3386, "step": 36730 }, { "epoch": 21.385331781140863, "grad_norm": 0.7074242234230042, "learning_rate": 3.576224480786506e-05, "loss": 0.3414, "step": 36740 }, { "epoch": 21.391152502910362, "grad_norm": 0.6888415813446045, "learning_rate": 3.573583002740393e-05, "loss": 0.3517, "step": 36750 }, { "epoch": 21.39697322467986, "grad_norm": 0.7234404683113098, "learning_rate": 3.570941958002103e-05, "loss": 0.3388, "step": 36760 }, { "epoch": 21.40279394644936, "grad_norm": 0.6451737880706787, "learning_rate": 3.568301347373912e-05, "loss": 0.3273, "step": 36770 }, { "epoch": 21.40861466821886, "grad_norm": 0.7831461429595947, "learning_rate": 3.5656611716579726e-05, "loss": 0.3392, "step": 36780 }, { "epoch": 21.41443538998836, "grad_norm": 0.739866316318512, "learning_rate": 3.5630214316562946e-05, "loss": 0.3381, "step": 36790 }, { "epoch": 21.42025611175786, "grad_norm": 0.7374833822250366, "learning_rate": 3.560382128170766e-05, "loss": 0.3382, "step": 36800 }, { "epoch": 21.426076833527357, "grad_norm": 0.6527644395828247, "learning_rate": 3.5577432620031374e-05, "loss": 0.3393, "step": 36810 }, { "epoch": 21.431897555296857, "grad_norm": 0.7740932703018188, "learning_rate": 3.5551048339550216e-05, "loss": 0.3413, "step": 36820 }, { "epoch": 21.437718277066356, "grad_norm": 0.7012261152267456, "learning_rate": 3.55246684482791e-05, "loss": 0.3289, "step": 36830 }, { "epoch": 21.443538998835855, "grad_norm": 0.7964479923248291, "learning_rate": 3.5498292954231496e-05, "loss": 0.3432, "step": 36840 }, { "epoch": 21.449359720605354, "grad_norm": 0.8228639364242554, "learning_rate": 3.54719218654196e-05, "loss": 0.3421, "step": 36850 }, { "epoch": 21.455180442374854, "grad_norm": 0.8354920744895935, "learning_rate": 3.544555518985425e-05, "loss": 0.3353, "step": 36860 }, { "epoch": 21.461001164144353, "grad_norm": 0.7703284025192261, "learning_rate": 3.541919293554494e-05, "loss": 0.3265, "step": 36870 }, { "epoch": 21.466821885913852, "grad_norm": 0.6666184663772583, "learning_rate": 3.539283511049985e-05, "loss": 0.3289, "step": 36880 }, { "epoch": 21.47264260768335, "grad_norm": 0.7977504730224609, "learning_rate": 3.5366481722725755e-05, "loss": 0.3461, "step": 36890 }, { "epoch": 21.47846332945285, "grad_norm": 0.7281312942504883, "learning_rate": 3.534013278022816e-05, "loss": 0.3327, "step": 36900 }, { "epoch": 21.484284051222353, "grad_norm": 0.7029562592506409, "learning_rate": 3.531378829101113e-05, "loss": 0.3362, "step": 36910 }, { "epoch": 21.490104772991852, "grad_norm": 0.705998420715332, "learning_rate": 3.528744826307746e-05, "loss": 0.3462, "step": 36920 }, { "epoch": 21.49592549476135, "grad_norm": 0.7375112771987915, "learning_rate": 3.5261112704428554e-05, "loss": 0.3362, "step": 36930 }, { "epoch": 21.50174621653085, "grad_norm": 0.6063780188560486, "learning_rate": 3.523478162306443e-05, "loss": 0.3243, "step": 36940 }, { "epoch": 21.50756693830035, "grad_norm": 0.6493676900863647, "learning_rate": 3.520845502698381e-05, "loss": 0.3379, "step": 36950 }, { "epoch": 21.51338766006985, "grad_norm": 0.6468377709388733, "learning_rate": 3.5182132924184005e-05, "loss": 0.3315, "step": 36960 }, { "epoch": 21.51920838183935, "grad_norm": 0.8080008029937744, "learning_rate": 3.5155815322660966e-05, "loss": 0.3321, "step": 36970 }, { "epoch": 21.525029103608848, "grad_norm": 0.7664111852645874, "learning_rate": 3.512950223040931e-05, "loss": 0.3285, "step": 36980 }, { "epoch": 21.530849825378347, "grad_norm": 0.7571009397506714, "learning_rate": 3.5103193655422216e-05, "loss": 0.3368, "step": 36990 }, { "epoch": 21.536670547147846, "grad_norm": 0.7228159308433533, "learning_rate": 3.5076889605691596e-05, "loss": 0.3407, "step": 37000 }, { "epoch": 21.542491268917345, "grad_norm": 0.7355550527572632, "learning_rate": 3.505059008920787e-05, "loss": 0.3449, "step": 37010 }, { "epoch": 21.548311990686845, "grad_norm": 0.6424346566200256, "learning_rate": 3.502429511396016e-05, "loss": 0.3413, "step": 37020 }, { "epoch": 21.554132712456344, "grad_norm": 0.6502785682678223, "learning_rate": 3.4998004687936196e-05, "loss": 0.3506, "step": 37030 }, { "epoch": 21.559953434225843, "grad_norm": 0.6797012090682983, "learning_rate": 3.497171881912229e-05, "loss": 0.3367, "step": 37040 }, { "epoch": 21.565774155995342, "grad_norm": 0.6663801074028015, "learning_rate": 3.494543751550342e-05, "loss": 0.3306, "step": 37050 }, { "epoch": 21.57159487776484, "grad_norm": 0.7086769342422485, "learning_rate": 3.491916078506313e-05, "loss": 0.3327, "step": 37060 }, { "epoch": 21.57741559953434, "grad_norm": 0.9260947108268738, "learning_rate": 3.489288863578361e-05, "loss": 0.3409, "step": 37070 }, { "epoch": 21.583236321303843, "grad_norm": 0.7362348437309265, "learning_rate": 3.4866621075645646e-05, "loss": 0.34, "step": 37080 }, { "epoch": 21.589057043073343, "grad_norm": 0.6804860234260559, "learning_rate": 3.4840358112628614e-05, "loss": 0.3383, "step": 37090 }, { "epoch": 21.594877764842842, "grad_norm": 0.6987407207489014, "learning_rate": 3.481409975471053e-05, "loss": 0.331, "step": 37100 }, { "epoch": 21.60069848661234, "grad_norm": 0.7223803400993347, "learning_rate": 3.4787846009867986e-05, "loss": 0.35, "step": 37110 }, { "epoch": 21.60651920838184, "grad_norm": 0.7456451058387756, "learning_rate": 3.476159688607615e-05, "loss": 0.3403, "step": 37120 }, { "epoch": 21.61233993015134, "grad_norm": 0.6306881308555603, "learning_rate": 3.4735352391308854e-05, "loss": 0.3337, "step": 37130 }, { "epoch": 21.61816065192084, "grad_norm": 0.6561478972434998, "learning_rate": 3.4709112533538446e-05, "loss": 0.3542, "step": 37140 }, { "epoch": 21.623981373690338, "grad_norm": 0.6695461869239807, "learning_rate": 3.4682877320735934e-05, "loss": 0.3376, "step": 37150 }, { "epoch": 21.629802095459837, "grad_norm": 0.637485682964325, "learning_rate": 3.465664676087085e-05, "loss": 0.3259, "step": 37160 }, { "epoch": 21.635622817229336, "grad_norm": 0.6366626024246216, "learning_rate": 3.463042086191136e-05, "loss": 0.3374, "step": 37170 }, { "epoch": 21.641443538998836, "grad_norm": 0.7192228436470032, "learning_rate": 3.460419963182423e-05, "loss": 0.3332, "step": 37180 }, { "epoch": 21.647264260768335, "grad_norm": 0.7737410664558411, "learning_rate": 3.457798307857473e-05, "loss": 0.3326, "step": 37190 }, { "epoch": 21.653084982537834, "grad_norm": 0.7489336133003235, "learning_rate": 3.455177121012678e-05, "loss": 0.3318, "step": 37200 }, { "epoch": 21.658905704307333, "grad_norm": 0.6922765970230103, "learning_rate": 3.452556403444285e-05, "loss": 0.3447, "step": 37210 }, { "epoch": 21.664726426076832, "grad_norm": 0.6859753727912903, "learning_rate": 3.4499361559483975e-05, "loss": 0.3384, "step": 37220 }, { "epoch": 21.67054714784633, "grad_norm": 0.7989771962165833, "learning_rate": 3.44731637932098e-05, "loss": 0.324, "step": 37230 }, { "epoch": 21.67636786961583, "grad_norm": 0.66962069272995, "learning_rate": 3.44469707435785e-05, "loss": 0.3273, "step": 37240 }, { "epoch": 21.682188591385334, "grad_norm": 0.6218221187591553, "learning_rate": 3.4420782418546835e-05, "loss": 0.3285, "step": 37250 }, { "epoch": 21.688009313154833, "grad_norm": 0.7458308935165405, "learning_rate": 3.439459882607012e-05, "loss": 0.3349, "step": 37260 }, { "epoch": 21.693830034924332, "grad_norm": 0.6749855875968933, "learning_rate": 3.436841997410225e-05, "loss": 0.3252, "step": 37270 }, { "epoch": 21.69965075669383, "grad_norm": 0.6006194949150085, "learning_rate": 3.434224587059567e-05, "loss": 0.3293, "step": 37280 }, { "epoch": 21.70547147846333, "grad_norm": 0.662146270275116, "learning_rate": 3.431607652350136e-05, "loss": 0.3219, "step": 37290 }, { "epoch": 21.71129220023283, "grad_norm": 0.6484871506690979, "learning_rate": 3.428991194076891e-05, "loss": 0.3327, "step": 37300 }, { "epoch": 21.71711292200233, "grad_norm": 0.7047139406204224, "learning_rate": 3.4263752130346394e-05, "loss": 0.3339, "step": 37310 }, { "epoch": 21.722933643771828, "grad_norm": 0.6524621844291687, "learning_rate": 3.4237597100180515e-05, "loss": 0.3259, "step": 37320 }, { "epoch": 21.728754365541327, "grad_norm": 0.747145414352417, "learning_rate": 3.4211446858216427e-05, "loss": 0.3256, "step": 37330 }, { "epoch": 21.734575087310827, "grad_norm": 0.6529291868209839, "learning_rate": 3.4185301412397915e-05, "loss": 0.3272, "step": 37340 }, { "epoch": 21.740395809080326, "grad_norm": 0.6768674850463867, "learning_rate": 3.415916077066729e-05, "loss": 0.3409, "step": 37350 }, { "epoch": 21.746216530849825, "grad_norm": 0.742368757724762, "learning_rate": 3.413302494096535e-05, "loss": 0.3351, "step": 37360 }, { "epoch": 21.752037252619324, "grad_norm": 0.6805161237716675, "learning_rate": 3.410689393123151e-05, "loss": 0.3322, "step": 37370 }, { "epoch": 21.757857974388823, "grad_norm": 0.7121198773384094, "learning_rate": 3.408076774940364e-05, "loss": 0.3438, "step": 37380 }, { "epoch": 21.763678696158323, "grad_norm": 0.671546459197998, "learning_rate": 3.40546464034182e-05, "loss": 0.3343, "step": 37390 }, { "epoch": 21.769499417927822, "grad_norm": 0.7978819012641907, "learning_rate": 3.4028529901210185e-05, "loss": 0.3461, "step": 37400 }, { "epoch": 21.77532013969732, "grad_norm": 0.7832465767860413, "learning_rate": 3.4002418250713086e-05, "loss": 0.3287, "step": 37410 }, { "epoch": 21.78114086146682, "grad_norm": 0.8794256448745728, "learning_rate": 3.3976311459858936e-05, "loss": 0.3329, "step": 37420 }, { "epoch": 21.78696158323632, "grad_norm": 0.6708558201789856, "learning_rate": 3.395020953657826e-05, "loss": 0.3418, "step": 37430 }, { "epoch": 21.792782305005822, "grad_norm": 0.6610600352287292, "learning_rate": 3.3924112488800165e-05, "loss": 0.3297, "step": 37440 }, { "epoch": 21.79860302677532, "grad_norm": 0.7060136198997498, "learning_rate": 3.389802032445225e-05, "loss": 0.3339, "step": 37450 }, { "epoch": 21.80442374854482, "grad_norm": 0.6384645700454712, "learning_rate": 3.38719330514606e-05, "loss": 0.3302, "step": 37460 }, { "epoch": 21.81024447031432, "grad_norm": 0.797792375087738, "learning_rate": 3.3845850677749866e-05, "loss": 0.3362, "step": 37470 }, { "epoch": 21.81606519208382, "grad_norm": 0.6716716289520264, "learning_rate": 3.3819773211243157e-05, "loss": 0.3364, "step": 37480 }, { "epoch": 21.82188591385332, "grad_norm": 0.6788004040718079, "learning_rate": 3.379370065986213e-05, "loss": 0.3458, "step": 37490 }, { "epoch": 21.827706635622818, "grad_norm": 0.7191587090492249, "learning_rate": 3.3767633031526955e-05, "loss": 0.3325, "step": 37500 }, { "epoch": 21.833527357392317, "grad_norm": 0.6809973120689392, "learning_rate": 3.374157033415626e-05, "loss": 0.3311, "step": 37510 }, { "epoch": 21.839348079161816, "grad_norm": 0.7087807059288025, "learning_rate": 3.371551257566723e-05, "loss": 0.3378, "step": 37520 }, { "epoch": 21.845168800931315, "grad_norm": 0.7498581409454346, "learning_rate": 3.36894597639755e-05, "loss": 0.3407, "step": 37530 }, { "epoch": 21.850989522700814, "grad_norm": 0.6927168965339661, "learning_rate": 3.366341190699523e-05, "loss": 0.3335, "step": 37540 }, { "epoch": 21.856810244470314, "grad_norm": 0.7308482527732849, "learning_rate": 3.36373690126391e-05, "loss": 0.3454, "step": 37550 }, { "epoch": 21.862630966239813, "grad_norm": 0.6413902044296265, "learning_rate": 3.3611331088818234e-05, "loss": 0.3306, "step": 37560 }, { "epoch": 21.868451688009312, "grad_norm": 0.6479231715202332, "learning_rate": 3.3585298143442265e-05, "loss": 0.3379, "step": 37570 }, { "epoch": 21.87427240977881, "grad_norm": 0.6431404948234558, "learning_rate": 3.35592701844193e-05, "loss": 0.3223, "step": 37580 }, { "epoch": 21.88009313154831, "grad_norm": 0.755004346370697, "learning_rate": 3.353324721965596e-05, "loss": 0.3477, "step": 37590 }, { "epoch": 21.88591385331781, "grad_norm": 0.7623006701469421, "learning_rate": 3.350722925705736e-05, "loss": 0.3415, "step": 37600 }, { "epoch": 21.891734575087312, "grad_norm": 0.7011147737503052, "learning_rate": 3.348121630452703e-05, "loss": 0.3452, "step": 37610 }, { "epoch": 21.89755529685681, "grad_norm": 0.6727545857429504, "learning_rate": 3.3455208369967044e-05, "loss": 0.331, "step": 37620 }, { "epoch": 21.90337601862631, "grad_norm": 0.6959089636802673, "learning_rate": 3.34292054612779e-05, "loss": 0.3272, "step": 37630 }, { "epoch": 21.90919674039581, "grad_norm": 0.7594817280769348, "learning_rate": 3.340320758635861e-05, "loss": 0.3339, "step": 37640 }, { "epoch": 21.91501746216531, "grad_norm": 0.6352834701538086, "learning_rate": 3.337721475310666e-05, "loss": 0.3328, "step": 37650 }, { "epoch": 21.92083818393481, "grad_norm": 0.7151809930801392, "learning_rate": 3.335122696941795e-05, "loss": 0.3385, "step": 37660 }, { "epoch": 21.926658905704308, "grad_norm": 0.6596654653549194, "learning_rate": 3.332524424318692e-05, "loss": 0.337, "step": 37670 }, { "epoch": 21.932479627473807, "grad_norm": 0.7745105028152466, "learning_rate": 3.32992665823064e-05, "loss": 0.3372, "step": 37680 }, { "epoch": 21.938300349243306, "grad_norm": 0.7224583625793457, "learning_rate": 3.327329399466774e-05, "loss": 0.3376, "step": 37690 }, { "epoch": 21.944121071012805, "grad_norm": 0.7358686923980713, "learning_rate": 3.324732648816072e-05, "loss": 0.3365, "step": 37700 }, { "epoch": 21.949941792782305, "grad_norm": 0.7624585032463074, "learning_rate": 3.322136407067358e-05, "loss": 0.3313, "step": 37710 }, { "epoch": 21.955762514551804, "grad_norm": 0.6433084011077881, "learning_rate": 3.3195406750093036e-05, "loss": 0.3317, "step": 37720 }, { "epoch": 21.961583236321303, "grad_norm": 0.6247460842132568, "learning_rate": 3.3169454534304205e-05, "loss": 0.3344, "step": 37730 }, { "epoch": 21.967403958090802, "grad_norm": 0.6696667075157166, "learning_rate": 3.3143507431190725e-05, "loss": 0.3297, "step": 37740 }, { "epoch": 21.9732246798603, "grad_norm": 0.6740946769714355, "learning_rate": 3.311756544863459e-05, "loss": 0.3301, "step": 37750 }, { "epoch": 21.9790454016298, "grad_norm": 0.644978940486908, "learning_rate": 3.309162859451633e-05, "loss": 0.3319, "step": 37760 }, { "epoch": 21.9848661233993, "grad_norm": 0.5983217358589172, "learning_rate": 3.306569687671487e-05, "loss": 0.339, "step": 37770 }, { "epoch": 21.990686845168803, "grad_norm": 0.6693963408470154, "learning_rate": 3.303977030310756e-05, "loss": 0.3255, "step": 37780 }, { "epoch": 21.996507566938302, "grad_norm": 0.6554827690124512, "learning_rate": 3.3013848881570245e-05, "loss": 0.3358, "step": 37790 }, { "epoch": 22.0023282887078, "grad_norm": 0.6846328973770142, "learning_rate": 3.298793261997712e-05, "loss": 0.3294, "step": 37800 }, { "epoch": 22.0081490104773, "grad_norm": 0.6713434457778931, "learning_rate": 3.2962021526200893e-05, "loss": 0.3216, "step": 37810 }, { "epoch": 22.0139697322468, "grad_norm": 0.6644225120544434, "learning_rate": 3.293611560811268e-05, "loss": 0.3372, "step": 37820 }, { "epoch": 22.0197904540163, "grad_norm": 0.6653372049331665, "learning_rate": 3.291021487358199e-05, "loss": 0.3359, "step": 37830 }, { "epoch": 22.025611175785798, "grad_norm": 0.8842939734458923, "learning_rate": 3.28843193304768e-05, "loss": 0.3285, "step": 37840 }, { "epoch": 22.031431897555297, "grad_norm": 0.6719329953193665, "learning_rate": 3.2858428986663456e-05, "loss": 0.33, "step": 37850 }, { "epoch": 22.037252619324796, "grad_norm": 0.6134235858917236, "learning_rate": 3.283254385000681e-05, "loss": 0.3326, "step": 37860 }, { "epoch": 22.043073341094296, "grad_norm": 0.7307459115982056, "learning_rate": 3.2806663928370076e-05, "loss": 0.3312, "step": 37870 }, { "epoch": 22.048894062863795, "grad_norm": 0.654157817363739, "learning_rate": 3.278078922961485e-05, "loss": 0.3129, "step": 37880 }, { "epoch": 22.054714784633294, "grad_norm": 0.6841602325439453, "learning_rate": 3.275491976160123e-05, "loss": 0.3232, "step": 37890 }, { "epoch": 22.060535506402793, "grad_norm": 0.7854149341583252, "learning_rate": 3.2729055532187645e-05, "loss": 0.3355, "step": 37900 }, { "epoch": 22.066356228172292, "grad_norm": 0.7315659523010254, "learning_rate": 3.270319654923097e-05, "loss": 0.3293, "step": 37910 }, { "epoch": 22.07217694994179, "grad_norm": 0.7051692008972168, "learning_rate": 3.2677342820586506e-05, "loss": 0.3388, "step": 37920 }, { "epoch": 22.07799767171129, "grad_norm": 0.6883712410926819, "learning_rate": 3.2651494354107905e-05, "loss": 0.3422, "step": 37930 }, { "epoch": 22.08381839348079, "grad_norm": 0.6675841212272644, "learning_rate": 3.2625651157647266e-05, "loss": 0.3328, "step": 37940 }, { "epoch": 22.08963911525029, "grad_norm": 0.7146363258361816, "learning_rate": 3.259981323905505e-05, "loss": 0.3386, "step": 37950 }, { "epoch": 22.095459837019792, "grad_norm": 0.7060887217521667, "learning_rate": 3.257398060618014e-05, "loss": 0.3385, "step": 37960 }, { "epoch": 22.10128055878929, "grad_norm": 0.8119245767593384, "learning_rate": 3.254815326686983e-05, "loss": 0.3317, "step": 37970 }, { "epoch": 22.10710128055879, "grad_norm": 0.7216388583183289, "learning_rate": 3.2522331228969774e-05, "loss": 0.33, "step": 37980 }, { "epoch": 22.11292200232829, "grad_norm": 0.6715294122695923, "learning_rate": 3.2496514500324006e-05, "loss": 0.3279, "step": 37990 }, { "epoch": 22.11874272409779, "grad_norm": 0.7167013883590698, "learning_rate": 3.247070308877498e-05, "loss": 0.3301, "step": 38000 }, { "epoch": 22.124563445867288, "grad_norm": 0.706569254398346, "learning_rate": 3.2444897002163515e-05, "loss": 0.3393, "step": 38010 }, { "epoch": 22.130384167636787, "grad_norm": 0.6007323265075684, "learning_rate": 3.241909624832885e-05, "loss": 0.3246, "step": 38020 }, { "epoch": 22.136204889406287, "grad_norm": 0.6744288802146912, "learning_rate": 3.239330083510852e-05, "loss": 0.3229, "step": 38030 }, { "epoch": 22.142025611175786, "grad_norm": 0.7173184156417847, "learning_rate": 3.236751077033855e-05, "loss": 0.3355, "step": 38040 }, { "epoch": 22.147846332945285, "grad_norm": 0.6656613945960999, "learning_rate": 3.234172606185322e-05, "loss": 0.3467, "step": 38050 }, { "epoch": 22.153667054714784, "grad_norm": 0.6305857300758362, "learning_rate": 3.231594671748528e-05, "loss": 0.3469, "step": 38060 }, { "epoch": 22.159487776484283, "grad_norm": 0.6713994741439819, "learning_rate": 3.2290172745065815e-05, "loss": 0.3426, "step": 38070 }, { "epoch": 22.165308498253783, "grad_norm": 0.6866316795349121, "learning_rate": 3.226440415242426e-05, "loss": 0.3506, "step": 38080 }, { "epoch": 22.171129220023282, "grad_norm": 0.6972302794456482, "learning_rate": 3.223864094738846e-05, "loss": 0.3336, "step": 38090 }, { "epoch": 22.17694994179278, "grad_norm": 0.6838639974594116, "learning_rate": 3.221288313778456e-05, "loss": 0.3319, "step": 38100 }, { "epoch": 22.18277066356228, "grad_norm": 0.7337770462036133, "learning_rate": 3.2187130731437125e-05, "loss": 0.3454, "step": 38110 }, { "epoch": 22.18859138533178, "grad_norm": 0.7251157164573669, "learning_rate": 3.216138373616905e-05, "loss": 0.3379, "step": 38120 }, { "epoch": 22.194412107101282, "grad_norm": 0.6329199075698853, "learning_rate": 3.21356421598016e-05, "loss": 0.3326, "step": 38130 }, { "epoch": 22.20023282887078, "grad_norm": 0.6695235371589661, "learning_rate": 3.210990601015438e-05, "loss": 0.335, "step": 38140 }, { "epoch": 22.20605355064028, "grad_norm": 0.6160730123519897, "learning_rate": 3.208417529504535e-05, "loss": 0.3408, "step": 38150 }, { "epoch": 22.21187427240978, "grad_norm": 0.6568275094032288, "learning_rate": 3.205845002229084e-05, "loss": 0.34, "step": 38160 }, { "epoch": 22.21769499417928, "grad_norm": 0.6632944941520691, "learning_rate": 3.203273019970547e-05, "loss": 0.3308, "step": 38170 }, { "epoch": 22.22351571594878, "grad_norm": 0.6472814083099365, "learning_rate": 3.200701583510227e-05, "loss": 0.3267, "step": 38180 }, { "epoch": 22.229336437718278, "grad_norm": 0.7141303420066833, "learning_rate": 3.198130693629261e-05, "loss": 0.335, "step": 38190 }, { "epoch": 22.235157159487777, "grad_norm": 0.6795303821563721, "learning_rate": 3.195560351108612e-05, "loss": 0.3426, "step": 38200 }, { "epoch": 22.240977881257276, "grad_norm": 0.7235453128814697, "learning_rate": 3.1929905567290865e-05, "loss": 0.3519, "step": 38210 }, { "epoch": 22.246798603026775, "grad_norm": 0.7229896187782288, "learning_rate": 3.1904213112713164e-05, "loss": 0.3367, "step": 38220 }, { "epoch": 22.252619324796274, "grad_norm": 0.7003626823425293, "learning_rate": 3.187852615515774e-05, "loss": 0.3349, "step": 38230 }, { "epoch": 22.258440046565774, "grad_norm": 0.6886194348335266, "learning_rate": 3.1852844702427606e-05, "loss": 0.3305, "step": 38240 }, { "epoch": 22.264260768335273, "grad_norm": 0.7490849494934082, "learning_rate": 3.18271687623241e-05, "loss": 0.3426, "step": 38250 }, { "epoch": 22.270081490104772, "grad_norm": 0.6742451786994934, "learning_rate": 3.1801498342646896e-05, "loss": 0.3385, "step": 38260 }, { "epoch": 22.27590221187427, "grad_norm": 0.7026007771492004, "learning_rate": 3.177583345119398e-05, "loss": 0.3242, "step": 38270 }, { "epoch": 22.28172293364377, "grad_norm": 0.8662603497505188, "learning_rate": 3.17501740957617e-05, "loss": 0.3511, "step": 38280 }, { "epoch": 22.28754365541327, "grad_norm": 0.8518390655517578, "learning_rate": 3.172452028414467e-05, "loss": 0.341, "step": 38290 }, { "epoch": 22.29336437718277, "grad_norm": 0.7386759519577026, "learning_rate": 3.169887202413583e-05, "loss": 0.3334, "step": 38300 }, { "epoch": 22.29918509895227, "grad_norm": 0.777918815612793, "learning_rate": 3.167322932352646e-05, "loss": 0.3472, "step": 38310 }, { "epoch": 22.30500582072177, "grad_norm": 0.8053515553474426, "learning_rate": 3.164759219010613e-05, "loss": 0.3307, "step": 38320 }, { "epoch": 22.31082654249127, "grad_norm": 0.6550808548927307, "learning_rate": 3.1621960631662725e-05, "loss": 0.3221, "step": 38330 }, { "epoch": 22.31664726426077, "grad_norm": 0.6546236872673035, "learning_rate": 3.159633465598245e-05, "loss": 0.3258, "step": 38340 }, { "epoch": 22.32246798603027, "grad_norm": 0.8396480679512024, "learning_rate": 3.1570714270849767e-05, "loss": 0.3455, "step": 38350 }, { "epoch": 22.328288707799768, "grad_norm": 0.5917834043502808, "learning_rate": 3.1545099484047516e-05, "loss": 0.3251, "step": 38360 }, { "epoch": 22.334109429569267, "grad_norm": 0.6772609353065491, "learning_rate": 3.151949030335674e-05, "loss": 0.3426, "step": 38370 }, { "epoch": 22.339930151338766, "grad_norm": 0.6964541077613831, "learning_rate": 3.149388673655687e-05, "loss": 0.3291, "step": 38380 }, { "epoch": 22.345750873108265, "grad_norm": 0.7115899920463562, "learning_rate": 3.146828879142559e-05, "loss": 0.3485, "step": 38390 }, { "epoch": 22.351571594877765, "grad_norm": 0.7444673776626587, "learning_rate": 3.1442696475738866e-05, "loss": 0.3395, "step": 38400 }, { "epoch": 22.357392316647264, "grad_norm": 0.7197073698043823, "learning_rate": 3.141710979727098e-05, "loss": 0.3261, "step": 38410 }, { "epoch": 22.363213038416763, "grad_norm": 0.5924409031867981, "learning_rate": 3.139152876379447e-05, "loss": 0.3269, "step": 38420 }, { "epoch": 22.369033760186262, "grad_norm": 0.6481055617332458, "learning_rate": 3.1365953383080214e-05, "loss": 0.3215, "step": 38430 }, { "epoch": 22.37485448195576, "grad_norm": 0.7591788172721863, "learning_rate": 3.134038366289731e-05, "loss": 0.3383, "step": 38440 }, { "epoch": 22.38067520372526, "grad_norm": 0.8185418844223022, "learning_rate": 3.131481961101317e-05, "loss": 0.3407, "step": 38450 }, { "epoch": 22.38649592549476, "grad_norm": 0.7385023236274719, "learning_rate": 3.128926123519349e-05, "loss": 0.3268, "step": 38460 }, { "epoch": 22.39231664726426, "grad_norm": 0.6877036094665527, "learning_rate": 3.1263708543202194e-05, "loss": 0.3242, "step": 38470 }, { "epoch": 22.398137369033762, "grad_norm": 0.6711561679840088, "learning_rate": 3.123816154280155e-05, "loss": 0.3455, "step": 38480 }, { "epoch": 22.40395809080326, "grad_norm": 0.6663245558738708, "learning_rate": 3.121262024175207e-05, "loss": 0.3375, "step": 38490 }, { "epoch": 22.40977881257276, "grad_norm": 0.7620108723640442, "learning_rate": 3.118708464781248e-05, "loss": 0.3613, "step": 38500 }, { "epoch": 22.41559953434226, "grad_norm": 0.766984224319458, "learning_rate": 3.116155476873987e-05, "loss": 0.3319, "step": 38510 }, { "epoch": 22.42142025611176, "grad_norm": 0.6278165578842163, "learning_rate": 3.11360306122895e-05, "loss": 0.3398, "step": 38520 }, { "epoch": 22.427240977881258, "grad_norm": 0.7882767915725708, "learning_rate": 3.1110512186214975e-05, "loss": 0.3201, "step": 38530 }, { "epoch": 22.433061699650757, "grad_norm": 0.7037577629089355, "learning_rate": 3.1084999498268095e-05, "loss": 0.3298, "step": 38540 }, { "epoch": 22.438882421420256, "grad_norm": 0.7482263445854187, "learning_rate": 3.1059492556198934e-05, "loss": 0.3416, "step": 38550 }, { "epoch": 22.444703143189756, "grad_norm": 0.733949601650238, "learning_rate": 3.103399136775586e-05, "loss": 0.3594, "step": 38560 }, { "epoch": 22.450523864959255, "grad_norm": 0.6905362010002136, "learning_rate": 3.100849594068541e-05, "loss": 0.3302, "step": 38570 }, { "epoch": 22.456344586728754, "grad_norm": 0.6680791974067688, "learning_rate": 3.0983006282732484e-05, "loss": 0.3306, "step": 38580 }, { "epoch": 22.462165308498253, "grad_norm": 0.6853116750717163, "learning_rate": 3.0957522401640116e-05, "loss": 0.3297, "step": 38590 }, { "epoch": 22.467986030267753, "grad_norm": 0.6925538182258606, "learning_rate": 3.0932044305149645e-05, "loss": 0.3321, "step": 38600 }, { "epoch": 22.47380675203725, "grad_norm": 0.7285367846488953, "learning_rate": 3.090657200100068e-05, "loss": 0.3366, "step": 38610 }, { "epoch": 22.47962747380675, "grad_norm": 0.660686194896698, "learning_rate": 3.088110549693099e-05, "loss": 0.329, "step": 38620 }, { "epoch": 22.48544819557625, "grad_norm": 0.6876624822616577, "learning_rate": 3.085564480067667e-05, "loss": 0.335, "step": 38630 }, { "epoch": 22.49126891734575, "grad_norm": 0.7062268257141113, "learning_rate": 3.0830189919971955e-05, "loss": 0.3318, "step": 38640 }, { "epoch": 22.49708963911525, "grad_norm": 0.7710912227630615, "learning_rate": 3.080474086254939e-05, "loss": 0.3327, "step": 38650 }, { "epoch": 22.50291036088475, "grad_norm": 0.7079597115516663, "learning_rate": 3.077929763613975e-05, "loss": 0.329, "step": 38660 }, { "epoch": 22.50873108265425, "grad_norm": 0.6451923847198486, "learning_rate": 3.075386024847198e-05, "loss": 0.3263, "step": 38670 }, { "epoch": 22.51455180442375, "grad_norm": 0.7416017055511475, "learning_rate": 3.072842870727331e-05, "loss": 0.3348, "step": 38680 }, { "epoch": 22.52037252619325, "grad_norm": 0.6284575462341309, "learning_rate": 3.070300302026916e-05, "loss": 0.3365, "step": 38690 }, { "epoch": 22.52619324796275, "grad_norm": 0.6481934785842896, "learning_rate": 3.067758319518318e-05, "loss": 0.3316, "step": 38700 }, { "epoch": 22.532013969732247, "grad_norm": 0.7146244645118713, "learning_rate": 3.065216923973725e-05, "loss": 0.3371, "step": 38710 }, { "epoch": 22.537834691501747, "grad_norm": 0.5997951626777649, "learning_rate": 3.062676116165145e-05, "loss": 0.3203, "step": 38720 }, { "epoch": 22.543655413271246, "grad_norm": 0.8422952890396118, "learning_rate": 3.06013589686441e-05, "loss": 0.3293, "step": 38730 }, { "epoch": 22.549476135040745, "grad_norm": 0.6853670477867126, "learning_rate": 3.05759626684317e-05, "loss": 0.3484, "step": 38740 }, { "epoch": 22.555296856810244, "grad_norm": 0.6306631565093994, "learning_rate": 3.055057226872896e-05, "loss": 0.3252, "step": 38750 }, { "epoch": 22.561117578579744, "grad_norm": 0.6272226572036743, "learning_rate": 3.052518777724887e-05, "loss": 0.3339, "step": 38760 }, { "epoch": 22.566938300349243, "grad_norm": 0.6595916152000427, "learning_rate": 3.04998092017025e-05, "loss": 0.3339, "step": 38770 }, { "epoch": 22.572759022118742, "grad_norm": 0.7975491285324097, "learning_rate": 3.0474436549799246e-05, "loss": 0.3338, "step": 38780 }, { "epoch": 22.57857974388824, "grad_norm": 0.6985238194465637, "learning_rate": 3.044906982924661e-05, "loss": 0.3351, "step": 38790 }, { "epoch": 22.58440046565774, "grad_norm": 0.6973143815994263, "learning_rate": 3.0423709047750337e-05, "loss": 0.3209, "step": 38800 }, { "epoch": 22.59022118742724, "grad_norm": 0.7389690279960632, "learning_rate": 3.03983542130144e-05, "loss": 0.3385, "step": 38810 }, { "epoch": 22.59604190919674, "grad_norm": 0.6473512053489685, "learning_rate": 3.0373005332740877e-05, "loss": 0.335, "step": 38820 }, { "epoch": 22.601862630966238, "grad_norm": 0.6390177011489868, "learning_rate": 3.034766241463013e-05, "loss": 0.3304, "step": 38830 }, { "epoch": 22.60768335273574, "grad_norm": 0.7053418159484863, "learning_rate": 3.032232546638064e-05, "loss": 0.3308, "step": 38840 }, { "epoch": 22.61350407450524, "grad_norm": 0.6877267360687256, "learning_rate": 3.0296994495689114e-05, "loss": 0.3328, "step": 38850 }, { "epoch": 22.61932479627474, "grad_norm": 0.7352542281150818, "learning_rate": 3.0271669510250444e-05, "loss": 0.326, "step": 38860 }, { "epoch": 22.62514551804424, "grad_norm": 0.6539514660835266, "learning_rate": 3.024635051775766e-05, "loss": 0.342, "step": 38870 }, { "epoch": 22.630966239813738, "grad_norm": 0.7866979837417603, "learning_rate": 3.022103752590205e-05, "loss": 0.3479, "step": 38880 }, { "epoch": 22.636786961583237, "grad_norm": 0.6687204241752625, "learning_rate": 3.0195730542372992e-05, "loss": 0.3261, "step": 38890 }, { "epoch": 22.642607683352736, "grad_norm": 0.6576656103134155, "learning_rate": 3.0170429574858084e-05, "loss": 0.3203, "step": 38900 }, { "epoch": 22.648428405122235, "grad_norm": 0.6343047618865967, "learning_rate": 3.0145134631043127e-05, "loss": 0.3288, "step": 38910 }, { "epoch": 22.654249126891735, "grad_norm": 0.6583399176597595, "learning_rate": 3.0119845718612018e-05, "loss": 0.3379, "step": 38920 }, { "epoch": 22.660069848661234, "grad_norm": 0.6817423701286316, "learning_rate": 3.009456284524688e-05, "loss": 0.3322, "step": 38930 }, { "epoch": 22.665890570430733, "grad_norm": 0.7840295433998108, "learning_rate": 3.0069286018627967e-05, "loss": 0.3272, "step": 38940 }, { "epoch": 22.671711292200232, "grad_norm": 0.779201328754425, "learning_rate": 3.0044015246433743e-05, "loss": 0.325, "step": 38950 }, { "epoch": 22.67753201396973, "grad_norm": 0.6365430355072021, "learning_rate": 3.0018750536340755e-05, "loss": 0.3328, "step": 38960 }, { "epoch": 22.68335273573923, "grad_norm": 0.6553083062171936, "learning_rate": 2.999349189602378e-05, "loss": 0.3286, "step": 38970 }, { "epoch": 22.68917345750873, "grad_norm": 0.6615903973579407, "learning_rate": 2.9968239333155733e-05, "loss": 0.3274, "step": 38980 }, { "epoch": 22.69499417927823, "grad_norm": 0.683250904083252, "learning_rate": 2.994299285540767e-05, "loss": 0.3246, "step": 38990 }, { "epoch": 22.70081490104773, "grad_norm": 0.6495421528816223, "learning_rate": 2.9917752470448813e-05, "loss": 0.3226, "step": 39000 }, { "epoch": 22.70663562281723, "grad_norm": 0.8674367070198059, "learning_rate": 2.9892518185946495e-05, "loss": 0.3449, "step": 39010 }, { "epoch": 22.71245634458673, "grad_norm": 0.7207105755805969, "learning_rate": 2.986729000956624e-05, "loss": 0.3473, "step": 39020 }, { "epoch": 22.71827706635623, "grad_norm": 0.6300368905067444, "learning_rate": 2.9842067948971736e-05, "loss": 0.3347, "step": 39030 }, { "epoch": 22.72409778812573, "grad_norm": 0.7853327393531799, "learning_rate": 2.9816852011824727e-05, "loss": 0.3389, "step": 39040 }, { "epoch": 22.729918509895228, "grad_norm": 0.6664361953735352, "learning_rate": 2.979164220578519e-05, "loss": 0.3381, "step": 39050 }, { "epoch": 22.735739231664727, "grad_norm": 0.8559459447860718, "learning_rate": 2.9766438538511165e-05, "loss": 0.3302, "step": 39060 }, { "epoch": 22.741559953434226, "grad_norm": 0.7077667713165283, "learning_rate": 2.9741241017658873e-05, "loss": 0.3327, "step": 39070 }, { "epoch": 22.747380675203726, "grad_norm": 0.6288335919380188, "learning_rate": 2.971604965088267e-05, "loss": 0.3147, "step": 39080 }, { "epoch": 22.753201396973225, "grad_norm": 0.7446673512458801, "learning_rate": 2.9690864445835008e-05, "loss": 0.3462, "step": 39090 }, { "epoch": 22.759022118742724, "grad_norm": 0.8352145552635193, "learning_rate": 2.966568541016651e-05, "loss": 0.323, "step": 39100 }, { "epoch": 22.764842840512223, "grad_norm": 0.8620668649673462, "learning_rate": 2.9640512551525867e-05, "loss": 0.3385, "step": 39110 }, { "epoch": 22.770663562281722, "grad_norm": 0.6781274080276489, "learning_rate": 2.961534587755995e-05, "loss": 0.314, "step": 39120 }, { "epoch": 22.77648428405122, "grad_norm": 0.6552201509475708, "learning_rate": 2.959018539591375e-05, "loss": 0.3305, "step": 39130 }, { "epoch": 22.78230500582072, "grad_norm": 0.6558501124382019, "learning_rate": 2.9565031114230325e-05, "loss": 0.3514, "step": 39140 }, { "epoch": 22.78812572759022, "grad_norm": 0.6931834816932678, "learning_rate": 2.9539883040150895e-05, "loss": 0.3264, "step": 39150 }, { "epoch": 22.79394644935972, "grad_norm": 0.8221969604492188, "learning_rate": 2.9514741181314774e-05, "loss": 0.3306, "step": 39160 }, { "epoch": 22.79976717112922, "grad_norm": 0.6565084457397461, "learning_rate": 2.94896055453594e-05, "loss": 0.3318, "step": 39170 }, { "epoch": 22.80558789289872, "grad_norm": 0.6954407691955566, "learning_rate": 2.9464476139920332e-05, "loss": 0.3161, "step": 39180 }, { "epoch": 22.81140861466822, "grad_norm": 0.7280017137527466, "learning_rate": 2.9439352972631186e-05, "loss": 0.3234, "step": 39190 }, { "epoch": 22.81722933643772, "grad_norm": 0.7761529684066772, "learning_rate": 2.9414236051123757e-05, "loss": 0.3359, "step": 39200 }, { "epoch": 22.82305005820722, "grad_norm": 0.6899221539497375, "learning_rate": 2.938912538302785e-05, "loss": 0.3333, "step": 39210 }, { "epoch": 22.828870779976718, "grad_norm": 0.6234305500984192, "learning_rate": 2.9364020975971464e-05, "loss": 0.3302, "step": 39220 }, { "epoch": 22.834691501746217, "grad_norm": 0.6647295355796814, "learning_rate": 2.9338922837580657e-05, "loss": 0.3247, "step": 39230 }, { "epoch": 22.840512223515717, "grad_norm": 0.6465012431144714, "learning_rate": 2.931383097547955e-05, "loss": 0.3392, "step": 39240 }, { "epoch": 22.846332945285216, "grad_norm": 0.6742829084396362, "learning_rate": 2.928874539729043e-05, "loss": 0.3297, "step": 39250 }, { "epoch": 22.852153667054715, "grad_norm": 0.6260968446731567, "learning_rate": 2.926366611063358e-05, "loss": 0.3208, "step": 39260 }, { "epoch": 22.857974388824214, "grad_norm": 0.6927050948143005, "learning_rate": 2.9238593123127463e-05, "loss": 0.3246, "step": 39270 }, { "epoch": 22.863795110593713, "grad_norm": 0.6517464518547058, "learning_rate": 2.9213526442388583e-05, "loss": 0.331, "step": 39280 }, { "epoch": 22.869615832363213, "grad_norm": 0.6933287978172302, "learning_rate": 2.9188466076031545e-05, "loss": 0.3452, "step": 39290 }, { "epoch": 22.875436554132712, "grad_norm": 0.6756855249404907, "learning_rate": 2.9163412031669012e-05, "loss": 0.3211, "step": 39300 }, { "epoch": 22.88125727590221, "grad_norm": 0.5991000533103943, "learning_rate": 2.913836431691175e-05, "loss": 0.3372, "step": 39310 }, { "epoch": 22.88707799767171, "grad_norm": 0.6881455779075623, "learning_rate": 2.9113322939368583e-05, "loss": 0.3264, "step": 39320 }, { "epoch": 22.89289871944121, "grad_norm": 0.6553789377212524, "learning_rate": 2.9088287906646427e-05, "loss": 0.3261, "step": 39330 }, { "epoch": 22.89871944121071, "grad_norm": 0.7290276288986206, "learning_rate": 2.906325922635024e-05, "loss": 0.353, "step": 39340 }, { "epoch": 22.904540162980208, "grad_norm": 0.7938579320907593, "learning_rate": 2.903823690608313e-05, "loss": 0.3304, "step": 39350 }, { "epoch": 22.91036088474971, "grad_norm": 0.7323189377784729, "learning_rate": 2.9013220953446174e-05, "loss": 0.3285, "step": 39360 }, { "epoch": 22.91618160651921, "grad_norm": 0.6657153964042664, "learning_rate": 2.8988211376038564e-05, "loss": 0.3195, "step": 39370 }, { "epoch": 22.92200232828871, "grad_norm": 0.6830253601074219, "learning_rate": 2.8963208181457564e-05, "loss": 0.3361, "step": 39380 }, { "epoch": 22.92782305005821, "grad_norm": 0.6577671766281128, "learning_rate": 2.8938211377298453e-05, "loss": 0.3221, "step": 39390 }, { "epoch": 22.933643771827708, "grad_norm": 0.6414768695831299, "learning_rate": 2.8913220971154652e-05, "loss": 0.3277, "step": 39400 }, { "epoch": 22.939464493597207, "grad_norm": 0.5950871109962463, "learning_rate": 2.888823697061753e-05, "loss": 0.324, "step": 39410 }, { "epoch": 22.945285215366706, "grad_norm": 0.5944444537162781, "learning_rate": 2.8863259383276618e-05, "loss": 0.3219, "step": 39420 }, { "epoch": 22.951105937136205, "grad_norm": 0.6308544278144836, "learning_rate": 2.8838288216719395e-05, "loss": 0.3264, "step": 39430 }, { "epoch": 22.956926658905704, "grad_norm": 0.6565899848937988, "learning_rate": 2.8813323478531484e-05, "loss": 0.3214, "step": 39440 }, { "epoch": 22.962747380675204, "grad_norm": 0.5810573697090149, "learning_rate": 2.8788365176296496e-05, "loss": 0.3252, "step": 39450 }, { "epoch": 22.968568102444703, "grad_norm": 0.7014142870903015, "learning_rate": 2.876341331759611e-05, "loss": 0.3244, "step": 39460 }, { "epoch": 22.974388824214202, "grad_norm": 0.6343380808830261, "learning_rate": 2.8738467910010036e-05, "loss": 0.327, "step": 39470 }, { "epoch": 22.9802095459837, "grad_norm": 0.5658296346664429, "learning_rate": 2.8713528961116032e-05, "loss": 0.3265, "step": 39480 }, { "epoch": 22.9860302677532, "grad_norm": 0.6942610740661621, "learning_rate": 2.8688596478489875e-05, "loss": 0.3421, "step": 39490 }, { "epoch": 22.9918509895227, "grad_norm": 0.5951147079467773, "learning_rate": 2.8663670469705434e-05, "loss": 0.3303, "step": 39500 }, { "epoch": 22.9976717112922, "grad_norm": 0.6446461081504822, "learning_rate": 2.8638750942334546e-05, "loss": 0.342, "step": 39510 }, { "epoch": 23.003492433061698, "grad_norm": 0.5962966084480286, "learning_rate": 2.8613837903947115e-05, "loss": 0.3352, "step": 39520 }, { "epoch": 23.009313154831197, "grad_norm": 0.6972450017929077, "learning_rate": 2.858893136211106e-05, "loss": 0.3342, "step": 39530 }, { "epoch": 23.0151338766007, "grad_norm": 0.6976050138473511, "learning_rate": 2.8564031324392315e-05, "loss": 0.3412, "step": 39540 }, { "epoch": 23.0209545983702, "grad_norm": 0.6520242691040039, "learning_rate": 2.85391377983549e-05, "loss": 0.3326, "step": 39550 }, { "epoch": 23.0267753201397, "grad_norm": 0.7362035512924194, "learning_rate": 2.851425079156075e-05, "loss": 0.3314, "step": 39560 }, { "epoch": 23.032596041909198, "grad_norm": 0.7666410207748413, "learning_rate": 2.848937031156994e-05, "loss": 0.345, "step": 39570 }, { "epoch": 23.038416763678697, "grad_norm": 0.6469203233718872, "learning_rate": 2.846449636594044e-05, "loss": 0.3338, "step": 39580 }, { "epoch": 23.044237485448196, "grad_norm": 0.6081802248954773, "learning_rate": 2.843962896222836e-05, "loss": 0.3344, "step": 39590 }, { "epoch": 23.050058207217695, "grad_norm": 0.64919114112854, "learning_rate": 2.8414768107987722e-05, "loss": 0.3312, "step": 39600 }, { "epoch": 23.055878928987195, "grad_norm": 0.6575202941894531, "learning_rate": 2.838991381077061e-05, "loss": 0.3384, "step": 39610 }, { "epoch": 23.061699650756694, "grad_norm": 0.6985563635826111, "learning_rate": 2.83650660781271e-05, "loss": 0.3258, "step": 39620 }, { "epoch": 23.067520372526193, "grad_norm": 0.6907888054847717, "learning_rate": 2.8340224917605285e-05, "loss": 0.3339, "step": 39630 }, { "epoch": 23.073341094295692, "grad_norm": 0.5949862599372864, "learning_rate": 2.831539033675122e-05, "loss": 0.3322, "step": 39640 }, { "epoch": 23.07916181606519, "grad_norm": 0.6172815561294556, "learning_rate": 2.8290562343109038e-05, "loss": 0.3295, "step": 39650 }, { "epoch": 23.08498253783469, "grad_norm": 0.6668429374694824, "learning_rate": 2.826574094422082e-05, "loss": 0.3332, "step": 39660 }, { "epoch": 23.09080325960419, "grad_norm": 0.6206480860710144, "learning_rate": 2.8240926147626645e-05, "loss": 0.332, "step": 39670 }, { "epoch": 23.09662398137369, "grad_norm": 0.5780577659606934, "learning_rate": 2.8216117960864586e-05, "loss": 0.3275, "step": 39680 }, { "epoch": 23.10244470314319, "grad_norm": 0.6354905366897583, "learning_rate": 2.8191316391470703e-05, "loss": 0.3277, "step": 39690 }, { "epoch": 23.108265424912688, "grad_norm": 0.6385529637336731, "learning_rate": 2.816652144697911e-05, "loss": 0.3306, "step": 39700 }, { "epoch": 23.11408614668219, "grad_norm": 0.5980204343795776, "learning_rate": 2.8141733134921783e-05, "loss": 0.333, "step": 39710 }, { "epoch": 23.11990686845169, "grad_norm": 0.6360259652137756, "learning_rate": 2.811695146282884e-05, "loss": 0.3269, "step": 39720 }, { "epoch": 23.12572759022119, "grad_norm": 0.6186702847480774, "learning_rate": 2.8092176438228212e-05, "loss": 0.3444, "step": 39730 }, { "epoch": 23.131548311990688, "grad_norm": 0.6348996162414551, "learning_rate": 2.806740806864598e-05, "loss": 0.33, "step": 39740 }, { "epoch": 23.137369033760187, "grad_norm": 0.6674802899360657, "learning_rate": 2.804264636160604e-05, "loss": 0.3384, "step": 39750 }, { "epoch": 23.143189755529686, "grad_norm": 0.6064538955688477, "learning_rate": 2.8017891324630402e-05, "loss": 0.3299, "step": 39760 }, { "epoch": 23.149010477299186, "grad_norm": 0.6769959330558777, "learning_rate": 2.7993142965238976e-05, "loss": 0.3275, "step": 39770 }, { "epoch": 23.154831199068685, "grad_norm": 0.6661591529846191, "learning_rate": 2.7968401290949665e-05, "loss": 0.3341, "step": 39780 }, { "epoch": 23.160651920838184, "grad_norm": 0.730217456817627, "learning_rate": 2.7943666309278328e-05, "loss": 0.3291, "step": 39790 }, { "epoch": 23.166472642607683, "grad_norm": 0.6662181615829468, "learning_rate": 2.7918938027738783e-05, "loss": 0.3173, "step": 39800 }, { "epoch": 23.172293364377182, "grad_norm": 0.7315805554389954, "learning_rate": 2.789421645384287e-05, "loss": 0.3495, "step": 39810 }, { "epoch": 23.17811408614668, "grad_norm": 0.6225489974021912, "learning_rate": 2.786950159510032e-05, "loss": 0.323, "step": 39820 }, { "epoch": 23.18393480791618, "grad_norm": 0.5722255706787109, "learning_rate": 2.7844793459018876e-05, "loss": 0.3262, "step": 39830 }, { "epoch": 23.18975552968568, "grad_norm": 0.6287586688995361, "learning_rate": 2.7820092053104195e-05, "loss": 0.3369, "step": 39840 }, { "epoch": 23.19557625145518, "grad_norm": 0.6789332032203674, "learning_rate": 2.7795397384859933e-05, "loss": 0.3452, "step": 39850 }, { "epoch": 23.20139697322468, "grad_norm": 0.6904298067092896, "learning_rate": 2.7770709461787638e-05, "loss": 0.3367, "step": 39860 }, { "epoch": 23.207217694994178, "grad_norm": 0.6491016149520874, "learning_rate": 2.7746028291386915e-05, "loss": 0.332, "step": 39870 }, { "epoch": 23.213038416763677, "grad_norm": 0.6175217628479004, "learning_rate": 2.772135388115519e-05, "loss": 0.3252, "step": 39880 }, { "epoch": 23.21885913853318, "grad_norm": 0.6040270924568176, "learning_rate": 2.7696686238587945e-05, "loss": 0.3255, "step": 39890 }, { "epoch": 23.22467986030268, "grad_norm": 0.6444884538650513, "learning_rate": 2.7672025371178505e-05, "loss": 0.3418, "step": 39900 }, { "epoch": 23.230500582072178, "grad_norm": 0.6483582854270935, "learning_rate": 2.7647371286418238e-05, "loss": 0.333, "step": 39910 }, { "epoch": 23.236321303841677, "grad_norm": 0.7698671221733093, "learning_rate": 2.762272399179639e-05, "loss": 0.3433, "step": 39920 }, { "epoch": 23.242142025611177, "grad_norm": 0.8120139837265015, "learning_rate": 2.7598083494800154e-05, "loss": 0.3206, "step": 39930 }, { "epoch": 23.247962747380676, "grad_norm": 0.7380738258361816, "learning_rate": 2.7573449802914664e-05, "loss": 0.3333, "step": 39940 }, { "epoch": 23.253783469150175, "grad_norm": 0.6060009598731995, "learning_rate": 2.7548822923622964e-05, "loss": 0.3236, "step": 39950 }, { "epoch": 23.259604190919674, "grad_norm": 0.6552690863609314, "learning_rate": 2.752420286440609e-05, "loss": 0.3365, "step": 39960 }, { "epoch": 23.265424912689173, "grad_norm": 0.6701632738113403, "learning_rate": 2.749958963274295e-05, "loss": 0.324, "step": 39970 }, { "epoch": 23.271245634458673, "grad_norm": 0.6771714091300964, "learning_rate": 2.747498323611039e-05, "loss": 0.3239, "step": 39980 }, { "epoch": 23.277066356228172, "grad_norm": 0.6820157170295715, "learning_rate": 2.7450383681983184e-05, "loss": 0.3308, "step": 39990 }, { "epoch": 23.28288707799767, "grad_norm": 0.6291298270225525, "learning_rate": 2.742579097783403e-05, "loss": 0.3389, "step": 40000 } ], "logging_steps": 10, "max_steps": 60000, "num_input_tokens_seen": 0, "num_train_epochs": 35, "save_steps": 10000, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 0.0, "train_batch_size": 16, "trial_name": null, "trial_params": null }