diff --git "a/checkpoints/GR00T/StackCube/500/checkpoint-100000/trainer_state.json" "b/checkpoints/GR00T/StackCube/500/checkpoint-100000/trainer_state.json" new file mode 100644--- /dev/null +++ "b/checkpoints/GR00T/StackCube/500/checkpoint-100000/trainer_state.json" @@ -0,0 +1,70034 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 15.532774153463809, + "eval_steps": 500, + "global_step": 100000, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.0015532774153463808, + "grad_norm": 10.520315170288086, + "learning_rate": 1.8e-07, + "loss": 0.862, + "step": 10 + }, + { + "epoch": 0.0031065548306927616, + "grad_norm": 8.840570449829102, + "learning_rate": 3.8e-07, + "loss": 0.9135, + "step": 20 + }, + { + "epoch": 0.004659832246039142, + "grad_norm": 9.000713348388672, + "learning_rate": 5.8e-07, + "loss": 0.8912, + "step": 30 + }, + { + "epoch": 0.006213109661385523, + "grad_norm": 8.670975685119629, + "learning_rate": 7.8e-07, + "loss": 0.7828, + "step": 40 + }, + { + "epoch": 0.007766387076731904, + "grad_norm": 5.826544761657715, + "learning_rate": 9.8e-07, + "loss": 0.6409, + "step": 50 + }, + { + "epoch": 0.009319664492078284, + "grad_norm": 2.992053985595703, + "learning_rate": 1.18e-06, + "loss": 0.5705, + "step": 60 + }, + { + "epoch": 0.010872941907424665, + "grad_norm": 3.2876100540161133, + "learning_rate": 1.3800000000000001e-06, + "loss": 0.4171, + "step": 70 + }, + { + "epoch": 0.012426219322771047, + "grad_norm": 1.8181413412094116, + "learning_rate": 1.5800000000000003e-06, + "loss": 0.3746, + "step": 80 + }, + { + "epoch": 0.013979496738117428, + "grad_norm": 1.7137627601623535, + "learning_rate": 1.7800000000000001e-06, + "loss": 0.3036, + "step": 90 + }, + { + "epoch": 0.015532774153463809, + "grad_norm": 1.6369364261627197, + "learning_rate": 1.98e-06, + "loss": 0.2483, + "step": 100 + }, + { + "epoch": 0.017086051568810188, + "grad_norm": 0.9569445848464966, + "learning_rate": 2.1800000000000003e-06, + "loss": 0.1973, + "step": 110 + }, + { + "epoch": 0.01863932898415657, + "grad_norm": 0.9038352370262146, + "learning_rate": 2.38e-06, + "loss": 0.1931, + "step": 120 + }, + { + "epoch": 0.02019260639950295, + "grad_norm": 0.9611704349517822, + "learning_rate": 2.5800000000000003e-06, + "loss": 0.1662, + "step": 130 + }, + { + "epoch": 0.02174588381484933, + "grad_norm": 0.9995275139808655, + "learning_rate": 2.78e-06, + "loss": 0.1683, + "step": 140 + }, + { + "epoch": 0.023299161230195712, + "grad_norm": 1.523456335067749, + "learning_rate": 2.9800000000000003e-06, + "loss": 0.1359, + "step": 150 + }, + { + "epoch": 0.024852438645542093, + "grad_norm": 1.578064203262329, + "learning_rate": 3.1800000000000005e-06, + "loss": 0.1265, + "step": 160 + }, + { + "epoch": 0.026405716060888474, + "grad_norm": 0.8265012502670288, + "learning_rate": 3.38e-06, + "loss": 0.1153, + "step": 170 + }, + { + "epoch": 0.027958993476234855, + "grad_norm": 1.146382212638855, + "learning_rate": 3.58e-06, + "loss": 0.1272, + "step": 180 + }, + { + "epoch": 0.029512270891581236, + "grad_norm": 1.2577002048492432, + "learning_rate": 3.7800000000000002e-06, + "loss": 0.0915, + "step": 190 + }, + { + "epoch": 0.031065548306927617, + "grad_norm": 1.0999211072921753, + "learning_rate": 3.98e-06, + "loss": 0.097, + "step": 200 + }, + { + "epoch": 0.032618825722273995, + "grad_norm": 1.0050560235977173, + "learning_rate": 4.18e-06, + "loss": 0.091, + "step": 210 + }, + { + "epoch": 0.034172103137620376, + "grad_norm": 1.0925753116607666, + "learning_rate": 4.38e-06, + "loss": 0.1003, + "step": 220 + }, + { + "epoch": 0.03572538055296676, + "grad_norm": 1.0926908254623413, + "learning_rate": 4.58e-06, + "loss": 0.0768, + "step": 230 + }, + { + "epoch": 0.03727865796831314, + "grad_norm": 1.0615990161895752, + "learning_rate": 4.780000000000001e-06, + "loss": 0.0848, + "step": 240 + }, + { + "epoch": 0.03883193538365952, + "grad_norm": 1.4504543542861938, + "learning_rate": 4.98e-06, + "loss": 0.072, + "step": 250 + }, + { + "epoch": 0.0403852127990059, + "grad_norm": 0.6974684000015259, + "learning_rate": 5.18e-06, + "loss": 0.061, + "step": 260 + }, + { + "epoch": 0.04193849021435228, + "grad_norm": 0.8194519281387329, + "learning_rate": 5.38e-06, + "loss": 0.0746, + "step": 270 + }, + { + "epoch": 0.04349176762969866, + "grad_norm": 2.718522310256958, + "learning_rate": 5.580000000000001e-06, + "loss": 0.0727, + "step": 280 + }, + { + "epoch": 0.04504504504504504, + "grad_norm": 0.9042211771011353, + "learning_rate": 5.78e-06, + "loss": 0.0768, + "step": 290 + }, + { + "epoch": 0.046598322460391424, + "grad_norm": 0.7680805325508118, + "learning_rate": 5.98e-06, + "loss": 0.0645, + "step": 300 + }, + { + "epoch": 0.048151599875737805, + "grad_norm": 0.8747005462646484, + "learning_rate": 6.18e-06, + "loss": 0.0635, + "step": 310 + }, + { + "epoch": 0.049704877291084186, + "grad_norm": 1.0177611112594604, + "learning_rate": 6.38e-06, + "loss": 0.0642, + "step": 320 + }, + { + "epoch": 0.05125815470643057, + "grad_norm": 1.2854441404342651, + "learning_rate": 6.58e-06, + "loss": 0.071, + "step": 330 + }, + { + "epoch": 0.05281143212177695, + "grad_norm": 0.9579643607139587, + "learning_rate": 6.78e-06, + "loss": 0.0584, + "step": 340 + }, + { + "epoch": 0.05436470953712333, + "grad_norm": 0.944252073764801, + "learning_rate": 6.98e-06, + "loss": 0.0613, + "step": 350 + }, + { + "epoch": 0.05591798695246971, + "grad_norm": 0.8837812542915344, + "learning_rate": 7.180000000000001e-06, + "loss": 0.0485, + "step": 360 + }, + { + "epoch": 0.05747126436781609, + "grad_norm": 0.9329251646995544, + "learning_rate": 7.3800000000000005e-06, + "loss": 0.0542, + "step": 370 + }, + { + "epoch": 0.05902454178316247, + "grad_norm": 0.7117618322372437, + "learning_rate": 7.580000000000001e-06, + "loss": 0.0546, + "step": 380 + }, + { + "epoch": 0.06057781919850885, + "grad_norm": 1.0061190128326416, + "learning_rate": 7.78e-06, + "loss": 0.051, + "step": 390 + }, + { + "epoch": 0.062131096613855234, + "grad_norm": 2.044189214706421, + "learning_rate": 7.98e-06, + "loss": 0.0525, + "step": 400 + }, + { + "epoch": 0.06368437402920162, + "grad_norm": 0.9586120843887329, + "learning_rate": 8.18e-06, + "loss": 0.0625, + "step": 410 + }, + { + "epoch": 0.06523765144454799, + "grad_norm": 1.048461675643921, + "learning_rate": 8.380000000000001e-06, + "loss": 0.0532, + "step": 420 + }, + { + "epoch": 0.06679092885989438, + "grad_norm": 0.9063873887062073, + "learning_rate": 8.580000000000001e-06, + "loss": 0.0502, + "step": 430 + }, + { + "epoch": 0.06834420627524075, + "grad_norm": 0.9286721348762512, + "learning_rate": 8.78e-06, + "loss": 0.0458, + "step": 440 + }, + { + "epoch": 0.06989748369058714, + "grad_norm": 0.815371036529541, + "learning_rate": 8.98e-06, + "loss": 0.0466, + "step": 450 + }, + { + "epoch": 0.07145076110593351, + "grad_norm": 0.6235409379005432, + "learning_rate": 9.180000000000002e-06, + "loss": 0.0506, + "step": 460 + }, + { + "epoch": 0.0730040385212799, + "grad_norm": 0.6477558612823486, + "learning_rate": 9.38e-06, + "loss": 0.0448, + "step": 470 + }, + { + "epoch": 0.07455731593662628, + "grad_norm": 0.8587325811386108, + "learning_rate": 9.58e-06, + "loss": 0.0462, + "step": 480 + }, + { + "epoch": 0.07611059335197266, + "grad_norm": 0.6987264752388, + "learning_rate": 9.78e-06, + "loss": 0.0404, + "step": 490 + }, + { + "epoch": 0.07766387076731904, + "grad_norm": 0.4682886004447937, + "learning_rate": 9.980000000000001e-06, + "loss": 0.0396, + "step": 500 + }, + { + "epoch": 0.07921714818266543, + "grad_norm": 0.6321015954017639, + "learning_rate": 1.018e-05, + "loss": 0.0429, + "step": 510 + }, + { + "epoch": 0.0807704255980118, + "grad_norm": 0.7103893756866455, + "learning_rate": 1.038e-05, + "loss": 0.041, + "step": 520 + }, + { + "epoch": 0.08232370301335819, + "grad_norm": 0.915998637676239, + "learning_rate": 1.058e-05, + "loss": 0.04, + "step": 530 + }, + { + "epoch": 0.08387698042870456, + "grad_norm": 0.9832605719566345, + "learning_rate": 1.0780000000000002e-05, + "loss": 0.0394, + "step": 540 + }, + { + "epoch": 0.08543025784405095, + "grad_norm": 0.8607420325279236, + "learning_rate": 1.098e-05, + "loss": 0.0447, + "step": 550 + }, + { + "epoch": 0.08698353525939732, + "grad_norm": 0.6145164370536804, + "learning_rate": 1.118e-05, + "loss": 0.035, + "step": 560 + }, + { + "epoch": 0.08853681267474371, + "grad_norm": 0.8953571915626526, + "learning_rate": 1.1380000000000001e-05, + "loss": 0.0375, + "step": 570 + }, + { + "epoch": 0.09009009009009009, + "grad_norm": 0.6377115845680237, + "learning_rate": 1.1580000000000001e-05, + "loss": 0.0382, + "step": 580 + }, + { + "epoch": 0.09164336750543647, + "grad_norm": 0.63884037733078, + "learning_rate": 1.178e-05, + "loss": 0.0374, + "step": 590 + }, + { + "epoch": 0.09319664492078285, + "grad_norm": 0.5148037075996399, + "learning_rate": 1.198e-05, + "loss": 0.0374, + "step": 600 + }, + { + "epoch": 0.09474992233612924, + "grad_norm": 0.709943950176239, + "learning_rate": 1.2180000000000002e-05, + "loss": 0.04, + "step": 610 + }, + { + "epoch": 0.09630319975147561, + "grad_norm": 0.9653154015541077, + "learning_rate": 1.238e-05, + "loss": 0.0387, + "step": 620 + }, + { + "epoch": 0.097856477166822, + "grad_norm": 0.8201069235801697, + "learning_rate": 1.258e-05, + "loss": 0.0339, + "step": 630 + }, + { + "epoch": 0.09940975458216837, + "grad_norm": 0.5168118476867676, + "learning_rate": 1.278e-05, + "loss": 0.0382, + "step": 640 + }, + { + "epoch": 0.10096303199751476, + "grad_norm": 0.8511587977409363, + "learning_rate": 1.2980000000000001e-05, + "loss": 0.0431, + "step": 650 + }, + { + "epoch": 0.10251630941286113, + "grad_norm": 0.9344369173049927, + "learning_rate": 1.3180000000000001e-05, + "loss": 0.0338, + "step": 660 + }, + { + "epoch": 0.10406958682820752, + "grad_norm": 0.7085831761360168, + "learning_rate": 1.338e-05, + "loss": 0.0409, + "step": 670 + }, + { + "epoch": 0.1056228642435539, + "grad_norm": 0.6299055218696594, + "learning_rate": 1.358e-05, + "loss": 0.0362, + "step": 680 + }, + { + "epoch": 0.10717614165890028, + "grad_norm": 0.9340008497238159, + "learning_rate": 1.3780000000000002e-05, + "loss": 0.033, + "step": 690 + }, + { + "epoch": 0.10872941907424666, + "grad_norm": 0.7196189165115356, + "learning_rate": 1.3980000000000002e-05, + "loss": 0.0335, + "step": 700 + }, + { + "epoch": 0.11028269648959305, + "grad_norm": 0.5900976061820984, + "learning_rate": 1.4180000000000001e-05, + "loss": 0.0431, + "step": 710 + }, + { + "epoch": 0.11183597390493942, + "grad_norm": 0.8505810499191284, + "learning_rate": 1.4380000000000001e-05, + "loss": 0.0378, + "step": 720 + }, + { + "epoch": 0.11338925132028581, + "grad_norm": 0.5741161704063416, + "learning_rate": 1.4580000000000003e-05, + "loss": 0.0402, + "step": 730 + }, + { + "epoch": 0.11494252873563218, + "grad_norm": 0.7534946799278259, + "learning_rate": 1.4779999999999999e-05, + "loss": 0.0327, + "step": 740 + }, + { + "epoch": 0.11649580615097857, + "grad_norm": 0.7736256718635559, + "learning_rate": 1.4979999999999999e-05, + "loss": 0.035, + "step": 750 + }, + { + "epoch": 0.11804908356632494, + "grad_norm": 0.622745931148529, + "learning_rate": 1.518e-05, + "loss": 0.0271, + "step": 760 + }, + { + "epoch": 0.11960236098167133, + "grad_norm": 0.5686683058738708, + "learning_rate": 1.538e-05, + "loss": 0.0375, + "step": 770 + }, + { + "epoch": 0.1211556383970177, + "grad_norm": 0.6631718873977661, + "learning_rate": 1.558e-05, + "loss": 0.0352, + "step": 780 + }, + { + "epoch": 0.1227089158123641, + "grad_norm": 0.4889935553073883, + "learning_rate": 1.578e-05, + "loss": 0.0365, + "step": 790 + }, + { + "epoch": 0.12426219322771047, + "grad_norm": 0.4553696811199188, + "learning_rate": 1.598e-05, + "loss": 0.0375, + "step": 800 + }, + { + "epoch": 0.12581547064305684, + "grad_norm": 0.3102133274078369, + "learning_rate": 1.618e-05, + "loss": 0.0293, + "step": 810 + }, + { + "epoch": 0.12736874805840323, + "grad_norm": 0.43393126130104065, + "learning_rate": 1.6380000000000002e-05, + "loss": 0.0329, + "step": 820 + }, + { + "epoch": 0.12892202547374962, + "grad_norm": 0.38528746366500854, + "learning_rate": 1.658e-05, + "loss": 0.0326, + "step": 830 + }, + { + "epoch": 0.13047530288909598, + "grad_norm": 0.47710102796554565, + "learning_rate": 1.6780000000000002e-05, + "loss": 0.0307, + "step": 840 + }, + { + "epoch": 0.13202858030444237, + "grad_norm": 0.550912082195282, + "learning_rate": 1.698e-05, + "loss": 0.0304, + "step": 850 + }, + { + "epoch": 0.13358185771978875, + "grad_norm": 0.6287652850151062, + "learning_rate": 1.718e-05, + "loss": 0.0332, + "step": 860 + }, + { + "epoch": 0.13513513513513514, + "grad_norm": 0.8281686902046204, + "learning_rate": 1.7380000000000003e-05, + "loss": 0.0337, + "step": 870 + }, + { + "epoch": 0.1366884125504815, + "grad_norm": 0.7580209970474243, + "learning_rate": 1.758e-05, + "loss": 0.0309, + "step": 880 + }, + { + "epoch": 0.1382416899658279, + "grad_norm": 0.6278378367424011, + "learning_rate": 1.7780000000000003e-05, + "loss": 0.0356, + "step": 890 + }, + { + "epoch": 0.13979496738117428, + "grad_norm": 0.5295566916465759, + "learning_rate": 1.798e-05, + "loss": 0.0419, + "step": 900 + }, + { + "epoch": 0.14134824479652067, + "grad_norm": 0.40549537539482117, + "learning_rate": 1.818e-05, + "loss": 0.0316, + "step": 910 + }, + { + "epoch": 0.14290152221186703, + "grad_norm": 0.9405691623687744, + "learning_rate": 1.838e-05, + "loss": 0.0298, + "step": 920 + }, + { + "epoch": 0.14445479962721341, + "grad_norm": 0.4894973039627075, + "learning_rate": 1.858e-05, + "loss": 0.0266, + "step": 930 + }, + { + "epoch": 0.1460080770425598, + "grad_norm": 0.5776618719100952, + "learning_rate": 1.878e-05, + "loss": 0.0243, + "step": 940 + }, + { + "epoch": 0.1475613544579062, + "grad_norm": 0.4884398877620697, + "learning_rate": 1.898e-05, + "loss": 0.0322, + "step": 950 + }, + { + "epoch": 0.14911463187325255, + "grad_norm": 0.5399041771888733, + "learning_rate": 1.918e-05, + "loss": 0.0266, + "step": 960 + }, + { + "epoch": 0.15066790928859894, + "grad_norm": 0.578637421131134, + "learning_rate": 1.938e-05, + "loss": 0.0266, + "step": 970 + }, + { + "epoch": 0.15222118670394533, + "grad_norm": 0.5606555342674255, + "learning_rate": 1.9580000000000002e-05, + "loss": 0.0255, + "step": 980 + }, + { + "epoch": 0.15377446411929171, + "grad_norm": 0.6661631464958191, + "learning_rate": 1.978e-05, + "loss": 0.033, + "step": 990 + }, + { + "epoch": 0.15532774153463808, + "grad_norm": 0.3831423223018646, + "learning_rate": 1.9980000000000002e-05, + "loss": 0.028, + "step": 1000 + }, + { + "epoch": 0.15688101894998446, + "grad_norm": 0.4626130759716034, + "learning_rate": 2.0180000000000003e-05, + "loss": 0.024, + "step": 1010 + }, + { + "epoch": 0.15843429636533085, + "grad_norm": 0.4741255044937134, + "learning_rate": 2.038e-05, + "loss": 0.0256, + "step": 1020 + }, + { + "epoch": 0.15998757378067724, + "grad_norm": 0.6987898945808411, + "learning_rate": 2.0580000000000003e-05, + "loss": 0.0267, + "step": 1030 + }, + { + "epoch": 0.1615408511960236, + "grad_norm": 0.5332676768302917, + "learning_rate": 2.078e-05, + "loss": 0.0242, + "step": 1040 + }, + { + "epoch": 0.16309412861137, + "grad_norm": 0.47977954149246216, + "learning_rate": 2.098e-05, + "loss": 0.0243, + "step": 1050 + }, + { + "epoch": 0.16464740602671638, + "grad_norm": 0.8516312837600708, + "learning_rate": 2.118e-05, + "loss": 0.0334, + "step": 1060 + }, + { + "epoch": 0.16620068344206276, + "grad_norm": 0.5538998246192932, + "learning_rate": 2.138e-05, + "loss": 0.0289, + "step": 1070 + }, + { + "epoch": 0.16775396085740912, + "grad_norm": 0.6159148216247559, + "learning_rate": 2.158e-05, + "loss": 0.0314, + "step": 1080 + }, + { + "epoch": 0.1693072382727555, + "grad_norm": 0.6107079386711121, + "learning_rate": 2.178e-05, + "loss": 0.021, + "step": 1090 + }, + { + "epoch": 0.1708605156881019, + "grad_norm": 0.4216293692588806, + "learning_rate": 2.198e-05, + "loss": 0.0237, + "step": 1100 + }, + { + "epoch": 0.1724137931034483, + "grad_norm": 0.561742901802063, + "learning_rate": 2.218e-05, + "loss": 0.0239, + "step": 1110 + }, + { + "epoch": 0.17396707051879465, + "grad_norm": 0.440642774105072, + "learning_rate": 2.2380000000000003e-05, + "loss": 0.0226, + "step": 1120 + }, + { + "epoch": 0.17552034793414104, + "grad_norm": 0.4446122348308563, + "learning_rate": 2.258e-05, + "loss": 0.0319, + "step": 1130 + }, + { + "epoch": 0.17707362534948742, + "grad_norm": 0.524465799331665, + "learning_rate": 2.2780000000000002e-05, + "loss": 0.0242, + "step": 1140 + }, + { + "epoch": 0.1786269027648338, + "grad_norm": 0.4178518056869507, + "learning_rate": 2.298e-05, + "loss": 0.0236, + "step": 1150 + }, + { + "epoch": 0.18018018018018017, + "grad_norm": 0.4855317175388336, + "learning_rate": 2.318e-05, + "loss": 0.0258, + "step": 1160 + }, + { + "epoch": 0.18173345759552656, + "grad_norm": 0.5085672736167908, + "learning_rate": 2.3380000000000003e-05, + "loss": 0.0239, + "step": 1170 + }, + { + "epoch": 0.18328673501087295, + "grad_norm": 0.5359751582145691, + "learning_rate": 2.358e-05, + "loss": 0.0228, + "step": 1180 + }, + { + "epoch": 0.18484001242621934, + "grad_norm": 0.722781777381897, + "learning_rate": 2.3780000000000003e-05, + "loss": 0.0216, + "step": 1190 + }, + { + "epoch": 0.1863932898415657, + "grad_norm": 0.5163661241531372, + "learning_rate": 2.398e-05, + "loss": 0.0261, + "step": 1200 + }, + { + "epoch": 0.18794656725691208, + "grad_norm": 0.4173479676246643, + "learning_rate": 2.418e-05, + "loss": 0.0257, + "step": 1210 + }, + { + "epoch": 0.18949984467225847, + "grad_norm": 0.40653157234191895, + "learning_rate": 2.438e-05, + "loss": 0.0209, + "step": 1220 + }, + { + "epoch": 0.19105312208760486, + "grad_norm": 0.45858967304229736, + "learning_rate": 2.4580000000000002e-05, + "loss": 0.0241, + "step": 1230 + }, + { + "epoch": 0.19260639950295122, + "grad_norm": 0.6160152554512024, + "learning_rate": 2.478e-05, + "loss": 0.0211, + "step": 1240 + }, + { + "epoch": 0.1941596769182976, + "grad_norm": 0.8483706116676331, + "learning_rate": 2.498e-05, + "loss": 0.0254, + "step": 1250 + }, + { + "epoch": 0.195712954333644, + "grad_norm": 0.4370304346084595, + "learning_rate": 2.5180000000000003e-05, + "loss": 0.0228, + "step": 1260 + }, + { + "epoch": 0.19726623174899036, + "grad_norm": 0.641593337059021, + "learning_rate": 2.5380000000000004e-05, + "loss": 0.0227, + "step": 1270 + }, + { + "epoch": 0.19881950916433674, + "grad_norm": 0.7113029956817627, + "learning_rate": 2.5580000000000002e-05, + "loss": 0.026, + "step": 1280 + }, + { + "epoch": 0.20037278657968313, + "grad_norm": 0.7529031038284302, + "learning_rate": 2.5779999999999997e-05, + "loss": 0.0349, + "step": 1290 + }, + { + "epoch": 0.20192606399502952, + "grad_norm": 0.603486180305481, + "learning_rate": 2.598e-05, + "loss": 0.0264, + "step": 1300 + }, + { + "epoch": 0.20347934141037588, + "grad_norm": 0.3281537592411041, + "learning_rate": 2.618e-05, + "loss": 0.0263, + "step": 1310 + }, + { + "epoch": 0.20503261882572227, + "grad_norm": 0.6985621452331543, + "learning_rate": 2.6379999999999998e-05, + "loss": 0.0243, + "step": 1320 + }, + { + "epoch": 0.20658589624106866, + "grad_norm": 0.7481710910797119, + "learning_rate": 2.658e-05, + "loss": 0.0218, + "step": 1330 + }, + { + "epoch": 0.20813917365641504, + "grad_norm": 0.6842254996299744, + "learning_rate": 2.678e-05, + "loss": 0.027, + "step": 1340 + }, + { + "epoch": 0.2096924510717614, + "grad_norm": 0.38610678911209106, + "learning_rate": 2.698e-05, + "loss": 0.0237, + "step": 1350 + }, + { + "epoch": 0.2112457284871078, + "grad_norm": 0.4756002426147461, + "learning_rate": 2.718e-05, + "loss": 0.0273, + "step": 1360 + }, + { + "epoch": 0.21279900590245418, + "grad_norm": 0.7587177157402039, + "learning_rate": 2.738e-05, + "loss": 0.0239, + "step": 1370 + }, + { + "epoch": 0.21435228331780057, + "grad_norm": 0.39397162199020386, + "learning_rate": 2.758e-05, + "loss": 0.0247, + "step": 1380 + }, + { + "epoch": 0.21590556073314693, + "grad_norm": 0.439784437417984, + "learning_rate": 2.778e-05, + "loss": 0.0245, + "step": 1390 + }, + { + "epoch": 0.21745883814849332, + "grad_norm": 0.48084908723831177, + "learning_rate": 2.798e-05, + "loss": 0.0248, + "step": 1400 + }, + { + "epoch": 0.2190121155638397, + "grad_norm": 0.44974952936172485, + "learning_rate": 2.818e-05, + "loss": 0.0232, + "step": 1410 + }, + { + "epoch": 0.2205653929791861, + "grad_norm": 0.8996325135231018, + "learning_rate": 2.8380000000000003e-05, + "loss": 0.0252, + "step": 1420 + }, + { + "epoch": 0.22211867039453245, + "grad_norm": 0.8257855772972107, + "learning_rate": 2.858e-05, + "loss": 0.0235, + "step": 1430 + }, + { + "epoch": 0.22367194780987884, + "grad_norm": 0.4148997962474823, + "learning_rate": 2.8780000000000002e-05, + "loss": 0.0224, + "step": 1440 + }, + { + "epoch": 0.22522522522522523, + "grad_norm": 0.6024894714355469, + "learning_rate": 2.898e-05, + "loss": 0.029, + "step": 1450 + }, + { + "epoch": 0.22677850264057162, + "grad_norm": 0.41765958070755005, + "learning_rate": 2.9180000000000002e-05, + "loss": 0.0248, + "step": 1460 + }, + { + "epoch": 0.22833178005591798, + "grad_norm": 0.6799632906913757, + "learning_rate": 2.9380000000000003e-05, + "loss": 0.028, + "step": 1470 + }, + { + "epoch": 0.22988505747126436, + "grad_norm": 0.7635379433631897, + "learning_rate": 2.958e-05, + "loss": 0.0249, + "step": 1480 + }, + { + "epoch": 0.23143833488661075, + "grad_norm": 0.673995316028595, + "learning_rate": 2.9780000000000003e-05, + "loss": 0.0187, + "step": 1490 + }, + { + "epoch": 0.23299161230195714, + "grad_norm": 0.6772050857543945, + "learning_rate": 2.998e-05, + "loss": 0.0341, + "step": 1500 + }, + { + "epoch": 0.2345448897173035, + "grad_norm": 0.651953399181366, + "learning_rate": 3.0180000000000002e-05, + "loss": 0.0215, + "step": 1510 + }, + { + "epoch": 0.2360981671326499, + "grad_norm": 0.41470155119895935, + "learning_rate": 3.0380000000000004e-05, + "loss": 0.025, + "step": 1520 + }, + { + "epoch": 0.23765144454799628, + "grad_norm": 0.6485040187835693, + "learning_rate": 3.058e-05, + "loss": 0.0255, + "step": 1530 + }, + { + "epoch": 0.23920472196334266, + "grad_norm": 0.538318932056427, + "learning_rate": 3.078e-05, + "loss": 0.0306, + "step": 1540 + }, + { + "epoch": 0.24075799937868902, + "grad_norm": 0.39871543645858765, + "learning_rate": 3.0980000000000005e-05, + "loss": 0.028, + "step": 1550 + }, + { + "epoch": 0.2423112767940354, + "grad_norm": 0.4559617340564728, + "learning_rate": 3.118e-05, + "loss": 0.0235, + "step": 1560 + }, + { + "epoch": 0.2438645542093818, + "grad_norm": 0.3996533751487732, + "learning_rate": 3.138e-05, + "loss": 0.0222, + "step": 1570 + }, + { + "epoch": 0.2454178316247282, + "grad_norm": 0.5798981785774231, + "learning_rate": 3.1580000000000006e-05, + "loss": 0.0236, + "step": 1580 + }, + { + "epoch": 0.24697110904007455, + "grad_norm": 0.4334715008735657, + "learning_rate": 3.1780000000000004e-05, + "loss": 0.021, + "step": 1590 + }, + { + "epoch": 0.24852438645542094, + "grad_norm": 0.5486367344856262, + "learning_rate": 3.198e-05, + "loss": 0.0212, + "step": 1600 + }, + { + "epoch": 0.2500776638707673, + "grad_norm": 0.4994879364967346, + "learning_rate": 3.218e-05, + "loss": 0.0231, + "step": 1610 + }, + { + "epoch": 0.2516309412861137, + "grad_norm": 0.3925953805446625, + "learning_rate": 3.238e-05, + "loss": 0.0182, + "step": 1620 + }, + { + "epoch": 0.2531842187014601, + "grad_norm": 0.5672652721405029, + "learning_rate": 3.2579999999999996e-05, + "loss": 0.0204, + "step": 1630 + }, + { + "epoch": 0.25473749611680646, + "grad_norm": 0.7377145886421204, + "learning_rate": 3.278e-05, + "loss": 0.0229, + "step": 1640 + }, + { + "epoch": 0.25629077353215285, + "grad_norm": 0.4721059203147888, + "learning_rate": 3.298e-05, + "loss": 0.0231, + "step": 1650 + }, + { + "epoch": 0.25784405094749924, + "grad_norm": 0.42174381017684937, + "learning_rate": 3.318e-05, + "loss": 0.0179, + "step": 1660 + }, + { + "epoch": 0.2593973283628456, + "grad_norm": 0.5939203500747681, + "learning_rate": 3.338e-05, + "loss": 0.0229, + "step": 1670 + }, + { + "epoch": 0.26095060577819196, + "grad_norm": 0.4673738181591034, + "learning_rate": 3.358e-05, + "loss": 0.0199, + "step": 1680 + }, + { + "epoch": 0.26250388319353835, + "grad_norm": 0.42138174176216125, + "learning_rate": 3.378e-05, + "loss": 0.0229, + "step": 1690 + }, + { + "epoch": 0.26405716060888473, + "grad_norm": 0.5729798078536987, + "learning_rate": 3.398e-05, + "loss": 0.0218, + "step": 1700 + }, + { + "epoch": 0.2656104380242311, + "grad_norm": 0.6073079705238342, + "learning_rate": 3.418e-05, + "loss": 0.0224, + "step": 1710 + }, + { + "epoch": 0.2671637154395775, + "grad_norm": 0.3984299600124359, + "learning_rate": 3.438e-05, + "loss": 0.0218, + "step": 1720 + }, + { + "epoch": 0.2687169928549239, + "grad_norm": 0.48615676164627075, + "learning_rate": 3.4580000000000004e-05, + "loss": 0.0223, + "step": 1730 + }, + { + "epoch": 0.2702702702702703, + "grad_norm": 0.4167632460594177, + "learning_rate": 3.478e-05, + "loss": 0.0254, + "step": 1740 + }, + { + "epoch": 0.2718235476856167, + "grad_norm": 0.47593551874160767, + "learning_rate": 3.498e-05, + "loss": 0.0252, + "step": 1750 + }, + { + "epoch": 0.273376825100963, + "grad_norm": 0.3084787428379059, + "learning_rate": 3.518e-05, + "loss": 0.0216, + "step": 1760 + }, + { + "epoch": 0.2749301025163094, + "grad_norm": 0.507323682308197, + "learning_rate": 3.5380000000000003e-05, + "loss": 0.021, + "step": 1770 + }, + { + "epoch": 0.2764833799316558, + "grad_norm": 0.3943825960159302, + "learning_rate": 3.558e-05, + "loss": 0.021, + "step": 1780 + }, + { + "epoch": 0.27803665734700217, + "grad_norm": 0.31532153487205505, + "learning_rate": 3.578e-05, + "loss": 0.0259, + "step": 1790 + }, + { + "epoch": 0.27958993476234856, + "grad_norm": 0.4185903072357178, + "learning_rate": 3.5980000000000004e-05, + "loss": 0.02, + "step": 1800 + }, + { + "epoch": 0.28114321217769495, + "grad_norm": 0.5347543954849243, + "learning_rate": 3.618e-05, + "loss": 0.0208, + "step": 1810 + }, + { + "epoch": 0.28269648959304133, + "grad_norm": 0.3923000395298004, + "learning_rate": 3.638e-05, + "loss": 0.0167, + "step": 1820 + }, + { + "epoch": 0.2842497670083877, + "grad_norm": 0.3414539694786072, + "learning_rate": 3.6580000000000006e-05, + "loss": 0.0225, + "step": 1830 + }, + { + "epoch": 0.28580304442373405, + "grad_norm": 0.35896119475364685, + "learning_rate": 3.6780000000000004e-05, + "loss": 0.0217, + "step": 1840 + }, + { + "epoch": 0.28735632183908044, + "grad_norm": 0.4595933258533478, + "learning_rate": 3.698e-05, + "loss": 0.0193, + "step": 1850 + }, + { + "epoch": 0.28890959925442683, + "grad_norm": 0.5662511587142944, + "learning_rate": 3.7180000000000007e-05, + "loss": 0.0226, + "step": 1860 + }, + { + "epoch": 0.2904628766697732, + "grad_norm": 0.2846526801586151, + "learning_rate": 3.7380000000000005e-05, + "loss": 0.0218, + "step": 1870 + }, + { + "epoch": 0.2920161540851196, + "grad_norm": 0.4547501802444458, + "learning_rate": 3.758e-05, + "loss": 0.0206, + "step": 1880 + }, + { + "epoch": 0.293569431500466, + "grad_norm": 0.47103607654571533, + "learning_rate": 3.778000000000001e-05, + "loss": 0.0171, + "step": 1890 + }, + { + "epoch": 0.2951227089158124, + "grad_norm": 0.5516520738601685, + "learning_rate": 3.7980000000000006e-05, + "loss": 0.0176, + "step": 1900 + }, + { + "epoch": 0.29667598633115877, + "grad_norm": 0.6942699551582336, + "learning_rate": 3.818e-05, + "loss": 0.0243, + "step": 1910 + }, + { + "epoch": 0.2982292637465051, + "grad_norm": 0.7388509511947632, + "learning_rate": 3.838e-05, + "loss": 0.0182, + "step": 1920 + }, + { + "epoch": 0.2997825411618515, + "grad_norm": 0.6073606610298157, + "learning_rate": 3.858e-05, + "loss": 0.0166, + "step": 1930 + }, + { + "epoch": 0.3013358185771979, + "grad_norm": 0.4602995216846466, + "learning_rate": 3.878e-05, + "loss": 0.024, + "step": 1940 + }, + { + "epoch": 0.30288909599254427, + "grad_norm": 0.43171101808547974, + "learning_rate": 3.898e-05, + "loss": 0.0196, + "step": 1950 + }, + { + "epoch": 0.30444237340789065, + "grad_norm": 0.47005200386047363, + "learning_rate": 3.918e-05, + "loss": 0.0192, + "step": 1960 + }, + { + "epoch": 0.30599565082323704, + "grad_norm": 0.4441761076450348, + "learning_rate": 3.938e-05, + "loss": 0.0209, + "step": 1970 + }, + { + "epoch": 0.30754892823858343, + "grad_norm": 0.456447035074234, + "learning_rate": 3.958e-05, + "loss": 0.025, + "step": 1980 + }, + { + "epoch": 0.3091022056539298, + "grad_norm": 0.4949875771999359, + "learning_rate": 3.978e-05, + "loss": 0.025, + "step": 1990 + }, + { + "epoch": 0.31065548306927615, + "grad_norm": 0.6502448320388794, + "learning_rate": 3.998e-05, + "loss": 0.0218, + "step": 2000 + }, + { + "epoch": 0.31220876048462254, + "grad_norm": 0.4644138514995575, + "learning_rate": 4.018e-05, + "loss": 0.0245, + "step": 2010 + }, + { + "epoch": 0.3137620378999689, + "grad_norm": 0.6796231269836426, + "learning_rate": 4.038e-05, + "loss": 0.0252, + "step": 2020 + }, + { + "epoch": 0.3153153153153153, + "grad_norm": 0.3754326105117798, + "learning_rate": 4.058e-05, + "loss": 0.0202, + "step": 2030 + }, + { + "epoch": 0.3168685927306617, + "grad_norm": 0.5195795297622681, + "learning_rate": 4.078e-05, + "loss": 0.0287, + "step": 2040 + }, + { + "epoch": 0.3184218701460081, + "grad_norm": 0.48421284556388855, + "learning_rate": 4.0980000000000004e-05, + "loss": 0.0227, + "step": 2050 + }, + { + "epoch": 0.3199751475613545, + "grad_norm": 0.5843702554702759, + "learning_rate": 4.118e-05, + "loss": 0.0207, + "step": 2060 + }, + { + "epoch": 0.32152842497670087, + "grad_norm": 0.584635853767395, + "learning_rate": 4.138e-05, + "loss": 0.0203, + "step": 2070 + }, + { + "epoch": 0.3230817023920472, + "grad_norm": 0.3669419288635254, + "learning_rate": 4.1580000000000005e-05, + "loss": 0.0214, + "step": 2080 + }, + { + "epoch": 0.3246349798073936, + "grad_norm": 0.565021812915802, + "learning_rate": 4.178e-05, + "loss": 0.0195, + "step": 2090 + }, + { + "epoch": 0.32618825722274, + "grad_norm": 0.5717247724533081, + "learning_rate": 4.198e-05, + "loss": 0.0222, + "step": 2100 + }, + { + "epoch": 0.32774153463808636, + "grad_norm": 0.35646867752075195, + "learning_rate": 4.2180000000000006e-05, + "loss": 0.0208, + "step": 2110 + }, + { + "epoch": 0.32929481205343275, + "grad_norm": 0.5235041975975037, + "learning_rate": 4.2380000000000004e-05, + "loss": 0.0193, + "step": 2120 + }, + { + "epoch": 0.33084808946877914, + "grad_norm": 0.27224472165107727, + "learning_rate": 4.258e-05, + "loss": 0.0188, + "step": 2130 + }, + { + "epoch": 0.3324013668841255, + "grad_norm": 0.5316774249076843, + "learning_rate": 4.278e-05, + "loss": 0.0154, + "step": 2140 + }, + { + "epoch": 0.33395464429947186, + "grad_norm": 0.35477688908576965, + "learning_rate": 4.2980000000000005e-05, + "loss": 0.0233, + "step": 2150 + }, + { + "epoch": 0.33550792171481825, + "grad_norm": 0.48373332619667053, + "learning_rate": 4.318e-05, + "loss": 0.0217, + "step": 2160 + }, + { + "epoch": 0.33706119913016463, + "grad_norm": 0.3084259033203125, + "learning_rate": 4.338e-05, + "loss": 0.0236, + "step": 2170 + }, + { + "epoch": 0.338614476545511, + "grad_norm": 0.48776939511299133, + "learning_rate": 4.3580000000000006e-05, + "loss": 0.0196, + "step": 2180 + }, + { + "epoch": 0.3401677539608574, + "grad_norm": 0.4808599352836609, + "learning_rate": 4.3780000000000004e-05, + "loss": 0.0175, + "step": 2190 + }, + { + "epoch": 0.3417210313762038, + "grad_norm": 0.2790966331958771, + "learning_rate": 4.398e-05, + "loss": 0.0251, + "step": 2200 + }, + { + "epoch": 0.3432743087915502, + "grad_norm": 0.3371812403202057, + "learning_rate": 4.418000000000001e-05, + "loss": 0.019, + "step": 2210 + }, + { + "epoch": 0.3448275862068966, + "grad_norm": 0.561122477054596, + "learning_rate": 4.438e-05, + "loss": 0.0201, + "step": 2220 + }, + { + "epoch": 0.3463808636222429, + "grad_norm": 0.5104467868804932, + "learning_rate": 4.458e-05, + "loss": 0.0252, + "step": 2230 + }, + { + "epoch": 0.3479341410375893, + "grad_norm": 0.32455939054489136, + "learning_rate": 4.478e-05, + "loss": 0.0189, + "step": 2240 + }, + { + "epoch": 0.3494874184529357, + "grad_norm": 0.2533765733242035, + "learning_rate": 4.498e-05, + "loss": 0.0246, + "step": 2250 + }, + { + "epoch": 0.35104069586828207, + "grad_norm": 0.46057406067848206, + "learning_rate": 4.518e-05, + "loss": 0.0216, + "step": 2260 + }, + { + "epoch": 0.35259397328362846, + "grad_norm": 0.503017783164978, + "learning_rate": 4.538e-05, + "loss": 0.0171, + "step": 2270 + }, + { + "epoch": 0.35414725069897485, + "grad_norm": 0.6312004923820496, + "learning_rate": 4.558e-05, + "loss": 0.0233, + "step": 2280 + }, + { + "epoch": 0.35570052811432123, + "grad_norm": 0.6088271737098694, + "learning_rate": 4.578e-05, + "loss": 0.0224, + "step": 2290 + }, + { + "epoch": 0.3572538055296676, + "grad_norm": 0.41161689162254333, + "learning_rate": 4.5980000000000004e-05, + "loss": 0.0183, + "step": 2300 + }, + { + "epoch": 0.35880708294501396, + "grad_norm": 0.4746848940849304, + "learning_rate": 4.618e-05, + "loss": 0.0208, + "step": 2310 + }, + { + "epoch": 0.36036036036036034, + "grad_norm": 0.3317146599292755, + "learning_rate": 4.638e-05, + "loss": 0.0202, + "step": 2320 + }, + { + "epoch": 0.36191363777570673, + "grad_norm": 0.3712039589881897, + "learning_rate": 4.6580000000000005e-05, + "loss": 0.0242, + "step": 2330 + }, + { + "epoch": 0.3634669151910531, + "grad_norm": 0.26181116700172424, + "learning_rate": 4.678e-05, + "loss": 0.0167, + "step": 2340 + }, + { + "epoch": 0.3650201926063995, + "grad_norm": 0.3496483266353607, + "learning_rate": 4.698e-05, + "loss": 0.0269, + "step": 2350 + }, + { + "epoch": 0.3665734700217459, + "grad_norm": 0.5396166443824768, + "learning_rate": 4.718e-05, + "loss": 0.0226, + "step": 2360 + }, + { + "epoch": 0.3681267474370923, + "grad_norm": 0.4764587879180908, + "learning_rate": 4.7380000000000004e-05, + "loss": 0.0296, + "step": 2370 + }, + { + "epoch": 0.36968002485243867, + "grad_norm": 0.39688241481781006, + "learning_rate": 4.758e-05, + "loss": 0.0167, + "step": 2380 + }, + { + "epoch": 0.371233302267785, + "grad_norm": 0.4072597622871399, + "learning_rate": 4.778e-05, + "loss": 0.0215, + "step": 2390 + }, + { + "epoch": 0.3727865796831314, + "grad_norm": 0.614445686340332, + "learning_rate": 4.7980000000000005e-05, + "loss": 0.0209, + "step": 2400 + }, + { + "epoch": 0.3743398570984778, + "grad_norm": 0.5894325375556946, + "learning_rate": 4.818e-05, + "loss": 0.0182, + "step": 2410 + }, + { + "epoch": 0.37589313451382417, + "grad_norm": 0.2753789722919464, + "learning_rate": 4.838e-05, + "loss": 0.0171, + "step": 2420 + }, + { + "epoch": 0.37744641192917056, + "grad_norm": 0.49310123920440674, + "learning_rate": 4.8580000000000006e-05, + "loss": 0.0205, + "step": 2430 + }, + { + "epoch": 0.37899968934451694, + "grad_norm": 0.34585824608802795, + "learning_rate": 4.8780000000000004e-05, + "loss": 0.0261, + "step": 2440 + }, + { + "epoch": 0.38055296675986333, + "grad_norm": 0.3830241858959198, + "learning_rate": 4.898e-05, + "loss": 0.013, + "step": 2450 + }, + { + "epoch": 0.3821062441752097, + "grad_norm": 0.5801122784614563, + "learning_rate": 4.918000000000001e-05, + "loss": 0.0198, + "step": 2460 + }, + { + "epoch": 0.38365952159055605, + "grad_norm": 0.5538902878761292, + "learning_rate": 4.9380000000000005e-05, + "loss": 0.0196, + "step": 2470 + }, + { + "epoch": 0.38521279900590244, + "grad_norm": 0.5190449357032776, + "learning_rate": 4.958e-05, + "loss": 0.0199, + "step": 2480 + }, + { + "epoch": 0.38676607642124883, + "grad_norm": 0.487604558467865, + "learning_rate": 4.978e-05, + "loss": 0.0168, + "step": 2490 + }, + { + "epoch": 0.3883193538365952, + "grad_norm": 0.4221518933773041, + "learning_rate": 4.9980000000000006e-05, + "loss": 0.0224, + "step": 2500 + }, + { + "epoch": 0.3898726312519416, + "grad_norm": 0.2814960777759552, + "learning_rate": 5.0180000000000004e-05, + "loss": 0.0179, + "step": 2510 + }, + { + "epoch": 0.391425908667288, + "grad_norm": 0.3462064564228058, + "learning_rate": 5.038e-05, + "loss": 0.0185, + "step": 2520 + }, + { + "epoch": 0.3929791860826344, + "grad_norm": 0.3584158420562744, + "learning_rate": 5.058000000000001e-05, + "loss": 0.0203, + "step": 2530 + }, + { + "epoch": 0.3945324634979807, + "grad_norm": 0.4177609980106354, + "learning_rate": 5.0780000000000005e-05, + "loss": 0.0257, + "step": 2540 + }, + { + "epoch": 0.3960857409133271, + "grad_norm": 0.5184181928634644, + "learning_rate": 5.098e-05, + "loss": 0.0229, + "step": 2550 + }, + { + "epoch": 0.3976390183286735, + "grad_norm": 0.3848685026168823, + "learning_rate": 5.118000000000001e-05, + "loss": 0.0205, + "step": 2560 + }, + { + "epoch": 0.3991922957440199, + "grad_norm": 0.3953237533569336, + "learning_rate": 5.1380000000000006e-05, + "loss": 0.0211, + "step": 2570 + }, + { + "epoch": 0.40074557315936626, + "grad_norm": 0.2951180636882782, + "learning_rate": 5.1580000000000004e-05, + "loss": 0.0185, + "step": 2580 + }, + { + "epoch": 0.40229885057471265, + "grad_norm": 0.4956471621990204, + "learning_rate": 5.178000000000001e-05, + "loss": 0.0185, + "step": 2590 + }, + { + "epoch": 0.40385212799005904, + "grad_norm": 0.689478874206543, + "learning_rate": 5.198000000000001e-05, + "loss": 0.0213, + "step": 2600 + }, + { + "epoch": 0.40540540540540543, + "grad_norm": 0.38176196813583374, + "learning_rate": 5.2180000000000005e-05, + "loss": 0.0189, + "step": 2610 + }, + { + "epoch": 0.40695868282075176, + "grad_norm": 0.45360425114631653, + "learning_rate": 5.238000000000001e-05, + "loss": 0.0234, + "step": 2620 + }, + { + "epoch": 0.40851196023609815, + "grad_norm": 0.36548545956611633, + "learning_rate": 5.258000000000001e-05, + "loss": 0.0158, + "step": 2630 + }, + { + "epoch": 0.41006523765144454, + "grad_norm": 0.34779226779937744, + "learning_rate": 5.2780000000000006e-05, + "loss": 0.0178, + "step": 2640 + }, + { + "epoch": 0.4116185150667909, + "grad_norm": 0.34819281101226807, + "learning_rate": 5.2980000000000004e-05, + "loss": 0.0178, + "step": 2650 + }, + { + "epoch": 0.4131717924821373, + "grad_norm": 0.32191672921180725, + "learning_rate": 5.318000000000001e-05, + "loss": 0.0232, + "step": 2660 + }, + { + "epoch": 0.4147250698974837, + "grad_norm": 0.23577529191970825, + "learning_rate": 5.338000000000001e-05, + "loss": 0.0179, + "step": 2670 + }, + { + "epoch": 0.4162783473128301, + "grad_norm": 0.3067491054534912, + "learning_rate": 5.3580000000000005e-05, + "loss": 0.0233, + "step": 2680 + }, + { + "epoch": 0.4178316247281765, + "grad_norm": 0.3084315359592438, + "learning_rate": 5.378e-05, + "loss": 0.018, + "step": 2690 + }, + { + "epoch": 0.4193849021435228, + "grad_norm": 0.40202921628952026, + "learning_rate": 5.3979999999999995e-05, + "loss": 0.0162, + "step": 2700 + }, + { + "epoch": 0.4209381795588692, + "grad_norm": 0.2072431892156601, + "learning_rate": 5.418e-05, + "loss": 0.0183, + "step": 2710 + }, + { + "epoch": 0.4224914569742156, + "grad_norm": 0.3502729833126068, + "learning_rate": 5.438e-05, + "loss": 0.0198, + "step": 2720 + }, + { + "epoch": 0.424044734389562, + "grad_norm": 0.37454554438591003, + "learning_rate": 5.4579999999999996e-05, + "loss": 0.0153, + "step": 2730 + }, + { + "epoch": 0.42559801180490836, + "grad_norm": 0.3113163709640503, + "learning_rate": 5.478e-05, + "loss": 0.017, + "step": 2740 + }, + { + "epoch": 0.42715128922025475, + "grad_norm": 0.32778725028038025, + "learning_rate": 5.498e-05, + "loss": 0.0145, + "step": 2750 + }, + { + "epoch": 0.42870456663560114, + "grad_norm": 0.3498435318470001, + "learning_rate": 5.518e-05, + "loss": 0.0133, + "step": 2760 + }, + { + "epoch": 0.4302578440509475, + "grad_norm": 0.3738687336444855, + "learning_rate": 5.538e-05, + "loss": 0.0132, + "step": 2770 + }, + { + "epoch": 0.43181112146629386, + "grad_norm": 0.3395034372806549, + "learning_rate": 5.558e-05, + "loss": 0.0191, + "step": 2780 + }, + { + "epoch": 0.43336439888164024, + "grad_norm": 0.3324732780456543, + "learning_rate": 5.578e-05, + "loss": 0.0158, + "step": 2790 + }, + { + "epoch": 0.43491767629698663, + "grad_norm": 0.521379828453064, + "learning_rate": 5.5979999999999996e-05, + "loss": 0.0203, + "step": 2800 + }, + { + "epoch": 0.436470953712333, + "grad_norm": 0.3804153800010681, + "learning_rate": 5.618e-05, + "loss": 0.0194, + "step": 2810 + }, + { + "epoch": 0.4380242311276794, + "grad_norm": 0.39825838804244995, + "learning_rate": 5.638e-05, + "loss": 0.0181, + "step": 2820 + }, + { + "epoch": 0.4395775085430258, + "grad_norm": 0.5336761474609375, + "learning_rate": 5.658e-05, + "loss": 0.0229, + "step": 2830 + }, + { + "epoch": 0.4411307859583722, + "grad_norm": 0.4386591613292694, + "learning_rate": 5.678e-05, + "loss": 0.0226, + "step": 2840 + }, + { + "epoch": 0.4426840633737186, + "grad_norm": 0.5435610413551331, + "learning_rate": 5.698e-05, + "loss": 0.0209, + "step": 2850 + }, + { + "epoch": 0.4442373407890649, + "grad_norm": 0.43159303069114685, + "learning_rate": 5.718e-05, + "loss": 0.0233, + "step": 2860 + }, + { + "epoch": 0.4457906182044113, + "grad_norm": 0.482189416885376, + "learning_rate": 5.738e-05, + "loss": 0.0195, + "step": 2870 + }, + { + "epoch": 0.4473438956197577, + "grad_norm": 0.668435275554657, + "learning_rate": 5.758e-05, + "loss": 0.0298, + "step": 2880 + }, + { + "epoch": 0.44889717303510407, + "grad_norm": 0.4698132276535034, + "learning_rate": 5.778e-05, + "loss": 0.0241, + "step": 2890 + }, + { + "epoch": 0.45045045045045046, + "grad_norm": 0.3543587923049927, + "learning_rate": 5.7980000000000004e-05, + "loss": 0.0244, + "step": 2900 + }, + { + "epoch": 0.45200372786579684, + "grad_norm": 0.43093013763427734, + "learning_rate": 5.818e-05, + "loss": 0.0263, + "step": 2910 + }, + { + "epoch": 0.45355700528114323, + "grad_norm": 0.41805657744407654, + "learning_rate": 5.838e-05, + "loss": 0.0246, + "step": 2920 + }, + { + "epoch": 0.45511028269648957, + "grad_norm": 0.3754327595233917, + "learning_rate": 5.858e-05, + "loss": 0.0213, + "step": 2930 + }, + { + "epoch": 0.45666356011183595, + "grad_norm": 0.40588903427124023, + "learning_rate": 5.878e-05, + "loss": 0.0189, + "step": 2940 + }, + { + "epoch": 0.45821683752718234, + "grad_norm": 0.608607292175293, + "learning_rate": 5.898e-05, + "loss": 0.0209, + "step": 2950 + }, + { + "epoch": 0.45977011494252873, + "grad_norm": 0.6140630841255188, + "learning_rate": 5.918e-05, + "loss": 0.0232, + "step": 2960 + }, + { + "epoch": 0.4613233923578751, + "grad_norm": 0.48643240332603455, + "learning_rate": 5.9380000000000004e-05, + "loss": 0.0198, + "step": 2970 + }, + { + "epoch": 0.4628766697732215, + "grad_norm": 0.5996144413948059, + "learning_rate": 5.958e-05, + "loss": 0.019, + "step": 2980 + }, + { + "epoch": 0.4644299471885679, + "grad_norm": 0.3854431211948395, + "learning_rate": 5.978e-05, + "loss": 0.0203, + "step": 2990 + }, + { + "epoch": 0.4659832246039143, + "grad_norm": 0.5484134554862976, + "learning_rate": 5.9980000000000005e-05, + "loss": 0.0226, + "step": 3000 + }, + { + "epoch": 0.4675365020192606, + "grad_norm": 0.33298459649086, + "learning_rate": 6.018e-05, + "loss": 0.0192, + "step": 3010 + }, + { + "epoch": 0.469089779434607, + "grad_norm": 0.5966723561286926, + "learning_rate": 6.038e-05, + "loss": 0.0199, + "step": 3020 + }, + { + "epoch": 0.4706430568499534, + "grad_norm": 0.5782431364059448, + "learning_rate": 6.0580000000000006e-05, + "loss": 0.0242, + "step": 3030 + }, + { + "epoch": 0.4721963342652998, + "grad_norm": 0.34788626432418823, + "learning_rate": 6.0780000000000004e-05, + "loss": 0.0198, + "step": 3040 + }, + { + "epoch": 0.47374961168064617, + "grad_norm": 0.5638571381568909, + "learning_rate": 6.098e-05, + "loss": 0.0246, + "step": 3050 + }, + { + "epoch": 0.47530288909599255, + "grad_norm": 0.3690008521080017, + "learning_rate": 6.118000000000001e-05, + "loss": 0.0225, + "step": 3060 + }, + { + "epoch": 0.47685616651133894, + "grad_norm": 0.4141135811805725, + "learning_rate": 6.138e-05, + "loss": 0.0234, + "step": 3070 + }, + { + "epoch": 0.47840944392668533, + "grad_norm": 0.31733399629592896, + "learning_rate": 6.158e-05, + "loss": 0.0178, + "step": 3080 + }, + { + "epoch": 0.47996272134203166, + "grad_norm": 0.3163086175918579, + "learning_rate": 6.178000000000001e-05, + "loss": 0.0199, + "step": 3090 + }, + { + "epoch": 0.48151599875737805, + "grad_norm": 0.5276801586151123, + "learning_rate": 6.198e-05, + "loss": 0.0191, + "step": 3100 + }, + { + "epoch": 0.48306927617272444, + "grad_norm": 0.47531530261039734, + "learning_rate": 6.218e-05, + "loss": 0.0207, + "step": 3110 + }, + { + "epoch": 0.4846225535880708, + "grad_norm": 0.33650848269462585, + "learning_rate": 6.238000000000001e-05, + "loss": 0.0141, + "step": 3120 + }, + { + "epoch": 0.4861758310034172, + "grad_norm": 0.3144567012786865, + "learning_rate": 6.258e-05, + "loss": 0.0184, + "step": 3130 + }, + { + "epoch": 0.4877291084187636, + "grad_norm": 0.39560994505882263, + "learning_rate": 6.278e-05, + "loss": 0.0168, + "step": 3140 + }, + { + "epoch": 0.48928238583411, + "grad_norm": 0.361379474401474, + "learning_rate": 6.298000000000001e-05, + "loss": 0.0189, + "step": 3150 + }, + { + "epoch": 0.4908356632494564, + "grad_norm": 0.3056003451347351, + "learning_rate": 6.318e-05, + "loss": 0.0176, + "step": 3160 + }, + { + "epoch": 0.4923889406648027, + "grad_norm": 0.4193875789642334, + "learning_rate": 6.338e-05, + "loss": 0.0199, + "step": 3170 + }, + { + "epoch": 0.4939422180801491, + "grad_norm": 0.4213220179080963, + "learning_rate": 6.358000000000001e-05, + "loss": 0.0212, + "step": 3180 + }, + { + "epoch": 0.4954954954954955, + "grad_norm": 0.4388089179992676, + "learning_rate": 6.378e-05, + "loss": 0.0277, + "step": 3190 + }, + { + "epoch": 0.4970487729108419, + "grad_norm": 0.6495473384857178, + "learning_rate": 6.398000000000001e-05, + "loss": 0.0235, + "step": 3200 + }, + { + "epoch": 0.49860205032618826, + "grad_norm": 0.36453118920326233, + "learning_rate": 6.418000000000001e-05, + "loss": 0.0219, + "step": 3210 + }, + { + "epoch": 0.5001553277415346, + "grad_norm": 0.35854142904281616, + "learning_rate": 6.438e-05, + "loss": 0.0194, + "step": 3220 + }, + { + "epoch": 0.501708605156881, + "grad_norm": 0.37633413076400757, + "learning_rate": 6.458000000000001e-05, + "loss": 0.0175, + "step": 3230 + }, + { + "epoch": 0.5032618825722274, + "grad_norm": 0.4141024351119995, + "learning_rate": 6.478000000000001e-05, + "loss": 0.0148, + "step": 3240 + }, + { + "epoch": 0.5048151599875738, + "grad_norm": 0.22727330029010773, + "learning_rate": 6.498e-05, + "loss": 0.0141, + "step": 3250 + }, + { + "epoch": 0.5063684374029201, + "grad_norm": 0.5006093978881836, + "learning_rate": 6.518000000000001e-05, + "loss": 0.0196, + "step": 3260 + }, + { + "epoch": 0.5079217148182665, + "grad_norm": 0.5491776466369629, + "learning_rate": 6.538000000000001e-05, + "loss": 0.0206, + "step": 3270 + }, + { + "epoch": 0.5094749922336129, + "grad_norm": 0.2808894217014313, + "learning_rate": 6.558e-05, + "loss": 0.0154, + "step": 3280 + }, + { + "epoch": 0.5110282696489593, + "grad_norm": 0.4756491482257843, + "learning_rate": 6.578000000000001e-05, + "loss": 0.0175, + "step": 3290 + }, + { + "epoch": 0.5125815470643057, + "grad_norm": 0.48110395669937134, + "learning_rate": 6.598e-05, + "loss": 0.0205, + "step": 3300 + }, + { + "epoch": 0.5141348244796521, + "grad_norm": 0.41848501563072205, + "learning_rate": 6.618e-05, + "loss": 0.0231, + "step": 3310 + }, + { + "epoch": 0.5156881018949985, + "grad_norm": 0.7108784914016724, + "learning_rate": 6.638e-05, + "loss": 0.0201, + "step": 3320 + }, + { + "epoch": 0.5172413793103449, + "grad_norm": 0.42968615889549255, + "learning_rate": 6.658e-05, + "loss": 0.026, + "step": 3330 + }, + { + "epoch": 0.5187946567256913, + "grad_norm": 0.5412790179252625, + "learning_rate": 6.678e-05, + "loss": 0.0192, + "step": 3340 + }, + { + "epoch": 0.5203479341410376, + "grad_norm": 0.3705042004585266, + "learning_rate": 6.698e-05, + "loss": 0.0243, + "step": 3350 + }, + { + "epoch": 0.5219012115563839, + "grad_norm": 0.3972005844116211, + "learning_rate": 6.718e-05, + "loss": 0.0174, + "step": 3360 + }, + { + "epoch": 0.5234544889717303, + "grad_norm": 0.431543231010437, + "learning_rate": 6.738e-05, + "loss": 0.0228, + "step": 3370 + }, + { + "epoch": 0.5250077663870767, + "grad_norm": 0.401349812746048, + "learning_rate": 6.758e-05, + "loss": 0.0186, + "step": 3380 + }, + { + "epoch": 0.5265610438024231, + "grad_norm": 0.5084168910980225, + "learning_rate": 6.778e-05, + "loss": 0.026, + "step": 3390 + }, + { + "epoch": 0.5281143212177695, + "grad_norm": 0.5007063150405884, + "learning_rate": 6.798e-05, + "loss": 0.0216, + "step": 3400 + }, + { + "epoch": 0.5296675986331159, + "grad_norm": 0.5127500295639038, + "learning_rate": 6.818e-05, + "loss": 0.0181, + "step": 3410 + }, + { + "epoch": 0.5312208760484622, + "grad_norm": 0.3325512111186981, + "learning_rate": 6.838e-05, + "loss": 0.0172, + "step": 3420 + }, + { + "epoch": 0.5327741534638086, + "grad_norm": 0.4098396301269531, + "learning_rate": 6.858e-05, + "loss": 0.0194, + "step": 3430 + }, + { + "epoch": 0.534327430879155, + "grad_norm": 0.43595871329307556, + "learning_rate": 6.878e-05, + "loss": 0.0208, + "step": 3440 + }, + { + "epoch": 0.5358807082945014, + "grad_norm": 0.3599027097225189, + "learning_rate": 6.898e-05, + "loss": 0.0186, + "step": 3450 + }, + { + "epoch": 0.5374339857098478, + "grad_norm": 0.34921860694885254, + "learning_rate": 6.918e-05, + "loss": 0.02, + "step": 3460 + }, + { + "epoch": 0.5389872631251942, + "grad_norm": 0.42308565974235535, + "learning_rate": 6.938e-05, + "loss": 0.0216, + "step": 3470 + }, + { + "epoch": 0.5405405405405406, + "grad_norm": 0.43401485681533813, + "learning_rate": 6.958e-05, + "loss": 0.0187, + "step": 3480 + }, + { + "epoch": 0.542093817955887, + "grad_norm": 0.44572073221206665, + "learning_rate": 6.978e-05, + "loss": 0.0169, + "step": 3490 + }, + { + "epoch": 0.5436470953712333, + "grad_norm": 0.5118770003318787, + "learning_rate": 6.998e-05, + "loss": 0.0209, + "step": 3500 + }, + { + "epoch": 0.5452003727865797, + "grad_norm": 0.33984583616256714, + "learning_rate": 7.018e-05, + "loss": 0.0202, + "step": 3510 + }, + { + "epoch": 0.546753650201926, + "grad_norm": 0.4228861331939697, + "learning_rate": 7.038e-05, + "loss": 0.0272, + "step": 3520 + }, + { + "epoch": 0.5483069276172724, + "grad_norm": 0.34555670619010925, + "learning_rate": 7.058e-05, + "loss": 0.018, + "step": 3530 + }, + { + "epoch": 0.5498602050326188, + "grad_norm": 0.21500937640666962, + "learning_rate": 7.078e-05, + "loss": 0.0205, + "step": 3540 + }, + { + "epoch": 0.5514134824479652, + "grad_norm": 0.3815748393535614, + "learning_rate": 7.098e-05, + "loss": 0.0188, + "step": 3550 + }, + { + "epoch": 0.5529667598633116, + "grad_norm": 0.3594578504562378, + "learning_rate": 7.118e-05, + "loss": 0.0161, + "step": 3560 + }, + { + "epoch": 0.554520037278658, + "grad_norm": 0.382590651512146, + "learning_rate": 7.138e-05, + "loss": 0.0164, + "step": 3570 + }, + { + "epoch": 0.5560733146940043, + "grad_norm": 0.31296059489250183, + "learning_rate": 7.158e-05, + "loss": 0.0172, + "step": 3580 + }, + { + "epoch": 0.5576265921093507, + "grad_norm": 0.25910648703575134, + "learning_rate": 7.178000000000001e-05, + "loss": 0.0147, + "step": 3590 + }, + { + "epoch": 0.5591798695246971, + "grad_norm": 0.5239319801330566, + "learning_rate": 7.198e-05, + "loss": 0.0202, + "step": 3600 + }, + { + "epoch": 0.5607331469400435, + "grad_norm": 0.3936024606227875, + "learning_rate": 7.218e-05, + "loss": 0.0192, + "step": 3610 + }, + { + "epoch": 0.5622864243553899, + "grad_norm": 0.39537519216537476, + "learning_rate": 7.238000000000001e-05, + "loss": 0.0167, + "step": 3620 + }, + { + "epoch": 0.5638397017707363, + "grad_norm": 0.34344062209129333, + "learning_rate": 7.258e-05, + "loss": 0.0152, + "step": 3630 + }, + { + "epoch": 0.5653929791860827, + "grad_norm": 0.5325593948364258, + "learning_rate": 7.278e-05, + "loss": 0.0234, + "step": 3640 + }, + { + "epoch": 0.566946256601429, + "grad_norm": 0.4423092007637024, + "learning_rate": 7.298000000000001e-05, + "loss": 0.0184, + "step": 3650 + }, + { + "epoch": 0.5684995340167754, + "grad_norm": 0.3030710518360138, + "learning_rate": 7.318e-05, + "loss": 0.0166, + "step": 3660 + }, + { + "epoch": 0.5700528114321218, + "grad_norm": 0.3966892957687378, + "learning_rate": 7.338e-05, + "loss": 0.0169, + "step": 3670 + }, + { + "epoch": 0.5716060888474681, + "grad_norm": 0.4060599207878113, + "learning_rate": 7.358000000000001e-05, + "loss": 0.0196, + "step": 3680 + }, + { + "epoch": 0.5731593662628145, + "grad_norm": 0.3652462363243103, + "learning_rate": 7.378e-05, + "loss": 0.0196, + "step": 3690 + }, + { + "epoch": 0.5747126436781609, + "grad_norm": 0.4788101315498352, + "learning_rate": 7.398e-05, + "loss": 0.0199, + "step": 3700 + }, + { + "epoch": 0.5762659210935073, + "grad_norm": 0.21746200323104858, + "learning_rate": 7.418000000000001e-05, + "loss": 0.0201, + "step": 3710 + }, + { + "epoch": 0.5778191985088537, + "grad_norm": 0.5130853652954102, + "learning_rate": 7.438e-05, + "loss": 0.0187, + "step": 3720 + }, + { + "epoch": 0.5793724759242, + "grad_norm": 0.2750910520553589, + "learning_rate": 7.458000000000001e-05, + "loss": 0.017, + "step": 3730 + }, + { + "epoch": 0.5809257533395464, + "grad_norm": 0.37744763493537903, + "learning_rate": 7.478e-05, + "loss": 0.0204, + "step": 3740 + }, + { + "epoch": 0.5824790307548928, + "grad_norm": 0.3738090693950653, + "learning_rate": 7.498e-05, + "loss": 0.0229, + "step": 3750 + }, + { + "epoch": 0.5840323081702392, + "grad_norm": 0.3108985722064972, + "learning_rate": 7.518000000000001e-05, + "loss": 0.0197, + "step": 3760 + }, + { + "epoch": 0.5855855855855856, + "grad_norm": 0.3083738684654236, + "learning_rate": 7.538e-05, + "loss": 0.0205, + "step": 3770 + }, + { + "epoch": 0.587138863000932, + "grad_norm": 0.5192537307739258, + "learning_rate": 7.558e-05, + "loss": 0.0209, + "step": 3780 + }, + { + "epoch": 0.5886921404162784, + "grad_norm": 0.5239286422729492, + "learning_rate": 7.578000000000001e-05, + "loss": 0.0213, + "step": 3790 + }, + { + "epoch": 0.5902454178316248, + "grad_norm": 0.43401089310646057, + "learning_rate": 7.598e-05, + "loss": 0.0222, + "step": 3800 + }, + { + "epoch": 0.5917986952469712, + "grad_norm": 0.5793190598487854, + "learning_rate": 7.618e-05, + "loss": 0.0174, + "step": 3810 + }, + { + "epoch": 0.5933519726623175, + "grad_norm": 0.32581451535224915, + "learning_rate": 7.638000000000001e-05, + "loss": 0.0215, + "step": 3820 + }, + { + "epoch": 0.5949052500776638, + "grad_norm": 0.5040780305862427, + "learning_rate": 7.658e-05, + "loss": 0.0206, + "step": 3830 + }, + { + "epoch": 0.5964585274930102, + "grad_norm": 0.3911091387271881, + "learning_rate": 7.678000000000001e-05, + "loss": 0.0189, + "step": 3840 + }, + { + "epoch": 0.5980118049083566, + "grad_norm": 0.24763505160808563, + "learning_rate": 7.698000000000001e-05, + "loss": 0.0177, + "step": 3850 + }, + { + "epoch": 0.599565082323703, + "grad_norm": 0.2768484354019165, + "learning_rate": 7.718e-05, + "loss": 0.0195, + "step": 3860 + }, + { + "epoch": 0.6011183597390494, + "grad_norm": 0.6096818447113037, + "learning_rate": 7.738000000000001e-05, + "loss": 0.0266, + "step": 3870 + }, + { + "epoch": 0.6026716371543958, + "grad_norm": 0.4644545316696167, + "learning_rate": 7.758000000000001e-05, + "loss": 0.0206, + "step": 3880 + }, + { + "epoch": 0.6042249145697421, + "grad_norm": 0.5047130584716797, + "learning_rate": 7.778e-05, + "loss": 0.0188, + "step": 3890 + }, + { + "epoch": 0.6057781919850885, + "grad_norm": 0.5871754884719849, + "learning_rate": 7.798000000000001e-05, + "loss": 0.0226, + "step": 3900 + }, + { + "epoch": 0.6073314694004349, + "grad_norm": 0.47881296277046204, + "learning_rate": 7.818000000000001e-05, + "loss": 0.0198, + "step": 3910 + }, + { + "epoch": 0.6088847468157813, + "grad_norm": 0.39559173583984375, + "learning_rate": 7.838e-05, + "loss": 0.0159, + "step": 3920 + }, + { + "epoch": 0.6104380242311277, + "grad_norm": 0.2875054180622101, + "learning_rate": 7.858000000000001e-05, + "loss": 0.0184, + "step": 3930 + }, + { + "epoch": 0.6119913016464741, + "grad_norm": 0.5121833086013794, + "learning_rate": 7.878e-05, + "loss": 0.019, + "step": 3940 + }, + { + "epoch": 0.6135445790618205, + "grad_norm": 0.3585139214992523, + "learning_rate": 7.897999999999999e-05, + "loss": 0.0214, + "step": 3950 + }, + { + "epoch": 0.6150978564771669, + "grad_norm": 0.30111926794052124, + "learning_rate": 7.918e-05, + "loss": 0.0164, + "step": 3960 + }, + { + "epoch": 0.6166511338925132, + "grad_norm": 0.25344574451446533, + "learning_rate": 7.938e-05, + "loss": 0.0182, + "step": 3970 + }, + { + "epoch": 0.6182044113078596, + "grad_norm": 0.5045255422592163, + "learning_rate": 7.958e-05, + "loss": 0.0194, + "step": 3980 + }, + { + "epoch": 0.6197576887232059, + "grad_norm": 0.4531000852584839, + "learning_rate": 7.978e-05, + "loss": 0.018, + "step": 3990 + }, + { + "epoch": 0.6213109661385523, + "grad_norm": 0.44806063175201416, + "learning_rate": 7.998e-05, + "loss": 0.0219, + "step": 4000 + }, + { + "epoch": 0.6228642435538987, + "grad_norm": 0.37562626600265503, + "learning_rate": 8.018e-05, + "loss": 0.0246, + "step": 4010 + }, + { + "epoch": 0.6244175209692451, + "grad_norm": 0.46248066425323486, + "learning_rate": 8.038e-05, + "loss": 0.0181, + "step": 4020 + }, + { + "epoch": 0.6259707983845915, + "grad_norm": 0.4201142191886902, + "learning_rate": 8.058e-05, + "loss": 0.0209, + "step": 4030 + }, + { + "epoch": 0.6275240757999379, + "grad_norm": 0.4213736951351166, + "learning_rate": 8.078e-05, + "loss": 0.0214, + "step": 4040 + }, + { + "epoch": 0.6290773532152842, + "grad_norm": 0.4433459937572479, + "learning_rate": 8.098e-05, + "loss": 0.0167, + "step": 4050 + }, + { + "epoch": 0.6306306306306306, + "grad_norm": 0.2448280155658722, + "learning_rate": 8.118e-05, + "loss": 0.0184, + "step": 4060 + }, + { + "epoch": 0.632183908045977, + "grad_norm": 0.31338390707969666, + "learning_rate": 8.138e-05, + "loss": 0.0187, + "step": 4070 + }, + { + "epoch": 0.6337371854613234, + "grad_norm": 0.4781652092933655, + "learning_rate": 8.158e-05, + "loss": 0.017, + "step": 4080 + }, + { + "epoch": 0.6352904628766698, + "grad_norm": 0.32836902141571045, + "learning_rate": 8.178e-05, + "loss": 0.0211, + "step": 4090 + }, + { + "epoch": 0.6368437402920162, + "grad_norm": 0.3929983079433441, + "learning_rate": 8.198e-05, + "loss": 0.024, + "step": 4100 + }, + { + "epoch": 0.6383970177073626, + "grad_norm": 0.3772599697113037, + "learning_rate": 8.218e-05, + "loss": 0.0221, + "step": 4110 + }, + { + "epoch": 0.639950295122709, + "grad_norm": 0.47064247727394104, + "learning_rate": 8.238000000000001e-05, + "loss": 0.023, + "step": 4120 + }, + { + "epoch": 0.6415035725380553, + "grad_norm": 0.36506521701812744, + "learning_rate": 8.258e-05, + "loss": 0.0187, + "step": 4130 + }, + { + "epoch": 0.6430568499534017, + "grad_norm": 0.5088617205619812, + "learning_rate": 8.278e-05, + "loss": 0.0164, + "step": 4140 + }, + { + "epoch": 0.644610127368748, + "grad_norm": 0.44516104459762573, + "learning_rate": 8.298000000000001e-05, + "loss": 0.0176, + "step": 4150 + }, + { + "epoch": 0.6461634047840944, + "grad_norm": 0.42838045954704285, + "learning_rate": 8.318e-05, + "loss": 0.0159, + "step": 4160 + }, + { + "epoch": 0.6477166821994408, + "grad_norm": 0.25981611013412476, + "learning_rate": 8.338e-05, + "loss": 0.0194, + "step": 4170 + }, + { + "epoch": 0.6492699596147872, + "grad_norm": 0.34499168395996094, + "learning_rate": 8.358e-05, + "loss": 0.0227, + "step": 4180 + }, + { + "epoch": 0.6508232370301336, + "grad_norm": 0.33225059509277344, + "learning_rate": 8.378e-05, + "loss": 0.0225, + "step": 4190 + }, + { + "epoch": 0.65237651444548, + "grad_norm": 0.3082224428653717, + "learning_rate": 8.398e-05, + "loss": 0.0176, + "step": 4200 + }, + { + "epoch": 0.6539297918608263, + "grad_norm": 0.43976566195487976, + "learning_rate": 8.418e-05, + "loss": 0.0157, + "step": 4210 + }, + { + "epoch": 0.6554830692761727, + "grad_norm": 0.48467379808425903, + "learning_rate": 8.438e-05, + "loss": 0.019, + "step": 4220 + }, + { + "epoch": 0.6570363466915191, + "grad_norm": 0.42133691906929016, + "learning_rate": 8.458e-05, + "loss": 0.0144, + "step": 4230 + }, + { + "epoch": 0.6585896241068655, + "grad_norm": 0.4173579812049866, + "learning_rate": 8.478e-05, + "loss": 0.0201, + "step": 4240 + }, + { + "epoch": 0.6601429015222119, + "grad_norm": 0.40235626697540283, + "learning_rate": 8.498e-05, + "loss": 0.0202, + "step": 4250 + }, + { + "epoch": 0.6616961789375583, + "grad_norm": 0.5169976949691772, + "learning_rate": 8.518000000000001e-05, + "loss": 0.0248, + "step": 4260 + }, + { + "epoch": 0.6632494563529047, + "grad_norm": 0.2483949512243271, + "learning_rate": 8.538e-05, + "loss": 0.0178, + "step": 4270 + }, + { + "epoch": 0.664802733768251, + "grad_norm": 0.3231925368309021, + "learning_rate": 8.558e-05, + "loss": 0.0154, + "step": 4280 + }, + { + "epoch": 0.6663560111835974, + "grad_norm": 0.3675335943698883, + "learning_rate": 8.578000000000001e-05, + "loss": 0.018, + "step": 4290 + }, + { + "epoch": 0.6679092885989437, + "grad_norm": 0.2966977655887604, + "learning_rate": 8.598e-05, + "loss": 0.0192, + "step": 4300 + }, + { + "epoch": 0.6694625660142901, + "grad_norm": 0.29981523752212524, + "learning_rate": 8.618e-05, + "loss": 0.015, + "step": 4310 + }, + { + "epoch": 0.6710158434296365, + "grad_norm": 0.34870874881744385, + "learning_rate": 8.638000000000001e-05, + "loss": 0.0196, + "step": 4320 + }, + { + "epoch": 0.6725691208449829, + "grad_norm": 0.3813667297363281, + "learning_rate": 8.658e-05, + "loss": 0.0193, + "step": 4330 + }, + { + "epoch": 0.6741223982603293, + "grad_norm": 0.36019906401634216, + "learning_rate": 8.678e-05, + "loss": 0.0188, + "step": 4340 + }, + { + "epoch": 0.6756756756756757, + "grad_norm": 0.5204485654830933, + "learning_rate": 8.698000000000001e-05, + "loss": 0.0143, + "step": 4350 + }, + { + "epoch": 0.677228953091022, + "grad_norm": 0.49940329790115356, + "learning_rate": 8.718e-05, + "loss": 0.0142, + "step": 4360 + }, + { + "epoch": 0.6787822305063684, + "grad_norm": 0.3257567286491394, + "learning_rate": 8.738000000000001e-05, + "loss": 0.0166, + "step": 4370 + }, + { + "epoch": 0.6803355079217148, + "grad_norm": 0.407619446516037, + "learning_rate": 8.758000000000001e-05, + "loss": 0.0199, + "step": 4380 + }, + { + "epoch": 0.6818887853370612, + "grad_norm": 0.286165326833725, + "learning_rate": 8.778e-05, + "loss": 0.0203, + "step": 4390 + }, + { + "epoch": 0.6834420627524076, + "grad_norm": 0.3630989193916321, + "learning_rate": 8.798000000000001e-05, + "loss": 0.0246, + "step": 4400 + }, + { + "epoch": 0.684995340167754, + "grad_norm": 0.3156430125236511, + "learning_rate": 8.818000000000001e-05, + "loss": 0.0187, + "step": 4410 + }, + { + "epoch": 0.6865486175831004, + "grad_norm": 0.435322105884552, + "learning_rate": 8.838e-05, + "loss": 0.0176, + "step": 4420 + }, + { + "epoch": 0.6881018949984468, + "grad_norm": 0.3547166585922241, + "learning_rate": 8.858000000000001e-05, + "loss": 0.0215, + "step": 4430 + }, + { + "epoch": 0.6896551724137931, + "grad_norm": 0.439145028591156, + "learning_rate": 8.878000000000001e-05, + "loss": 0.0252, + "step": 4440 + }, + { + "epoch": 0.6912084498291395, + "grad_norm": 0.3124251961708069, + "learning_rate": 8.898e-05, + "loss": 0.0168, + "step": 4450 + }, + { + "epoch": 0.6927617272444858, + "grad_norm": 0.3023993968963623, + "learning_rate": 8.918000000000001e-05, + "loss": 0.0161, + "step": 4460 + }, + { + "epoch": 0.6943150046598322, + "grad_norm": 0.42634400725364685, + "learning_rate": 8.938e-05, + "loss": 0.0186, + "step": 4470 + }, + { + "epoch": 0.6958682820751786, + "grad_norm": 0.3319787383079529, + "learning_rate": 8.958e-05, + "loss": 0.0213, + "step": 4480 + }, + { + "epoch": 0.697421559490525, + "grad_norm": 0.4231663942337036, + "learning_rate": 8.978000000000001e-05, + "loss": 0.0213, + "step": 4490 + }, + { + "epoch": 0.6989748369058714, + "grad_norm": 0.4354873299598694, + "learning_rate": 8.998e-05, + "loss": 0.0235, + "step": 4500 + }, + { + "epoch": 0.7005281143212178, + "grad_norm": 0.4831709861755371, + "learning_rate": 9.018000000000001e-05, + "loss": 0.0234, + "step": 4510 + }, + { + "epoch": 0.7020813917365641, + "grad_norm": 0.28796571493148804, + "learning_rate": 9.038000000000001e-05, + "loss": 0.0187, + "step": 4520 + }, + { + "epoch": 0.7036346691519105, + "grad_norm": 0.4441050887107849, + "learning_rate": 9.058e-05, + "loss": 0.02, + "step": 4530 + }, + { + "epoch": 0.7051879465672569, + "grad_norm": 0.35589665174484253, + "learning_rate": 9.078000000000001e-05, + "loss": 0.0174, + "step": 4540 + }, + { + "epoch": 0.7067412239826033, + "grad_norm": 0.3405473232269287, + "learning_rate": 9.098000000000001e-05, + "loss": 0.0194, + "step": 4550 + }, + { + "epoch": 0.7082945013979497, + "grad_norm": 0.44914329051971436, + "learning_rate": 9.118e-05, + "loss": 0.0165, + "step": 4560 + }, + { + "epoch": 0.7098477788132961, + "grad_norm": 0.3398020267486572, + "learning_rate": 9.138e-05, + "loss": 0.0205, + "step": 4570 + }, + { + "epoch": 0.7114010562286425, + "grad_norm": 0.3878951966762543, + "learning_rate": 9.158e-05, + "loss": 0.0224, + "step": 4580 + }, + { + "epoch": 0.7129543336439889, + "grad_norm": 0.23215351998806, + "learning_rate": 9.178e-05, + "loss": 0.0175, + "step": 4590 + }, + { + "epoch": 0.7145076110593352, + "grad_norm": 0.3993667662143707, + "learning_rate": 9.198e-05, + "loss": 0.0173, + "step": 4600 + }, + { + "epoch": 0.7160608884746815, + "grad_norm": 0.4023281931877136, + "learning_rate": 9.218e-05, + "loss": 0.0193, + "step": 4610 + }, + { + "epoch": 0.7176141658900279, + "grad_norm": 0.3193744421005249, + "learning_rate": 9.238e-05, + "loss": 0.0219, + "step": 4620 + }, + { + "epoch": 0.7191674433053743, + "grad_norm": 0.3987332582473755, + "learning_rate": 9.258e-05, + "loss": 0.0175, + "step": 4630 + }, + { + "epoch": 0.7207207207207207, + "grad_norm": 0.29884573817253113, + "learning_rate": 9.278e-05, + "loss": 0.0208, + "step": 4640 + }, + { + "epoch": 0.7222739981360671, + "grad_norm": 0.40280893445014954, + "learning_rate": 9.298e-05, + "loss": 0.0195, + "step": 4650 + }, + { + "epoch": 0.7238272755514135, + "grad_norm": 0.41402533650398254, + "learning_rate": 9.318e-05, + "loss": 0.0248, + "step": 4660 + }, + { + "epoch": 0.7253805529667599, + "grad_norm": 0.26802965998649597, + "learning_rate": 9.338e-05, + "loss": 0.0155, + "step": 4670 + }, + { + "epoch": 0.7269338303821062, + "grad_norm": 0.2491266429424286, + "learning_rate": 9.358e-05, + "loss": 0.0153, + "step": 4680 + }, + { + "epoch": 0.7284871077974526, + "grad_norm": 0.5178234577178955, + "learning_rate": 9.378e-05, + "loss": 0.0193, + "step": 4690 + }, + { + "epoch": 0.730040385212799, + "grad_norm": 0.6372799873352051, + "learning_rate": 9.398e-05, + "loss": 0.0233, + "step": 4700 + }, + { + "epoch": 0.7315936626281454, + "grad_norm": 0.4429604113101959, + "learning_rate": 9.418e-05, + "loss": 0.0161, + "step": 4710 + }, + { + "epoch": 0.7331469400434918, + "grad_norm": 0.4115501046180725, + "learning_rate": 9.438e-05, + "loss": 0.0203, + "step": 4720 + }, + { + "epoch": 0.7347002174588382, + "grad_norm": 0.27570152282714844, + "learning_rate": 9.458e-05, + "loss": 0.0179, + "step": 4730 + }, + { + "epoch": 0.7362534948741846, + "grad_norm": 0.46651196479797363, + "learning_rate": 9.478e-05, + "loss": 0.0209, + "step": 4740 + }, + { + "epoch": 0.737806772289531, + "grad_norm": 0.3809497356414795, + "learning_rate": 9.498e-05, + "loss": 0.0193, + "step": 4750 + }, + { + "epoch": 0.7393600497048773, + "grad_norm": 0.3887590169906616, + "learning_rate": 9.518000000000001e-05, + "loss": 0.0284, + "step": 4760 + }, + { + "epoch": 0.7409133271202236, + "grad_norm": 0.2755822241306305, + "learning_rate": 9.538e-05, + "loss": 0.0219, + "step": 4770 + }, + { + "epoch": 0.74246660453557, + "grad_norm": 0.37884753942489624, + "learning_rate": 9.558e-05, + "loss": 0.0232, + "step": 4780 + }, + { + "epoch": 0.7440198819509164, + "grad_norm": 0.5830638408660889, + "learning_rate": 9.578000000000001e-05, + "loss": 0.0207, + "step": 4790 + }, + { + "epoch": 0.7455731593662628, + "grad_norm": 0.3067241907119751, + "learning_rate": 9.598e-05, + "loss": 0.0195, + "step": 4800 + }, + { + "epoch": 0.7471264367816092, + "grad_norm": 0.5548604726791382, + "learning_rate": 9.618e-05, + "loss": 0.0243, + "step": 4810 + }, + { + "epoch": 0.7486797141969556, + "grad_norm": 0.353014200925827, + "learning_rate": 9.638000000000001e-05, + "loss": 0.0197, + "step": 4820 + }, + { + "epoch": 0.750232991612302, + "grad_norm": 0.3387666642665863, + "learning_rate": 9.658e-05, + "loss": 0.0216, + "step": 4830 + }, + { + "epoch": 0.7517862690276483, + "grad_norm": 0.4228551387786865, + "learning_rate": 9.678e-05, + "loss": 0.0255, + "step": 4840 + }, + { + "epoch": 0.7533395464429947, + "grad_norm": 0.31531909108161926, + "learning_rate": 9.698000000000001e-05, + "loss": 0.0276, + "step": 4850 + }, + { + "epoch": 0.7548928238583411, + "grad_norm": 0.4276667535305023, + "learning_rate": 9.718e-05, + "loss": 0.0231, + "step": 4860 + }, + { + "epoch": 0.7564461012736875, + "grad_norm": 0.3621021807193756, + "learning_rate": 9.738e-05, + "loss": 0.0185, + "step": 4870 + }, + { + "epoch": 0.7579993786890339, + "grad_norm": 0.32501593232154846, + "learning_rate": 9.758000000000001e-05, + "loss": 0.018, + "step": 4880 + }, + { + "epoch": 0.7595526561043803, + "grad_norm": 0.32458144426345825, + "learning_rate": 9.778e-05, + "loss": 0.019, + "step": 4890 + }, + { + "epoch": 0.7611059335197267, + "grad_norm": 0.29764431715011597, + "learning_rate": 9.798000000000001e-05, + "loss": 0.0203, + "step": 4900 + }, + { + "epoch": 0.762659210935073, + "grad_norm": 0.34730517864227295, + "learning_rate": 9.818000000000001e-05, + "loss": 0.0183, + "step": 4910 + }, + { + "epoch": 0.7642124883504194, + "grad_norm": 0.3498704433441162, + "learning_rate": 9.838e-05, + "loss": 0.02, + "step": 4920 + }, + { + "epoch": 0.7657657657657657, + "grad_norm": 0.34347331523895264, + "learning_rate": 9.858000000000001e-05, + "loss": 0.0217, + "step": 4930 + }, + { + "epoch": 0.7673190431811121, + "grad_norm": 0.3170455992221832, + "learning_rate": 9.878e-05, + "loss": 0.017, + "step": 4940 + }, + { + "epoch": 0.7688723205964585, + "grad_norm": 0.273774653673172, + "learning_rate": 9.898e-05, + "loss": 0.0197, + "step": 4950 + }, + { + "epoch": 0.7704255980118049, + "grad_norm": 0.2755628228187561, + "learning_rate": 9.918000000000001e-05, + "loss": 0.0143, + "step": 4960 + }, + { + "epoch": 0.7719788754271513, + "grad_norm": 0.47582414746284485, + "learning_rate": 9.938e-05, + "loss": 0.0226, + "step": 4970 + }, + { + "epoch": 0.7735321528424977, + "grad_norm": 0.27605265378952026, + "learning_rate": 9.958e-05, + "loss": 0.0253, + "step": 4980 + }, + { + "epoch": 0.775085430257844, + "grad_norm": 0.32398170232772827, + "learning_rate": 9.978000000000001e-05, + "loss": 0.0184, + "step": 4990 + }, + { + "epoch": 0.7766387076731904, + "grad_norm": 0.4195055365562439, + "learning_rate": 9.998e-05, + "loss": 0.0268, + "step": 5000 + }, + { + "epoch": 0.7781919850885368, + "grad_norm": 0.49962925910949707, + "learning_rate": 9.999999778549045e-05, + "loss": 0.0206, + "step": 5010 + }, + { + "epoch": 0.7797452625038832, + "grad_norm": 0.3984353840351105, + "learning_rate": 9.999999013039593e-05, + "loss": 0.0227, + "step": 5020 + }, + { + "epoch": 0.7812985399192296, + "grad_norm": 0.34977152943611145, + "learning_rate": 9.999997700737766e-05, + "loss": 0.017, + "step": 5030 + }, + { + "epoch": 0.782851817334576, + "grad_norm": 0.38614553213119507, + "learning_rate": 9.999995841643709e-05, + "loss": 0.0186, + "step": 5040 + }, + { + "epoch": 0.7844050947499224, + "grad_norm": 0.26361149549484253, + "learning_rate": 9.999993435757623e-05, + "loss": 0.0185, + "step": 5050 + }, + { + "epoch": 0.7859583721652688, + "grad_norm": 0.36877477169036865, + "learning_rate": 9.999990483079773e-05, + "loss": 0.0163, + "step": 5060 + }, + { + "epoch": 0.7875116495806151, + "grad_norm": 0.28381046652793884, + "learning_rate": 9.999986983610481e-05, + "loss": 0.0174, + "step": 5070 + }, + { + "epoch": 0.7890649269959614, + "grad_norm": 0.28600403666496277, + "learning_rate": 9.99998293735013e-05, + "loss": 0.0182, + "step": 5080 + }, + { + "epoch": 0.7906182044113078, + "grad_norm": 0.246056467294693, + "learning_rate": 9.999978344299161e-05, + "loss": 0.0217, + "step": 5090 + }, + { + "epoch": 0.7921714818266542, + "grad_norm": 0.19279707968235016, + "learning_rate": 9.99997320445808e-05, + "loss": 0.0186, + "step": 5100 + }, + { + "epoch": 0.7937247592420006, + "grad_norm": 0.44312459230422974, + "learning_rate": 9.999967517827444e-05, + "loss": 0.018, + "step": 5110 + }, + { + "epoch": 0.795278036657347, + "grad_norm": 0.46537938714027405, + "learning_rate": 9.999961284407879e-05, + "loss": 0.0239, + "step": 5120 + }, + { + "epoch": 0.7968313140726934, + "grad_norm": 0.2998867928981781, + "learning_rate": 9.999954504200067e-05, + "loss": 0.016, + "step": 5130 + }, + { + "epoch": 0.7983845914880398, + "grad_norm": 0.45792466402053833, + "learning_rate": 9.999947177204744e-05, + "loss": 0.0201, + "step": 5140 + }, + { + "epoch": 0.7999378689033861, + "grad_norm": 0.5906203985214233, + "learning_rate": 9.999939303422718e-05, + "loss": 0.0209, + "step": 5150 + }, + { + "epoch": 0.8014911463187325, + "grad_norm": 0.5923995971679688, + "learning_rate": 9.999930882854847e-05, + "loss": 0.021, + "step": 5160 + }, + { + "epoch": 0.8030444237340789, + "grad_norm": 0.3864830434322357, + "learning_rate": 9.999921915502051e-05, + "loss": 0.0135, + "step": 5170 + }, + { + "epoch": 0.8045977011494253, + "grad_norm": 0.4023837745189667, + "learning_rate": 9.99991240136531e-05, + "loss": 0.0184, + "step": 5180 + }, + { + "epoch": 0.8061509785647717, + "grad_norm": 0.4698565602302551, + "learning_rate": 9.999902340445668e-05, + "loss": 0.0191, + "step": 5190 + }, + { + "epoch": 0.8077042559801181, + "grad_norm": 0.44033360481262207, + "learning_rate": 9.999891732744224e-05, + "loss": 0.0207, + "step": 5200 + }, + { + "epoch": 0.8092575333954645, + "grad_norm": 0.48180603981018066, + "learning_rate": 9.999880578262135e-05, + "loss": 0.021, + "step": 5210 + }, + { + "epoch": 0.8108108108108109, + "grad_norm": 0.3052384853363037, + "learning_rate": 9.999868877000624e-05, + "loss": 0.0225, + "step": 5220 + }, + { + "epoch": 0.8123640882261572, + "grad_norm": 0.2826438546180725, + "learning_rate": 9.99985662896097e-05, + "loss": 0.0219, + "step": 5230 + }, + { + "epoch": 0.8139173656415035, + "grad_norm": 0.3504715859889984, + "learning_rate": 9.999843834144513e-05, + "loss": 0.0203, + "step": 5240 + }, + { + "epoch": 0.8154706430568499, + "grad_norm": 0.46667152643203735, + "learning_rate": 9.99983049255265e-05, + "loss": 0.0178, + "step": 5250 + }, + { + "epoch": 0.8170239204721963, + "grad_norm": 0.4184421896934509, + "learning_rate": 9.999816604186843e-05, + "loss": 0.0203, + "step": 5260 + }, + { + "epoch": 0.8185771978875427, + "grad_norm": 0.3795464336872101, + "learning_rate": 9.999802169048609e-05, + "loss": 0.0199, + "step": 5270 + }, + { + "epoch": 0.8201304753028891, + "grad_norm": 0.3482944965362549, + "learning_rate": 9.999787187139527e-05, + "loss": 0.0234, + "step": 5280 + }, + { + "epoch": 0.8216837527182355, + "grad_norm": 0.2763994336128235, + "learning_rate": 9.999771658461234e-05, + "loss": 0.017, + "step": 5290 + }, + { + "epoch": 0.8232370301335818, + "grad_norm": 0.29315200448036194, + "learning_rate": 9.999755583015431e-05, + "loss": 0.0164, + "step": 5300 + }, + { + "epoch": 0.8247903075489282, + "grad_norm": 0.37575140595436096, + "learning_rate": 9.999738960803874e-05, + "loss": 0.0199, + "step": 5310 + }, + { + "epoch": 0.8263435849642746, + "grad_norm": 0.3647838532924652, + "learning_rate": 9.99972179182838e-05, + "loss": 0.0202, + "step": 5320 + }, + { + "epoch": 0.827896862379621, + "grad_norm": 0.3591662049293518, + "learning_rate": 9.99970407609083e-05, + "loss": 0.0189, + "step": 5330 + }, + { + "epoch": 0.8294501397949674, + "grad_norm": 0.36736661195755005, + "learning_rate": 9.999685813593159e-05, + "loss": 0.0221, + "step": 5340 + }, + { + "epoch": 0.8310034172103138, + "grad_norm": 0.42795246839523315, + "learning_rate": 9.999667004337362e-05, + "loss": 0.0206, + "step": 5350 + }, + { + "epoch": 0.8325566946256602, + "grad_norm": 0.39783039689064026, + "learning_rate": 9.9996476483255e-05, + "loss": 0.0224, + "step": 5360 + }, + { + "epoch": 0.8341099720410066, + "grad_norm": 0.42989394068717957, + "learning_rate": 9.999627745559688e-05, + "loss": 0.0188, + "step": 5370 + }, + { + "epoch": 0.835663249456353, + "grad_norm": 0.2490014135837555, + "learning_rate": 9.999607296042101e-05, + "loss": 0.0168, + "step": 5380 + }, + { + "epoch": 0.8372165268716993, + "grad_norm": 0.426862895488739, + "learning_rate": 9.99958629977498e-05, + "loss": 0.0177, + "step": 5390 + }, + { + "epoch": 0.8387698042870456, + "grad_norm": 0.19840911030769348, + "learning_rate": 9.999564756760615e-05, + "loss": 0.013, + "step": 5400 + }, + { + "epoch": 0.840323081702392, + "grad_norm": 0.31534600257873535, + "learning_rate": 9.999542667001366e-05, + "loss": 0.0163, + "step": 5410 + }, + { + "epoch": 0.8418763591177384, + "grad_norm": 0.21565191447734833, + "learning_rate": 9.999520030499647e-05, + "loss": 0.0133, + "step": 5420 + }, + { + "epoch": 0.8434296365330848, + "grad_norm": 0.27959680557250977, + "learning_rate": 9.999496847257936e-05, + "loss": 0.015, + "step": 5430 + }, + { + "epoch": 0.8449829139484312, + "grad_norm": 0.39910680055618286, + "learning_rate": 9.999473117278764e-05, + "loss": 0.0175, + "step": 5440 + }, + { + "epoch": 0.8465361913637776, + "grad_norm": 0.4373420774936676, + "learning_rate": 9.999448840564731e-05, + "loss": 0.0212, + "step": 5450 + }, + { + "epoch": 0.848089468779124, + "grad_norm": 0.24881170690059662, + "learning_rate": 9.999424017118488e-05, + "loss": 0.0188, + "step": 5460 + }, + { + "epoch": 0.8496427461944703, + "grad_norm": 0.521437406539917, + "learning_rate": 9.999398646942751e-05, + "loss": 0.0219, + "step": 5470 + }, + { + "epoch": 0.8511960236098167, + "grad_norm": 0.2967197895050049, + "learning_rate": 9.999372730040296e-05, + "loss": 0.0201, + "step": 5480 + }, + { + "epoch": 0.8527493010251631, + "grad_norm": 0.30900272727012634, + "learning_rate": 9.999346266413953e-05, + "loss": 0.0164, + "step": 5490 + }, + { + "epoch": 0.8543025784405095, + "grad_norm": 0.2953626811504364, + "learning_rate": 9.99931925606662e-05, + "loss": 0.0155, + "step": 5500 + }, + { + "epoch": 0.8558558558558559, + "grad_norm": 0.4015914499759674, + "learning_rate": 9.99929169900125e-05, + "loss": 0.0184, + "step": 5510 + }, + { + "epoch": 0.8574091332712023, + "grad_norm": 0.4269683063030243, + "learning_rate": 9.999263595220855e-05, + "loss": 0.017, + "step": 5520 + }, + { + "epoch": 0.8589624106865487, + "grad_norm": 0.4371897876262665, + "learning_rate": 9.99923494472851e-05, + "loss": 0.0237, + "step": 5530 + }, + { + "epoch": 0.860515688101895, + "grad_norm": 0.43023189902305603, + "learning_rate": 9.999205747527348e-05, + "loss": 0.0183, + "step": 5540 + }, + { + "epoch": 0.8620689655172413, + "grad_norm": 0.6745098829269409, + "learning_rate": 9.999176003620561e-05, + "loss": 0.0221, + "step": 5550 + }, + { + "epoch": 0.8636222429325877, + "grad_norm": 0.3654077649116516, + "learning_rate": 9.999145713011405e-05, + "loss": 0.0214, + "step": 5560 + }, + { + "epoch": 0.8651755203479341, + "grad_norm": 0.2959289252758026, + "learning_rate": 9.999114875703186e-05, + "loss": 0.0192, + "step": 5570 + }, + { + "epoch": 0.8667287977632805, + "grad_norm": 0.27987805008888245, + "learning_rate": 9.999083491699281e-05, + "loss": 0.0208, + "step": 5580 + }, + { + "epoch": 0.8682820751786269, + "grad_norm": 0.31854942440986633, + "learning_rate": 9.999051561003123e-05, + "loss": 0.023, + "step": 5590 + }, + { + "epoch": 0.8698353525939733, + "grad_norm": 0.31920871138572693, + "learning_rate": 9.999019083618202e-05, + "loss": 0.0214, + "step": 5600 + }, + { + "epoch": 0.8713886300093197, + "grad_norm": 0.29555726051330566, + "learning_rate": 9.99898605954807e-05, + "loss": 0.0203, + "step": 5610 + }, + { + "epoch": 0.872941907424666, + "grad_norm": 0.5261118412017822, + "learning_rate": 9.998952488796338e-05, + "loss": 0.022, + "step": 5620 + }, + { + "epoch": 0.8744951848400124, + "grad_norm": 0.413279265165329, + "learning_rate": 9.998918371366676e-05, + "loss": 0.0198, + "step": 5630 + }, + { + "epoch": 0.8760484622553588, + "grad_norm": 0.5236073732376099, + "learning_rate": 9.99888370726282e-05, + "loss": 0.0238, + "step": 5640 + }, + { + "epoch": 0.8776017396707052, + "grad_norm": 0.41230425238609314, + "learning_rate": 9.998848496488556e-05, + "loss": 0.0189, + "step": 5650 + }, + { + "epoch": 0.8791550170860516, + "grad_norm": 0.3116629123687744, + "learning_rate": 9.998812739047736e-05, + "loss": 0.0243, + "step": 5660 + }, + { + "epoch": 0.880708294501398, + "grad_norm": 0.5200260877609253, + "learning_rate": 9.99877643494427e-05, + "loss": 0.0222, + "step": 5670 + }, + { + "epoch": 0.8822615719167444, + "grad_norm": 0.3421640992164612, + "learning_rate": 9.998739584182128e-05, + "loss": 0.0212, + "step": 5680 + }, + { + "epoch": 0.8838148493320908, + "grad_norm": 0.5181063413619995, + "learning_rate": 9.998702186765342e-05, + "loss": 0.0195, + "step": 5690 + }, + { + "epoch": 0.8853681267474371, + "grad_norm": 0.3798215985298157, + "learning_rate": 9.998664242698e-05, + "loss": 0.0163, + "step": 5700 + }, + { + "epoch": 0.8869214041627834, + "grad_norm": 0.34588822722435, + "learning_rate": 9.998625751984251e-05, + "loss": 0.019, + "step": 5710 + }, + { + "epoch": 0.8884746815781298, + "grad_norm": 0.4281695485115051, + "learning_rate": 9.998586714628307e-05, + "loss": 0.0189, + "step": 5720 + }, + { + "epoch": 0.8900279589934762, + "grad_norm": 0.37254616618156433, + "learning_rate": 9.998547130634432e-05, + "loss": 0.0173, + "step": 5730 + }, + { + "epoch": 0.8915812364088226, + "grad_norm": 0.38244301080703735, + "learning_rate": 9.99850700000696e-05, + "loss": 0.0178, + "step": 5740 + }, + { + "epoch": 0.893134513824169, + "grad_norm": 0.3610047698020935, + "learning_rate": 9.998466322750278e-05, + "loss": 0.0184, + "step": 5750 + }, + { + "epoch": 0.8946877912395154, + "grad_norm": 0.2922575771808624, + "learning_rate": 9.998425098868834e-05, + "loss": 0.0194, + "step": 5760 + }, + { + "epoch": 0.8962410686548617, + "grad_norm": 0.43177950382232666, + "learning_rate": 9.998383328367136e-05, + "loss": 0.0182, + "step": 5770 + }, + { + "epoch": 0.8977943460702081, + "grad_norm": 0.33594071865081787, + "learning_rate": 9.99834101124975e-05, + "loss": 0.0194, + "step": 5780 + }, + { + "epoch": 0.8993476234855545, + "grad_norm": 0.2617932856082916, + "learning_rate": 9.998298147521309e-05, + "loss": 0.0177, + "step": 5790 + }, + { + "epoch": 0.9009009009009009, + "grad_norm": 0.2744154632091522, + "learning_rate": 9.998254737186496e-05, + "loss": 0.0158, + "step": 5800 + }, + { + "epoch": 0.9024541783162473, + "grad_norm": 0.3040594458580017, + "learning_rate": 9.99821078025006e-05, + "loss": 0.0168, + "step": 5810 + }, + { + "epoch": 0.9040074557315937, + "grad_norm": 0.31313011050224304, + "learning_rate": 9.998166276716807e-05, + "loss": 0.0129, + "step": 5820 + }, + { + "epoch": 0.9055607331469401, + "grad_norm": 0.5319452285766602, + "learning_rate": 9.998121226591606e-05, + "loss": 0.019, + "step": 5830 + }, + { + "epoch": 0.9071140105622865, + "grad_norm": 0.35954758524894714, + "learning_rate": 9.998075629879382e-05, + "loss": 0.0155, + "step": 5840 + }, + { + "epoch": 0.9086672879776329, + "grad_norm": 0.6281808018684387, + "learning_rate": 9.99802948658512e-05, + "loss": 0.031, + "step": 5850 + }, + { + "epoch": 0.9102205653929791, + "grad_norm": 0.5640057921409607, + "learning_rate": 9.99798279671387e-05, + "loss": 0.0202, + "step": 5860 + }, + { + "epoch": 0.9117738428083255, + "grad_norm": 0.47710713744163513, + "learning_rate": 9.997935560270734e-05, + "loss": 0.0208, + "step": 5870 + }, + { + "epoch": 0.9133271202236719, + "grad_norm": 0.4223294258117676, + "learning_rate": 9.997887777260879e-05, + "loss": 0.026, + "step": 5880 + }, + { + "epoch": 0.9148803976390183, + "grad_norm": 0.5207691192626953, + "learning_rate": 9.997839447689532e-05, + "loss": 0.0183, + "step": 5890 + }, + { + "epoch": 0.9164336750543647, + "grad_norm": 0.32846391201019287, + "learning_rate": 9.997790571561978e-05, + "loss": 0.0227, + "step": 5900 + }, + { + "epoch": 0.9179869524697111, + "grad_norm": 0.4321678578853607, + "learning_rate": 9.99774114888356e-05, + "loss": 0.0173, + "step": 5910 + }, + { + "epoch": 0.9195402298850575, + "grad_norm": 0.36262258887290955, + "learning_rate": 9.997691179659684e-05, + "loss": 0.0199, + "step": 5920 + }, + { + "epoch": 0.9210935073004038, + "grad_norm": 0.42490583658218384, + "learning_rate": 9.997640663895815e-05, + "loss": 0.0154, + "step": 5930 + }, + { + "epoch": 0.9226467847157502, + "grad_norm": 0.4503207802772522, + "learning_rate": 9.997589601597477e-05, + "loss": 0.0184, + "step": 5940 + }, + { + "epoch": 0.9242000621310966, + "grad_norm": 0.24310779571533203, + "learning_rate": 9.997537992770252e-05, + "loss": 0.0154, + "step": 5950 + }, + { + "epoch": 0.925753339546443, + "grad_norm": 0.5075423717498779, + "learning_rate": 9.997485837419788e-05, + "loss": 0.0155, + "step": 5960 + }, + { + "epoch": 0.9273066169617894, + "grad_norm": 0.3087664842605591, + "learning_rate": 9.997433135551786e-05, + "loss": 0.0165, + "step": 5970 + }, + { + "epoch": 0.9288598943771358, + "grad_norm": 0.2795964777469635, + "learning_rate": 9.997379887172009e-05, + "loss": 0.0202, + "step": 5980 + }, + { + "epoch": 0.9304131717924822, + "grad_norm": 0.32519736886024475, + "learning_rate": 9.997326092286281e-05, + "loss": 0.0185, + "step": 5990 + }, + { + "epoch": 0.9319664492078286, + "grad_norm": 0.2983795404434204, + "learning_rate": 9.997271750900486e-05, + "loss": 0.0178, + "step": 6000 + }, + { + "epoch": 0.933519726623175, + "grad_norm": 0.35272425413131714, + "learning_rate": 9.997216863020565e-05, + "loss": 0.0167, + "step": 6010 + }, + { + "epoch": 0.9350730040385212, + "grad_norm": 0.22682850062847137, + "learning_rate": 9.99716142865252e-05, + "loss": 0.0184, + "step": 6020 + }, + { + "epoch": 0.9366262814538676, + "grad_norm": 0.354120135307312, + "learning_rate": 9.997105447802415e-05, + "loss": 0.0166, + "step": 6030 + }, + { + "epoch": 0.938179558869214, + "grad_norm": 0.3732014298439026, + "learning_rate": 9.997048920476373e-05, + "loss": 0.0154, + "step": 6040 + }, + { + "epoch": 0.9397328362845604, + "grad_norm": 0.536556601524353, + "learning_rate": 9.996991846680572e-05, + "loss": 0.0127, + "step": 6050 + }, + { + "epoch": 0.9412861136999068, + "grad_norm": 0.49011772871017456, + "learning_rate": 9.996934226421257e-05, + "loss": 0.0192, + "step": 6060 + }, + { + "epoch": 0.9428393911152532, + "grad_norm": 0.6354818940162659, + "learning_rate": 9.996876059704726e-05, + "loss": 0.0173, + "step": 6070 + }, + { + "epoch": 0.9443926685305996, + "grad_norm": 0.5354923605918884, + "learning_rate": 9.996817346537343e-05, + "loss": 0.0197, + "step": 6080 + }, + { + "epoch": 0.9459459459459459, + "grad_norm": 0.25008535385131836, + "learning_rate": 9.996758086925526e-05, + "loss": 0.0193, + "step": 6090 + }, + { + "epoch": 0.9474992233612923, + "grad_norm": 0.4074191749095917, + "learning_rate": 9.996698280875759e-05, + "loss": 0.0158, + "step": 6100 + }, + { + "epoch": 0.9490525007766387, + "grad_norm": 0.2638848125934601, + "learning_rate": 9.99663792839458e-05, + "loss": 0.0185, + "step": 6110 + }, + { + "epoch": 0.9506057781919851, + "grad_norm": 0.34890350699424744, + "learning_rate": 9.99657702948859e-05, + "loss": 0.0184, + "step": 6120 + }, + { + "epoch": 0.9521590556073315, + "grad_norm": 0.35434606671333313, + "learning_rate": 9.996515584164448e-05, + "loss": 0.0197, + "step": 6130 + }, + { + "epoch": 0.9537123330226779, + "grad_norm": 0.4577362537384033, + "learning_rate": 9.996453592428873e-05, + "loss": 0.0144, + "step": 6140 + }, + { + "epoch": 0.9552656104380243, + "grad_norm": 0.23075363039970398, + "learning_rate": 9.996391054288646e-05, + "loss": 0.0176, + "step": 6150 + }, + { + "epoch": 0.9568188878533707, + "grad_norm": 0.179216206073761, + "learning_rate": 9.996327969750605e-05, + "loss": 0.0181, + "step": 6160 + }, + { + "epoch": 0.958372165268717, + "grad_norm": 0.2669224441051483, + "learning_rate": 9.996264338821649e-05, + "loss": 0.0156, + "step": 6170 + }, + { + "epoch": 0.9599254426840633, + "grad_norm": 0.27921393513679504, + "learning_rate": 9.996200161508735e-05, + "loss": 0.0168, + "step": 6180 + }, + { + "epoch": 0.9614787200994097, + "grad_norm": 0.31630295515060425, + "learning_rate": 9.996135437818885e-05, + "loss": 0.0191, + "step": 6190 + }, + { + "epoch": 0.9630319975147561, + "grad_norm": 0.33314740657806396, + "learning_rate": 9.996070167759175e-05, + "loss": 0.0168, + "step": 6200 + }, + { + "epoch": 0.9645852749301025, + "grad_norm": 0.48492223024368286, + "learning_rate": 9.996004351336743e-05, + "loss": 0.0159, + "step": 6210 + }, + { + "epoch": 0.9661385523454489, + "grad_norm": 0.32416772842407227, + "learning_rate": 9.995937988558785e-05, + "loss": 0.0169, + "step": 6220 + }, + { + "epoch": 0.9676918297607953, + "grad_norm": 0.26955047249794006, + "learning_rate": 9.995871079432561e-05, + "loss": 0.0153, + "step": 6230 + }, + { + "epoch": 0.9692451071761417, + "grad_norm": 0.3047213852405548, + "learning_rate": 9.995803623965389e-05, + "loss": 0.0172, + "step": 6240 + }, + { + "epoch": 0.970798384591488, + "grad_norm": 0.49126601219177246, + "learning_rate": 9.995735622164641e-05, + "loss": 0.0178, + "step": 6250 + }, + { + "epoch": 0.9723516620068344, + "grad_norm": 0.38471412658691406, + "learning_rate": 9.995667074037758e-05, + "loss": 0.018, + "step": 6260 + }, + { + "epoch": 0.9739049394221808, + "grad_norm": 0.3738669157028198, + "learning_rate": 9.995597979592232e-05, + "loss": 0.0177, + "step": 6270 + }, + { + "epoch": 0.9754582168375272, + "grad_norm": 0.4019346535205841, + "learning_rate": 9.995528338835625e-05, + "loss": 0.0168, + "step": 6280 + }, + { + "epoch": 0.9770114942528736, + "grad_norm": 0.4047831892967224, + "learning_rate": 9.995458151775547e-05, + "loss": 0.0194, + "step": 6290 + }, + { + "epoch": 0.97856477166822, + "grad_norm": 0.32513511180877686, + "learning_rate": 9.995387418419677e-05, + "loss": 0.0184, + "step": 6300 + }, + { + "epoch": 0.9801180490835664, + "grad_norm": 0.3433535695075989, + "learning_rate": 9.99531613877575e-05, + "loss": 0.0168, + "step": 6310 + }, + { + "epoch": 0.9816713264989128, + "grad_norm": 0.37518540024757385, + "learning_rate": 9.995244312851559e-05, + "loss": 0.0163, + "step": 6320 + }, + { + "epoch": 0.983224603914259, + "grad_norm": 0.37061384320259094, + "learning_rate": 9.995171940654961e-05, + "loss": 0.0188, + "step": 6330 + }, + { + "epoch": 0.9847778813296054, + "grad_norm": 0.5047083497047424, + "learning_rate": 9.995099022193871e-05, + "loss": 0.0216, + "step": 6340 + }, + { + "epoch": 0.9863311587449518, + "grad_norm": 0.26485079526901245, + "learning_rate": 9.995025557476261e-05, + "loss": 0.0173, + "step": 6350 + }, + { + "epoch": 0.9878844361602982, + "grad_norm": 0.3691689074039459, + "learning_rate": 9.994951546510165e-05, + "loss": 0.02, + "step": 6360 + }, + { + "epoch": 0.9894377135756446, + "grad_norm": 0.32936593890190125, + "learning_rate": 9.994876989303679e-05, + "loss": 0.0187, + "step": 6370 + }, + { + "epoch": 0.990990990990991, + "grad_norm": 0.31786420941352844, + "learning_rate": 9.994801885864955e-05, + "loss": 0.018, + "step": 6380 + }, + { + "epoch": 0.9925442684063374, + "grad_norm": 0.4727120101451874, + "learning_rate": 9.994726236202205e-05, + "loss": 0.018, + "step": 6390 + }, + { + "epoch": 0.9940975458216837, + "grad_norm": 0.38210999965667725, + "learning_rate": 9.994650040323704e-05, + "loss": 0.0198, + "step": 6400 + }, + { + "epoch": 0.9956508232370301, + "grad_norm": 0.29737892746925354, + "learning_rate": 9.994573298237784e-05, + "loss": 0.0205, + "step": 6410 + }, + { + "epoch": 0.9972041006523765, + "grad_norm": 0.35102900862693787, + "learning_rate": 9.994496009952837e-05, + "loss": 0.0165, + "step": 6420 + }, + { + "epoch": 0.9987573780677229, + "grad_norm": 0.32245099544525146, + "learning_rate": 9.994418175477316e-05, + "loss": 0.016, + "step": 6430 + }, + { + "epoch": 1.0003106554830692, + "grad_norm": 0.3356991708278656, + "learning_rate": 9.994339794819733e-05, + "loss": 0.0169, + "step": 6440 + }, + { + "epoch": 1.0018639328984156, + "grad_norm": 0.30640697479248047, + "learning_rate": 9.994260867988658e-05, + "loss": 0.0164, + "step": 6450 + }, + { + "epoch": 1.003417210313762, + "grad_norm": 0.3921419680118561, + "learning_rate": 9.994181394992723e-05, + "loss": 0.0181, + "step": 6460 + }, + { + "epoch": 1.0049704877291084, + "grad_norm": 0.4090765118598938, + "learning_rate": 9.994101375840618e-05, + "loss": 0.0179, + "step": 6470 + }, + { + "epoch": 1.0065237651444547, + "grad_norm": 0.3990112841129303, + "learning_rate": 9.994020810541098e-05, + "loss": 0.0168, + "step": 6480 + }, + { + "epoch": 1.0080770425598011, + "grad_norm": 0.39592981338500977, + "learning_rate": 9.99393969910297e-05, + "loss": 0.016, + "step": 6490 + }, + { + "epoch": 1.0096303199751475, + "grad_norm": 0.5027446746826172, + "learning_rate": 9.993858041535104e-05, + "loss": 0.013, + "step": 6500 + }, + { + "epoch": 1.011183597390494, + "grad_norm": 0.3200371265411377, + "learning_rate": 9.99377583784643e-05, + "loss": 0.0191, + "step": 6510 + }, + { + "epoch": 1.0127368748058403, + "grad_norm": 0.3155459463596344, + "learning_rate": 9.993693088045939e-05, + "loss": 0.0196, + "step": 6520 + }, + { + "epoch": 1.0142901522211867, + "grad_norm": 0.3452337384223938, + "learning_rate": 9.99360979214268e-05, + "loss": 0.0201, + "step": 6530 + }, + { + "epoch": 1.015843429636533, + "grad_norm": 0.35117754340171814, + "learning_rate": 9.99352595014576e-05, + "loss": 0.0225, + "step": 6540 + }, + { + "epoch": 1.0173967070518795, + "grad_norm": 0.48641908168792725, + "learning_rate": 9.993441562064354e-05, + "loss": 0.0189, + "step": 6550 + }, + { + "epoch": 1.0189499844672258, + "grad_norm": 0.29876571893692017, + "learning_rate": 9.993356627907685e-05, + "loss": 0.0191, + "step": 6560 + }, + { + "epoch": 1.0205032618825722, + "grad_norm": 0.3222905993461609, + "learning_rate": 9.99327114768504e-05, + "loss": 0.0187, + "step": 6570 + }, + { + "epoch": 1.0220565392979186, + "grad_norm": 0.30776363611221313, + "learning_rate": 9.99318512140577e-05, + "loss": 0.021, + "step": 6580 + }, + { + "epoch": 1.023609816713265, + "grad_norm": 0.3782219886779785, + "learning_rate": 9.993098549079284e-05, + "loss": 0.017, + "step": 6590 + }, + { + "epoch": 1.0251630941286114, + "grad_norm": 0.27767637372016907, + "learning_rate": 9.993011430715047e-05, + "loss": 0.0164, + "step": 6600 + }, + { + "epoch": 1.0267163715439578, + "grad_norm": 0.36127999424934387, + "learning_rate": 9.992923766322586e-05, + "loss": 0.0191, + "step": 6610 + }, + { + "epoch": 1.0282696489593042, + "grad_norm": 0.41528084874153137, + "learning_rate": 9.99283555591149e-05, + "loss": 0.0163, + "step": 6620 + }, + { + "epoch": 1.0298229263746506, + "grad_norm": 0.3935718238353729, + "learning_rate": 9.992746799491404e-05, + "loss": 0.0214, + "step": 6630 + }, + { + "epoch": 1.031376203789997, + "grad_norm": 0.33298933506011963, + "learning_rate": 9.992657497072033e-05, + "loss": 0.0157, + "step": 6640 + }, + { + "epoch": 1.0329294812053433, + "grad_norm": 0.4433385729789734, + "learning_rate": 9.992567648663147e-05, + "loss": 0.0166, + "step": 6650 + }, + { + "epoch": 1.0344827586206897, + "grad_norm": 0.3985118567943573, + "learning_rate": 9.992477254274568e-05, + "loss": 0.0159, + "step": 6660 + }, + { + "epoch": 1.0360360360360361, + "grad_norm": 0.3494184613227844, + "learning_rate": 9.992386313916183e-05, + "loss": 0.0176, + "step": 6670 + }, + { + "epoch": 1.0375893134513825, + "grad_norm": 0.373420774936676, + "learning_rate": 9.992294827597934e-05, + "loss": 0.0187, + "step": 6680 + }, + { + "epoch": 1.0391425908667289, + "grad_norm": 0.3657410144805908, + "learning_rate": 9.992202795329831e-05, + "loss": 0.0143, + "step": 6690 + }, + { + "epoch": 1.0406958682820753, + "grad_norm": 0.3566342294216156, + "learning_rate": 9.992110217121936e-05, + "loss": 0.0172, + "step": 6700 + }, + { + "epoch": 1.0422491456974217, + "grad_norm": 0.22829951345920563, + "learning_rate": 9.992017092984372e-05, + "loss": 0.0154, + "step": 6710 + }, + { + "epoch": 1.0438024231127678, + "grad_norm": 0.2941630780696869, + "learning_rate": 9.991923422927326e-05, + "loss": 0.0185, + "step": 6720 + }, + { + "epoch": 1.0453557005281142, + "grad_norm": 0.40444087982177734, + "learning_rate": 9.991829206961037e-05, + "loss": 0.0189, + "step": 6730 + }, + { + "epoch": 1.0469089779434606, + "grad_norm": 0.2612919211387634, + "learning_rate": 9.991734445095813e-05, + "loss": 0.0199, + "step": 6740 + }, + { + "epoch": 1.048462255358807, + "grad_norm": 0.31991633772850037, + "learning_rate": 9.991639137342015e-05, + "loss": 0.0163, + "step": 6750 + }, + { + "epoch": 1.0500155327741534, + "grad_norm": 0.44674965739250183, + "learning_rate": 9.991543283710064e-05, + "loss": 0.017, + "step": 6760 + }, + { + "epoch": 1.0515688101894998, + "grad_norm": 0.3704787790775299, + "learning_rate": 9.991446884210445e-05, + "loss": 0.0161, + "step": 6770 + }, + { + "epoch": 1.0531220876048462, + "grad_norm": 0.29208335280418396, + "learning_rate": 9.9913499388537e-05, + "loss": 0.0156, + "step": 6780 + }, + { + "epoch": 1.0546753650201925, + "grad_norm": 0.4363374412059784, + "learning_rate": 9.99125244765043e-05, + "loss": 0.0166, + "step": 6790 + }, + { + "epoch": 1.056228642435539, + "grad_norm": 0.3936074376106262, + "learning_rate": 9.991154410611296e-05, + "loss": 0.0188, + "step": 6800 + }, + { + "epoch": 1.0577819198508853, + "grad_norm": 0.26997190713882446, + "learning_rate": 9.99105582774702e-05, + "loss": 0.0177, + "step": 6810 + }, + { + "epoch": 1.0593351972662317, + "grad_norm": 0.3069198429584503, + "learning_rate": 9.990956699068384e-05, + "loss": 0.0163, + "step": 6820 + }, + { + "epoch": 1.060888474681578, + "grad_norm": 0.2344360053539276, + "learning_rate": 9.990857024586224e-05, + "loss": 0.0194, + "step": 6830 + }, + { + "epoch": 1.0624417520969245, + "grad_norm": 0.3164210021495819, + "learning_rate": 9.990756804311446e-05, + "loss": 0.0121, + "step": 6840 + }, + { + "epoch": 1.0639950295122709, + "grad_norm": 0.2180115431547165, + "learning_rate": 9.990656038255006e-05, + "loss": 0.0149, + "step": 6850 + }, + { + "epoch": 1.0655483069276173, + "grad_norm": 0.31404751539230347, + "learning_rate": 9.990554726427926e-05, + "loss": 0.0169, + "step": 6860 + }, + { + "epoch": 1.0671015843429636, + "grad_norm": 0.2733767032623291, + "learning_rate": 9.990452868841284e-05, + "loss": 0.02, + "step": 6870 + }, + { + "epoch": 1.06865486175831, + "grad_norm": 0.49669310450553894, + "learning_rate": 9.99035046550622e-05, + "loss": 0.0204, + "step": 6880 + }, + { + "epoch": 1.0702081391736564, + "grad_norm": 0.3467947840690613, + "learning_rate": 9.99024751643393e-05, + "loss": 0.017, + "step": 6890 + }, + { + "epoch": 1.0717614165890028, + "grad_norm": 0.3023830056190491, + "learning_rate": 9.990144021635677e-05, + "loss": 0.0189, + "step": 6900 + }, + { + "epoch": 1.0733146940043492, + "grad_norm": 0.31852760910987854, + "learning_rate": 9.990039981122775e-05, + "loss": 0.0202, + "step": 6910 + }, + { + "epoch": 1.0748679714196956, + "grad_norm": 0.2784249484539032, + "learning_rate": 9.989935394906602e-05, + "loss": 0.0238, + "step": 6920 + }, + { + "epoch": 1.076421248835042, + "grad_norm": 0.34835079312324524, + "learning_rate": 9.989830262998598e-05, + "loss": 0.0127, + "step": 6930 + }, + { + "epoch": 1.0779745262503884, + "grad_norm": 0.30136239528656006, + "learning_rate": 9.989724585410259e-05, + "loss": 0.0151, + "step": 6940 + }, + { + "epoch": 1.0795278036657348, + "grad_norm": 0.471185564994812, + "learning_rate": 9.989618362153139e-05, + "loss": 0.0161, + "step": 6950 + }, + { + "epoch": 1.0810810810810811, + "grad_norm": 0.4087601900100708, + "learning_rate": 9.989511593238859e-05, + "loss": 0.0202, + "step": 6960 + }, + { + "epoch": 1.0826343584964275, + "grad_norm": 0.21011652052402496, + "learning_rate": 9.98940427867909e-05, + "loss": 0.0191, + "step": 6970 + }, + { + "epoch": 1.084187635911774, + "grad_norm": 0.25809380412101746, + "learning_rate": 9.989296418485573e-05, + "loss": 0.0162, + "step": 6980 + }, + { + "epoch": 1.0857409133271203, + "grad_norm": 0.30935874581336975, + "learning_rate": 9.989188012670101e-05, + "loss": 0.0195, + "step": 6990 + }, + { + "epoch": 1.0872941907424667, + "grad_norm": 0.2712854743003845, + "learning_rate": 9.989079061244528e-05, + "loss": 0.0151, + "step": 7000 + }, + { + "epoch": 1.088847468157813, + "grad_norm": 0.5930557250976562, + "learning_rate": 9.988969564220769e-05, + "loss": 0.0149, + "step": 7010 + }, + { + "epoch": 1.0904007455731595, + "grad_norm": 0.36446622014045715, + "learning_rate": 9.988859521610801e-05, + "loss": 0.0145, + "step": 7020 + }, + { + "epoch": 1.0919540229885056, + "grad_norm": 0.40819108486175537, + "learning_rate": 9.988748933426656e-05, + "loss": 0.0148, + "step": 7030 + }, + { + "epoch": 1.0935073004038522, + "grad_norm": 0.40872350335121155, + "learning_rate": 9.988637799680428e-05, + "loss": 0.0174, + "step": 7040 + }, + { + "epoch": 1.0950605778191984, + "grad_norm": 0.3175908625125885, + "learning_rate": 9.98852612038427e-05, + "loss": 0.0232, + "step": 7050 + }, + { + "epoch": 1.0966138552345448, + "grad_norm": 0.3920903503894806, + "learning_rate": 9.988413895550397e-05, + "loss": 0.0185, + "step": 7060 + }, + { + "epoch": 1.0981671326498912, + "grad_norm": 0.3184399902820587, + "learning_rate": 9.98830112519108e-05, + "loss": 0.0175, + "step": 7070 + }, + { + "epoch": 1.0997204100652376, + "grad_norm": 0.7046704888343811, + "learning_rate": 9.98818780931865e-05, + "loss": 0.02, + "step": 7080 + }, + { + "epoch": 1.101273687480584, + "grad_norm": 0.4409906566143036, + "learning_rate": 9.988073947945502e-05, + "loss": 0.0164, + "step": 7090 + }, + { + "epoch": 1.1028269648959304, + "grad_norm": 0.37105226516723633, + "learning_rate": 9.987959541084087e-05, + "loss": 0.0189, + "step": 7100 + }, + { + "epoch": 1.1043802423112767, + "grad_norm": 0.20455071330070496, + "learning_rate": 9.987844588746915e-05, + "loss": 0.0163, + "step": 7110 + }, + { + "epoch": 1.1059335197266231, + "grad_norm": 0.5389917492866516, + "learning_rate": 9.987729090946558e-05, + "loss": 0.0183, + "step": 7120 + }, + { + "epoch": 1.1074867971419695, + "grad_norm": 0.44545549154281616, + "learning_rate": 9.987613047695647e-05, + "loss": 0.02, + "step": 7130 + }, + { + "epoch": 1.109040074557316, + "grad_norm": 0.24384364485740662, + "learning_rate": 9.987496459006871e-05, + "loss": 0.0169, + "step": 7140 + }, + { + "epoch": 1.1105933519726623, + "grad_norm": 0.44987204670906067, + "learning_rate": 9.987379324892982e-05, + "loss": 0.0202, + "step": 7150 + }, + { + "epoch": 1.1121466293880087, + "grad_norm": 0.20075516402721405, + "learning_rate": 9.987261645366788e-05, + "loss": 0.0191, + "step": 7160 + }, + { + "epoch": 1.113699906803355, + "grad_norm": 0.24144484102725983, + "learning_rate": 9.987143420441158e-05, + "loss": 0.0186, + "step": 7170 + }, + { + "epoch": 1.1152531842187015, + "grad_norm": 0.33520159125328064, + "learning_rate": 9.987024650129022e-05, + "loss": 0.0183, + "step": 7180 + }, + { + "epoch": 1.1168064616340478, + "grad_norm": 0.3396528959274292, + "learning_rate": 9.986905334443368e-05, + "loss": 0.0212, + "step": 7190 + }, + { + "epoch": 1.1183597390493942, + "grad_norm": 0.3367615342140198, + "learning_rate": 9.986785473397245e-05, + "loss": 0.0165, + "step": 7200 + }, + { + "epoch": 1.1199130164647406, + "grad_norm": 0.3289222717285156, + "learning_rate": 9.98666506700376e-05, + "loss": 0.0129, + "step": 7210 + }, + { + "epoch": 1.121466293880087, + "grad_norm": 0.4373385012149811, + "learning_rate": 9.986544115276081e-05, + "loss": 0.0176, + "step": 7220 + }, + { + "epoch": 1.1230195712954334, + "grad_norm": 0.4226805567741394, + "learning_rate": 9.986422618227433e-05, + "loss": 0.0156, + "step": 7230 + }, + { + "epoch": 1.1245728487107798, + "grad_norm": 0.23161908984184265, + "learning_rate": 9.986300575871106e-05, + "loss": 0.0129, + "step": 7240 + }, + { + "epoch": 1.1261261261261262, + "grad_norm": 0.22154763340950012, + "learning_rate": 9.986177988220444e-05, + "loss": 0.0127, + "step": 7250 + }, + { + "epoch": 1.1276794035414726, + "grad_norm": 0.2634393572807312, + "learning_rate": 9.986054855288856e-05, + "loss": 0.0143, + "step": 7260 + }, + { + "epoch": 1.129232680956819, + "grad_norm": 0.46643778681755066, + "learning_rate": 9.985931177089802e-05, + "loss": 0.0154, + "step": 7270 + }, + { + "epoch": 1.1307859583721653, + "grad_norm": 0.24202823638916016, + "learning_rate": 9.985806953636814e-05, + "loss": 0.0163, + "step": 7280 + }, + { + "epoch": 1.1323392357875117, + "grad_norm": 0.38412684202194214, + "learning_rate": 9.985682184943471e-05, + "loss": 0.0157, + "step": 7290 + }, + { + "epoch": 1.133892513202858, + "grad_norm": 0.2598573565483093, + "learning_rate": 9.98555687102342e-05, + "loss": 0.0149, + "step": 7300 + }, + { + "epoch": 1.1354457906182045, + "grad_norm": 0.2904340624809265, + "learning_rate": 9.985431011890367e-05, + "loss": 0.0186, + "step": 7310 + }, + { + "epoch": 1.1369990680335509, + "grad_norm": 0.5105819702148438, + "learning_rate": 9.985304607558075e-05, + "loss": 0.0143, + "step": 7320 + }, + { + "epoch": 1.1385523454488973, + "grad_norm": 0.3893519937992096, + "learning_rate": 9.985177658040364e-05, + "loss": 0.0206, + "step": 7330 + }, + { + "epoch": 1.1401056228642434, + "grad_norm": 0.3756752610206604, + "learning_rate": 9.985050163351119e-05, + "loss": 0.0145, + "step": 7340 + }, + { + "epoch": 1.14165890027959, + "grad_norm": 0.28373485803604126, + "learning_rate": 9.984922123504286e-05, + "loss": 0.015, + "step": 7350 + }, + { + "epoch": 1.1432121776949362, + "grad_norm": 0.3340517282485962, + "learning_rate": 9.984793538513862e-05, + "loss": 0.0157, + "step": 7360 + }, + { + "epoch": 1.1447654551102826, + "grad_norm": 0.29003044962882996, + "learning_rate": 9.984664408393912e-05, + "loss": 0.0151, + "step": 7370 + }, + { + "epoch": 1.146318732525629, + "grad_norm": 0.3689163029193878, + "learning_rate": 9.984534733158556e-05, + "loss": 0.0114, + "step": 7380 + }, + { + "epoch": 1.1478720099409754, + "grad_norm": 0.18055206537246704, + "learning_rate": 9.984404512821977e-05, + "loss": 0.0178, + "step": 7390 + }, + { + "epoch": 1.1494252873563218, + "grad_norm": 0.3216051459312439, + "learning_rate": 9.984273747398411e-05, + "loss": 0.0172, + "step": 7400 + }, + { + "epoch": 1.1509785647716682, + "grad_norm": 0.24145548045635223, + "learning_rate": 9.984142436902165e-05, + "loss": 0.0148, + "step": 7410 + }, + { + "epoch": 1.1525318421870145, + "grad_norm": 0.2502581477165222, + "learning_rate": 9.984010581347596e-05, + "loss": 0.0162, + "step": 7420 + }, + { + "epoch": 1.154085119602361, + "grad_norm": 0.33213433623313904, + "learning_rate": 9.983878180749121e-05, + "loss": 0.0189, + "step": 7430 + }, + { + "epoch": 1.1556383970177073, + "grad_norm": 0.271272212266922, + "learning_rate": 9.983745235121222e-05, + "loss": 0.019, + "step": 7440 + }, + { + "epoch": 1.1571916744330537, + "grad_norm": 0.30055496096611023, + "learning_rate": 9.983611744478438e-05, + "loss": 0.0128, + "step": 7450 + }, + { + "epoch": 1.1587449518484, + "grad_norm": 0.23646429181098938, + "learning_rate": 9.983477708835365e-05, + "loss": 0.0144, + "step": 7460 + }, + { + "epoch": 1.1602982292637465, + "grad_norm": 0.31205177307128906, + "learning_rate": 9.983343128206664e-05, + "loss": 0.0175, + "step": 7470 + }, + { + "epoch": 1.1618515066790929, + "grad_norm": 0.36643821001052856, + "learning_rate": 9.983208002607049e-05, + "loss": 0.0183, + "step": 7480 + }, + { + "epoch": 1.1634047840944393, + "grad_norm": 0.26584482192993164, + "learning_rate": 9.9830723320513e-05, + "loss": 0.0167, + "step": 7490 + }, + { + "epoch": 1.1649580615097856, + "grad_norm": 0.2748718857765198, + "learning_rate": 9.982936116554254e-05, + "loss": 0.0157, + "step": 7500 + }, + { + "epoch": 1.166511338925132, + "grad_norm": 0.5795344710350037, + "learning_rate": 9.982799356130803e-05, + "loss": 0.0225, + "step": 7510 + }, + { + "epoch": 1.1680646163404784, + "grad_norm": 0.4612305760383606, + "learning_rate": 9.982662050795908e-05, + "loss": 0.0205, + "step": 7520 + }, + { + "epoch": 1.1696178937558248, + "grad_norm": 0.3150828778743744, + "learning_rate": 9.982524200564583e-05, + "loss": 0.0207, + "step": 7530 + }, + { + "epoch": 1.1711711711711712, + "grad_norm": 0.2751251757144928, + "learning_rate": 9.982385805451901e-05, + "loss": 0.0169, + "step": 7540 + }, + { + "epoch": 1.1727244485865176, + "grad_norm": 0.36012545228004456, + "learning_rate": 9.982246865472998e-05, + "loss": 0.0147, + "step": 7550 + }, + { + "epoch": 1.174277726001864, + "grad_norm": 0.2633851170539856, + "learning_rate": 9.982107380643069e-05, + "loss": 0.0149, + "step": 7560 + }, + { + "epoch": 1.1758310034172104, + "grad_norm": 0.2528384327888489, + "learning_rate": 9.981967350977368e-05, + "loss": 0.015, + "step": 7570 + }, + { + "epoch": 1.1773842808325568, + "grad_norm": 0.470086008310318, + "learning_rate": 9.981826776491208e-05, + "loss": 0.0241, + "step": 7580 + }, + { + "epoch": 1.1789375582479031, + "grad_norm": 0.24368254840373993, + "learning_rate": 9.98168565719996e-05, + "loss": 0.0137, + "step": 7590 + }, + { + "epoch": 1.1804908356632495, + "grad_norm": 0.37256336212158203, + "learning_rate": 9.98154399311906e-05, + "loss": 0.0206, + "step": 7600 + }, + { + "epoch": 1.182044113078596, + "grad_norm": 0.32334524393081665, + "learning_rate": 9.981401784263997e-05, + "loss": 0.0163, + "step": 7610 + }, + { + "epoch": 1.1835973904939423, + "grad_norm": 0.3433332145214081, + "learning_rate": 9.981259030650326e-05, + "loss": 0.0179, + "step": 7620 + }, + { + "epoch": 1.1851506679092887, + "grad_norm": 0.21497423946857452, + "learning_rate": 9.981115732293655e-05, + "loss": 0.0184, + "step": 7630 + }, + { + "epoch": 1.186703945324635, + "grad_norm": 0.3469467759132385, + "learning_rate": 9.980971889209659e-05, + "loss": 0.0174, + "step": 7640 + }, + { + "epoch": 1.1882572227399812, + "grad_norm": 0.4563831388950348, + "learning_rate": 9.980827501414064e-05, + "loss": 0.0191, + "step": 7650 + }, + { + "epoch": 1.1898105001553279, + "grad_norm": 0.6407505869865417, + "learning_rate": 9.980682568922663e-05, + "loss": 0.0152, + "step": 7660 + }, + { + "epoch": 1.191363777570674, + "grad_norm": 0.3707462251186371, + "learning_rate": 9.980537091751304e-05, + "loss": 0.0192, + "step": 7670 + }, + { + "epoch": 1.1929170549860204, + "grad_norm": 0.29078441858291626, + "learning_rate": 9.980391069915897e-05, + "loss": 0.0189, + "step": 7680 + }, + { + "epoch": 1.1944703324013668, + "grad_norm": 0.42212405800819397, + "learning_rate": 9.98024450343241e-05, + "loss": 0.019, + "step": 7690 + }, + { + "epoch": 1.1960236098167132, + "grad_norm": 0.4724245071411133, + "learning_rate": 9.980097392316872e-05, + "loss": 0.0176, + "step": 7700 + }, + { + "epoch": 1.1975768872320596, + "grad_norm": 0.5354567170143127, + "learning_rate": 9.97994973658537e-05, + "loss": 0.0184, + "step": 7710 + }, + { + "epoch": 1.199130164647406, + "grad_norm": 0.4048025608062744, + "learning_rate": 9.979801536254054e-05, + "loss": 0.0186, + "step": 7720 + }, + { + "epoch": 1.2006834420627523, + "grad_norm": 0.20856048166751862, + "learning_rate": 9.979652791339127e-05, + "loss": 0.0176, + "step": 7730 + }, + { + "epoch": 1.2022367194780987, + "grad_norm": 0.28858667612075806, + "learning_rate": 9.97950350185686e-05, + "loss": 0.0171, + "step": 7740 + }, + { + "epoch": 1.2037899968934451, + "grad_norm": 0.2554994821548462, + "learning_rate": 9.979353667823574e-05, + "loss": 0.0165, + "step": 7750 + }, + { + "epoch": 1.2053432743087915, + "grad_norm": 0.3354344367980957, + "learning_rate": 9.979203289255658e-05, + "loss": 0.0171, + "step": 7760 + }, + { + "epoch": 1.206896551724138, + "grad_norm": 0.18710562586784363, + "learning_rate": 9.979052366169557e-05, + "loss": 0.0108, + "step": 7770 + }, + { + "epoch": 1.2084498291394843, + "grad_norm": 0.31552937626838684, + "learning_rate": 9.978900898581775e-05, + "loss": 0.0169, + "step": 7780 + }, + { + "epoch": 1.2100031065548307, + "grad_norm": 0.2555586099624634, + "learning_rate": 9.978748886508875e-05, + "loss": 0.0159, + "step": 7790 + }, + { + "epoch": 1.211556383970177, + "grad_norm": 0.3226424753665924, + "learning_rate": 9.978596329967484e-05, + "loss": 0.0193, + "step": 7800 + }, + { + "epoch": 1.2131096613855235, + "grad_norm": 0.2536790668964386, + "learning_rate": 9.978443228974284e-05, + "loss": 0.0174, + "step": 7810 + }, + { + "epoch": 1.2146629388008698, + "grad_norm": 0.4822244942188263, + "learning_rate": 9.978289583546015e-05, + "loss": 0.0213, + "step": 7820 + }, + { + "epoch": 1.2162162162162162, + "grad_norm": 0.31640732288360596, + "learning_rate": 9.978135393699484e-05, + "loss": 0.021, + "step": 7830 + }, + { + "epoch": 1.2177694936315626, + "grad_norm": 0.34465089440345764, + "learning_rate": 9.977980659451548e-05, + "loss": 0.0165, + "step": 7840 + }, + { + "epoch": 1.219322771046909, + "grad_norm": 0.3761986494064331, + "learning_rate": 9.977825380819135e-05, + "loss": 0.0204, + "step": 7850 + }, + { + "epoch": 1.2208760484622554, + "grad_norm": 0.33944863080978394, + "learning_rate": 9.97766955781922e-05, + "loss": 0.013, + "step": 7860 + }, + { + "epoch": 1.2224293258776018, + "grad_norm": 0.4596520960330963, + "learning_rate": 9.977513190468848e-05, + "loss": 0.0195, + "step": 7870 + }, + { + "epoch": 1.2239826032929482, + "grad_norm": 0.40122222900390625, + "learning_rate": 9.977356278785116e-05, + "loss": 0.0168, + "step": 7880 + }, + { + "epoch": 1.2255358807082946, + "grad_norm": 0.27352824807167053, + "learning_rate": 9.977198822785184e-05, + "loss": 0.0146, + "step": 7890 + }, + { + "epoch": 1.227089158123641, + "grad_norm": 0.39735284447669983, + "learning_rate": 9.977040822486273e-05, + "loss": 0.017, + "step": 7900 + }, + { + "epoch": 1.2286424355389873, + "grad_norm": 0.41000816226005554, + "learning_rate": 9.97688227790566e-05, + "loss": 0.0163, + "step": 7910 + }, + { + "epoch": 1.2301957129543337, + "grad_norm": 0.31276199221611023, + "learning_rate": 9.976723189060684e-05, + "loss": 0.0147, + "step": 7920 + }, + { + "epoch": 1.23174899036968, + "grad_norm": 0.3164517283439636, + "learning_rate": 9.976563555968742e-05, + "loss": 0.0152, + "step": 7930 + }, + { + "epoch": 1.2333022677850265, + "grad_norm": 0.3896557092666626, + "learning_rate": 9.976403378647292e-05, + "loss": 0.019, + "step": 7940 + }, + { + "epoch": 1.2348555452003729, + "grad_norm": 0.39297956228256226, + "learning_rate": 9.97624265711385e-05, + "loss": 0.0254, + "step": 7950 + }, + { + "epoch": 1.236408822615719, + "grad_norm": 0.5398741364479065, + "learning_rate": 9.976081391385993e-05, + "loss": 0.0195, + "step": 7960 + }, + { + "epoch": 1.2379621000310657, + "grad_norm": 0.29761239886283875, + "learning_rate": 9.975919581481356e-05, + "loss": 0.0137, + "step": 7970 + }, + { + "epoch": 1.2395153774464118, + "grad_norm": 0.26741668581962585, + "learning_rate": 9.975757227417634e-05, + "loss": 0.0194, + "step": 7980 + }, + { + "epoch": 1.2410686548617582, + "grad_norm": 0.31487807631492615, + "learning_rate": 9.975594329212586e-05, + "loss": 0.0156, + "step": 7990 + }, + { + "epoch": 1.2426219322771046, + "grad_norm": 0.4104226529598236, + "learning_rate": 9.97543088688402e-05, + "loss": 0.0143, + "step": 8000 + }, + { + "epoch": 1.244175209692451, + "grad_norm": 0.3838336765766144, + "learning_rate": 9.975266900449814e-05, + "loss": 0.0179, + "step": 8010 + }, + { + "epoch": 1.2457284871077974, + "grad_norm": 0.42922329902648926, + "learning_rate": 9.975102369927898e-05, + "loss": 0.0174, + "step": 8020 + }, + { + "epoch": 1.2472817645231438, + "grad_norm": 0.40781643986701965, + "learning_rate": 9.974937295336269e-05, + "loss": 0.0172, + "step": 8030 + }, + { + "epoch": 1.2488350419384902, + "grad_norm": 0.5519856214523315, + "learning_rate": 9.974771676692975e-05, + "loss": 0.0164, + "step": 8040 + }, + { + "epoch": 1.2503883193538365, + "grad_norm": 0.4289715588092804, + "learning_rate": 9.974605514016131e-05, + "loss": 0.0162, + "step": 8050 + }, + { + "epoch": 1.251941596769183, + "grad_norm": 0.2688447833061218, + "learning_rate": 9.974438807323907e-05, + "loss": 0.0183, + "step": 8060 + }, + { + "epoch": 1.2534948741845293, + "grad_norm": 0.4818929135799408, + "learning_rate": 9.974271556634535e-05, + "loss": 0.0174, + "step": 8070 + }, + { + "epoch": 1.2550481515998757, + "grad_norm": 0.4088376462459564, + "learning_rate": 9.974103761966302e-05, + "loss": 0.0144, + "step": 8080 + }, + { + "epoch": 1.256601429015222, + "grad_norm": 0.22220325469970703, + "learning_rate": 9.973935423337563e-05, + "loss": 0.0163, + "step": 8090 + }, + { + "epoch": 1.2581547064305685, + "grad_norm": 0.5348305106163025, + "learning_rate": 9.973766540766722e-05, + "loss": 0.0149, + "step": 8100 + }, + { + "epoch": 1.2597079838459149, + "grad_norm": 0.335131973028183, + "learning_rate": 9.97359711427225e-05, + "loss": 0.016, + "step": 8110 + }, + { + "epoch": 1.2612612612612613, + "grad_norm": 0.4181067943572998, + "learning_rate": 9.973427143872677e-05, + "loss": 0.0155, + "step": 8120 + }, + { + "epoch": 1.2628145386766076, + "grad_norm": 0.4384543001651764, + "learning_rate": 9.973256629586589e-05, + "loss": 0.0172, + "step": 8130 + }, + { + "epoch": 1.264367816091954, + "grad_norm": 0.2145700305700302, + "learning_rate": 9.973085571432632e-05, + "loss": 0.0165, + "step": 8140 + }, + { + "epoch": 1.2659210935073004, + "grad_norm": 0.30765897035598755, + "learning_rate": 9.972913969429513e-05, + "loss": 0.016, + "step": 8150 + }, + { + "epoch": 1.2674743709226468, + "grad_norm": 0.27688372135162354, + "learning_rate": 9.972741823596e-05, + "loss": 0.0168, + "step": 8160 + }, + { + "epoch": 1.2690276483379932, + "grad_norm": 0.38305509090423584, + "learning_rate": 9.972569133950917e-05, + "loss": 0.0158, + "step": 8170 + }, + { + "epoch": 1.2705809257533396, + "grad_norm": 0.2817029058933258, + "learning_rate": 9.972395900513151e-05, + "loss": 0.0141, + "step": 8180 + }, + { + "epoch": 1.272134203168686, + "grad_norm": 0.28056272864341736, + "learning_rate": 9.972222123301645e-05, + "loss": 0.015, + "step": 8190 + }, + { + "epoch": 1.2736874805840324, + "grad_norm": 0.3753458559513092, + "learning_rate": 9.972047802335403e-05, + "loss": 0.0172, + "step": 8200 + }, + { + "epoch": 1.2752407579993787, + "grad_norm": 0.4194175601005554, + "learning_rate": 9.971872937633488e-05, + "loss": 0.0156, + "step": 8210 + }, + { + "epoch": 1.2767940354147251, + "grad_norm": 0.4183066189289093, + "learning_rate": 9.971697529215024e-05, + "loss": 0.0153, + "step": 8220 + }, + { + "epoch": 1.2783473128300715, + "grad_norm": 0.5250834226608276, + "learning_rate": 9.971521577099192e-05, + "loss": 0.0164, + "step": 8230 + }, + { + "epoch": 1.279900590245418, + "grad_norm": 0.36224794387817383, + "learning_rate": 9.971345081305236e-05, + "loss": 0.012, + "step": 8240 + }, + { + "epoch": 1.281453867660764, + "grad_norm": 0.41289451718330383, + "learning_rate": 9.971168041852456e-05, + "loss": 0.0142, + "step": 8250 + }, + { + "epoch": 1.2830071450761107, + "grad_norm": 0.4915356934070587, + "learning_rate": 9.970990458760215e-05, + "loss": 0.0147, + "step": 8260 + }, + { + "epoch": 1.2845604224914569, + "grad_norm": 0.43263867497444153, + "learning_rate": 9.970812332047929e-05, + "loss": 0.0177, + "step": 8270 + }, + { + "epoch": 1.2861136999068035, + "grad_norm": 0.5074807405471802, + "learning_rate": 9.97063366173508e-05, + "loss": 0.0146, + "step": 8280 + }, + { + "epoch": 1.2876669773221496, + "grad_norm": 0.37113043665885925, + "learning_rate": 9.970454447841207e-05, + "loss": 0.0176, + "step": 8290 + }, + { + "epoch": 1.2892202547374962, + "grad_norm": 0.31569188833236694, + "learning_rate": 9.970274690385909e-05, + "loss": 0.0163, + "step": 8300 + }, + { + "epoch": 1.2907735321528424, + "grad_norm": 0.26400870084762573, + "learning_rate": 9.970094389388844e-05, + "loss": 0.0187, + "step": 8310 + }, + { + "epoch": 1.292326809568189, + "grad_norm": 0.22425170242786407, + "learning_rate": 9.969913544869728e-05, + "loss": 0.0204, + "step": 8320 + }, + { + "epoch": 1.2938800869835352, + "grad_norm": 0.38154277205467224, + "learning_rate": 9.96973215684834e-05, + "loss": 0.0183, + "step": 8330 + }, + { + "epoch": 1.2954333643988816, + "grad_norm": 0.6668344736099243, + "learning_rate": 9.969550225344513e-05, + "loss": 0.0178, + "step": 8340 + }, + { + "epoch": 1.296986641814228, + "grad_norm": 0.4948280155658722, + "learning_rate": 9.969367750378147e-05, + "loss": 0.0174, + "step": 8350 + }, + { + "epoch": 1.2985399192295743, + "grad_norm": 0.47617673873901367, + "learning_rate": 9.969184731969194e-05, + "loss": 0.0169, + "step": 8360 + }, + { + "epoch": 1.3000931966449207, + "grad_norm": 0.27045583724975586, + "learning_rate": 9.96900117013767e-05, + "loss": 0.012, + "step": 8370 + }, + { + "epoch": 1.3016464740602671, + "grad_norm": 0.48912015557289124, + "learning_rate": 9.96881706490365e-05, + "loss": 0.0175, + "step": 8380 + }, + { + "epoch": 1.3031997514756135, + "grad_norm": 0.5312341451644897, + "learning_rate": 9.968632416287265e-05, + "loss": 0.0196, + "step": 8390 + }, + { + "epoch": 1.30475302889096, + "grad_norm": 0.17496709525585175, + "learning_rate": 9.96844722430871e-05, + "loss": 0.0117, + "step": 8400 + }, + { + "epoch": 1.3063063063063063, + "grad_norm": 0.249140664935112, + "learning_rate": 9.968261488988235e-05, + "loss": 0.0148, + "step": 8410 + }, + { + "epoch": 1.3078595837216527, + "grad_norm": 0.22597871720790863, + "learning_rate": 9.968075210346155e-05, + "loss": 0.0116, + "step": 8420 + }, + { + "epoch": 1.309412861136999, + "grad_norm": 0.2035120278596878, + "learning_rate": 9.967888388402839e-05, + "loss": 0.0156, + "step": 8430 + }, + { + "epoch": 1.3109661385523455, + "grad_norm": 0.3695875406265259, + "learning_rate": 9.967701023178717e-05, + "loss": 0.0158, + "step": 8440 + }, + { + "epoch": 1.3125194159676918, + "grad_norm": 0.28234195709228516, + "learning_rate": 9.967513114694282e-05, + "loss": 0.016, + "step": 8450 + }, + { + "epoch": 1.3140726933830382, + "grad_norm": 0.3503580689430237, + "learning_rate": 9.967324662970079e-05, + "loss": 0.0131, + "step": 8460 + }, + { + "epoch": 1.3156259707983846, + "grad_norm": 0.24213959276676178, + "learning_rate": 9.96713566802672e-05, + "loss": 0.0144, + "step": 8470 + }, + { + "epoch": 1.317179248213731, + "grad_norm": 0.2769998013973236, + "learning_rate": 9.966946129884873e-05, + "loss": 0.0195, + "step": 8480 + }, + { + "epoch": 1.3187325256290774, + "grad_norm": 0.2774069011211395, + "learning_rate": 9.966756048565265e-05, + "loss": 0.0167, + "step": 8490 + }, + { + "epoch": 1.3202858030444238, + "grad_norm": 0.5085484981536865, + "learning_rate": 9.966565424088681e-05, + "loss": 0.0168, + "step": 8500 + }, + { + "epoch": 1.3218390804597702, + "grad_norm": 0.3553764820098877, + "learning_rate": 9.96637425647597e-05, + "loss": 0.0178, + "step": 8510 + }, + { + "epoch": 1.3233923578751166, + "grad_norm": 0.30803221464157104, + "learning_rate": 9.966182545748038e-05, + "loss": 0.0146, + "step": 8520 + }, + { + "epoch": 1.324945635290463, + "grad_norm": 0.5888924598693848, + "learning_rate": 9.96599029192585e-05, + "loss": 0.0144, + "step": 8530 + }, + { + "epoch": 1.3264989127058093, + "grad_norm": 0.24805815517902374, + "learning_rate": 9.965797495030428e-05, + "loss": 0.019, + "step": 8540 + }, + { + "epoch": 1.3280521901211557, + "grad_norm": 0.3487095236778259, + "learning_rate": 9.96560415508286e-05, + "loss": 0.0161, + "step": 8550 + }, + { + "epoch": 1.3296054675365019, + "grad_norm": 0.26895055174827576, + "learning_rate": 9.965410272104286e-05, + "loss": 0.0167, + "step": 8560 + }, + { + "epoch": 1.3311587449518485, + "grad_norm": 0.3590301275253296, + "learning_rate": 9.96521584611591e-05, + "loss": 0.0144, + "step": 8570 + }, + { + "epoch": 1.3327120223671947, + "grad_norm": 0.4134247601032257, + "learning_rate": 9.965020877138994e-05, + "loss": 0.0164, + "step": 8580 + }, + { + "epoch": 1.3342652997825413, + "grad_norm": 0.37142398953437805, + "learning_rate": 9.964825365194861e-05, + "loss": 0.0149, + "step": 8590 + }, + { + "epoch": 1.3358185771978874, + "grad_norm": 0.38487711548805237, + "learning_rate": 9.96462931030489e-05, + "loss": 0.0186, + "step": 8600 + }, + { + "epoch": 1.337371854613234, + "grad_norm": 0.35946446657180786, + "learning_rate": 9.96443271249052e-05, + "loss": 0.0171, + "step": 8610 + }, + { + "epoch": 1.3389251320285802, + "grad_norm": 0.3276605010032654, + "learning_rate": 9.964235571773255e-05, + "loss": 0.0176, + "step": 8620 + }, + { + "epoch": 1.3404784094439268, + "grad_norm": 0.6908092498779297, + "learning_rate": 9.96403788817465e-05, + "loss": 0.013, + "step": 8630 + }, + { + "epoch": 1.342031686859273, + "grad_norm": 0.26217830181121826, + "learning_rate": 9.963839661716325e-05, + "loss": 0.0156, + "step": 8640 + }, + { + "epoch": 1.3435849642746194, + "grad_norm": 0.4615252614021301, + "learning_rate": 9.963640892419958e-05, + "loss": 0.0176, + "step": 8650 + }, + { + "epoch": 1.3451382416899658, + "grad_norm": 0.33472633361816406, + "learning_rate": 9.963441580307286e-05, + "loss": 0.0162, + "step": 8660 + }, + { + "epoch": 1.3466915191053122, + "grad_norm": 0.3033904433250427, + "learning_rate": 9.963241725400104e-05, + "loss": 0.0182, + "step": 8670 + }, + { + "epoch": 1.3482447965206585, + "grad_norm": 0.36833086609840393, + "learning_rate": 9.963041327720271e-05, + "loss": 0.0175, + "step": 8680 + }, + { + "epoch": 1.349798073936005, + "grad_norm": 0.37706124782562256, + "learning_rate": 9.962840387289697e-05, + "loss": 0.0142, + "step": 8690 + }, + { + "epoch": 1.3513513513513513, + "grad_norm": 0.3331157863140106, + "learning_rate": 9.962638904130363e-05, + "loss": 0.0189, + "step": 8700 + }, + { + "epoch": 1.3529046287666977, + "grad_norm": 0.2747398912906647, + "learning_rate": 9.962436878264298e-05, + "loss": 0.0143, + "step": 8710 + }, + { + "epoch": 1.354457906182044, + "grad_norm": 0.27341294288635254, + "learning_rate": 9.962234309713598e-05, + "loss": 0.0176, + "step": 8720 + }, + { + "epoch": 1.3560111835973905, + "grad_norm": 0.33686259388923645, + "learning_rate": 9.962031198500414e-05, + "loss": 0.0169, + "step": 8730 + }, + { + "epoch": 1.3575644610127369, + "grad_norm": 0.32250097393989563, + "learning_rate": 9.961827544646958e-05, + "loss": 0.0136, + "step": 8740 + }, + { + "epoch": 1.3591177384280833, + "grad_norm": 0.3414478898048401, + "learning_rate": 9.961623348175501e-05, + "loss": 0.0153, + "step": 8750 + }, + { + "epoch": 1.3606710158434296, + "grad_norm": 0.42749693989753723, + "learning_rate": 9.961418609108377e-05, + "loss": 0.0184, + "step": 8760 + }, + { + "epoch": 1.362224293258776, + "grad_norm": 0.5277280807495117, + "learning_rate": 9.961213327467971e-05, + "loss": 0.0153, + "step": 8770 + }, + { + "epoch": 1.3637775706741224, + "grad_norm": 0.44496792554855347, + "learning_rate": 9.961007503276736e-05, + "loss": 0.0139, + "step": 8780 + }, + { + "epoch": 1.3653308480894688, + "grad_norm": 0.25884342193603516, + "learning_rate": 9.960801136557179e-05, + "loss": 0.0141, + "step": 8790 + }, + { + "epoch": 1.3668841255048152, + "grad_norm": 0.5373533964157104, + "learning_rate": 9.960594227331866e-05, + "loss": 0.0159, + "step": 8800 + }, + { + "epoch": 1.3684374029201616, + "grad_norm": 0.29431861639022827, + "learning_rate": 9.960386775623429e-05, + "loss": 0.0167, + "step": 8810 + }, + { + "epoch": 1.369990680335508, + "grad_norm": 0.5235652923583984, + "learning_rate": 9.96017878145455e-05, + "loss": 0.0154, + "step": 8820 + }, + { + "epoch": 1.3715439577508544, + "grad_norm": 0.3219124674797058, + "learning_rate": 9.959970244847977e-05, + "loss": 0.0166, + "step": 8830 + }, + { + "epoch": 1.3730972351662007, + "grad_norm": 0.3323783278465271, + "learning_rate": 9.959761165826518e-05, + "loss": 0.0176, + "step": 8840 + }, + { + "epoch": 1.3746505125815471, + "grad_norm": 0.49701783061027527, + "learning_rate": 9.959551544413033e-05, + "loss": 0.0158, + "step": 8850 + }, + { + "epoch": 1.3762037899968935, + "grad_norm": 0.31064170598983765, + "learning_rate": 9.959341380630448e-05, + "loss": 0.0153, + "step": 8860 + }, + { + "epoch": 1.3777570674122397, + "grad_norm": 0.37811097502708435, + "learning_rate": 9.959130674501746e-05, + "loss": 0.0166, + "step": 8870 + }, + { + "epoch": 1.3793103448275863, + "grad_norm": 0.3092891275882721, + "learning_rate": 9.958919426049968e-05, + "loss": 0.0146, + "step": 8880 + }, + { + "epoch": 1.3808636222429325, + "grad_norm": 0.27060750126838684, + "learning_rate": 9.958707635298219e-05, + "loss": 0.0162, + "step": 8890 + }, + { + "epoch": 1.382416899658279, + "grad_norm": 0.3955566883087158, + "learning_rate": 9.958495302269657e-05, + "loss": 0.0192, + "step": 8900 + }, + { + "epoch": 1.3839701770736252, + "grad_norm": 0.2686033248901367, + "learning_rate": 9.958282426987503e-05, + "loss": 0.0142, + "step": 8910 + }, + { + "epoch": 1.3855234544889719, + "grad_norm": 0.3138161599636078, + "learning_rate": 9.95806900947504e-05, + "loss": 0.0119, + "step": 8920 + }, + { + "epoch": 1.387076731904318, + "grad_norm": 0.33933010697364807, + "learning_rate": 9.957855049755604e-05, + "loss": 0.0148, + "step": 8930 + }, + { + "epoch": 1.3886300093196646, + "grad_norm": 0.22111253440380096, + "learning_rate": 9.957640547852593e-05, + "loss": 0.0137, + "step": 8940 + }, + { + "epoch": 1.3901832867350108, + "grad_norm": 0.4493364691734314, + "learning_rate": 9.957425503789466e-05, + "loss": 0.0197, + "step": 8950 + }, + { + "epoch": 1.3917365641503572, + "grad_norm": 0.3273703157901764, + "learning_rate": 9.957209917589738e-05, + "loss": 0.0146, + "step": 8960 + }, + { + "epoch": 1.3932898415657036, + "grad_norm": 0.4371713399887085, + "learning_rate": 9.956993789276987e-05, + "loss": 0.0129, + "step": 8970 + }, + { + "epoch": 1.39484311898105, + "grad_norm": 0.2751273214817047, + "learning_rate": 9.956777118874847e-05, + "loss": 0.0164, + "step": 8980 + }, + { + "epoch": 1.3963963963963963, + "grad_norm": 0.22952383756637573, + "learning_rate": 9.956559906407016e-05, + "loss": 0.017, + "step": 8990 + }, + { + "epoch": 1.3979496738117427, + "grad_norm": 0.22934183478355408, + "learning_rate": 9.956342151897245e-05, + "loss": 0.0133, + "step": 9000 + }, + { + "epoch": 1.3995029512270891, + "grad_norm": 0.24275462329387665, + "learning_rate": 9.956123855369346e-05, + "loss": 0.0119, + "step": 9010 + }, + { + "epoch": 1.4010562286424355, + "grad_norm": 0.29522305727005005, + "learning_rate": 9.955905016847196e-05, + "loss": 0.0139, + "step": 9020 + }, + { + "epoch": 1.402609506057782, + "grad_norm": 0.38670527935028076, + "learning_rate": 9.955685636354723e-05, + "loss": 0.0131, + "step": 9030 + }, + { + "epoch": 1.4041627834731283, + "grad_norm": 0.28863754868507385, + "learning_rate": 9.95546571391592e-05, + "loss": 0.0126, + "step": 9040 + }, + { + "epoch": 1.4057160608884747, + "grad_norm": 0.2782325744628906, + "learning_rate": 9.955245249554837e-05, + "loss": 0.0162, + "step": 9050 + }, + { + "epoch": 1.407269338303821, + "grad_norm": 0.22019727528095245, + "learning_rate": 9.955024243295582e-05, + "loss": 0.0148, + "step": 9060 + }, + { + "epoch": 1.4088226157191674, + "grad_norm": 0.28896504640579224, + "learning_rate": 9.954802695162328e-05, + "loss": 0.0121, + "step": 9070 + }, + { + "epoch": 1.4103758931345138, + "grad_norm": 0.3450586497783661, + "learning_rate": 9.954580605179302e-05, + "loss": 0.019, + "step": 9080 + }, + { + "epoch": 1.4119291705498602, + "grad_norm": 0.23975376784801483, + "learning_rate": 9.954357973370788e-05, + "loss": 0.0169, + "step": 9090 + }, + { + "epoch": 1.4134824479652066, + "grad_norm": 0.35354459285736084, + "learning_rate": 9.954134799761135e-05, + "loss": 0.0161, + "step": 9100 + }, + { + "epoch": 1.415035725380553, + "grad_norm": 0.3178999722003937, + "learning_rate": 9.953911084374748e-05, + "loss": 0.0147, + "step": 9110 + }, + { + "epoch": 1.4165890027958994, + "grad_norm": 0.4045858085155487, + "learning_rate": 9.953686827236093e-05, + "loss": 0.0154, + "step": 9120 + }, + { + "epoch": 1.4181422802112458, + "grad_norm": 0.214920312166214, + "learning_rate": 9.953462028369695e-05, + "loss": 0.0143, + "step": 9130 + }, + { + "epoch": 1.4196955576265922, + "grad_norm": 0.28926336765289307, + "learning_rate": 9.953236687800136e-05, + "loss": 0.0147, + "step": 9140 + }, + { + "epoch": 1.4212488350419386, + "grad_norm": 0.3791525959968567, + "learning_rate": 9.95301080555206e-05, + "loss": 0.0162, + "step": 9150 + }, + { + "epoch": 1.422802112457285, + "grad_norm": 0.50425124168396, + "learning_rate": 9.952784381650171e-05, + "loss": 0.0151, + "step": 9160 + }, + { + "epoch": 1.4243553898726313, + "grad_norm": 0.5183606147766113, + "learning_rate": 9.952557416119226e-05, + "loss": 0.0143, + "step": 9170 + }, + { + "epoch": 1.4259086672879777, + "grad_norm": 0.36352968215942383, + "learning_rate": 9.95232990898405e-05, + "loss": 0.0162, + "step": 9180 + }, + { + "epoch": 1.427461944703324, + "grad_norm": 0.2501816153526306, + "learning_rate": 9.95210186026952e-05, + "loss": 0.016, + "step": 9190 + }, + { + "epoch": 1.4290152221186703, + "grad_norm": 0.39076825976371765, + "learning_rate": 9.951873270000576e-05, + "loss": 0.0198, + "step": 9200 + }, + { + "epoch": 1.4305684995340169, + "grad_norm": 0.717907726764679, + "learning_rate": 9.951644138202216e-05, + "loss": 0.0161, + "step": 9210 + }, + { + "epoch": 1.432121776949363, + "grad_norm": 0.46513524651527405, + "learning_rate": 9.951414464899498e-05, + "loss": 0.0238, + "step": 9220 + }, + { + "epoch": 1.4336750543647097, + "grad_norm": 0.39053118228912354, + "learning_rate": 9.951184250117538e-05, + "loss": 0.0168, + "step": 9230 + }, + { + "epoch": 1.4352283317800558, + "grad_norm": 0.40763354301452637, + "learning_rate": 9.950953493881513e-05, + "loss": 0.0189, + "step": 9240 + }, + { + "epoch": 1.4367816091954024, + "grad_norm": 0.31702491641044617, + "learning_rate": 9.950722196216658e-05, + "loss": 0.0147, + "step": 9250 + }, + { + "epoch": 1.4383348866107486, + "grad_norm": 0.4224323332309723, + "learning_rate": 9.950490357148265e-05, + "loss": 0.0191, + "step": 9260 + }, + { + "epoch": 1.439888164026095, + "grad_norm": 0.3443903625011444, + "learning_rate": 9.950257976701692e-05, + "loss": 0.0163, + "step": 9270 + }, + { + "epoch": 1.4414414414414414, + "grad_norm": 0.3258558213710785, + "learning_rate": 9.950025054902348e-05, + "loss": 0.0142, + "step": 9280 + }, + { + "epoch": 1.4429947188567878, + "grad_norm": 0.31932905316352844, + "learning_rate": 9.949791591775706e-05, + "loss": 0.0145, + "step": 9290 + }, + { + "epoch": 1.4445479962721341, + "grad_norm": 0.34017977118492126, + "learning_rate": 9.949557587347298e-05, + "loss": 0.0164, + "step": 9300 + }, + { + "epoch": 1.4461012736874805, + "grad_norm": 0.3894646465778351, + "learning_rate": 9.949323041642713e-05, + "loss": 0.0157, + "step": 9310 + }, + { + "epoch": 1.447654551102827, + "grad_norm": 0.2983075678348541, + "learning_rate": 9.949087954687602e-05, + "loss": 0.016, + "step": 9320 + }, + { + "epoch": 1.4492078285181733, + "grad_norm": 0.45601215958595276, + "learning_rate": 9.948852326507672e-05, + "loss": 0.0153, + "step": 9330 + }, + { + "epoch": 1.4507611059335197, + "grad_norm": 0.20842038094997406, + "learning_rate": 9.948616157128694e-05, + "loss": 0.0154, + "step": 9340 + }, + { + "epoch": 1.452314383348866, + "grad_norm": 0.15462933480739594, + "learning_rate": 9.948379446576493e-05, + "loss": 0.0128, + "step": 9350 + }, + { + "epoch": 1.4538676607642125, + "grad_norm": 0.24067743122577667, + "learning_rate": 9.948142194876952e-05, + "loss": 0.0181, + "step": 9360 + }, + { + "epoch": 1.4554209381795589, + "grad_norm": 0.39767634868621826, + "learning_rate": 9.947904402056024e-05, + "loss": 0.0166, + "step": 9370 + }, + { + "epoch": 1.4569742155949053, + "grad_norm": 0.3803887367248535, + "learning_rate": 9.947666068139708e-05, + "loss": 0.0175, + "step": 9380 + }, + { + "epoch": 1.4585274930102516, + "grad_norm": 0.2448764443397522, + "learning_rate": 9.947427193154071e-05, + "loss": 0.0148, + "step": 9390 + }, + { + "epoch": 1.460080770425598, + "grad_norm": 0.3308228552341461, + "learning_rate": 9.947187777125233e-05, + "loss": 0.0137, + "step": 9400 + }, + { + "epoch": 1.4616340478409444, + "grad_norm": 0.2494431883096695, + "learning_rate": 9.946947820079377e-05, + "loss": 0.0174, + "step": 9410 + }, + { + "epoch": 1.4631873252562908, + "grad_norm": 0.220392107963562, + "learning_rate": 9.946707322042747e-05, + "loss": 0.0191, + "step": 9420 + }, + { + "epoch": 1.4647406026716372, + "grad_norm": 0.3174278736114502, + "learning_rate": 9.94646628304164e-05, + "loss": 0.0139, + "step": 9430 + }, + { + "epoch": 1.4662938800869836, + "grad_norm": 0.4605891704559326, + "learning_rate": 9.946224703102418e-05, + "loss": 0.0143, + "step": 9440 + }, + { + "epoch": 1.46784715750233, + "grad_norm": 0.24672187864780426, + "learning_rate": 9.945982582251498e-05, + "loss": 0.0155, + "step": 9450 + }, + { + "epoch": 1.4694004349176764, + "grad_norm": 0.3509500026702881, + "learning_rate": 9.94573992051536e-05, + "loss": 0.0174, + "step": 9460 + }, + { + "epoch": 1.4709537123330227, + "grad_norm": 0.35315772891044617, + "learning_rate": 9.94549671792054e-05, + "loss": 0.0165, + "step": 9470 + }, + { + "epoch": 1.4725069897483691, + "grad_norm": 0.397389680147171, + "learning_rate": 9.945252974493635e-05, + "loss": 0.0154, + "step": 9480 + }, + { + "epoch": 1.4740602671637155, + "grad_norm": 0.2170630395412445, + "learning_rate": 9.9450086902613e-05, + "loss": 0.0127, + "step": 9490 + }, + { + "epoch": 1.475613544579062, + "grad_norm": 0.27053818106651306, + "learning_rate": 9.944763865250248e-05, + "loss": 0.0147, + "step": 9500 + }, + { + "epoch": 1.477166821994408, + "grad_norm": 0.2970278561115265, + "learning_rate": 9.944518499487254e-05, + "loss": 0.0141, + "step": 9510 + }, + { + "epoch": 1.4787200994097547, + "grad_norm": 0.2296913117170334, + "learning_rate": 9.944272592999151e-05, + "loss": 0.015, + "step": 9520 + }, + { + "epoch": 1.4802733768251009, + "grad_norm": 0.2973487973213196, + "learning_rate": 9.94402614581283e-05, + "loss": 0.0151, + "step": 9530 + }, + { + "epoch": 1.4818266542404475, + "grad_norm": 0.11543003469705582, + "learning_rate": 9.943779157955244e-05, + "loss": 0.0146, + "step": 9540 + }, + { + "epoch": 1.4833799316557936, + "grad_norm": 0.35809093713760376, + "learning_rate": 9.943531629453403e-05, + "loss": 0.0169, + "step": 9550 + }, + { + "epoch": 1.4849332090711402, + "grad_norm": 0.26831549406051636, + "learning_rate": 9.943283560334375e-05, + "loss": 0.0153, + "step": 9560 + }, + { + "epoch": 1.4864864864864864, + "grad_norm": 0.29438266158103943, + "learning_rate": 9.943034950625288e-05, + "loss": 0.0143, + "step": 9570 + }, + { + "epoch": 1.4880397639018328, + "grad_norm": 0.42799460887908936, + "learning_rate": 9.942785800353332e-05, + "loss": 0.0126, + "step": 9580 + }, + { + "epoch": 1.4895930413171792, + "grad_norm": 0.2948685586452484, + "learning_rate": 9.942536109545751e-05, + "loss": 0.0119, + "step": 9590 + }, + { + "epoch": 1.4911463187325256, + "grad_norm": 0.1868930160999298, + "learning_rate": 9.942285878229853e-05, + "loss": 0.0122, + "step": 9600 + }, + { + "epoch": 1.492699596147872, + "grad_norm": 0.392236590385437, + "learning_rate": 9.942035106433001e-05, + "loss": 0.0164, + "step": 9610 + }, + { + "epoch": 1.4942528735632183, + "grad_norm": 0.20153376460075378, + "learning_rate": 9.94178379418262e-05, + "loss": 0.0175, + "step": 9620 + }, + { + "epoch": 1.4958061509785647, + "grad_norm": 0.2883988320827484, + "learning_rate": 9.941531941506194e-05, + "loss": 0.0155, + "step": 9630 + }, + { + "epoch": 1.4973594283939111, + "grad_norm": 0.37317800521850586, + "learning_rate": 9.941279548431263e-05, + "loss": 0.0163, + "step": 9640 + }, + { + "epoch": 1.4989127058092575, + "grad_norm": 0.22636477649211884, + "learning_rate": 9.941026614985431e-05, + "loss": 0.0144, + "step": 9650 + }, + { + "epoch": 1.500465983224604, + "grad_norm": 0.3345864415168762, + "learning_rate": 9.940773141196357e-05, + "loss": 0.0177, + "step": 9660 + }, + { + "epoch": 1.5020192606399503, + "grad_norm": 0.3195659816265106, + "learning_rate": 9.94051912709176e-05, + "loss": 0.0145, + "step": 9670 + }, + { + "epoch": 1.5035725380552967, + "grad_norm": 0.47580862045288086, + "learning_rate": 9.940264572699421e-05, + "loss": 0.0186, + "step": 9680 + }, + { + "epoch": 1.505125815470643, + "grad_norm": 0.5058879256248474, + "learning_rate": 9.940009478047174e-05, + "loss": 0.0145, + "step": 9690 + }, + { + "epoch": 1.5066790928859894, + "grad_norm": 0.5173896551132202, + "learning_rate": 9.939753843162918e-05, + "loss": 0.0137, + "step": 9700 + }, + { + "epoch": 1.5082323703013358, + "grad_norm": 0.4186009466648102, + "learning_rate": 9.939497668074609e-05, + "loss": 0.0158, + "step": 9710 + }, + { + "epoch": 1.5097856477166822, + "grad_norm": 0.21258720755577087, + "learning_rate": 9.93924095281026e-05, + "loss": 0.0154, + "step": 9720 + }, + { + "epoch": 1.5113389251320286, + "grad_norm": 0.25344234704971313, + "learning_rate": 9.938983697397948e-05, + "loss": 0.0142, + "step": 9730 + }, + { + "epoch": 1.512892202547375, + "grad_norm": 0.20465999841690063, + "learning_rate": 9.938725901865805e-05, + "loss": 0.0118, + "step": 9740 + }, + { + "epoch": 1.5144454799627214, + "grad_norm": 0.2601820230484009, + "learning_rate": 9.93846756624202e-05, + "loss": 0.0144, + "step": 9750 + }, + { + "epoch": 1.5159987573780678, + "grad_norm": 0.44564709067344666, + "learning_rate": 9.938208690554849e-05, + "loss": 0.0128, + "step": 9760 + }, + { + "epoch": 1.5175520347934142, + "grad_norm": 0.45532456040382385, + "learning_rate": 9.9379492748326e-05, + "loss": 0.0168, + "step": 9770 + }, + { + "epoch": 1.5191053122087603, + "grad_norm": 0.38467103242874146, + "learning_rate": 9.937689319103641e-05, + "loss": 0.0147, + "step": 9780 + }, + { + "epoch": 1.520658589624107, + "grad_norm": 0.2731381952762604, + "learning_rate": 9.937428823396404e-05, + "loss": 0.0166, + "step": 9790 + }, + { + "epoch": 1.522211867039453, + "grad_norm": 0.37091922760009766, + "learning_rate": 9.937167787739372e-05, + "loss": 0.012, + "step": 9800 + }, + { + "epoch": 1.5237651444547997, + "grad_norm": 0.2871425747871399, + "learning_rate": 9.936906212161095e-05, + "loss": 0.019, + "step": 9810 + }, + { + "epoch": 1.5253184218701459, + "grad_norm": 0.2245483696460724, + "learning_rate": 9.936644096690176e-05, + "loss": 0.0116, + "step": 9820 + }, + { + "epoch": 1.5268716992854925, + "grad_norm": 0.2986292541027069, + "learning_rate": 9.936381441355282e-05, + "loss": 0.02, + "step": 9830 + }, + { + "epoch": 1.5284249767008387, + "grad_norm": 0.14258934557437897, + "learning_rate": 9.936118246185136e-05, + "loss": 0.0163, + "step": 9840 + }, + { + "epoch": 1.5299782541161853, + "grad_norm": 0.35909250378608704, + "learning_rate": 9.935854511208518e-05, + "loss": 0.016, + "step": 9850 + }, + { + "epoch": 1.5315315315315314, + "grad_norm": 0.4008188247680664, + "learning_rate": 9.935590236454272e-05, + "loss": 0.015, + "step": 9860 + }, + { + "epoch": 1.533084808946878, + "grad_norm": 0.23237092792987823, + "learning_rate": 9.935325421951298e-05, + "loss": 0.0125, + "step": 9870 + }, + { + "epoch": 1.5346380863622242, + "grad_norm": 0.2698978781700134, + "learning_rate": 9.935060067728557e-05, + "loss": 0.0153, + "step": 9880 + }, + { + "epoch": 1.5361913637775708, + "grad_norm": 0.2639816701412201, + "learning_rate": 9.934794173815067e-05, + "loss": 0.019, + "step": 9890 + }, + { + "epoch": 1.537744641192917, + "grad_norm": 0.36631903052330017, + "learning_rate": 9.934527740239906e-05, + "loss": 0.014, + "step": 9900 + }, + { + "epoch": 1.5392979186082636, + "grad_norm": 0.41027456521987915, + "learning_rate": 9.934260767032209e-05, + "loss": 0.0145, + "step": 9910 + }, + { + "epoch": 1.5408511960236098, + "grad_norm": 0.31867915391921997, + "learning_rate": 9.933993254221172e-05, + "loss": 0.016, + "step": 9920 + }, + { + "epoch": 1.5424044734389561, + "grad_norm": 0.33149874210357666, + "learning_rate": 9.933725201836053e-05, + "loss": 0.0114, + "step": 9930 + }, + { + "epoch": 1.5439577508543025, + "grad_norm": 0.32059815526008606, + "learning_rate": 9.933456609906162e-05, + "loss": 0.0149, + "step": 9940 + }, + { + "epoch": 1.545511028269649, + "grad_norm": 0.38014984130859375, + "learning_rate": 9.933187478460875e-05, + "loss": 0.0171, + "step": 9950 + }, + { + "epoch": 1.5470643056849953, + "grad_norm": 0.28725916147232056, + "learning_rate": 9.93291780752962e-05, + "loss": 0.0165, + "step": 9960 + }, + { + "epoch": 1.5486175831003417, + "grad_norm": 0.2851012647151947, + "learning_rate": 9.932647597141893e-05, + "loss": 0.0144, + "step": 9970 + }, + { + "epoch": 1.550170860515688, + "grad_norm": 0.3285042941570282, + "learning_rate": 9.932376847327239e-05, + "loss": 0.0194, + "step": 9980 + }, + { + "epoch": 1.5517241379310345, + "grad_norm": 0.3430781960487366, + "learning_rate": 9.932105558115268e-05, + "loss": 0.0149, + "step": 9990 + }, + { + "epoch": 1.5532774153463809, + "grad_norm": 0.3310341536998749, + "learning_rate": 9.931833729535651e-05, + "loss": 0.0172, + "step": 10000 + }, + { + "epoch": 1.5548306927617273, + "grad_norm": 0.35465508699417114, + "learning_rate": 9.931561361618111e-05, + "loss": 0.0157, + "step": 10010 + }, + { + "epoch": 1.5563839701770736, + "grad_norm": 0.25908032059669495, + "learning_rate": 9.931288454392435e-05, + "loss": 0.0192, + "step": 10020 + }, + { + "epoch": 1.55793724759242, + "grad_norm": 0.3224649131298065, + "learning_rate": 9.931015007888467e-05, + "loss": 0.0148, + "step": 10030 + }, + { + "epoch": 1.5594905250077664, + "grad_norm": 0.26478680968284607, + "learning_rate": 9.930741022136112e-05, + "loss": 0.0191, + "step": 10040 + }, + { + "epoch": 1.5610438024231128, + "grad_norm": 0.3597929775714874, + "learning_rate": 9.930466497165333e-05, + "loss": 0.0164, + "step": 10050 + }, + { + "epoch": 1.5625970798384592, + "grad_norm": 0.4054081439971924, + "learning_rate": 9.93019143300615e-05, + "loss": 0.0169, + "step": 10060 + }, + { + "epoch": 1.5641503572538056, + "grad_norm": 0.2255040556192398, + "learning_rate": 9.929915829688644e-05, + "loss": 0.0177, + "step": 10070 + }, + { + "epoch": 1.565703634669152, + "grad_norm": 0.2871943414211273, + "learning_rate": 9.929639687242955e-05, + "loss": 0.0162, + "step": 10080 + }, + { + "epoch": 1.5672569120844981, + "grad_norm": 0.37801751494407654, + "learning_rate": 9.929363005699281e-05, + "loss": 0.0136, + "step": 10090 + }, + { + "epoch": 1.5688101894998447, + "grad_norm": 0.32905399799346924, + "learning_rate": 9.92908578508788e-05, + "loss": 0.014, + "step": 10100 + }, + { + "epoch": 1.570363466915191, + "grad_norm": 0.3224899172782898, + "learning_rate": 9.928808025439069e-05, + "loss": 0.0138, + "step": 10110 + }, + { + "epoch": 1.5719167443305375, + "grad_norm": 0.3332047462463379, + "learning_rate": 9.928529726783223e-05, + "loss": 0.0146, + "step": 10120 + }, + { + "epoch": 1.5734700217458837, + "grad_norm": 0.3010748326778412, + "learning_rate": 9.928250889150774e-05, + "loss": 0.0125, + "step": 10130 + }, + { + "epoch": 1.5750232991612303, + "grad_norm": 0.5180928111076355, + "learning_rate": 9.92797151257222e-05, + "loss": 0.0133, + "step": 10140 + }, + { + "epoch": 1.5765765765765765, + "grad_norm": 0.3486071228981018, + "learning_rate": 9.927691597078108e-05, + "loss": 0.0178, + "step": 10150 + }, + { + "epoch": 1.578129853991923, + "grad_norm": 0.37138083577156067, + "learning_rate": 9.927411142699053e-05, + "loss": 0.0129, + "step": 10160 + }, + { + "epoch": 1.5796831314072692, + "grad_norm": 0.35563725233078003, + "learning_rate": 9.927130149465725e-05, + "loss": 0.0154, + "step": 10170 + }, + { + "epoch": 1.5812364088226158, + "grad_norm": 0.35734423995018005, + "learning_rate": 9.92684861740885e-05, + "loss": 0.0102, + "step": 10180 + }, + { + "epoch": 1.582789686237962, + "grad_norm": 0.354963481426239, + "learning_rate": 9.926566546559217e-05, + "loss": 0.0155, + "step": 10190 + }, + { + "epoch": 1.5843429636533086, + "grad_norm": 0.39097440242767334, + "learning_rate": 9.926283936947673e-05, + "loss": 0.0149, + "step": 10200 + }, + { + "epoch": 1.5858962410686548, + "grad_norm": 0.48775404691696167, + "learning_rate": 9.926000788605126e-05, + "loss": 0.0171, + "step": 10210 + }, + { + "epoch": 1.5874495184840014, + "grad_norm": 0.2683597207069397, + "learning_rate": 9.92571710156254e-05, + "loss": 0.0111, + "step": 10220 + }, + { + "epoch": 1.5890027958993476, + "grad_norm": 0.303055077791214, + "learning_rate": 9.925432875850936e-05, + "loss": 0.0162, + "step": 10230 + }, + { + "epoch": 1.590556073314694, + "grad_norm": 0.317962110042572, + "learning_rate": 9.925148111501396e-05, + "loss": 0.0127, + "step": 10240 + }, + { + "epoch": 1.5921093507300403, + "grad_norm": 0.3267838656902313, + "learning_rate": 9.924862808545066e-05, + "loss": 0.0152, + "step": 10250 + }, + { + "epoch": 1.5936626281453867, + "grad_norm": 0.35026809573173523, + "learning_rate": 9.924576967013141e-05, + "loss": 0.0146, + "step": 10260 + }, + { + "epoch": 1.5952159055607331, + "grad_norm": 0.4110967516899109, + "learning_rate": 9.924290586936887e-05, + "loss": 0.0145, + "step": 10270 + }, + { + "epoch": 1.5967691829760795, + "grad_norm": 0.38687002658843994, + "learning_rate": 9.924003668347614e-05, + "loss": 0.0186, + "step": 10280 + }, + { + "epoch": 1.598322460391426, + "grad_norm": 0.44465896487236023, + "learning_rate": 9.923716211276704e-05, + "loss": 0.0184, + "step": 10290 + }, + { + "epoch": 1.5998757378067723, + "grad_norm": 0.6582725644111633, + "learning_rate": 9.923428215755594e-05, + "loss": 0.013, + "step": 10300 + }, + { + "epoch": 1.6014290152221187, + "grad_norm": 0.42459264397621155, + "learning_rate": 9.923139681815775e-05, + "loss": 0.0131, + "step": 10310 + }, + { + "epoch": 1.602982292637465, + "grad_norm": 0.33421778678894043, + "learning_rate": 9.922850609488801e-05, + "loss": 0.0174, + "step": 10320 + }, + { + "epoch": 1.6045355700528114, + "grad_norm": 0.22062747180461884, + "learning_rate": 9.922560998806287e-05, + "loss": 0.0124, + "step": 10330 + }, + { + "epoch": 1.6060888474681578, + "grad_norm": 0.335239052772522, + "learning_rate": 9.922270849799905e-05, + "loss": 0.013, + "step": 10340 + }, + { + "epoch": 1.6076421248835042, + "grad_norm": 0.29178911447525024, + "learning_rate": 9.92198016250138e-05, + "loss": 0.014, + "step": 10350 + }, + { + "epoch": 1.6091954022988506, + "grad_norm": 0.24123895168304443, + "learning_rate": 9.921688936942506e-05, + "loss": 0.013, + "step": 10360 + }, + { + "epoch": 1.610748679714197, + "grad_norm": 0.31549888849258423, + "learning_rate": 9.921397173155129e-05, + "loss": 0.0165, + "step": 10370 + }, + { + "epoch": 1.6123019571295434, + "grad_norm": 0.2645583152770996, + "learning_rate": 9.921104871171157e-05, + "loss": 0.0155, + "step": 10380 + }, + { + "epoch": 1.6138552345448898, + "grad_norm": 0.32306531071662903, + "learning_rate": 9.920812031022554e-05, + "loss": 0.0147, + "step": 10390 + }, + { + "epoch": 1.615408511960236, + "grad_norm": 0.34770479798316956, + "learning_rate": 9.920518652741348e-05, + "loss": 0.0156, + "step": 10400 + }, + { + "epoch": 1.6169617893755825, + "grad_norm": 0.2791750729084015, + "learning_rate": 9.920224736359618e-05, + "loss": 0.0145, + "step": 10410 + }, + { + "epoch": 1.6185150667909287, + "grad_norm": 0.33591556549072266, + "learning_rate": 9.91993028190951e-05, + "loss": 0.0172, + "step": 10420 + }, + { + "epoch": 1.6200683442062753, + "grad_norm": 0.30298730731010437, + "learning_rate": 9.919635289423222e-05, + "loss": 0.0146, + "step": 10430 + }, + { + "epoch": 1.6216216216216215, + "grad_norm": 0.5605953931808472, + "learning_rate": 9.919339758933015e-05, + "loss": 0.0186, + "step": 10440 + }, + { + "epoch": 1.623174899036968, + "grad_norm": 0.5088539123535156, + "learning_rate": 9.919043690471209e-05, + "loss": 0.0158, + "step": 10450 + }, + { + "epoch": 1.6247281764523143, + "grad_norm": 0.34683331847190857, + "learning_rate": 9.91874708407018e-05, + "loss": 0.0161, + "step": 10460 + }, + { + "epoch": 1.6262814538676609, + "grad_norm": 0.5190711617469788, + "learning_rate": 9.918449939762367e-05, + "loss": 0.019, + "step": 10470 + }, + { + "epoch": 1.627834731283007, + "grad_norm": 0.31883570551872253, + "learning_rate": 9.91815225758026e-05, + "loss": 0.0153, + "step": 10480 + }, + { + "epoch": 1.6293880086983537, + "grad_norm": 0.14477255940437317, + "learning_rate": 9.917854037556419e-05, + "loss": 0.0145, + "step": 10490 + }, + { + "epoch": 1.6309412861136998, + "grad_norm": 0.2599961459636688, + "learning_rate": 9.917555279723454e-05, + "loss": 0.0153, + "step": 10500 + }, + { + "epoch": 1.6324945635290464, + "grad_norm": 0.18952277302742004, + "learning_rate": 9.917255984114036e-05, + "loss": 0.0154, + "step": 10510 + }, + { + "epoch": 1.6340478409443926, + "grad_norm": 0.2522837519645691, + "learning_rate": 9.916956150760896e-05, + "loss": 0.0181, + "step": 10520 + }, + { + "epoch": 1.6356011183597392, + "grad_norm": 0.2978634834289551, + "learning_rate": 9.916655779696826e-05, + "loss": 0.0169, + "step": 10530 + }, + { + "epoch": 1.6371543957750854, + "grad_norm": 0.25787103176116943, + "learning_rate": 9.916354870954671e-05, + "loss": 0.0157, + "step": 10540 + }, + { + "epoch": 1.6387076731904318, + "grad_norm": 0.33212384581565857, + "learning_rate": 9.91605342456734e-05, + "loss": 0.0146, + "step": 10550 + }, + { + "epoch": 1.6402609506057781, + "grad_norm": 0.22246508300304413, + "learning_rate": 9.915751440567795e-05, + "loss": 0.0117, + "step": 10560 + }, + { + "epoch": 1.6418142280211245, + "grad_norm": 0.309709370136261, + "learning_rate": 9.915448918989066e-05, + "loss": 0.0139, + "step": 10570 + }, + { + "epoch": 1.643367505436471, + "grad_norm": 0.14263159036636353, + "learning_rate": 9.915145859864232e-05, + "loss": 0.0126, + "step": 10580 + }, + { + "epoch": 1.6449207828518173, + "grad_norm": 0.29086533188819885, + "learning_rate": 9.914842263226437e-05, + "loss": 0.0131, + "step": 10590 + }, + { + "epoch": 1.6464740602671637, + "grad_norm": 0.3211687207221985, + "learning_rate": 9.914538129108882e-05, + "loss": 0.0164, + "step": 10600 + }, + { + "epoch": 1.64802733768251, + "grad_norm": 0.2965916097164154, + "learning_rate": 9.914233457544825e-05, + "loss": 0.0104, + "step": 10610 + }, + { + "epoch": 1.6495806150978565, + "grad_norm": 0.22153028845787048, + "learning_rate": 9.913928248567586e-05, + "loss": 0.0134, + "step": 10620 + }, + { + "epoch": 1.6511338925132029, + "grad_norm": 0.35677871108055115, + "learning_rate": 9.913622502210542e-05, + "loss": 0.0128, + "step": 10630 + }, + { + "epoch": 1.6526871699285492, + "grad_norm": 0.30285418033599854, + "learning_rate": 9.913316218507128e-05, + "loss": 0.0122, + "step": 10640 + }, + { + "epoch": 1.6542404473438956, + "grad_norm": 0.35710641741752625, + "learning_rate": 9.91300939749084e-05, + "loss": 0.013, + "step": 10650 + }, + { + "epoch": 1.655793724759242, + "grad_norm": 0.16188757121562958, + "learning_rate": 9.91270203919523e-05, + "loss": 0.0161, + "step": 10660 + }, + { + "epoch": 1.6573470021745884, + "grad_norm": 0.2342321276664734, + "learning_rate": 9.912394143653912e-05, + "loss": 0.0117, + "step": 10670 + }, + { + "epoch": 1.6589002795899348, + "grad_norm": 0.415877103805542, + "learning_rate": 9.912085710900555e-05, + "loss": 0.0154, + "step": 10680 + }, + { + "epoch": 1.6604535570052812, + "grad_norm": 0.46460703015327454, + "learning_rate": 9.911776740968892e-05, + "loss": 0.0144, + "step": 10690 + }, + { + "epoch": 1.6620068344206276, + "grad_norm": 0.5741338133811951, + "learning_rate": 9.911467233892709e-05, + "loss": 0.013, + "step": 10700 + }, + { + "epoch": 1.6635601118359737, + "grad_norm": 0.43400225043296814, + "learning_rate": 9.911157189705853e-05, + "loss": 0.0181, + "step": 10710 + }, + { + "epoch": 1.6651133892513204, + "grad_norm": 0.2960217595100403, + "learning_rate": 9.910846608442229e-05, + "loss": 0.0137, + "step": 10720 + }, + { + "epoch": 1.6666666666666665, + "grad_norm": 0.3495723605155945, + "learning_rate": 9.910535490135805e-05, + "loss": 0.0163, + "step": 10730 + }, + { + "epoch": 1.6682199440820131, + "grad_norm": 0.49671709537506104, + "learning_rate": 9.910223834820603e-05, + "loss": 0.017, + "step": 10740 + }, + { + "epoch": 1.6697732214973593, + "grad_norm": 0.30548352003097534, + "learning_rate": 9.909911642530703e-05, + "loss": 0.0146, + "step": 10750 + }, + { + "epoch": 1.671326498912706, + "grad_norm": 0.2857396602630615, + "learning_rate": 9.909598913300249e-05, + "loss": 0.0146, + "step": 10760 + }, + { + "epoch": 1.672879776328052, + "grad_norm": 0.5669980049133301, + "learning_rate": 9.909285647163438e-05, + "loss": 0.0191, + "step": 10770 + }, + { + "epoch": 1.6744330537433987, + "grad_norm": 0.44672486186027527, + "learning_rate": 9.908971844154531e-05, + "loss": 0.0145, + "step": 10780 + }, + { + "epoch": 1.6759863311587448, + "grad_norm": 0.3184344470500946, + "learning_rate": 9.908657504307843e-05, + "loss": 0.0188, + "step": 10790 + }, + { + "epoch": 1.6775396085740915, + "grad_norm": 0.44894370436668396, + "learning_rate": 9.908342627657751e-05, + "loss": 0.0142, + "step": 10800 + }, + { + "epoch": 1.6790928859894376, + "grad_norm": 0.2671568691730499, + "learning_rate": 9.908027214238689e-05, + "loss": 0.014, + "step": 10810 + }, + { + "epoch": 1.6806461634047842, + "grad_norm": 0.3005874752998352, + "learning_rate": 9.90771126408515e-05, + "loss": 0.0131, + "step": 10820 + }, + { + "epoch": 1.6821994408201304, + "grad_norm": 0.40342622995376587, + "learning_rate": 9.907394777231685e-05, + "loss": 0.014, + "step": 10830 + }, + { + "epoch": 1.683752718235477, + "grad_norm": 0.22057725489139557, + "learning_rate": 9.907077753712905e-05, + "loss": 0.0163, + "step": 10840 + }, + { + "epoch": 1.6853059956508232, + "grad_norm": 0.6012116074562073, + "learning_rate": 9.906760193563482e-05, + "loss": 0.0146, + "step": 10850 + }, + { + "epoch": 1.6868592730661698, + "grad_norm": 0.31450292468070984, + "learning_rate": 9.906442096818139e-05, + "loss": 0.015, + "step": 10860 + }, + { + "epoch": 1.688412550481516, + "grad_norm": 0.257682740688324, + "learning_rate": 9.906123463511665e-05, + "loss": 0.0152, + "step": 10870 + }, + { + "epoch": 1.6899658278968623, + "grad_norm": 0.23654873669147491, + "learning_rate": 9.905804293678907e-05, + "loss": 0.012, + "step": 10880 + }, + { + "epoch": 1.6915191053122087, + "grad_norm": 0.21518097817897797, + "learning_rate": 9.905484587354766e-05, + "loss": 0.014, + "step": 10890 + }, + { + "epoch": 1.6930723827275551, + "grad_norm": 0.3484296202659607, + "learning_rate": 9.905164344574205e-05, + "loss": 0.0112, + "step": 10900 + }, + { + "epoch": 1.6946256601429015, + "grad_norm": 0.24349945783615112, + "learning_rate": 9.904843565372248e-05, + "loss": 0.0164, + "step": 10910 + }, + { + "epoch": 1.696178937558248, + "grad_norm": 0.21182280778884888, + "learning_rate": 9.904522249783972e-05, + "loss": 0.0143, + "step": 10920 + }, + { + "epoch": 1.6977322149735943, + "grad_norm": 0.2816213369369507, + "learning_rate": 9.904200397844517e-05, + "loss": 0.0131, + "step": 10930 + }, + { + "epoch": 1.6992854923889407, + "grad_norm": 0.366150826215744, + "learning_rate": 9.903878009589078e-05, + "loss": 0.0124, + "step": 10940 + }, + { + "epoch": 1.700838769804287, + "grad_norm": 0.2911145091056824, + "learning_rate": 9.903555085052915e-05, + "loss": 0.0126, + "step": 10950 + }, + { + "epoch": 1.7023920472196334, + "grad_norm": 0.25882720947265625, + "learning_rate": 9.903231624271338e-05, + "loss": 0.0128, + "step": 10960 + }, + { + "epoch": 1.7039453246349798, + "grad_norm": 0.2501699924468994, + "learning_rate": 9.902907627279724e-05, + "loss": 0.0146, + "step": 10970 + }, + { + "epoch": 1.7054986020503262, + "grad_norm": 0.23334112763404846, + "learning_rate": 9.902583094113504e-05, + "loss": 0.0165, + "step": 10980 + }, + { + "epoch": 1.7070518794656726, + "grad_norm": 0.35120099782943726, + "learning_rate": 9.902258024808168e-05, + "loss": 0.0135, + "step": 10990 + }, + { + "epoch": 1.708605156881019, + "grad_norm": 0.2753477394580841, + "learning_rate": 9.901932419399264e-05, + "loss": 0.0133, + "step": 11000 + }, + { + "epoch": 1.7101584342963654, + "grad_norm": 0.2314516007900238, + "learning_rate": 9.9016062779224e-05, + "loss": 0.0117, + "step": 11010 + }, + { + "epoch": 1.7117117117117115, + "grad_norm": 0.28952547907829285, + "learning_rate": 9.901279600413242e-05, + "loss": 0.014, + "step": 11020 + }, + { + "epoch": 1.7132649891270582, + "grad_norm": 0.2753351926803589, + "learning_rate": 9.900952386907518e-05, + "loss": 0.013, + "step": 11030 + }, + { + "epoch": 1.7148182665424043, + "grad_norm": 0.21597273647785187, + "learning_rate": 9.90062463744101e-05, + "loss": 0.0139, + "step": 11040 + }, + { + "epoch": 1.716371543957751, + "grad_norm": 0.3236696720123291, + "learning_rate": 9.900296352049558e-05, + "loss": 0.0122, + "step": 11050 + }, + { + "epoch": 1.717924821373097, + "grad_norm": 0.2940537631511688, + "learning_rate": 9.899967530769065e-05, + "loss": 0.0108, + "step": 11060 + }, + { + "epoch": 1.7194780987884437, + "grad_norm": 0.2145911306142807, + "learning_rate": 9.899638173635489e-05, + "loss": 0.0187, + "step": 11070 + }, + { + "epoch": 1.7210313762037899, + "grad_norm": 0.29431453347206116, + "learning_rate": 9.899308280684849e-05, + "loss": 0.0127, + "step": 11080 + }, + { + "epoch": 1.7225846536191365, + "grad_norm": 0.2471170723438263, + "learning_rate": 9.898977851953222e-05, + "loss": 0.0154, + "step": 11090 + }, + { + "epoch": 1.7241379310344827, + "grad_norm": 0.21834518015384674, + "learning_rate": 9.898646887476741e-05, + "loss": 0.0122, + "step": 11100 + }, + { + "epoch": 1.7256912084498293, + "grad_norm": 0.45959925651550293, + "learning_rate": 9.898315387291603e-05, + "loss": 0.0209, + "step": 11110 + }, + { + "epoch": 1.7272444858651754, + "grad_norm": 0.3219773769378662, + "learning_rate": 9.89798335143406e-05, + "loss": 0.0189, + "step": 11120 + }, + { + "epoch": 1.728797763280522, + "grad_norm": 0.23789498209953308, + "learning_rate": 9.897650779940419e-05, + "loss": 0.0158, + "step": 11130 + }, + { + "epoch": 1.7303510406958682, + "grad_norm": 0.3663710057735443, + "learning_rate": 9.897317672847054e-05, + "loss": 0.0174, + "step": 11140 + }, + { + "epoch": 1.7319043181112148, + "grad_norm": 0.3635806739330292, + "learning_rate": 9.89698403019039e-05, + "loss": 0.021, + "step": 11150 + }, + { + "epoch": 1.733457595526561, + "grad_norm": 0.2899337410926819, + "learning_rate": 9.896649852006917e-05, + "loss": 0.0207, + "step": 11160 + }, + { + "epoch": 1.7350108729419076, + "grad_norm": 0.3331875801086426, + "learning_rate": 9.896315138333177e-05, + "loss": 0.0159, + "step": 11170 + }, + { + "epoch": 1.7365641503572538, + "grad_norm": 0.3070160150527954, + "learning_rate": 9.895979889205774e-05, + "loss": 0.0153, + "step": 11180 + }, + { + "epoch": 1.7381174277726001, + "grad_norm": 0.44996434450149536, + "learning_rate": 9.895644104661372e-05, + "loss": 0.0188, + "step": 11190 + }, + { + "epoch": 1.7396707051879465, + "grad_norm": 0.2959100604057312, + "learning_rate": 9.895307784736691e-05, + "loss": 0.0169, + "step": 11200 + }, + { + "epoch": 1.741223982603293, + "grad_norm": 0.4868946671485901, + "learning_rate": 9.894970929468512e-05, + "loss": 0.0161, + "step": 11210 + }, + { + "epoch": 1.7427772600186393, + "grad_norm": 0.2831275761127472, + "learning_rate": 9.89463353889367e-05, + "loss": 0.0149, + "step": 11220 + }, + { + "epoch": 1.7443305374339857, + "grad_norm": 0.27055054903030396, + "learning_rate": 9.894295613049065e-05, + "loss": 0.0153, + "step": 11230 + }, + { + "epoch": 1.745883814849332, + "grad_norm": 0.26354891061782837, + "learning_rate": 9.893957151971649e-05, + "loss": 0.0126, + "step": 11240 + }, + { + "epoch": 1.7474370922646785, + "grad_norm": 0.5179353356361389, + "learning_rate": 9.893618155698436e-05, + "loss": 0.0176, + "step": 11250 + }, + { + "epoch": 1.7489903696800249, + "grad_norm": 0.2863048315048218, + "learning_rate": 9.8932786242665e-05, + "loss": 0.013, + "step": 11260 + }, + { + "epoch": 1.7505436470953712, + "grad_norm": 0.3526429235935211, + "learning_rate": 9.89293855771297e-05, + "loss": 0.0126, + "step": 11270 + }, + { + "epoch": 1.7520969245107176, + "grad_norm": 0.19362322986125946, + "learning_rate": 9.892597956075036e-05, + "loss": 0.0141, + "step": 11280 + }, + { + "epoch": 1.753650201926064, + "grad_norm": 0.2638789415359497, + "learning_rate": 9.892256819389947e-05, + "loss": 0.0128, + "step": 11290 + }, + { + "epoch": 1.7552034793414104, + "grad_norm": 0.31091761589050293, + "learning_rate": 9.891915147695006e-05, + "loss": 0.0141, + "step": 11300 + }, + { + "epoch": 1.7567567567567568, + "grad_norm": 0.2733268439769745, + "learning_rate": 9.891572941027577e-05, + "loss": 0.0139, + "step": 11310 + }, + { + "epoch": 1.7583100341721032, + "grad_norm": 0.49007412791252136, + "learning_rate": 9.89123019942509e-05, + "loss": 0.0114, + "step": 11320 + }, + { + "epoch": 1.7598633115874494, + "grad_norm": 0.29070383310317993, + "learning_rate": 9.89088692292502e-05, + "loss": 0.0152, + "step": 11330 + }, + { + "epoch": 1.761416589002796, + "grad_norm": 0.33346110582351685, + "learning_rate": 9.89054311156491e-05, + "loss": 0.0155, + "step": 11340 + }, + { + "epoch": 1.7629698664181421, + "grad_norm": 0.2686813473701477, + "learning_rate": 9.890198765382357e-05, + "loss": 0.0154, + "step": 11350 + }, + { + "epoch": 1.7645231438334887, + "grad_norm": 0.31774356961250305, + "learning_rate": 9.889853884415021e-05, + "loss": 0.0137, + "step": 11360 + }, + { + "epoch": 1.766076421248835, + "grad_norm": 0.24117225408554077, + "learning_rate": 9.889508468700614e-05, + "loss": 0.019, + "step": 11370 + }, + { + "epoch": 1.7676296986641815, + "grad_norm": 0.247414693236351, + "learning_rate": 9.889162518276915e-05, + "loss": 0.0153, + "step": 11380 + }, + { + "epoch": 1.7691829760795277, + "grad_norm": 0.22794316709041595, + "learning_rate": 9.888816033181752e-05, + "loss": 0.0121, + "step": 11390 + }, + { + "epoch": 1.7707362534948743, + "grad_norm": 0.3274255692958832, + "learning_rate": 9.888469013453018e-05, + "loss": 0.0092, + "step": 11400 + }, + { + "epoch": 1.7722895309102205, + "grad_norm": 0.24636074900627136, + "learning_rate": 9.888121459128663e-05, + "loss": 0.0138, + "step": 11410 + }, + { + "epoch": 1.773842808325567, + "grad_norm": 0.28550589084625244, + "learning_rate": 9.887773370246693e-05, + "loss": 0.0143, + "step": 11420 + }, + { + "epoch": 1.7753960857409132, + "grad_norm": 0.403823584318161, + "learning_rate": 9.887424746845177e-05, + "loss": 0.0177, + "step": 11430 + }, + { + "epoch": 1.7769493631562598, + "grad_norm": 0.24222561717033386, + "learning_rate": 9.887075588962239e-05, + "loss": 0.0127, + "step": 11440 + }, + { + "epoch": 1.778502640571606, + "grad_norm": 0.5680859684944153, + "learning_rate": 9.88672589663606e-05, + "loss": 0.0138, + "step": 11450 + }, + { + "epoch": 1.7800559179869526, + "grad_norm": 0.3311220407485962, + "learning_rate": 9.886375669904886e-05, + "loss": 0.018, + "step": 11460 + }, + { + "epoch": 1.7816091954022988, + "grad_norm": 0.18942208588123322, + "learning_rate": 9.886024908807014e-05, + "loss": 0.0145, + "step": 11470 + }, + { + "epoch": 1.7831624728176454, + "grad_norm": 0.26785218715667725, + "learning_rate": 9.885673613380806e-05, + "loss": 0.0153, + "step": 11480 + }, + { + "epoch": 1.7847157502329916, + "grad_norm": 0.20637986063957214, + "learning_rate": 9.885321783664676e-05, + "loss": 0.0132, + "step": 11490 + }, + { + "epoch": 1.786269027648338, + "grad_norm": 0.22372344136238098, + "learning_rate": 9.884969419697101e-05, + "loss": 0.0122, + "step": 11500 + }, + { + "epoch": 1.7878223050636843, + "grad_norm": 0.2016613632440567, + "learning_rate": 9.884616521516614e-05, + "loss": 0.0129, + "step": 11510 + }, + { + "epoch": 1.7893755824790307, + "grad_norm": 0.6435093283653259, + "learning_rate": 9.88426308916181e-05, + "loss": 0.0155, + "step": 11520 + }, + { + "epoch": 1.7909288598943771, + "grad_norm": 0.3849537670612335, + "learning_rate": 9.883909122671335e-05, + "loss": 0.0144, + "step": 11530 + }, + { + "epoch": 1.7924821373097235, + "grad_norm": 0.26871246099472046, + "learning_rate": 9.883554622083904e-05, + "loss": 0.0153, + "step": 11540 + }, + { + "epoch": 1.7940354147250699, + "grad_norm": 0.37777698040008545, + "learning_rate": 9.88319958743828e-05, + "loss": 0.0153, + "step": 11550 + }, + { + "epoch": 1.7955886921404163, + "grad_norm": 0.2860715985298157, + "learning_rate": 9.882844018773291e-05, + "loss": 0.0166, + "step": 11560 + }, + { + "epoch": 1.7971419695557627, + "grad_norm": 0.22712601721286774, + "learning_rate": 9.882487916127823e-05, + "loss": 0.0149, + "step": 11570 + }, + { + "epoch": 1.798695246971109, + "grad_norm": 0.30010849237442017, + "learning_rate": 9.882131279540815e-05, + "loss": 0.0137, + "step": 11580 + }, + { + "epoch": 1.8002485243864554, + "grad_norm": 0.3470853865146637, + "learning_rate": 9.881774109051271e-05, + "loss": 0.0187, + "step": 11590 + }, + { + "epoch": 1.8018018018018018, + "grad_norm": 0.2508409917354584, + "learning_rate": 9.881416404698252e-05, + "loss": 0.0132, + "step": 11600 + }, + { + "epoch": 1.8033550792171482, + "grad_norm": 0.41145578026771545, + "learning_rate": 9.881058166520873e-05, + "loss": 0.0146, + "step": 11610 + }, + { + "epoch": 1.8049083566324946, + "grad_norm": 0.228557288646698, + "learning_rate": 9.880699394558311e-05, + "loss": 0.0141, + "step": 11620 + }, + { + "epoch": 1.806461634047841, + "grad_norm": 0.20208345353603363, + "learning_rate": 9.880340088849801e-05, + "loss": 0.0128, + "step": 11630 + }, + { + "epoch": 1.8080149114631874, + "grad_norm": 0.6907563209533691, + "learning_rate": 9.879980249434637e-05, + "loss": 0.0166, + "step": 11640 + }, + { + "epoch": 1.8095681888785338, + "grad_norm": 0.2660052180290222, + "learning_rate": 9.879619876352168e-05, + "loss": 0.0159, + "step": 11650 + }, + { + "epoch": 1.81112146629388, + "grad_norm": 0.4010213613510132, + "learning_rate": 9.879258969641809e-05, + "loss": 0.0156, + "step": 11660 + }, + { + "epoch": 1.8126747437092265, + "grad_norm": 0.39970600605010986, + "learning_rate": 9.878897529343023e-05, + "loss": 0.0143, + "step": 11670 + }, + { + "epoch": 1.8142280211245727, + "grad_norm": 0.20888054370880127, + "learning_rate": 9.878535555495338e-05, + "loss": 0.0139, + "step": 11680 + }, + { + "epoch": 1.8157812985399193, + "grad_norm": 0.1796136200428009, + "learning_rate": 9.87817304813834e-05, + "loss": 0.011, + "step": 11690 + }, + { + "epoch": 1.8173345759552655, + "grad_norm": 0.17413458228111267, + "learning_rate": 9.877810007311671e-05, + "loss": 0.013, + "step": 11700 + }, + { + "epoch": 1.818887853370612, + "grad_norm": 0.17001208662986755, + "learning_rate": 9.877446433055035e-05, + "loss": 0.0176, + "step": 11710 + }, + { + "epoch": 1.8204411307859583, + "grad_norm": 0.25068843364715576, + "learning_rate": 9.877082325408191e-05, + "loss": 0.0143, + "step": 11720 + }, + { + "epoch": 1.8219944082013049, + "grad_norm": 0.2851838171482086, + "learning_rate": 9.876717684410954e-05, + "loss": 0.0158, + "step": 11730 + }, + { + "epoch": 1.823547685616651, + "grad_norm": 0.38438326120376587, + "learning_rate": 9.876352510103204e-05, + "loss": 0.0143, + "step": 11740 + }, + { + "epoch": 1.8251009630319976, + "grad_norm": 0.26116296648979187, + "learning_rate": 9.875986802524875e-05, + "loss": 0.0179, + "step": 11750 + }, + { + "epoch": 1.8266542404473438, + "grad_norm": 0.26457834243774414, + "learning_rate": 9.87562056171596e-05, + "loss": 0.0152, + "step": 11760 + }, + { + "epoch": 1.8282075178626904, + "grad_norm": 0.31846776604652405, + "learning_rate": 9.875253787716511e-05, + "loss": 0.014, + "step": 11770 + }, + { + "epoch": 1.8297607952780366, + "grad_norm": 0.34290897846221924, + "learning_rate": 9.874886480566637e-05, + "loss": 0.0173, + "step": 11780 + }, + { + "epoch": 1.8313140726933832, + "grad_norm": 0.29616761207580566, + "learning_rate": 9.874518640306507e-05, + "loss": 0.0161, + "step": 11790 + }, + { + "epoch": 1.8328673501087294, + "grad_norm": 0.3395690321922302, + "learning_rate": 9.874150266976347e-05, + "loss": 0.0141, + "step": 11800 + }, + { + "epoch": 1.8344206275240758, + "grad_norm": 0.3654578924179077, + "learning_rate": 9.873781360616443e-05, + "loss": 0.0143, + "step": 11810 + }, + { + "epoch": 1.8359739049394221, + "grad_norm": 0.2631981074810028, + "learning_rate": 9.873411921267137e-05, + "loss": 0.0197, + "step": 11820 + }, + { + "epoch": 1.8375271823547685, + "grad_norm": 0.4015713334083557, + "learning_rate": 9.873041948968829e-05, + "loss": 0.0168, + "step": 11830 + }, + { + "epoch": 1.839080459770115, + "grad_norm": 0.4183467924594879, + "learning_rate": 9.872671443761981e-05, + "loss": 0.0136, + "step": 11840 + }, + { + "epoch": 1.8406337371854613, + "grad_norm": 0.30078843235969543, + "learning_rate": 9.872300405687109e-05, + "loss": 0.0201, + "step": 11850 + }, + { + "epoch": 1.8421870146008077, + "grad_norm": 0.22767935693264008, + "learning_rate": 9.871928834784792e-05, + "loss": 0.016, + "step": 11860 + }, + { + "epoch": 1.843740292016154, + "grad_norm": 0.5135709643363953, + "learning_rate": 9.871556731095661e-05, + "loss": 0.0182, + "step": 11870 + }, + { + "epoch": 1.8452935694315005, + "grad_norm": 0.3457010090351105, + "learning_rate": 9.871184094660411e-05, + "loss": 0.0118, + "step": 11880 + }, + { + "epoch": 1.8468468468468469, + "grad_norm": 0.3898541331291199, + "learning_rate": 9.870810925519791e-05, + "loss": 0.0131, + "step": 11890 + }, + { + "epoch": 1.8484001242621932, + "grad_norm": 0.20598269999027252, + "learning_rate": 9.870437223714612e-05, + "loss": 0.0151, + "step": 11900 + }, + { + "epoch": 1.8499534016775396, + "grad_norm": 0.25065872073173523, + "learning_rate": 9.87006298928574e-05, + "loss": 0.0192, + "step": 11910 + }, + { + "epoch": 1.851506679092886, + "grad_norm": 0.4425346553325653, + "learning_rate": 9.869688222274103e-05, + "loss": 0.0106, + "step": 11920 + }, + { + "epoch": 1.8530599565082324, + "grad_norm": 0.34262949228286743, + "learning_rate": 9.869312922720681e-05, + "loss": 0.0171, + "step": 11930 + }, + { + "epoch": 1.8546132339235788, + "grad_norm": 0.34090322256088257, + "learning_rate": 9.868937090666521e-05, + "loss": 0.0166, + "step": 11940 + }, + { + "epoch": 1.8561665113389252, + "grad_norm": 0.370144784450531, + "learning_rate": 9.86856072615272e-05, + "loss": 0.014, + "step": 11950 + }, + { + "epoch": 1.8577197887542716, + "grad_norm": 0.4881286919116974, + "learning_rate": 9.868183829220438e-05, + "loss": 0.0178, + "step": 11960 + }, + { + "epoch": 1.8592730661696177, + "grad_norm": 0.4429682493209839, + "learning_rate": 9.867806399910893e-05, + "loss": 0.0192, + "step": 11970 + }, + { + "epoch": 1.8608263435849643, + "grad_norm": 0.36893999576568604, + "learning_rate": 9.867428438265356e-05, + "loss": 0.0152, + "step": 11980 + }, + { + "epoch": 1.8623796210003105, + "grad_norm": 0.2420511692762375, + "learning_rate": 9.867049944325165e-05, + "loss": 0.0124, + "step": 11990 + }, + { + "epoch": 1.8639328984156571, + "grad_norm": 0.4092390239238739, + "learning_rate": 9.86667091813171e-05, + "loss": 0.0138, + "step": 12000 + }, + { + "epoch": 1.8654861758310033, + "grad_norm": 0.5588946342468262, + "learning_rate": 9.866291359726438e-05, + "loss": 0.0128, + "step": 12010 + }, + { + "epoch": 1.86703945324635, + "grad_norm": 0.30820152163505554, + "learning_rate": 9.865911269150861e-05, + "loss": 0.0161, + "step": 12020 + }, + { + "epoch": 1.868592730661696, + "grad_norm": 0.129241481423378, + "learning_rate": 9.865530646446544e-05, + "loss": 0.0134, + "step": 12030 + }, + { + "epoch": 1.8701460080770427, + "grad_norm": 0.24093350768089294, + "learning_rate": 9.86514949165511e-05, + "loss": 0.014, + "step": 12040 + }, + { + "epoch": 1.8716992854923888, + "grad_norm": 0.265403687953949, + "learning_rate": 9.864767804818243e-05, + "loss": 0.0143, + "step": 12050 + }, + { + "epoch": 1.8732525629077355, + "grad_norm": 0.3874143362045288, + "learning_rate": 9.86438558597768e-05, + "loss": 0.0146, + "step": 12060 + }, + { + "epoch": 1.8748058403230816, + "grad_norm": 0.24985696375370026, + "learning_rate": 9.864002835175225e-05, + "loss": 0.0198, + "step": 12070 + }, + { + "epoch": 1.8763591177384282, + "grad_norm": 0.20545673370361328, + "learning_rate": 9.863619552452734e-05, + "loss": 0.0135, + "step": 12080 + }, + { + "epoch": 1.8779123951537744, + "grad_norm": 0.5185883641242981, + "learning_rate": 9.863235737852119e-05, + "loss": 0.0155, + "step": 12090 + }, + { + "epoch": 1.879465672569121, + "grad_norm": 0.25621160864830017, + "learning_rate": 9.862851391415356e-05, + "loss": 0.013, + "step": 12100 + }, + { + "epoch": 1.8810189499844672, + "grad_norm": 0.29373615980148315, + "learning_rate": 9.862466513184477e-05, + "loss": 0.0109, + "step": 12110 + }, + { + "epoch": 1.8825722273998136, + "grad_norm": 0.4556388854980469, + "learning_rate": 9.86208110320157e-05, + "loss": 0.0164, + "step": 12120 + }, + { + "epoch": 1.88412550481516, + "grad_norm": 0.27095654606819153, + "learning_rate": 9.861695161508784e-05, + "loss": 0.0123, + "step": 12130 + }, + { + "epoch": 1.8856787822305063, + "grad_norm": 0.31013786792755127, + "learning_rate": 9.861308688148324e-05, + "loss": 0.0135, + "step": 12140 + }, + { + "epoch": 1.8872320596458527, + "grad_norm": 0.3064512610435486, + "learning_rate": 9.860921683162455e-05, + "loss": 0.0125, + "step": 12150 + }, + { + "epoch": 1.888785337061199, + "grad_norm": 0.34376034140586853, + "learning_rate": 9.860534146593499e-05, + "loss": 0.0176, + "step": 12160 + }, + { + "epoch": 1.8903386144765455, + "grad_norm": 0.4108976125717163, + "learning_rate": 9.860146078483836e-05, + "loss": 0.0146, + "step": 12170 + }, + { + "epoch": 1.8918918918918919, + "grad_norm": 0.29663142561912537, + "learning_rate": 9.859757478875905e-05, + "loss": 0.0167, + "step": 12180 + }, + { + "epoch": 1.8934451693072383, + "grad_norm": 0.3589955270290375, + "learning_rate": 9.859368347812204e-05, + "loss": 0.0149, + "step": 12190 + }, + { + "epoch": 1.8949984467225847, + "grad_norm": 0.3675268292427063, + "learning_rate": 9.858978685335285e-05, + "loss": 0.0144, + "step": 12200 + }, + { + "epoch": 1.896551724137931, + "grad_norm": 0.4716379642486572, + "learning_rate": 9.858588491487763e-05, + "loss": 0.0148, + "step": 12210 + }, + { + "epoch": 1.8981050015532774, + "grad_norm": 0.28929492831230164, + "learning_rate": 9.858197766312308e-05, + "loss": 0.0141, + "step": 12220 + }, + { + "epoch": 1.8996582789686238, + "grad_norm": 0.47330349683761597, + "learning_rate": 9.857806509851649e-05, + "loss": 0.0172, + "step": 12230 + }, + { + "epoch": 1.9012115563839702, + "grad_norm": 0.483556866645813, + "learning_rate": 9.857414722148574e-05, + "loss": 0.016, + "step": 12240 + }, + { + "epoch": 1.9027648337993166, + "grad_norm": 0.5114333033561707, + "learning_rate": 9.857022403245928e-05, + "loss": 0.0136, + "step": 12250 + }, + { + "epoch": 1.904318111214663, + "grad_norm": 0.19761058688163757, + "learning_rate": 9.856629553186615e-05, + "loss": 0.0171, + "step": 12260 + }, + { + "epoch": 1.9058713886300094, + "grad_norm": 0.26558852195739746, + "learning_rate": 9.856236172013595e-05, + "loss": 0.0162, + "step": 12270 + }, + { + "epoch": 1.9074246660453555, + "grad_norm": 0.3281805217266083, + "learning_rate": 9.85584225976989e-05, + "loss": 0.0157, + "step": 12280 + }, + { + "epoch": 1.9089779434607022, + "grad_norm": 0.3274213969707489, + "learning_rate": 9.855447816498575e-05, + "loss": 0.0149, + "step": 12290 + }, + { + "epoch": 1.9105312208760483, + "grad_norm": 0.4070855677127838, + "learning_rate": 9.855052842242787e-05, + "loss": 0.0126, + "step": 12300 + }, + { + "epoch": 1.912084498291395, + "grad_norm": 0.45005494356155396, + "learning_rate": 9.85465733704572e-05, + "loss": 0.0161, + "step": 12310 + }, + { + "epoch": 1.913637775706741, + "grad_norm": 0.1503768116235733, + "learning_rate": 9.854261300950624e-05, + "loss": 0.0139, + "step": 12320 + }, + { + "epoch": 1.9151910531220877, + "grad_norm": 0.3132115602493286, + "learning_rate": 9.853864734000813e-05, + "loss": 0.012, + "step": 12330 + }, + { + "epoch": 1.9167443305374339, + "grad_norm": 0.5219056606292725, + "learning_rate": 9.85346763623965e-05, + "loss": 0.0136, + "step": 12340 + }, + { + "epoch": 1.9182976079527805, + "grad_norm": 0.2993656396865845, + "learning_rate": 9.853070007710564e-05, + "loss": 0.013, + "step": 12350 + }, + { + "epoch": 1.9198508853681266, + "grad_norm": 0.40840092301368713, + "learning_rate": 9.85267184845704e-05, + "loss": 0.0107, + "step": 12360 + }, + { + "epoch": 1.9214041627834733, + "grad_norm": 0.3155345022678375, + "learning_rate": 9.852273158522616e-05, + "loss": 0.0118, + "step": 12370 + }, + { + "epoch": 1.9229574401988194, + "grad_norm": 0.2769092321395874, + "learning_rate": 9.851873937950896e-05, + "loss": 0.0202, + "step": 12380 + }, + { + "epoch": 1.924510717614166, + "grad_norm": 0.48541203141212463, + "learning_rate": 9.851474186785537e-05, + "loss": 0.0159, + "step": 12390 + }, + { + "epoch": 1.9260639950295122, + "grad_norm": 0.337337464094162, + "learning_rate": 9.851073905070254e-05, + "loss": 0.0125, + "step": 12400 + }, + { + "epoch": 1.9276172724448588, + "grad_norm": 0.22497773170471191, + "learning_rate": 9.850673092848824e-05, + "loss": 0.0137, + "step": 12410 + }, + { + "epoch": 1.929170549860205, + "grad_norm": 0.21915742754936218, + "learning_rate": 9.850271750165077e-05, + "loss": 0.0119, + "step": 12420 + }, + { + "epoch": 1.9307238272755514, + "grad_norm": 0.3788263201713562, + "learning_rate": 9.849869877062902e-05, + "loss": 0.0139, + "step": 12430 + }, + { + "epoch": 1.9322771046908978, + "grad_norm": 0.25957435369491577, + "learning_rate": 9.849467473586252e-05, + "loss": 0.0102, + "step": 12440 + }, + { + "epoch": 1.9338303821062441, + "grad_norm": 0.24391841888427734, + "learning_rate": 9.849064539779127e-05, + "loss": 0.0106, + "step": 12450 + }, + { + "epoch": 1.9353836595215905, + "grad_norm": 0.2485792487859726, + "learning_rate": 9.848661075685594e-05, + "loss": 0.0172, + "step": 12460 + }, + { + "epoch": 1.936936936936937, + "grad_norm": 0.19192522764205933, + "learning_rate": 9.848257081349778e-05, + "loss": 0.0117, + "step": 12470 + }, + { + "epoch": 1.9384902143522833, + "grad_norm": 0.4180350601673126, + "learning_rate": 9.847852556815856e-05, + "loss": 0.0165, + "step": 12480 + }, + { + "epoch": 1.9400434917676297, + "grad_norm": 0.32827967405319214, + "learning_rate": 9.847447502128067e-05, + "loss": 0.0157, + "step": 12490 + }, + { + "epoch": 1.941596769182976, + "grad_norm": 0.20994390547275543, + "learning_rate": 9.847041917330708e-05, + "loss": 0.0132, + "step": 12500 + }, + { + "epoch": 1.9431500465983225, + "grad_norm": 0.2308572232723236, + "learning_rate": 9.846635802468132e-05, + "loss": 0.0158, + "step": 12510 + }, + { + "epoch": 1.9447033240136689, + "grad_norm": 0.29315385222435, + "learning_rate": 9.84622915758475e-05, + "loss": 0.0125, + "step": 12520 + }, + { + "epoch": 1.9462566014290152, + "grad_norm": 0.3045196831226349, + "learning_rate": 9.845821982725034e-05, + "loss": 0.0137, + "step": 12530 + }, + { + "epoch": 1.9478098788443616, + "grad_norm": 0.35516223311424255, + "learning_rate": 9.845414277933514e-05, + "loss": 0.0113, + "step": 12540 + }, + { + "epoch": 1.949363156259708, + "grad_norm": 0.27106499671936035, + "learning_rate": 9.845006043254771e-05, + "loss": 0.017, + "step": 12550 + }, + { + "epoch": 1.9509164336750544, + "grad_norm": 0.24038876593112946, + "learning_rate": 9.844597278733451e-05, + "loss": 0.0148, + "step": 12560 + }, + { + "epoch": 1.9524697110904008, + "grad_norm": 0.24547798931598663, + "learning_rate": 9.844187984414259e-05, + "loss": 0.0147, + "step": 12570 + }, + { + "epoch": 1.9540229885057472, + "grad_norm": 0.4026097357273102, + "learning_rate": 9.84377816034195e-05, + "loss": 0.0147, + "step": 12580 + }, + { + "epoch": 1.9555762659210933, + "grad_norm": 0.2684248983860016, + "learning_rate": 9.843367806561345e-05, + "loss": 0.0146, + "step": 12590 + }, + { + "epoch": 1.95712954333644, + "grad_norm": 0.29501277208328247, + "learning_rate": 9.842956923117317e-05, + "loss": 0.0127, + "step": 12600 + }, + { + "epoch": 1.9586828207517861, + "grad_norm": 0.3273240029811859, + "learning_rate": 9.842545510054802e-05, + "loss": 0.0147, + "step": 12610 + }, + { + "epoch": 1.9602360981671327, + "grad_norm": 0.3453301191329956, + "learning_rate": 9.842133567418792e-05, + "loss": 0.0161, + "step": 12620 + }, + { + "epoch": 1.961789375582479, + "grad_norm": 0.38221266865730286, + "learning_rate": 9.841721095254333e-05, + "loss": 0.0143, + "step": 12630 + }, + { + "epoch": 1.9633426529978255, + "grad_norm": 0.21378526091575623, + "learning_rate": 9.841308093606537e-05, + "loss": 0.012, + "step": 12640 + }, + { + "epoch": 1.9648959304131717, + "grad_norm": 0.5065209269523621, + "learning_rate": 9.840894562520565e-05, + "loss": 0.0136, + "step": 12650 + }, + { + "epoch": 1.9664492078285183, + "grad_norm": 0.42984265089035034, + "learning_rate": 9.840480502041642e-05, + "loss": 0.0125, + "step": 12660 + }, + { + "epoch": 1.9680024852438645, + "grad_norm": 0.3190993368625641, + "learning_rate": 9.840065912215049e-05, + "loss": 0.0115, + "step": 12670 + }, + { + "epoch": 1.969555762659211, + "grad_norm": 0.20274078845977783, + "learning_rate": 9.839650793086124e-05, + "loss": 0.0105, + "step": 12680 + }, + { + "epoch": 1.9711090400745572, + "grad_norm": 0.33617034554481506, + "learning_rate": 9.839235144700265e-05, + "loss": 0.0161, + "step": 12690 + }, + { + "epoch": 1.9726623174899038, + "grad_norm": 0.1822579801082611, + "learning_rate": 9.838818967102926e-05, + "loss": 0.0152, + "step": 12700 + }, + { + "epoch": 1.97421559490525, + "grad_norm": 0.2877594530582428, + "learning_rate": 9.83840226033962e-05, + "loss": 0.0112, + "step": 12710 + }, + { + "epoch": 1.9757688723205966, + "grad_norm": 0.42716169357299805, + "learning_rate": 9.837985024455918e-05, + "loss": 0.0149, + "step": 12720 + }, + { + "epoch": 1.9773221497359428, + "grad_norm": 0.3210507035255432, + "learning_rate": 9.837567259497447e-05, + "loss": 0.013, + "step": 12730 + }, + { + "epoch": 1.9788754271512892, + "grad_norm": 0.34443825483322144, + "learning_rate": 9.837148965509894e-05, + "loss": 0.0133, + "step": 12740 + }, + { + "epoch": 1.9804287045666356, + "grad_norm": 0.35667768120765686, + "learning_rate": 9.836730142539001e-05, + "loss": 0.0151, + "step": 12750 + }, + { + "epoch": 1.981981981981982, + "grad_norm": 1.5943374633789062, + "learning_rate": 9.836310790630574e-05, + "loss": 0.0135, + "step": 12760 + }, + { + "epoch": 1.9835352593973283, + "grad_norm": 0.23242539167404175, + "learning_rate": 9.83589090983047e-05, + "loss": 0.0103, + "step": 12770 + }, + { + "epoch": 1.9850885368126747, + "grad_norm": 0.29163387417793274, + "learning_rate": 9.835470500184605e-05, + "loss": 0.0169, + "step": 12780 + }, + { + "epoch": 1.986641814228021, + "grad_norm": 0.3990844488143921, + "learning_rate": 9.835049561738957e-05, + "loss": 0.016, + "step": 12790 + }, + { + "epoch": 1.9881950916433675, + "grad_norm": 0.39342179894447327, + "learning_rate": 9.834628094539558e-05, + "loss": 0.0164, + "step": 12800 + }, + { + "epoch": 1.9897483690587139, + "grad_norm": 0.2541835308074951, + "learning_rate": 9.834206098632499e-05, + "loss": 0.0163, + "step": 12810 + }, + { + "epoch": 1.9913016464740603, + "grad_norm": 0.3945200741291046, + "learning_rate": 9.833783574063931e-05, + "loss": 0.0163, + "step": 12820 + }, + { + "epoch": 1.9928549238894067, + "grad_norm": 0.2991805076599121, + "learning_rate": 9.833360520880058e-05, + "loss": 0.0148, + "step": 12830 + }, + { + "epoch": 1.994408201304753, + "grad_norm": 0.2775903344154358, + "learning_rate": 9.832936939127144e-05, + "loss": 0.0143, + "step": 12840 + }, + { + "epoch": 1.9959614787200994, + "grad_norm": 0.31346824765205383, + "learning_rate": 9.832512828851515e-05, + "loss": 0.0127, + "step": 12850 + }, + { + "epoch": 1.9975147561354458, + "grad_norm": 0.35409030318260193, + "learning_rate": 9.832088190099546e-05, + "loss": 0.0178, + "step": 12860 + }, + { + "epoch": 1.9990680335507922, + "grad_norm": 0.2777476906776428, + "learning_rate": 9.831663022917679e-05, + "loss": 0.0133, + "step": 12870 + }, + { + "epoch": 2.0006213109661384, + "grad_norm": 0.3078782558441162, + "learning_rate": 9.831237327352407e-05, + "loss": 0.0138, + "step": 12880 + }, + { + "epoch": 2.002174588381485, + "grad_norm": 0.39905938506126404, + "learning_rate": 9.830811103450286e-05, + "loss": 0.0107, + "step": 12890 + }, + { + "epoch": 2.003727865796831, + "grad_norm": 0.31053584814071655, + "learning_rate": 9.830384351257924e-05, + "loss": 0.0119, + "step": 12900 + }, + { + "epoch": 2.0052811432121778, + "grad_norm": 0.2651865482330322, + "learning_rate": 9.829957070821993e-05, + "loss": 0.0103, + "step": 12910 + }, + { + "epoch": 2.006834420627524, + "grad_norm": 0.39772099256515503, + "learning_rate": 9.829529262189218e-05, + "loss": 0.0155, + "step": 12920 + }, + { + "epoch": 2.0083876980428705, + "grad_norm": 0.3280603289604187, + "learning_rate": 9.829100925406385e-05, + "loss": 0.0136, + "step": 12930 + }, + { + "epoch": 2.0099409754582167, + "grad_norm": 0.4283953309059143, + "learning_rate": 9.828672060520333e-05, + "loss": 0.0091, + "step": 12940 + }, + { + "epoch": 2.0114942528735633, + "grad_norm": 0.4071812927722931, + "learning_rate": 9.828242667577966e-05, + "loss": 0.014, + "step": 12950 + }, + { + "epoch": 2.0130475302889095, + "grad_norm": 0.37846773862838745, + "learning_rate": 9.82781274662624e-05, + "loss": 0.0145, + "step": 12960 + }, + { + "epoch": 2.014600807704256, + "grad_norm": 0.5168367028236389, + "learning_rate": 9.82738229771217e-05, + "loss": 0.0146, + "step": 12970 + }, + { + "epoch": 2.0161540851196023, + "grad_norm": 0.25568118691444397, + "learning_rate": 9.826951320882829e-05, + "loss": 0.016, + "step": 12980 + }, + { + "epoch": 2.017707362534949, + "grad_norm": 0.33470773696899414, + "learning_rate": 9.826519816185351e-05, + "loss": 0.0134, + "step": 12990 + }, + { + "epoch": 2.019260639950295, + "grad_norm": 0.37861168384552, + "learning_rate": 9.826087783666921e-05, + "loss": 0.0148, + "step": 13000 + }, + { + "epoch": 2.0208139173656416, + "grad_norm": 0.44316351413726807, + "learning_rate": 9.825655223374787e-05, + "loss": 0.0177, + "step": 13010 + }, + { + "epoch": 2.022367194780988, + "grad_norm": 0.25743964314460754, + "learning_rate": 9.825222135356253e-05, + "loss": 0.0135, + "step": 13020 + }, + { + "epoch": 2.0239204721963344, + "grad_norm": 0.2714439630508423, + "learning_rate": 9.82478851965868e-05, + "loss": 0.0144, + "step": 13030 + }, + { + "epoch": 2.0254737496116806, + "grad_norm": 0.4788552224636078, + "learning_rate": 9.82435437632949e-05, + "loss": 0.014, + "step": 13040 + }, + { + "epoch": 2.027027027027027, + "grad_norm": 0.2622465491294861, + "learning_rate": 9.823919705416158e-05, + "loss": 0.0124, + "step": 13050 + }, + { + "epoch": 2.0285803044423734, + "grad_norm": 0.40123382210731506, + "learning_rate": 9.82348450696622e-05, + "loss": 0.0132, + "step": 13060 + }, + { + "epoch": 2.03013358185772, + "grad_norm": 0.2944749891757965, + "learning_rate": 9.823048781027268e-05, + "loss": 0.0125, + "step": 13070 + }, + { + "epoch": 2.031686859273066, + "grad_norm": 0.24467480182647705, + "learning_rate": 9.822612527646953e-05, + "loss": 0.0132, + "step": 13080 + }, + { + "epoch": 2.0332401366884127, + "grad_norm": 0.2738519310951233, + "learning_rate": 9.822175746872984e-05, + "loss": 0.0124, + "step": 13090 + }, + { + "epoch": 2.034793414103759, + "grad_norm": 0.46693551540374756, + "learning_rate": 9.821738438753123e-05, + "loss": 0.0122, + "step": 13100 + }, + { + "epoch": 2.0363466915191055, + "grad_norm": 0.22979436814785004, + "learning_rate": 9.821300603335196e-05, + "loss": 0.015, + "step": 13110 + }, + { + "epoch": 2.0378999689344517, + "grad_norm": 0.21508069336414337, + "learning_rate": 9.820862240667085e-05, + "loss": 0.0161, + "step": 13120 + }, + { + "epoch": 2.0394532463497983, + "grad_norm": 0.3473801612854004, + "learning_rate": 9.820423350796726e-05, + "loss": 0.0174, + "step": 13130 + }, + { + "epoch": 2.0410065237651445, + "grad_norm": 0.15030115842819214, + "learning_rate": 9.819983933772118e-05, + "loss": 0.0107, + "step": 13140 + }, + { + "epoch": 2.0425598011804906, + "grad_norm": 0.28895825147628784, + "learning_rate": 9.819543989641314e-05, + "loss": 0.015, + "step": 13150 + }, + { + "epoch": 2.0441130785958372, + "grad_norm": 0.3666306734085083, + "learning_rate": 9.819103518452423e-05, + "loss": 0.0148, + "step": 13160 + }, + { + "epoch": 2.0456663560111834, + "grad_norm": 0.24220244586467743, + "learning_rate": 9.818662520253618e-05, + "loss": 0.0152, + "step": 13170 + }, + { + "epoch": 2.04721963342653, + "grad_norm": 0.14396004378795624, + "learning_rate": 9.818220995093126e-05, + "loss": 0.0127, + "step": 13180 + }, + { + "epoch": 2.048772910841876, + "grad_norm": 0.34982454776763916, + "learning_rate": 9.817778943019228e-05, + "loss": 0.0126, + "step": 13190 + }, + { + "epoch": 2.050326188257223, + "grad_norm": 0.4462582468986511, + "learning_rate": 9.81733636408027e-05, + "loss": 0.015, + "step": 13200 + }, + { + "epoch": 2.051879465672569, + "grad_norm": 0.39449894428253174, + "learning_rate": 9.816893258324649e-05, + "loss": 0.0128, + "step": 13210 + }, + { + "epoch": 2.0534327430879156, + "grad_norm": 0.3983723223209381, + "learning_rate": 9.816449625800823e-05, + "loss": 0.0128, + "step": 13220 + }, + { + "epoch": 2.0549860205032617, + "grad_norm": 0.3119395673274994, + "learning_rate": 9.816005466557308e-05, + "loss": 0.0159, + "step": 13230 + }, + { + "epoch": 2.0565392979186083, + "grad_norm": 0.23894605040550232, + "learning_rate": 9.815560780642674e-05, + "loss": 0.0152, + "step": 13240 + }, + { + "epoch": 2.0580925753339545, + "grad_norm": 0.19377082586288452, + "learning_rate": 9.815115568105555e-05, + "loss": 0.0132, + "step": 13250 + }, + { + "epoch": 2.059645852749301, + "grad_norm": 0.3007720708847046, + "learning_rate": 9.814669828994638e-05, + "loss": 0.0159, + "step": 13260 + }, + { + "epoch": 2.0611991301646473, + "grad_norm": 0.3847615718841553, + "learning_rate": 9.814223563358665e-05, + "loss": 0.0145, + "step": 13270 + }, + { + "epoch": 2.062752407579994, + "grad_norm": 0.3416416049003601, + "learning_rate": 9.813776771246443e-05, + "loss": 0.0153, + "step": 13280 + }, + { + "epoch": 2.06430568499534, + "grad_norm": 0.24457475543022156, + "learning_rate": 9.813329452706829e-05, + "loss": 0.0156, + "step": 13290 + }, + { + "epoch": 2.0658589624106867, + "grad_norm": 0.3087742328643799, + "learning_rate": 9.812881607788744e-05, + "loss": 0.0196, + "step": 13300 + }, + { + "epoch": 2.067412239826033, + "grad_norm": 0.3458199203014374, + "learning_rate": 9.812433236541163e-05, + "loss": 0.0143, + "step": 13310 + }, + { + "epoch": 2.0689655172413794, + "grad_norm": 0.25594449043273926, + "learning_rate": 9.811984339013116e-05, + "loss": 0.0133, + "step": 13320 + }, + { + "epoch": 2.0705187946567256, + "grad_norm": 0.3160611391067505, + "learning_rate": 9.811534915253698e-05, + "loss": 0.0175, + "step": 13330 + }, + { + "epoch": 2.0720720720720722, + "grad_norm": 0.3036614954471588, + "learning_rate": 9.811084965312056e-05, + "loss": 0.0134, + "step": 13340 + }, + { + "epoch": 2.0736253494874184, + "grad_norm": 0.18928807973861694, + "learning_rate": 9.810634489237396e-05, + "loss": 0.0136, + "step": 13350 + }, + { + "epoch": 2.075178626902765, + "grad_norm": 0.32245343923568726, + "learning_rate": 9.81018348707898e-05, + "loss": 0.0135, + "step": 13360 + }, + { + "epoch": 2.076731904318111, + "grad_norm": 0.18813581764698029, + "learning_rate": 9.809731958886131e-05, + "loss": 0.0126, + "step": 13370 + }, + { + "epoch": 2.0782851817334578, + "grad_norm": 0.25022444128990173, + "learning_rate": 9.809279904708224e-05, + "loss": 0.017, + "step": 13380 + }, + { + "epoch": 2.079838459148804, + "grad_norm": 0.3615023195743561, + "learning_rate": 9.808827324594699e-05, + "loss": 0.0252, + "step": 13390 + }, + { + "epoch": 2.0813917365641506, + "grad_norm": 0.2965255081653595, + "learning_rate": 9.808374218595046e-05, + "loss": 0.0159, + "step": 13400 + }, + { + "epoch": 2.0829450139794967, + "grad_norm": 0.42309609055519104, + "learning_rate": 9.80792058675882e-05, + "loss": 0.0143, + "step": 13410 + }, + { + "epoch": 2.0844982913948433, + "grad_norm": 0.2894909977912903, + "learning_rate": 9.807466429135627e-05, + "loss": 0.0149, + "step": 13420 + }, + { + "epoch": 2.0860515688101895, + "grad_norm": 0.36917322874069214, + "learning_rate": 9.807011745775132e-05, + "loss": 0.0167, + "step": 13430 + }, + { + "epoch": 2.0876048462255357, + "grad_norm": 0.44244417548179626, + "learning_rate": 9.806556536727061e-05, + "loss": 0.0176, + "step": 13440 + }, + { + "epoch": 2.0891581236408823, + "grad_norm": 0.2835220396518707, + "learning_rate": 9.806100802041193e-05, + "loss": 0.0155, + "step": 13450 + }, + { + "epoch": 2.0907114010562284, + "grad_norm": 0.2442038357257843, + "learning_rate": 9.805644541767368e-05, + "loss": 0.0167, + "step": 13460 + }, + { + "epoch": 2.092264678471575, + "grad_norm": 0.17360958456993103, + "learning_rate": 9.805187755955478e-05, + "loss": 0.0139, + "step": 13470 + }, + { + "epoch": 2.093817955886921, + "grad_norm": 0.2356140911579132, + "learning_rate": 9.804730444655483e-05, + "loss": 0.0137, + "step": 13480 + }, + { + "epoch": 2.095371233302268, + "grad_norm": 0.42833319306373596, + "learning_rate": 9.804272607917388e-05, + "loss": 0.0217, + "step": 13490 + }, + { + "epoch": 2.096924510717614, + "grad_norm": 0.2545282244682312, + "learning_rate": 9.803814245791265e-05, + "loss": 0.0197, + "step": 13500 + }, + { + "epoch": 2.0984777881329606, + "grad_norm": 0.35604193806648254, + "learning_rate": 9.803355358327239e-05, + "loss": 0.0161, + "step": 13510 + }, + { + "epoch": 2.1000310655483068, + "grad_norm": 0.24332645535469055, + "learning_rate": 9.802895945575492e-05, + "loss": 0.0152, + "step": 13520 + }, + { + "epoch": 2.1015843429636534, + "grad_norm": 0.2519283592700958, + "learning_rate": 9.802436007586266e-05, + "loss": 0.0159, + "step": 13530 + }, + { + "epoch": 2.1031376203789995, + "grad_norm": 0.30917903780937195, + "learning_rate": 9.801975544409858e-05, + "loss": 0.0155, + "step": 13540 + }, + { + "epoch": 2.104690897794346, + "grad_norm": 0.5678055882453918, + "learning_rate": 9.801514556096625e-05, + "loss": 0.0147, + "step": 13550 + }, + { + "epoch": 2.1062441752096923, + "grad_norm": 0.38022440671920776, + "learning_rate": 9.801053042696977e-05, + "loss": 0.0164, + "step": 13560 + }, + { + "epoch": 2.107797452625039, + "grad_norm": 0.29564571380615234, + "learning_rate": 9.800591004261388e-05, + "loss": 0.0116, + "step": 13570 + }, + { + "epoch": 2.109350730040385, + "grad_norm": 0.40859508514404297, + "learning_rate": 9.800128440840385e-05, + "loss": 0.0174, + "step": 13580 + }, + { + "epoch": 2.1109040074557317, + "grad_norm": 0.3424898684024811, + "learning_rate": 9.799665352484552e-05, + "loss": 0.0134, + "step": 13590 + }, + { + "epoch": 2.112457284871078, + "grad_norm": 0.4232414662837982, + "learning_rate": 9.799201739244532e-05, + "loss": 0.014, + "step": 13600 + }, + { + "epoch": 2.1140105622864245, + "grad_norm": 0.34474873542785645, + "learning_rate": 9.798737601171025e-05, + "loss": 0.0135, + "step": 13610 + }, + { + "epoch": 2.1155638397017706, + "grad_norm": 0.47152793407440186, + "learning_rate": 9.79827293831479e-05, + "loss": 0.0135, + "step": 13620 + }, + { + "epoch": 2.1171171171171173, + "grad_norm": 0.3486926555633545, + "learning_rate": 9.797807750726638e-05, + "loss": 0.0139, + "step": 13630 + }, + { + "epoch": 2.1186703945324634, + "grad_norm": 0.317797988653183, + "learning_rate": 9.797342038457446e-05, + "loss": 0.0174, + "step": 13640 + }, + { + "epoch": 2.12022367194781, + "grad_norm": 0.29568910598754883, + "learning_rate": 9.796875801558141e-05, + "loss": 0.0124, + "step": 13650 + }, + { + "epoch": 2.121776949363156, + "grad_norm": 0.25911492109298706, + "learning_rate": 9.79640904007971e-05, + "loss": 0.0135, + "step": 13660 + }, + { + "epoch": 2.123330226778503, + "grad_norm": 0.3711915910243988, + "learning_rate": 9.795941754073199e-05, + "loss": 0.0113, + "step": 13670 + }, + { + "epoch": 2.124883504193849, + "grad_norm": 0.2332984060049057, + "learning_rate": 9.795473943589705e-05, + "loss": 0.0164, + "step": 13680 + }, + { + "epoch": 2.1264367816091956, + "grad_norm": 0.3209548890590668, + "learning_rate": 9.795005608680394e-05, + "loss": 0.0139, + "step": 13690 + }, + { + "epoch": 2.1279900590245417, + "grad_norm": 0.16813968122005463, + "learning_rate": 9.794536749396477e-05, + "loss": 0.0118, + "step": 13700 + }, + { + "epoch": 2.1295433364398884, + "grad_norm": 0.45157694816589355, + "learning_rate": 9.79406736578923e-05, + "loss": 0.0138, + "step": 13710 + }, + { + "epoch": 2.1310966138552345, + "grad_norm": 0.29740041494369507, + "learning_rate": 9.793597457909984e-05, + "loss": 0.0122, + "step": 13720 + }, + { + "epoch": 2.132649891270581, + "grad_norm": 0.45295798778533936, + "learning_rate": 9.793127025810127e-05, + "loss": 0.0174, + "step": 13730 + }, + { + "epoch": 2.1342031686859273, + "grad_norm": 0.45100387930870056, + "learning_rate": 9.792656069541104e-05, + "loss": 0.0165, + "step": 13740 + }, + { + "epoch": 2.135756446101274, + "grad_norm": 0.3191740810871124, + "learning_rate": 9.79218458915442e-05, + "loss": 0.0153, + "step": 13750 + }, + { + "epoch": 2.13730972351662, + "grad_norm": 0.43842068314552307, + "learning_rate": 9.791712584701634e-05, + "loss": 0.014, + "step": 13760 + }, + { + "epoch": 2.1388630009319662, + "grad_norm": 0.26254868507385254, + "learning_rate": 9.791240056234364e-05, + "loss": 0.0134, + "step": 13770 + }, + { + "epoch": 2.140416278347313, + "grad_norm": 0.24207712709903717, + "learning_rate": 9.790767003804283e-05, + "loss": 0.0125, + "step": 13780 + }, + { + "epoch": 2.141969555762659, + "grad_norm": 0.26398414373397827, + "learning_rate": 9.790293427463126e-05, + "loss": 0.0151, + "step": 13790 + }, + { + "epoch": 2.1435228331780056, + "grad_norm": 0.41865259408950806, + "learning_rate": 9.789819327262684e-05, + "loss": 0.0157, + "step": 13800 + }, + { + "epoch": 2.145076110593352, + "grad_norm": 0.5024716854095459, + "learning_rate": 9.7893447032548e-05, + "loss": 0.0192, + "step": 13810 + }, + { + "epoch": 2.1466293880086984, + "grad_norm": 0.46439671516418457, + "learning_rate": 9.78886955549138e-05, + "loss": 0.0125, + "step": 13820 + }, + { + "epoch": 2.1481826654240446, + "grad_norm": 0.3791353106498718, + "learning_rate": 9.788393884024387e-05, + "loss": 0.0131, + "step": 13830 + }, + { + "epoch": 2.149735942839391, + "grad_norm": 0.37531912326812744, + "learning_rate": 9.787917688905836e-05, + "loss": 0.0135, + "step": 13840 + }, + { + "epoch": 2.1512892202547373, + "grad_norm": 0.34737175703048706, + "learning_rate": 9.787440970187807e-05, + "loss": 0.0127, + "step": 13850 + }, + { + "epoch": 2.152842497670084, + "grad_norm": 0.2810872197151184, + "learning_rate": 9.786963727922429e-05, + "loss": 0.0167, + "step": 13860 + }, + { + "epoch": 2.15439577508543, + "grad_norm": 0.32190677523612976, + "learning_rate": 9.786485962161897e-05, + "loss": 0.0102, + "step": 13870 + }, + { + "epoch": 2.1559490525007767, + "grad_norm": 0.2752133905887604, + "learning_rate": 9.786007672958455e-05, + "loss": 0.0181, + "step": 13880 + }, + { + "epoch": 2.157502329916123, + "grad_norm": 0.2765721082687378, + "learning_rate": 9.78552886036441e-05, + "loss": 0.0118, + "step": 13890 + }, + { + "epoch": 2.1590556073314695, + "grad_norm": 0.2598908841609955, + "learning_rate": 9.785049524432124e-05, + "loss": 0.0108, + "step": 13900 + }, + { + "epoch": 2.1606088847468157, + "grad_norm": 0.2565263509750366, + "learning_rate": 9.784569665214016e-05, + "loss": 0.0123, + "step": 13910 + }, + { + "epoch": 2.1621621621621623, + "grad_norm": 0.26543888449668884, + "learning_rate": 9.784089282762563e-05, + "loss": 0.0114, + "step": 13920 + }, + { + "epoch": 2.1637154395775084, + "grad_norm": 0.2745278775691986, + "learning_rate": 9.7836083771303e-05, + "loss": 0.0119, + "step": 13930 + }, + { + "epoch": 2.165268716992855, + "grad_norm": 0.20877893269062042, + "learning_rate": 9.783126948369817e-05, + "loss": 0.0173, + "step": 13940 + }, + { + "epoch": 2.1668219944082012, + "grad_norm": 0.17738816142082214, + "learning_rate": 9.78264499653376e-05, + "loss": 0.0134, + "step": 13950 + }, + { + "epoch": 2.168375271823548, + "grad_norm": 0.34231671690940857, + "learning_rate": 9.782162521674838e-05, + "loss": 0.0119, + "step": 13960 + }, + { + "epoch": 2.169928549238894, + "grad_norm": 0.22761917114257812, + "learning_rate": 9.781679523845812e-05, + "loss": 0.0119, + "step": 13970 + }, + { + "epoch": 2.1714818266542406, + "grad_norm": 0.6802140474319458, + "learning_rate": 9.781196003099502e-05, + "loss": 0.0149, + "step": 13980 + }, + { + "epoch": 2.1730351040695868, + "grad_norm": 0.35607239603996277, + "learning_rate": 9.780711959488786e-05, + "loss": 0.0156, + "step": 13990 + }, + { + "epoch": 2.1745883814849334, + "grad_norm": 0.1504765897989273, + "learning_rate": 9.780227393066599e-05, + "loss": 0.009, + "step": 14000 + }, + { + "epoch": 2.1761416589002796, + "grad_norm": 0.2421502321958542, + "learning_rate": 9.77974230388593e-05, + "loss": 0.012, + "step": 14010 + }, + { + "epoch": 2.177694936315626, + "grad_norm": 0.27549728751182556, + "learning_rate": 9.779256691999829e-05, + "loss": 0.0122, + "step": 14020 + }, + { + "epoch": 2.1792482137309723, + "grad_norm": 0.36897012591362, + "learning_rate": 9.778770557461403e-05, + "loss": 0.0127, + "step": 14030 + }, + { + "epoch": 2.180801491146319, + "grad_norm": 0.29670676589012146, + "learning_rate": 9.778283900323812e-05, + "loss": 0.016, + "step": 14040 + }, + { + "epoch": 2.182354768561665, + "grad_norm": 0.1790635585784912, + "learning_rate": 9.777796720640277e-05, + "loss": 0.0123, + "step": 14050 + }, + { + "epoch": 2.1839080459770113, + "grad_norm": 0.33207815885543823, + "learning_rate": 9.777309018464078e-05, + "loss": 0.0099, + "step": 14060 + }, + { + "epoch": 2.185461323392358, + "grad_norm": 0.23777256906032562, + "learning_rate": 9.776820793848547e-05, + "loss": 0.0125, + "step": 14070 + }, + { + "epoch": 2.1870146008077045, + "grad_norm": 0.639934241771698, + "learning_rate": 9.776332046847075e-05, + "loss": 0.0132, + "step": 14080 + }, + { + "epoch": 2.1885678782230507, + "grad_norm": 0.4759534001350403, + "learning_rate": 9.775842777513111e-05, + "loss": 0.0132, + "step": 14090 + }, + { + "epoch": 2.190121155638397, + "grad_norm": 0.2557327449321747, + "learning_rate": 9.775352985900163e-05, + "loss": 0.0127, + "step": 14100 + }, + { + "epoch": 2.1916744330537434, + "grad_norm": 0.6805595755577087, + "learning_rate": 9.774862672061791e-05, + "loss": 0.0101, + "step": 14110 + }, + { + "epoch": 2.1932277104690896, + "grad_norm": 0.5148204565048218, + "learning_rate": 9.774371836051616e-05, + "loss": 0.0134, + "step": 14120 + }, + { + "epoch": 2.194780987884436, + "grad_norm": 0.28258490562438965, + "learning_rate": 9.773880477923315e-05, + "loss": 0.015, + "step": 14130 + }, + { + "epoch": 2.1963342652997824, + "grad_norm": 0.22563563287258148, + "learning_rate": 9.773388597730623e-05, + "loss": 0.0141, + "step": 14140 + }, + { + "epoch": 2.197887542715129, + "grad_norm": 0.3877614736557007, + "learning_rate": 9.77289619552733e-05, + "loss": 0.0118, + "step": 14150 + }, + { + "epoch": 2.199440820130475, + "grad_norm": 0.41872328519821167, + "learning_rate": 9.772403271367285e-05, + "loss": 0.0164, + "step": 14160 + }, + { + "epoch": 2.2009940975458218, + "grad_norm": 0.3974515497684479, + "learning_rate": 9.771909825304396e-05, + "loss": 0.0154, + "step": 14170 + }, + { + "epoch": 2.202547374961168, + "grad_norm": 0.2851545512676239, + "learning_rate": 9.771415857392619e-05, + "loss": 0.0117, + "step": 14180 + }, + { + "epoch": 2.2041006523765145, + "grad_norm": 0.3933827579021454, + "learning_rate": 9.770921367685978e-05, + "loss": 0.0134, + "step": 14190 + }, + { + "epoch": 2.2056539297918607, + "grad_norm": 0.3050946593284607, + "learning_rate": 9.770426356238551e-05, + "loss": 0.0128, + "step": 14200 + }, + { + "epoch": 2.2072072072072073, + "grad_norm": 0.3217598795890808, + "learning_rate": 9.769930823104469e-05, + "loss": 0.0127, + "step": 14210 + }, + { + "epoch": 2.2087604846225535, + "grad_norm": 0.1802087426185608, + "learning_rate": 9.769434768337926e-05, + "loss": 0.0148, + "step": 14220 + }, + { + "epoch": 2.2103137620379, + "grad_norm": 0.2802642583847046, + "learning_rate": 9.768938191993164e-05, + "loss": 0.0127, + "step": 14230 + }, + { + "epoch": 2.2118670394532463, + "grad_norm": 0.20305560529232025, + "learning_rate": 9.768441094124494e-05, + "loss": 0.0095, + "step": 14240 + }, + { + "epoch": 2.213420316868593, + "grad_norm": 0.4513525664806366, + "learning_rate": 9.767943474786275e-05, + "loss": 0.0111, + "step": 14250 + }, + { + "epoch": 2.214973594283939, + "grad_norm": 0.2779688835144043, + "learning_rate": 9.767445334032923e-05, + "loss": 0.0155, + "step": 14260 + }, + { + "epoch": 2.2165268716992856, + "grad_norm": 0.24542053043842316, + "learning_rate": 9.766946671918919e-05, + "loss": 0.0108, + "step": 14270 + }, + { + "epoch": 2.218080149114632, + "grad_norm": 0.2462293952703476, + "learning_rate": 9.766447488498796e-05, + "loss": 0.0122, + "step": 14280 + }, + { + "epoch": 2.2196334265299784, + "grad_norm": 0.4415476620197296, + "learning_rate": 9.765947783827139e-05, + "loss": 0.0101, + "step": 14290 + }, + { + "epoch": 2.2211867039453246, + "grad_norm": 0.4391491115093231, + "learning_rate": 9.765447557958599e-05, + "loss": 0.0124, + "step": 14300 + }, + { + "epoch": 2.222739981360671, + "grad_norm": 0.2572084665298462, + "learning_rate": 9.764946810947879e-05, + "loss": 0.0101, + "step": 14310 + }, + { + "epoch": 2.2242932587760174, + "grad_norm": 0.31012192368507385, + "learning_rate": 9.764445542849738e-05, + "loss": 0.0154, + "step": 14320 + }, + { + "epoch": 2.225846536191364, + "grad_norm": 0.36681580543518066, + "learning_rate": 9.763943753718998e-05, + "loss": 0.0154, + "step": 14330 + }, + { + "epoch": 2.22739981360671, + "grad_norm": 0.2931648790836334, + "learning_rate": 9.76344144361053e-05, + "loss": 0.0148, + "step": 14340 + }, + { + "epoch": 2.2289530910220567, + "grad_norm": 0.2462402880191803, + "learning_rate": 9.762938612579269e-05, + "loss": 0.0149, + "step": 14350 + }, + { + "epoch": 2.230506368437403, + "grad_norm": 0.23577119410037994, + "learning_rate": 9.762435260680202e-05, + "loss": 0.0119, + "step": 14360 + }, + { + "epoch": 2.2320596458527495, + "grad_norm": 0.20211368799209595, + "learning_rate": 9.761931387968373e-05, + "loss": 0.0147, + "step": 14370 + }, + { + "epoch": 2.2336129232680957, + "grad_norm": 0.30690789222717285, + "learning_rate": 9.76142699449889e-05, + "loss": 0.0133, + "step": 14380 + }, + { + "epoch": 2.235166200683442, + "grad_norm": 0.2597759962081909, + "learning_rate": 9.760922080326908e-05, + "loss": 0.0125, + "step": 14390 + }, + { + "epoch": 2.2367194780987885, + "grad_norm": 0.2953495383262634, + "learning_rate": 9.760416645507644e-05, + "loss": 0.0105, + "step": 14400 + }, + { + "epoch": 2.2382727555141346, + "grad_norm": 0.23623548448085785, + "learning_rate": 9.759910690096375e-05, + "loss": 0.0123, + "step": 14410 + }, + { + "epoch": 2.2398260329294812, + "grad_norm": 0.39663735032081604, + "learning_rate": 9.759404214148429e-05, + "loss": 0.0152, + "step": 14420 + }, + { + "epoch": 2.2413793103448274, + "grad_norm": 0.22406871616840363, + "learning_rate": 9.758897217719191e-05, + "loss": 0.0136, + "step": 14430 + }, + { + "epoch": 2.242932587760174, + "grad_norm": 0.1833263486623764, + "learning_rate": 9.758389700864113e-05, + "loss": 0.0114, + "step": 14440 + }, + { + "epoch": 2.24448586517552, + "grad_norm": 0.24832619726657867, + "learning_rate": 9.757881663638688e-05, + "loss": 0.0183, + "step": 14450 + }, + { + "epoch": 2.246039142590867, + "grad_norm": 0.5701901316642761, + "learning_rate": 9.757373106098478e-05, + "loss": 0.0129, + "step": 14460 + }, + { + "epoch": 2.247592420006213, + "grad_norm": 0.20222721993923187, + "learning_rate": 9.756864028299097e-05, + "loss": 0.0173, + "step": 14470 + }, + { + "epoch": 2.2491456974215596, + "grad_norm": 0.4099556803703308, + "learning_rate": 9.75635443029622e-05, + "loss": 0.0138, + "step": 14480 + }, + { + "epoch": 2.2506989748369057, + "grad_norm": 0.3622557520866394, + "learning_rate": 9.755844312145572e-05, + "loss": 0.0117, + "step": 14490 + }, + { + "epoch": 2.2522522522522523, + "grad_norm": 0.5453993082046509, + "learning_rate": 9.755333673902941e-05, + "loss": 0.0123, + "step": 14500 + }, + { + "epoch": 2.2538055296675985, + "grad_norm": 0.4363246560096741, + "learning_rate": 9.75482251562417e-05, + "loss": 0.0118, + "step": 14510 + }, + { + "epoch": 2.255358807082945, + "grad_norm": 0.3537389039993286, + "learning_rate": 9.754310837365155e-05, + "loss": 0.0126, + "step": 14520 + }, + { + "epoch": 2.2569120844982913, + "grad_norm": 0.5232810974121094, + "learning_rate": 9.753798639181856e-05, + "loss": 0.0169, + "step": 14530 + }, + { + "epoch": 2.258465361913638, + "grad_norm": 0.3708614110946655, + "learning_rate": 9.753285921130286e-05, + "loss": 0.0126, + "step": 14540 + }, + { + "epoch": 2.260018639328984, + "grad_norm": 0.41515928506851196, + "learning_rate": 9.752772683266512e-05, + "loss": 0.0153, + "step": 14550 + }, + { + "epoch": 2.2615719167443307, + "grad_norm": 0.4122466444969177, + "learning_rate": 9.752258925646665e-05, + "loss": 0.013, + "step": 14560 + }, + { + "epoch": 2.263125194159677, + "grad_norm": 0.284193217754364, + "learning_rate": 9.751744648326926e-05, + "loss": 0.0116, + "step": 14570 + }, + { + "epoch": 2.2646784715750234, + "grad_norm": 0.29903754591941833, + "learning_rate": 9.751229851363536e-05, + "loss": 0.0111, + "step": 14580 + }, + { + "epoch": 2.2662317489903696, + "grad_norm": 0.25595059990882874, + "learning_rate": 9.750714534812793e-05, + "loss": 0.0113, + "step": 14590 + }, + { + "epoch": 2.267785026405716, + "grad_norm": 0.43521541357040405, + "learning_rate": 9.750198698731053e-05, + "loss": 0.0142, + "step": 14600 + }, + { + "epoch": 2.2693383038210624, + "grad_norm": 0.2519705891609192, + "learning_rate": 9.749682343174722e-05, + "loss": 0.0139, + "step": 14610 + }, + { + "epoch": 2.270891581236409, + "grad_norm": 0.34966331720352173, + "learning_rate": 9.749165468200272e-05, + "loss": 0.0123, + "step": 14620 + }, + { + "epoch": 2.272444858651755, + "grad_norm": 0.40294885635375977, + "learning_rate": 9.748648073864229e-05, + "loss": 0.0157, + "step": 14630 + }, + { + "epoch": 2.2739981360671018, + "grad_norm": 0.5084206461906433, + "learning_rate": 9.748130160223168e-05, + "loss": 0.0161, + "step": 14640 + }, + { + "epoch": 2.275551413482448, + "grad_norm": 0.24511860311031342, + "learning_rate": 9.747611727333734e-05, + "loss": 0.0119, + "step": 14650 + }, + { + "epoch": 2.2771046908977945, + "grad_norm": 0.32857486605644226, + "learning_rate": 9.74709277525262e-05, + "loss": 0.0137, + "step": 14660 + }, + { + "epoch": 2.2786579683131407, + "grad_norm": 0.2524584233760834, + "learning_rate": 9.746573304036576e-05, + "loss": 0.0147, + "step": 14670 + }, + { + "epoch": 2.280211245728487, + "grad_norm": 0.510093629360199, + "learning_rate": 9.746053313742412e-05, + "loss": 0.0151, + "step": 14680 + }, + { + "epoch": 2.2817645231438335, + "grad_norm": 0.4134204685688019, + "learning_rate": 9.745532804426994e-05, + "loss": 0.0128, + "step": 14690 + }, + { + "epoch": 2.28331780055918, + "grad_norm": 0.39944159984588623, + "learning_rate": 9.745011776147242e-05, + "loss": 0.0171, + "step": 14700 + }, + { + "epoch": 2.2848710779745263, + "grad_norm": 0.3652400076389313, + "learning_rate": 9.744490228960138e-05, + "loss": 0.0173, + "step": 14710 + }, + { + "epoch": 2.2864243553898724, + "grad_norm": 0.2540777325630188, + "learning_rate": 9.743968162922713e-05, + "loss": 0.0128, + "step": 14720 + }, + { + "epoch": 2.287977632805219, + "grad_norm": 0.3105280101299286, + "learning_rate": 9.743445578092064e-05, + "loss": 0.014, + "step": 14730 + }, + { + "epoch": 2.289530910220565, + "grad_norm": 0.29926422238349915, + "learning_rate": 9.742922474525338e-05, + "loss": 0.0147, + "step": 14740 + }, + { + "epoch": 2.291084187635912, + "grad_norm": 0.3901355564594269, + "learning_rate": 9.742398852279741e-05, + "loss": 0.0132, + "step": 14750 + }, + { + "epoch": 2.292637465051258, + "grad_norm": 0.3008464276790619, + "learning_rate": 9.741874711412535e-05, + "loss": 0.0137, + "step": 14760 + }, + { + "epoch": 2.2941907424666046, + "grad_norm": 0.4431647062301636, + "learning_rate": 9.741350051981042e-05, + "loss": 0.0163, + "step": 14770 + }, + { + "epoch": 2.2957440198819508, + "grad_norm": 0.23918361961841583, + "learning_rate": 9.740824874042633e-05, + "loss": 0.0143, + "step": 14780 + }, + { + "epoch": 2.2972972972972974, + "grad_norm": 0.2715335488319397, + "learning_rate": 9.740299177654746e-05, + "loss": 0.0141, + "step": 14790 + }, + { + "epoch": 2.2988505747126435, + "grad_norm": 0.4067401587963104, + "learning_rate": 9.739772962874867e-05, + "loss": 0.0114, + "step": 14800 + }, + { + "epoch": 2.30040385212799, + "grad_norm": 0.23552657663822174, + "learning_rate": 9.739246229760541e-05, + "loss": 0.0113, + "step": 14810 + }, + { + "epoch": 2.3019571295433363, + "grad_norm": 0.30400750041007996, + "learning_rate": 9.738718978369376e-05, + "loss": 0.0104, + "step": 14820 + }, + { + "epoch": 2.303510406958683, + "grad_norm": 0.29764121770858765, + "learning_rate": 9.738191208759025e-05, + "loss": 0.0174, + "step": 14830 + }, + { + "epoch": 2.305063684374029, + "grad_norm": 0.2732667028903961, + "learning_rate": 9.73766292098721e-05, + "loss": 0.01, + "step": 14840 + }, + { + "epoch": 2.3066169617893757, + "grad_norm": 0.3641069233417511, + "learning_rate": 9.737134115111699e-05, + "loss": 0.0113, + "step": 14850 + }, + { + "epoch": 2.308170239204722, + "grad_norm": 0.2671750783920288, + "learning_rate": 9.736604791190323e-05, + "loss": 0.0114, + "step": 14860 + }, + { + "epoch": 2.3097235166200685, + "grad_norm": 0.3126373887062073, + "learning_rate": 9.73607494928097e-05, + "loss": 0.0136, + "step": 14870 + }, + { + "epoch": 2.3112767940354146, + "grad_norm": 0.4269578456878662, + "learning_rate": 9.735544589441581e-05, + "loss": 0.0133, + "step": 14880 + }, + { + "epoch": 2.3128300714507612, + "grad_norm": 0.35884302854537964, + "learning_rate": 9.735013711730154e-05, + "loss": 0.0127, + "step": 14890 + }, + { + "epoch": 2.3143833488661074, + "grad_norm": 0.23623616993427277, + "learning_rate": 9.734482316204747e-05, + "loss": 0.0133, + "step": 14900 + }, + { + "epoch": 2.315936626281454, + "grad_norm": 0.30775076150894165, + "learning_rate": 9.733950402923473e-05, + "loss": 0.0131, + "step": 14910 + }, + { + "epoch": 2.3174899036968, + "grad_norm": 0.4161844551563263, + "learning_rate": 9.7334179719445e-05, + "loss": 0.0148, + "step": 14920 + }, + { + "epoch": 2.319043181112147, + "grad_norm": 0.23816479742527008, + "learning_rate": 9.732885023326053e-05, + "loss": 0.0151, + "step": 14930 + }, + { + "epoch": 2.320596458527493, + "grad_norm": 0.3343961834907532, + "learning_rate": 9.732351557126418e-05, + "loss": 0.011, + "step": 14940 + }, + { + "epoch": 2.3221497359428396, + "grad_norm": 0.44793370366096497, + "learning_rate": 9.731817573403929e-05, + "loss": 0.0129, + "step": 14950 + }, + { + "epoch": 2.3237030133581857, + "grad_norm": 0.2906535863876343, + "learning_rate": 9.731283072216985e-05, + "loss": 0.011, + "step": 14960 + }, + { + "epoch": 2.325256290773532, + "grad_norm": 0.277584046125412, + "learning_rate": 9.730748053624039e-05, + "loss": 0.0136, + "step": 14970 + }, + { + "epoch": 2.3268095681888785, + "grad_norm": 0.2834952771663666, + "learning_rate": 9.730212517683598e-05, + "loss": 0.0118, + "step": 14980 + }, + { + "epoch": 2.328362845604225, + "grad_norm": 0.3405815064907074, + "learning_rate": 9.729676464454228e-05, + "loss": 0.0111, + "step": 14990 + }, + { + "epoch": 2.3299161230195713, + "grad_norm": 0.45104968547821045, + "learning_rate": 9.72913989399455e-05, + "loss": 0.0131, + "step": 15000 + }, + { + "epoch": 2.3314694004349175, + "grad_norm": 0.1768372356891632, + "learning_rate": 9.728602806363242e-05, + "loss": 0.0123, + "step": 15010 + }, + { + "epoch": 2.333022677850264, + "grad_norm": 0.22963687777519226, + "learning_rate": 9.728065201619043e-05, + "loss": 0.0127, + "step": 15020 + }, + { + "epoch": 2.3345759552656107, + "grad_norm": 0.2346068024635315, + "learning_rate": 9.727527079820742e-05, + "loss": 0.015, + "step": 15030 + }, + { + "epoch": 2.336129232680957, + "grad_norm": 0.22660639882087708, + "learning_rate": 9.726988441027186e-05, + "loss": 0.0129, + "step": 15040 + }, + { + "epoch": 2.337682510096303, + "grad_norm": 0.2547760605812073, + "learning_rate": 9.726449285297281e-05, + "loss": 0.0146, + "step": 15050 + }, + { + "epoch": 2.3392357875116496, + "grad_norm": 0.21363379061222076, + "learning_rate": 9.72590961268999e-05, + "loss": 0.0121, + "step": 15060 + }, + { + "epoch": 2.340789064926996, + "grad_norm": 0.2984892725944519, + "learning_rate": 9.725369423264328e-05, + "loss": 0.0138, + "step": 15070 + }, + { + "epoch": 2.3423423423423424, + "grad_norm": 0.1913643479347229, + "learning_rate": 9.72482871707937e-05, + "loss": 0.0135, + "step": 15080 + }, + { + "epoch": 2.3438956197576886, + "grad_norm": 0.211838960647583, + "learning_rate": 9.724287494194247e-05, + "loss": 0.015, + "step": 15090 + }, + { + "epoch": 2.345448897173035, + "grad_norm": 0.23223671317100525, + "learning_rate": 9.723745754668147e-05, + "loss": 0.013, + "step": 15100 + }, + { + "epoch": 2.3470021745883813, + "grad_norm": 0.5213814973831177, + "learning_rate": 9.723203498560313e-05, + "loss": 0.0125, + "step": 15110 + }, + { + "epoch": 2.348555452003728, + "grad_norm": 0.4562344551086426, + "learning_rate": 9.722660725930046e-05, + "loss": 0.0144, + "step": 15120 + }, + { + "epoch": 2.350108729419074, + "grad_norm": 0.32967546582221985, + "learning_rate": 9.722117436836702e-05, + "loss": 0.0136, + "step": 15130 + }, + { + "epoch": 2.3516620068344207, + "grad_norm": 0.32894057035446167, + "learning_rate": 9.721573631339696e-05, + "loss": 0.0131, + "step": 15140 + }, + { + "epoch": 2.353215284249767, + "grad_norm": 0.27182283997535706, + "learning_rate": 9.721029309498494e-05, + "loss": 0.0119, + "step": 15150 + }, + { + "epoch": 2.3547685616651135, + "grad_norm": 0.2502163052558899, + "learning_rate": 9.720484471372627e-05, + "loss": 0.0122, + "step": 15160 + }, + { + "epoch": 2.3563218390804597, + "grad_norm": 0.18011966347694397, + "learning_rate": 9.719939117021673e-05, + "loss": 0.0113, + "step": 15170 + }, + { + "epoch": 2.3578751164958063, + "grad_norm": 0.1472717672586441, + "learning_rate": 9.719393246505275e-05, + "loss": 0.0119, + "step": 15180 + }, + { + "epoch": 2.3594283939111524, + "grad_norm": 0.2419448047876358, + "learning_rate": 9.718846859883128e-05, + "loss": 0.0131, + "step": 15190 + }, + { + "epoch": 2.360981671326499, + "grad_norm": 0.4918290078639984, + "learning_rate": 9.718299957214982e-05, + "loss": 0.0158, + "step": 15200 + }, + { + "epoch": 2.362534948741845, + "grad_norm": 0.2916393280029297, + "learning_rate": 9.717752538560646e-05, + "loss": 0.0128, + "step": 15210 + }, + { + "epoch": 2.364088226157192, + "grad_norm": 0.18598248064517975, + "learning_rate": 9.717204603979986e-05, + "loss": 0.0136, + "step": 15220 + }, + { + "epoch": 2.365641503572538, + "grad_norm": 0.15871936082839966, + "learning_rate": 9.716656153532922e-05, + "loss": 0.0133, + "step": 15230 + }, + { + "epoch": 2.3671947809878846, + "grad_norm": 0.2667124271392822, + "learning_rate": 9.716107187279434e-05, + "loss": 0.0165, + "step": 15240 + }, + { + "epoch": 2.3687480584032308, + "grad_norm": 0.16835938394069672, + "learning_rate": 9.715557705279555e-05, + "loss": 0.0141, + "step": 15250 + }, + { + "epoch": 2.3703013358185774, + "grad_norm": 0.36407461762428284, + "learning_rate": 9.715007707593372e-05, + "loss": 0.012, + "step": 15260 + }, + { + "epoch": 2.3718546132339235, + "grad_norm": 0.28827401995658875, + "learning_rate": 9.714457194281036e-05, + "loss": 0.0176, + "step": 15270 + }, + { + "epoch": 2.37340789064927, + "grad_norm": 0.34721609950065613, + "learning_rate": 9.713906165402751e-05, + "loss": 0.0143, + "step": 15280 + }, + { + "epoch": 2.3749611680646163, + "grad_norm": 0.2920253872871399, + "learning_rate": 9.713354621018774e-05, + "loss": 0.018, + "step": 15290 + }, + { + "epoch": 2.3765144454799625, + "grad_norm": 0.3537023365497589, + "learning_rate": 9.712802561189422e-05, + "loss": 0.0193, + "step": 15300 + }, + { + "epoch": 2.378067722895309, + "grad_norm": 0.3130330443382263, + "learning_rate": 9.712249985975069e-05, + "loss": 0.0145, + "step": 15310 + }, + { + "epoch": 2.3796210003106557, + "grad_norm": 0.2726975381374359, + "learning_rate": 9.71169689543614e-05, + "loss": 0.014, + "step": 15320 + }, + { + "epoch": 2.381174277726002, + "grad_norm": 0.32283318042755127, + "learning_rate": 9.711143289633123e-05, + "loss": 0.013, + "step": 15330 + }, + { + "epoch": 2.382727555141348, + "grad_norm": 0.34135159850120544, + "learning_rate": 9.710589168626561e-05, + "loss": 0.0121, + "step": 15340 + }, + { + "epoch": 2.3842808325566947, + "grad_norm": 0.36000874638557434, + "learning_rate": 9.710034532477048e-05, + "loss": 0.0121, + "step": 15350 + }, + { + "epoch": 2.385834109972041, + "grad_norm": 0.23756249248981476, + "learning_rate": 9.709479381245239e-05, + "loss": 0.0129, + "step": 15360 + }, + { + "epoch": 2.3873873873873874, + "grad_norm": 0.29960086941719055, + "learning_rate": 9.708923714991847e-05, + "loss": 0.0099, + "step": 15370 + }, + { + "epoch": 2.3889406648027336, + "grad_norm": 0.6723942756652832, + "learning_rate": 9.708367533777638e-05, + "loss": 0.016, + "step": 15380 + }, + { + "epoch": 2.39049394221808, + "grad_norm": 0.2469257116317749, + "learning_rate": 9.707810837663431e-05, + "loss": 0.0164, + "step": 15390 + }, + { + "epoch": 2.3920472196334264, + "grad_norm": 0.16810427606105804, + "learning_rate": 9.707253626710113e-05, + "loss": 0.0116, + "step": 15400 + }, + { + "epoch": 2.393600497048773, + "grad_norm": 0.37786388397216797, + "learning_rate": 9.706695900978613e-05, + "loss": 0.0133, + "step": 15410 + }, + { + "epoch": 2.395153774464119, + "grad_norm": 0.27390119433403015, + "learning_rate": 9.706137660529926e-05, + "loss": 0.0154, + "step": 15420 + }, + { + "epoch": 2.3967070518794658, + "grad_norm": 0.32927387952804565, + "learning_rate": 9.705578905425101e-05, + "loss": 0.0111, + "step": 15430 + }, + { + "epoch": 2.398260329294812, + "grad_norm": 0.3628624677658081, + "learning_rate": 9.705019635725241e-05, + "loss": 0.0107, + "step": 15440 + }, + { + "epoch": 2.3998136067101585, + "grad_norm": 0.47843384742736816, + "learning_rate": 9.704459851491508e-05, + "loss": 0.0155, + "step": 15450 + }, + { + "epoch": 2.4013668841255047, + "grad_norm": 0.15768331289291382, + "learning_rate": 9.703899552785118e-05, + "loss": 0.0096, + "step": 15460 + }, + { + "epoch": 2.4029201615408513, + "grad_norm": 0.3205975890159607, + "learning_rate": 9.703338739667346e-05, + "loss": 0.0113, + "step": 15470 + }, + { + "epoch": 2.4044734389561975, + "grad_norm": 0.14893433451652527, + "learning_rate": 9.70277741219952e-05, + "loss": 0.0113, + "step": 15480 + }, + { + "epoch": 2.406026716371544, + "grad_norm": 0.311143696308136, + "learning_rate": 9.702215570443027e-05, + "loss": 0.0175, + "step": 15490 + }, + { + "epoch": 2.4075799937868902, + "grad_norm": 0.2783229649066925, + "learning_rate": 9.701653214459309e-05, + "loss": 0.0154, + "step": 15500 + }, + { + "epoch": 2.409133271202237, + "grad_norm": 0.22703152894973755, + "learning_rate": 9.701090344309865e-05, + "loss": 0.0121, + "step": 15510 + }, + { + "epoch": 2.410686548617583, + "grad_norm": 0.433180034160614, + "learning_rate": 9.700526960056247e-05, + "loss": 0.013, + "step": 15520 + }, + { + "epoch": 2.4122398260329296, + "grad_norm": 0.21792744100093842, + "learning_rate": 9.699963061760068e-05, + "loss": 0.0126, + "step": 15530 + }, + { + "epoch": 2.413793103448276, + "grad_norm": 0.2216700315475464, + "learning_rate": 9.699398649482997e-05, + "loss": 0.0137, + "step": 15540 + }, + { + "epoch": 2.4153463808636224, + "grad_norm": 0.3800995647907257, + "learning_rate": 9.698833723286753e-05, + "loss": 0.0102, + "step": 15550 + }, + { + "epoch": 2.4168996582789686, + "grad_norm": 0.15031485259532928, + "learning_rate": 9.698268283233118e-05, + "loss": 0.013, + "step": 15560 + }, + { + "epoch": 2.418452935694315, + "grad_norm": 0.45413562655448914, + "learning_rate": 9.697702329383929e-05, + "loss": 0.0096, + "step": 15570 + }, + { + "epoch": 2.4200062131096614, + "grad_norm": 0.3635943830013275, + "learning_rate": 9.697135861801074e-05, + "loss": 0.0137, + "step": 15580 + }, + { + "epoch": 2.4215594905250075, + "grad_norm": 0.20678608119487762, + "learning_rate": 9.696568880546505e-05, + "loss": 0.0127, + "step": 15590 + }, + { + "epoch": 2.423112767940354, + "grad_norm": 0.20693418383598328, + "learning_rate": 9.696001385682223e-05, + "loss": 0.0156, + "step": 15600 + }, + { + "epoch": 2.4246660453557007, + "grad_norm": 0.4167131781578064, + "learning_rate": 9.695433377270291e-05, + "loss": 0.0123, + "step": 15610 + }, + { + "epoch": 2.426219322771047, + "grad_norm": 0.16012750566005707, + "learning_rate": 9.694864855372824e-05, + "loss": 0.011, + "step": 15620 + }, + { + "epoch": 2.427772600186393, + "grad_norm": 0.4748728573322296, + "learning_rate": 9.694295820051995e-05, + "loss": 0.0085, + "step": 15630 + }, + { + "epoch": 2.4293258776017397, + "grad_norm": 0.3614254891872406, + "learning_rate": 9.693726271370032e-05, + "loss": 0.0099, + "step": 15640 + }, + { + "epoch": 2.4308791550170863, + "grad_norm": 0.15810371935367584, + "learning_rate": 9.693156209389221e-05, + "loss": 0.0129, + "step": 15650 + }, + { + "epoch": 2.4324324324324325, + "grad_norm": 0.24884504079818726, + "learning_rate": 9.692585634171905e-05, + "loss": 0.0125, + "step": 15660 + }, + { + "epoch": 2.4339857098477786, + "grad_norm": 0.3929746448993683, + "learning_rate": 9.692014545780476e-05, + "loss": 0.0105, + "step": 15670 + }, + { + "epoch": 2.4355389872631252, + "grad_norm": 0.1757364422082901, + "learning_rate": 9.691442944277393e-05, + "loss": 0.0095, + "step": 15680 + }, + { + "epoch": 2.4370922646784714, + "grad_norm": 0.3209347128868103, + "learning_rate": 9.690870829725162e-05, + "loss": 0.0171, + "step": 15690 + }, + { + "epoch": 2.438645542093818, + "grad_norm": 0.29814669489860535, + "learning_rate": 9.69029820218635e-05, + "loss": 0.0101, + "step": 15700 + }, + { + "epoch": 2.440198819509164, + "grad_norm": 0.28712713718414307, + "learning_rate": 9.689725061723579e-05, + "loss": 0.0113, + "step": 15710 + }, + { + "epoch": 2.441752096924511, + "grad_norm": 0.3818059265613556, + "learning_rate": 9.689151408399527e-05, + "loss": 0.0152, + "step": 15720 + }, + { + "epoch": 2.443305374339857, + "grad_norm": 0.2972812354564667, + "learning_rate": 9.688577242276924e-05, + "loss": 0.0111, + "step": 15730 + }, + { + "epoch": 2.4448586517552036, + "grad_norm": 0.548615574836731, + "learning_rate": 9.688002563418566e-05, + "loss": 0.016, + "step": 15740 + }, + { + "epoch": 2.4464119291705497, + "grad_norm": 0.31793344020843506, + "learning_rate": 9.687427371887293e-05, + "loss": 0.0105, + "step": 15750 + }, + { + "epoch": 2.4479652065858963, + "grad_norm": 0.3160152733325958, + "learning_rate": 9.686851667746012e-05, + "loss": 0.0122, + "step": 15760 + }, + { + "epoch": 2.4495184840012425, + "grad_norm": 0.5344441533088684, + "learning_rate": 9.686275451057677e-05, + "loss": 0.0151, + "step": 15770 + }, + { + "epoch": 2.451071761416589, + "grad_norm": 0.3031679391860962, + "learning_rate": 9.685698721885308e-05, + "loss": 0.0117, + "step": 15780 + }, + { + "epoch": 2.4526250388319353, + "grad_norm": 0.23459912836551666, + "learning_rate": 9.68512148029197e-05, + "loss": 0.0132, + "step": 15790 + }, + { + "epoch": 2.454178316247282, + "grad_norm": 0.43599367141723633, + "learning_rate": 9.684543726340791e-05, + "loss": 0.0134, + "step": 15800 + }, + { + "epoch": 2.455731593662628, + "grad_norm": 0.415313184261322, + "learning_rate": 9.683965460094952e-05, + "loss": 0.015, + "step": 15810 + }, + { + "epoch": 2.4572848710779747, + "grad_norm": 0.3635603189468384, + "learning_rate": 9.683386681617694e-05, + "loss": 0.0131, + "step": 15820 + }, + { + "epoch": 2.458838148493321, + "grad_norm": 0.5075594782829285, + "learning_rate": 9.68280739097231e-05, + "loss": 0.0122, + "step": 15830 + }, + { + "epoch": 2.4603914259086674, + "grad_norm": 0.19502206146717072, + "learning_rate": 9.682227588222148e-05, + "loss": 0.0109, + "step": 15840 + }, + { + "epoch": 2.4619447033240136, + "grad_norm": 0.23813964426517487, + "learning_rate": 9.681647273430618e-05, + "loss": 0.0127, + "step": 15850 + }, + { + "epoch": 2.46349798073936, + "grad_norm": 0.23629991710186005, + "learning_rate": 9.681066446661182e-05, + "loss": 0.0147, + "step": 15860 + }, + { + "epoch": 2.4650512581547064, + "grad_norm": 0.3482632339000702, + "learning_rate": 9.680485107977357e-05, + "loss": 0.0157, + "step": 15870 + }, + { + "epoch": 2.466604535570053, + "grad_norm": 0.41187572479248047, + "learning_rate": 9.679903257442716e-05, + "loss": 0.0114, + "step": 15880 + }, + { + "epoch": 2.468157812985399, + "grad_norm": 0.3281382918357849, + "learning_rate": 9.679320895120891e-05, + "loss": 0.0125, + "step": 15890 + }, + { + "epoch": 2.4697110904007458, + "grad_norm": 0.5302166938781738, + "learning_rate": 9.67873802107557e-05, + "loss": 0.0142, + "step": 15900 + }, + { + "epoch": 2.471264367816092, + "grad_norm": 0.5512802600860596, + "learning_rate": 9.67815463537049e-05, + "loss": 0.0175, + "step": 15910 + }, + { + "epoch": 2.472817645231438, + "grad_norm": 0.24065513908863068, + "learning_rate": 9.677570738069457e-05, + "loss": 0.012, + "step": 15920 + }, + { + "epoch": 2.4743709226467847, + "grad_norm": 0.27866971492767334, + "learning_rate": 9.676986329236318e-05, + "loss": 0.0188, + "step": 15930 + }, + { + "epoch": 2.4759242000621313, + "grad_norm": 0.2867230176925659, + "learning_rate": 9.676401408934987e-05, + "loss": 0.0094, + "step": 15940 + }, + { + "epoch": 2.4774774774774775, + "grad_norm": 0.26930803060531616, + "learning_rate": 9.675815977229428e-05, + "loss": 0.0137, + "step": 15950 + }, + { + "epoch": 2.4790307548928237, + "grad_norm": 0.37579455971717834, + "learning_rate": 9.675230034183664e-05, + "loss": 0.0147, + "step": 15960 + }, + { + "epoch": 2.4805840323081703, + "grad_norm": 0.2776697278022766, + "learning_rate": 9.674643579861773e-05, + "loss": 0.0121, + "step": 15970 + }, + { + "epoch": 2.4821373097235164, + "grad_norm": 0.3094950318336487, + "learning_rate": 9.674056614327886e-05, + "loss": 0.0129, + "step": 15980 + }, + { + "epoch": 2.483690587138863, + "grad_norm": 0.35946527123451233, + "learning_rate": 9.673469137646198e-05, + "loss": 0.0142, + "step": 15990 + }, + { + "epoch": 2.485243864554209, + "grad_norm": 0.5290329456329346, + "learning_rate": 9.67288114988095e-05, + "loss": 0.0138, + "step": 16000 + }, + { + "epoch": 2.486797141969556, + "grad_norm": 0.32905688881874084, + "learning_rate": 9.672292651096447e-05, + "loss": 0.0132, + "step": 16010 + }, + { + "epoch": 2.488350419384902, + "grad_norm": 0.2515275478363037, + "learning_rate": 9.671703641357042e-05, + "loss": 0.0105, + "step": 16020 + }, + { + "epoch": 2.4899036968002486, + "grad_norm": 0.4086555242538452, + "learning_rate": 9.67111412072715e-05, + "loss": 0.0142, + "step": 16030 + }, + { + "epoch": 2.4914569742155948, + "grad_norm": 0.3881515562534332, + "learning_rate": 9.670524089271242e-05, + "loss": 0.0126, + "step": 16040 + }, + { + "epoch": 2.4930102516309414, + "grad_norm": 0.30003416538238525, + "learning_rate": 9.669933547053842e-05, + "loss": 0.0121, + "step": 16050 + }, + { + "epoch": 2.4945635290462875, + "grad_norm": 0.2852894067764282, + "learning_rate": 9.669342494139531e-05, + "loss": 0.0126, + "step": 16060 + }, + { + "epoch": 2.496116806461634, + "grad_norm": 0.19833143055438995, + "learning_rate": 9.668750930592943e-05, + "loss": 0.012, + "step": 16070 + }, + { + "epoch": 2.4976700838769803, + "grad_norm": 0.20153272151947021, + "learning_rate": 9.668158856478775e-05, + "loss": 0.0142, + "step": 16080 + }, + { + "epoch": 2.499223361292327, + "grad_norm": 0.24986353516578674, + "learning_rate": 9.66756627186177e-05, + "loss": 0.0101, + "step": 16090 + }, + { + "epoch": 2.500776638707673, + "grad_norm": 0.30410975217819214, + "learning_rate": 9.666973176806737e-05, + "loss": 0.013, + "step": 16100 + }, + { + "epoch": 2.5023299161230197, + "grad_norm": 0.2702200710773468, + "learning_rate": 9.666379571378534e-05, + "loss": 0.0117, + "step": 16110 + }, + { + "epoch": 2.503883193538366, + "grad_norm": 0.24804244935512543, + "learning_rate": 9.665785455642076e-05, + "loss": 0.0111, + "step": 16120 + }, + { + "epoch": 2.5054364709537125, + "grad_norm": 0.47765350341796875, + "learning_rate": 9.665190829662337e-05, + "loss": 0.013, + "step": 16130 + }, + { + "epoch": 2.5069897483690586, + "grad_norm": 0.7213842868804932, + "learning_rate": 9.664595693504342e-05, + "loss": 0.0166, + "step": 16140 + }, + { + "epoch": 2.5085430257844052, + "grad_norm": 0.2859228253364563, + "learning_rate": 9.664000047233175e-05, + "loss": 0.0124, + "step": 16150 + }, + { + "epoch": 2.5100963031997514, + "grad_norm": 0.4866156578063965, + "learning_rate": 9.663403890913976e-05, + "loss": 0.0158, + "step": 16160 + }, + { + "epoch": 2.511649580615098, + "grad_norm": 0.569245457649231, + "learning_rate": 9.662807224611938e-05, + "loss": 0.0158, + "step": 16170 + }, + { + "epoch": 2.513202858030444, + "grad_norm": 0.3881169557571411, + "learning_rate": 9.662210048392311e-05, + "loss": 0.0128, + "step": 16180 + }, + { + "epoch": 2.514756135445791, + "grad_norm": 0.4377520680427551, + "learning_rate": 9.661612362320405e-05, + "loss": 0.0139, + "step": 16190 + }, + { + "epoch": 2.516309412861137, + "grad_norm": 0.27828195691108704, + "learning_rate": 9.661014166461579e-05, + "loss": 0.0137, + "step": 16200 + }, + { + "epoch": 2.517862690276483, + "grad_norm": 0.3415865898132324, + "learning_rate": 9.66041546088125e-05, + "loss": 0.0109, + "step": 16210 + }, + { + "epoch": 2.5194159676918297, + "grad_norm": 0.27297666668891907, + "learning_rate": 9.659816245644895e-05, + "loss": 0.0119, + "step": 16220 + }, + { + "epoch": 2.5209692451071763, + "grad_norm": 0.3088768720626831, + "learning_rate": 9.65921652081804e-05, + "loss": 0.0094, + "step": 16230 + }, + { + "epoch": 2.5225225225225225, + "grad_norm": 0.24727098643779755, + "learning_rate": 9.658616286466271e-05, + "loss": 0.0084, + "step": 16240 + }, + { + "epoch": 2.5240757999378687, + "grad_norm": 0.19642116129398346, + "learning_rate": 9.65801554265523e-05, + "loss": 0.0101, + "step": 16250 + }, + { + "epoch": 2.5256290773532153, + "grad_norm": 0.37657681107521057, + "learning_rate": 9.657414289450612e-05, + "loss": 0.0148, + "step": 16260 + }, + { + "epoch": 2.527182354768562, + "grad_norm": 0.26968368887901306, + "learning_rate": 9.656812526918171e-05, + "loss": 0.0107, + "step": 16270 + }, + { + "epoch": 2.528735632183908, + "grad_norm": 0.2681080102920532, + "learning_rate": 9.656210255123712e-05, + "loss": 0.0103, + "step": 16280 + }, + { + "epoch": 2.5302889095992542, + "grad_norm": 0.5129451155662537, + "learning_rate": 9.6556074741331e-05, + "loss": 0.0116, + "step": 16290 + }, + { + "epoch": 2.531842187014601, + "grad_norm": 0.2825247645378113, + "learning_rate": 9.655004184012256e-05, + "loss": 0.0119, + "step": 16300 + }, + { + "epoch": 2.5333954644299475, + "grad_norm": 0.31687480211257935, + "learning_rate": 9.654400384827152e-05, + "loss": 0.0154, + "step": 16310 + }, + { + "epoch": 2.5349487418452936, + "grad_norm": 0.3325687348842621, + "learning_rate": 9.653796076643818e-05, + "loss": 0.0107, + "step": 16320 + }, + { + "epoch": 2.53650201926064, + "grad_norm": 0.27310851216316223, + "learning_rate": 9.653191259528344e-05, + "loss": 0.0142, + "step": 16330 + }, + { + "epoch": 2.5380552966759864, + "grad_norm": 0.2852237820625305, + "learning_rate": 9.65258593354687e-05, + "loss": 0.0112, + "step": 16340 + }, + { + "epoch": 2.5396085740913326, + "grad_norm": 0.23947498202323914, + "learning_rate": 9.651980098765591e-05, + "loss": 0.0104, + "step": 16350 + }, + { + "epoch": 2.541161851506679, + "grad_norm": 0.2502205967903137, + "learning_rate": 9.651373755250765e-05, + "loss": 0.0152, + "step": 16360 + }, + { + "epoch": 2.5427151289220253, + "grad_norm": 0.5061002373695374, + "learning_rate": 9.650766903068697e-05, + "loss": 0.0114, + "step": 16370 + }, + { + "epoch": 2.544268406337372, + "grad_norm": 0.31435301899909973, + "learning_rate": 9.650159542285753e-05, + "loss": 0.0119, + "step": 16380 + }, + { + "epoch": 2.545821683752718, + "grad_norm": 0.23674017190933228, + "learning_rate": 9.649551672968353e-05, + "loss": 0.0121, + "step": 16390 + }, + { + "epoch": 2.5473749611680647, + "grad_norm": 0.21190522611141205, + "learning_rate": 9.648943295182973e-05, + "loss": 0.0117, + "step": 16400 + }, + { + "epoch": 2.548928238583411, + "grad_norm": 0.3097875416278839, + "learning_rate": 9.648334408996144e-05, + "loss": 0.0144, + "step": 16410 + }, + { + "epoch": 2.5504815159987575, + "grad_norm": 0.36286672949790955, + "learning_rate": 9.647725014474452e-05, + "loss": 0.0137, + "step": 16420 + }, + { + "epoch": 2.5520347934141037, + "grad_norm": 0.21566888689994812, + "learning_rate": 9.64711511168454e-05, + "loss": 0.0087, + "step": 16430 + }, + { + "epoch": 2.5535880708294503, + "grad_norm": 0.3184620141983032, + "learning_rate": 9.646504700693108e-05, + "loss": 0.0123, + "step": 16440 + }, + { + "epoch": 2.5551413482447964, + "grad_norm": 0.22329457104206085, + "learning_rate": 9.645893781566907e-05, + "loss": 0.0098, + "step": 16450 + }, + { + "epoch": 2.556694625660143, + "grad_norm": 0.2976702153682709, + "learning_rate": 9.645282354372744e-05, + "loss": 0.0111, + "step": 16460 + }, + { + "epoch": 2.558247903075489, + "grad_norm": 0.3536278009414673, + "learning_rate": 9.644670419177491e-05, + "loss": 0.0114, + "step": 16470 + }, + { + "epoch": 2.559801180490836, + "grad_norm": 0.4325779676437378, + "learning_rate": 9.644057976048062e-05, + "loss": 0.0124, + "step": 16480 + }, + { + "epoch": 2.561354457906182, + "grad_norm": 0.22731275856494904, + "learning_rate": 9.643445025051435e-05, + "loss": 0.0137, + "step": 16490 + }, + { + "epoch": 2.562907735321528, + "grad_norm": 0.24685965478420258, + "learning_rate": 9.642831566254641e-05, + "loss": 0.0124, + "step": 16500 + }, + { + "epoch": 2.5644610127368748, + "grad_norm": 0.2714805603027344, + "learning_rate": 9.642217599724769e-05, + "loss": 0.0122, + "step": 16510 + }, + { + "epoch": 2.5660142901522214, + "grad_norm": 0.22981297969818115, + "learning_rate": 9.64160312552896e-05, + "loss": 0.0115, + "step": 16520 + }, + { + "epoch": 2.5675675675675675, + "grad_norm": 0.24046199023723602, + "learning_rate": 9.64098814373441e-05, + "loss": 0.0147, + "step": 16530 + }, + { + "epoch": 2.5691208449829137, + "grad_norm": 0.22848282754421234, + "learning_rate": 9.640372654408374e-05, + "loss": 0.0111, + "step": 16540 + }, + { + "epoch": 2.5706741223982603, + "grad_norm": 0.3916492462158203, + "learning_rate": 9.639756657618162e-05, + "loss": 0.0117, + "step": 16550 + }, + { + "epoch": 2.572227399813607, + "grad_norm": 0.19405686855316162, + "learning_rate": 9.639140153431138e-05, + "loss": 0.0132, + "step": 16560 + }, + { + "epoch": 2.573780677228953, + "grad_norm": 0.3235650658607483, + "learning_rate": 9.638523141914721e-05, + "loss": 0.0099, + "step": 16570 + }, + { + "epoch": 2.5753339546442993, + "grad_norm": 0.27469775080680847, + "learning_rate": 9.637905623136388e-05, + "loss": 0.016, + "step": 16580 + }, + { + "epoch": 2.576887232059646, + "grad_norm": 0.4833729565143585, + "learning_rate": 9.637287597163669e-05, + "loss": 0.0098, + "step": 16590 + }, + { + "epoch": 2.5784405094749925, + "grad_norm": 0.33667731285095215, + "learning_rate": 9.63666906406415e-05, + "loss": 0.016, + "step": 16600 + }, + { + "epoch": 2.5799937868903386, + "grad_norm": 0.20911747217178345, + "learning_rate": 9.636050023905473e-05, + "loss": 0.0143, + "step": 16610 + }, + { + "epoch": 2.581547064305685, + "grad_norm": 0.36941492557525635, + "learning_rate": 9.635430476755336e-05, + "loss": 0.0132, + "step": 16620 + }, + { + "epoch": 2.5831003417210314, + "grad_norm": 0.3245101571083069, + "learning_rate": 9.63481042268149e-05, + "loss": 0.0148, + "step": 16630 + }, + { + "epoch": 2.584653619136378, + "grad_norm": 0.48457208275794983, + "learning_rate": 9.634189861751745e-05, + "loss": 0.0155, + "step": 16640 + }, + { + "epoch": 2.586206896551724, + "grad_norm": 0.37330201268196106, + "learning_rate": 9.633568794033967e-05, + "loss": 0.0156, + "step": 16650 + }, + { + "epoch": 2.5877601739670704, + "grad_norm": 0.2811684012413025, + "learning_rate": 9.63294721959607e-05, + "loss": 0.0142, + "step": 16660 + }, + { + "epoch": 2.589313451382417, + "grad_norm": 0.3234114348888397, + "learning_rate": 9.63232513850603e-05, + "loss": 0.0119, + "step": 16670 + }, + { + "epoch": 2.590866728797763, + "grad_norm": 0.3291720747947693, + "learning_rate": 9.631702550831878e-05, + "loss": 0.0137, + "step": 16680 + }, + { + "epoch": 2.5924200062131098, + "grad_norm": 0.1873646229505539, + "learning_rate": 9.631079456641698e-05, + "loss": 0.0135, + "step": 16690 + }, + { + "epoch": 2.593973283628456, + "grad_norm": 0.7227073907852173, + "learning_rate": 9.630455856003632e-05, + "loss": 0.0135, + "step": 16700 + }, + { + "epoch": 2.5955265610438025, + "grad_norm": 0.21923208236694336, + "learning_rate": 9.629831748985876e-05, + "loss": 0.0131, + "step": 16710 + }, + { + "epoch": 2.5970798384591487, + "grad_norm": 0.2388484627008438, + "learning_rate": 9.629207135656679e-05, + "loss": 0.014, + "step": 16720 + }, + { + "epoch": 2.5986331158744953, + "grad_norm": 0.3268207311630249, + "learning_rate": 9.628582016084353e-05, + "loss": 0.0087, + "step": 16730 + }, + { + "epoch": 2.6001863932898415, + "grad_norm": 0.4959701597690582, + "learning_rate": 9.627956390337254e-05, + "loss": 0.0118, + "step": 16740 + }, + { + "epoch": 2.601739670705188, + "grad_norm": 0.27025556564331055, + "learning_rate": 9.627330258483802e-05, + "loss": 0.0122, + "step": 16750 + }, + { + "epoch": 2.6032929481205342, + "grad_norm": 0.18521520495414734, + "learning_rate": 9.62670362059247e-05, + "loss": 0.012, + "step": 16760 + }, + { + "epoch": 2.604846225535881, + "grad_norm": 0.302435964345932, + "learning_rate": 9.626076476731786e-05, + "loss": 0.0145, + "step": 16770 + }, + { + "epoch": 2.606399502951227, + "grad_norm": 0.34895044565200806, + "learning_rate": 9.625448826970336e-05, + "loss": 0.0123, + "step": 16780 + }, + { + "epoch": 2.6079527803665736, + "grad_norm": 0.10295545309782028, + "learning_rate": 9.624820671376755e-05, + "loss": 0.0126, + "step": 16790 + }, + { + "epoch": 2.60950605778192, + "grad_norm": 0.28426700830459595, + "learning_rate": 9.62419201001974e-05, + "loss": 0.016, + "step": 16800 + }, + { + "epoch": 2.6110593351972664, + "grad_norm": 0.22943837940692902, + "learning_rate": 9.623562842968037e-05, + "loss": 0.0193, + "step": 16810 + }, + { + "epoch": 2.6126126126126126, + "grad_norm": 0.2045169472694397, + "learning_rate": 9.622933170290454e-05, + "loss": 0.0149, + "step": 16820 + }, + { + "epoch": 2.6141658900279587, + "grad_norm": 0.393023818731308, + "learning_rate": 9.622302992055849e-05, + "loss": 0.0154, + "step": 16830 + }, + { + "epoch": 2.6157191674433053, + "grad_norm": 0.38193702697753906, + "learning_rate": 9.62167230833314e-05, + "loss": 0.0144, + "step": 16840 + }, + { + "epoch": 2.617272444858652, + "grad_norm": 0.20674818754196167, + "learning_rate": 9.621041119191295e-05, + "loss": 0.0128, + "step": 16850 + }, + { + "epoch": 2.618825722273998, + "grad_norm": 0.2866438031196594, + "learning_rate": 9.620409424699342e-05, + "loss": 0.0118, + "step": 16860 + }, + { + "epoch": 2.6203789996893443, + "grad_norm": 0.36442217230796814, + "learning_rate": 9.619777224926359e-05, + "loss": 0.0139, + "step": 16870 + }, + { + "epoch": 2.621932277104691, + "grad_norm": 0.37413090467453003, + "learning_rate": 9.619144519941485e-05, + "loss": 0.0179, + "step": 16880 + }, + { + "epoch": 2.6234855545200375, + "grad_norm": 0.43354976177215576, + "learning_rate": 9.618511309813912e-05, + "loss": 0.014, + "step": 16890 + }, + { + "epoch": 2.6250388319353837, + "grad_norm": 0.6840096712112427, + "learning_rate": 9.617877594612886e-05, + "loss": 0.0162, + "step": 16900 + }, + { + "epoch": 2.62659210935073, + "grad_norm": 0.4480349123477936, + "learning_rate": 9.617243374407707e-05, + "loss": 0.0171, + "step": 16910 + }, + { + "epoch": 2.6281453867660765, + "grad_norm": 0.3141748011112213, + "learning_rate": 9.616608649267736e-05, + "loss": 0.0155, + "step": 16920 + }, + { + "epoch": 2.629698664181423, + "grad_norm": 0.20386669039726257, + "learning_rate": 9.615973419262385e-05, + "loss": 0.0122, + "step": 16930 + }, + { + "epoch": 2.6312519415967692, + "grad_norm": 0.23242461681365967, + "learning_rate": 9.615337684461119e-05, + "loss": 0.0166, + "step": 16940 + }, + { + "epoch": 2.6328052190121154, + "grad_norm": 0.3753695487976074, + "learning_rate": 9.614701444933465e-05, + "loss": 0.0111, + "step": 16950 + }, + { + "epoch": 2.634358496427462, + "grad_norm": 0.69097900390625, + "learning_rate": 9.614064700748997e-05, + "loss": 0.0135, + "step": 16960 + }, + { + "epoch": 2.635911773842808, + "grad_norm": 0.2438003271818161, + "learning_rate": 9.613427451977352e-05, + "loss": 0.0137, + "step": 16970 + }, + { + "epoch": 2.637465051258155, + "grad_norm": 0.183570995926857, + "learning_rate": 9.612789698688216e-05, + "loss": 0.0111, + "step": 16980 + }, + { + "epoch": 2.639018328673501, + "grad_norm": 0.23791739344596863, + "learning_rate": 9.612151440951334e-05, + "loss": 0.0155, + "step": 16990 + }, + { + "epoch": 2.6405716060888476, + "grad_norm": 0.30592232942581177, + "learning_rate": 9.611512678836506e-05, + "loss": 0.013, + "step": 17000 + }, + { + "epoch": 2.6421248835041937, + "grad_norm": 0.21742235124111176, + "learning_rate": 9.610873412413584e-05, + "loss": 0.0113, + "step": 17010 + }, + { + "epoch": 2.6436781609195403, + "grad_norm": 0.2161146104335785, + "learning_rate": 9.610233641752476e-05, + "loss": 0.015, + "step": 17020 + }, + { + "epoch": 2.6452314383348865, + "grad_norm": 0.26678234338760376, + "learning_rate": 9.609593366923151e-05, + "loss": 0.01, + "step": 17030 + }, + { + "epoch": 2.646784715750233, + "grad_norm": 0.2648029923439026, + "learning_rate": 9.608952587995625e-05, + "loss": 0.0104, + "step": 17040 + }, + { + "epoch": 2.6483379931655793, + "grad_norm": 0.298827588558197, + "learning_rate": 9.608311305039972e-05, + "loss": 0.0127, + "step": 17050 + }, + { + "epoch": 2.649891270580926, + "grad_norm": 0.26711976528167725, + "learning_rate": 9.607669518126326e-05, + "loss": 0.01, + "step": 17060 + }, + { + "epoch": 2.651444547996272, + "grad_norm": 0.3923009932041168, + "learning_rate": 9.607027227324866e-05, + "loss": 0.0108, + "step": 17070 + }, + { + "epoch": 2.6529978254116187, + "grad_norm": 0.30316683650016785, + "learning_rate": 9.606384432705837e-05, + "loss": 0.0148, + "step": 17080 + }, + { + "epoch": 2.654551102826965, + "grad_norm": 0.25990837812423706, + "learning_rate": 9.60574113433953e-05, + "loss": 0.0101, + "step": 17090 + }, + { + "epoch": 2.6561043802423114, + "grad_norm": 0.23450316488742828, + "learning_rate": 9.6050973322963e-05, + "loss": 0.0148, + "step": 17100 + }, + { + "epoch": 2.6576576576576576, + "grad_norm": 0.3410149812698364, + "learning_rate": 9.604453026646547e-05, + "loss": 0.0115, + "step": 17110 + }, + { + "epoch": 2.6592109350730038, + "grad_norm": 0.36587846279144287, + "learning_rate": 9.603808217460735e-05, + "loss": 0.013, + "step": 17120 + }, + { + "epoch": 2.6607642124883504, + "grad_norm": 0.39739561080932617, + "learning_rate": 9.603162904809377e-05, + "loss": 0.0145, + "step": 17130 + }, + { + "epoch": 2.662317489903697, + "grad_norm": 0.3332538306713104, + "learning_rate": 9.602517088763045e-05, + "loss": 0.015, + "step": 17140 + }, + { + "epoch": 2.663870767319043, + "grad_norm": 0.3687618374824524, + "learning_rate": 9.601870769392365e-05, + "loss": 0.0094, + "step": 17150 + }, + { + "epoch": 2.6654240447343893, + "grad_norm": 0.15086185932159424, + "learning_rate": 9.601223946768017e-05, + "loss": 0.0127, + "step": 17160 + }, + { + "epoch": 2.666977322149736, + "grad_norm": 0.2799261808395386, + "learning_rate": 9.600576620960734e-05, + "loss": 0.0128, + "step": 17170 + }, + { + "epoch": 2.6685305995650825, + "grad_norm": 0.2839941680431366, + "learning_rate": 9.599928792041308e-05, + "loss": 0.0126, + "step": 17180 + }, + { + "epoch": 2.6700838769804287, + "grad_norm": 0.23261983692646027, + "learning_rate": 9.599280460080587e-05, + "loss": 0.0119, + "step": 17190 + }, + { + "epoch": 2.671637154395775, + "grad_norm": 0.3179191052913666, + "learning_rate": 9.59863162514947e-05, + "loss": 0.0112, + "step": 17200 + }, + { + "epoch": 2.6731904318111215, + "grad_norm": 0.24609677493572235, + "learning_rate": 9.597982287318911e-05, + "loss": 0.012, + "step": 17210 + }, + { + "epoch": 2.674743709226468, + "grad_norm": 0.2527526319026947, + "learning_rate": 9.597332446659923e-05, + "loss": 0.0105, + "step": 17220 + }, + { + "epoch": 2.6762969866418143, + "grad_norm": 0.3544883728027344, + "learning_rate": 9.59668210324357e-05, + "loss": 0.0102, + "step": 17230 + }, + { + "epoch": 2.6778502640571604, + "grad_norm": 0.2488439977169037, + "learning_rate": 9.596031257140974e-05, + "loss": 0.0112, + "step": 17240 + }, + { + "epoch": 2.679403541472507, + "grad_norm": 0.3232794404029846, + "learning_rate": 9.59537990842331e-05, + "loss": 0.0155, + "step": 17250 + }, + { + "epoch": 2.6809568188878536, + "grad_norm": 0.6495294570922852, + "learning_rate": 9.594728057161806e-05, + "loss": 0.0149, + "step": 17260 + }, + { + "epoch": 2.6825100963032, + "grad_norm": 0.2568696141242981, + "learning_rate": 9.594075703427752e-05, + "loss": 0.0118, + "step": 17270 + }, + { + "epoch": 2.684063373718546, + "grad_norm": 0.39379754662513733, + "learning_rate": 9.593422847292486e-05, + "loss": 0.0171, + "step": 17280 + }, + { + "epoch": 2.6856166511338926, + "grad_norm": 0.44010332226753235, + "learning_rate": 9.592769488827402e-05, + "loss": 0.0108, + "step": 17290 + }, + { + "epoch": 2.6871699285492388, + "grad_norm": 0.4911414086818695, + "learning_rate": 9.592115628103952e-05, + "loss": 0.0124, + "step": 17300 + }, + { + "epoch": 2.6887232059645854, + "grad_norm": 0.3760782778263092, + "learning_rate": 9.591461265193643e-05, + "loss": 0.0108, + "step": 17310 + }, + { + "epoch": 2.6902764833799315, + "grad_norm": 0.3584495484828949, + "learning_rate": 9.590806400168032e-05, + "loss": 0.0115, + "step": 17320 + }, + { + "epoch": 2.691829760795278, + "grad_norm": 0.3758566677570343, + "learning_rate": 9.590151033098735e-05, + "loss": 0.0109, + "step": 17330 + }, + { + "epoch": 2.6933830382106243, + "grad_norm": 0.3328604996204376, + "learning_rate": 9.589495164057423e-05, + "loss": 0.0116, + "step": 17340 + }, + { + "epoch": 2.694936315625971, + "grad_norm": 0.2381529062986374, + "learning_rate": 9.58883879311582e-05, + "loss": 0.0116, + "step": 17350 + }, + { + "epoch": 2.696489593041317, + "grad_norm": 0.26207271218299866, + "learning_rate": 9.588181920345705e-05, + "loss": 0.0133, + "step": 17360 + }, + { + "epoch": 2.6980428704566637, + "grad_norm": 0.27246108651161194, + "learning_rate": 9.587524545818913e-05, + "loss": 0.0116, + "step": 17370 + }, + { + "epoch": 2.69959614787201, + "grad_norm": 0.3336132764816284, + "learning_rate": 9.586866669607335e-05, + "loss": 0.0127, + "step": 17380 + }, + { + "epoch": 2.7011494252873565, + "grad_norm": 0.45943358540534973, + "learning_rate": 9.586208291782915e-05, + "loss": 0.0113, + "step": 17390 + }, + { + "epoch": 2.7027027027027026, + "grad_norm": 0.26499539613723755, + "learning_rate": 9.58554941241765e-05, + "loss": 0.0133, + "step": 17400 + }, + { + "epoch": 2.7042559801180492, + "grad_norm": 0.27061569690704346, + "learning_rate": 9.584890031583596e-05, + "loss": 0.0112, + "step": 17410 + }, + { + "epoch": 2.7058092575333954, + "grad_norm": 0.5797892212867737, + "learning_rate": 9.584230149352861e-05, + "loss": 0.0126, + "step": 17420 + }, + { + "epoch": 2.707362534948742, + "grad_norm": 0.2391386479139328, + "learning_rate": 9.58356976579761e-05, + "loss": 0.0189, + "step": 17430 + }, + { + "epoch": 2.708915812364088, + "grad_norm": 0.24562036991119385, + "learning_rate": 9.58290888099006e-05, + "loss": 0.0107, + "step": 17440 + }, + { + "epoch": 2.7104690897794343, + "grad_norm": 0.2500697374343872, + "learning_rate": 9.582247495002486e-05, + "loss": 0.0115, + "step": 17450 + }, + { + "epoch": 2.712022367194781, + "grad_norm": 0.38843947649002075, + "learning_rate": 9.581585607907214e-05, + "loss": 0.011, + "step": 17460 + }, + { + "epoch": 2.7135756446101276, + "grad_norm": 0.6079829335212708, + "learning_rate": 9.580923219776628e-05, + "loss": 0.0116, + "step": 17470 + }, + { + "epoch": 2.7151289220254737, + "grad_norm": 0.27849674224853516, + "learning_rate": 9.580260330683167e-05, + "loss": 0.0157, + "step": 17480 + }, + { + "epoch": 2.71668219944082, + "grad_norm": 0.296875923871994, + "learning_rate": 9.579596940699322e-05, + "loss": 0.0128, + "step": 17490 + }, + { + "epoch": 2.7182354768561665, + "grad_norm": 0.3370267152786255, + "learning_rate": 9.578933049897643e-05, + "loss": 0.0137, + "step": 17500 + }, + { + "epoch": 2.719788754271513, + "grad_norm": 0.25123339891433716, + "learning_rate": 9.578268658350728e-05, + "loss": 0.014, + "step": 17510 + }, + { + "epoch": 2.7213420316868593, + "grad_norm": 0.2360130399465561, + "learning_rate": 9.577603766131235e-05, + "loss": 0.0111, + "step": 17520 + }, + { + "epoch": 2.7228953091022055, + "grad_norm": 0.28583383560180664, + "learning_rate": 9.576938373311878e-05, + "loss": 0.0139, + "step": 17530 + }, + { + "epoch": 2.724448586517552, + "grad_norm": 0.3385310173034668, + "learning_rate": 9.576272479965421e-05, + "loss": 0.0157, + "step": 17540 + }, + { + "epoch": 2.7260018639328987, + "grad_norm": 0.2581496238708496, + "learning_rate": 9.575606086164687e-05, + "loss": 0.0119, + "step": 17550 + }, + { + "epoch": 2.727555141348245, + "grad_norm": 0.4972870945930481, + "learning_rate": 9.57493919198255e-05, + "loss": 0.0126, + "step": 17560 + }, + { + "epoch": 2.729108418763591, + "grad_norm": 0.4363095164299011, + "learning_rate": 9.57427179749194e-05, + "loss": 0.0116, + "step": 17570 + }, + { + "epoch": 2.7306616961789376, + "grad_norm": 0.4796726107597351, + "learning_rate": 9.573603902765846e-05, + "loss": 0.0105, + "step": 17580 + }, + { + "epoch": 2.732214973594284, + "grad_norm": 0.2900858223438263, + "learning_rate": 9.572935507877304e-05, + "loss": 0.0173, + "step": 17590 + }, + { + "epoch": 2.7337682510096304, + "grad_norm": 0.2036392092704773, + "learning_rate": 9.57226661289941e-05, + "loss": 0.0114, + "step": 17600 + }, + { + "epoch": 2.7353215284249766, + "grad_norm": 0.36201202869415283, + "learning_rate": 9.571597217905315e-05, + "loss": 0.0125, + "step": 17610 + }, + { + "epoch": 2.736874805840323, + "grad_norm": 0.3062945306301117, + "learning_rate": 9.57092732296822e-05, + "loss": 0.0108, + "step": 17620 + }, + { + "epoch": 2.7384280832556693, + "grad_norm": 0.37316662073135376, + "learning_rate": 9.570256928161385e-05, + "loss": 0.0128, + "step": 17630 + }, + { + "epoch": 2.739981360671016, + "grad_norm": 0.5169488787651062, + "learning_rate": 9.569586033558126e-05, + "loss": 0.0112, + "step": 17640 + }, + { + "epoch": 2.741534638086362, + "grad_norm": 0.33902299404144287, + "learning_rate": 9.568914639231807e-05, + "loss": 0.0113, + "step": 17650 + }, + { + "epoch": 2.7430879155017087, + "grad_norm": 0.4483776390552521, + "learning_rate": 9.568242745255852e-05, + "loss": 0.0133, + "step": 17660 + }, + { + "epoch": 2.744641192917055, + "grad_norm": 0.31292030215263367, + "learning_rate": 9.567570351703739e-05, + "loss": 0.0167, + "step": 17670 + }, + { + "epoch": 2.7461944703324015, + "grad_norm": 0.23766650259494781, + "learning_rate": 9.566897458649001e-05, + "loss": 0.011, + "step": 17680 + }, + { + "epoch": 2.7477477477477477, + "grad_norm": 0.24850428104400635, + "learning_rate": 9.566224066165221e-05, + "loss": 0.0116, + "step": 17690 + }, + { + "epoch": 2.7493010251630943, + "grad_norm": 0.22649802267551422, + "learning_rate": 9.565550174326043e-05, + "loss": 0.0136, + "step": 17700 + }, + { + "epoch": 2.7508543025784404, + "grad_norm": 0.18827788531780243, + "learning_rate": 9.564875783205162e-05, + "loss": 0.0145, + "step": 17710 + }, + { + "epoch": 2.752407579993787, + "grad_norm": 0.17951886355876923, + "learning_rate": 9.564200892876328e-05, + "loss": 0.0105, + "step": 17720 + }, + { + "epoch": 2.753960857409133, + "grad_norm": 0.31421610713005066, + "learning_rate": 9.563525503413348e-05, + "loss": 0.014, + "step": 17730 + }, + { + "epoch": 2.7555141348244794, + "grad_norm": 0.4491939842700958, + "learning_rate": 9.562849614890079e-05, + "loss": 0.016, + "step": 17740 + }, + { + "epoch": 2.757067412239826, + "grad_norm": 0.24065694212913513, + "learning_rate": 9.562173227380436e-05, + "loss": 0.0138, + "step": 17750 + }, + { + "epoch": 2.7586206896551726, + "grad_norm": 0.3920459747314453, + "learning_rate": 9.561496340958389e-05, + "loss": 0.0126, + "step": 17760 + }, + { + "epoch": 2.7601739670705188, + "grad_norm": 0.29658040404319763, + "learning_rate": 9.560818955697959e-05, + "loss": 0.0141, + "step": 17770 + }, + { + "epoch": 2.761727244485865, + "grad_norm": 0.20982946455478668, + "learning_rate": 9.560141071673228e-05, + "loss": 0.0103, + "step": 17780 + }, + { + "epoch": 2.7632805219012115, + "grad_norm": 0.21938556432724, + "learning_rate": 9.559462688958323e-05, + "loss": 0.012, + "step": 17790 + }, + { + "epoch": 2.764833799316558, + "grad_norm": 0.35098856687545776, + "learning_rate": 9.558783807627434e-05, + "loss": 0.0145, + "step": 17800 + }, + { + "epoch": 2.7663870767319043, + "grad_norm": 0.3692220449447632, + "learning_rate": 9.558104427754801e-05, + "loss": 0.0109, + "step": 17810 + }, + { + "epoch": 2.7679403541472505, + "grad_norm": 0.25227242708206177, + "learning_rate": 9.557424549414722e-05, + "loss": 0.009, + "step": 17820 + }, + { + "epoch": 2.769493631562597, + "grad_norm": 0.45073503255844116, + "learning_rate": 9.556744172681546e-05, + "loss": 0.0131, + "step": 17830 + }, + { + "epoch": 2.7710469089779437, + "grad_norm": 0.23533488810062408, + "learning_rate": 9.556063297629677e-05, + "loss": 0.0124, + "step": 17840 + }, + { + "epoch": 2.77260018639329, + "grad_norm": 0.3996090888977051, + "learning_rate": 9.555381924333578e-05, + "loss": 0.013, + "step": 17850 + }, + { + "epoch": 2.774153463808636, + "grad_norm": 0.18904824554920197, + "learning_rate": 9.554700052867758e-05, + "loss": 0.0096, + "step": 17860 + }, + { + "epoch": 2.7757067412239826, + "grad_norm": 0.6416215896606445, + "learning_rate": 9.554017683306789e-05, + "loss": 0.0124, + "step": 17870 + }, + { + "epoch": 2.7772600186393293, + "grad_norm": 0.2994936406612396, + "learning_rate": 9.553334815725294e-05, + "loss": 0.0121, + "step": 17880 + }, + { + "epoch": 2.7788132960546754, + "grad_norm": 0.42242249846458435, + "learning_rate": 9.552651450197949e-05, + "loss": 0.0121, + "step": 17890 + }, + { + "epoch": 2.7803665734700216, + "grad_norm": 0.20810051262378693, + "learning_rate": 9.551967586799486e-05, + "loss": 0.0124, + "step": 17900 + }, + { + "epoch": 2.781919850885368, + "grad_norm": 0.2735881507396698, + "learning_rate": 9.551283225604692e-05, + "loss": 0.0136, + "step": 17910 + }, + { + "epoch": 2.7834731283007144, + "grad_norm": 0.37609750032424927, + "learning_rate": 9.550598366688406e-05, + "loss": 0.0166, + "step": 17920 + }, + { + "epoch": 2.785026405716061, + "grad_norm": 0.3570910394191742, + "learning_rate": 9.549913010125526e-05, + "loss": 0.0127, + "step": 17930 + }, + { + "epoch": 2.786579683131407, + "grad_norm": 0.2493283450603485, + "learning_rate": 9.549227155990999e-05, + "loss": 0.0131, + "step": 17940 + }, + { + "epoch": 2.7881329605467537, + "grad_norm": 0.1853734701871872, + "learning_rate": 9.548540804359828e-05, + "loss": 0.0108, + "step": 17950 + }, + { + "epoch": 2.7896862379621, + "grad_norm": 0.29866278171539307, + "learning_rate": 9.547853955307077e-05, + "loss": 0.0154, + "step": 17960 + }, + { + "epoch": 2.7912395153774465, + "grad_norm": 0.3705228269100189, + "learning_rate": 9.547166608907853e-05, + "loss": 0.0143, + "step": 17970 + }, + { + "epoch": 2.7927927927927927, + "grad_norm": 0.24370421469211578, + "learning_rate": 9.546478765237326e-05, + "loss": 0.0114, + "step": 17980 + }, + { + "epoch": 2.7943460702081393, + "grad_norm": 0.30981069803237915, + "learning_rate": 9.545790424370715e-05, + "loss": 0.0171, + "step": 17990 + }, + { + "epoch": 2.7958993476234855, + "grad_norm": 0.29411736130714417, + "learning_rate": 9.5451015863833e-05, + "loss": 0.0138, + "step": 18000 + }, + { + "epoch": 2.797452625038832, + "grad_norm": 0.24574346840381622, + "learning_rate": 9.544412251350408e-05, + "loss": 0.0152, + "step": 18010 + }, + { + "epoch": 2.7990059024541782, + "grad_norm": 0.2822411060333252, + "learning_rate": 9.543722419347422e-05, + "loss": 0.0194, + "step": 18020 + }, + { + "epoch": 2.800559179869525, + "grad_norm": 0.36165305972099304, + "learning_rate": 9.543032090449788e-05, + "loss": 0.0144, + "step": 18030 + }, + { + "epoch": 2.802112457284871, + "grad_norm": 0.3031671941280365, + "learning_rate": 9.542341264732992e-05, + "loss": 0.0175, + "step": 18040 + }, + { + "epoch": 2.8036657347002176, + "grad_norm": 0.21350154280662537, + "learning_rate": 9.541649942272585e-05, + "loss": 0.0125, + "step": 18050 + }, + { + "epoch": 2.805219012115564, + "grad_norm": 0.20317164063453674, + "learning_rate": 9.54095812314417e-05, + "loss": 0.0139, + "step": 18060 + }, + { + "epoch": 2.80677228953091, + "grad_norm": 0.2933678925037384, + "learning_rate": 9.540265807423401e-05, + "loss": 0.0121, + "step": 18070 + }, + { + "epoch": 2.8083255669462566, + "grad_norm": 0.393731951713562, + "learning_rate": 9.53957299518599e-05, + "loss": 0.012, + "step": 18080 + }, + { + "epoch": 2.809878844361603, + "grad_norm": 0.25301823019981384, + "learning_rate": 9.5388796865077e-05, + "loss": 0.0096, + "step": 18090 + }, + { + "epoch": 2.8114321217769493, + "grad_norm": 0.21467985212802887, + "learning_rate": 9.538185881464353e-05, + "loss": 0.0162, + "step": 18100 + }, + { + "epoch": 2.8129853991922955, + "grad_norm": 0.3032570481300354, + "learning_rate": 9.537491580131821e-05, + "loss": 0.0132, + "step": 18110 + }, + { + "epoch": 2.814538676607642, + "grad_norm": 0.33205145597457886, + "learning_rate": 9.53679678258603e-05, + "loss": 0.01, + "step": 18120 + }, + { + "epoch": 2.8160919540229887, + "grad_norm": 0.37848806381225586, + "learning_rate": 9.536101488902966e-05, + "loss": 0.0152, + "step": 18130 + }, + { + "epoch": 2.817645231438335, + "grad_norm": 0.293568879365921, + "learning_rate": 9.535405699158663e-05, + "loss": 0.0164, + "step": 18140 + }, + { + "epoch": 2.819198508853681, + "grad_norm": 0.26691731810569763, + "learning_rate": 9.53470941342921e-05, + "loss": 0.0126, + "step": 18150 + }, + { + "epoch": 2.8207517862690277, + "grad_norm": 0.2711033523082733, + "learning_rate": 9.534012631790756e-05, + "loss": 0.0124, + "step": 18160 + }, + { + "epoch": 2.8223050636843743, + "grad_norm": 0.22603954374790192, + "learning_rate": 9.533315354319494e-05, + "loss": 0.0116, + "step": 18170 + }, + { + "epoch": 2.8238583410997204, + "grad_norm": 0.31326842308044434, + "learning_rate": 9.532617581091682e-05, + "loss": 0.0135, + "step": 18180 + }, + { + "epoch": 2.8254116185150666, + "grad_norm": 0.3621816337108612, + "learning_rate": 9.531919312183629e-05, + "loss": 0.0152, + "step": 18190 + }, + { + "epoch": 2.8269648959304132, + "grad_norm": 0.2147025167942047, + "learning_rate": 9.531220547671688e-05, + "loss": 0.0098, + "step": 18200 + }, + { + "epoch": 2.8285181733457594, + "grad_norm": 0.22634613513946533, + "learning_rate": 9.530521287632285e-05, + "loss": 0.0103, + "step": 18210 + }, + { + "epoch": 2.830071450761106, + "grad_norm": 0.5167091488838196, + "learning_rate": 9.529821532141884e-05, + "loss": 0.0128, + "step": 18220 + }, + { + "epoch": 2.831624728176452, + "grad_norm": 0.5344523787498474, + "learning_rate": 9.52912128127701e-05, + "loss": 0.0175, + "step": 18230 + }, + { + "epoch": 2.8331780055917988, + "grad_norm": 0.364193856716156, + "learning_rate": 9.528420535114244e-05, + "loss": 0.0109, + "step": 18240 + }, + { + "epoch": 2.834731283007145, + "grad_norm": 0.3555275499820709, + "learning_rate": 9.527719293730215e-05, + "loss": 0.0136, + "step": 18250 + }, + { + "epoch": 2.8362845604224916, + "grad_norm": 0.27027493715286255, + "learning_rate": 9.527017557201611e-05, + "loss": 0.0121, + "step": 18260 + }, + { + "epoch": 2.8378378378378377, + "grad_norm": 0.27386388182640076, + "learning_rate": 9.526315325605176e-05, + "loss": 0.0137, + "step": 18270 + }, + { + "epoch": 2.8393911152531843, + "grad_norm": 0.27432698011398315, + "learning_rate": 9.525612599017699e-05, + "loss": 0.0111, + "step": 18280 + }, + { + "epoch": 2.8409443926685305, + "grad_norm": 0.2675636112689972, + "learning_rate": 9.524909377516033e-05, + "loss": 0.0128, + "step": 18290 + }, + { + "epoch": 2.842497670083877, + "grad_norm": 0.23969359695911407, + "learning_rate": 9.524205661177081e-05, + "loss": 0.0153, + "step": 18300 + }, + { + "epoch": 2.8440509474992233, + "grad_norm": 0.27205219864845276, + "learning_rate": 9.523501450077801e-05, + "loss": 0.0112, + "step": 18310 + }, + { + "epoch": 2.84560422491457, + "grad_norm": 0.15997314453125, + "learning_rate": 9.522796744295202e-05, + "loss": 0.0116, + "step": 18320 + }, + { + "epoch": 2.847157502329916, + "grad_norm": 0.36606040596961975, + "learning_rate": 9.522091543906352e-05, + "loss": 0.0097, + "step": 18330 + }, + { + "epoch": 2.8487107797452627, + "grad_norm": 0.21237586438655853, + "learning_rate": 9.521385848988369e-05, + "loss": 0.0136, + "step": 18340 + }, + { + "epoch": 2.850264057160609, + "grad_norm": 0.2421288937330246, + "learning_rate": 9.520679659618428e-05, + "loss": 0.0113, + "step": 18350 + }, + { + "epoch": 2.8518173345759554, + "grad_norm": 0.2712749242782593, + "learning_rate": 9.519972975873754e-05, + "loss": 0.0112, + "step": 18360 + }, + { + "epoch": 2.8533706119913016, + "grad_norm": 0.2383471131324768, + "learning_rate": 9.519265797831633e-05, + "loss": 0.0101, + "step": 18370 + }, + { + "epoch": 2.854923889406648, + "grad_norm": 0.7882151007652283, + "learning_rate": 9.518558125569399e-05, + "loss": 0.0115, + "step": 18380 + }, + { + "epoch": 2.8564771668219944, + "grad_norm": 0.3889428675174713, + "learning_rate": 9.517849959164442e-05, + "loss": 0.0099, + "step": 18390 + }, + { + "epoch": 2.8580304442373405, + "grad_norm": 0.2483224719762802, + "learning_rate": 9.517141298694205e-05, + "loss": 0.0134, + "step": 18400 + }, + { + "epoch": 2.859583721652687, + "grad_norm": 0.3188333511352539, + "learning_rate": 9.516432144236188e-05, + "loss": 0.0159, + "step": 18410 + }, + { + "epoch": 2.8611369990680338, + "grad_norm": 0.25816550850868225, + "learning_rate": 9.515722495867941e-05, + "loss": 0.013, + "step": 18420 + }, + { + "epoch": 2.86269027648338, + "grad_norm": 0.29084140062332153, + "learning_rate": 9.515012353667072e-05, + "loss": 0.0119, + "step": 18430 + }, + { + "epoch": 2.864243553898726, + "grad_norm": 0.22622434794902802, + "learning_rate": 9.51430171771124e-05, + "loss": 0.0093, + "step": 18440 + }, + { + "epoch": 2.8657968313140727, + "grad_norm": 0.24139732122421265, + "learning_rate": 9.513590588078159e-05, + "loss": 0.0133, + "step": 18450 + }, + { + "epoch": 2.8673501087294193, + "grad_norm": 0.21718953549861908, + "learning_rate": 9.512878964845597e-05, + "loss": 0.0127, + "step": 18460 + }, + { + "epoch": 2.8689033861447655, + "grad_norm": 0.2912321984767914, + "learning_rate": 9.512166848091377e-05, + "loss": 0.0128, + "step": 18470 + }, + { + "epoch": 2.8704566635601116, + "grad_norm": 0.2571829855442047, + "learning_rate": 9.511454237893376e-05, + "loss": 0.0121, + "step": 18480 + }, + { + "epoch": 2.8720099409754583, + "grad_norm": 0.26689413189888, + "learning_rate": 9.51074113432952e-05, + "loss": 0.0136, + "step": 18490 + }, + { + "epoch": 2.873563218390805, + "grad_norm": 0.21859589219093323, + "learning_rate": 9.510027537477797e-05, + "loss": 0.016, + "step": 18500 + }, + { + "epoch": 2.875116495806151, + "grad_norm": 0.23000536859035492, + "learning_rate": 9.509313447416242e-05, + "loss": 0.0128, + "step": 18510 + }, + { + "epoch": 2.876669773221497, + "grad_norm": 0.17780470848083496, + "learning_rate": 9.508598864222949e-05, + "loss": 0.0158, + "step": 18520 + }, + { + "epoch": 2.878223050636844, + "grad_norm": 0.2424098253250122, + "learning_rate": 9.507883787976062e-05, + "loss": 0.0129, + "step": 18530 + }, + { + "epoch": 2.87977632805219, + "grad_norm": 0.2550719678401947, + "learning_rate": 9.507168218753781e-05, + "loss": 0.0184, + "step": 18540 + }, + { + "epoch": 2.8813296054675366, + "grad_norm": 0.21634788811206818, + "learning_rate": 9.506452156634362e-05, + "loss": 0.0133, + "step": 18550 + }, + { + "epoch": 2.8828828828828827, + "grad_norm": 0.3126607835292816, + "learning_rate": 9.505735601696109e-05, + "loss": 0.0152, + "step": 18560 + }, + { + "epoch": 2.8844361602982294, + "grad_norm": 0.2470981478691101, + "learning_rate": 9.505018554017385e-05, + "loss": 0.012, + "step": 18570 + }, + { + "epoch": 2.8859894377135755, + "grad_norm": 0.32953035831451416, + "learning_rate": 9.504301013676604e-05, + "loss": 0.0122, + "step": 18580 + }, + { + "epoch": 2.887542715128922, + "grad_norm": 0.3486786484718323, + "learning_rate": 9.503582980752238e-05, + "loss": 0.0147, + "step": 18590 + }, + { + "epoch": 2.8890959925442683, + "grad_norm": 0.3113548755645752, + "learning_rate": 9.502864455322809e-05, + "loss": 0.0119, + "step": 18600 + }, + { + "epoch": 2.890649269959615, + "grad_norm": 0.32099664211273193, + "learning_rate": 9.502145437466891e-05, + "loss": 0.0146, + "step": 18610 + }, + { + "epoch": 2.892202547374961, + "grad_norm": 0.28158316016197205, + "learning_rate": 9.501425927263116e-05, + "loss": 0.0142, + "step": 18620 + }, + { + "epoch": 2.8937558247903077, + "grad_norm": 0.23080046474933624, + "learning_rate": 9.500705924790172e-05, + "loss": 0.0116, + "step": 18630 + }, + { + "epoch": 2.895309102205654, + "grad_norm": 0.18200159072875977, + "learning_rate": 9.499985430126794e-05, + "loss": 0.0136, + "step": 18640 + }, + { + "epoch": 2.8968623796210005, + "grad_norm": 0.23076112568378448, + "learning_rate": 9.499264443351775e-05, + "loss": 0.0119, + "step": 18650 + }, + { + "epoch": 2.8984156570363466, + "grad_norm": 0.3826637864112854, + "learning_rate": 9.498542964543961e-05, + "loss": 0.0112, + "step": 18660 + }, + { + "epoch": 2.8999689344516932, + "grad_norm": 0.21169844269752502, + "learning_rate": 9.497820993782252e-05, + "loss": 0.0198, + "step": 18670 + }, + { + "epoch": 2.9015222118670394, + "grad_norm": 0.198900505900383, + "learning_rate": 9.497098531145601e-05, + "loss": 0.0112, + "step": 18680 + }, + { + "epoch": 2.9030754892823856, + "grad_norm": 0.6476576328277588, + "learning_rate": 9.496375576713017e-05, + "loss": 0.0135, + "step": 18690 + }, + { + "epoch": 2.904628766697732, + "grad_norm": 0.29089683294296265, + "learning_rate": 9.49565213056356e-05, + "loss": 0.0146, + "step": 18700 + }, + { + "epoch": 2.906182044113079, + "grad_norm": 0.18277356028556824, + "learning_rate": 9.494928192776342e-05, + "loss": 0.0083, + "step": 18710 + }, + { + "epoch": 2.907735321528425, + "grad_norm": 0.2327483594417572, + "learning_rate": 9.494203763430538e-05, + "loss": 0.009, + "step": 18720 + }, + { + "epoch": 2.909288598943771, + "grad_norm": 0.2083003669977188, + "learning_rate": 9.493478842605366e-05, + "loss": 0.0117, + "step": 18730 + }, + { + "epoch": 2.9108418763591177, + "grad_norm": 0.3121630549430847, + "learning_rate": 9.492753430380105e-05, + "loss": 0.0113, + "step": 18740 + }, + { + "epoch": 2.9123951537744643, + "grad_norm": 0.19776159524917603, + "learning_rate": 9.492027526834083e-05, + "loss": 0.0101, + "step": 18750 + }, + { + "epoch": 2.9139484311898105, + "grad_norm": 0.24807654321193695, + "learning_rate": 9.491301132046684e-05, + "loss": 0.0102, + "step": 18760 + }, + { + "epoch": 2.9155017086051567, + "grad_norm": 0.3085631728172302, + "learning_rate": 9.490574246097345e-05, + "loss": 0.0134, + "step": 18770 + }, + { + "epoch": 2.9170549860205033, + "grad_norm": 0.26046222448349, + "learning_rate": 9.48984686906556e-05, + "loss": 0.0119, + "step": 18780 + }, + { + "epoch": 2.91860826343585, + "grad_norm": 0.2034911811351776, + "learning_rate": 9.489119001030871e-05, + "loss": 0.0137, + "step": 18790 + }, + { + "epoch": 2.920161540851196, + "grad_norm": 0.31858739256858826, + "learning_rate": 9.488390642072878e-05, + "loss": 0.0115, + "step": 18800 + }, + { + "epoch": 2.9217148182665422, + "grad_norm": 0.2329019010066986, + "learning_rate": 9.48766179227123e-05, + "loss": 0.0095, + "step": 18810 + }, + { + "epoch": 2.923268095681889, + "grad_norm": 0.42213088274002075, + "learning_rate": 9.486932451705636e-05, + "loss": 0.0154, + "step": 18820 + }, + { + "epoch": 2.924821373097235, + "grad_norm": 0.3092533349990845, + "learning_rate": 9.486202620455857e-05, + "loss": 0.0116, + "step": 18830 + }, + { + "epoch": 2.9263746505125816, + "grad_norm": 0.35486382246017456, + "learning_rate": 9.485472298601704e-05, + "loss": 0.0097, + "step": 18840 + }, + { + "epoch": 2.9279279279279278, + "grad_norm": 0.43649041652679443, + "learning_rate": 9.484741486223043e-05, + "loss": 0.014, + "step": 18850 + }, + { + "epoch": 2.9294812053432744, + "grad_norm": 0.4185431897640228, + "learning_rate": 9.484010183399797e-05, + "loss": 0.0136, + "step": 18860 + }, + { + "epoch": 2.9310344827586206, + "grad_norm": 0.41346824169158936, + "learning_rate": 9.483278390211938e-05, + "loss": 0.0124, + "step": 18870 + }, + { + "epoch": 2.932587760173967, + "grad_norm": 0.2714039981365204, + "learning_rate": 9.482546106739496e-05, + "loss": 0.0104, + "step": 18880 + }, + { + "epoch": 2.9341410375893133, + "grad_norm": 0.297014445066452, + "learning_rate": 9.48181333306255e-05, + "loss": 0.0105, + "step": 18890 + }, + { + "epoch": 2.93569431500466, + "grad_norm": 0.41303515434265137, + "learning_rate": 9.481080069261237e-05, + "loss": 0.0121, + "step": 18900 + }, + { + "epoch": 2.937247592420006, + "grad_norm": 0.31028953194618225, + "learning_rate": 9.480346315415745e-05, + "loss": 0.0111, + "step": 18910 + }, + { + "epoch": 2.9388008698353527, + "grad_norm": 0.2570362985134125, + "learning_rate": 9.479612071606314e-05, + "loss": 0.0139, + "step": 18920 + }, + { + "epoch": 2.940354147250699, + "grad_norm": 0.38641786575317383, + "learning_rate": 9.478877337913244e-05, + "loss": 0.0141, + "step": 18930 + }, + { + "epoch": 2.9419074246660455, + "grad_norm": 0.4110334813594818, + "learning_rate": 9.478142114416881e-05, + "loss": 0.0125, + "step": 18940 + }, + { + "epoch": 2.9434607020813917, + "grad_norm": 0.29259490966796875, + "learning_rate": 9.47740640119763e-05, + "loss": 0.0112, + "step": 18950 + }, + { + "epoch": 2.9450139794967383, + "grad_norm": 0.3107304871082306, + "learning_rate": 9.476670198335947e-05, + "loss": 0.0099, + "step": 18960 + }, + { + "epoch": 2.9465672569120844, + "grad_norm": 0.48729339241981506, + "learning_rate": 9.47593350591234e-05, + "loss": 0.0113, + "step": 18970 + }, + { + "epoch": 2.948120534327431, + "grad_norm": 0.32369399070739746, + "learning_rate": 9.475196324007376e-05, + "loss": 0.0127, + "step": 18980 + }, + { + "epoch": 2.949673811742777, + "grad_norm": 0.2408108115196228, + "learning_rate": 9.474458652701669e-05, + "loss": 0.0062, + "step": 18990 + }, + { + "epoch": 2.951227089158124, + "grad_norm": 0.19254474341869354, + "learning_rate": 9.473720492075892e-05, + "loss": 0.0092, + "step": 19000 + }, + { + "epoch": 2.95278036657347, + "grad_norm": 0.2353518307209015, + "learning_rate": 9.472981842210768e-05, + "loss": 0.0088, + "step": 19010 + }, + { + "epoch": 2.954333643988816, + "grad_norm": 0.34170466661453247, + "learning_rate": 9.472242703187074e-05, + "loss": 0.0136, + "step": 19020 + }, + { + "epoch": 2.9558869214041628, + "grad_norm": 0.24583151936531067, + "learning_rate": 9.471503075085643e-05, + "loss": 0.0131, + "step": 19030 + }, + { + "epoch": 2.9574401988195094, + "grad_norm": 0.21525827050209045, + "learning_rate": 9.470762957987359e-05, + "loss": 0.0086, + "step": 19040 + }, + { + "epoch": 2.9589934762348555, + "grad_norm": 0.28503307700157166, + "learning_rate": 9.470022351973158e-05, + "loss": 0.0117, + "step": 19050 + }, + { + "epoch": 2.9605467536502017, + "grad_norm": 0.17607992887496948, + "learning_rate": 9.469281257124034e-05, + "loss": 0.0086, + "step": 19060 + }, + { + "epoch": 2.9621000310655483, + "grad_norm": 0.3137458860874176, + "learning_rate": 9.46853967352103e-05, + "loss": 0.0125, + "step": 19070 + }, + { + "epoch": 2.963653308480895, + "grad_norm": 0.4404831826686859, + "learning_rate": 9.467797601245246e-05, + "loss": 0.0108, + "step": 19080 + }, + { + "epoch": 2.965206585896241, + "grad_norm": 0.3242684304714203, + "learning_rate": 9.467055040377834e-05, + "loss": 0.0087, + "step": 19090 + }, + { + "epoch": 2.9667598633115873, + "grad_norm": 0.3226889669895172, + "learning_rate": 9.466311990999999e-05, + "loss": 0.0118, + "step": 19100 + }, + { + "epoch": 2.968313140726934, + "grad_norm": 0.3348540961742401, + "learning_rate": 9.465568453193e-05, + "loss": 0.0091, + "step": 19110 + }, + { + "epoch": 2.9698664181422805, + "grad_norm": 0.3566052317619324, + "learning_rate": 9.464824427038148e-05, + "loss": 0.0148, + "step": 19120 + }, + { + "epoch": 2.9714196955576266, + "grad_norm": 0.2936733365058899, + "learning_rate": 9.46407991261681e-05, + "loss": 0.012, + "step": 19130 + }, + { + "epoch": 2.972972972972973, + "grad_norm": 0.25623783469200134, + "learning_rate": 9.463334910010404e-05, + "loss": 0.0113, + "step": 19140 + }, + { + "epoch": 2.9745262503883194, + "grad_norm": 0.21039102971553802, + "learning_rate": 9.462589419300403e-05, + "loss": 0.0137, + "step": 19150 + }, + { + "epoch": 2.9760795278036656, + "grad_norm": 0.3283661901950836, + "learning_rate": 9.461843440568333e-05, + "loss": 0.0118, + "step": 19160 + }, + { + "epoch": 2.977632805219012, + "grad_norm": 0.2491186559200287, + "learning_rate": 9.461096973895773e-05, + "loss": 0.0153, + "step": 19170 + }, + { + "epoch": 2.9791860826343584, + "grad_norm": 0.2496190220117569, + "learning_rate": 9.460350019364355e-05, + "loss": 0.0131, + "step": 19180 + }, + { + "epoch": 2.980739360049705, + "grad_norm": 0.41105425357818604, + "learning_rate": 9.459602577055764e-05, + "loss": 0.0136, + "step": 19190 + }, + { + "epoch": 2.982292637465051, + "grad_norm": 0.3463808298110962, + "learning_rate": 9.45885464705174e-05, + "loss": 0.0121, + "step": 19200 + }, + { + "epoch": 2.9838459148803977, + "grad_norm": 0.2472551465034485, + "learning_rate": 9.458106229434076e-05, + "loss": 0.0105, + "step": 19210 + }, + { + "epoch": 2.985399192295744, + "grad_norm": 0.2884030342102051, + "learning_rate": 9.457357324284617e-05, + "loss": 0.0117, + "step": 19220 + }, + { + "epoch": 2.9869524697110905, + "grad_norm": 0.7362286448478699, + "learning_rate": 9.456607931685262e-05, + "loss": 0.0123, + "step": 19230 + }, + { + "epoch": 2.9885057471264367, + "grad_norm": 0.27597883343696594, + "learning_rate": 9.455858051717965e-05, + "loss": 0.0134, + "step": 19240 + }, + { + "epoch": 2.9900590245417833, + "grad_norm": 0.24286429584026337, + "learning_rate": 9.45510768446473e-05, + "loss": 0.0112, + "step": 19250 + }, + { + "epoch": 2.9916123019571295, + "grad_norm": 0.18219305574893951, + "learning_rate": 9.454356830007618e-05, + "loss": 0.0121, + "step": 19260 + }, + { + "epoch": 2.993165579372476, + "grad_norm": 0.4376879334449768, + "learning_rate": 9.45360548842874e-05, + "loss": 0.0103, + "step": 19270 + }, + { + "epoch": 2.9947188567878222, + "grad_norm": 0.19992595911026, + "learning_rate": 9.452853659810261e-05, + "loss": 0.0108, + "step": 19280 + }, + { + "epoch": 2.996272134203169, + "grad_norm": 0.2621522843837738, + "learning_rate": 9.452101344234401e-05, + "loss": 0.0104, + "step": 19290 + }, + { + "epoch": 2.997825411618515, + "grad_norm": 0.3175516724586487, + "learning_rate": 9.451348541783431e-05, + "loss": 0.0111, + "step": 19300 + }, + { + "epoch": 2.999378689033861, + "grad_norm": 0.2838965654373169, + "learning_rate": 9.450595252539678e-05, + "loss": 0.011, + "step": 19310 + }, + { + "epoch": 3.000931966449208, + "grad_norm": 0.1815713793039322, + "learning_rate": 9.449841476585518e-05, + "loss": 0.0098, + "step": 19320 + }, + { + "epoch": 3.0024852438645544, + "grad_norm": 0.4007812738418579, + "learning_rate": 9.449087214003384e-05, + "loss": 0.012, + "step": 19330 + }, + { + "epoch": 3.0040385212799006, + "grad_norm": 0.21935118734836578, + "learning_rate": 9.448332464875765e-05, + "loss": 0.0128, + "step": 19340 + }, + { + "epoch": 3.005591798695247, + "grad_norm": 0.26544272899627686, + "learning_rate": 9.447577229285192e-05, + "loss": 0.01, + "step": 19350 + }, + { + "epoch": 3.0071450761105933, + "grad_norm": 0.29527780413627625, + "learning_rate": 9.446821507314261e-05, + "loss": 0.0125, + "step": 19360 + }, + { + "epoch": 3.00869835352594, + "grad_norm": 0.42588144540786743, + "learning_rate": 9.446065299045617e-05, + "loss": 0.0113, + "step": 19370 + }, + { + "epoch": 3.010251630941286, + "grad_norm": 0.27514922618865967, + "learning_rate": 9.445308604561955e-05, + "loss": 0.0113, + "step": 19380 + }, + { + "epoch": 3.0118049083566323, + "grad_norm": 0.33496496081352234, + "learning_rate": 9.444551423946028e-05, + "loss": 0.0113, + "step": 19390 + }, + { + "epoch": 3.013358185771979, + "grad_norm": 0.28873828053474426, + "learning_rate": 9.443793757280638e-05, + "loss": 0.013, + "step": 19400 + }, + { + "epoch": 3.014911463187325, + "grad_norm": 0.23120366036891937, + "learning_rate": 9.443035604648646e-05, + "loss": 0.0114, + "step": 19410 + }, + { + "epoch": 3.0164647406026717, + "grad_norm": 0.22900083661079407, + "learning_rate": 9.44227696613296e-05, + "loss": 0.0101, + "step": 19420 + }, + { + "epoch": 3.018018018018018, + "grad_norm": 0.29566672444343567, + "learning_rate": 9.441517841816542e-05, + "loss": 0.0111, + "step": 19430 + }, + { + "epoch": 3.0195712954333644, + "grad_norm": 0.2680406868457794, + "learning_rate": 9.440758231782413e-05, + "loss": 0.0118, + "step": 19440 + }, + { + "epoch": 3.0211245728487106, + "grad_norm": 0.24159018695354462, + "learning_rate": 9.439998136113639e-05, + "loss": 0.012, + "step": 19450 + }, + { + "epoch": 3.022677850264057, + "grad_norm": 0.24651426076889038, + "learning_rate": 9.439237554893344e-05, + "loss": 0.0116, + "step": 19460 + }, + { + "epoch": 3.0242311276794034, + "grad_norm": 0.34677401185035706, + "learning_rate": 9.438476488204705e-05, + "loss": 0.0111, + "step": 19470 + }, + { + "epoch": 3.02578440509475, + "grad_norm": 0.22353598475456238, + "learning_rate": 9.43771493613095e-05, + "loss": 0.0126, + "step": 19480 + }, + { + "epoch": 3.027337682510096, + "grad_norm": 0.27751702070236206, + "learning_rate": 9.436952898755362e-05, + "loss": 0.0132, + "step": 19490 + }, + { + "epoch": 3.0288909599254428, + "grad_norm": 0.2564186155796051, + "learning_rate": 9.436190376161276e-05, + "loss": 0.0113, + "step": 19500 + }, + { + "epoch": 3.030444237340789, + "grad_norm": 0.2304266095161438, + "learning_rate": 9.43542736843208e-05, + "loss": 0.0136, + "step": 19510 + }, + { + "epoch": 3.0319975147561355, + "grad_norm": 0.2970900535583496, + "learning_rate": 9.434663875651216e-05, + "loss": 0.0115, + "step": 19520 + }, + { + "epoch": 3.0335507921714817, + "grad_norm": 0.3123922049999237, + "learning_rate": 9.433899897902177e-05, + "loss": 0.0115, + "step": 19530 + }, + { + "epoch": 3.0351040695868283, + "grad_norm": 0.3022879660129547, + "learning_rate": 9.433135435268511e-05, + "loss": 0.0083, + "step": 19540 + }, + { + "epoch": 3.0366573470021745, + "grad_norm": 0.2719579339027405, + "learning_rate": 9.432370487833819e-05, + "loss": 0.0166, + "step": 19550 + }, + { + "epoch": 3.038210624417521, + "grad_norm": 0.1858803629875183, + "learning_rate": 9.431605055681756e-05, + "loss": 0.0146, + "step": 19560 + }, + { + "epoch": 3.0397639018328673, + "grad_norm": 0.32513049244880676, + "learning_rate": 9.430839138896026e-05, + "loss": 0.0173, + "step": 19570 + }, + { + "epoch": 3.041317179248214, + "grad_norm": 0.2647060453891754, + "learning_rate": 9.43007273756039e-05, + "loss": 0.0115, + "step": 19580 + }, + { + "epoch": 3.04287045666356, + "grad_norm": 0.30064430832862854, + "learning_rate": 9.429305851758658e-05, + "loss": 0.0123, + "step": 19590 + }, + { + "epoch": 3.0444237340789067, + "grad_norm": 0.2685665190219879, + "learning_rate": 9.428538481574699e-05, + "loss": 0.011, + "step": 19600 + }, + { + "epoch": 3.045977011494253, + "grad_norm": 0.21263734996318817, + "learning_rate": 9.42777062709243e-05, + "loss": 0.0099, + "step": 19610 + }, + { + "epoch": 3.0475302889095994, + "grad_norm": 0.39744436740875244, + "learning_rate": 9.427002288395821e-05, + "loss": 0.0121, + "step": 19620 + }, + { + "epoch": 3.0490835663249456, + "grad_norm": 0.5306727886199951, + "learning_rate": 9.426233465568898e-05, + "loss": 0.015, + "step": 19630 + }, + { + "epoch": 3.050636843740292, + "grad_norm": 0.2937623858451843, + "learning_rate": 9.42546415869574e-05, + "loss": 0.0119, + "step": 19640 + }, + { + "epoch": 3.0521901211556384, + "grad_norm": 0.21801349520683289, + "learning_rate": 9.424694367860473e-05, + "loss": 0.0117, + "step": 19650 + }, + { + "epoch": 3.053743398570985, + "grad_norm": 0.6240372061729431, + "learning_rate": 9.423924093147284e-05, + "loss": 0.0148, + "step": 19660 + }, + { + "epoch": 3.055296675986331, + "grad_norm": 0.251109778881073, + "learning_rate": 9.423153334640407e-05, + "loss": 0.0118, + "step": 19670 + }, + { + "epoch": 3.0568499534016778, + "grad_norm": 0.2646167278289795, + "learning_rate": 9.42238209242413e-05, + "loss": 0.0134, + "step": 19680 + }, + { + "epoch": 3.058403230817024, + "grad_norm": 0.2500948905944824, + "learning_rate": 9.421610366582798e-05, + "loss": 0.0111, + "step": 19690 + }, + { + "epoch": 3.05995650823237, + "grad_norm": 0.3422153890132904, + "learning_rate": 9.420838157200803e-05, + "loss": 0.0147, + "step": 19700 + }, + { + "epoch": 3.0615097856477167, + "grad_norm": 0.2159009873867035, + "learning_rate": 9.420065464362594e-05, + "loss": 0.0143, + "step": 19710 + }, + { + "epoch": 3.063063063063063, + "grad_norm": 0.33432242274284363, + "learning_rate": 9.419292288152673e-05, + "loss": 0.0129, + "step": 19720 + }, + { + "epoch": 3.0646163404784095, + "grad_norm": 0.2619082033634186, + "learning_rate": 9.418518628655588e-05, + "loss": 0.011, + "step": 19730 + }, + { + "epoch": 3.0661696178937556, + "grad_norm": 0.20173026621341705, + "learning_rate": 9.417744485955951e-05, + "loss": 0.0108, + "step": 19740 + }, + { + "epoch": 3.0677228953091022, + "grad_norm": 0.25106507539749146, + "learning_rate": 9.41696986013842e-05, + "loss": 0.0136, + "step": 19750 + }, + { + "epoch": 3.0692761727244484, + "grad_norm": 0.1945876032114029, + "learning_rate": 9.416194751287705e-05, + "loss": 0.0144, + "step": 19760 + }, + { + "epoch": 3.070829450139795, + "grad_norm": 0.2704847753047943, + "learning_rate": 9.415419159488572e-05, + "loss": 0.0119, + "step": 19770 + }, + { + "epoch": 3.072382727555141, + "grad_norm": 0.24541018903255463, + "learning_rate": 9.414643084825837e-05, + "loss": 0.0149, + "step": 19780 + }, + { + "epoch": 3.073936004970488, + "grad_norm": 0.27127906680107117, + "learning_rate": 9.413866527384372e-05, + "loss": 0.0109, + "step": 19790 + }, + { + "epoch": 3.075489282385834, + "grad_norm": 0.3081499934196472, + "learning_rate": 9.4130894872491e-05, + "loss": 0.0121, + "step": 19800 + }, + { + "epoch": 3.0770425598011806, + "grad_norm": 0.29673898220062256, + "learning_rate": 9.412311964504998e-05, + "loss": 0.0121, + "step": 19810 + }, + { + "epoch": 3.0785958372165267, + "grad_norm": 0.26920297741889954, + "learning_rate": 9.411533959237091e-05, + "loss": 0.0103, + "step": 19820 + }, + { + "epoch": 3.0801491146318734, + "grad_norm": 0.31113672256469727, + "learning_rate": 9.410755471530464e-05, + "loss": 0.0135, + "step": 19830 + }, + { + "epoch": 3.0817023920472195, + "grad_norm": 0.25350257754325867, + "learning_rate": 9.40997650147025e-05, + "loss": 0.0126, + "step": 19840 + }, + { + "epoch": 3.083255669462566, + "grad_norm": 0.2561637759208679, + "learning_rate": 9.409197049141637e-05, + "loss": 0.0101, + "step": 19850 + }, + { + "epoch": 3.0848089468779123, + "grad_norm": 0.2942717373371124, + "learning_rate": 9.408417114629863e-05, + "loss": 0.015, + "step": 19860 + }, + { + "epoch": 3.086362224293259, + "grad_norm": 0.23123545944690704, + "learning_rate": 9.40763669802022e-05, + "loss": 0.0113, + "step": 19870 + }, + { + "epoch": 3.087915501708605, + "grad_norm": 0.2028033584356308, + "learning_rate": 9.406855799398056e-05, + "loss": 0.0115, + "step": 19880 + }, + { + "epoch": 3.0894687791239517, + "grad_norm": 0.27862441539764404, + "learning_rate": 9.406074418848767e-05, + "loss": 0.0122, + "step": 19890 + }, + { + "epoch": 3.091022056539298, + "grad_norm": 0.4030202031135559, + "learning_rate": 9.405292556457805e-05, + "loss": 0.0153, + "step": 19900 + }, + { + "epoch": 3.0925753339546445, + "grad_norm": 0.2521499693393707, + "learning_rate": 9.404510212310671e-05, + "loss": 0.0137, + "step": 19910 + }, + { + "epoch": 3.0941286113699906, + "grad_norm": 0.2969946563243866, + "learning_rate": 9.403727386492924e-05, + "loss": 0.0103, + "step": 19920 + }, + { + "epoch": 3.0956818887853372, + "grad_norm": 0.28713247179985046, + "learning_rate": 9.40294407909017e-05, + "loss": 0.0136, + "step": 19930 + }, + { + "epoch": 3.0972351662006834, + "grad_norm": 0.3595671057701111, + "learning_rate": 9.40216029018807e-05, + "loss": 0.0099, + "step": 19940 + }, + { + "epoch": 3.09878844361603, + "grad_norm": 0.328769713640213, + "learning_rate": 9.401376019872338e-05, + "loss": 0.0123, + "step": 19950 + }, + { + "epoch": 3.100341721031376, + "grad_norm": 0.2586689591407776, + "learning_rate": 9.400591268228746e-05, + "loss": 0.0116, + "step": 19960 + }, + { + "epoch": 3.101894998446723, + "grad_norm": 0.31035125255584717, + "learning_rate": 9.399806035343106e-05, + "loss": 0.0125, + "step": 19970 + }, + { + "epoch": 3.103448275862069, + "grad_norm": 0.3865697681903839, + "learning_rate": 9.399020321301294e-05, + "loss": 0.0098, + "step": 19980 + }, + { + "epoch": 3.1050015532774156, + "grad_norm": 0.3203708231449127, + "learning_rate": 9.398234126189234e-05, + "loss": 0.0122, + "step": 19990 + }, + { + "epoch": 3.1065548306927617, + "grad_norm": 0.3414400815963745, + "learning_rate": 9.397447450092902e-05, + "loss": 0.0103, + "step": 20000 + }, + { + "epoch": 3.108108108108108, + "grad_norm": 0.2497052252292633, + "learning_rate": 9.39666029309833e-05, + "loss": 0.0124, + "step": 20010 + }, + { + "epoch": 3.1096613855234545, + "grad_norm": 0.22898992896080017, + "learning_rate": 9.395872655291596e-05, + "loss": 0.0074, + "step": 20020 + }, + { + "epoch": 3.1112146629388007, + "grad_norm": 0.32783377170562744, + "learning_rate": 9.395084536758838e-05, + "loss": 0.0125, + "step": 20030 + }, + { + "epoch": 3.1127679403541473, + "grad_norm": 0.3119792938232422, + "learning_rate": 9.394295937586243e-05, + "loss": 0.0112, + "step": 20040 + }, + { + "epoch": 3.1143212177694934, + "grad_norm": 0.2999776601791382, + "learning_rate": 9.393506857860052e-05, + "loss": 0.0104, + "step": 20050 + }, + { + "epoch": 3.11587449518484, + "grad_norm": 0.42066580057144165, + "learning_rate": 9.392717297666555e-05, + "loss": 0.0123, + "step": 20060 + }, + { + "epoch": 3.117427772600186, + "grad_norm": 0.5834600329399109, + "learning_rate": 9.391927257092101e-05, + "loss": 0.0122, + "step": 20070 + }, + { + "epoch": 3.118981050015533, + "grad_norm": 0.2610912024974823, + "learning_rate": 9.391136736223085e-05, + "loss": 0.0117, + "step": 20080 + }, + { + "epoch": 3.120534327430879, + "grad_norm": 0.6126047372817993, + "learning_rate": 9.390345735145956e-05, + "loss": 0.0117, + "step": 20090 + }, + { + "epoch": 3.1220876048462256, + "grad_norm": 0.2454865723848343, + "learning_rate": 9.389554253947219e-05, + "loss": 0.0111, + "step": 20100 + }, + { + "epoch": 3.1236408822615718, + "grad_norm": 0.18731960654258728, + "learning_rate": 9.388762292713428e-05, + "loss": 0.0108, + "step": 20110 + }, + { + "epoch": 3.1251941596769184, + "grad_norm": 0.1814521849155426, + "learning_rate": 9.38796985153119e-05, + "loss": 0.009, + "step": 20120 + }, + { + "epoch": 3.1267474370922645, + "grad_norm": 0.2008492350578308, + "learning_rate": 9.387176930487169e-05, + "loss": 0.0102, + "step": 20130 + }, + { + "epoch": 3.128300714507611, + "grad_norm": 0.27331680059432983, + "learning_rate": 9.386383529668072e-05, + "loss": 0.0112, + "step": 20140 + }, + { + "epoch": 3.1298539919229573, + "grad_norm": 0.23787230253219604, + "learning_rate": 9.385589649160669e-05, + "loss": 0.0134, + "step": 20150 + }, + { + "epoch": 3.131407269338304, + "grad_norm": 0.2379443496465683, + "learning_rate": 9.384795289051775e-05, + "loss": 0.0099, + "step": 20160 + }, + { + "epoch": 3.13296054675365, + "grad_norm": 0.2288418561220169, + "learning_rate": 9.384000449428261e-05, + "loss": 0.0104, + "step": 20170 + }, + { + "epoch": 3.1345138241689967, + "grad_norm": 0.31677353382110596, + "learning_rate": 9.383205130377048e-05, + "loss": 0.0128, + "step": 20180 + }, + { + "epoch": 3.136067101584343, + "grad_norm": 0.20908957719802856, + "learning_rate": 9.382409331985114e-05, + "loss": 0.0112, + "step": 20190 + }, + { + "epoch": 3.1376203789996895, + "grad_norm": 0.24375629425048828, + "learning_rate": 9.381613054339482e-05, + "loss": 0.013, + "step": 20200 + }, + { + "epoch": 3.1391736564150357, + "grad_norm": 0.2426828294992447, + "learning_rate": 9.380816297527235e-05, + "loss": 0.0094, + "step": 20210 + }, + { + "epoch": 3.1407269338303823, + "grad_norm": 0.2953178286552429, + "learning_rate": 9.380019061635506e-05, + "loss": 0.0094, + "step": 20220 + }, + { + "epoch": 3.1422802112457284, + "grad_norm": 0.23979954421520233, + "learning_rate": 9.379221346751474e-05, + "loss": 0.0114, + "step": 20230 + }, + { + "epoch": 3.143833488661075, + "grad_norm": 0.2571571171283722, + "learning_rate": 9.378423152962382e-05, + "loss": 0.0084, + "step": 20240 + }, + { + "epoch": 3.145386766076421, + "grad_norm": 0.25140005350112915, + "learning_rate": 9.377624480355517e-05, + "loss": 0.0084, + "step": 20250 + }, + { + "epoch": 3.146940043491768, + "grad_norm": 0.22560931742191315, + "learning_rate": 9.376825329018219e-05, + "loss": 0.0112, + "step": 20260 + }, + { + "epoch": 3.148493320907114, + "grad_norm": 0.17148323357105255, + "learning_rate": 9.376025699037884e-05, + "loss": 0.0084, + "step": 20270 + }, + { + "epoch": 3.1500465983224606, + "grad_norm": 0.27792325615882874, + "learning_rate": 9.37522559050196e-05, + "loss": 0.0109, + "step": 20280 + }, + { + "epoch": 3.1515998757378068, + "grad_norm": 0.2261061817407608, + "learning_rate": 9.37442500349794e-05, + "loss": 0.0103, + "step": 20290 + }, + { + "epoch": 3.153153153153153, + "grad_norm": 0.10395192354917526, + "learning_rate": 9.373623938113381e-05, + "loss": 0.0098, + "step": 20300 + }, + { + "epoch": 3.1547064305684995, + "grad_norm": 0.1917177140712738, + "learning_rate": 9.372822394435883e-05, + "loss": 0.0115, + "step": 20310 + }, + { + "epoch": 3.156259707983846, + "grad_norm": 0.15007278323173523, + "learning_rate": 9.372020372553102e-05, + "loss": 0.0124, + "step": 20320 + }, + { + "epoch": 3.1578129853991923, + "grad_norm": 0.3122901916503906, + "learning_rate": 9.371217872552746e-05, + "loss": 0.0114, + "step": 20330 + }, + { + "epoch": 3.1593662628145385, + "grad_norm": 0.29330167174339294, + "learning_rate": 9.370414894522576e-05, + "loss": 0.013, + "step": 20340 + }, + { + "epoch": 3.160919540229885, + "grad_norm": 0.29678335785865784, + "learning_rate": 9.369611438550406e-05, + "loss": 0.0117, + "step": 20350 + }, + { + "epoch": 3.1624728176452312, + "grad_norm": 0.2590521574020386, + "learning_rate": 9.368807504724095e-05, + "loss": 0.0136, + "step": 20360 + }, + { + "epoch": 3.164026095060578, + "grad_norm": 0.2275535613298416, + "learning_rate": 9.368003093131565e-05, + "loss": 0.0103, + "step": 20370 + }, + { + "epoch": 3.165579372475924, + "grad_norm": 0.34555989503860474, + "learning_rate": 9.367198203860785e-05, + "loss": 0.0075, + "step": 20380 + }, + { + "epoch": 3.1671326498912706, + "grad_norm": 0.3062836825847626, + "learning_rate": 9.366392836999774e-05, + "loss": 0.0157, + "step": 20390 + }, + { + "epoch": 3.168685927306617, + "grad_norm": 0.1386418342590332, + "learning_rate": 9.365586992636607e-05, + "loss": 0.0101, + "step": 20400 + }, + { + "epoch": 3.1702392047219634, + "grad_norm": 0.392526775598526, + "learning_rate": 9.364780670859412e-05, + "loss": 0.014, + "step": 20410 + }, + { + "epoch": 3.1717924821373096, + "grad_norm": 0.40599122643470764, + "learning_rate": 9.363973871756364e-05, + "loss": 0.0114, + "step": 20420 + }, + { + "epoch": 3.173345759552656, + "grad_norm": 0.47228074073791504, + "learning_rate": 9.363166595415696e-05, + "loss": 0.0103, + "step": 20430 + }, + { + "epoch": 3.1748990369680024, + "grad_norm": 0.461098849773407, + "learning_rate": 9.362358841925686e-05, + "loss": 0.0159, + "step": 20440 + }, + { + "epoch": 3.176452314383349, + "grad_norm": 0.3781289756298065, + "learning_rate": 9.361550611374674e-05, + "loss": 0.0141, + "step": 20450 + }, + { + "epoch": 3.178005591798695, + "grad_norm": 0.20647141337394714, + "learning_rate": 9.360741903851043e-05, + "loss": 0.0162, + "step": 20460 + }, + { + "epoch": 3.1795588692140417, + "grad_norm": 0.16266575455665588, + "learning_rate": 9.359932719443236e-05, + "loss": 0.0132, + "step": 20470 + }, + { + "epoch": 3.181112146629388, + "grad_norm": 0.3278332054615021, + "learning_rate": 9.35912305823974e-05, + "loss": 0.0108, + "step": 20480 + }, + { + "epoch": 3.1826654240447345, + "grad_norm": 0.17238345742225647, + "learning_rate": 9.358312920329101e-05, + "loss": 0.0099, + "step": 20490 + }, + { + "epoch": 3.1842187014600807, + "grad_norm": 0.24751286208629608, + "learning_rate": 9.357502305799914e-05, + "loss": 0.0121, + "step": 20500 + }, + { + "epoch": 3.1857719788754273, + "grad_norm": 0.18504881858825684, + "learning_rate": 9.356691214740824e-05, + "loss": 0.015, + "step": 20510 + }, + { + "epoch": 3.1873252562907735, + "grad_norm": 0.20581232011318207, + "learning_rate": 9.355879647240535e-05, + "loss": 0.0121, + "step": 20520 + }, + { + "epoch": 3.18887853370612, + "grad_norm": 0.22940124571323395, + "learning_rate": 9.355067603387798e-05, + "loss": 0.0135, + "step": 20530 + }, + { + "epoch": 3.1904318111214662, + "grad_norm": 0.21317455172538757, + "learning_rate": 9.354255083271412e-05, + "loss": 0.0113, + "step": 20540 + }, + { + "epoch": 3.191985088536813, + "grad_norm": 0.20817597210407257, + "learning_rate": 9.353442086980239e-05, + "loss": 0.0116, + "step": 20550 + }, + { + "epoch": 3.193538365952159, + "grad_norm": 0.17577233910560608, + "learning_rate": 9.352628614603185e-05, + "loss": 0.0139, + "step": 20560 + }, + { + "epoch": 3.1950916433675056, + "grad_norm": 0.31584134697914124, + "learning_rate": 9.351814666229209e-05, + "loss": 0.0132, + "step": 20570 + }, + { + "epoch": 3.196644920782852, + "grad_norm": 0.24763967096805573, + "learning_rate": 9.351000241947324e-05, + "loss": 0.0158, + "step": 20580 + }, + { + "epoch": 3.1981981981981984, + "grad_norm": 0.29731103777885437, + "learning_rate": 9.350185341846594e-05, + "loss": 0.0171, + "step": 20590 + }, + { + "epoch": 3.1997514756135446, + "grad_norm": 0.35281920433044434, + "learning_rate": 9.349369966016134e-05, + "loss": 0.0107, + "step": 20600 + }, + { + "epoch": 3.201304753028891, + "grad_norm": 0.29651114344596863, + "learning_rate": 9.348554114545117e-05, + "loss": 0.0116, + "step": 20610 + }, + { + "epoch": 3.2028580304442373, + "grad_norm": 0.28440821170806885, + "learning_rate": 9.347737787522758e-05, + "loss": 0.0135, + "step": 20620 + }, + { + "epoch": 3.2044113078595835, + "grad_norm": 0.2358158826828003, + "learning_rate": 9.346920985038332e-05, + "loss": 0.0111, + "step": 20630 + }, + { + "epoch": 3.20596458527493, + "grad_norm": 0.3200501501560211, + "learning_rate": 9.346103707181162e-05, + "loss": 0.0114, + "step": 20640 + }, + { + "epoch": 3.2075178626902763, + "grad_norm": 0.24987971782684326, + "learning_rate": 9.345285954040626e-05, + "loss": 0.0127, + "step": 20650 + }, + { + "epoch": 3.209071140105623, + "grad_norm": 0.35491499304771423, + "learning_rate": 9.34446772570615e-05, + "loss": 0.0114, + "step": 20660 + }, + { + "epoch": 3.210624417520969, + "grad_norm": 0.3705187141895294, + "learning_rate": 9.343649022267214e-05, + "loss": 0.013, + "step": 20670 + }, + { + "epoch": 3.2121776949363157, + "grad_norm": 0.40135708451271057, + "learning_rate": 9.342829843813353e-05, + "loss": 0.0098, + "step": 20680 + }, + { + "epoch": 3.213730972351662, + "grad_norm": 0.24173596501350403, + "learning_rate": 9.342010190434149e-05, + "loss": 0.0118, + "step": 20690 + }, + { + "epoch": 3.2152842497670084, + "grad_norm": 0.3063809871673584, + "learning_rate": 9.34119006221924e-05, + "loss": 0.0101, + "step": 20700 + }, + { + "epoch": 3.2168375271823546, + "grad_norm": 0.3397006690502167, + "learning_rate": 9.340369459258313e-05, + "loss": 0.0097, + "step": 20710 + }, + { + "epoch": 3.218390804597701, + "grad_norm": 0.20273754000663757, + "learning_rate": 9.339548381641106e-05, + "loss": 0.0096, + "step": 20720 + }, + { + "epoch": 3.2199440820130474, + "grad_norm": 0.3981679379940033, + "learning_rate": 9.338726829457413e-05, + "loss": 0.0112, + "step": 20730 + }, + { + "epoch": 3.221497359428394, + "grad_norm": 0.362443208694458, + "learning_rate": 9.337904802797078e-05, + "loss": 0.0112, + "step": 20740 + }, + { + "epoch": 3.22305063684374, + "grad_norm": 0.3804469704627991, + "learning_rate": 9.337082301749993e-05, + "loss": 0.0102, + "step": 20750 + }, + { + "epoch": 3.2246039142590868, + "grad_norm": 0.3034236431121826, + "learning_rate": 9.336259326406109e-05, + "loss": 0.0106, + "step": 20760 + }, + { + "epoch": 3.226157191674433, + "grad_norm": 0.339146226644516, + "learning_rate": 9.335435876855427e-05, + "loss": 0.0125, + "step": 20770 + }, + { + "epoch": 3.2277104690897795, + "grad_norm": 0.4225268065929413, + "learning_rate": 9.334611953187994e-05, + "loss": 0.0132, + "step": 20780 + }, + { + "epoch": 3.2292637465051257, + "grad_norm": 0.4951165020465851, + "learning_rate": 9.333787555493914e-05, + "loss": 0.0096, + "step": 20790 + }, + { + "epoch": 3.2308170239204723, + "grad_norm": 0.2666996419429779, + "learning_rate": 9.332962683863345e-05, + "loss": 0.0153, + "step": 20800 + }, + { + "epoch": 3.2323703013358185, + "grad_norm": 0.32522353529930115, + "learning_rate": 9.332137338386489e-05, + "loss": 0.01, + "step": 20810 + }, + { + "epoch": 3.233923578751165, + "grad_norm": 0.1960015445947647, + "learning_rate": 9.33131151915361e-05, + "loss": 0.0123, + "step": 20820 + }, + { + "epoch": 3.2354768561665113, + "grad_norm": 0.3885432183742523, + "learning_rate": 9.330485226255012e-05, + "loss": 0.0094, + "step": 20830 + }, + { + "epoch": 3.237030133581858, + "grad_norm": 0.31178832054138184, + "learning_rate": 9.329658459781061e-05, + "loss": 0.0125, + "step": 20840 + }, + { + "epoch": 3.238583410997204, + "grad_norm": 0.28320810198783875, + "learning_rate": 9.328831219822172e-05, + "loss": 0.0108, + "step": 20850 + }, + { + "epoch": 3.2401366884125506, + "grad_norm": 0.31837740540504456, + "learning_rate": 9.328003506468808e-05, + "loss": 0.0098, + "step": 20860 + }, + { + "epoch": 3.241689965827897, + "grad_norm": 0.20931397378444672, + "learning_rate": 9.327175319811488e-05, + "loss": 0.0108, + "step": 20870 + }, + { + "epoch": 3.2432432432432434, + "grad_norm": 0.2660629153251648, + "learning_rate": 9.326346659940781e-05, + "loss": 0.0118, + "step": 20880 + }, + { + "epoch": 3.2447965206585896, + "grad_norm": 0.39587876200675964, + "learning_rate": 9.325517526947308e-05, + "loss": 0.0095, + "step": 20890 + }, + { + "epoch": 3.246349798073936, + "grad_norm": 0.2595386505126953, + "learning_rate": 9.32468792092174e-05, + "loss": 0.0119, + "step": 20900 + }, + { + "epoch": 3.2479030754892824, + "grad_norm": 0.20803725719451904, + "learning_rate": 9.323857841954803e-05, + "loss": 0.01, + "step": 20910 + }, + { + "epoch": 3.2494563529046285, + "grad_norm": 0.7638009786605835, + "learning_rate": 9.323027290137276e-05, + "loss": 0.0136, + "step": 20920 + }, + { + "epoch": 3.251009630319975, + "grad_norm": 0.30796581506729126, + "learning_rate": 9.322196265559981e-05, + "loss": 0.0117, + "step": 20930 + }, + { + "epoch": 3.2525629077353218, + "grad_norm": 0.22048047184944153, + "learning_rate": 9.321364768313803e-05, + "loss": 0.0122, + "step": 20940 + }, + { + "epoch": 3.254116185150668, + "grad_norm": 0.21845073997974396, + "learning_rate": 9.32053279848967e-05, + "loss": 0.0112, + "step": 20950 + }, + { + "epoch": 3.255669462566014, + "grad_norm": 0.2943372130393982, + "learning_rate": 9.319700356178567e-05, + "loss": 0.0149, + "step": 20960 + }, + { + "epoch": 3.2572227399813607, + "grad_norm": 0.26217031478881836, + "learning_rate": 9.318867441471527e-05, + "loss": 0.015, + "step": 20970 + }, + { + "epoch": 3.258776017396707, + "grad_norm": 0.28291210532188416, + "learning_rate": 9.318034054459637e-05, + "loss": 0.0123, + "step": 20980 + }, + { + "epoch": 3.2603292948120535, + "grad_norm": 0.35096850991249084, + "learning_rate": 9.317200195234034e-05, + "loss": 0.0127, + "step": 20990 + }, + { + "epoch": 3.2618825722273996, + "grad_norm": 0.22059714794158936, + "learning_rate": 9.316365863885909e-05, + "loss": 0.0153, + "step": 21000 + }, + { + "epoch": 3.2634358496427462, + "grad_norm": 0.2634607255458832, + "learning_rate": 9.315531060506502e-05, + "loss": 0.0103, + "step": 21010 + }, + { + "epoch": 3.2649891270580924, + "grad_norm": 0.23862551152706146, + "learning_rate": 9.314695785187108e-05, + "loss": 0.0103, + "step": 21020 + }, + { + "epoch": 3.266542404473439, + "grad_norm": 0.3108724057674408, + "learning_rate": 9.313860038019069e-05, + "loss": 0.011, + "step": 21030 + }, + { + "epoch": 3.268095681888785, + "grad_norm": 0.3598577678203583, + "learning_rate": 9.313023819093782e-05, + "loss": 0.0114, + "step": 21040 + }, + { + "epoch": 3.269648959304132, + "grad_norm": 0.38132357597351074, + "learning_rate": 9.312187128502695e-05, + "loss": 0.0114, + "step": 21050 + }, + { + "epoch": 3.271202236719478, + "grad_norm": 0.42112982273101807, + "learning_rate": 9.311349966337307e-05, + "loss": 0.0143, + "step": 21060 + }, + { + "epoch": 3.2727555141348246, + "grad_norm": 0.27375665307044983, + "learning_rate": 9.310512332689169e-05, + "loss": 0.0106, + "step": 21070 + }, + { + "epoch": 3.2743087915501707, + "grad_norm": 0.19041644036769867, + "learning_rate": 9.309674227649883e-05, + "loss": 0.0133, + "step": 21080 + }, + { + "epoch": 3.2758620689655173, + "grad_norm": 0.337337464094162, + "learning_rate": 9.308835651311103e-05, + "loss": 0.0107, + "step": 21090 + }, + { + "epoch": 3.2774153463808635, + "grad_norm": 0.157277911901474, + "learning_rate": 9.307996603764533e-05, + "loss": 0.0119, + "step": 21100 + }, + { + "epoch": 3.27896862379621, + "grad_norm": 0.3370499014854431, + "learning_rate": 9.307157085101932e-05, + "loss": 0.0132, + "step": 21110 + }, + { + "epoch": 3.2805219012115563, + "grad_norm": 0.1741078943014145, + "learning_rate": 9.306317095415109e-05, + "loss": 0.0105, + "step": 21120 + }, + { + "epoch": 3.282075178626903, + "grad_norm": 0.2832064926624298, + "learning_rate": 9.305476634795922e-05, + "loss": 0.01, + "step": 21130 + }, + { + "epoch": 3.283628456042249, + "grad_norm": 0.6301528215408325, + "learning_rate": 9.304635703336284e-05, + "loss": 0.0104, + "step": 21140 + }, + { + "epoch": 3.2851817334575957, + "grad_norm": 0.22531495988368988, + "learning_rate": 9.303794301128157e-05, + "loss": 0.0123, + "step": 21150 + }, + { + "epoch": 3.286735010872942, + "grad_norm": 0.19159966707229614, + "learning_rate": 9.302952428263555e-05, + "loss": 0.0102, + "step": 21160 + }, + { + "epoch": 3.2882882882882885, + "grad_norm": 0.21782776713371277, + "learning_rate": 9.302110084834545e-05, + "loss": 0.015, + "step": 21170 + }, + { + "epoch": 3.2898415657036346, + "grad_norm": 0.5264155864715576, + "learning_rate": 9.301267270933245e-05, + "loss": 0.0119, + "step": 21180 + }, + { + "epoch": 3.2913948431189812, + "grad_norm": 0.9097236394882202, + "learning_rate": 9.300423986651823e-05, + "loss": 0.0112, + "step": 21190 + }, + { + "epoch": 3.2929481205343274, + "grad_norm": 0.30733874440193176, + "learning_rate": 9.299580232082501e-05, + "loss": 0.0126, + "step": 21200 + }, + { + "epoch": 3.294501397949674, + "grad_norm": 0.4566698968410492, + "learning_rate": 9.298736007317547e-05, + "loss": 0.0115, + "step": 21210 + }, + { + "epoch": 3.29605467536502, + "grad_norm": 0.3381621837615967, + "learning_rate": 9.297891312449288e-05, + "loss": 0.0097, + "step": 21220 + }, + { + "epoch": 3.297607952780367, + "grad_norm": 0.1842970997095108, + "learning_rate": 9.297046147570094e-05, + "loss": 0.0108, + "step": 21230 + }, + { + "epoch": 3.299161230195713, + "grad_norm": 0.3247500956058502, + "learning_rate": 9.296200512772396e-05, + "loss": 0.0126, + "step": 21240 + }, + { + "epoch": 3.300714507611059, + "grad_norm": 0.20950007438659668, + "learning_rate": 9.295354408148668e-05, + "loss": 0.0092, + "step": 21250 + }, + { + "epoch": 3.3022677850264057, + "grad_norm": 0.1678164005279541, + "learning_rate": 9.294507833791441e-05, + "loss": 0.011, + "step": 21260 + }, + { + "epoch": 3.3038210624417523, + "grad_norm": 0.31954076886177063, + "learning_rate": 9.293660789793295e-05, + "loss": 0.0136, + "step": 21270 + }, + { + "epoch": 3.3053743398570985, + "grad_norm": 0.3302403390407562, + "learning_rate": 9.292813276246858e-05, + "loss": 0.013, + "step": 21280 + }, + { + "epoch": 3.3069276172724447, + "grad_norm": 0.19310924410820007, + "learning_rate": 9.291965293244816e-05, + "loss": 0.0119, + "step": 21290 + }, + { + "epoch": 3.3084808946877913, + "grad_norm": 0.3959212303161621, + "learning_rate": 9.291116840879904e-05, + "loss": 0.0163, + "step": 21300 + }, + { + "epoch": 3.3100341721031374, + "grad_norm": 0.36570265889167786, + "learning_rate": 9.290267919244904e-05, + "loss": 0.0119, + "step": 21310 + }, + { + "epoch": 3.311587449518484, + "grad_norm": 0.2009919285774231, + "learning_rate": 9.289418528432655e-05, + "loss": 0.0107, + "step": 21320 + }, + { + "epoch": 3.31314072693383, + "grad_norm": 0.20176465809345245, + "learning_rate": 9.288568668536045e-05, + "loss": 0.0108, + "step": 21330 + }, + { + "epoch": 3.314694004349177, + "grad_norm": 0.2186717540025711, + "learning_rate": 9.287718339648013e-05, + "loss": 0.0083, + "step": 21340 + }, + { + "epoch": 3.316247281764523, + "grad_norm": 0.268530011177063, + "learning_rate": 9.28686754186155e-05, + "loss": 0.0129, + "step": 21350 + }, + { + "epoch": 3.3178005591798696, + "grad_norm": 0.3437795341014862, + "learning_rate": 9.286016275269698e-05, + "loss": 0.017, + "step": 21360 + }, + { + "epoch": 3.3193538365952158, + "grad_norm": 0.474628210067749, + "learning_rate": 9.285164539965551e-05, + "loss": 0.0137, + "step": 21370 + }, + { + "epoch": 3.3209071140105624, + "grad_norm": 0.2630443572998047, + "learning_rate": 9.284312336042251e-05, + "loss": 0.0095, + "step": 21380 + }, + { + "epoch": 3.3224603914259085, + "grad_norm": 0.33571526408195496, + "learning_rate": 9.283459663592996e-05, + "loss": 0.0103, + "step": 21390 + }, + { + "epoch": 3.324013668841255, + "grad_norm": 0.25545382499694824, + "learning_rate": 9.282606522711033e-05, + "loss": 0.0109, + "step": 21400 + }, + { + "epoch": 3.3255669462566013, + "grad_norm": 0.29619356989860535, + "learning_rate": 9.281752913489657e-05, + "loss": 0.0127, + "step": 21410 + }, + { + "epoch": 3.327120223671948, + "grad_norm": 0.4554675221443176, + "learning_rate": 9.280898836022222e-05, + "loss": 0.012, + "step": 21420 + }, + { + "epoch": 3.328673501087294, + "grad_norm": 0.2782634496688843, + "learning_rate": 9.280044290402126e-05, + "loss": 0.0107, + "step": 21430 + }, + { + "epoch": 3.3302267785026407, + "grad_norm": 0.20981848239898682, + "learning_rate": 9.279189276722821e-05, + "loss": 0.0124, + "step": 21440 + }, + { + "epoch": 3.331780055917987, + "grad_norm": 0.26505348086357117, + "learning_rate": 9.278333795077812e-05, + "loss": 0.0119, + "step": 21450 + }, + { + "epoch": 3.3333333333333335, + "grad_norm": 0.23100294172763824, + "learning_rate": 9.27747784556065e-05, + "loss": 0.0104, + "step": 21460 + }, + { + "epoch": 3.3348866107486796, + "grad_norm": 0.38956478238105774, + "learning_rate": 9.276621428264942e-05, + "loss": 0.0166, + "step": 21470 + }, + { + "epoch": 3.3364398881640263, + "grad_norm": 0.32060718536376953, + "learning_rate": 9.275764543284345e-05, + "loss": 0.0112, + "step": 21480 + }, + { + "epoch": 3.3379931655793724, + "grad_norm": 0.29849064350128174, + "learning_rate": 9.274907190712566e-05, + "loss": 0.0094, + "step": 21490 + }, + { + "epoch": 3.339546442994719, + "grad_norm": 0.23415283858776093, + "learning_rate": 9.274049370643363e-05, + "loss": 0.0126, + "step": 21500 + }, + { + "epoch": 3.341099720410065, + "grad_norm": 0.1659853160381317, + "learning_rate": 9.273191083170547e-05, + "loss": 0.0109, + "step": 21510 + }, + { + "epoch": 3.342652997825412, + "grad_norm": 0.1903878152370453, + "learning_rate": 9.27233232838798e-05, + "loss": 0.0101, + "step": 21520 + }, + { + "epoch": 3.344206275240758, + "grad_norm": 0.27012237906455994, + "learning_rate": 9.27147310638957e-05, + "loss": 0.0072, + "step": 21530 + }, + { + "epoch": 3.345759552656104, + "grad_norm": 0.20241659879684448, + "learning_rate": 9.270613417269286e-05, + "loss": 0.0135, + "step": 21540 + }, + { + "epoch": 3.3473128300714508, + "grad_norm": 0.11945385485887527, + "learning_rate": 9.269753261121138e-05, + "loss": 0.01, + "step": 21550 + }, + { + "epoch": 3.3488661074867974, + "grad_norm": 0.14638814330101013, + "learning_rate": 9.268892638039194e-05, + "loss": 0.0108, + "step": 21560 + }, + { + "epoch": 3.3504193849021435, + "grad_norm": 0.20786800980567932, + "learning_rate": 9.268031548117569e-05, + "loss": 0.0094, + "step": 21570 + }, + { + "epoch": 3.3519726623174897, + "grad_norm": 0.27537596225738525, + "learning_rate": 9.26716999145043e-05, + "loss": 0.0149, + "step": 21580 + }, + { + "epoch": 3.3535259397328363, + "grad_norm": 0.40828052163124084, + "learning_rate": 9.266307968131998e-05, + "loss": 0.0121, + "step": 21590 + }, + { + "epoch": 3.355079217148183, + "grad_norm": 0.31639373302459717, + "learning_rate": 9.26544547825654e-05, + "loss": 0.0093, + "step": 21600 + }, + { + "epoch": 3.356632494563529, + "grad_norm": 0.445125937461853, + "learning_rate": 9.264582521918376e-05, + "loss": 0.011, + "step": 21610 + }, + { + "epoch": 3.3581857719788752, + "grad_norm": 0.2243729829788208, + "learning_rate": 9.263719099211881e-05, + "loss": 0.0107, + "step": 21620 + }, + { + "epoch": 3.359739049394222, + "grad_norm": 0.3422379195690155, + "learning_rate": 9.262855210231476e-05, + "loss": 0.011, + "step": 21630 + }, + { + "epoch": 3.361292326809568, + "grad_norm": 0.26820704340934753, + "learning_rate": 9.261990855071633e-05, + "loss": 0.0095, + "step": 21640 + }, + { + "epoch": 3.3628456042249146, + "grad_norm": 0.3542826771736145, + "learning_rate": 9.261126033826878e-05, + "loss": 0.0104, + "step": 21650 + }, + { + "epoch": 3.364398881640261, + "grad_norm": 0.23100493848323822, + "learning_rate": 9.260260746591786e-05, + "loss": 0.0093, + "step": 21660 + }, + { + "epoch": 3.3659521590556074, + "grad_norm": 0.18831293284893036, + "learning_rate": 9.259394993460985e-05, + "loss": 0.0099, + "step": 21670 + }, + { + "epoch": 3.3675054364709536, + "grad_norm": 0.3103720545768738, + "learning_rate": 9.258528774529151e-05, + "loss": 0.0088, + "step": 21680 + }, + { + "epoch": 3.3690587138863, + "grad_norm": 0.26059380173683167, + "learning_rate": 9.257662089891013e-05, + "loss": 0.0118, + "step": 21690 + }, + { + "epoch": 3.3706119913016463, + "grad_norm": 0.30802983045578003, + "learning_rate": 9.25679493964135e-05, + "loss": 0.0111, + "step": 21700 + }, + { + "epoch": 3.372165268716993, + "grad_norm": 0.30463939905166626, + "learning_rate": 9.255927323874994e-05, + "loss": 0.0095, + "step": 21710 + }, + { + "epoch": 3.373718546132339, + "grad_norm": 0.3637387156486511, + "learning_rate": 9.255059242686822e-05, + "loss": 0.0108, + "step": 21720 + }, + { + "epoch": 3.3752718235476857, + "grad_norm": 0.4189542829990387, + "learning_rate": 9.254190696171769e-05, + "loss": 0.0103, + "step": 21730 + }, + { + "epoch": 3.376825100963032, + "grad_norm": 0.357070654630661, + "learning_rate": 9.25332168442482e-05, + "loss": 0.0096, + "step": 21740 + }, + { + "epoch": 3.3783783783783785, + "grad_norm": 0.26173993945121765, + "learning_rate": 9.252452207541004e-05, + "loss": 0.0103, + "step": 21750 + }, + { + "epoch": 3.3799316557937247, + "grad_norm": 0.33165764808654785, + "learning_rate": 9.251582265615409e-05, + "loss": 0.0122, + "step": 21760 + }, + { + "epoch": 3.3814849332090713, + "grad_norm": 0.4832715094089508, + "learning_rate": 9.250711858743169e-05, + "loss": 0.0103, + "step": 21770 + }, + { + "epoch": 3.3830382106244175, + "grad_norm": 0.30511802434921265, + "learning_rate": 9.24984098701947e-05, + "loss": 0.0107, + "step": 21780 + }, + { + "epoch": 3.384591488039764, + "grad_norm": 0.35127854347229004, + "learning_rate": 9.248969650539552e-05, + "loss": 0.0116, + "step": 21790 + }, + { + "epoch": 3.3861447654551102, + "grad_norm": 0.20866145193576813, + "learning_rate": 9.2480978493987e-05, + "loss": 0.0108, + "step": 21800 + }, + { + "epoch": 3.387698042870457, + "grad_norm": 0.33291566371917725, + "learning_rate": 9.247225583692256e-05, + "loss": 0.0128, + "step": 21810 + }, + { + "epoch": 3.389251320285803, + "grad_norm": 0.2970622777938843, + "learning_rate": 9.246352853515607e-05, + "loss": 0.0118, + "step": 21820 + }, + { + "epoch": 3.3908045977011496, + "grad_norm": 0.3660731613636017, + "learning_rate": 9.245479658964194e-05, + "loss": 0.0115, + "step": 21830 + }, + { + "epoch": 3.392357875116496, + "grad_norm": 0.30892401933670044, + "learning_rate": 9.244606000133507e-05, + "loss": 0.0075, + "step": 21840 + }, + { + "epoch": 3.3939111525318424, + "grad_norm": 0.38255640864372253, + "learning_rate": 9.24373187711909e-05, + "loss": 0.0116, + "step": 21850 + }, + { + "epoch": 3.3954644299471886, + "grad_norm": 0.25128892064094543, + "learning_rate": 9.242857290016537e-05, + "loss": 0.01, + "step": 21860 + }, + { + "epoch": 3.3970177073625347, + "grad_norm": 0.28193822503089905, + "learning_rate": 9.241982238921488e-05, + "loss": 0.0135, + "step": 21870 + }, + { + "epoch": 3.3985709847778813, + "grad_norm": 0.36392268538475037, + "learning_rate": 9.24110672392964e-05, + "loss": 0.0093, + "step": 21880 + }, + { + "epoch": 3.400124262193228, + "grad_norm": 0.30007627606391907, + "learning_rate": 9.240230745136737e-05, + "loss": 0.0119, + "step": 21890 + }, + { + "epoch": 3.401677539608574, + "grad_norm": 0.22436250746250153, + "learning_rate": 9.239354302638575e-05, + "loss": 0.0117, + "step": 21900 + }, + { + "epoch": 3.4032308170239203, + "grad_norm": 0.2664126753807068, + "learning_rate": 9.238477396531e-05, + "loss": 0.0096, + "step": 21910 + }, + { + "epoch": 3.404784094439267, + "grad_norm": 0.22797876596450806, + "learning_rate": 9.23760002690991e-05, + "loss": 0.0081, + "step": 21920 + }, + { + "epoch": 3.406337371854613, + "grad_norm": 0.6350820064544678, + "learning_rate": 9.236722193871252e-05, + "loss": 0.0127, + "step": 21930 + }, + { + "epoch": 3.4078906492699597, + "grad_norm": 0.22916360199451447, + "learning_rate": 9.235843897511023e-05, + "loss": 0.0106, + "step": 21940 + }, + { + "epoch": 3.409443926685306, + "grad_norm": 0.46882954239845276, + "learning_rate": 9.234965137925276e-05, + "loss": 0.0109, + "step": 21950 + }, + { + "epoch": 3.4109972041006524, + "grad_norm": 0.5030233860015869, + "learning_rate": 9.234085915210108e-05, + "loss": 0.0113, + "step": 21960 + }, + { + "epoch": 3.4125504815159986, + "grad_norm": 0.2769275903701782, + "learning_rate": 9.23320622946167e-05, + "loss": 0.0098, + "step": 21970 + }, + { + "epoch": 3.414103758931345, + "grad_norm": 0.177702859044075, + "learning_rate": 9.232326080776163e-05, + "loss": 0.0136, + "step": 21980 + }, + { + "epoch": 3.4156570363466914, + "grad_norm": 0.3752250373363495, + "learning_rate": 9.23144546924984e-05, + "loss": 0.0104, + "step": 21990 + }, + { + "epoch": 3.417210313762038, + "grad_norm": 0.2277209460735321, + "learning_rate": 9.230564394979e-05, + "loss": 0.0088, + "step": 22000 + }, + { + "epoch": 3.418763591177384, + "grad_norm": 0.3153318762779236, + "learning_rate": 9.22968285806e-05, + "loss": 0.0115, + "step": 22010 + }, + { + "epoch": 3.4203168685927308, + "grad_norm": 0.3879602551460266, + "learning_rate": 9.228800858589242e-05, + "loss": 0.0097, + "step": 22020 + }, + { + "epoch": 3.421870146008077, + "grad_norm": 0.24705475568771362, + "learning_rate": 9.227918396663179e-05, + "loss": 0.0125, + "step": 22030 + }, + { + "epoch": 3.4234234234234235, + "grad_norm": 0.22815905511379242, + "learning_rate": 9.227035472378319e-05, + "loss": 0.0113, + "step": 22040 + }, + { + "epoch": 3.4249767008387697, + "grad_norm": 0.3003162145614624, + "learning_rate": 9.226152085831213e-05, + "loss": 0.01, + "step": 22050 + }, + { + "epoch": 3.4265299782541163, + "grad_norm": 0.1404826045036316, + "learning_rate": 9.22526823711847e-05, + "loss": 0.0084, + "step": 22060 + }, + { + "epoch": 3.4280832556694625, + "grad_norm": 0.17410433292388916, + "learning_rate": 9.224383926336745e-05, + "loss": 0.0117, + "step": 22070 + }, + { + "epoch": 3.429636533084809, + "grad_norm": 0.3394884169101715, + "learning_rate": 9.223499153582744e-05, + "loss": 0.0086, + "step": 22080 + }, + { + "epoch": 3.4311898105001553, + "grad_norm": 0.1363336741924286, + "learning_rate": 9.222613918953226e-05, + "loss": 0.0102, + "step": 22090 + }, + { + "epoch": 3.432743087915502, + "grad_norm": 0.19305124878883362, + "learning_rate": 9.221728222544999e-05, + "loss": 0.0103, + "step": 22100 + }, + { + "epoch": 3.434296365330848, + "grad_norm": 0.3223602771759033, + "learning_rate": 9.22084206445492e-05, + "loss": 0.011, + "step": 22110 + }, + { + "epoch": 3.4358496427461946, + "grad_norm": 0.23286935687065125, + "learning_rate": 9.2199554447799e-05, + "loss": 0.0111, + "step": 22120 + }, + { + "epoch": 3.437402920161541, + "grad_norm": 0.23369406163692474, + "learning_rate": 9.219068363616897e-05, + "loss": 0.0105, + "step": 22130 + }, + { + "epoch": 3.4389561975768874, + "grad_norm": 0.21121846139431, + "learning_rate": 9.218180821062919e-05, + "loss": 0.0078, + "step": 22140 + }, + { + "epoch": 3.4405094749922336, + "grad_norm": 0.2656073272228241, + "learning_rate": 9.21729281721503e-05, + "loss": 0.0099, + "step": 22150 + }, + { + "epoch": 3.4420627524075798, + "grad_norm": 0.15762197971343994, + "learning_rate": 9.216404352170339e-05, + "loss": 0.0075, + "step": 22160 + }, + { + "epoch": 3.4436160298229264, + "grad_norm": 0.22662512958049774, + "learning_rate": 9.215515426026007e-05, + "loss": 0.0081, + "step": 22170 + }, + { + "epoch": 3.445169307238273, + "grad_norm": 0.20441512763500214, + "learning_rate": 9.214626038879246e-05, + "loss": 0.0084, + "step": 22180 + }, + { + "epoch": 3.446722584653619, + "grad_norm": 0.2743140757083893, + "learning_rate": 9.21373619082732e-05, + "loss": 0.0099, + "step": 22190 + }, + { + "epoch": 3.4482758620689653, + "grad_norm": 0.3131599426269531, + "learning_rate": 9.212845881967535e-05, + "loss": 0.01, + "step": 22200 + }, + { + "epoch": 3.449829139484312, + "grad_norm": 0.2371591180562973, + "learning_rate": 9.211955112397262e-05, + "loss": 0.0082, + "step": 22210 + }, + { + "epoch": 3.4513824168996585, + "grad_norm": 0.3772068917751312, + "learning_rate": 9.211063882213909e-05, + "loss": 0.0146, + "step": 22220 + }, + { + "epoch": 3.4529356943150047, + "grad_norm": 0.3878268599510193, + "learning_rate": 9.210172191514942e-05, + "loss": 0.0143, + "step": 22230 + }, + { + "epoch": 3.454488971730351, + "grad_norm": 0.23504343628883362, + "learning_rate": 9.209280040397874e-05, + "loss": 0.0123, + "step": 22240 + }, + { + "epoch": 3.4560422491456975, + "grad_norm": 0.22084349393844604, + "learning_rate": 9.208387428960268e-05, + "loss": 0.0094, + "step": 22250 + }, + { + "epoch": 3.4575955265610436, + "grad_norm": 0.45775556564331055, + "learning_rate": 9.20749435729974e-05, + "loss": 0.012, + "step": 22260 + }, + { + "epoch": 3.4591488039763902, + "grad_norm": 0.28957223892211914, + "learning_rate": 9.206600825513957e-05, + "loss": 0.0093, + "step": 22270 + }, + { + "epoch": 3.4607020813917364, + "grad_norm": 0.35902857780456543, + "learning_rate": 9.20570683370063e-05, + "loss": 0.0122, + "step": 22280 + }, + { + "epoch": 3.462255358807083, + "grad_norm": 0.15304149687290192, + "learning_rate": 9.204812381957528e-05, + "loss": 0.009, + "step": 22290 + }, + { + "epoch": 3.463808636222429, + "grad_norm": 0.26230543851852417, + "learning_rate": 9.203917470382465e-05, + "loss": 0.008, + "step": 22300 + }, + { + "epoch": 3.465361913637776, + "grad_norm": 0.27568763494491577, + "learning_rate": 9.203022099073309e-05, + "loss": 0.0091, + "step": 22310 + }, + { + "epoch": 3.466915191053122, + "grad_norm": 0.27380502223968506, + "learning_rate": 9.202126268127976e-05, + "loss": 0.01, + "step": 22320 + }, + { + "epoch": 3.4684684684684686, + "grad_norm": 0.39896082878112793, + "learning_rate": 9.20122997764443e-05, + "loss": 0.0124, + "step": 22330 + }, + { + "epoch": 3.4700217458838147, + "grad_norm": 0.22949884831905365, + "learning_rate": 9.200333227720692e-05, + "loss": 0.0098, + "step": 22340 + }, + { + "epoch": 3.4715750232991613, + "grad_norm": 0.2954805791378021, + "learning_rate": 9.199436018454826e-05, + "loss": 0.0117, + "step": 22350 + }, + { + "epoch": 3.4731283007145075, + "grad_norm": 0.3703261911869049, + "learning_rate": 9.198538349944952e-05, + "loss": 0.0123, + "step": 22360 + }, + { + "epoch": 3.474681578129854, + "grad_norm": 0.26493215560913086, + "learning_rate": 9.197640222289234e-05, + "loss": 0.0132, + "step": 22370 + }, + { + "epoch": 3.4762348555452003, + "grad_norm": 0.36670246720314026, + "learning_rate": 9.196741635585895e-05, + "loss": 0.013, + "step": 22380 + }, + { + "epoch": 3.477788132960547, + "grad_norm": 0.23658829927444458, + "learning_rate": 9.195842589933199e-05, + "loss": 0.0123, + "step": 22390 + }, + { + "epoch": 3.479341410375893, + "grad_norm": 0.1914675384759903, + "learning_rate": 9.194943085429466e-05, + "loss": 0.0133, + "step": 22400 + }, + { + "epoch": 3.4808946877912397, + "grad_norm": 0.2732207477092743, + "learning_rate": 9.194043122173065e-05, + "loss": 0.0153, + "step": 22410 + }, + { + "epoch": 3.482447965206586, + "grad_norm": 0.3779652416706085, + "learning_rate": 9.193142700262413e-05, + "loss": 0.0098, + "step": 22420 + }, + { + "epoch": 3.4840012426219324, + "grad_norm": 0.38273313641548157, + "learning_rate": 9.192241819795979e-05, + "loss": 0.0102, + "step": 22430 + }, + { + "epoch": 3.4855545200372786, + "grad_norm": 0.2951861321926117, + "learning_rate": 9.191340480872284e-05, + "loss": 0.0106, + "step": 22440 + }, + { + "epoch": 3.4871077974526252, + "grad_norm": 0.30607280135154724, + "learning_rate": 9.190438683589895e-05, + "loss": 0.0166, + "step": 22450 + }, + { + "epoch": 3.4886610748679714, + "grad_norm": 0.42411115765571594, + "learning_rate": 9.189536428047432e-05, + "loss": 0.0113, + "step": 22460 + }, + { + "epoch": 3.490214352283318, + "grad_norm": 0.41566258668899536, + "learning_rate": 9.188633714343564e-05, + "loss": 0.0092, + "step": 22470 + }, + { + "epoch": 3.491767629698664, + "grad_norm": 0.2075221687555313, + "learning_rate": 9.18773054257701e-05, + "loss": 0.0085, + "step": 22480 + }, + { + "epoch": 3.4933209071140103, + "grad_norm": 0.5350593328475952, + "learning_rate": 9.18682691284654e-05, + "loss": 0.0126, + "step": 22490 + }, + { + "epoch": 3.494874184529357, + "grad_norm": 0.28852379322052, + "learning_rate": 9.185922825250974e-05, + "loss": 0.0111, + "step": 22500 + }, + { + "epoch": 3.4964274619447036, + "grad_norm": 0.28343185782432556, + "learning_rate": 9.185018279889181e-05, + "loss": 0.01, + "step": 22510 + }, + { + "epoch": 3.4979807393600497, + "grad_norm": 0.1941060870885849, + "learning_rate": 9.184113276860082e-05, + "loss": 0.011, + "step": 22520 + }, + { + "epoch": 3.499534016775396, + "grad_norm": 0.4355437159538269, + "learning_rate": 9.183207816262645e-05, + "loss": 0.0109, + "step": 22530 + }, + { + "epoch": 3.5010872941907425, + "grad_norm": 0.15806297957897186, + "learning_rate": 9.182301898195891e-05, + "loss": 0.0092, + "step": 22540 + }, + { + "epoch": 3.502640571606089, + "grad_norm": 0.29890474677085876, + "learning_rate": 9.181395522758889e-05, + "loss": 0.0104, + "step": 22550 + }, + { + "epoch": 3.5041938490214353, + "grad_norm": 0.23715123534202576, + "learning_rate": 9.180488690050759e-05, + "loss": 0.0109, + "step": 22560 + }, + { + "epoch": 3.5057471264367814, + "grad_norm": 0.34035852551460266, + "learning_rate": 9.179581400170671e-05, + "loss": 0.0096, + "step": 22570 + }, + { + "epoch": 3.507300403852128, + "grad_norm": 0.247608482837677, + "learning_rate": 9.178673653217845e-05, + "loss": 0.01, + "step": 22580 + }, + { + "epoch": 3.508853681267474, + "grad_norm": 0.20027774572372437, + "learning_rate": 9.177765449291551e-05, + "loss": 0.0115, + "step": 22590 + }, + { + "epoch": 3.510406958682821, + "grad_norm": 0.28895822167396545, + "learning_rate": 9.176856788491109e-05, + "loss": 0.0114, + "step": 22600 + }, + { + "epoch": 3.511960236098167, + "grad_norm": 0.5804033279418945, + "learning_rate": 9.175947670915887e-05, + "loss": 0.01, + "step": 22610 + }, + { + "epoch": 3.5135135135135136, + "grad_norm": 0.3269624412059784, + "learning_rate": 9.175038096665309e-05, + "loss": 0.0104, + "step": 22620 + }, + { + "epoch": 3.5150667909288598, + "grad_norm": 0.24534666538238525, + "learning_rate": 9.17412806583884e-05, + "loss": 0.008, + "step": 22630 + }, + { + "epoch": 3.5166200683442064, + "grad_norm": 0.18567027151584625, + "learning_rate": 9.173217578536002e-05, + "loss": 0.0099, + "step": 22640 + }, + { + "epoch": 3.5181733457595525, + "grad_norm": 0.6278121471405029, + "learning_rate": 9.172306634856362e-05, + "loss": 0.0105, + "step": 22650 + }, + { + "epoch": 3.519726623174899, + "grad_norm": 0.3402174115180969, + "learning_rate": 9.171395234899545e-05, + "loss": 0.0106, + "step": 22660 + }, + { + "epoch": 3.5212799005902453, + "grad_norm": 0.23403280973434448, + "learning_rate": 9.170483378765214e-05, + "loss": 0.0107, + "step": 22670 + }, + { + "epoch": 3.522833178005592, + "grad_norm": 0.475827693939209, + "learning_rate": 9.169571066553091e-05, + "loss": 0.014, + "step": 22680 + }, + { + "epoch": 3.524386455420938, + "grad_norm": 0.2102651447057724, + "learning_rate": 9.168658298362946e-05, + "loss": 0.0118, + "step": 22690 + }, + { + "epoch": 3.5259397328362847, + "grad_norm": 0.37682604789733887, + "learning_rate": 9.167745074294598e-05, + "loss": 0.0136, + "step": 22700 + }, + { + "epoch": 3.527493010251631, + "grad_norm": 0.4006369709968567, + "learning_rate": 9.166831394447913e-05, + "loss": 0.012, + "step": 22710 + }, + { + "epoch": 3.5290462876669775, + "grad_norm": 0.19253984093666077, + "learning_rate": 9.165917258922812e-05, + "loss": 0.0094, + "step": 22720 + }, + { + "epoch": 3.5305995650823236, + "grad_norm": 0.32695257663726807, + "learning_rate": 9.165002667819262e-05, + "loss": 0.0108, + "step": 22730 + }, + { + "epoch": 3.53215284249767, + "grad_norm": 0.28626227378845215, + "learning_rate": 9.164087621237282e-05, + "loss": 0.0141, + "step": 22740 + }, + { + "epoch": 3.5337061199130164, + "grad_norm": 0.39707109332084656, + "learning_rate": 9.163172119276942e-05, + "loss": 0.0088, + "step": 22750 + }, + { + "epoch": 3.535259397328363, + "grad_norm": 0.20582641661167145, + "learning_rate": 9.162256162038358e-05, + "loss": 0.0132, + "step": 22760 + }, + { + "epoch": 3.536812674743709, + "grad_norm": 0.2306606024503708, + "learning_rate": 9.161339749621698e-05, + "loss": 0.0167, + "step": 22770 + }, + { + "epoch": 3.5383659521590554, + "grad_norm": 0.47989779710769653, + "learning_rate": 9.160422882127177e-05, + "loss": 0.0117, + "step": 22780 + }, + { + "epoch": 3.539919229574402, + "grad_norm": 0.3749353587627411, + "learning_rate": 9.159505559655069e-05, + "loss": 0.013, + "step": 22790 + }, + { + "epoch": 3.5414725069897486, + "grad_norm": 0.373311311006546, + "learning_rate": 9.158587782305684e-05, + "loss": 0.0126, + "step": 22800 + }, + { + "epoch": 3.5430257844050947, + "grad_norm": 0.40823888778686523, + "learning_rate": 9.157669550179391e-05, + "loss": 0.0127, + "step": 22810 + }, + { + "epoch": 3.544579061820441, + "grad_norm": 0.46002185344696045, + "learning_rate": 9.156750863376609e-05, + "loss": 0.0112, + "step": 22820 + }, + { + "epoch": 3.5461323392357875, + "grad_norm": 0.16695807874202728, + "learning_rate": 9.155831721997801e-05, + "loss": 0.0122, + "step": 22830 + }, + { + "epoch": 3.547685616651134, + "grad_norm": 0.23942841589450836, + "learning_rate": 9.154912126143484e-05, + "loss": 0.01, + "step": 22840 + }, + { + "epoch": 3.5492388940664803, + "grad_norm": 0.20609937608242035, + "learning_rate": 9.153992075914224e-05, + "loss": 0.0115, + "step": 22850 + }, + { + "epoch": 3.5507921714818265, + "grad_norm": 0.4051339030265808, + "learning_rate": 9.153071571410635e-05, + "loss": 0.0107, + "step": 22860 + }, + { + "epoch": 3.552345448897173, + "grad_norm": 0.37442728877067566, + "learning_rate": 9.152150612733384e-05, + "loss": 0.0098, + "step": 22870 + }, + { + "epoch": 3.5538987263125197, + "grad_norm": 0.4696243405342102, + "learning_rate": 9.151229199983184e-05, + "loss": 0.0126, + "step": 22880 + }, + { + "epoch": 3.555452003727866, + "grad_norm": 0.28631311655044556, + "learning_rate": 9.150307333260802e-05, + "loss": 0.0131, + "step": 22890 + }, + { + "epoch": 3.557005281143212, + "grad_norm": 0.29137611389160156, + "learning_rate": 9.149385012667048e-05, + "loss": 0.0112, + "step": 22900 + }, + { + "epoch": 3.5585585585585586, + "grad_norm": 0.2950848639011383, + "learning_rate": 9.148462238302788e-05, + "loss": 0.0133, + "step": 22910 + }, + { + "epoch": 3.560111835973905, + "grad_norm": 0.21781034767627716, + "learning_rate": 9.147539010268936e-05, + "loss": 0.0104, + "step": 22920 + }, + { + "epoch": 3.5616651133892514, + "grad_norm": 0.25956761837005615, + "learning_rate": 9.14661532866645e-05, + "loss": 0.0095, + "step": 22930 + }, + { + "epoch": 3.5632183908045976, + "grad_norm": 0.22973224520683289, + "learning_rate": 9.145691193596348e-05, + "loss": 0.0114, + "step": 22940 + }, + { + "epoch": 3.564771668219944, + "grad_norm": 0.8476009368896484, + "learning_rate": 9.144766605159691e-05, + "loss": 0.0147, + "step": 22950 + }, + { + "epoch": 3.5663249456352903, + "grad_norm": 0.30230191349983215, + "learning_rate": 9.14384156345759e-05, + "loss": 0.0121, + "step": 22960 + }, + { + "epoch": 3.567878223050637, + "grad_norm": 0.27737027406692505, + "learning_rate": 9.142916068591204e-05, + "loss": 0.0099, + "step": 22970 + }, + { + "epoch": 3.569431500465983, + "grad_norm": 0.25020715594291687, + "learning_rate": 9.141990120661746e-05, + "loss": 0.0086, + "step": 22980 + }, + { + "epoch": 3.5709847778813297, + "grad_norm": 0.23484109342098236, + "learning_rate": 9.141063719770475e-05, + "loss": 0.0116, + "step": 22990 + }, + { + "epoch": 3.572538055296676, + "grad_norm": 0.1574685424566269, + "learning_rate": 9.140136866018704e-05, + "loss": 0.0099, + "step": 23000 + }, + { + "epoch": 3.5740913327120225, + "grad_norm": 0.2605140209197998, + "learning_rate": 9.139209559507788e-05, + "loss": 0.0112, + "step": 23010 + }, + { + "epoch": 3.5756446101273687, + "grad_norm": 0.230982705950737, + "learning_rate": 9.13828180033914e-05, + "loss": 0.0096, + "step": 23020 + }, + { + "epoch": 3.5771978875427153, + "grad_norm": 0.2242240458726883, + "learning_rate": 9.137353588614212e-05, + "loss": 0.0122, + "step": 23030 + }, + { + "epoch": 3.5787511649580614, + "grad_norm": 0.26983344554901123, + "learning_rate": 9.136424924434519e-05, + "loss": 0.0105, + "step": 23040 + }, + { + "epoch": 3.580304442373408, + "grad_norm": 0.31141388416290283, + "learning_rate": 9.135495807901615e-05, + "loss": 0.0142, + "step": 23050 + }, + { + "epoch": 3.5818577197887542, + "grad_norm": 0.2500792145729065, + "learning_rate": 9.134566239117108e-05, + "loss": 0.0122, + "step": 23060 + }, + { + "epoch": 3.5834109972041004, + "grad_norm": 0.40980759263038635, + "learning_rate": 9.13363621818265e-05, + "loss": 0.0139, + "step": 23070 + }, + { + "epoch": 3.584964274619447, + "grad_norm": 0.3037922978401184, + "learning_rate": 9.132705745199953e-05, + "loss": 0.0105, + "step": 23080 + }, + { + "epoch": 3.5865175520347936, + "grad_norm": 0.2150699347257614, + "learning_rate": 9.131774820270768e-05, + "loss": 0.0092, + "step": 23090 + }, + { + "epoch": 3.5880708294501398, + "grad_norm": 0.3410148620605469, + "learning_rate": 9.130843443496901e-05, + "loss": 0.0121, + "step": 23100 + }, + { + "epoch": 3.589624106865486, + "grad_norm": 0.40345051884651184, + "learning_rate": 9.129911614980206e-05, + "loss": 0.0121, + "step": 23110 + }, + { + "epoch": 3.5911773842808326, + "grad_norm": 0.2773008942604065, + "learning_rate": 9.128979334822584e-05, + "loss": 0.0126, + "step": 23120 + }, + { + "epoch": 3.592730661696179, + "grad_norm": 0.30397769808769226, + "learning_rate": 9.128046603125992e-05, + "loss": 0.0107, + "step": 23130 + }, + { + "epoch": 3.5942839391115253, + "grad_norm": 0.37290629744529724, + "learning_rate": 9.12711341999243e-05, + "loss": 0.0114, + "step": 23140 + }, + { + "epoch": 3.5958372165268715, + "grad_norm": 0.3458348214626312, + "learning_rate": 9.12617978552395e-05, + "loss": 0.0131, + "step": 23150 + }, + { + "epoch": 3.597390493942218, + "grad_norm": 0.22297339141368866, + "learning_rate": 9.12524569982265e-05, + "loss": 0.0119, + "step": 23160 + }, + { + "epoch": 3.5989437713575647, + "grad_norm": 0.25473958253860474, + "learning_rate": 9.124311162990684e-05, + "loss": 0.01, + "step": 23170 + }, + { + "epoch": 3.600497048772911, + "grad_norm": 0.3022042512893677, + "learning_rate": 9.12337617513025e-05, + "loss": 0.0137, + "step": 23180 + }, + { + "epoch": 3.602050326188257, + "grad_norm": 0.1634209156036377, + "learning_rate": 9.122440736343596e-05, + "loss": 0.0082, + "step": 23190 + }, + { + "epoch": 3.6036036036036037, + "grad_norm": 0.29381972551345825, + "learning_rate": 9.12150484673302e-05, + "loss": 0.0099, + "step": 23200 + }, + { + "epoch": 3.60515688101895, + "grad_norm": 0.49219879508018494, + "learning_rate": 9.120568506400873e-05, + "loss": 0.0168, + "step": 23210 + }, + { + "epoch": 3.6067101584342964, + "grad_norm": 0.2669578492641449, + "learning_rate": 9.119631715449548e-05, + "loss": 0.0107, + "step": 23220 + }, + { + "epoch": 3.6082634358496426, + "grad_norm": 0.3711036145687103, + "learning_rate": 9.118694473981493e-05, + "loss": 0.0144, + "step": 23230 + }, + { + "epoch": 3.609816713264989, + "grad_norm": 0.20045976340770721, + "learning_rate": 9.117756782099203e-05, + "loss": 0.0138, + "step": 23240 + }, + { + "epoch": 3.6113699906803354, + "grad_norm": 0.26282578706741333, + "learning_rate": 9.11681863990522e-05, + "loss": 0.0111, + "step": 23250 + }, + { + "epoch": 3.612923268095682, + "grad_norm": 0.32370829582214355, + "learning_rate": 9.115880047502142e-05, + "loss": 0.014, + "step": 23260 + }, + { + "epoch": 3.614476545511028, + "grad_norm": 0.24864086508750916, + "learning_rate": 9.114941004992609e-05, + "loss": 0.01, + "step": 23270 + }, + { + "epoch": 3.6160298229263748, + "grad_norm": 0.28307145833969116, + "learning_rate": 9.114001512479317e-05, + "loss": 0.009, + "step": 23280 + }, + { + "epoch": 3.617583100341721, + "grad_norm": 0.40682825446128845, + "learning_rate": 9.113061570065003e-05, + "loss": 0.0149, + "step": 23290 + }, + { + "epoch": 3.6191363777570675, + "grad_norm": 0.17764389514923096, + "learning_rate": 9.112121177852459e-05, + "loss": 0.0089, + "step": 23300 + }, + { + "epoch": 3.6206896551724137, + "grad_norm": 0.2686454653739929, + "learning_rate": 9.111180335944527e-05, + "loss": 0.0074, + "step": 23310 + }, + { + "epoch": 3.6222429325877603, + "grad_norm": 0.3231373429298401, + "learning_rate": 9.110239044444093e-05, + "loss": 0.0102, + "step": 23320 + }, + { + "epoch": 3.6237962100031065, + "grad_norm": 0.9655479788780212, + "learning_rate": 9.109297303454099e-05, + "loss": 0.0122, + "step": 23330 + }, + { + "epoch": 3.625349487418453, + "grad_norm": 0.18100644648075104, + "learning_rate": 9.108355113077526e-05, + "loss": 0.0075, + "step": 23340 + }, + { + "epoch": 3.6269027648337993, + "grad_norm": 0.18114757537841797, + "learning_rate": 9.107412473417419e-05, + "loss": 0.0085, + "step": 23350 + }, + { + "epoch": 3.628456042249146, + "grad_norm": 0.24247650802135468, + "learning_rate": 9.106469384576858e-05, + "loss": 0.0105, + "step": 23360 + }, + { + "epoch": 3.630009319664492, + "grad_norm": 0.25328323245048523, + "learning_rate": 9.105525846658978e-05, + "loss": 0.0113, + "step": 23370 + }, + { + "epoch": 3.6315625970798386, + "grad_norm": 0.2409656047821045, + "learning_rate": 9.104581859766965e-05, + "loss": 0.0113, + "step": 23380 + }, + { + "epoch": 3.633115874495185, + "grad_norm": 0.21672822535037994, + "learning_rate": 9.10363742400405e-05, + "loss": 0.0099, + "step": 23390 + }, + { + "epoch": 3.634669151910531, + "grad_norm": 0.3111449182033539, + "learning_rate": 9.102692539473518e-05, + "loss": 0.0076, + "step": 23400 + }, + { + "epoch": 3.6362224293258776, + "grad_norm": 0.21536919474601746, + "learning_rate": 9.101747206278697e-05, + "loss": 0.0126, + "step": 23410 + }, + { + "epoch": 3.637775706741224, + "grad_norm": 0.21603921055793762, + "learning_rate": 9.100801424522968e-05, + "loss": 0.0105, + "step": 23420 + }, + { + "epoch": 3.6393289841565704, + "grad_norm": 0.2872626483440399, + "learning_rate": 9.099855194309762e-05, + "loss": 0.01, + "step": 23430 + }, + { + "epoch": 3.6408822615719165, + "grad_norm": 0.13094943761825562, + "learning_rate": 9.098908515742554e-05, + "loss": 0.0092, + "step": 23440 + }, + { + "epoch": 3.642435538987263, + "grad_norm": 0.24070437252521515, + "learning_rate": 9.097961388924873e-05, + "loss": 0.0117, + "step": 23450 + }, + { + "epoch": 3.6439888164026097, + "grad_norm": 0.4158906638622284, + "learning_rate": 9.097013813960298e-05, + "loss": 0.0103, + "step": 23460 + }, + { + "epoch": 3.645542093817956, + "grad_norm": 0.3558915853500366, + "learning_rate": 9.09606579095245e-05, + "loss": 0.0101, + "step": 23470 + }, + { + "epoch": 3.647095371233302, + "grad_norm": 0.35797256231307983, + "learning_rate": 9.095117320005008e-05, + "loss": 0.0139, + "step": 23480 + }, + { + "epoch": 3.6486486486486487, + "grad_norm": 0.260377436876297, + "learning_rate": 9.094168401221691e-05, + "loss": 0.0113, + "step": 23490 + }, + { + "epoch": 3.6502019260639953, + "grad_norm": 0.22952868044376373, + "learning_rate": 9.093219034706273e-05, + "loss": 0.0112, + "step": 23500 + }, + { + "epoch": 3.6517552034793415, + "grad_norm": 0.33189037442207336, + "learning_rate": 9.092269220562577e-05, + "loss": 0.0115, + "step": 23510 + }, + { + "epoch": 3.6533084808946876, + "grad_norm": 0.2013804316520691, + "learning_rate": 9.09131895889447e-05, + "loss": 0.0133, + "step": 23520 + }, + { + "epoch": 3.6548617583100342, + "grad_norm": 0.34864991903305054, + "learning_rate": 9.090368249805873e-05, + "loss": 0.0114, + "step": 23530 + }, + { + "epoch": 3.6564150357253804, + "grad_norm": 0.24121619760990143, + "learning_rate": 9.089417093400754e-05, + "loss": 0.0143, + "step": 23540 + }, + { + "epoch": 3.657968313140727, + "grad_norm": 0.1507536768913269, + "learning_rate": 9.088465489783131e-05, + "loss": 0.0083, + "step": 23550 + }, + { + "epoch": 3.659521590556073, + "grad_norm": 0.20920392870903015, + "learning_rate": 9.087513439057068e-05, + "loss": 0.01, + "step": 23560 + }, + { + "epoch": 3.66107486797142, + "grad_norm": 0.3523212671279907, + "learning_rate": 9.08656094132668e-05, + "loss": 0.0084, + "step": 23570 + }, + { + "epoch": 3.662628145386766, + "grad_norm": 0.3292141258716583, + "learning_rate": 9.085607996696134e-05, + "loss": 0.0111, + "step": 23580 + }, + { + "epoch": 3.6641814228021126, + "grad_norm": 0.2696448862552643, + "learning_rate": 9.084654605269639e-05, + "loss": 0.0096, + "step": 23590 + }, + { + "epoch": 3.6657347002174587, + "grad_norm": 0.311996191740036, + "learning_rate": 9.083700767151457e-05, + "loss": 0.0108, + "step": 23600 + }, + { + "epoch": 3.6672879776328053, + "grad_norm": 0.3434617519378662, + "learning_rate": 9.082746482445898e-05, + "loss": 0.0108, + "step": 23610 + }, + { + "epoch": 3.6688412550481515, + "grad_norm": 0.47602859139442444, + "learning_rate": 9.081791751257325e-05, + "loss": 0.0118, + "step": 23620 + }, + { + "epoch": 3.670394532463498, + "grad_norm": 0.42373883724212646, + "learning_rate": 9.080836573690142e-05, + "loss": 0.0104, + "step": 23630 + }, + { + "epoch": 3.6719478098788443, + "grad_norm": 0.3963583707809448, + "learning_rate": 9.079880949848805e-05, + "loss": 0.0131, + "step": 23640 + }, + { + "epoch": 3.673501087294191, + "grad_norm": 0.5709888339042664, + "learning_rate": 9.078924879837822e-05, + "loss": 0.0127, + "step": 23650 + }, + { + "epoch": 3.675054364709537, + "grad_norm": 0.19206377863883972, + "learning_rate": 9.077968363761747e-05, + "loss": 0.0112, + "step": 23660 + }, + { + "epoch": 3.6766076421248837, + "grad_norm": 0.4502418041229248, + "learning_rate": 9.077011401725182e-05, + "loss": 0.013, + "step": 23670 + }, + { + "epoch": 3.67816091954023, + "grad_norm": 0.5531956553459167, + "learning_rate": 9.07605399383278e-05, + "loss": 0.011, + "step": 23680 + }, + { + "epoch": 3.679714196955576, + "grad_norm": 0.37876084446907043, + "learning_rate": 9.075096140189243e-05, + "loss": 0.012, + "step": 23690 + }, + { + "epoch": 3.6812674743709226, + "grad_norm": 0.32745423913002014, + "learning_rate": 9.074137840899318e-05, + "loss": 0.0143, + "step": 23700 + }, + { + "epoch": 3.682820751786269, + "grad_norm": 0.2831701338291168, + "learning_rate": 9.073179096067804e-05, + "loss": 0.011, + "step": 23710 + }, + { + "epoch": 3.6843740292016154, + "grad_norm": 0.2896466851234436, + "learning_rate": 9.072219905799549e-05, + "loss": 0.0142, + "step": 23720 + }, + { + "epoch": 3.6859273066169616, + "grad_norm": 0.3595406115055084, + "learning_rate": 9.071260270199447e-05, + "loss": 0.0172, + "step": 23730 + }, + { + "epoch": 3.687480584032308, + "grad_norm": 0.28201109170913696, + "learning_rate": 9.070300189372441e-05, + "loss": 0.0099, + "step": 23740 + }, + { + "epoch": 3.6890338614476548, + "grad_norm": 0.23809628188610077, + "learning_rate": 9.069339663423528e-05, + "loss": 0.0132, + "step": 23750 + }, + { + "epoch": 3.690587138863001, + "grad_norm": 0.2731495499610901, + "learning_rate": 9.068378692457747e-05, + "loss": 0.0128, + "step": 23760 + }, + { + "epoch": 3.692140416278347, + "grad_norm": 0.30699026584625244, + "learning_rate": 9.067417276580189e-05, + "loss": 0.0126, + "step": 23770 + }, + { + "epoch": 3.6936936936936937, + "grad_norm": 0.2717634439468384, + "learning_rate": 9.066455415895993e-05, + "loss": 0.0106, + "step": 23780 + }, + { + "epoch": 3.6952469711090403, + "grad_norm": 0.27314722537994385, + "learning_rate": 9.065493110510346e-05, + "loss": 0.013, + "step": 23790 + }, + { + "epoch": 3.6968002485243865, + "grad_norm": 0.21836578845977783, + "learning_rate": 9.064530360528484e-05, + "loss": 0.0111, + "step": 23800 + }, + { + "epoch": 3.6983535259397327, + "grad_norm": 0.31817400455474854, + "learning_rate": 9.063567166055695e-05, + "loss": 0.0101, + "step": 23810 + }, + { + "epoch": 3.6999068033550793, + "grad_norm": 0.2575538158416748, + "learning_rate": 9.062603527197308e-05, + "loss": 0.0114, + "step": 23820 + }, + { + "epoch": 3.7014600807704254, + "grad_norm": 0.3410029113292694, + "learning_rate": 9.06163944405871e-05, + "loss": 0.011, + "step": 23830 + }, + { + "epoch": 3.703013358185772, + "grad_norm": 0.19098079204559326, + "learning_rate": 9.060674916745327e-05, + "loss": 0.0111, + "step": 23840 + }, + { + "epoch": 3.704566635601118, + "grad_norm": 0.3030313551425934, + "learning_rate": 9.05970994536264e-05, + "loss": 0.0167, + "step": 23850 + }, + { + "epoch": 3.706119913016465, + "grad_norm": 0.47913292050361633, + "learning_rate": 9.05874453001618e-05, + "loss": 0.0107, + "step": 23860 + }, + { + "epoch": 3.707673190431811, + "grad_norm": 0.2570559084415436, + "learning_rate": 9.057778670811517e-05, + "loss": 0.0125, + "step": 23870 + }, + { + "epoch": 3.7092264678471576, + "grad_norm": 0.5299606919288635, + "learning_rate": 9.056812367854281e-05, + "loss": 0.0108, + "step": 23880 + }, + { + "epoch": 3.7107797452625038, + "grad_norm": 0.3638221025466919, + "learning_rate": 9.055845621250143e-05, + "loss": 0.0135, + "step": 23890 + }, + { + "epoch": 3.7123330226778504, + "grad_norm": 0.8614619970321655, + "learning_rate": 9.054878431104825e-05, + "loss": 0.01, + "step": 23900 + }, + { + "epoch": 3.7138863000931965, + "grad_norm": 0.2923581004142761, + "learning_rate": 9.0539107975241e-05, + "loss": 0.0106, + "step": 23910 + }, + { + "epoch": 3.715439577508543, + "grad_norm": 0.3718606233596802, + "learning_rate": 9.052942720613784e-05, + "loss": 0.0132, + "step": 23920 + }, + { + "epoch": 3.7169928549238893, + "grad_norm": 0.2701959013938904, + "learning_rate": 9.051974200479745e-05, + "loss": 0.0147, + "step": 23930 + }, + { + "epoch": 3.718546132339236, + "grad_norm": 0.18919779360294342, + "learning_rate": 9.051005237227901e-05, + "loss": 0.011, + "step": 23940 + }, + { + "epoch": 3.720099409754582, + "grad_norm": 0.33912891149520874, + "learning_rate": 9.050035830964215e-05, + "loss": 0.0099, + "step": 23950 + }, + { + "epoch": 3.7216526871699287, + "grad_norm": 0.36024752259254456, + "learning_rate": 9.049065981794698e-05, + "loss": 0.0093, + "step": 23960 + }, + { + "epoch": 3.723205964585275, + "grad_norm": 0.41528627276420593, + "learning_rate": 9.048095689825414e-05, + "loss": 0.0119, + "step": 23970 + }, + { + "epoch": 3.7247592420006215, + "grad_norm": 0.2904050052165985, + "learning_rate": 9.047124955162472e-05, + "loss": 0.0097, + "step": 23980 + }, + { + "epoch": 3.7263125194159676, + "grad_norm": 0.1622840017080307, + "learning_rate": 9.046153777912028e-05, + "loss": 0.0115, + "step": 23990 + }, + { + "epoch": 3.7278657968313142, + "grad_norm": 0.21902386844158173, + "learning_rate": 9.045182158180292e-05, + "loss": 0.0096, + "step": 24000 + }, + { + "epoch": 3.7294190742466604, + "grad_norm": 0.4002116024494171, + "learning_rate": 9.044210096073516e-05, + "loss": 0.0123, + "step": 24010 + }, + { + "epoch": 3.7309723516620066, + "grad_norm": 0.4695390462875366, + "learning_rate": 9.043237591698004e-05, + "loss": 0.0123, + "step": 24020 + }, + { + "epoch": 3.732525629077353, + "grad_norm": 0.33076444268226624, + "learning_rate": 9.04226464516011e-05, + "loss": 0.0125, + "step": 24030 + }, + { + "epoch": 3.7340789064927, + "grad_norm": 0.19490692019462585, + "learning_rate": 9.041291256566229e-05, + "loss": 0.0075, + "step": 24040 + }, + { + "epoch": 3.735632183908046, + "grad_norm": 0.16053488850593567, + "learning_rate": 9.040317426022814e-05, + "loss": 0.0109, + "step": 24050 + }, + { + "epoch": 3.737185461323392, + "grad_norm": 0.31854087114334106, + "learning_rate": 9.03934315363636e-05, + "loss": 0.0102, + "step": 24060 + }, + { + "epoch": 3.7387387387387387, + "grad_norm": 0.31978797912597656, + "learning_rate": 9.038368439513409e-05, + "loss": 0.0151, + "step": 24070 + }, + { + "epoch": 3.7402920161540854, + "grad_norm": 0.2127339243888855, + "learning_rate": 9.03739328376056e-05, + "loss": 0.0113, + "step": 24080 + }, + { + "epoch": 3.7418452935694315, + "grad_norm": 0.33404961228370667, + "learning_rate": 9.036417686484451e-05, + "loss": 0.0126, + "step": 24090 + }, + { + "epoch": 3.7433985709847777, + "grad_norm": 0.21611382067203522, + "learning_rate": 9.035441647791773e-05, + "loss": 0.0111, + "step": 24100 + }, + { + "epoch": 3.7449518484001243, + "grad_norm": 0.2665908634662628, + "learning_rate": 9.034465167789263e-05, + "loss": 0.0088, + "step": 24110 + }, + { + "epoch": 3.746505125815471, + "grad_norm": 0.2213902771472931, + "learning_rate": 9.033488246583706e-05, + "loss": 0.0111, + "step": 24120 + }, + { + "epoch": 3.748058403230817, + "grad_norm": 0.20923234522342682, + "learning_rate": 9.032510884281941e-05, + "loss": 0.0087, + "step": 24130 + }, + { + "epoch": 3.7496116806461632, + "grad_norm": 0.16770032048225403, + "learning_rate": 9.031533080990848e-05, + "loss": 0.0095, + "step": 24140 + }, + { + "epoch": 3.75116495806151, + "grad_norm": 0.23019331693649292, + "learning_rate": 9.030554836817358e-05, + "loss": 0.0121, + "step": 24150 + }, + { + "epoch": 3.752718235476856, + "grad_norm": 0.17820380628108978, + "learning_rate": 9.029576151868451e-05, + "loss": 0.0103, + "step": 24160 + }, + { + "epoch": 3.7542715128922026, + "grad_norm": 0.17760442197322845, + "learning_rate": 9.028597026251155e-05, + "loss": 0.0065, + "step": 24170 + }, + { + "epoch": 3.755824790307549, + "grad_norm": 0.24275676906108856, + "learning_rate": 9.027617460072547e-05, + "loss": 0.0103, + "step": 24180 + }, + { + "epoch": 3.7573780677228954, + "grad_norm": 0.1876344531774521, + "learning_rate": 9.026637453439745e-05, + "loss": 0.0109, + "step": 24190 + }, + { + "epoch": 3.7589313451382416, + "grad_norm": 0.23554745316505432, + "learning_rate": 9.025657006459927e-05, + "loss": 0.0087, + "step": 24200 + }, + { + "epoch": 3.760484622553588, + "grad_norm": 0.2918838858604431, + "learning_rate": 9.024676119240311e-05, + "loss": 0.0121, + "step": 24210 + }, + { + "epoch": 3.7620378999689343, + "grad_norm": 0.27073290944099426, + "learning_rate": 9.023694791888166e-05, + "loss": 0.0086, + "step": 24220 + }, + { + "epoch": 3.763591177384281, + "grad_norm": 0.32038286328315735, + "learning_rate": 9.022713024510808e-05, + "loss": 0.0126, + "step": 24230 + }, + { + "epoch": 3.765144454799627, + "grad_norm": 0.36868664622306824, + "learning_rate": 9.021730817215601e-05, + "loss": 0.0144, + "step": 24240 + }, + { + "epoch": 3.7666977322149737, + "grad_norm": 0.3822726905345917, + "learning_rate": 9.02074817010996e-05, + "loss": 0.0088, + "step": 24250 + }, + { + "epoch": 3.76825100963032, + "grad_norm": 0.19045203924179077, + "learning_rate": 9.019765083301342e-05, + "loss": 0.0086, + "step": 24260 + }, + { + "epoch": 3.7698042870456665, + "grad_norm": 0.41100427508354187, + "learning_rate": 9.01878155689726e-05, + "loss": 0.014, + "step": 24270 + }, + { + "epoch": 3.7713575644610127, + "grad_norm": 0.30352604389190674, + "learning_rate": 9.017797591005268e-05, + "loss": 0.0121, + "step": 24280 + }, + { + "epoch": 3.7729108418763593, + "grad_norm": 0.23721155524253845, + "learning_rate": 9.016813185732972e-05, + "loss": 0.0103, + "step": 24290 + }, + { + "epoch": 3.7744641192917054, + "grad_norm": 0.41766053438186646, + "learning_rate": 9.015828341188027e-05, + "loss": 0.0146, + "step": 24300 + }, + { + "epoch": 3.7760173967070516, + "grad_norm": 0.23496168851852417, + "learning_rate": 9.01484305747813e-05, + "loss": 0.0103, + "step": 24310 + }, + { + "epoch": 3.777570674122398, + "grad_norm": 0.7083344459533691, + "learning_rate": 9.013857334711033e-05, + "loss": 0.0133, + "step": 24320 + }, + { + "epoch": 3.779123951537745, + "grad_norm": 0.23415236175060272, + "learning_rate": 9.012871172994534e-05, + "loss": 0.01, + "step": 24330 + }, + { + "epoch": 3.780677228953091, + "grad_norm": 0.3157029449939728, + "learning_rate": 9.011884572436476e-05, + "loss": 0.0114, + "step": 24340 + }, + { + "epoch": 3.782230506368437, + "grad_norm": 0.22819329798221588, + "learning_rate": 9.010897533144754e-05, + "loss": 0.0103, + "step": 24350 + }, + { + "epoch": 3.7837837837837838, + "grad_norm": 0.3370586633682251, + "learning_rate": 9.009910055227306e-05, + "loss": 0.0101, + "step": 24360 + }, + { + "epoch": 3.7853370611991304, + "grad_norm": 0.18568874895572662, + "learning_rate": 9.008922138792124e-05, + "loss": 0.0109, + "step": 24370 + }, + { + "epoch": 3.7868903386144765, + "grad_norm": 0.24086825549602509, + "learning_rate": 9.007933783947244e-05, + "loss": 0.0091, + "step": 24380 + }, + { + "epoch": 3.7884436160298227, + "grad_norm": 0.215671569108963, + "learning_rate": 9.006944990800752e-05, + "loss": 0.014, + "step": 24390 + }, + { + "epoch": 3.7899968934451693, + "grad_norm": 0.19352129101753235, + "learning_rate": 9.005955759460779e-05, + "loss": 0.0106, + "step": 24400 + }, + { + "epoch": 3.791550170860516, + "grad_norm": 0.30985379219055176, + "learning_rate": 9.004966090035508e-05, + "loss": 0.0111, + "step": 24410 + }, + { + "epoch": 3.793103448275862, + "grad_norm": 0.33107686042785645, + "learning_rate": 9.003975982633166e-05, + "loss": 0.0114, + "step": 24420 + }, + { + "epoch": 3.7946567256912083, + "grad_norm": 0.22770509123802185, + "learning_rate": 9.00298543736203e-05, + "loss": 0.0128, + "step": 24430 + }, + { + "epoch": 3.796210003106555, + "grad_norm": 0.3192411959171295, + "learning_rate": 9.001994454330427e-05, + "loss": 0.0089, + "step": 24440 + }, + { + "epoch": 3.7977632805219015, + "grad_norm": 0.25310423970222473, + "learning_rate": 9.001003033646727e-05, + "loss": 0.0116, + "step": 24450 + }, + { + "epoch": 3.7993165579372477, + "grad_norm": 0.2813625633716583, + "learning_rate": 9.00001117541935e-05, + "loss": 0.0123, + "step": 24460 + }, + { + "epoch": 3.800869835352594, + "grad_norm": 0.3783665597438812, + "learning_rate": 8.999018879756764e-05, + "loss": 0.0108, + "step": 24470 + }, + { + "epoch": 3.8024231127679404, + "grad_norm": 0.23065844178199768, + "learning_rate": 8.998026146767487e-05, + "loss": 0.0086, + "step": 24480 + }, + { + "epoch": 3.8039763901832866, + "grad_norm": 0.27731841802597046, + "learning_rate": 8.99703297656008e-05, + "loss": 0.01, + "step": 24490 + }, + { + "epoch": 3.805529667598633, + "grad_norm": 0.20540432631969452, + "learning_rate": 8.996039369243156e-05, + "loss": 0.0088, + "step": 24500 + }, + { + "epoch": 3.8070829450139794, + "grad_norm": 0.14613407850265503, + "learning_rate": 8.995045324925378e-05, + "loss": 0.0109, + "step": 24510 + }, + { + "epoch": 3.808636222429326, + "grad_norm": 0.3185731768608093, + "learning_rate": 8.994050843715448e-05, + "loss": 0.0127, + "step": 24520 + }, + { + "epoch": 3.810189499844672, + "grad_norm": 0.24198320508003235, + "learning_rate": 8.993055925722121e-05, + "loss": 0.01, + "step": 24530 + }, + { + "epoch": 3.8117427772600188, + "grad_norm": 0.3547452390193939, + "learning_rate": 8.992060571054202e-05, + "loss": 0.0141, + "step": 24540 + }, + { + "epoch": 3.813296054675365, + "grad_norm": 0.38823479413986206, + "learning_rate": 8.991064779820542e-05, + "loss": 0.0095, + "step": 24550 + }, + { + "epoch": 3.8148493320907115, + "grad_norm": 0.18600568175315857, + "learning_rate": 8.990068552130036e-05, + "loss": 0.0084, + "step": 24560 + }, + { + "epoch": 3.8164026095060577, + "grad_norm": 0.3605685234069824, + "learning_rate": 8.989071888091634e-05, + "loss": 0.0113, + "step": 24570 + }, + { + "epoch": 3.8179558869214043, + "grad_norm": 0.27060797810554504, + "learning_rate": 8.988074787814329e-05, + "loss": 0.0084, + "step": 24580 + }, + { + "epoch": 3.8195091643367505, + "grad_norm": 0.2528657913208008, + "learning_rate": 8.987077251407158e-05, + "loss": 0.0087, + "step": 24590 + }, + { + "epoch": 3.821062441752097, + "grad_norm": 0.24219495058059692, + "learning_rate": 8.986079278979216e-05, + "loss": 0.0114, + "step": 24600 + }, + { + "epoch": 3.8226157191674432, + "grad_norm": 0.2830139696598053, + "learning_rate": 8.985080870639635e-05, + "loss": 0.0075, + "step": 24610 + }, + { + "epoch": 3.82416899658279, + "grad_norm": 0.41723161935806274, + "learning_rate": 8.984082026497603e-05, + "loss": 0.0106, + "step": 24620 + }, + { + "epoch": 3.825722273998136, + "grad_norm": 0.22488930821418762, + "learning_rate": 8.98308274666235e-05, + "loss": 0.009, + "step": 24630 + }, + { + "epoch": 3.827275551413482, + "grad_norm": 0.24578028917312622, + "learning_rate": 8.982083031243155e-05, + "loss": 0.0116, + "step": 24640 + }, + { + "epoch": 3.828828828828829, + "grad_norm": 0.4029971957206726, + "learning_rate": 8.98108288034935e-05, + "loss": 0.0097, + "step": 24650 + }, + { + "epoch": 3.8303821062441754, + "grad_norm": 0.4286789894104004, + "learning_rate": 8.980082294090305e-05, + "loss": 0.0109, + "step": 24660 + }, + { + "epoch": 3.8319353836595216, + "grad_norm": 0.25567519664764404, + "learning_rate": 8.979081272575443e-05, + "loss": 0.0098, + "step": 24670 + }, + { + "epoch": 3.8334886610748677, + "grad_norm": 0.19377973675727844, + "learning_rate": 8.978079815914236e-05, + "loss": 0.007, + "step": 24680 + }, + { + "epoch": 3.8350419384902144, + "grad_norm": 0.23998203873634338, + "learning_rate": 8.977077924216202e-05, + "loss": 0.0127, + "step": 24690 + }, + { + "epoch": 3.836595215905561, + "grad_norm": 0.21582503616809845, + "learning_rate": 8.976075597590905e-05, + "loss": 0.0095, + "step": 24700 + }, + { + "epoch": 3.838148493320907, + "grad_norm": 0.3221307694911957, + "learning_rate": 8.975072836147958e-05, + "loss": 0.0088, + "step": 24710 + }, + { + "epoch": 3.8397017707362533, + "grad_norm": 0.3324994444847107, + "learning_rate": 8.974069639997025e-05, + "loss": 0.0126, + "step": 24720 + }, + { + "epoch": 3.8412550481516, + "grad_norm": 0.2502076029777527, + "learning_rate": 8.973066009247808e-05, + "loss": 0.0129, + "step": 24730 + }, + { + "epoch": 3.8428083255669465, + "grad_norm": 0.24743586778640747, + "learning_rate": 8.972061944010066e-05, + "loss": 0.0113, + "step": 24740 + }, + { + "epoch": 3.8443616029822927, + "grad_norm": 0.19658684730529785, + "learning_rate": 8.971057444393603e-05, + "loss": 0.0105, + "step": 24750 + }, + { + "epoch": 3.845914880397639, + "grad_norm": 0.1969096064567566, + "learning_rate": 8.970052510508268e-05, + "loss": 0.01, + "step": 24760 + }, + { + "epoch": 3.8474681578129855, + "grad_norm": 0.15066437423229218, + "learning_rate": 8.969047142463959e-05, + "loss": 0.0109, + "step": 24770 + }, + { + "epoch": 3.8490214352283316, + "grad_norm": 0.27470386028289795, + "learning_rate": 8.968041340370621e-05, + "loss": 0.0122, + "step": 24780 + }, + { + "epoch": 3.8505747126436782, + "grad_norm": 0.16672064363956451, + "learning_rate": 8.96703510433825e-05, + "loss": 0.0125, + "step": 24790 + }, + { + "epoch": 3.8521279900590244, + "grad_norm": 0.5565047264099121, + "learning_rate": 8.966028434476883e-05, + "loss": 0.0111, + "step": 24800 + }, + { + "epoch": 3.853681267474371, + "grad_norm": 0.35457953810691833, + "learning_rate": 8.96502133089661e-05, + "loss": 0.0122, + "step": 24810 + }, + { + "epoch": 3.855234544889717, + "grad_norm": 0.2877291738986969, + "learning_rate": 8.964013793707564e-05, + "loss": 0.0114, + "step": 24820 + }, + { + "epoch": 3.856787822305064, + "grad_norm": 0.35003286600112915, + "learning_rate": 8.963005823019932e-05, + "loss": 0.0144, + "step": 24830 + }, + { + "epoch": 3.85834109972041, + "grad_norm": 0.23843717575073242, + "learning_rate": 8.961997418943939e-05, + "loss": 0.008, + "step": 24840 + }, + { + "epoch": 3.8598943771357566, + "grad_norm": 0.25304633378982544, + "learning_rate": 8.960988581589865e-05, + "loss": 0.0119, + "step": 24850 + }, + { + "epoch": 3.8614476545511027, + "grad_norm": 0.24912704527378082, + "learning_rate": 8.959979311068037e-05, + "loss": 0.0139, + "step": 24860 + }, + { + "epoch": 3.8630009319664493, + "grad_norm": 0.2501377761363983, + "learning_rate": 8.958969607488823e-05, + "loss": 0.0107, + "step": 24870 + }, + { + "epoch": 3.8645542093817955, + "grad_norm": 0.2822844684123993, + "learning_rate": 8.957959470962647e-05, + "loss": 0.014, + "step": 24880 + }, + { + "epoch": 3.866107486797142, + "grad_norm": 0.1681593954563141, + "learning_rate": 8.956948901599971e-05, + "loss": 0.0087, + "step": 24890 + }, + { + "epoch": 3.8676607642124883, + "grad_norm": 0.23125946521759033, + "learning_rate": 8.955937899511315e-05, + "loss": 0.0096, + "step": 24900 + }, + { + "epoch": 3.869214041627835, + "grad_norm": 0.301662415266037, + "learning_rate": 8.954926464807238e-05, + "loss": 0.0105, + "step": 24910 + }, + { + "epoch": 3.870767319043181, + "grad_norm": 0.2115197479724884, + "learning_rate": 8.953914597598347e-05, + "loss": 0.0084, + "step": 24920 + }, + { + "epoch": 3.872320596458527, + "grad_norm": 0.19678549468517303, + "learning_rate": 8.952902297995303e-05, + "loss": 0.0085, + "step": 24930 + }, + { + "epoch": 3.873873873873874, + "grad_norm": 0.27717161178588867, + "learning_rate": 8.951889566108804e-05, + "loss": 0.0113, + "step": 24940 + }, + { + "epoch": 3.8754271512892204, + "grad_norm": 0.2013668566942215, + "learning_rate": 8.950876402049606e-05, + "loss": 0.0116, + "step": 24950 + }, + { + "epoch": 3.8769804287045666, + "grad_norm": 0.1999899446964264, + "learning_rate": 8.949862805928504e-05, + "loss": 0.0082, + "step": 24960 + }, + { + "epoch": 3.8785337061199128, + "grad_norm": 0.3094763159751892, + "learning_rate": 8.948848777856343e-05, + "loss": 0.0137, + "step": 24970 + }, + { + "epoch": 3.8800869835352594, + "grad_norm": 0.2381165623664856, + "learning_rate": 8.947834317944017e-05, + "loss": 0.0117, + "step": 24980 + }, + { + "epoch": 3.881640260950606, + "grad_norm": 0.2524758577346802, + "learning_rate": 8.946819426302466e-05, + "loss": 0.0099, + "step": 24990 + }, + { + "epoch": 3.883193538365952, + "grad_norm": 0.21542534232139587, + "learning_rate": 8.945804103042676e-05, + "loss": 0.0092, + "step": 25000 + }, + { + "epoch": 3.8847468157812983, + "grad_norm": 0.5472559928894043, + "learning_rate": 8.944788348275681e-05, + "loss": 0.0127, + "step": 25010 + }, + { + "epoch": 3.886300093196645, + "grad_norm": 0.18684640526771545, + "learning_rate": 8.943772162112565e-05, + "loss": 0.0078, + "step": 25020 + }, + { + "epoch": 3.8878533706119915, + "grad_norm": 0.35619258880615234, + "learning_rate": 8.942755544664454e-05, + "loss": 0.0092, + "step": 25030 + }, + { + "epoch": 3.8894066480273377, + "grad_norm": 0.2946649193763733, + "learning_rate": 8.941738496042525e-05, + "loss": 0.0108, + "step": 25040 + }, + { + "epoch": 3.890959925442684, + "grad_norm": 0.29120102524757385, + "learning_rate": 8.940721016357999e-05, + "loss": 0.0126, + "step": 25050 + }, + { + "epoch": 3.8925132028580305, + "grad_norm": 0.25601744651794434, + "learning_rate": 8.939703105722148e-05, + "loss": 0.0127, + "step": 25060 + }, + { + "epoch": 3.894066480273377, + "grad_norm": 0.37372350692749023, + "learning_rate": 8.93868476424629e-05, + "loss": 0.0129, + "step": 25070 + }, + { + "epoch": 3.8956197576887233, + "grad_norm": 0.337014377117157, + "learning_rate": 8.937665992041786e-05, + "loss": 0.0103, + "step": 25080 + }, + { + "epoch": 3.8971730351040694, + "grad_norm": 0.4805167317390442, + "learning_rate": 8.93664678922005e-05, + "loss": 0.0107, + "step": 25090 + }, + { + "epoch": 3.898726312519416, + "grad_norm": 0.28153517842292786, + "learning_rate": 8.93562715589254e-05, + "loss": 0.0135, + "step": 25100 + }, + { + "epoch": 3.900279589934762, + "grad_norm": 0.24814003705978394, + "learning_rate": 8.934607092170762e-05, + "loss": 0.01, + "step": 25110 + }, + { + "epoch": 3.901832867350109, + "grad_norm": 0.4074574112892151, + "learning_rate": 8.933586598166266e-05, + "loss": 0.0144, + "step": 25120 + }, + { + "epoch": 3.903386144765455, + "grad_norm": 0.3043968379497528, + "learning_rate": 8.932565673990655e-05, + "loss": 0.0117, + "step": 25130 + }, + { + "epoch": 3.9049394221808016, + "grad_norm": 0.536807119846344, + "learning_rate": 8.931544319755574e-05, + "loss": 0.0129, + "step": 25140 + }, + { + "epoch": 3.9064926995961478, + "grad_norm": 0.25280192494392395, + "learning_rate": 8.930522535572718e-05, + "loss": 0.0139, + "step": 25150 + }, + { + "epoch": 3.9080459770114944, + "grad_norm": 0.22550326585769653, + "learning_rate": 8.929500321553826e-05, + "loss": 0.0131, + "step": 25160 + }, + { + "epoch": 3.9095992544268405, + "grad_norm": 0.2534283399581909, + "learning_rate": 8.928477677810686e-05, + "loss": 0.0107, + "step": 25170 + }, + { + "epoch": 3.911152531842187, + "grad_norm": 0.29138466715812683, + "learning_rate": 8.927454604455137e-05, + "loss": 0.01, + "step": 25180 + }, + { + "epoch": 3.9127058092575333, + "grad_norm": 0.26762932538986206, + "learning_rate": 8.926431101599053e-05, + "loss": 0.0121, + "step": 25190 + }, + { + "epoch": 3.91425908667288, + "grad_norm": 0.2789064049720764, + "learning_rate": 8.925407169354369e-05, + "loss": 0.0118, + "step": 25200 + }, + { + "epoch": 3.915812364088226, + "grad_norm": 0.3474670350551605, + "learning_rate": 8.92438280783306e-05, + "loss": 0.0086, + "step": 25210 + }, + { + "epoch": 3.9173656415035727, + "grad_norm": 0.520646870136261, + "learning_rate": 8.923358017147146e-05, + "loss": 0.0114, + "step": 25220 + }, + { + "epoch": 3.918918918918919, + "grad_norm": 0.25099804997444153, + "learning_rate": 8.922332797408697e-05, + "loss": 0.0102, + "step": 25230 + }, + { + "epoch": 3.9204721963342655, + "grad_norm": 0.23914232850074768, + "learning_rate": 8.921307148729831e-05, + "loss": 0.0121, + "step": 25240 + }, + { + "epoch": 3.9220254737496116, + "grad_norm": 0.47815752029418945, + "learning_rate": 8.920281071222712e-05, + "loss": 0.0134, + "step": 25250 + }, + { + "epoch": 3.923578751164958, + "grad_norm": 0.3972282409667969, + "learning_rate": 8.919254564999548e-05, + "loss": 0.0123, + "step": 25260 + }, + { + "epoch": 3.9251320285803044, + "grad_norm": 0.22188925743103027, + "learning_rate": 8.918227630172598e-05, + "loss": 0.0135, + "step": 25270 + }, + { + "epoch": 3.926685305995651, + "grad_norm": 0.3008710443973541, + "learning_rate": 8.917200266854165e-05, + "loss": 0.0112, + "step": 25280 + }, + { + "epoch": 3.928238583410997, + "grad_norm": 0.2387913018465042, + "learning_rate": 8.9161724751566e-05, + "loss": 0.0118, + "step": 25290 + }, + { + "epoch": 3.9297918608263434, + "grad_norm": 0.2221137285232544, + "learning_rate": 8.915144255192302e-05, + "loss": 0.0116, + "step": 25300 + }, + { + "epoch": 3.93134513824169, + "grad_norm": 0.2891944944858551, + "learning_rate": 8.914115607073714e-05, + "loss": 0.0106, + "step": 25310 + }, + { + "epoch": 3.9328984156570366, + "grad_norm": 0.22891299426555634, + "learning_rate": 8.913086530913327e-05, + "loss": 0.0134, + "step": 25320 + }, + { + "epoch": 3.9344516930723827, + "grad_norm": 0.2379780411720276, + "learning_rate": 8.912057026823681e-05, + "loss": 0.0109, + "step": 25330 + }, + { + "epoch": 3.936004970487729, + "grad_norm": 0.1904175579547882, + "learning_rate": 8.91102709491736e-05, + "loss": 0.012, + "step": 25340 + }, + { + "epoch": 3.9375582479030755, + "grad_norm": 0.34377360343933105, + "learning_rate": 8.909996735306996e-05, + "loss": 0.0109, + "step": 25350 + }, + { + "epoch": 3.939111525318422, + "grad_norm": 0.32390594482421875, + "learning_rate": 8.908965948105268e-05, + "loss": 0.0076, + "step": 25360 + }, + { + "epoch": 3.9406648027337683, + "grad_norm": 0.2386329621076584, + "learning_rate": 8.907934733424901e-05, + "loss": 0.01, + "step": 25370 + }, + { + "epoch": 3.9422180801491145, + "grad_norm": 0.15039651095867157, + "learning_rate": 8.906903091378666e-05, + "loss": 0.01, + "step": 25380 + }, + { + "epoch": 3.943771357564461, + "grad_norm": 0.25573593378067017, + "learning_rate": 8.905871022079384e-05, + "loss": 0.013, + "step": 25390 + }, + { + "epoch": 3.9453246349798072, + "grad_norm": 0.7918702363967896, + "learning_rate": 8.90483852563992e-05, + "loss": 0.011, + "step": 25400 + }, + { + "epoch": 3.946877912395154, + "grad_norm": 0.21792973577976227, + "learning_rate": 8.903805602173185e-05, + "loss": 0.0116, + "step": 25410 + }, + { + "epoch": 3.9484311898105, + "grad_norm": 0.19511279463768005, + "learning_rate": 8.902772251792137e-05, + "loss": 0.009, + "step": 25420 + }, + { + "epoch": 3.9499844672258466, + "grad_norm": 0.21952615678310394, + "learning_rate": 8.901738474609786e-05, + "loss": 0.0118, + "step": 25430 + }, + { + "epoch": 3.951537744641193, + "grad_norm": 0.19715069234371185, + "learning_rate": 8.900704270739179e-05, + "loss": 0.0102, + "step": 25440 + }, + { + "epoch": 3.9530910220565394, + "grad_norm": 0.361738383769989, + "learning_rate": 8.89966964029342e-05, + "loss": 0.012, + "step": 25450 + }, + { + "epoch": 3.9546442994718856, + "grad_norm": 0.33274978399276733, + "learning_rate": 8.898634583385652e-05, + "loss": 0.013, + "step": 25460 + }, + { + "epoch": 3.956197576887232, + "grad_norm": 0.273517370223999, + "learning_rate": 8.897599100129065e-05, + "loss": 0.0105, + "step": 25470 + }, + { + "epoch": 3.9577508543025783, + "grad_norm": 0.29032620787620544, + "learning_rate": 8.896563190636903e-05, + "loss": 0.0128, + "step": 25480 + }, + { + "epoch": 3.959304131717925, + "grad_norm": 0.18526996672153473, + "learning_rate": 8.895526855022448e-05, + "loss": 0.0105, + "step": 25490 + }, + { + "epoch": 3.960857409133271, + "grad_norm": 0.32549458742141724, + "learning_rate": 8.894490093399033e-05, + "loss": 0.0139, + "step": 25500 + }, + { + "epoch": 3.9624106865486177, + "grad_norm": 0.33519938588142395, + "learning_rate": 8.893452905880035e-05, + "loss": 0.0135, + "step": 25510 + }, + { + "epoch": 3.963963963963964, + "grad_norm": 0.2424570769071579, + "learning_rate": 8.892415292578883e-05, + "loss": 0.0121, + "step": 25520 + }, + { + "epoch": 3.9655172413793105, + "grad_norm": 0.288763165473938, + "learning_rate": 8.891377253609046e-05, + "loss": 0.0133, + "step": 25530 + }, + { + "epoch": 3.9670705187946567, + "grad_norm": 0.22682048380374908, + "learning_rate": 8.890338789084043e-05, + "loss": 0.011, + "step": 25540 + }, + { + "epoch": 3.968623796210003, + "grad_norm": 0.22693127393722534, + "learning_rate": 8.88929989911744e-05, + "loss": 0.0094, + "step": 25550 + }, + { + "epoch": 3.9701770736253494, + "grad_norm": 0.3066685199737549, + "learning_rate": 8.888260583822847e-05, + "loss": 0.0117, + "step": 25560 + }, + { + "epoch": 3.971730351040696, + "grad_norm": 0.34770911931991577, + "learning_rate": 8.887220843313921e-05, + "loss": 0.0113, + "step": 25570 + }, + { + "epoch": 3.973283628456042, + "grad_norm": 0.23799602687358856, + "learning_rate": 8.88618067770437e-05, + "loss": 0.0123, + "step": 25580 + }, + { + "epoch": 3.9748369058713884, + "grad_norm": 0.33474910259246826, + "learning_rate": 8.885140087107942e-05, + "loss": 0.0105, + "step": 25590 + }, + { + "epoch": 3.976390183286735, + "grad_norm": 0.3924482762813568, + "learning_rate": 8.884099071638436e-05, + "loss": 0.011, + "step": 25600 + }, + { + "epoch": 3.9779434607020816, + "grad_norm": 0.22412709891796112, + "learning_rate": 8.883057631409695e-05, + "loss": 0.0101, + "step": 25610 + }, + { + "epoch": 3.9794967381174278, + "grad_norm": 0.1708299219608307, + "learning_rate": 8.882015766535608e-05, + "loss": 0.0095, + "step": 25620 + }, + { + "epoch": 3.981050015532774, + "grad_norm": 0.2970634400844574, + "learning_rate": 8.880973477130115e-05, + "loss": 0.011, + "step": 25630 + }, + { + "epoch": 3.9826032929481205, + "grad_norm": 0.2267591804265976, + "learning_rate": 8.879930763307197e-05, + "loss": 0.0112, + "step": 25640 + }, + { + "epoch": 3.984156570363467, + "grad_norm": 0.4111708104610443, + "learning_rate": 8.878887625180884e-05, + "loss": 0.0106, + "step": 25650 + }, + { + "epoch": 3.9857098477788133, + "grad_norm": 0.1682768613100052, + "learning_rate": 8.877844062865253e-05, + "loss": 0.0104, + "step": 25660 + }, + { + "epoch": 3.9872631251941595, + "grad_norm": 0.30186155438423157, + "learning_rate": 8.876800076474424e-05, + "loss": 0.0123, + "step": 25670 + }, + { + "epoch": 3.988816402609506, + "grad_norm": 0.22091805934906006, + "learning_rate": 8.875755666122568e-05, + "loss": 0.0104, + "step": 25680 + }, + { + "epoch": 3.9903696800248527, + "grad_norm": 0.43025559186935425, + "learning_rate": 8.8747108319239e-05, + "loss": 0.0122, + "step": 25690 + }, + { + "epoch": 3.991922957440199, + "grad_norm": 0.4503591060638428, + "learning_rate": 8.87366557399268e-05, + "loss": 0.0094, + "step": 25700 + }, + { + "epoch": 3.993476234855545, + "grad_norm": 0.30399540066719055, + "learning_rate": 8.872619892443217e-05, + "loss": 0.0112, + "step": 25710 + }, + { + "epoch": 3.9950295122708916, + "grad_norm": 0.17710712552070618, + "learning_rate": 8.871573787389865e-05, + "loss": 0.01, + "step": 25720 + }, + { + "epoch": 3.996582789686238, + "grad_norm": 0.24861639738082886, + "learning_rate": 8.870527258947024e-05, + "loss": 0.0102, + "step": 25730 + }, + { + "epoch": 3.9981360671015844, + "grad_norm": 1.3647173643112183, + "learning_rate": 8.869480307229143e-05, + "loss": 0.012, + "step": 25740 + }, + { + "epoch": 3.9996893445169306, + "grad_norm": 0.2943127751350403, + "learning_rate": 8.868432932350712e-05, + "loss": 0.0113, + "step": 25750 + }, + { + "epoch": 4.001242621932277, + "grad_norm": 0.2405327409505844, + "learning_rate": 8.867385134426272e-05, + "loss": 0.0132, + "step": 25760 + }, + { + "epoch": 4.002795899347624, + "grad_norm": 0.28516408801078796, + "learning_rate": 8.866336913570407e-05, + "loss": 0.0104, + "step": 25770 + }, + { + "epoch": 4.00434917676297, + "grad_norm": 0.2722184658050537, + "learning_rate": 8.865288269897751e-05, + "loss": 0.0122, + "step": 25780 + }, + { + "epoch": 4.005902454178316, + "grad_norm": 0.22543387115001678, + "learning_rate": 8.864239203522981e-05, + "loss": 0.0123, + "step": 25790 + }, + { + "epoch": 4.007455731593662, + "grad_norm": 0.2077832818031311, + "learning_rate": 8.863189714560822e-05, + "loss": 0.009, + "step": 25800 + }, + { + "epoch": 4.009009009009009, + "grad_norm": 0.3851434886455536, + "learning_rate": 8.862139803126043e-05, + "loss": 0.0103, + "step": 25810 + }, + { + "epoch": 4.0105622864243555, + "grad_norm": 0.33362695574760437, + "learning_rate": 8.861089469333463e-05, + "loss": 0.0102, + "step": 25820 + }, + { + "epoch": 4.012115563839702, + "grad_norm": 0.28746941685676575, + "learning_rate": 8.860038713297944e-05, + "loss": 0.0132, + "step": 25830 + }, + { + "epoch": 4.013668841255048, + "grad_norm": 0.3765944838523865, + "learning_rate": 8.858987535134394e-05, + "loss": 0.0145, + "step": 25840 + }, + { + "epoch": 4.015222118670395, + "grad_norm": 0.41737404465675354, + "learning_rate": 8.857935934957769e-05, + "loss": 0.0138, + "step": 25850 + }, + { + "epoch": 4.016775396085741, + "grad_norm": 0.4369918704032898, + "learning_rate": 8.856883912883071e-05, + "loss": 0.0136, + "step": 25860 + }, + { + "epoch": 4.018328673501087, + "grad_norm": 0.557256817817688, + "learning_rate": 8.855831469025346e-05, + "loss": 0.0138, + "step": 25870 + }, + { + "epoch": 4.019881950916433, + "grad_norm": 0.35637351870536804, + "learning_rate": 8.854778603499689e-05, + "loss": 0.0094, + "step": 25880 + }, + { + "epoch": 4.0214352283317805, + "grad_norm": 0.4275244176387787, + "learning_rate": 8.85372531642124e-05, + "loss": 0.0108, + "step": 25890 + }, + { + "epoch": 4.022988505747127, + "grad_norm": 0.23809601366519928, + "learning_rate": 8.852671607905185e-05, + "loss": 0.0098, + "step": 25900 + }, + { + "epoch": 4.024541783162473, + "grad_norm": 0.3180028200149536, + "learning_rate": 8.851617478066754e-05, + "loss": 0.0144, + "step": 25910 + }, + { + "epoch": 4.026095060577819, + "grad_norm": 0.2760772407054901, + "learning_rate": 8.850562927021227e-05, + "loss": 0.0123, + "step": 25920 + }, + { + "epoch": 4.027648337993166, + "grad_norm": 0.23964755237102509, + "learning_rate": 8.849507954883928e-05, + "loss": 0.0108, + "step": 25930 + }, + { + "epoch": 4.029201615408512, + "grad_norm": 0.699877917766571, + "learning_rate": 8.848452561770226e-05, + "loss": 0.0115, + "step": 25940 + }, + { + "epoch": 4.030754892823858, + "grad_norm": 0.3516925573348999, + "learning_rate": 8.847396747795538e-05, + "loss": 0.0105, + "step": 25950 + }, + { + "epoch": 4.0323081702392045, + "grad_norm": 0.2606875002384186, + "learning_rate": 8.846340513075327e-05, + "loss": 0.0106, + "step": 25960 + }, + { + "epoch": 4.033861447654551, + "grad_norm": 0.1901322454214096, + "learning_rate": 8.845283857725099e-05, + "loss": 0.0097, + "step": 25970 + }, + { + "epoch": 4.035414725069898, + "grad_norm": 0.2545694410800934, + "learning_rate": 8.844226781860409e-05, + "loss": 0.0088, + "step": 25980 + }, + { + "epoch": 4.036968002485244, + "grad_norm": 0.4766624867916107, + "learning_rate": 8.84316928559686e-05, + "loss": 0.0139, + "step": 25990 + }, + { + "epoch": 4.03852127990059, + "grad_norm": 0.3356645107269287, + "learning_rate": 8.842111369050094e-05, + "loss": 0.0093, + "step": 26000 + }, + { + "epoch": 4.040074557315936, + "grad_norm": 0.2625381052494049, + "learning_rate": 8.841053032335808e-05, + "loss": 0.0099, + "step": 26010 + }, + { + "epoch": 4.041627834731283, + "grad_norm": 0.38582348823547363, + "learning_rate": 8.839994275569735e-05, + "loss": 0.0098, + "step": 26020 + }, + { + "epoch": 4.0431811121466295, + "grad_norm": 0.6784653067588806, + "learning_rate": 8.838935098867662e-05, + "loss": 0.0106, + "step": 26030 + }, + { + "epoch": 4.044734389561976, + "grad_norm": 0.30421268939971924, + "learning_rate": 8.837875502345418e-05, + "loss": 0.0119, + "step": 26040 + }, + { + "epoch": 4.046287666977322, + "grad_norm": 0.422463983297348, + "learning_rate": 8.83681548611888e-05, + "loss": 0.0099, + "step": 26050 + }, + { + "epoch": 4.047840944392669, + "grad_norm": 0.17532013356685638, + "learning_rate": 8.835755050303969e-05, + "loss": 0.0098, + "step": 26060 + }, + { + "epoch": 4.049394221808015, + "grad_norm": 0.26758480072021484, + "learning_rate": 8.834694195016653e-05, + "loss": 0.0119, + "step": 26070 + }, + { + "epoch": 4.050947499223361, + "grad_norm": 0.26299262046813965, + "learning_rate": 8.833632920372942e-05, + "loss": 0.011, + "step": 26080 + }, + { + "epoch": 4.052500776638707, + "grad_norm": 0.21729132533073425, + "learning_rate": 8.832571226488903e-05, + "loss": 0.0124, + "step": 26090 + }, + { + "epoch": 4.054054054054054, + "grad_norm": 0.22503264248371124, + "learning_rate": 8.831509113480634e-05, + "loss": 0.0111, + "step": 26100 + }, + { + "epoch": 4.055607331469401, + "grad_norm": 0.1864558458328247, + "learning_rate": 8.83044658146429e-05, + "loss": 0.0105, + "step": 26110 + }, + { + "epoch": 4.057160608884747, + "grad_norm": 0.2270529866218567, + "learning_rate": 8.829383630556067e-05, + "loss": 0.0104, + "step": 26120 + }, + { + "epoch": 4.058713886300093, + "grad_norm": 0.23184806108474731, + "learning_rate": 8.828320260872207e-05, + "loss": 0.0096, + "step": 26130 + }, + { + "epoch": 4.06026716371544, + "grad_norm": 0.20580445230007172, + "learning_rate": 8.827256472529e-05, + "loss": 0.007, + "step": 26140 + }, + { + "epoch": 4.061820441130786, + "grad_norm": 0.20794256031513214, + "learning_rate": 8.826192265642778e-05, + "loss": 0.0097, + "step": 26150 + }, + { + "epoch": 4.063373718546132, + "grad_norm": 0.2990153729915619, + "learning_rate": 8.825127640329923e-05, + "loss": 0.0098, + "step": 26160 + }, + { + "epoch": 4.064926995961478, + "grad_norm": 0.25416895747184753, + "learning_rate": 8.824062596706861e-05, + "loss": 0.0104, + "step": 26170 + }, + { + "epoch": 4.0664802733768255, + "grad_norm": 0.3324458599090576, + "learning_rate": 8.822997134890062e-05, + "loss": 0.0128, + "step": 26180 + }, + { + "epoch": 4.068033550792172, + "grad_norm": 0.3597685396671295, + "learning_rate": 8.821931254996044e-05, + "loss": 0.0115, + "step": 26190 + }, + { + "epoch": 4.069586828207518, + "grad_norm": 0.6617540717124939, + "learning_rate": 8.82086495714137e-05, + "loss": 0.0093, + "step": 26200 + }, + { + "epoch": 4.071140105622864, + "grad_norm": 0.225095734000206, + "learning_rate": 8.81979824144265e-05, + "loss": 0.01, + "step": 26210 + }, + { + "epoch": 4.072693383038211, + "grad_norm": 0.20842081308364868, + "learning_rate": 8.818731108016536e-05, + "loss": 0.0114, + "step": 26220 + }, + { + "epoch": 4.074246660453557, + "grad_norm": 0.31950289011001587, + "learning_rate": 8.81766355697973e-05, + "loss": 0.0108, + "step": 26230 + }, + { + "epoch": 4.075799937868903, + "grad_norm": 0.21985672414302826, + "learning_rate": 8.816595588448977e-05, + "loss": 0.0075, + "step": 26240 + }, + { + "epoch": 4.0773532152842495, + "grad_norm": 0.338119775056839, + "learning_rate": 8.81552720254107e-05, + "loss": 0.0118, + "step": 26250 + }, + { + "epoch": 4.078906492699597, + "grad_norm": 0.2077280879020691, + "learning_rate": 8.814458399372842e-05, + "loss": 0.0129, + "step": 26260 + }, + { + "epoch": 4.080459770114943, + "grad_norm": 0.17363905906677246, + "learning_rate": 8.813389179061181e-05, + "loss": 0.009, + "step": 26270 + }, + { + "epoch": 4.082013047530289, + "grad_norm": 0.1778227686882019, + "learning_rate": 8.812319541723012e-05, + "loss": 0.0093, + "step": 26280 + }, + { + "epoch": 4.083566324945635, + "grad_norm": 0.12196733802556992, + "learning_rate": 8.811249487475309e-05, + "loss": 0.0084, + "step": 26290 + }, + { + "epoch": 4.085119602360981, + "grad_norm": 0.13716167211532593, + "learning_rate": 8.810179016435092e-05, + "loss": 0.0085, + "step": 26300 + }, + { + "epoch": 4.086672879776328, + "grad_norm": 0.4707149863243103, + "learning_rate": 8.809108128719428e-05, + "loss": 0.0109, + "step": 26310 + }, + { + "epoch": 4.0882261571916745, + "grad_norm": 0.25525885820388794, + "learning_rate": 8.808036824445424e-05, + "loss": 0.011, + "step": 26320 + }, + { + "epoch": 4.089779434607021, + "grad_norm": 0.4163278341293335, + "learning_rate": 8.806965103730238e-05, + "loss": 0.0088, + "step": 26330 + }, + { + "epoch": 4.091332712022367, + "grad_norm": 0.33526530861854553, + "learning_rate": 8.805892966691074e-05, + "loss": 0.0103, + "step": 26340 + }, + { + "epoch": 4.092885989437714, + "grad_norm": 0.29982054233551025, + "learning_rate": 8.804820413445175e-05, + "loss": 0.0081, + "step": 26350 + }, + { + "epoch": 4.09443926685306, + "grad_norm": 0.27063143253326416, + "learning_rate": 8.803747444109837e-05, + "loss": 0.0102, + "step": 26360 + }, + { + "epoch": 4.095992544268406, + "grad_norm": 0.42837411165237427, + "learning_rate": 8.802674058802399e-05, + "loss": 0.0111, + "step": 26370 + }, + { + "epoch": 4.097545821683752, + "grad_norm": 0.19062399864196777, + "learning_rate": 8.801600257640241e-05, + "loss": 0.0118, + "step": 26380 + }, + { + "epoch": 4.099099099099099, + "grad_norm": 0.2949468493461609, + "learning_rate": 8.800526040740795e-05, + "loss": 0.0101, + "step": 26390 + }, + { + "epoch": 4.100652376514446, + "grad_norm": 0.19501811265945435, + "learning_rate": 8.799451408221535e-05, + "loss": 0.0096, + "step": 26400 + }, + { + "epoch": 4.102205653929792, + "grad_norm": 0.2261330485343933, + "learning_rate": 8.798376360199982e-05, + "loss": 0.0115, + "step": 26410 + }, + { + "epoch": 4.103758931345138, + "grad_norm": 0.3032586872577667, + "learning_rate": 8.797300896793701e-05, + "loss": 0.0102, + "step": 26420 + }, + { + "epoch": 4.105312208760485, + "grad_norm": 0.37445974349975586, + "learning_rate": 8.796225018120302e-05, + "loss": 0.0117, + "step": 26430 + }, + { + "epoch": 4.106865486175831, + "grad_norm": 0.3304964601993561, + "learning_rate": 8.795148724297444e-05, + "loss": 0.0076, + "step": 26440 + }, + { + "epoch": 4.108418763591177, + "grad_norm": 0.394551157951355, + "learning_rate": 8.794072015442825e-05, + "loss": 0.0115, + "step": 26450 + }, + { + "epoch": 4.1099720410065235, + "grad_norm": 0.23947186768054962, + "learning_rate": 8.792994891674198e-05, + "loss": 0.0107, + "step": 26460 + }, + { + "epoch": 4.1115253184218705, + "grad_norm": 0.22495624423027039, + "learning_rate": 8.79191735310935e-05, + "loss": 0.009, + "step": 26470 + }, + { + "epoch": 4.113078595837217, + "grad_norm": 0.3074114918708801, + "learning_rate": 8.790839399866122e-05, + "loss": 0.0135, + "step": 26480 + }, + { + "epoch": 4.114631873252563, + "grad_norm": 0.18291887640953064, + "learning_rate": 8.789761032062397e-05, + "loss": 0.007, + "step": 26490 + }, + { + "epoch": 4.116185150667909, + "grad_norm": 0.3226327896118164, + "learning_rate": 8.788682249816103e-05, + "loss": 0.0104, + "step": 26500 + }, + { + "epoch": 4.117738428083256, + "grad_norm": 0.26781436800956726, + "learning_rate": 8.787603053245215e-05, + "loss": 0.0095, + "step": 26510 + }, + { + "epoch": 4.119291705498602, + "grad_norm": 0.26340052485466003, + "learning_rate": 8.78652344246775e-05, + "loss": 0.0093, + "step": 26520 + }, + { + "epoch": 4.120844982913948, + "grad_norm": 0.16798873245716095, + "learning_rate": 8.785443417601776e-05, + "loss": 0.0055, + "step": 26530 + }, + { + "epoch": 4.122398260329295, + "grad_norm": 0.23376111686229706, + "learning_rate": 8.784362978765401e-05, + "loss": 0.0112, + "step": 26540 + }, + { + "epoch": 4.123951537744642, + "grad_norm": 0.48119091987609863, + "learning_rate": 8.783282126076779e-05, + "loss": 0.0103, + "step": 26550 + }, + { + "epoch": 4.125504815159988, + "grad_norm": 0.21944378316402435, + "learning_rate": 8.782200859654112e-05, + "loss": 0.0083, + "step": 26560 + }, + { + "epoch": 4.127058092575334, + "grad_norm": 0.3793306052684784, + "learning_rate": 8.781119179615646e-05, + "loss": 0.0107, + "step": 26570 + }, + { + "epoch": 4.12861136999068, + "grad_norm": 0.23484860360622406, + "learning_rate": 8.780037086079674e-05, + "loss": 0.0117, + "step": 26580 + }, + { + "epoch": 4.130164647406026, + "grad_norm": 0.4005507528781891, + "learning_rate": 8.778954579164527e-05, + "loss": 0.0106, + "step": 26590 + }, + { + "epoch": 4.131717924821373, + "grad_norm": 0.4685084819793701, + "learning_rate": 8.777871658988588e-05, + "loss": 0.0111, + "step": 26600 + }, + { + "epoch": 4.1332712022367195, + "grad_norm": 0.4133804142475128, + "learning_rate": 8.776788325670285e-05, + "loss": 0.0112, + "step": 26610 + }, + { + "epoch": 4.134824479652066, + "grad_norm": 0.3481087386608124, + "learning_rate": 8.775704579328089e-05, + "loss": 0.008, + "step": 26620 + }, + { + "epoch": 4.136377757067412, + "grad_norm": 0.38422664999961853, + "learning_rate": 8.774620420080517e-05, + "loss": 0.01, + "step": 26630 + }, + { + "epoch": 4.137931034482759, + "grad_norm": 0.5089642405509949, + "learning_rate": 8.773535848046131e-05, + "loss": 0.0144, + "step": 26640 + }, + { + "epoch": 4.139484311898105, + "grad_norm": 0.38427412509918213, + "learning_rate": 8.772450863343538e-05, + "loss": 0.0127, + "step": 26650 + }, + { + "epoch": 4.141037589313451, + "grad_norm": 0.19446836411952972, + "learning_rate": 8.77136546609139e-05, + "loss": 0.011, + "step": 26660 + }, + { + "epoch": 4.142590866728797, + "grad_norm": 0.20245088636875153, + "learning_rate": 8.770279656408385e-05, + "loss": 0.0106, + "step": 26670 + }, + { + "epoch": 4.1441441441441444, + "grad_norm": 0.25516703724861145, + "learning_rate": 8.769193434413265e-05, + "loss": 0.0085, + "step": 26680 + }, + { + "epoch": 4.145697421559491, + "grad_norm": 0.23846903443336487, + "learning_rate": 8.76810680022482e-05, + "loss": 0.0118, + "step": 26690 + }, + { + "epoch": 4.147250698974837, + "grad_norm": 0.266055166721344, + "learning_rate": 8.767019753961878e-05, + "loss": 0.0122, + "step": 26700 + }, + { + "epoch": 4.148803976390183, + "grad_norm": 0.2942397892475128, + "learning_rate": 8.765932295743321e-05, + "loss": 0.0135, + "step": 26710 + }, + { + "epoch": 4.15035725380553, + "grad_norm": 0.3203454315662384, + "learning_rate": 8.764844425688068e-05, + "loss": 0.0125, + "step": 26720 + }, + { + "epoch": 4.151910531220876, + "grad_norm": 0.3913573622703552, + "learning_rate": 8.763756143915092e-05, + "loss": 0.0126, + "step": 26730 + }, + { + "epoch": 4.153463808636222, + "grad_norm": 0.31446972489356995, + "learning_rate": 8.7626674505434e-05, + "loss": 0.0124, + "step": 26740 + }, + { + "epoch": 4.1550170860515685, + "grad_norm": 0.8651464581489563, + "learning_rate": 8.761578345692053e-05, + "loss": 0.0123, + "step": 26750 + }, + { + "epoch": 4.1565703634669156, + "grad_norm": 0.39737623929977417, + "learning_rate": 8.760488829480156e-05, + "loss": 0.0098, + "step": 26760 + }, + { + "epoch": 4.158123640882262, + "grad_norm": 0.3397308588027954, + "learning_rate": 8.759398902026854e-05, + "loss": 0.0118, + "step": 26770 + }, + { + "epoch": 4.159676918297608, + "grad_norm": 0.37220609188079834, + "learning_rate": 8.758308563451339e-05, + "loss": 0.0108, + "step": 26780 + }, + { + "epoch": 4.161230195712954, + "grad_norm": 0.3386576473712921, + "learning_rate": 8.75721781387285e-05, + "loss": 0.013, + "step": 26790 + }, + { + "epoch": 4.162783473128301, + "grad_norm": 0.23816560208797455, + "learning_rate": 8.75612665341067e-05, + "loss": 0.0106, + "step": 26800 + }, + { + "epoch": 4.164336750543647, + "grad_norm": 0.28871220350265503, + "learning_rate": 8.755035082184126e-05, + "loss": 0.0106, + "step": 26810 + }, + { + "epoch": 4.165890027958993, + "grad_norm": 0.20686101913452148, + "learning_rate": 8.753943100312592e-05, + "loss": 0.0095, + "step": 26820 + }, + { + "epoch": 4.16744330537434, + "grad_norm": 0.29717808961868286, + "learning_rate": 8.752850707915484e-05, + "loss": 0.0097, + "step": 26830 + }, + { + "epoch": 4.168996582789687, + "grad_norm": 0.3389827311038971, + "learning_rate": 8.751757905112264e-05, + "loss": 0.0105, + "step": 26840 + }, + { + "epoch": 4.170549860205033, + "grad_norm": 0.23110045492649078, + "learning_rate": 8.75066469202244e-05, + "loss": 0.0068, + "step": 26850 + }, + { + "epoch": 4.172103137620379, + "grad_norm": 1.1369657516479492, + "learning_rate": 8.749571068765567e-05, + "loss": 0.0114, + "step": 26860 + }, + { + "epoch": 4.173656415035725, + "grad_norm": 0.41711997985839844, + "learning_rate": 8.748477035461238e-05, + "loss": 0.0121, + "step": 26870 + }, + { + "epoch": 4.175209692451071, + "grad_norm": 0.2981308400630951, + "learning_rate": 8.747382592229095e-05, + "loss": 0.0139, + "step": 26880 + }, + { + "epoch": 4.176762969866418, + "grad_norm": 0.2487267106771469, + "learning_rate": 8.746287739188828e-05, + "loss": 0.0094, + "step": 26890 + }, + { + "epoch": 4.1783162472817645, + "grad_norm": 0.23986934125423431, + "learning_rate": 8.745192476460165e-05, + "loss": 0.0116, + "step": 26900 + }, + { + "epoch": 4.179869524697111, + "grad_norm": 0.2001851499080658, + "learning_rate": 8.744096804162882e-05, + "loss": 0.0128, + "step": 26910 + }, + { + "epoch": 4.181422802112457, + "grad_norm": 0.22737166285514832, + "learning_rate": 8.743000722416804e-05, + "loss": 0.0114, + "step": 26920 + }, + { + "epoch": 4.182976079527804, + "grad_norm": 0.2587239742279053, + "learning_rate": 8.741904231341793e-05, + "loss": 0.0088, + "step": 26930 + }, + { + "epoch": 4.18452935694315, + "grad_norm": 0.20237065851688385, + "learning_rate": 8.740807331057762e-05, + "loss": 0.0106, + "step": 26940 + }, + { + "epoch": 4.186082634358496, + "grad_norm": 0.255438894033432, + "learning_rate": 8.739710021684667e-05, + "loss": 0.0106, + "step": 26950 + }, + { + "epoch": 4.187635911773842, + "grad_norm": 0.356117844581604, + "learning_rate": 8.738612303342503e-05, + "loss": 0.0129, + "step": 26960 + }, + { + "epoch": 4.1891891891891895, + "grad_norm": 0.2961748540401459, + "learning_rate": 8.73751417615132e-05, + "loss": 0.0105, + "step": 26970 + }, + { + "epoch": 4.190742466604536, + "grad_norm": 0.27660655975341797, + "learning_rate": 8.736415640231208e-05, + "loss": 0.0083, + "step": 26980 + }, + { + "epoch": 4.192295744019882, + "grad_norm": 0.22245408594608307, + "learning_rate": 8.735316695702297e-05, + "loss": 0.0118, + "step": 26990 + }, + { + "epoch": 4.193849021435228, + "grad_norm": 0.36561325192451477, + "learning_rate": 8.734217342684769e-05, + "loss": 0.0106, + "step": 27000 + }, + { + "epoch": 4.195402298850575, + "grad_norm": 0.21226204931735992, + "learning_rate": 8.733117581298847e-05, + "loss": 0.0105, + "step": 27010 + }, + { + "epoch": 4.196955576265921, + "grad_norm": 0.1969742327928543, + "learning_rate": 8.732017411664796e-05, + "loss": 0.0086, + "step": 27020 + }, + { + "epoch": 4.198508853681267, + "grad_norm": 0.30170589685440063, + "learning_rate": 8.730916833902936e-05, + "loss": 0.0083, + "step": 27030 + }, + { + "epoch": 4.2000621310966135, + "grad_norm": 0.2216322124004364, + "learning_rate": 8.729815848133618e-05, + "loss": 0.0091, + "step": 27040 + }, + { + "epoch": 4.201615408511961, + "grad_norm": 0.22927029430866241, + "learning_rate": 8.728714454477247e-05, + "loss": 0.01, + "step": 27050 + }, + { + "epoch": 4.203168685927307, + "grad_norm": 0.20858632028102875, + "learning_rate": 8.727612653054269e-05, + "loss": 0.0085, + "step": 27060 + }, + { + "epoch": 4.204721963342653, + "grad_norm": 0.25757288932800293, + "learning_rate": 8.726510443985176e-05, + "loss": 0.0085, + "step": 27070 + }, + { + "epoch": 4.206275240757999, + "grad_norm": 0.2555765211582184, + "learning_rate": 8.725407827390503e-05, + "loss": 0.0097, + "step": 27080 + }, + { + "epoch": 4.207828518173346, + "grad_norm": 0.2576732635498047, + "learning_rate": 8.724304803390833e-05, + "loss": 0.0121, + "step": 27090 + }, + { + "epoch": 4.209381795588692, + "grad_norm": 0.3679388165473938, + "learning_rate": 8.723201372106788e-05, + "loss": 0.012, + "step": 27100 + }, + { + "epoch": 4.2109350730040385, + "grad_norm": 0.328489750623703, + "learning_rate": 8.722097533659038e-05, + "loss": 0.0125, + "step": 27110 + }, + { + "epoch": 4.212488350419385, + "grad_norm": 0.3499918580055237, + "learning_rate": 8.720993288168299e-05, + "loss": 0.008, + "step": 27120 + }, + { + "epoch": 4.214041627834732, + "grad_norm": 1.3002455234527588, + "learning_rate": 8.719888635755327e-05, + "loss": 0.0109, + "step": 27130 + }, + { + "epoch": 4.215594905250078, + "grad_norm": 0.24494433403015137, + "learning_rate": 8.718783576540928e-05, + "loss": 0.01, + "step": 27140 + }, + { + "epoch": 4.217148182665424, + "grad_norm": 0.3208041489124298, + "learning_rate": 8.717678110645948e-05, + "loss": 0.0125, + "step": 27150 + }, + { + "epoch": 4.21870146008077, + "grad_norm": 0.1988697648048401, + "learning_rate": 8.716572238191279e-05, + "loss": 0.009, + "step": 27160 + }, + { + "epoch": 4.220254737496117, + "grad_norm": 0.3015557527542114, + "learning_rate": 8.715465959297857e-05, + "loss": 0.0082, + "step": 27170 + }, + { + "epoch": 4.221808014911463, + "grad_norm": 0.20649652183055878, + "learning_rate": 8.714359274086665e-05, + "loss": 0.0125, + "step": 27180 + }, + { + "epoch": 4.22336129232681, + "grad_norm": 0.3108457922935486, + "learning_rate": 8.713252182678726e-05, + "loss": 0.012, + "step": 27190 + }, + { + "epoch": 4.224914569742156, + "grad_norm": 0.23401814699172974, + "learning_rate": 8.712144685195112e-05, + "loss": 0.0114, + "step": 27200 + }, + { + "epoch": 4.226467847157502, + "grad_norm": 0.20038282871246338, + "learning_rate": 8.711036781756936e-05, + "loss": 0.0089, + "step": 27210 + }, + { + "epoch": 4.228021124572849, + "grad_norm": 0.27527451515197754, + "learning_rate": 8.709928472485357e-05, + "loss": 0.0096, + "step": 27220 + }, + { + "epoch": 4.229574401988195, + "grad_norm": 0.30440977215766907, + "learning_rate": 8.708819757501579e-05, + "loss": 0.0096, + "step": 27230 + }, + { + "epoch": 4.231127679403541, + "grad_norm": 0.32694190740585327, + "learning_rate": 8.707710636926846e-05, + "loss": 0.009, + "step": 27240 + }, + { + "epoch": 4.2326809568188875, + "grad_norm": 0.20158331096172333, + "learning_rate": 8.706601110882455e-05, + "loss": 0.0135, + "step": 27250 + }, + { + "epoch": 4.2342342342342345, + "grad_norm": 0.33561593294143677, + "learning_rate": 8.705491179489738e-05, + "loss": 0.0126, + "step": 27260 + }, + { + "epoch": 4.235787511649581, + "grad_norm": 0.22883471846580505, + "learning_rate": 8.704380842870077e-05, + "loss": 0.0079, + "step": 27270 + }, + { + "epoch": 4.237340789064927, + "grad_norm": 0.31381693482398987, + "learning_rate": 8.703270101144895e-05, + "loss": 0.0085, + "step": 27280 + }, + { + "epoch": 4.238894066480273, + "grad_norm": 0.3057848811149597, + "learning_rate": 8.702158954435664e-05, + "loss": 0.0117, + "step": 27290 + }, + { + "epoch": 4.24044734389562, + "grad_norm": 0.21099326014518738, + "learning_rate": 8.701047402863896e-05, + "loss": 0.011, + "step": 27300 + }, + { + "epoch": 4.242000621310966, + "grad_norm": 0.26181891560554504, + "learning_rate": 8.699935446551148e-05, + "loss": 0.0114, + "step": 27310 + }, + { + "epoch": 4.243553898726312, + "grad_norm": 0.21261043846607208, + "learning_rate": 8.698823085619022e-05, + "loss": 0.0137, + "step": 27320 + }, + { + "epoch": 4.245107176141659, + "grad_norm": 0.2379794865846634, + "learning_rate": 8.697710320189166e-05, + "loss": 0.012, + "step": 27330 + }, + { + "epoch": 4.246660453557006, + "grad_norm": 0.28181248903274536, + "learning_rate": 8.696597150383268e-05, + "loss": 0.0107, + "step": 27340 + }, + { + "epoch": 4.248213730972352, + "grad_norm": 0.42979148030281067, + "learning_rate": 8.695483576323063e-05, + "loss": 0.0139, + "step": 27350 + }, + { + "epoch": 4.249767008387698, + "grad_norm": 0.6043772101402283, + "learning_rate": 8.69436959813033e-05, + "loss": 0.0112, + "step": 27360 + }, + { + "epoch": 4.251320285803044, + "grad_norm": 0.24529743194580078, + "learning_rate": 8.693255215926892e-05, + "loss": 0.0111, + "step": 27370 + }, + { + "epoch": 4.252873563218391, + "grad_norm": 0.3021644055843353, + "learning_rate": 8.692140429834617e-05, + "loss": 0.0126, + "step": 27380 + }, + { + "epoch": 4.254426840633737, + "grad_norm": 0.20741386711597443, + "learning_rate": 8.691025239975415e-05, + "loss": 0.0092, + "step": 27390 + }, + { + "epoch": 4.2559801180490835, + "grad_norm": 0.20903001725673676, + "learning_rate": 8.689909646471243e-05, + "loss": 0.011, + "step": 27400 + }, + { + "epoch": 4.25753339546443, + "grad_norm": 0.1911519467830658, + "learning_rate": 8.688793649444099e-05, + "loss": 0.0093, + "step": 27410 + }, + { + "epoch": 4.259086672879777, + "grad_norm": 0.4572712481021881, + "learning_rate": 8.687677249016029e-05, + "loss": 0.0103, + "step": 27420 + }, + { + "epoch": 4.260639950295123, + "grad_norm": 0.25243061780929565, + "learning_rate": 8.686560445309118e-05, + "loss": 0.0093, + "step": 27430 + }, + { + "epoch": 4.262193227710469, + "grad_norm": 0.22536365687847137, + "learning_rate": 8.685443238445499e-05, + "loss": 0.0085, + "step": 27440 + }, + { + "epoch": 4.263746505125815, + "grad_norm": 0.24140147864818573, + "learning_rate": 8.68432562854735e-05, + "loss": 0.0101, + "step": 27450 + }, + { + "epoch": 4.265299782541162, + "grad_norm": 0.5709446668624878, + "learning_rate": 8.683207615736887e-05, + "loss": 0.0155, + "step": 27460 + }, + { + "epoch": 4.266853059956508, + "grad_norm": 0.3240688145160675, + "learning_rate": 8.682089200136379e-05, + "loss": 0.0103, + "step": 27470 + }, + { + "epoch": 4.268406337371855, + "grad_norm": 0.2931091785430908, + "learning_rate": 8.680970381868132e-05, + "loss": 0.0115, + "step": 27480 + }, + { + "epoch": 4.269959614787201, + "grad_norm": 0.3986968994140625, + "learning_rate": 8.679851161054498e-05, + "loss": 0.0123, + "step": 27490 + }, + { + "epoch": 4.271512892202548, + "grad_norm": 0.290676474571228, + "learning_rate": 8.678731537817873e-05, + "loss": 0.0097, + "step": 27500 + }, + { + "epoch": 4.273066169617894, + "grad_norm": 0.2173018902540207, + "learning_rate": 8.677611512280697e-05, + "loss": 0.0114, + "step": 27510 + }, + { + "epoch": 4.27461944703324, + "grad_norm": 0.36577630043029785, + "learning_rate": 8.676491084565457e-05, + "loss": 0.011, + "step": 27520 + }, + { + "epoch": 4.276172724448586, + "grad_norm": 0.4633292257785797, + "learning_rate": 8.675370254794678e-05, + "loss": 0.0092, + "step": 27530 + }, + { + "epoch": 4.2777260018639325, + "grad_norm": 0.20974749326705933, + "learning_rate": 8.674249023090935e-05, + "loss": 0.009, + "step": 27540 + }, + { + "epoch": 4.2792792792792795, + "grad_norm": 0.21844184398651123, + "learning_rate": 8.673127389576843e-05, + "loss": 0.0084, + "step": 27550 + }, + { + "epoch": 4.280832556694626, + "grad_norm": 0.26549232006073, + "learning_rate": 8.67200535437506e-05, + "loss": 0.0102, + "step": 27560 + }, + { + "epoch": 4.282385834109972, + "grad_norm": 0.2327466756105423, + "learning_rate": 8.670882917608296e-05, + "loss": 0.011, + "step": 27570 + }, + { + "epoch": 4.283939111525318, + "grad_norm": 0.20930948853492737, + "learning_rate": 8.669760079399292e-05, + "loss": 0.0083, + "step": 27580 + }, + { + "epoch": 4.285492388940665, + "grad_norm": 0.2428317666053772, + "learning_rate": 8.668636839870845e-05, + "loss": 0.0104, + "step": 27590 + }, + { + "epoch": 4.287045666356011, + "grad_norm": 0.5992583632469177, + "learning_rate": 8.667513199145789e-05, + "loss": 0.012, + "step": 27600 + }, + { + "epoch": 4.288598943771357, + "grad_norm": 0.3265366554260254, + "learning_rate": 8.666389157347002e-05, + "loss": 0.0109, + "step": 27610 + }, + { + "epoch": 4.290152221186704, + "grad_norm": 0.37226974964141846, + "learning_rate": 8.66526471459741e-05, + "loss": 0.0084, + "step": 27620 + }, + { + "epoch": 4.291705498602051, + "grad_norm": 0.32580187916755676, + "learning_rate": 8.66413987101998e-05, + "loss": 0.0092, + "step": 27630 + }, + { + "epoch": 4.293258776017397, + "grad_norm": 0.20853270590305328, + "learning_rate": 8.663014626737723e-05, + "loss": 0.01, + "step": 27640 + }, + { + "epoch": 4.294812053432743, + "grad_norm": 0.29566171765327454, + "learning_rate": 8.661888981873691e-05, + "loss": 0.013, + "step": 27650 + }, + { + "epoch": 4.296365330848089, + "grad_norm": 0.3686099946498871, + "learning_rate": 8.660762936550988e-05, + "loss": 0.0105, + "step": 27660 + }, + { + "epoch": 4.297918608263436, + "grad_norm": 0.2603360116481781, + "learning_rate": 8.659636490892753e-05, + "loss": 0.0146, + "step": 27670 + }, + { + "epoch": 4.299471885678782, + "grad_norm": 0.2379627376794815, + "learning_rate": 8.658509645022174e-05, + "loss": 0.0105, + "step": 27680 + }, + { + "epoch": 4.3010251630941285, + "grad_norm": 0.2719678282737732, + "learning_rate": 8.657382399062481e-05, + "loss": 0.0084, + "step": 27690 + }, + { + "epoch": 4.302578440509475, + "grad_norm": 0.18398065865039825, + "learning_rate": 8.656254753136946e-05, + "loss": 0.0092, + "step": 27700 + }, + { + "epoch": 4.304131717924822, + "grad_norm": 0.19533342123031616, + "learning_rate": 8.655126707368891e-05, + "loss": 0.0125, + "step": 27710 + }, + { + "epoch": 4.305684995340168, + "grad_norm": 0.24650630354881287, + "learning_rate": 8.653998261881672e-05, + "loss": 0.0096, + "step": 27720 + }, + { + "epoch": 4.307238272755514, + "grad_norm": 0.21396535634994507, + "learning_rate": 8.652869416798699e-05, + "loss": 0.0125, + "step": 27730 + }, + { + "epoch": 4.30879155017086, + "grad_norm": 0.24923405051231384, + "learning_rate": 8.651740172243417e-05, + "loss": 0.0101, + "step": 27740 + }, + { + "epoch": 4.310344827586207, + "grad_norm": 0.2211877405643463, + "learning_rate": 8.65061052833932e-05, + "loss": 0.0105, + "step": 27750 + }, + { + "epoch": 4.3118981050015535, + "grad_norm": 0.22116230428218842, + "learning_rate": 8.649480485209945e-05, + "loss": 0.0085, + "step": 27760 + }, + { + "epoch": 4.3134513824169, + "grad_norm": 0.528913676738739, + "learning_rate": 8.64835004297887e-05, + "loss": 0.0094, + "step": 27770 + }, + { + "epoch": 4.315004659832246, + "grad_norm": 0.3609791696071625, + "learning_rate": 8.64721920176972e-05, + "loss": 0.0085, + "step": 27780 + }, + { + "epoch": 4.316557937247593, + "grad_norm": 0.4972170293331146, + "learning_rate": 8.646087961706164e-05, + "loss": 0.0125, + "step": 27790 + }, + { + "epoch": 4.318111214662939, + "grad_norm": 0.2546960711479187, + "learning_rate": 8.644956322911908e-05, + "loss": 0.0085, + "step": 27800 + }, + { + "epoch": 4.319664492078285, + "grad_norm": 0.30179232358932495, + "learning_rate": 8.643824285510709e-05, + "loss": 0.0118, + "step": 27810 + }, + { + "epoch": 4.321217769493631, + "grad_norm": 0.2615557014942169, + "learning_rate": 8.642691849626364e-05, + "loss": 0.0117, + "step": 27820 + }, + { + "epoch": 4.3227710469089775, + "grad_norm": 0.4213070869445801, + "learning_rate": 8.641559015382717e-05, + "loss": 0.0122, + "step": 27830 + }, + { + "epoch": 4.324324324324325, + "grad_norm": 0.24562326073646545, + "learning_rate": 8.640425782903649e-05, + "loss": 0.0086, + "step": 27840 + }, + { + "epoch": 4.325877601739671, + "grad_norm": 0.21290798485279083, + "learning_rate": 8.639292152313091e-05, + "loss": 0.0125, + "step": 27850 + }, + { + "epoch": 4.327430879155017, + "grad_norm": 0.1999458372592926, + "learning_rate": 8.638158123735015e-05, + "loss": 0.0082, + "step": 27860 + }, + { + "epoch": 4.328984156570363, + "grad_norm": 0.22996550798416138, + "learning_rate": 8.637023697293436e-05, + "loss": 0.0083, + "step": 27870 + }, + { + "epoch": 4.33053743398571, + "grad_norm": 0.20581656694412231, + "learning_rate": 8.635888873112414e-05, + "loss": 0.0104, + "step": 27880 + }, + { + "epoch": 4.332090711401056, + "grad_norm": 0.4621066451072693, + "learning_rate": 8.634753651316052e-05, + "loss": 0.0114, + "step": 27890 + }, + { + "epoch": 4.3336439888164024, + "grad_norm": 0.353061705827713, + "learning_rate": 8.633618032028496e-05, + "loss": 0.0111, + "step": 27900 + }, + { + "epoch": 4.335197266231749, + "grad_norm": 0.35136839747428894, + "learning_rate": 8.632482015373934e-05, + "loss": 0.0084, + "step": 27910 + }, + { + "epoch": 4.336750543647096, + "grad_norm": 0.20413903892040253, + "learning_rate": 8.6313456014766e-05, + "loss": 0.0098, + "step": 27920 + }, + { + "epoch": 4.338303821062442, + "grad_norm": 0.16001178324222565, + "learning_rate": 8.630208790460771e-05, + "loss": 0.0114, + "step": 27930 + }, + { + "epoch": 4.339857098477788, + "grad_norm": 0.26592177152633667, + "learning_rate": 8.629071582450768e-05, + "loss": 0.0133, + "step": 27940 + }, + { + "epoch": 4.341410375893134, + "grad_norm": 0.28810450434684753, + "learning_rate": 8.62793397757095e-05, + "loss": 0.011, + "step": 27950 + }, + { + "epoch": 4.342963653308481, + "grad_norm": 0.2768503427505493, + "learning_rate": 8.626795975945729e-05, + "loss": 0.0139, + "step": 27960 + }, + { + "epoch": 4.344516930723827, + "grad_norm": 0.2100571095943451, + "learning_rate": 8.625657577699551e-05, + "loss": 0.0089, + "step": 27970 + }, + { + "epoch": 4.3460702081391736, + "grad_norm": 0.36544227600097656, + "learning_rate": 8.624518782956914e-05, + "loss": 0.0099, + "step": 27980 + }, + { + "epoch": 4.34762348555452, + "grad_norm": 0.2865404486656189, + "learning_rate": 8.62337959184235e-05, + "loss": 0.0106, + "step": 27990 + }, + { + "epoch": 4.349176762969867, + "grad_norm": 0.22928491234779358, + "learning_rate": 8.622240004480441e-05, + "loss": 0.0094, + "step": 28000 + }, + { + "epoch": 4.350730040385213, + "grad_norm": 0.5176405906677246, + "learning_rate": 8.621100020995814e-05, + "loss": 0.0103, + "step": 28010 + }, + { + "epoch": 4.352283317800559, + "grad_norm": 0.24855273962020874, + "learning_rate": 8.619959641513132e-05, + "loss": 0.0116, + "step": 28020 + }, + { + "epoch": 4.353836595215905, + "grad_norm": 0.15535973012447357, + "learning_rate": 8.618818866157105e-05, + "loss": 0.014, + "step": 28030 + }, + { + "epoch": 4.355389872631252, + "grad_norm": 0.3573170602321625, + "learning_rate": 8.617677695052487e-05, + "loss": 0.0076, + "step": 28040 + }, + { + "epoch": 4.3569431500465985, + "grad_norm": 0.4400773048400879, + "learning_rate": 8.616536128324078e-05, + "loss": 0.01, + "step": 28050 + }, + { + "epoch": 4.358496427461945, + "grad_norm": 0.2179245799779892, + "learning_rate": 8.615394166096712e-05, + "loss": 0.0089, + "step": 28060 + }, + { + "epoch": 4.360049704877291, + "grad_norm": 0.2164163738489151, + "learning_rate": 8.614251808495279e-05, + "loss": 0.0075, + "step": 28070 + }, + { + "epoch": 4.361602982292638, + "grad_norm": 0.23472419381141663, + "learning_rate": 8.6131090556447e-05, + "loss": 0.0125, + "step": 28080 + }, + { + "epoch": 4.363156259707984, + "grad_norm": 0.25656381249427795, + "learning_rate": 8.611965907669947e-05, + "loss": 0.0125, + "step": 28090 + }, + { + "epoch": 4.36470953712333, + "grad_norm": 0.4279372990131378, + "learning_rate": 8.610822364696034e-05, + "loss": 0.013, + "step": 28100 + }, + { + "epoch": 4.366262814538676, + "grad_norm": 0.30041393637657166, + "learning_rate": 8.609678426848015e-05, + "loss": 0.0175, + "step": 28110 + }, + { + "epoch": 4.3678160919540225, + "grad_norm": 0.39621150493621826, + "learning_rate": 8.60853409425099e-05, + "loss": 0.0133, + "step": 28120 + }, + { + "epoch": 4.36936936936937, + "grad_norm": 0.24698469042778015, + "learning_rate": 8.607389367030104e-05, + "loss": 0.0097, + "step": 28130 + }, + { + "epoch": 4.370922646784716, + "grad_norm": 0.2909940183162689, + "learning_rate": 8.606244245310538e-05, + "loss": 0.0098, + "step": 28140 + }, + { + "epoch": 4.372475924200062, + "grad_norm": 0.3502448797225952, + "learning_rate": 8.605098729217525e-05, + "loss": 0.0079, + "step": 28150 + }, + { + "epoch": 4.374029201615409, + "grad_norm": 0.2694975435733795, + "learning_rate": 8.603952818876335e-05, + "loss": 0.0075, + "step": 28160 + }, + { + "epoch": 4.375582479030755, + "grad_norm": 0.19490091502666473, + "learning_rate": 8.602806514412281e-05, + "loss": 0.0104, + "step": 28170 + }, + { + "epoch": 4.377135756446101, + "grad_norm": 0.29678481817245483, + "learning_rate": 8.601659815950726e-05, + "loss": 0.0085, + "step": 28180 + }, + { + "epoch": 4.3786890338614475, + "grad_norm": 0.19394166767597198, + "learning_rate": 8.600512723617067e-05, + "loss": 0.0093, + "step": 28190 + }, + { + "epoch": 4.380242311276794, + "grad_norm": 0.1782529354095459, + "learning_rate": 8.59936523753675e-05, + "loss": 0.0109, + "step": 28200 + }, + { + "epoch": 4.381795588692141, + "grad_norm": 0.3304471969604492, + "learning_rate": 8.598217357835264e-05, + "loss": 0.0109, + "step": 28210 + }, + { + "epoch": 4.383348866107487, + "grad_norm": 0.24337264895439148, + "learning_rate": 8.597069084638135e-05, + "loss": 0.0094, + "step": 28220 + }, + { + "epoch": 4.384902143522833, + "grad_norm": 0.5109863877296448, + "learning_rate": 8.595920418070939e-05, + "loss": 0.0107, + "step": 28230 + }, + { + "epoch": 4.386455420938179, + "grad_norm": 0.329284131526947, + "learning_rate": 8.594771358259295e-05, + "loss": 0.0102, + "step": 28240 + }, + { + "epoch": 4.388008698353526, + "grad_norm": 0.2995052635669708, + "learning_rate": 8.593621905328858e-05, + "loss": 0.009, + "step": 28250 + }, + { + "epoch": 4.389561975768872, + "grad_norm": 0.34430649876594543, + "learning_rate": 8.592472059405333e-05, + "loss": 0.0119, + "step": 28260 + }, + { + "epoch": 4.391115253184219, + "grad_norm": 0.23743566870689392, + "learning_rate": 8.591321820614464e-05, + "loss": 0.0095, + "step": 28270 + }, + { + "epoch": 4.392668530599565, + "grad_norm": 0.25621262192726135, + "learning_rate": 8.590171189082041e-05, + "loss": 0.0075, + "step": 28280 + }, + { + "epoch": 4.394221808014912, + "grad_norm": 0.1586337834596634, + "learning_rate": 8.589020164933894e-05, + "loss": 0.008, + "step": 28290 + }, + { + "epoch": 4.395775085430258, + "grad_norm": 0.3561062812805176, + "learning_rate": 8.587868748295898e-05, + "loss": 0.009, + "step": 28300 + }, + { + "epoch": 4.397328362845604, + "grad_norm": 0.20366743206977844, + "learning_rate": 8.586716939293971e-05, + "loss": 0.0092, + "step": 28310 + }, + { + "epoch": 4.39888164026095, + "grad_norm": 0.13328053057193756, + "learning_rate": 8.58556473805407e-05, + "loss": 0.0096, + "step": 28320 + }, + { + "epoch": 4.400434917676297, + "grad_norm": 0.3458876609802246, + "learning_rate": 8.584412144702202e-05, + "loss": 0.0107, + "step": 28330 + }, + { + "epoch": 4.4019881950916435, + "grad_norm": 0.3298887610435486, + "learning_rate": 8.58325915936441e-05, + "loss": 0.0104, + "step": 28340 + }, + { + "epoch": 4.40354147250699, + "grad_norm": 0.2675412893295288, + "learning_rate": 8.582105782166783e-05, + "loss": 0.0088, + "step": 28350 + }, + { + "epoch": 4.405094749922336, + "grad_norm": 0.2858831584453583, + "learning_rate": 8.580952013235455e-05, + "loss": 0.0084, + "step": 28360 + }, + { + "epoch": 4.406648027337683, + "grad_norm": 0.16477815806865692, + "learning_rate": 8.579797852696596e-05, + "loss": 0.0085, + "step": 28370 + }, + { + "epoch": 4.408201304753029, + "grad_norm": 0.24305929243564606, + "learning_rate": 8.578643300676428e-05, + "loss": 0.0112, + "step": 28380 + }, + { + "epoch": 4.409754582168375, + "grad_norm": 0.2686951756477356, + "learning_rate": 8.577488357301209e-05, + "loss": 0.008, + "step": 28390 + }, + { + "epoch": 4.411307859583721, + "grad_norm": 0.17871896922588348, + "learning_rate": 8.576333022697242e-05, + "loss": 0.0098, + "step": 28400 + }, + { + "epoch": 4.412861136999068, + "grad_norm": 0.3945523500442505, + "learning_rate": 8.575177296990873e-05, + "loss": 0.0093, + "step": 28410 + }, + { + "epoch": 4.414414414414415, + "grad_norm": 0.1605132818222046, + "learning_rate": 8.574021180308489e-05, + "loss": 0.0101, + "step": 28420 + }, + { + "epoch": 4.415967691829761, + "grad_norm": 0.1635628044605255, + "learning_rate": 8.572864672776523e-05, + "loss": 0.0097, + "step": 28430 + }, + { + "epoch": 4.417520969245107, + "grad_norm": 0.2514294683933258, + "learning_rate": 8.571707774521447e-05, + "loss": 0.0113, + "step": 28440 + }, + { + "epoch": 4.419074246660454, + "grad_norm": 0.24078145623207092, + "learning_rate": 8.57055048566978e-05, + "loss": 0.0111, + "step": 28450 + }, + { + "epoch": 4.4206275240758, + "grad_norm": 0.30480432510375977, + "learning_rate": 8.569392806348078e-05, + "loss": 0.0108, + "step": 28460 + }, + { + "epoch": 4.422180801491146, + "grad_norm": 0.1397799253463745, + "learning_rate": 8.568234736682947e-05, + "loss": 0.0089, + "step": 28470 + }, + { + "epoch": 4.4237340789064925, + "grad_norm": 0.27092549204826355, + "learning_rate": 8.567076276801029e-05, + "loss": 0.0084, + "step": 28480 + }, + { + "epoch": 4.425287356321839, + "grad_norm": 0.23487086594104767, + "learning_rate": 8.565917426829013e-05, + "loss": 0.0116, + "step": 28490 + }, + { + "epoch": 4.426840633737186, + "grad_norm": 0.9666440486907959, + "learning_rate": 8.564758186893628e-05, + "loss": 0.012, + "step": 28500 + }, + { + "epoch": 4.428393911152532, + "grad_norm": 0.3206287920475006, + "learning_rate": 8.563598557121649e-05, + "loss": 0.0091, + "step": 28510 + }, + { + "epoch": 4.429947188567878, + "grad_norm": 0.2541646957397461, + "learning_rate": 8.562438537639888e-05, + "loss": 0.0082, + "step": 28520 + }, + { + "epoch": 4.431500465983224, + "grad_norm": 0.3480575978755951, + "learning_rate": 8.561278128575206e-05, + "loss": 0.0089, + "step": 28530 + }, + { + "epoch": 4.433053743398571, + "grad_norm": 0.23702281713485718, + "learning_rate": 8.5601173300545e-05, + "loss": 0.012, + "step": 28540 + }, + { + "epoch": 4.434607020813917, + "grad_norm": 0.26086971163749695, + "learning_rate": 8.558956142204717e-05, + "loss": 0.0088, + "step": 28550 + }, + { + "epoch": 4.436160298229264, + "grad_norm": 0.18044021725654602, + "learning_rate": 8.55779456515284e-05, + "loss": 0.0068, + "step": 28560 + }, + { + "epoch": 4.43771357564461, + "grad_norm": 0.3705526888370514, + "learning_rate": 8.556632599025898e-05, + "loss": 0.0078, + "step": 28570 + }, + { + "epoch": 4.439266853059957, + "grad_norm": 0.2727314829826355, + "learning_rate": 8.555470243950964e-05, + "loss": 0.0081, + "step": 28580 + }, + { + "epoch": 4.440820130475303, + "grad_norm": 0.2696983516216278, + "learning_rate": 8.554307500055148e-05, + "loss": 0.0073, + "step": 28590 + }, + { + "epoch": 4.442373407890649, + "grad_norm": 0.2494751513004303, + "learning_rate": 8.553144367465609e-05, + "loss": 0.008, + "step": 28600 + }, + { + "epoch": 4.443926685305995, + "grad_norm": 0.19970279932022095, + "learning_rate": 8.551980846309544e-05, + "loss": 0.009, + "step": 28610 + }, + { + "epoch": 4.445479962721342, + "grad_norm": 0.30506181716918945, + "learning_rate": 8.550816936714193e-05, + "loss": 0.0151, + "step": 28620 + }, + { + "epoch": 4.4470332401366885, + "grad_norm": 0.28541961312294006, + "learning_rate": 8.549652638806841e-05, + "loss": 0.0086, + "step": 28630 + }, + { + "epoch": 4.448586517552035, + "grad_norm": 0.3602825701236725, + "learning_rate": 8.548487952714812e-05, + "loss": 0.0141, + "step": 28640 + }, + { + "epoch": 4.450139794967381, + "grad_norm": 0.14547154307365417, + "learning_rate": 8.547322878565478e-05, + "loss": 0.011, + "step": 28650 + }, + { + "epoch": 4.451693072382728, + "grad_norm": 0.4600008726119995, + "learning_rate": 8.546157416486245e-05, + "loss": 0.0089, + "step": 28660 + }, + { + "epoch": 4.453246349798074, + "grad_norm": 0.3944399654865265, + "learning_rate": 8.54499156660457e-05, + "loss": 0.0094, + "step": 28670 + }, + { + "epoch": 4.45479962721342, + "grad_norm": 0.2727721035480499, + "learning_rate": 8.543825329047947e-05, + "loss": 0.0135, + "step": 28680 + }, + { + "epoch": 4.456352904628766, + "grad_norm": 0.7028254866600037, + "learning_rate": 8.542658703943913e-05, + "loss": 0.0102, + "step": 28690 + }, + { + "epoch": 4.4579061820441135, + "grad_norm": 0.4322689473628998, + "learning_rate": 8.541491691420051e-05, + "loss": 0.0094, + "step": 28700 + }, + { + "epoch": 4.45945945945946, + "grad_norm": 0.3783259987831116, + "learning_rate": 8.54032429160398e-05, + "loss": 0.0088, + "step": 28710 + }, + { + "epoch": 4.461012736874806, + "grad_norm": 0.20531898736953735, + "learning_rate": 8.539156504623369e-05, + "loss": 0.0098, + "step": 28720 + }, + { + "epoch": 4.462566014290152, + "grad_norm": 0.1631566733121872, + "learning_rate": 8.537988330605923e-05, + "loss": 0.0086, + "step": 28730 + }, + { + "epoch": 4.464119291705499, + "grad_norm": 0.515092134475708, + "learning_rate": 8.536819769679393e-05, + "loss": 0.0066, + "step": 28740 + }, + { + "epoch": 4.465672569120845, + "grad_norm": 0.3079532980918884, + "learning_rate": 8.53565082197157e-05, + "loss": 0.0076, + "step": 28750 + }, + { + "epoch": 4.467225846536191, + "grad_norm": 0.3468378186225891, + "learning_rate": 8.534481487610289e-05, + "loss": 0.0089, + "step": 28760 + }, + { + "epoch": 4.4687791239515375, + "grad_norm": 0.3643616735935211, + "learning_rate": 8.533311766723428e-05, + "loss": 0.0118, + "step": 28770 + }, + { + "epoch": 4.470332401366884, + "grad_norm": 0.26345184445381165, + "learning_rate": 8.532141659438901e-05, + "loss": 0.0084, + "step": 28780 + }, + { + "epoch": 4.471885678782231, + "grad_norm": 0.29311537742614746, + "learning_rate": 8.530971165884675e-05, + "loss": 0.0082, + "step": 28790 + }, + { + "epoch": 4.473438956197577, + "grad_norm": 0.19494515657424927, + "learning_rate": 8.529800286188752e-05, + "loss": 0.0092, + "step": 28800 + }, + { + "epoch": 4.474992233612923, + "grad_norm": 0.184655100107193, + "learning_rate": 8.528629020479175e-05, + "loss": 0.0074, + "step": 28810 + }, + { + "epoch": 4.476545511028269, + "grad_norm": 0.26188984513282776, + "learning_rate": 8.527457368884033e-05, + "loss": 0.0065, + "step": 28820 + }, + { + "epoch": 4.478098788443616, + "grad_norm": 0.22769004106521606, + "learning_rate": 8.526285331531458e-05, + "loss": 0.0067, + "step": 28830 + }, + { + "epoch": 4.4796520658589625, + "grad_norm": 0.23611269891262054, + "learning_rate": 8.525112908549621e-05, + "loss": 0.0104, + "step": 28840 + }, + { + "epoch": 4.481205343274309, + "grad_norm": 0.39262351393699646, + "learning_rate": 8.523940100066735e-05, + "loss": 0.0102, + "step": 28850 + }, + { + "epoch": 4.482758620689655, + "grad_norm": 0.24412886798381805, + "learning_rate": 8.52276690621106e-05, + "loss": 0.008, + "step": 28860 + }, + { + "epoch": 4.484311898105002, + "grad_norm": 0.2585740089416504, + "learning_rate": 8.521593327110889e-05, + "loss": 0.0078, + "step": 28870 + }, + { + "epoch": 4.485865175520348, + "grad_norm": 0.1704898327589035, + "learning_rate": 8.520419362894569e-05, + "loss": 0.0109, + "step": 28880 + }, + { + "epoch": 4.487418452935694, + "grad_norm": 0.38076239824295044, + "learning_rate": 8.51924501369048e-05, + "loss": 0.0098, + "step": 28890 + }, + { + "epoch": 4.48897173035104, + "grad_norm": 0.3581386208534241, + "learning_rate": 8.518070279627047e-05, + "loss": 0.0113, + "step": 28900 + }, + { + "epoch": 4.490525007766387, + "grad_norm": 0.23473280668258667, + "learning_rate": 8.516895160832737e-05, + "loss": 0.0091, + "step": 28910 + }, + { + "epoch": 4.492078285181734, + "grad_norm": 0.2633395195007324, + "learning_rate": 8.515719657436061e-05, + "loss": 0.0127, + "step": 28920 + }, + { + "epoch": 4.49363156259708, + "grad_norm": 0.3683812618255615, + "learning_rate": 8.514543769565568e-05, + "loss": 0.0105, + "step": 28930 + }, + { + "epoch": 4.495184840012426, + "grad_norm": 0.2017602026462555, + "learning_rate": 8.513367497349853e-05, + "loss": 0.0078, + "step": 28940 + }, + { + "epoch": 4.496738117427773, + "grad_norm": 0.1938939392566681, + "learning_rate": 8.51219084091755e-05, + "loss": 0.0079, + "step": 28950 + }, + { + "epoch": 4.498291394843119, + "grad_norm": 0.18463122844696045, + "learning_rate": 8.511013800397338e-05, + "loss": 0.0094, + "step": 28960 + }, + { + "epoch": 4.499844672258465, + "grad_norm": 1.0882956981658936, + "learning_rate": 8.509836375917937e-05, + "loss": 0.0106, + "step": 28970 + }, + { + "epoch": 4.5013979496738115, + "grad_norm": 0.18893010914325714, + "learning_rate": 8.508658567608104e-05, + "loss": 0.0108, + "step": 28980 + }, + { + "epoch": 4.502951227089158, + "grad_norm": 0.30375880002975464, + "learning_rate": 8.507480375596647e-05, + "loss": 0.0097, + "step": 28990 + }, + { + "epoch": 4.504504504504505, + "grad_norm": 0.513602077960968, + "learning_rate": 8.506301800012408e-05, + "loss": 0.0082, + "step": 29000 + }, + { + "epoch": 4.506057781919851, + "grad_norm": 0.28405165672302246, + "learning_rate": 8.505122840984278e-05, + "loss": 0.0131, + "step": 29010 + }, + { + "epoch": 4.507611059335197, + "grad_norm": 0.8591514825820923, + "learning_rate": 8.503943498641182e-05, + "loss": 0.0127, + "step": 29020 + }, + { + "epoch": 4.509164336750544, + "grad_norm": 0.4698464870452881, + "learning_rate": 8.502763773112095e-05, + "loss": 0.0093, + "step": 29030 + }, + { + "epoch": 4.51071761416589, + "grad_norm": 0.3286454975605011, + "learning_rate": 8.501583664526026e-05, + "loss": 0.0127, + "step": 29040 + }, + { + "epoch": 4.512270891581236, + "grad_norm": 0.2979861795902252, + "learning_rate": 8.500403173012032e-05, + "loss": 0.0094, + "step": 29050 + }, + { + "epoch": 4.513824168996583, + "grad_norm": 0.30872106552124023, + "learning_rate": 8.499222298699211e-05, + "loss": 0.0114, + "step": 29060 + }, + { + "epoch": 4.515377446411929, + "grad_norm": 0.27028679847717285, + "learning_rate": 8.498041041716701e-05, + "loss": 0.0101, + "step": 29070 + }, + { + "epoch": 4.516930723827276, + "grad_norm": 0.2922718822956085, + "learning_rate": 8.496859402193681e-05, + "loss": 0.0108, + "step": 29080 + }, + { + "epoch": 4.518484001242622, + "grad_norm": 0.205459326505661, + "learning_rate": 8.495677380259374e-05, + "loss": 0.0129, + "step": 29090 + }, + { + "epoch": 4.520037278657968, + "grad_norm": 0.13214120268821716, + "learning_rate": 8.494494976043045e-05, + "loss": 0.0108, + "step": 29100 + }, + { + "epoch": 4.521590556073315, + "grad_norm": 0.2522391378879547, + "learning_rate": 8.493312189673998e-05, + "loss": 0.0107, + "step": 29110 + }, + { + "epoch": 4.523143833488661, + "grad_norm": 0.319465696811676, + "learning_rate": 8.492129021281584e-05, + "loss": 0.0098, + "step": 29120 + }, + { + "epoch": 4.5246971109040075, + "grad_norm": 0.2196507602930069, + "learning_rate": 8.490945470995188e-05, + "loss": 0.0107, + "step": 29130 + }, + { + "epoch": 4.526250388319354, + "grad_norm": 0.22734332084655762, + "learning_rate": 8.489761538944247e-05, + "loss": 0.0103, + "step": 29140 + }, + { + "epoch": 4.5278036657347, + "grad_norm": 0.2113359421491623, + "learning_rate": 8.48857722525823e-05, + "loss": 0.0093, + "step": 29150 + }, + { + "epoch": 4.529356943150047, + "grad_norm": 0.37265413999557495, + "learning_rate": 8.487392530066652e-05, + "loss": 0.0098, + "step": 29160 + }, + { + "epoch": 4.530910220565393, + "grad_norm": 0.23277713358402252, + "learning_rate": 8.486207453499069e-05, + "loss": 0.0086, + "step": 29170 + }, + { + "epoch": 4.532463497980739, + "grad_norm": 0.260431706905365, + "learning_rate": 8.485021995685082e-05, + "loss": 0.0094, + "step": 29180 + }, + { + "epoch": 4.534016775396085, + "grad_norm": 0.15543539822101593, + "learning_rate": 8.483836156754328e-05, + "loss": 0.0075, + "step": 29190 + }, + { + "epoch": 4.535570052811432, + "grad_norm": 0.20571787655353546, + "learning_rate": 8.482649936836491e-05, + "loss": 0.0096, + "step": 29200 + }, + { + "epoch": 4.537123330226779, + "grad_norm": 0.20520295202732086, + "learning_rate": 8.481463336061293e-05, + "loss": 0.008, + "step": 29210 + }, + { + "epoch": 4.538676607642125, + "grad_norm": 0.29079315066337585, + "learning_rate": 8.480276354558496e-05, + "loss": 0.0121, + "step": 29220 + }, + { + "epoch": 4.540229885057471, + "grad_norm": 0.36008909344673157, + "learning_rate": 8.479088992457913e-05, + "loss": 0.0099, + "step": 29230 + }, + { + "epoch": 4.541783162472818, + "grad_norm": 0.36116281151771545, + "learning_rate": 8.477901249889387e-05, + "loss": 0.0101, + "step": 29240 + }, + { + "epoch": 4.543336439888164, + "grad_norm": 0.23893488943576813, + "learning_rate": 8.47671312698281e-05, + "loss": 0.0083, + "step": 29250 + }, + { + "epoch": 4.54488971730351, + "grad_norm": 0.28217872977256775, + "learning_rate": 8.475524623868112e-05, + "loss": 0.0111, + "step": 29260 + }, + { + "epoch": 4.5464429947188565, + "grad_norm": 0.33618393540382385, + "learning_rate": 8.474335740675266e-05, + "loss": 0.0115, + "step": 29270 + }, + { + "epoch": 4.5479962721342035, + "grad_norm": 0.2358110249042511, + "learning_rate": 8.473146477534289e-05, + "loss": 0.0098, + "step": 29280 + }, + { + "epoch": 4.54954954954955, + "grad_norm": 0.2099287360906601, + "learning_rate": 8.471956834575232e-05, + "loss": 0.0088, + "step": 29290 + }, + { + "epoch": 4.551102826964896, + "grad_norm": 0.22095243632793427, + "learning_rate": 8.470766811928197e-05, + "loss": 0.0095, + "step": 29300 + }, + { + "epoch": 4.552656104380242, + "grad_norm": 0.3242826759815216, + "learning_rate": 8.469576409723323e-05, + "loss": 0.0119, + "step": 29310 + }, + { + "epoch": 4.554209381795589, + "grad_norm": 0.364574670791626, + "learning_rate": 8.468385628090788e-05, + "loss": 0.0106, + "step": 29320 + }, + { + "epoch": 4.555762659210935, + "grad_norm": 0.2710215747356415, + "learning_rate": 8.467194467160815e-05, + "loss": 0.0094, + "step": 29330 + }, + { + "epoch": 4.557315936626281, + "grad_norm": 0.3841574192047119, + "learning_rate": 8.466002927063667e-05, + "loss": 0.0102, + "step": 29340 + }, + { + "epoch": 4.558869214041628, + "grad_norm": 0.43021538853645325, + "learning_rate": 8.464811007929651e-05, + "loss": 0.0142, + "step": 29350 + }, + { + "epoch": 4.560422491456974, + "grad_norm": 0.5136095285415649, + "learning_rate": 8.463618709889114e-05, + "loss": 0.0112, + "step": 29360 + }, + { + "epoch": 4.561975768872321, + "grad_norm": 0.20163121819496155, + "learning_rate": 8.462426033072442e-05, + "loss": 0.0066, + "step": 29370 + }, + { + "epoch": 4.563529046287667, + "grad_norm": 0.3470019996166229, + "learning_rate": 8.461232977610061e-05, + "loss": 0.0117, + "step": 29380 + }, + { + "epoch": 4.565082323703013, + "grad_norm": 0.24573957920074463, + "learning_rate": 8.46003954363245e-05, + "loss": 0.0137, + "step": 29390 + }, + { + "epoch": 4.56663560111836, + "grad_norm": 0.44665127992630005, + "learning_rate": 8.458845731270115e-05, + "loss": 0.0137, + "step": 29400 + }, + { + "epoch": 4.568188878533706, + "grad_norm": 0.2872612774372101, + "learning_rate": 8.45765154065361e-05, + "loss": 0.0087, + "step": 29410 + }, + { + "epoch": 4.5697421559490525, + "grad_norm": 0.2117944061756134, + "learning_rate": 8.456456971913532e-05, + "loss": 0.0086, + "step": 29420 + }, + { + "epoch": 4.571295433364399, + "grad_norm": 0.32438430190086365, + "learning_rate": 8.455262025180517e-05, + "loss": 0.0098, + "step": 29430 + }, + { + "epoch": 4.572848710779745, + "grad_norm": 0.22574551403522491, + "learning_rate": 8.454066700585242e-05, + "loss": 0.0085, + "step": 29440 + }, + { + "epoch": 4.574401988195092, + "grad_norm": 0.2750547528266907, + "learning_rate": 8.452870998258423e-05, + "loss": 0.0104, + "step": 29450 + }, + { + "epoch": 4.575955265610438, + "grad_norm": 0.2715367078781128, + "learning_rate": 8.451674918330825e-05, + "loss": 0.0113, + "step": 29460 + }, + { + "epoch": 4.577508543025784, + "grad_norm": 0.2056458592414856, + "learning_rate": 8.450478460933246e-05, + "loss": 0.0112, + "step": 29470 + }, + { + "epoch": 4.57906182044113, + "grad_norm": 0.5016694664955139, + "learning_rate": 8.449281626196532e-05, + "loss": 0.0106, + "step": 29480 + }, + { + "epoch": 4.5806150978564775, + "grad_norm": 0.23517996072769165, + "learning_rate": 8.448084414251564e-05, + "loss": 0.0114, + "step": 29490 + }, + { + "epoch": 4.582168375271824, + "grad_norm": 0.11535537987947464, + "learning_rate": 8.446886825229271e-05, + "loss": 0.007, + "step": 29500 + }, + { + "epoch": 4.58372165268717, + "grad_norm": 0.3454356789588928, + "learning_rate": 8.445688859260615e-05, + "loss": 0.0109, + "step": 29510 + }, + { + "epoch": 4.585274930102516, + "grad_norm": 0.21828535199165344, + "learning_rate": 8.444490516476606e-05, + "loss": 0.0142, + "step": 29520 + }, + { + "epoch": 4.586828207517863, + "grad_norm": 0.30922529101371765, + "learning_rate": 8.443291797008293e-05, + "loss": 0.0104, + "step": 29530 + }, + { + "epoch": 4.588381484933209, + "grad_norm": 1.3637542724609375, + "learning_rate": 8.442092700986765e-05, + "loss": 0.0097, + "step": 29540 + }, + { + "epoch": 4.589934762348555, + "grad_norm": 0.2357684075832367, + "learning_rate": 8.440893228543156e-05, + "loss": 0.0085, + "step": 29550 + }, + { + "epoch": 4.5914880397639015, + "grad_norm": 0.28349825739860535, + "learning_rate": 8.439693379808638e-05, + "loss": 0.0117, + "step": 29560 + }, + { + "epoch": 4.593041317179249, + "grad_norm": 0.2594817876815796, + "learning_rate": 8.43849315491442e-05, + "loss": 0.0084, + "step": 29570 + }, + { + "epoch": 4.594594594594595, + "grad_norm": 0.339699923992157, + "learning_rate": 8.437292553991763e-05, + "loss": 0.0092, + "step": 29580 + }, + { + "epoch": 4.596147872009941, + "grad_norm": 0.18038402497768402, + "learning_rate": 8.436091577171959e-05, + "loss": 0.0109, + "step": 29590 + }, + { + "epoch": 4.597701149425287, + "grad_norm": 0.5774239301681519, + "learning_rate": 8.434890224586347e-05, + "loss": 0.0109, + "step": 29600 + }, + { + "epoch": 4.599254426840634, + "grad_norm": 0.2710758149623871, + "learning_rate": 8.433688496366303e-05, + "loss": 0.0088, + "step": 29610 + }, + { + "epoch": 4.60080770425598, + "grad_norm": 0.4441206455230713, + "learning_rate": 8.432486392643248e-05, + "loss": 0.008, + "step": 29620 + }, + { + "epoch": 4.6023609816713265, + "grad_norm": 0.2007865607738495, + "learning_rate": 8.431283913548643e-05, + "loss": 0.01, + "step": 29630 + }, + { + "epoch": 4.603914259086673, + "grad_norm": 0.34478819370269775, + "learning_rate": 8.430081059213985e-05, + "loss": 0.0105, + "step": 29640 + }, + { + "epoch": 4.605467536502019, + "grad_norm": 0.2378225326538086, + "learning_rate": 8.428877829770823e-05, + "loss": 0.0096, + "step": 29650 + }, + { + "epoch": 4.607020813917366, + "grad_norm": 0.269948810338974, + "learning_rate": 8.427674225350735e-05, + "loss": 0.0095, + "step": 29660 + }, + { + "epoch": 4.608574091332712, + "grad_norm": 0.2571246922016144, + "learning_rate": 8.426470246085347e-05, + "loss": 0.0093, + "step": 29670 + }, + { + "epoch": 4.610127368748058, + "grad_norm": 0.26461029052734375, + "learning_rate": 8.425265892106324e-05, + "loss": 0.0104, + "step": 29680 + }, + { + "epoch": 4.611680646163405, + "grad_norm": 0.23671045899391174, + "learning_rate": 8.424061163545374e-05, + "loss": 0.009, + "step": 29690 + }, + { + "epoch": 4.613233923578751, + "grad_norm": 0.2433784455060959, + "learning_rate": 8.422856060534243e-05, + "loss": 0.0112, + "step": 29700 + }, + { + "epoch": 4.614787200994098, + "grad_norm": 0.471150666475296, + "learning_rate": 8.421650583204718e-05, + "loss": 0.0092, + "step": 29710 + }, + { + "epoch": 4.616340478409444, + "grad_norm": 0.26211488246917725, + "learning_rate": 8.420444731688633e-05, + "loss": 0.0097, + "step": 29720 + }, + { + "epoch": 4.61789375582479, + "grad_norm": 0.21346305310726166, + "learning_rate": 8.419238506117852e-05, + "loss": 0.0119, + "step": 29730 + }, + { + "epoch": 4.619447033240137, + "grad_norm": 0.40709757804870605, + "learning_rate": 8.418031906624289e-05, + "loss": 0.0096, + "step": 29740 + }, + { + "epoch": 4.621000310655483, + "grad_norm": 0.32790160179138184, + "learning_rate": 8.416824933339898e-05, + "loss": 0.0097, + "step": 29750 + }, + { + "epoch": 4.622553588070829, + "grad_norm": 0.2601298391819, + "learning_rate": 8.415617586396667e-05, + "loss": 0.0102, + "step": 29760 + }, + { + "epoch": 4.624106865486175, + "grad_norm": 0.20110897719860077, + "learning_rate": 8.414409865926632e-05, + "loss": 0.0096, + "step": 29770 + }, + { + "epoch": 4.6256601429015225, + "grad_norm": 0.2633477747440338, + "learning_rate": 8.413201772061867e-05, + "loss": 0.0086, + "step": 29780 + }, + { + "epoch": 4.627213420316869, + "grad_norm": 0.2683119773864746, + "learning_rate": 8.411993304934488e-05, + "loss": 0.0092, + "step": 29790 + }, + { + "epoch": 4.628766697732215, + "grad_norm": 0.23726485669612885, + "learning_rate": 8.410784464676654e-05, + "loss": 0.0119, + "step": 29800 + }, + { + "epoch": 4.630319975147561, + "grad_norm": 0.21999303996562958, + "learning_rate": 8.409575251420556e-05, + "loss": 0.011, + "step": 29810 + }, + { + "epoch": 4.631873252562908, + "grad_norm": 0.27251726388931274, + "learning_rate": 8.408365665298435e-05, + "loss": 0.0103, + "step": 29820 + }, + { + "epoch": 4.633426529978254, + "grad_norm": 0.16444501280784607, + "learning_rate": 8.40715570644257e-05, + "loss": 0.0078, + "step": 29830 + }, + { + "epoch": 4.6349798073936, + "grad_norm": 0.17585885524749756, + "learning_rate": 8.40594537498528e-05, + "loss": 0.0109, + "step": 29840 + }, + { + "epoch": 4.6365330848089465, + "grad_norm": 0.5923117995262146, + "learning_rate": 8.404734671058924e-05, + "loss": 0.0106, + "step": 29850 + }, + { + "epoch": 4.638086362224294, + "grad_norm": 0.28110647201538086, + "learning_rate": 8.403523594795902e-05, + "loss": 0.0098, + "step": 29860 + }, + { + "epoch": 4.63963963963964, + "grad_norm": 0.23890750110149384, + "learning_rate": 8.402312146328659e-05, + "loss": 0.0095, + "step": 29870 + }, + { + "epoch": 4.641192917054986, + "grad_norm": 0.24070525169372559, + "learning_rate": 8.401100325789675e-05, + "loss": 0.0096, + "step": 29880 + }, + { + "epoch": 4.642746194470332, + "grad_norm": 0.1645163893699646, + "learning_rate": 8.399888133311472e-05, + "loss": 0.0099, + "step": 29890 + }, + { + "epoch": 4.644299471885679, + "grad_norm": 0.22407937049865723, + "learning_rate": 8.398675569026613e-05, + "loss": 0.0063, + "step": 29900 + }, + { + "epoch": 4.645852749301025, + "grad_norm": 0.20256172120571136, + "learning_rate": 8.397462633067705e-05, + "loss": 0.0095, + "step": 29910 + }, + { + "epoch": 4.6474060267163715, + "grad_norm": 0.2985793948173523, + "learning_rate": 8.396249325567392e-05, + "loss": 0.0088, + "step": 29920 + }, + { + "epoch": 4.648959304131718, + "grad_norm": 0.29858720302581787, + "learning_rate": 8.395035646658357e-05, + "loss": 0.0094, + "step": 29930 + }, + { + "epoch": 4.650512581547064, + "grad_norm": 0.41629716753959656, + "learning_rate": 8.39382159647333e-05, + "loss": 0.0087, + "step": 29940 + }, + { + "epoch": 4.652065858962411, + "grad_norm": 0.349469929933548, + "learning_rate": 8.392607175145075e-05, + "loss": 0.0115, + "step": 29950 + }, + { + "epoch": 4.653619136377757, + "grad_norm": 0.3311222195625305, + "learning_rate": 8.3913923828064e-05, + "loss": 0.0099, + "step": 29960 + }, + { + "epoch": 4.655172413793103, + "grad_norm": 0.25267294049263, + "learning_rate": 8.390177219590152e-05, + "loss": 0.0088, + "step": 29970 + }, + { + "epoch": 4.65672569120845, + "grad_norm": 0.148090198636055, + "learning_rate": 8.388961685629222e-05, + "loss": 0.0084, + "step": 29980 + }, + { + "epoch": 4.658278968623796, + "grad_norm": 0.23312103748321533, + "learning_rate": 8.387745781056536e-05, + "loss": 0.0104, + "step": 29990 + }, + { + "epoch": 4.659832246039143, + "grad_norm": 0.23858894407749176, + "learning_rate": 8.386529506005065e-05, + "loss": 0.0113, + "step": 30000 + }, + { + "epoch": 4.661385523454489, + "grad_norm": 0.17945510149002075, + "learning_rate": 8.38531286060782e-05, + "loss": 0.0118, + "step": 30010 + }, + { + "epoch": 4.662938800869835, + "grad_norm": 0.23961828649044037, + "learning_rate": 8.384095844997849e-05, + "loss": 0.0092, + "step": 30020 + }, + { + "epoch": 4.664492078285182, + "grad_norm": 0.2596084773540497, + "learning_rate": 8.382878459308245e-05, + "loss": 0.0124, + "step": 30030 + }, + { + "epoch": 4.666045355700528, + "grad_norm": 0.23687238991260529, + "learning_rate": 8.381660703672138e-05, + "loss": 0.0124, + "step": 30040 + }, + { + "epoch": 4.667598633115874, + "grad_norm": 0.29575127363204956, + "learning_rate": 8.380442578222702e-05, + "loss": 0.0121, + "step": 30050 + }, + { + "epoch": 4.669151910531221, + "grad_norm": 0.25149714946746826, + "learning_rate": 8.379224083093146e-05, + "loss": 0.0089, + "step": 30060 + }, + { + "epoch": 4.6707051879465675, + "grad_norm": 0.27577677369117737, + "learning_rate": 8.378005218416727e-05, + "loss": 0.0094, + "step": 30070 + }, + { + "epoch": 4.672258465361914, + "grad_norm": 0.2853153347969055, + "learning_rate": 8.376785984326735e-05, + "loss": 0.01, + "step": 30080 + }, + { + "epoch": 4.67381174277726, + "grad_norm": 0.41111233830451965, + "learning_rate": 8.375566380956506e-05, + "loss": 0.0113, + "step": 30090 + }, + { + "epoch": 4.675365020192606, + "grad_norm": 0.22990530729293823, + "learning_rate": 8.374346408439411e-05, + "loss": 0.0109, + "step": 30100 + }, + { + "epoch": 4.676918297607953, + "grad_norm": 0.18116866052150726, + "learning_rate": 8.373126066908868e-05, + "loss": 0.01, + "step": 30110 + }, + { + "epoch": 4.678471575023299, + "grad_norm": 0.24736405909061432, + "learning_rate": 8.371905356498326e-05, + "loss": 0.0111, + "step": 30120 + }, + { + "epoch": 4.680024852438645, + "grad_norm": 0.21136896312236786, + "learning_rate": 8.370684277341288e-05, + "loss": 0.0085, + "step": 30130 + }, + { + "epoch": 4.681578129853992, + "grad_norm": 0.1316426396369934, + "learning_rate": 8.369462829571282e-05, + "loss": 0.0086, + "step": 30140 + }, + { + "epoch": 4.683131407269339, + "grad_norm": 0.24126063287258148, + "learning_rate": 8.36824101332189e-05, + "loss": 0.01, + "step": 30150 + }, + { + "epoch": 4.684684684684685, + "grad_norm": 0.40505728125572205, + "learning_rate": 8.367018828726721e-05, + "loss": 0.0094, + "step": 30160 + }, + { + "epoch": 4.686237962100031, + "grad_norm": 0.23786939680576324, + "learning_rate": 8.365796275919438e-05, + "loss": 0.0076, + "step": 30170 + }, + { + "epoch": 4.687791239515377, + "grad_norm": 0.5901625752449036, + "learning_rate": 8.364573355033734e-05, + "loss": 0.0118, + "step": 30180 + }, + { + "epoch": 4.689344516930724, + "grad_norm": 0.2879071533679962, + "learning_rate": 8.363350066203346e-05, + "loss": 0.0103, + "step": 30190 + }, + { + "epoch": 4.69089779434607, + "grad_norm": 0.3308802843093872, + "learning_rate": 8.362126409562053e-05, + "loss": 0.0132, + "step": 30200 + }, + { + "epoch": 4.6924510717614165, + "grad_norm": 0.28215980529785156, + "learning_rate": 8.360902385243667e-05, + "loss": 0.0102, + "step": 30210 + }, + { + "epoch": 4.694004349176763, + "grad_norm": 0.3920758068561554, + "learning_rate": 8.359677993382052e-05, + "loss": 0.0132, + "step": 30220 + }, + { + "epoch": 4.69555762659211, + "grad_norm": 0.3076661229133606, + "learning_rate": 8.358453234111103e-05, + "loss": 0.0135, + "step": 30230 + }, + { + "epoch": 4.697110904007456, + "grad_norm": 0.31317338347435, + "learning_rate": 8.357228107564756e-05, + "loss": 0.0102, + "step": 30240 + }, + { + "epoch": 4.698664181422802, + "grad_norm": 0.16746364533901215, + "learning_rate": 8.356002613876993e-05, + "loss": 0.0077, + "step": 30250 + }, + { + "epoch": 4.700217458838148, + "grad_norm": 0.2787275016307831, + "learning_rate": 8.35477675318183e-05, + "loss": 0.0092, + "step": 30260 + }, + { + "epoch": 4.701770736253495, + "grad_norm": 0.5164092183113098, + "learning_rate": 8.353550525613323e-05, + "loss": 0.0086, + "step": 30270 + }, + { + "epoch": 4.7033240136688415, + "grad_norm": 0.2708331346511841, + "learning_rate": 8.352323931305572e-05, + "loss": 0.0092, + "step": 30280 + }, + { + "epoch": 4.704877291084188, + "grad_norm": 0.34558427333831787, + "learning_rate": 8.351096970392717e-05, + "loss": 0.0094, + "step": 30290 + }, + { + "epoch": 4.706430568499534, + "grad_norm": 0.2305053323507309, + "learning_rate": 8.349869643008937e-05, + "loss": 0.0118, + "step": 30300 + }, + { + "epoch": 4.70798384591488, + "grad_norm": 0.24648140370845795, + "learning_rate": 8.348641949288449e-05, + "loss": 0.01, + "step": 30310 + }, + { + "epoch": 4.709537123330227, + "grad_norm": 0.2365100383758545, + "learning_rate": 8.34741388936551e-05, + "loss": 0.0068, + "step": 30320 + }, + { + "epoch": 4.711090400745573, + "grad_norm": 0.8857736587524414, + "learning_rate": 8.346185463374423e-05, + "loss": 0.0097, + "step": 30330 + }, + { + "epoch": 4.712643678160919, + "grad_norm": 0.3894282579421997, + "learning_rate": 8.344956671449524e-05, + "loss": 0.0104, + "step": 30340 + }, + { + "epoch": 4.714196955576266, + "grad_norm": 0.3226231336593628, + "learning_rate": 8.343727513725192e-05, + "loss": 0.0098, + "step": 30350 + }, + { + "epoch": 4.715750232991613, + "grad_norm": 0.17384818196296692, + "learning_rate": 8.342497990335847e-05, + "loss": 0.0118, + "step": 30360 + }, + { + "epoch": 4.717303510406959, + "grad_norm": 0.2992291748523712, + "learning_rate": 8.341268101415946e-05, + "loss": 0.0097, + "step": 30370 + }, + { + "epoch": 4.718856787822305, + "grad_norm": 0.3102450668811798, + "learning_rate": 8.34003784709999e-05, + "loss": 0.0134, + "step": 30380 + }, + { + "epoch": 4.720410065237651, + "grad_norm": 0.27187249064445496, + "learning_rate": 8.338807227522516e-05, + "loss": 0.0087, + "step": 30390 + }, + { + "epoch": 4.721963342652998, + "grad_norm": 0.40100520849227905, + "learning_rate": 8.337576242818103e-05, + "loss": 0.0109, + "step": 30400 + }, + { + "epoch": 4.723516620068344, + "grad_norm": 0.2163499891757965, + "learning_rate": 8.336344893121372e-05, + "loss": 0.01, + "step": 30410 + }, + { + "epoch": 4.72506989748369, + "grad_norm": 0.21358317136764526, + "learning_rate": 8.335113178566977e-05, + "loss": 0.0112, + "step": 30420 + }, + { + "epoch": 4.726623174899037, + "grad_norm": 0.2121487706899643, + "learning_rate": 8.33388109928962e-05, + "loss": 0.0094, + "step": 30430 + }, + { + "epoch": 4.728176452314384, + "grad_norm": 0.4311557710170746, + "learning_rate": 8.33264865542404e-05, + "loss": 0.0094, + "step": 30440 + }, + { + "epoch": 4.72972972972973, + "grad_norm": 0.3620319962501526, + "learning_rate": 8.331415847105013e-05, + "loss": 0.0091, + "step": 30450 + }, + { + "epoch": 4.731283007145076, + "grad_norm": 0.2201693207025528, + "learning_rate": 8.330182674467357e-05, + "loss": 0.0096, + "step": 30460 + }, + { + "epoch": 4.732836284560422, + "grad_norm": 0.23415103554725647, + "learning_rate": 8.32894913764593e-05, + "loss": 0.0117, + "step": 30470 + }, + { + "epoch": 4.734389561975769, + "grad_norm": 0.3053898513317108, + "learning_rate": 8.327715236775633e-05, + "loss": 0.0095, + "step": 30480 + }, + { + "epoch": 4.735942839391115, + "grad_norm": 1.179787516593933, + "learning_rate": 8.326480971991398e-05, + "loss": 0.0128, + "step": 30490 + }, + { + "epoch": 4.7374961168064615, + "grad_norm": 0.1695004105567932, + "learning_rate": 8.325246343428206e-05, + "loss": 0.0114, + "step": 30500 + }, + { + "epoch": 4.739049394221808, + "grad_norm": 0.48486191034317017, + "learning_rate": 8.324011351221072e-05, + "loss": 0.0118, + "step": 30510 + }, + { + "epoch": 4.740602671637155, + "grad_norm": 0.38113319873809814, + "learning_rate": 8.322775995505057e-05, + "loss": 0.0104, + "step": 30520 + }, + { + "epoch": 4.742155949052501, + "grad_norm": 0.20123139023780823, + "learning_rate": 8.321540276415254e-05, + "loss": 0.0131, + "step": 30530 + }, + { + "epoch": 4.743709226467847, + "grad_norm": 0.5015894174575806, + "learning_rate": 8.320304194086798e-05, + "loss": 0.0103, + "step": 30540 + }, + { + "epoch": 4.745262503883193, + "grad_norm": 0.5461856126785278, + "learning_rate": 8.31906774865487e-05, + "loss": 0.0098, + "step": 30550 + }, + { + "epoch": 4.74681578129854, + "grad_norm": 0.30906009674072266, + "learning_rate": 8.317830940254682e-05, + "loss": 0.0087, + "step": 30560 + }, + { + "epoch": 4.7483690587138865, + "grad_norm": 0.30376142263412476, + "learning_rate": 8.316593769021491e-05, + "loss": 0.0083, + "step": 30570 + }, + { + "epoch": 4.749922336129233, + "grad_norm": 0.27703624963760376, + "learning_rate": 8.315356235090592e-05, + "loss": 0.0106, + "step": 30580 + }, + { + "epoch": 4.751475613544579, + "grad_norm": 0.26058152318000793, + "learning_rate": 8.314118338597319e-05, + "loss": 0.0104, + "step": 30590 + }, + { + "epoch": 4.753028890959925, + "grad_norm": 0.37437009811401367, + "learning_rate": 8.312880079677048e-05, + "loss": 0.0117, + "step": 30600 + }, + { + "epoch": 4.754582168375272, + "grad_norm": 0.21839000284671783, + "learning_rate": 8.311641458465191e-05, + "loss": 0.0102, + "step": 30610 + }, + { + "epoch": 4.756135445790618, + "grad_norm": 0.31302410364151, + "learning_rate": 8.310402475097205e-05, + "loss": 0.0091, + "step": 30620 + }, + { + "epoch": 4.757688723205964, + "grad_norm": 0.29703062772750854, + "learning_rate": 8.309163129708581e-05, + "loss": 0.0091, + "step": 30630 + }, + { + "epoch": 4.759242000621311, + "grad_norm": 0.18843717873096466, + "learning_rate": 8.307923422434852e-05, + "loss": 0.0097, + "step": 30640 + }, + { + "epoch": 4.760795278036658, + "grad_norm": 0.21235281229019165, + "learning_rate": 8.30668335341159e-05, + "loss": 0.0099, + "step": 30650 + }, + { + "epoch": 4.762348555452004, + "grad_norm": 0.588090717792511, + "learning_rate": 8.305442922774408e-05, + "loss": 0.0115, + "step": 30660 + }, + { + "epoch": 4.76390183286735, + "grad_norm": 0.15490667521953583, + "learning_rate": 8.304202130658959e-05, + "loss": 0.0117, + "step": 30670 + }, + { + "epoch": 4.765455110282696, + "grad_norm": 0.36919349431991577, + "learning_rate": 8.302960977200931e-05, + "loss": 0.0104, + "step": 30680 + }, + { + "epoch": 4.767008387698043, + "grad_norm": 0.27249106764793396, + "learning_rate": 8.301719462536058e-05, + "loss": 0.009, + "step": 30690 + }, + { + "epoch": 4.768561665113389, + "grad_norm": 0.17241954803466797, + "learning_rate": 8.300477586800108e-05, + "loss": 0.0099, + "step": 30700 + }, + { + "epoch": 4.7701149425287355, + "grad_norm": 0.22727914154529572, + "learning_rate": 8.299235350128892e-05, + "loss": 0.0085, + "step": 30710 + }, + { + "epoch": 4.771668219944082, + "grad_norm": 0.2613919973373413, + "learning_rate": 8.297992752658258e-05, + "loss": 0.0122, + "step": 30720 + }, + { + "epoch": 4.773221497359429, + "grad_norm": 0.2236497849225998, + "learning_rate": 8.296749794524094e-05, + "loss": 0.0141, + "step": 30730 + }, + { + "epoch": 4.774774774774775, + "grad_norm": 0.2821626365184784, + "learning_rate": 8.295506475862332e-05, + "loss": 0.0089, + "step": 30740 + }, + { + "epoch": 4.776328052190121, + "grad_norm": 0.22491154074668884, + "learning_rate": 8.294262796808933e-05, + "loss": 0.0118, + "step": 30750 + }, + { + "epoch": 4.777881329605467, + "grad_norm": 0.25117775797843933, + "learning_rate": 8.293018757499909e-05, + "loss": 0.0078, + "step": 30760 + }, + { + "epoch": 4.779434607020814, + "grad_norm": 0.23309548199176788, + "learning_rate": 8.291774358071305e-05, + "loss": 0.0101, + "step": 30770 + }, + { + "epoch": 4.78098788443616, + "grad_norm": 0.45569610595703125, + "learning_rate": 8.290529598659205e-05, + "loss": 0.0105, + "step": 30780 + }, + { + "epoch": 4.782541161851507, + "grad_norm": 0.20063911378383636, + "learning_rate": 8.289284479399738e-05, + "loss": 0.0091, + "step": 30790 + }, + { + "epoch": 4.784094439266853, + "grad_norm": 0.2796301543712616, + "learning_rate": 8.288039000429064e-05, + "loss": 0.0106, + "step": 30800 + }, + { + "epoch": 4.7856477166822, + "grad_norm": 0.13628588616847992, + "learning_rate": 8.286793161883388e-05, + "loss": 0.0079, + "step": 30810 + }, + { + "epoch": 4.787200994097546, + "grad_norm": 0.1576908975839615, + "learning_rate": 8.285546963898954e-05, + "loss": 0.007, + "step": 30820 + }, + { + "epoch": 4.788754271512892, + "grad_norm": 0.28617072105407715, + "learning_rate": 8.284300406612044e-05, + "loss": 0.0096, + "step": 30830 + }, + { + "epoch": 4.790307548928238, + "grad_norm": 0.4401049017906189, + "learning_rate": 8.283053490158978e-05, + "loss": 0.0118, + "step": 30840 + }, + { + "epoch": 4.791860826343585, + "grad_norm": 0.37648066878318787, + "learning_rate": 8.28180621467612e-05, + "loss": 0.0083, + "step": 30850 + }, + { + "epoch": 4.7934141037589315, + "grad_norm": 0.19740556180477142, + "learning_rate": 8.280558580299867e-05, + "loss": 0.0074, + "step": 30860 + }, + { + "epoch": 4.794967381174278, + "grad_norm": 0.25560468435287476, + "learning_rate": 8.279310587166661e-05, + "loss": 0.0084, + "step": 30870 + }, + { + "epoch": 4.796520658589624, + "grad_norm": 0.27360638976097107, + "learning_rate": 8.278062235412978e-05, + "loss": 0.0099, + "step": 30880 + }, + { + "epoch": 4.79807393600497, + "grad_norm": 0.1894545704126358, + "learning_rate": 8.276813525175339e-05, + "loss": 0.0069, + "step": 30890 + }, + { + "epoch": 4.799627213420317, + "grad_norm": 0.2973006069660187, + "learning_rate": 8.2755644565903e-05, + "loss": 0.007, + "step": 30900 + }, + { + "epoch": 4.801180490835663, + "grad_norm": 0.2512770891189575, + "learning_rate": 8.274315029794454e-05, + "loss": 0.0085, + "step": 30910 + }, + { + "epoch": 4.802733768251009, + "grad_norm": 0.2892058491706848, + "learning_rate": 8.273065244924443e-05, + "loss": 0.0083, + "step": 30920 + }, + { + "epoch": 4.8042870456663564, + "grad_norm": 0.37744829058647156, + "learning_rate": 8.271815102116936e-05, + "loss": 0.0067, + "step": 30930 + }, + { + "epoch": 4.805840323081703, + "grad_norm": 0.46280792355537415, + "learning_rate": 8.270564601508648e-05, + "loss": 0.0104, + "step": 30940 + }, + { + "epoch": 4.807393600497049, + "grad_norm": 0.25206050276756287, + "learning_rate": 8.269313743236333e-05, + "loss": 0.012, + "step": 30950 + }, + { + "epoch": 4.808946877912395, + "grad_norm": 0.27161094546318054, + "learning_rate": 8.268062527436783e-05, + "loss": 0.011, + "step": 30960 + }, + { + "epoch": 4.810500155327741, + "grad_norm": 0.1847987025976181, + "learning_rate": 8.266810954246828e-05, + "loss": 0.0107, + "step": 30970 + }, + { + "epoch": 4.812053432743088, + "grad_norm": 0.25065720081329346, + "learning_rate": 8.265559023803338e-05, + "loss": 0.0102, + "step": 30980 + }, + { + "epoch": 4.813606710158434, + "grad_norm": 0.2444182187318802, + "learning_rate": 8.264306736243223e-05, + "loss": 0.0123, + "step": 30990 + }, + { + "epoch": 4.8151599875737805, + "grad_norm": 0.49077731370925903, + "learning_rate": 8.263054091703432e-05, + "loss": 0.0122, + "step": 31000 + }, + { + "epoch": 4.8167132649891276, + "grad_norm": 0.5581703186035156, + "learning_rate": 8.26180109032095e-05, + "loss": 0.009, + "step": 31010 + }, + { + "epoch": 4.818266542404474, + "grad_norm": 0.2625753581523895, + "learning_rate": 8.260547732232807e-05, + "loss": 0.0077, + "step": 31020 + }, + { + "epoch": 4.81981981981982, + "grad_norm": 0.3291319012641907, + "learning_rate": 8.259294017576063e-05, + "loss": 0.01, + "step": 31030 + }, + { + "epoch": 4.821373097235166, + "grad_norm": 0.41132429242134094, + "learning_rate": 8.258039946487828e-05, + "loss": 0.0104, + "step": 31040 + }, + { + "epoch": 4.822926374650512, + "grad_norm": 0.21591484546661377, + "learning_rate": 8.256785519105241e-05, + "loss": 0.0077, + "step": 31050 + }, + { + "epoch": 4.824479652065859, + "grad_norm": 0.3034130334854126, + "learning_rate": 8.255530735565488e-05, + "loss": 0.0068, + "step": 31060 + }, + { + "epoch": 4.826032929481205, + "grad_norm": 0.19468404352664948, + "learning_rate": 8.254275596005787e-05, + "loss": 0.0098, + "step": 31070 + }, + { + "epoch": 4.827586206896552, + "grad_norm": 0.31437936425209045, + "learning_rate": 8.2530201005634e-05, + "loss": 0.0106, + "step": 31080 + }, + { + "epoch": 4.829139484311898, + "grad_norm": 0.5519593358039856, + "learning_rate": 8.251764249375626e-05, + "loss": 0.0112, + "step": 31090 + }, + { + "epoch": 4.830692761727245, + "grad_norm": 0.23072442412376404, + "learning_rate": 8.250508042579803e-05, + "loss": 0.0075, + "step": 31100 + }, + { + "epoch": 4.832246039142591, + "grad_norm": 0.19043490290641785, + "learning_rate": 8.249251480313307e-05, + "loss": 0.0111, + "step": 31110 + }, + { + "epoch": 4.833799316557937, + "grad_norm": 0.2693043649196625, + "learning_rate": 8.247994562713555e-05, + "loss": 0.0086, + "step": 31120 + }, + { + "epoch": 4.835352593973283, + "grad_norm": 0.2600656747817993, + "learning_rate": 8.246737289917998e-05, + "loss": 0.0103, + "step": 31130 + }, + { + "epoch": 4.83690587138863, + "grad_norm": 0.34305986762046814, + "learning_rate": 8.245479662064135e-05, + "loss": 0.0104, + "step": 31140 + }, + { + "epoch": 4.8384591488039765, + "grad_norm": 0.2515650689601898, + "learning_rate": 8.244221679289496e-05, + "loss": 0.0131, + "step": 31150 + }, + { + "epoch": 4.840012426219323, + "grad_norm": 0.27106672525405884, + "learning_rate": 8.242963341731652e-05, + "loss": 0.0123, + "step": 31160 + }, + { + "epoch": 4.841565703634669, + "grad_norm": 0.30703747272491455, + "learning_rate": 8.24170464952821e-05, + "loss": 0.0107, + "step": 31170 + }, + { + "epoch": 4.843118981050015, + "grad_norm": 0.22906947135925293, + "learning_rate": 8.240445602816824e-05, + "loss": 0.0079, + "step": 31180 + }, + { + "epoch": 4.844672258465362, + "grad_norm": 0.32080188393592834, + "learning_rate": 8.239186201735179e-05, + "loss": 0.0094, + "step": 31190 + }, + { + "epoch": 4.846225535880708, + "grad_norm": 0.29191967844963074, + "learning_rate": 8.237926446420998e-05, + "loss": 0.0096, + "step": 31200 + }, + { + "epoch": 4.847778813296054, + "grad_norm": 0.3669886887073517, + "learning_rate": 8.236666337012052e-05, + "loss": 0.0084, + "step": 31210 + }, + { + "epoch": 4.8493320907114015, + "grad_norm": 0.1533530205488205, + "learning_rate": 8.23540587364614e-05, + "loss": 0.0078, + "step": 31220 + }, + { + "epoch": 4.850885368126748, + "grad_norm": 0.24215532839298248, + "learning_rate": 8.234145056461107e-05, + "loss": 0.0079, + "step": 31230 + }, + { + "epoch": 4.852438645542094, + "grad_norm": 0.16273191571235657, + "learning_rate": 8.232883885594832e-05, + "loss": 0.0093, + "step": 31240 + }, + { + "epoch": 4.85399192295744, + "grad_norm": 0.23245786130428314, + "learning_rate": 8.231622361185236e-05, + "loss": 0.0104, + "step": 31250 + }, + { + "epoch": 4.855545200372786, + "grad_norm": 0.26824355125427246, + "learning_rate": 8.230360483370278e-05, + "loss": 0.0065, + "step": 31260 + }, + { + "epoch": 4.857098477788133, + "grad_norm": 0.17867878079414368, + "learning_rate": 8.229098252287953e-05, + "loss": 0.0116, + "step": 31270 + }, + { + "epoch": 4.858651755203479, + "grad_norm": 0.2526090443134308, + "learning_rate": 8.2278356680763e-05, + "loss": 0.0098, + "step": 31280 + }, + { + "epoch": 4.8602050326188255, + "grad_norm": 0.32187145948410034, + "learning_rate": 8.22657273087339e-05, + "loss": 0.0107, + "step": 31290 + }, + { + "epoch": 4.861758310034173, + "grad_norm": 0.34141770005226135, + "learning_rate": 8.225309440817336e-05, + "loss": 0.0087, + "step": 31300 + }, + { + "epoch": 4.863311587449519, + "grad_norm": 0.32365328073501587, + "learning_rate": 8.224045798046293e-05, + "loss": 0.0093, + "step": 31310 + }, + { + "epoch": 4.864864864864865, + "grad_norm": 0.3113953769207001, + "learning_rate": 8.22278180269845e-05, + "loss": 0.0095, + "step": 31320 + }, + { + "epoch": 4.866418142280211, + "grad_norm": 0.30999258160591125, + "learning_rate": 8.221517454912031e-05, + "loss": 0.0077, + "step": 31330 + }, + { + "epoch": 4.867971419695557, + "grad_norm": 0.25969552993774414, + "learning_rate": 8.220252754825308e-05, + "loss": 0.0079, + "step": 31340 + }, + { + "epoch": 4.869524697110904, + "grad_norm": 0.3588017523288727, + "learning_rate": 8.218987702576586e-05, + "loss": 0.0084, + "step": 31350 + }, + { + "epoch": 4.8710779745262505, + "grad_norm": 0.306344211101532, + "learning_rate": 8.217722298304207e-05, + "loss": 0.0105, + "step": 31360 + }, + { + "epoch": 4.872631251941597, + "grad_norm": 0.5451350808143616, + "learning_rate": 8.216456542146557e-05, + "loss": 0.0102, + "step": 31370 + }, + { + "epoch": 4.874184529356943, + "grad_norm": 0.5402668714523315, + "learning_rate": 8.215190434242055e-05, + "loss": 0.0083, + "step": 31380 + }, + { + "epoch": 4.87573780677229, + "grad_norm": 0.6126529574394226, + "learning_rate": 8.213923974729161e-05, + "loss": 0.0102, + "step": 31390 + }, + { + "epoch": 4.877291084187636, + "grad_norm": 0.4238356351852417, + "learning_rate": 8.212657163746373e-05, + "loss": 0.0114, + "step": 31400 + }, + { + "epoch": 4.878844361602982, + "grad_norm": 0.1596018373966217, + "learning_rate": 8.211390001432227e-05, + "loss": 0.0131, + "step": 31410 + }, + { + "epoch": 4.880397639018328, + "grad_norm": 0.20178698003292084, + "learning_rate": 8.210122487925297e-05, + "loss": 0.0095, + "step": 31420 + }, + { + "epoch": 4.881950916433675, + "grad_norm": 0.18243898451328278, + "learning_rate": 8.208854623364202e-05, + "loss": 0.0095, + "step": 31430 + }, + { + "epoch": 4.883504193849022, + "grad_norm": 0.21593862771987915, + "learning_rate": 8.207586407887589e-05, + "loss": 0.0093, + "step": 31440 + }, + { + "epoch": 4.885057471264368, + "grad_norm": 0.3341086208820343, + "learning_rate": 8.206317841634148e-05, + "loss": 0.0135, + "step": 31450 + }, + { + "epoch": 4.886610748679714, + "grad_norm": 0.2509872019290924, + "learning_rate": 8.205048924742609e-05, + "loss": 0.0129, + "step": 31460 + }, + { + "epoch": 4.888164026095061, + "grad_norm": 0.20394791662693024, + "learning_rate": 8.203779657351738e-05, + "loss": 0.0117, + "step": 31470 + }, + { + "epoch": 4.889717303510407, + "grad_norm": 0.22595728933811188, + "learning_rate": 8.20251003960034e-05, + "loss": 0.0089, + "step": 31480 + }, + { + "epoch": 4.891270580925753, + "grad_norm": 0.3352772891521454, + "learning_rate": 8.201240071627258e-05, + "loss": 0.0113, + "step": 31490 + }, + { + "epoch": 4.8928238583410995, + "grad_norm": 0.30953845381736755, + "learning_rate": 8.199969753571377e-05, + "loss": 0.0102, + "step": 31500 + }, + { + "epoch": 4.8943771357564465, + "grad_norm": 0.302259236574173, + "learning_rate": 8.198699085571615e-05, + "loss": 0.0117, + "step": 31510 + }, + { + "epoch": 4.895930413171793, + "grad_norm": 0.2521390914916992, + "learning_rate": 8.197428067766928e-05, + "loss": 0.0091, + "step": 31520 + }, + { + "epoch": 4.897483690587139, + "grad_norm": 0.24263040721416473, + "learning_rate": 8.196156700296316e-05, + "loss": 0.0109, + "step": 31530 + }, + { + "epoch": 4.899036968002485, + "grad_norm": 0.19680579006671906, + "learning_rate": 8.194884983298814e-05, + "loss": 0.012, + "step": 31540 + }, + { + "epoch": 4.900590245417831, + "grad_norm": 0.22390110790729523, + "learning_rate": 8.193612916913491e-05, + "loss": 0.0084, + "step": 31550 + }, + { + "epoch": 4.902143522833178, + "grad_norm": 0.27495232224464417, + "learning_rate": 8.192340501279463e-05, + "loss": 0.0083, + "step": 31560 + }, + { + "epoch": 4.903696800248524, + "grad_norm": 0.16605322062969208, + "learning_rate": 8.191067736535876e-05, + "loss": 0.0095, + "step": 31570 + }, + { + "epoch": 4.905250077663871, + "grad_norm": 0.40132808685302734, + "learning_rate": 8.18979462282192e-05, + "loss": 0.0102, + "step": 31580 + }, + { + "epoch": 4.906803355079218, + "grad_norm": 0.24396836757659912, + "learning_rate": 8.188521160276819e-05, + "loss": 0.0076, + "step": 31590 + }, + { + "epoch": 4.908356632494564, + "grad_norm": 0.32688358426094055, + "learning_rate": 8.187247349039837e-05, + "loss": 0.0114, + "step": 31600 + }, + { + "epoch": 4.90990990990991, + "grad_norm": 0.17226813733577728, + "learning_rate": 8.185973189250278e-05, + "loss": 0.009, + "step": 31610 + }, + { + "epoch": 4.911463187325256, + "grad_norm": 0.2437591850757599, + "learning_rate": 8.184698681047482e-05, + "loss": 0.0088, + "step": 31620 + }, + { + "epoch": 4.913016464740602, + "grad_norm": 0.15505053102970123, + "learning_rate": 8.183423824570827e-05, + "loss": 0.0088, + "step": 31630 + }, + { + "epoch": 4.914569742155949, + "grad_norm": 0.21978600323200226, + "learning_rate": 8.182148619959725e-05, + "loss": 0.0094, + "step": 31640 + }, + { + "epoch": 4.9161230195712955, + "grad_norm": 0.5389467477798462, + "learning_rate": 8.180873067353636e-05, + "loss": 0.0085, + "step": 31650 + }, + { + "epoch": 4.917676296986642, + "grad_norm": 0.18397489190101624, + "learning_rate": 8.179597166892052e-05, + "loss": 0.0091, + "step": 31660 + }, + { + "epoch": 4.919229574401988, + "grad_norm": 0.4707520008087158, + "learning_rate": 8.178320918714501e-05, + "loss": 0.0122, + "step": 31670 + }, + { + "epoch": 4.920782851817335, + "grad_norm": 0.32238641381263733, + "learning_rate": 8.177044322960554e-05, + "loss": 0.0107, + "step": 31680 + }, + { + "epoch": 4.922336129232681, + "grad_norm": 0.3932350277900696, + "learning_rate": 8.175767379769816e-05, + "loss": 0.0103, + "step": 31690 + }, + { + "epoch": 4.923889406648027, + "grad_norm": 0.21165895462036133, + "learning_rate": 8.174490089281932e-05, + "loss": 0.0138, + "step": 31700 + }, + { + "epoch": 4.925442684063373, + "grad_norm": 0.25981250405311584, + "learning_rate": 8.173212451636584e-05, + "loss": 0.0089, + "step": 31710 + }, + { + "epoch": 4.92699596147872, + "grad_norm": 0.30277660489082336, + "learning_rate": 8.171934466973493e-05, + "loss": 0.0115, + "step": 31720 + }, + { + "epoch": 4.928549238894067, + "grad_norm": 0.34015002846717834, + "learning_rate": 8.170656135432418e-05, + "loss": 0.0071, + "step": 31730 + }, + { + "epoch": 4.930102516309413, + "grad_norm": 0.2068876028060913, + "learning_rate": 8.169377457153155e-05, + "loss": 0.0103, + "step": 31740 + }, + { + "epoch": 4.931655793724759, + "grad_norm": 0.22134268283843994, + "learning_rate": 8.168098432275539e-05, + "loss": 0.0067, + "step": 31750 + }, + { + "epoch": 4.933209071140106, + "grad_norm": 0.2587699294090271, + "learning_rate": 8.166819060939442e-05, + "loss": 0.0093, + "step": 31760 + }, + { + "epoch": 4.934762348555452, + "grad_norm": 0.4414481222629547, + "learning_rate": 8.165539343284772e-05, + "loss": 0.0082, + "step": 31770 + }, + { + "epoch": 4.936315625970798, + "grad_norm": 0.2595367431640625, + "learning_rate": 8.16425927945148e-05, + "loss": 0.0145, + "step": 31780 + }, + { + "epoch": 4.9378689033861445, + "grad_norm": 0.17137779295444489, + "learning_rate": 8.162978869579551e-05, + "loss": 0.0093, + "step": 31790 + }, + { + "epoch": 4.9394221808014915, + "grad_norm": 0.8748199343681335, + "learning_rate": 8.161698113809007e-05, + "loss": 0.0103, + "step": 31800 + }, + { + "epoch": 4.940975458216838, + "grad_norm": 0.2803286015987396, + "learning_rate": 8.160417012279911e-05, + "loss": 0.0084, + "step": 31810 + }, + { + "epoch": 4.942528735632184, + "grad_norm": 0.34359997510910034, + "learning_rate": 8.159135565132363e-05, + "loss": 0.0127, + "step": 31820 + }, + { + "epoch": 4.94408201304753, + "grad_norm": 0.3010127544403076, + "learning_rate": 8.157853772506498e-05, + "loss": 0.0111, + "step": 31830 + }, + { + "epoch": 4.945635290462876, + "grad_norm": 0.43210896849632263, + "learning_rate": 8.156571634542494e-05, + "loss": 0.0103, + "step": 31840 + }, + { + "epoch": 4.947188567878223, + "grad_norm": 0.4077731966972351, + "learning_rate": 8.15528915138056e-05, + "loss": 0.0109, + "step": 31850 + }, + { + "epoch": 4.948741845293569, + "grad_norm": 0.16553641855716705, + "learning_rate": 8.154006323160949e-05, + "loss": 0.0098, + "step": 31860 + }, + { + "epoch": 4.950295122708916, + "grad_norm": 0.5904366970062256, + "learning_rate": 8.152723150023949e-05, + "loss": 0.0111, + "step": 31870 + }, + { + "epoch": 4.951848400124263, + "grad_norm": 0.3981987535953522, + "learning_rate": 8.151439632109886e-05, + "loss": 0.0084, + "step": 31880 + }, + { + "epoch": 4.953401677539609, + "grad_norm": 0.2524905800819397, + "learning_rate": 8.150155769559122e-05, + "loss": 0.0119, + "step": 31890 + }, + { + "epoch": 4.954954954954955, + "grad_norm": 0.2637515068054199, + "learning_rate": 8.148871562512058e-05, + "loss": 0.0132, + "step": 31900 + }, + { + "epoch": 4.956508232370301, + "grad_norm": 0.27945640683174133, + "learning_rate": 8.147587011109136e-05, + "loss": 0.0074, + "step": 31910 + }, + { + "epoch": 4.958061509785647, + "grad_norm": 0.2343638688325882, + "learning_rate": 8.14630211549083e-05, + "loss": 0.0102, + "step": 31920 + }, + { + "epoch": 4.959614787200994, + "grad_norm": 0.3132230341434479, + "learning_rate": 8.145016875797655e-05, + "loss": 0.0111, + "step": 31930 + }, + { + "epoch": 4.9611680646163405, + "grad_norm": 0.4538421332836151, + "learning_rate": 8.143731292170164e-05, + "loss": 0.0129, + "step": 31940 + }, + { + "epoch": 4.962721342031687, + "grad_norm": 0.3106589615345001, + "learning_rate": 8.142445364748944e-05, + "loss": 0.0066, + "step": 31950 + }, + { + "epoch": 4.964274619447033, + "grad_norm": 0.22680062055587769, + "learning_rate": 8.141159093674624e-05, + "loss": 0.01, + "step": 31960 + }, + { + "epoch": 4.96582789686238, + "grad_norm": 0.1948218047618866, + "learning_rate": 8.139872479087869e-05, + "loss": 0.0108, + "step": 31970 + }, + { + "epoch": 4.967381174277726, + "grad_norm": 0.21844889223575592, + "learning_rate": 8.13858552112938e-05, + "loss": 0.0114, + "step": 31980 + }, + { + "epoch": 4.968934451693072, + "grad_norm": 0.4541386365890503, + "learning_rate": 8.137298219939895e-05, + "loss": 0.0097, + "step": 31990 + }, + { + "epoch": 4.970487729108418, + "grad_norm": 0.26270169019699097, + "learning_rate": 8.136010575660196e-05, + "loss": 0.011, + "step": 32000 + }, + { + "epoch": 4.9720410065237655, + "grad_norm": 0.8937418460845947, + "learning_rate": 8.134722588431095e-05, + "loss": 0.0077, + "step": 32010 + }, + { + "epoch": 4.973594283939112, + "grad_norm": 0.36082354187965393, + "learning_rate": 8.133434258393444e-05, + "loss": 0.0076, + "step": 32020 + }, + { + "epoch": 4.975147561354458, + "grad_norm": 1.3312937021255493, + "learning_rate": 8.132145585688134e-05, + "loss": 0.0115, + "step": 32030 + }, + { + "epoch": 4.976700838769804, + "grad_norm": 0.14504843950271606, + "learning_rate": 8.130856570456093e-05, + "loss": 0.006, + "step": 32040 + }, + { + "epoch": 4.978254116185151, + "grad_norm": 0.32123202085494995, + "learning_rate": 8.129567212838283e-05, + "loss": 0.0105, + "step": 32050 + }, + { + "epoch": 4.979807393600497, + "grad_norm": 0.12227489799261093, + "learning_rate": 8.128277512975708e-05, + "loss": 0.0104, + "step": 32060 + }, + { + "epoch": 4.981360671015843, + "grad_norm": 0.19228902459144592, + "learning_rate": 8.126987471009408e-05, + "loss": 0.0089, + "step": 32070 + }, + { + "epoch": 4.9829139484311895, + "grad_norm": 0.26075127720832825, + "learning_rate": 8.125697087080459e-05, + "loss": 0.0105, + "step": 32080 + }, + { + "epoch": 4.984467225846537, + "grad_norm": 0.33151766657829285, + "learning_rate": 8.124406361329976e-05, + "loss": 0.0097, + "step": 32090 + }, + { + "epoch": 4.986020503261883, + "grad_norm": 0.35361120104789734, + "learning_rate": 8.123115293899111e-05, + "loss": 0.0086, + "step": 32100 + }, + { + "epoch": 4.987573780677229, + "grad_norm": 0.18322014808654785, + "learning_rate": 8.121823884929055e-05, + "loss": 0.0092, + "step": 32110 + }, + { + "epoch": 4.989127058092575, + "grad_norm": 0.4721601605415344, + "learning_rate": 8.12053213456103e-05, + "loss": 0.0093, + "step": 32120 + }, + { + "epoch": 4.990680335507921, + "grad_norm": 0.34402596950531006, + "learning_rate": 8.119240042936303e-05, + "loss": 0.0062, + "step": 32130 + }, + { + "epoch": 4.992233612923268, + "grad_norm": 0.3928764760494232, + "learning_rate": 8.117947610196175e-05, + "loss": 0.0069, + "step": 32140 + }, + { + "epoch": 4.9937868903386144, + "grad_norm": 0.3335276246070862, + "learning_rate": 8.116654836481982e-05, + "loss": 0.0085, + "step": 32150 + }, + { + "epoch": 4.995340167753961, + "grad_norm": 0.23428142070770264, + "learning_rate": 8.115361721935106e-05, + "loss": 0.0103, + "step": 32160 + }, + { + "epoch": 4.996893445169308, + "grad_norm": 0.21186412870883942, + "learning_rate": 8.114068266696953e-05, + "loss": 0.0125, + "step": 32170 + }, + { + "epoch": 4.998446722584654, + "grad_norm": 0.2500457465648651, + "learning_rate": 8.112774470908978e-05, + "loss": 0.0084, + "step": 32180 + }, + { + "epoch": 5.0, + "grad_norm": 0.30541855096817017, + "learning_rate": 8.111480334712665e-05, + "loss": 0.0097, + "step": 32190 + }, + { + "epoch": 5.001553277415346, + "grad_norm": 0.33834847807884216, + "learning_rate": 8.110185858249542e-05, + "loss": 0.009, + "step": 32200 + }, + { + "epoch": 5.003106554830692, + "grad_norm": 0.18957029283046722, + "learning_rate": 8.108891041661168e-05, + "loss": 0.0076, + "step": 32210 + }, + { + "epoch": 5.004659832246039, + "grad_norm": 0.22763372957706451, + "learning_rate": 8.107595885089146e-05, + "loss": 0.0074, + "step": 32220 + }, + { + "epoch": 5.0062131096613856, + "grad_norm": 0.23594854772090912, + "learning_rate": 8.106300388675108e-05, + "loss": 0.0119, + "step": 32230 + }, + { + "epoch": 5.007766387076732, + "grad_norm": 0.19903090596199036, + "learning_rate": 8.105004552560731e-05, + "loss": 0.0114, + "step": 32240 + }, + { + "epoch": 5.009319664492078, + "grad_norm": 0.41493892669677734, + "learning_rate": 8.103708376887724e-05, + "loss": 0.014, + "step": 32250 + }, + { + "epoch": 5.010872941907425, + "grad_norm": 0.2725399136543274, + "learning_rate": 8.102411861797836e-05, + "loss": 0.0091, + "step": 32260 + }, + { + "epoch": 5.012426219322771, + "grad_norm": 0.21194672584533691, + "learning_rate": 8.101115007432851e-05, + "loss": 0.0109, + "step": 32270 + }, + { + "epoch": 5.013979496738117, + "grad_norm": 0.2512771487236023, + "learning_rate": 8.09981781393459e-05, + "loss": 0.0097, + "step": 32280 + }, + { + "epoch": 5.015532774153463, + "grad_norm": 0.3892413079738617, + "learning_rate": 8.098520281444915e-05, + "loss": 0.0108, + "step": 32290 + }, + { + "epoch": 5.0170860515688105, + "grad_norm": 0.226112961769104, + "learning_rate": 8.09722241010572e-05, + "loss": 0.0094, + "step": 32300 + }, + { + "epoch": 5.018639328984157, + "grad_norm": 0.34675684571266174, + "learning_rate": 8.095924200058939e-05, + "loss": 0.0094, + "step": 32310 + }, + { + "epoch": 5.020192606399503, + "grad_norm": 0.17524079978466034, + "learning_rate": 8.094625651446541e-05, + "loss": 0.0085, + "step": 32320 + }, + { + "epoch": 5.021745883814849, + "grad_norm": 0.206590473651886, + "learning_rate": 8.093326764410536e-05, + "loss": 0.0068, + "step": 32330 + }, + { + "epoch": 5.023299161230196, + "grad_norm": 0.3600185811519623, + "learning_rate": 8.092027539092966e-05, + "loss": 0.0099, + "step": 32340 + }, + { + "epoch": 5.024852438645542, + "grad_norm": 0.3932145833969116, + "learning_rate": 8.090727975635913e-05, + "loss": 0.0095, + "step": 32350 + }, + { + "epoch": 5.026405716060888, + "grad_norm": 0.46331948041915894, + "learning_rate": 8.089428074181497e-05, + "loss": 0.0099, + "step": 32360 + }, + { + "epoch": 5.0279589934762345, + "grad_norm": 0.2980809807777405, + "learning_rate": 8.088127834871871e-05, + "loss": 0.0126, + "step": 32370 + }, + { + "epoch": 5.029512270891582, + "grad_norm": 0.39716678857803345, + "learning_rate": 8.086827257849226e-05, + "loss": 0.0098, + "step": 32380 + }, + { + "epoch": 5.031065548306928, + "grad_norm": 0.2773441970348358, + "learning_rate": 8.085526343255795e-05, + "loss": 0.0115, + "step": 32390 + }, + { + "epoch": 5.032618825722274, + "grad_norm": 0.20822501182556152, + "learning_rate": 8.084225091233842e-05, + "loss": 0.0101, + "step": 32400 + }, + { + "epoch": 5.03417210313762, + "grad_norm": 0.6044703125953674, + "learning_rate": 8.082923501925668e-05, + "loss": 0.0101, + "step": 32410 + }, + { + "epoch": 5.035725380552967, + "grad_norm": 0.2766653001308441, + "learning_rate": 8.081621575473617e-05, + "loss": 0.0079, + "step": 32420 + }, + { + "epoch": 5.037278657968313, + "grad_norm": 0.3638305366039276, + "learning_rate": 8.080319312020064e-05, + "loss": 0.0121, + "step": 32430 + }, + { + "epoch": 5.0388319353836595, + "grad_norm": 0.30974459648132324, + "learning_rate": 8.079016711707421e-05, + "loss": 0.0095, + "step": 32440 + }, + { + "epoch": 5.040385212799006, + "grad_norm": 0.36642616987228394, + "learning_rate": 8.077713774678139e-05, + "loss": 0.0085, + "step": 32450 + }, + { + "epoch": 5.041938490214353, + "grad_norm": 0.3124522864818573, + "learning_rate": 8.076410501074707e-05, + "loss": 0.0074, + "step": 32460 + }, + { + "epoch": 5.043491767629699, + "grad_norm": 0.23362044990062714, + "learning_rate": 8.075106891039647e-05, + "loss": 0.0086, + "step": 32470 + }, + { + "epoch": 5.045045045045045, + "grad_norm": 0.19235974550247192, + "learning_rate": 8.073802944715523e-05, + "loss": 0.0073, + "step": 32480 + }, + { + "epoch": 5.046598322460391, + "grad_norm": 0.3261791169643402, + "learning_rate": 8.072498662244929e-05, + "loss": 0.008, + "step": 32490 + }, + { + "epoch": 5.048151599875737, + "grad_norm": 0.4818001389503479, + "learning_rate": 8.0711940437705e-05, + "loss": 0.0102, + "step": 32500 + }, + { + "epoch": 5.049704877291084, + "grad_norm": 0.20176951587200165, + "learning_rate": 8.06988908943491e-05, + "loss": 0.0073, + "step": 32510 + }, + { + "epoch": 5.051258154706431, + "grad_norm": 0.22109241783618927, + "learning_rate": 8.068583799380863e-05, + "loss": 0.0095, + "step": 32520 + }, + { + "epoch": 5.052811432121777, + "grad_norm": 0.31776368618011475, + "learning_rate": 8.067278173751104e-05, + "loss": 0.008, + "step": 32530 + }, + { + "epoch": 5.054364709537123, + "grad_norm": 0.20315411686897278, + "learning_rate": 8.065972212688417e-05, + "loss": 0.0122, + "step": 32540 + }, + { + "epoch": 5.05591798695247, + "grad_norm": 0.2088657021522522, + "learning_rate": 8.064665916335618e-05, + "loss": 0.0104, + "step": 32550 + }, + { + "epoch": 5.057471264367816, + "grad_norm": 0.351691871881485, + "learning_rate": 8.063359284835564e-05, + "loss": 0.0098, + "step": 32560 + }, + { + "epoch": 5.059024541783162, + "grad_norm": 0.2924782633781433, + "learning_rate": 8.062052318331142e-05, + "loss": 0.0096, + "step": 32570 + }, + { + "epoch": 5.0605778191985085, + "grad_norm": 0.25119543075561523, + "learning_rate": 8.060745016965283e-05, + "loss": 0.0105, + "step": 32580 + }, + { + "epoch": 5.0621310966138555, + "grad_norm": 0.2241852581501007, + "learning_rate": 8.059437380880952e-05, + "loss": 0.0142, + "step": 32590 + }, + { + "epoch": 5.063684374029202, + "grad_norm": 0.5208736062049866, + "learning_rate": 8.058129410221146e-05, + "loss": 0.0113, + "step": 32600 + }, + { + "epoch": 5.065237651444548, + "grad_norm": 0.22488577663898468, + "learning_rate": 8.056821105128908e-05, + "loss": 0.008, + "step": 32610 + }, + { + "epoch": 5.066790928859894, + "grad_norm": 0.33201271295547485, + "learning_rate": 8.05551246574731e-05, + "loss": 0.0079, + "step": 32620 + }, + { + "epoch": 5.068344206275241, + "grad_norm": 0.3047802448272705, + "learning_rate": 8.05420349221946e-05, + "loss": 0.0087, + "step": 32630 + }, + { + "epoch": 5.069897483690587, + "grad_norm": 0.30412301421165466, + "learning_rate": 8.05289418468851e-05, + "loss": 0.0072, + "step": 32640 + }, + { + "epoch": 5.071450761105933, + "grad_norm": 0.23200227320194244, + "learning_rate": 8.051584543297642e-05, + "loss": 0.0083, + "step": 32650 + }, + { + "epoch": 5.07300403852128, + "grad_norm": 0.26561063528060913, + "learning_rate": 8.050274568190074e-05, + "loss": 0.0103, + "step": 32660 + }, + { + "epoch": 5.074557315936627, + "grad_norm": 0.17383603751659393, + "learning_rate": 8.048964259509067e-05, + "loss": 0.0103, + "step": 32670 + }, + { + "epoch": 5.076110593351973, + "grad_norm": 0.21528421342372894, + "learning_rate": 8.047653617397914e-05, + "loss": 0.0094, + "step": 32680 + }, + { + "epoch": 5.077663870767319, + "grad_norm": 0.5407341122627258, + "learning_rate": 8.046342641999941e-05, + "loss": 0.0123, + "step": 32690 + }, + { + "epoch": 5.079217148182665, + "grad_norm": 0.34818965196609497, + "learning_rate": 8.045031333458517e-05, + "loss": 0.0108, + "step": 32700 + }, + { + "epoch": 5.080770425598012, + "grad_norm": 0.3046015501022339, + "learning_rate": 8.043719691917047e-05, + "loss": 0.0115, + "step": 32710 + }, + { + "epoch": 5.082323703013358, + "grad_norm": 0.3676663041114807, + "learning_rate": 8.042407717518966e-05, + "loss": 0.0099, + "step": 32720 + }, + { + "epoch": 5.0838769804287045, + "grad_norm": 0.1900905817747116, + "learning_rate": 8.041095410407751e-05, + "loss": 0.0086, + "step": 32730 + }, + { + "epoch": 5.085430257844051, + "grad_norm": 0.4096532464027405, + "learning_rate": 8.039782770726913e-05, + "loss": 0.0105, + "step": 32740 + }, + { + "epoch": 5.086983535259398, + "grad_norm": 0.29241204261779785, + "learning_rate": 8.038469798620004e-05, + "loss": 0.0101, + "step": 32750 + }, + { + "epoch": 5.088536812674744, + "grad_norm": 0.19048193097114563, + "learning_rate": 8.037156494230604e-05, + "loss": 0.0093, + "step": 32760 + }, + { + "epoch": 5.09009009009009, + "grad_norm": 0.23536591231822968, + "learning_rate": 8.035842857702338e-05, + "loss": 0.01, + "step": 32770 + }, + { + "epoch": 5.091643367505436, + "grad_norm": 0.3535400927066803, + "learning_rate": 8.03452888917886e-05, + "loss": 0.0091, + "step": 32780 + }, + { + "epoch": 5.093196644920783, + "grad_norm": 0.20931997895240784, + "learning_rate": 8.033214588803866e-05, + "loss": 0.0094, + "step": 32790 + }, + { + "epoch": 5.094749922336129, + "grad_norm": 0.11933622509241104, + "learning_rate": 8.031899956721083e-05, + "loss": 0.0081, + "step": 32800 + }, + { + "epoch": 5.096303199751476, + "grad_norm": 0.16754019260406494, + "learning_rate": 8.030584993074282e-05, + "loss": 0.0097, + "step": 32810 + }, + { + "epoch": 5.097856477166822, + "grad_norm": 0.3049464821815491, + "learning_rate": 8.02926969800726e-05, + "loss": 0.0096, + "step": 32820 + }, + { + "epoch": 5.099409754582168, + "grad_norm": 0.22035405039787292, + "learning_rate": 8.027954071663859e-05, + "loss": 0.0102, + "step": 32830 + }, + { + "epoch": 5.100963031997515, + "grad_norm": 0.2347090095281601, + "learning_rate": 8.026638114187954e-05, + "loss": 0.0105, + "step": 32840 + }, + { + "epoch": 5.102516309412861, + "grad_norm": 0.15137609839439392, + "learning_rate": 8.025321825723456e-05, + "loss": 0.0118, + "step": 32850 + }, + { + "epoch": 5.104069586828207, + "grad_norm": 0.2459876388311386, + "learning_rate": 8.02400520641431e-05, + "loss": 0.0106, + "step": 32860 + }, + { + "epoch": 5.1056228642435535, + "grad_norm": 0.28098535537719727, + "learning_rate": 8.022688256404501e-05, + "loss": 0.0093, + "step": 32870 + }, + { + "epoch": 5.1071761416589005, + "grad_norm": 0.3416535556316376, + "learning_rate": 8.02137097583805e-05, + "loss": 0.0102, + "step": 32880 + }, + { + "epoch": 5.108729419074247, + "grad_norm": 0.2896071970462799, + "learning_rate": 8.02005336485901e-05, + "loss": 0.0068, + "step": 32890 + }, + { + "epoch": 5.110282696489593, + "grad_norm": 0.24129191040992737, + "learning_rate": 8.018735423611476e-05, + "loss": 0.009, + "step": 32900 + }, + { + "epoch": 5.111835973904939, + "grad_norm": 0.3035406470298767, + "learning_rate": 8.017417152239574e-05, + "loss": 0.0101, + "step": 32910 + }, + { + "epoch": 5.113389251320286, + "grad_norm": 0.28308597207069397, + "learning_rate": 8.01609855088747e-05, + "loss": 0.01, + "step": 32920 + }, + { + "epoch": 5.114942528735632, + "grad_norm": 0.291411817073822, + "learning_rate": 8.014779619699362e-05, + "loss": 0.0073, + "step": 32930 + }, + { + "epoch": 5.116495806150978, + "grad_norm": 0.3996778428554535, + "learning_rate": 8.013460358819489e-05, + "loss": 0.0107, + "step": 32940 + }, + { + "epoch": 5.118049083566325, + "grad_norm": 0.567974328994751, + "learning_rate": 8.01214076839212e-05, + "loss": 0.0112, + "step": 32950 + }, + { + "epoch": 5.119602360981672, + "grad_norm": 0.3510206639766693, + "learning_rate": 8.010820848561565e-05, + "loss": 0.011, + "step": 32960 + }, + { + "epoch": 5.121155638397018, + "grad_norm": 0.22813671827316284, + "learning_rate": 8.009500599472171e-05, + "loss": 0.0119, + "step": 32970 + }, + { + "epoch": 5.122708915812364, + "grad_norm": 0.24378103017807007, + "learning_rate": 8.008180021268314e-05, + "loss": 0.0097, + "step": 32980 + }, + { + "epoch": 5.12426219322771, + "grad_norm": 0.323882132768631, + "learning_rate": 8.006859114094414e-05, + "loss": 0.0116, + "step": 32990 + }, + { + "epoch": 5.125815470643057, + "grad_norm": 0.2670682370662689, + "learning_rate": 8.005537878094921e-05, + "loss": 0.0073, + "step": 33000 + }, + { + "epoch": 5.127368748058403, + "grad_norm": 0.4593847990036011, + "learning_rate": 8.004216313414323e-05, + "loss": 0.0098, + "step": 33010 + }, + { + "epoch": 5.1289220254737495, + "grad_norm": 0.25744763016700745, + "learning_rate": 8.002894420197149e-05, + "loss": 0.0076, + "step": 33020 + }, + { + "epoch": 5.130475302889096, + "grad_norm": 0.1841530054807663, + "learning_rate": 8.001572198587954e-05, + "loss": 0.009, + "step": 33030 + }, + { + "epoch": 5.132028580304443, + "grad_norm": 0.27506595849990845, + "learning_rate": 8.000249648731338e-05, + "loss": 0.0077, + "step": 33040 + }, + { + "epoch": 5.133581857719789, + "grad_norm": 0.6901695132255554, + "learning_rate": 7.998926770771928e-05, + "loss": 0.0104, + "step": 33050 + }, + { + "epoch": 5.135135135135135, + "grad_norm": 0.2795543372631073, + "learning_rate": 7.997603564854397e-05, + "loss": 0.0083, + "step": 33060 + }, + { + "epoch": 5.136688412550481, + "grad_norm": 0.27014511823654175, + "learning_rate": 7.996280031123448e-05, + "loss": 0.0124, + "step": 33070 + }, + { + "epoch": 5.138241689965828, + "grad_norm": 0.12232808023691177, + "learning_rate": 7.994956169723818e-05, + "loss": 0.0086, + "step": 33080 + }, + { + "epoch": 5.1397949673811745, + "grad_norm": 0.32639315724372864, + "learning_rate": 7.993631980800285e-05, + "loss": 0.008, + "step": 33090 + }, + { + "epoch": 5.141348244796521, + "grad_norm": 0.3453945815563202, + "learning_rate": 7.992307464497659e-05, + "loss": 0.0087, + "step": 33100 + }, + { + "epoch": 5.142901522211867, + "grad_norm": 0.3241943418979645, + "learning_rate": 7.990982620960787e-05, + "loss": 0.0095, + "step": 33110 + }, + { + "epoch": 5.144454799627214, + "grad_norm": 0.23020119965076447, + "learning_rate": 7.989657450334554e-05, + "loss": 0.0082, + "step": 33120 + }, + { + "epoch": 5.14600807704256, + "grad_norm": 0.4769623875617981, + "learning_rate": 7.988331952763877e-05, + "loss": 0.0118, + "step": 33130 + }, + { + "epoch": 5.147561354457906, + "grad_norm": 0.3865530788898468, + "learning_rate": 7.987006128393709e-05, + "loss": 0.0103, + "step": 33140 + }, + { + "epoch": 5.149114631873252, + "grad_norm": 0.3560277819633484, + "learning_rate": 7.985679977369043e-05, + "loss": 0.0115, + "step": 33150 + }, + { + "epoch": 5.1506679092885985, + "grad_norm": 0.3034554421901703, + "learning_rate": 7.984353499834902e-05, + "loss": 0.0106, + "step": 33160 + }, + { + "epoch": 5.152221186703946, + "grad_norm": 0.32863664627075195, + "learning_rate": 7.983026695936351e-05, + "loss": 0.0084, + "step": 33170 + }, + { + "epoch": 5.153774464119292, + "grad_norm": 0.2575600743293762, + "learning_rate": 7.981699565818486e-05, + "loss": 0.0087, + "step": 33180 + }, + { + "epoch": 5.155327741534638, + "grad_norm": 0.16117915511131287, + "learning_rate": 7.980372109626437e-05, + "loss": 0.0095, + "step": 33190 + }, + { + "epoch": 5.156881018949984, + "grad_norm": 0.24371598660945892, + "learning_rate": 7.979044327505375e-05, + "loss": 0.0129, + "step": 33200 + }, + { + "epoch": 5.158434296365331, + "grad_norm": 0.4866076707839966, + "learning_rate": 7.977716219600506e-05, + "loss": 0.0083, + "step": 33210 + }, + { + "epoch": 5.159987573780677, + "grad_norm": 0.17242300510406494, + "learning_rate": 7.97638778605707e-05, + "loss": 0.0077, + "step": 33220 + }, + { + "epoch": 5.1615408511960235, + "grad_norm": 0.30946627259254456, + "learning_rate": 7.975059027020338e-05, + "loss": 0.0082, + "step": 33230 + }, + { + "epoch": 5.16309412861137, + "grad_norm": 0.1827787160873413, + "learning_rate": 7.973729942635623e-05, + "loss": 0.0092, + "step": 33240 + }, + { + "epoch": 5.164647406026717, + "grad_norm": 0.2575138807296753, + "learning_rate": 7.972400533048273e-05, + "loss": 0.0081, + "step": 33250 + }, + { + "epoch": 5.166200683442063, + "grad_norm": 0.3561016023159027, + "learning_rate": 7.97107079840367e-05, + "loss": 0.0122, + "step": 33260 + }, + { + "epoch": 5.167753960857409, + "grad_norm": 0.20403650403022766, + "learning_rate": 7.969740738847231e-05, + "loss": 0.0067, + "step": 33270 + }, + { + "epoch": 5.169307238272755, + "grad_norm": 0.20262068510055542, + "learning_rate": 7.968410354524411e-05, + "loss": 0.0111, + "step": 33280 + }, + { + "epoch": 5.170860515688102, + "grad_norm": 0.2164425551891327, + "learning_rate": 7.967079645580697e-05, + "loss": 0.0064, + "step": 33290 + }, + { + "epoch": 5.172413793103448, + "grad_norm": 0.2570211887359619, + "learning_rate": 7.965748612161612e-05, + "loss": 0.0073, + "step": 33300 + }, + { + "epoch": 5.173967070518795, + "grad_norm": 0.279333233833313, + "learning_rate": 7.96441725441272e-05, + "loss": 0.0101, + "step": 33310 + }, + { + "epoch": 5.175520347934141, + "grad_norm": 0.2705356478691101, + "learning_rate": 7.963085572479614e-05, + "loss": 0.0098, + "step": 33320 + }, + { + "epoch": 5.177073625349488, + "grad_norm": 0.25865042209625244, + "learning_rate": 7.961753566507924e-05, + "loss": 0.0083, + "step": 33330 + }, + { + "epoch": 5.178626902764834, + "grad_norm": 0.24768999218940735, + "learning_rate": 7.960421236643316e-05, + "loss": 0.0103, + "step": 33340 + }, + { + "epoch": 5.18018018018018, + "grad_norm": 0.3549000322818756, + "learning_rate": 7.959088583031496e-05, + "loss": 0.0072, + "step": 33350 + }, + { + "epoch": 5.181733457595526, + "grad_norm": 0.21015632152557373, + "learning_rate": 7.957755605818194e-05, + "loss": 0.0084, + "step": 33360 + }, + { + "epoch": 5.183286735010873, + "grad_norm": 0.38742542266845703, + "learning_rate": 7.956422305149185e-05, + "loss": 0.0101, + "step": 33370 + }, + { + "epoch": 5.1848400124262195, + "grad_norm": 0.34552645683288574, + "learning_rate": 7.95508868117028e-05, + "loss": 0.0149, + "step": 33380 + }, + { + "epoch": 5.186393289841566, + "grad_norm": 0.19309170544147491, + "learning_rate": 7.953754734027318e-05, + "loss": 0.0102, + "step": 33390 + }, + { + "epoch": 5.187946567256912, + "grad_norm": 0.18394771218299866, + "learning_rate": 7.952420463866182e-05, + "loss": 0.0097, + "step": 33400 + }, + { + "epoch": 5.189499844672259, + "grad_norm": 0.40726637840270996, + "learning_rate": 7.951085870832782e-05, + "loss": 0.0082, + "step": 33410 + }, + { + "epoch": 5.191053122087605, + "grad_norm": 0.19244952499866486, + "learning_rate": 7.949750955073067e-05, + "loss": 0.0109, + "step": 33420 + }, + { + "epoch": 5.192606399502951, + "grad_norm": 0.30722156167030334, + "learning_rate": 7.948415716733022e-05, + "loss": 0.0104, + "step": 33430 + }, + { + "epoch": 5.194159676918297, + "grad_norm": 0.29191312193870544, + "learning_rate": 7.947080155958669e-05, + "loss": 0.0083, + "step": 33440 + }, + { + "epoch": 5.1957129543336436, + "grad_norm": 0.31551089882850647, + "learning_rate": 7.94574427289606e-05, + "loss": 0.0068, + "step": 33450 + }, + { + "epoch": 5.197266231748991, + "grad_norm": 0.5253422856330872, + "learning_rate": 7.944408067691284e-05, + "loss": 0.0084, + "step": 33460 + }, + { + "epoch": 5.198819509164337, + "grad_norm": 0.29516395926475525, + "learning_rate": 7.943071540490473e-05, + "loss": 0.0073, + "step": 33470 + }, + { + "epoch": 5.200372786579683, + "grad_norm": 0.279774010181427, + "learning_rate": 7.94173469143978e-05, + "loss": 0.0064, + "step": 33480 + }, + { + "epoch": 5.201926063995029, + "grad_norm": 0.17011874914169312, + "learning_rate": 7.940397520685406e-05, + "loss": 0.009, + "step": 33490 + }, + { + "epoch": 5.203479341410376, + "grad_norm": 0.4736815094947815, + "learning_rate": 7.939060028373577e-05, + "loss": 0.0092, + "step": 33500 + }, + { + "epoch": 5.205032618825722, + "grad_norm": 0.31094542145729065, + "learning_rate": 7.937722214650565e-05, + "loss": 0.0096, + "step": 33510 + }, + { + "epoch": 5.2065858962410685, + "grad_norm": 0.42709091305732727, + "learning_rate": 7.936384079662666e-05, + "loss": 0.0122, + "step": 33520 + }, + { + "epoch": 5.208139173656415, + "grad_norm": 0.2204519659280777, + "learning_rate": 7.93504562355622e-05, + "loss": 0.01, + "step": 33530 + }, + { + "epoch": 5.209692451071762, + "grad_norm": 0.3961595892906189, + "learning_rate": 7.933706846477599e-05, + "loss": 0.0156, + "step": 33540 + }, + { + "epoch": 5.211245728487108, + "grad_norm": 0.23671665787696838, + "learning_rate": 7.932367748573206e-05, + "loss": 0.0102, + "step": 33550 + }, + { + "epoch": 5.212799005902454, + "grad_norm": 0.2876068353652954, + "learning_rate": 7.931028329989485e-05, + "loss": 0.0081, + "step": 33560 + }, + { + "epoch": 5.2143522833178, + "grad_norm": 0.24508146941661835, + "learning_rate": 7.929688590872913e-05, + "loss": 0.0091, + "step": 33570 + }, + { + "epoch": 5.215905560733147, + "grad_norm": 0.34346169233322144, + "learning_rate": 7.928348531370003e-05, + "loss": 0.0118, + "step": 33580 + }, + { + "epoch": 5.217458838148493, + "grad_norm": 0.5493550896644592, + "learning_rate": 7.927008151627297e-05, + "loss": 0.013, + "step": 33590 + }, + { + "epoch": 5.21901211556384, + "grad_norm": 0.529228687286377, + "learning_rate": 7.925667451791383e-05, + "loss": 0.0114, + "step": 33600 + }, + { + "epoch": 5.220565392979186, + "grad_norm": 0.25746235251426697, + "learning_rate": 7.924326432008874e-05, + "loss": 0.009, + "step": 33610 + }, + { + "epoch": 5.222118670394533, + "grad_norm": 0.2264828234910965, + "learning_rate": 7.922985092426422e-05, + "loss": 0.0085, + "step": 33620 + }, + { + "epoch": 5.223671947809879, + "grad_norm": 0.17995744943618774, + "learning_rate": 7.921643433190717e-05, + "loss": 0.0119, + "step": 33630 + }, + { + "epoch": 5.225225225225225, + "grad_norm": 0.4388698637485504, + "learning_rate": 7.920301454448478e-05, + "loss": 0.0095, + "step": 33640 + }, + { + "epoch": 5.226778502640571, + "grad_norm": 0.2519659996032715, + "learning_rate": 7.918959156346461e-05, + "loss": 0.0123, + "step": 33650 + }, + { + "epoch": 5.228331780055918, + "grad_norm": 0.403739333152771, + "learning_rate": 7.91761653903146e-05, + "loss": 0.012, + "step": 33660 + }, + { + "epoch": 5.2298850574712645, + "grad_norm": 0.22221773862838745, + "learning_rate": 7.916273602650302e-05, + "loss": 0.0103, + "step": 33670 + }, + { + "epoch": 5.231438334886611, + "grad_norm": 0.26718220114707947, + "learning_rate": 7.914930347349847e-05, + "loss": 0.0099, + "step": 33680 + }, + { + "epoch": 5.232991612301957, + "grad_norm": 0.19220221042633057, + "learning_rate": 7.913586773276992e-05, + "loss": 0.0088, + "step": 33690 + }, + { + "epoch": 5.234544889717304, + "grad_norm": 0.4293426275253296, + "learning_rate": 7.912242880578667e-05, + "loss": 0.0089, + "step": 33700 + }, + { + "epoch": 5.23609816713265, + "grad_norm": 0.22428539395332336, + "learning_rate": 7.910898669401839e-05, + "loss": 0.0076, + "step": 33710 + }, + { + "epoch": 5.237651444547996, + "grad_norm": 0.29027146100997925, + "learning_rate": 7.909554139893511e-05, + "loss": 0.0072, + "step": 33720 + }, + { + "epoch": 5.239204721963342, + "grad_norm": 0.1449016034603119, + "learning_rate": 7.908209292200715e-05, + "loss": 0.0067, + "step": 33730 + }, + { + "epoch": 5.240757999378689, + "grad_norm": 0.2525518238544464, + "learning_rate": 7.906864126470523e-05, + "loss": 0.0083, + "step": 33740 + }, + { + "epoch": 5.242311276794036, + "grad_norm": 0.3264566659927368, + "learning_rate": 7.905518642850041e-05, + "loss": 0.0138, + "step": 33750 + }, + { + "epoch": 5.243864554209382, + "grad_norm": 0.2560894191265106, + "learning_rate": 7.904172841486409e-05, + "loss": 0.0094, + "step": 33760 + }, + { + "epoch": 5.245417831624728, + "grad_norm": 0.3513851463794708, + "learning_rate": 7.902826722526801e-05, + "loss": 0.0098, + "step": 33770 + }, + { + "epoch": 5.246971109040074, + "grad_norm": 0.32772547006607056, + "learning_rate": 7.901480286118427e-05, + "loss": 0.0123, + "step": 33780 + }, + { + "epoch": 5.248524386455421, + "grad_norm": 0.15154172480106354, + "learning_rate": 7.900133532408531e-05, + "loss": 0.0061, + "step": 33790 + }, + { + "epoch": 5.250077663870767, + "grad_norm": 0.33583301305770874, + "learning_rate": 7.898786461544395e-05, + "loss": 0.0084, + "step": 33800 + }, + { + "epoch": 5.2516309412861135, + "grad_norm": 0.15867209434509277, + "learning_rate": 7.897439073673325e-05, + "loss": 0.0085, + "step": 33810 + }, + { + "epoch": 5.25318421870146, + "grad_norm": 0.36491405963897705, + "learning_rate": 7.896091368942677e-05, + "loss": 0.009, + "step": 33820 + }, + { + "epoch": 5.254737496116807, + "grad_norm": 0.2282732129096985, + "learning_rate": 7.894743347499832e-05, + "loss": 0.0066, + "step": 33830 + }, + { + "epoch": 5.256290773532153, + "grad_norm": 0.5570105314254761, + "learning_rate": 7.893395009492203e-05, + "loss": 0.0113, + "step": 33840 + }, + { + "epoch": 5.257844050947499, + "grad_norm": 0.40332698822021484, + "learning_rate": 7.892046355067248e-05, + "loss": 0.008, + "step": 33850 + }, + { + "epoch": 5.259397328362845, + "grad_norm": 0.38372719287872314, + "learning_rate": 7.890697384372451e-05, + "loss": 0.0104, + "step": 33860 + }, + { + "epoch": 5.260950605778192, + "grad_norm": 0.19526712596416473, + "learning_rate": 7.889348097555336e-05, + "loss": 0.0102, + "step": 33870 + }, + { + "epoch": 5.2625038831935385, + "grad_norm": 0.14701077342033386, + "learning_rate": 7.887998494763455e-05, + "loss": 0.0089, + "step": 33880 + }, + { + "epoch": 5.264057160608885, + "grad_norm": 0.8586593866348267, + "learning_rate": 7.886648576144404e-05, + "loss": 0.0104, + "step": 33890 + }, + { + "epoch": 5.265610438024231, + "grad_norm": 0.16387952864170074, + "learning_rate": 7.885298341845802e-05, + "loss": 0.0085, + "step": 33900 + }, + { + "epoch": 5.267163715439578, + "grad_norm": 0.1875910609960556, + "learning_rate": 7.883947792015311e-05, + "loss": 0.0082, + "step": 33910 + }, + { + "epoch": 5.268716992854924, + "grad_norm": 0.17304818332195282, + "learning_rate": 7.882596926800628e-05, + "loss": 0.0071, + "step": 33920 + }, + { + "epoch": 5.27027027027027, + "grad_norm": 0.46108478307724, + "learning_rate": 7.881245746349477e-05, + "loss": 0.0087, + "step": 33930 + }, + { + "epoch": 5.271823547685616, + "grad_norm": 0.2569412291049957, + "learning_rate": 7.879894250809623e-05, + "loss": 0.0079, + "step": 33940 + }, + { + "epoch": 5.273376825100963, + "grad_norm": 0.2834755480289459, + "learning_rate": 7.878542440328865e-05, + "loss": 0.0088, + "step": 33950 + }, + { + "epoch": 5.27493010251631, + "grad_norm": 0.18165862560272217, + "learning_rate": 7.877190315055031e-05, + "loss": 0.0069, + "step": 33960 + }, + { + "epoch": 5.276483379931656, + "grad_norm": 0.22886043787002563, + "learning_rate": 7.875837875135991e-05, + "loss": 0.0121, + "step": 33970 + }, + { + "epoch": 5.278036657347002, + "grad_norm": 0.2697429060935974, + "learning_rate": 7.874485120719646e-05, + "loss": 0.0107, + "step": 33980 + }, + { + "epoch": 5.279589934762349, + "grad_norm": 0.17874310910701752, + "learning_rate": 7.873132051953928e-05, + "loss": 0.0071, + "step": 33990 + }, + { + "epoch": 5.281143212177695, + "grad_norm": 0.38873109221458435, + "learning_rate": 7.87177866898681e-05, + "loss": 0.0086, + "step": 34000 + }, + { + "epoch": 5.282696489593041, + "grad_norm": 0.2145019918680191, + "learning_rate": 7.870424971966294e-05, + "loss": 0.0081, + "step": 34010 + }, + { + "epoch": 5.284249767008387, + "grad_norm": 0.23763826489448547, + "learning_rate": 7.869070961040419e-05, + "loss": 0.0094, + "step": 34020 + }, + { + "epoch": 5.285803044423734, + "grad_norm": 0.23563125729560852, + "learning_rate": 7.867716636357257e-05, + "loss": 0.0099, + "step": 34030 + }, + { + "epoch": 5.287356321839081, + "grad_norm": 0.2437421977519989, + "learning_rate": 7.866361998064915e-05, + "loss": 0.0089, + "step": 34040 + }, + { + "epoch": 5.288909599254427, + "grad_norm": 0.2710791230201721, + "learning_rate": 7.865007046311534e-05, + "loss": 0.0093, + "step": 34050 + }, + { + "epoch": 5.290462876669773, + "grad_norm": 0.22734184563159943, + "learning_rate": 7.86365178124529e-05, + "loss": 0.0076, + "step": 34060 + }, + { + "epoch": 5.29201615408512, + "grad_norm": 0.25709718465805054, + "learning_rate": 7.862296203014394e-05, + "loss": 0.0114, + "step": 34070 + }, + { + "epoch": 5.293569431500466, + "grad_norm": 0.22938789427280426, + "learning_rate": 7.860940311767088e-05, + "loss": 0.0085, + "step": 34080 + }, + { + "epoch": 5.295122708915812, + "grad_norm": 0.23788344860076904, + "learning_rate": 7.85958410765165e-05, + "loss": 0.008, + "step": 34090 + }, + { + "epoch": 5.2966759863311585, + "grad_norm": 0.3801719844341278, + "learning_rate": 7.858227590816394e-05, + "loss": 0.0084, + "step": 34100 + }, + { + "epoch": 5.298229263746505, + "grad_norm": 0.15475940704345703, + "learning_rate": 7.856870761409664e-05, + "loss": 0.0113, + "step": 34110 + }, + { + "epoch": 5.299782541161852, + "grad_norm": 0.3263712525367737, + "learning_rate": 7.855513619579846e-05, + "loss": 0.0125, + "step": 34120 + }, + { + "epoch": 5.301335818577198, + "grad_norm": 0.2902296483516693, + "learning_rate": 7.85415616547535e-05, + "loss": 0.0086, + "step": 34130 + }, + { + "epoch": 5.302889095992544, + "grad_norm": 0.2650800347328186, + "learning_rate": 7.852798399244627e-05, + "loss": 0.0077, + "step": 34140 + }, + { + "epoch": 5.30444237340789, + "grad_norm": 0.19673305749893188, + "learning_rate": 7.851440321036161e-05, + "loss": 0.0086, + "step": 34150 + }, + { + "epoch": 5.305995650823237, + "grad_norm": 0.3799756169319153, + "learning_rate": 7.850081930998468e-05, + "loss": 0.008, + "step": 34160 + }, + { + "epoch": 5.3075489282385835, + "grad_norm": 0.3298395276069641, + "learning_rate": 7.8487232292801e-05, + "loss": 0.0073, + "step": 34170 + }, + { + "epoch": 5.30910220565393, + "grad_norm": 0.15770362317562103, + "learning_rate": 7.847364216029642e-05, + "loss": 0.0068, + "step": 34180 + }, + { + "epoch": 5.310655483069276, + "grad_norm": 0.3022010326385498, + "learning_rate": 7.846004891395716e-05, + "loss": 0.0073, + "step": 34190 + }, + { + "epoch": 5.312208760484623, + "grad_norm": 0.17537666857242584, + "learning_rate": 7.844645255526972e-05, + "loss": 0.0076, + "step": 34200 + }, + { + "epoch": 5.313762037899969, + "grad_norm": 0.2931879758834839, + "learning_rate": 7.843285308572101e-05, + "loss": 0.0084, + "step": 34210 + }, + { + "epoch": 5.315315315315315, + "grad_norm": 0.3234827518463135, + "learning_rate": 7.841925050679823e-05, + "loss": 0.009, + "step": 34220 + }, + { + "epoch": 5.316868592730661, + "grad_norm": 0.1946500539779663, + "learning_rate": 7.840564481998895e-05, + "loss": 0.0074, + "step": 34230 + }, + { + "epoch": 5.318421870146008, + "grad_norm": 0.8167595863342285, + "learning_rate": 7.839203602678105e-05, + "loss": 0.0077, + "step": 34240 + }, + { + "epoch": 5.319975147561355, + "grad_norm": 0.23156116902828217, + "learning_rate": 7.837842412866279e-05, + "loss": 0.009, + "step": 34250 + }, + { + "epoch": 5.321528424976701, + "grad_norm": 0.46614184975624084, + "learning_rate": 7.836480912712272e-05, + "loss": 0.0078, + "step": 34260 + }, + { + "epoch": 5.323081702392047, + "grad_norm": 0.2879970073699951, + "learning_rate": 7.835119102364976e-05, + "loss": 0.0096, + "step": 34270 + }, + { + "epoch": 5.324634979807394, + "grad_norm": 0.2578183710575104, + "learning_rate": 7.83375698197332e-05, + "loss": 0.0103, + "step": 34280 + }, + { + "epoch": 5.32618825722274, + "grad_norm": 0.38068127632141113, + "learning_rate": 7.83239455168626e-05, + "loss": 0.0111, + "step": 34290 + }, + { + "epoch": 5.327741534638086, + "grad_norm": 0.3558367192745209, + "learning_rate": 7.83103181165279e-05, + "loss": 0.0093, + "step": 34300 + }, + { + "epoch": 5.3292948120534325, + "grad_norm": 0.3981821835041046, + "learning_rate": 7.829668762021937e-05, + "loss": 0.0107, + "step": 34310 + }, + { + "epoch": 5.3308480894687795, + "grad_norm": 0.28433191776275635, + "learning_rate": 7.828305402942764e-05, + "loss": 0.0093, + "step": 34320 + }, + { + "epoch": 5.332401366884126, + "grad_norm": 0.251329630613327, + "learning_rate": 7.826941734564363e-05, + "loss": 0.0097, + "step": 34330 + }, + { + "epoch": 5.333954644299472, + "grad_norm": 0.22945918142795563, + "learning_rate": 7.825577757035865e-05, + "loss": 0.0076, + "step": 34340 + }, + { + "epoch": 5.335507921714818, + "grad_norm": 0.19252847135066986, + "learning_rate": 7.824213470506431e-05, + "loss": 0.0066, + "step": 34350 + }, + { + "epoch": 5.337061199130165, + "grad_norm": 0.4285889267921448, + "learning_rate": 7.822848875125257e-05, + "loss": 0.0103, + "step": 34360 + }, + { + "epoch": 5.338614476545511, + "grad_norm": 0.24423201382160187, + "learning_rate": 7.821483971041576e-05, + "loss": 0.0084, + "step": 34370 + }, + { + "epoch": 5.340167753960857, + "grad_norm": 0.224773570895195, + "learning_rate": 7.820118758404649e-05, + "loss": 0.0088, + "step": 34380 + }, + { + "epoch": 5.341721031376204, + "grad_norm": 0.14797046780586243, + "learning_rate": 7.818753237363776e-05, + "loss": 0.0089, + "step": 34390 + }, + { + "epoch": 5.34327430879155, + "grad_norm": 0.21032746136188507, + "learning_rate": 7.817387408068286e-05, + "loss": 0.008, + "step": 34400 + }, + { + "epoch": 5.344827586206897, + "grad_norm": 0.364835649728775, + "learning_rate": 7.816021270667544e-05, + "loss": 0.0091, + "step": 34410 + }, + { + "epoch": 5.346380863622243, + "grad_norm": 0.27370548248291016, + "learning_rate": 7.81465482531095e-05, + "loss": 0.0091, + "step": 34420 + }, + { + "epoch": 5.347934141037589, + "grad_norm": 0.6152792572975159, + "learning_rate": 7.813288072147938e-05, + "loss": 0.0103, + "step": 34430 + }, + { + "epoch": 5.349487418452935, + "grad_norm": 0.35896098613739014, + "learning_rate": 7.811921011327972e-05, + "loss": 0.0124, + "step": 34440 + }, + { + "epoch": 5.351040695868282, + "grad_norm": 0.3808845579624176, + "learning_rate": 7.810553643000549e-05, + "loss": 0.01, + "step": 34450 + }, + { + "epoch": 5.3525939732836285, + "grad_norm": 0.31089600920677185, + "learning_rate": 7.809185967315206e-05, + "loss": 0.0095, + "step": 34460 + }, + { + "epoch": 5.354147250698975, + "grad_norm": 0.33070650696754456, + "learning_rate": 7.80781798442151e-05, + "loss": 0.0077, + "step": 34470 + }, + { + "epoch": 5.355700528114321, + "grad_norm": 0.28593045473098755, + "learning_rate": 7.806449694469061e-05, + "loss": 0.0093, + "step": 34480 + }, + { + "epoch": 5.357253805529668, + "grad_norm": 0.3509925305843353, + "learning_rate": 7.805081097607492e-05, + "loss": 0.01, + "step": 34490 + }, + { + "epoch": 5.358807082945014, + "grad_norm": 0.14816847443580627, + "learning_rate": 7.803712193986474e-05, + "loss": 0.0093, + "step": 34500 + }, + { + "epoch": 5.36036036036036, + "grad_norm": 0.27043700218200684, + "learning_rate": 7.802342983755702e-05, + "loss": 0.0076, + "step": 34510 + }, + { + "epoch": 5.361913637775706, + "grad_norm": 0.1844574362039566, + "learning_rate": 7.800973467064918e-05, + "loss": 0.0103, + "step": 34520 + }, + { + "epoch": 5.3634669151910535, + "grad_norm": 0.33994218707084656, + "learning_rate": 7.799603644063884e-05, + "loss": 0.0089, + "step": 34530 + }, + { + "epoch": 5.3650201926064, + "grad_norm": 0.38284367322921753, + "learning_rate": 7.798233514902405e-05, + "loss": 0.0093, + "step": 34540 + }, + { + "epoch": 5.366573470021746, + "grad_norm": 0.3216160237789154, + "learning_rate": 7.796863079730318e-05, + "loss": 0.0063, + "step": 34550 + }, + { + "epoch": 5.368126747437092, + "grad_norm": 0.18751133978366852, + "learning_rate": 7.795492338697488e-05, + "loss": 0.0094, + "step": 34560 + }, + { + "epoch": 5.369680024852439, + "grad_norm": 0.24098928272724152, + "learning_rate": 7.794121291953819e-05, + "loss": 0.0085, + "step": 34570 + }, + { + "epoch": 5.371233302267785, + "grad_norm": 0.5696064233779907, + "learning_rate": 7.792749939649246e-05, + "loss": 0.0088, + "step": 34580 + }, + { + "epoch": 5.372786579683131, + "grad_norm": 0.2374163717031479, + "learning_rate": 7.79137828193374e-05, + "loss": 0.008, + "step": 34590 + }, + { + "epoch": 5.3743398570984775, + "grad_norm": 0.3493524193763733, + "learning_rate": 7.790006318957301e-05, + "loss": 0.0106, + "step": 34600 + }, + { + "epoch": 5.375893134513825, + "grad_norm": 0.1661684811115265, + "learning_rate": 7.788634050869965e-05, + "loss": 0.0082, + "step": 34610 + }, + { + "epoch": 5.377446411929171, + "grad_norm": 0.34257686138153076, + "learning_rate": 7.787261477821803e-05, + "loss": 0.0144, + "step": 34620 + }, + { + "epoch": 5.378999689344517, + "grad_norm": 0.16655109822750092, + "learning_rate": 7.785888599962916e-05, + "loss": 0.007, + "step": 34630 + }, + { + "epoch": 5.380552966759863, + "grad_norm": 0.3936655819416046, + "learning_rate": 7.784515417443439e-05, + "loss": 0.0075, + "step": 34640 + }, + { + "epoch": 5.38210624417521, + "grad_norm": 0.25367435812950134, + "learning_rate": 7.783141930413545e-05, + "loss": 0.0091, + "step": 34650 + }, + { + "epoch": 5.383659521590556, + "grad_norm": 0.5231650471687317, + "learning_rate": 7.78176813902343e-05, + "loss": 0.0092, + "step": 34660 + }, + { + "epoch": 5.385212799005902, + "grad_norm": 0.23860350251197815, + "learning_rate": 7.780394043423336e-05, + "loss": 0.0071, + "step": 34670 + }, + { + "epoch": 5.386766076421249, + "grad_norm": 0.18074239790439606, + "learning_rate": 7.77901964376353e-05, + "loss": 0.007, + "step": 34680 + }, + { + "epoch": 5.388319353836595, + "grad_norm": 0.2719387710094452, + "learning_rate": 7.777644940194316e-05, + "loss": 0.0062, + "step": 34690 + }, + { + "epoch": 5.389872631251942, + "grad_norm": 0.11410732567310333, + "learning_rate": 7.776269932866023e-05, + "loss": 0.0058, + "step": 34700 + }, + { + "epoch": 5.391425908667288, + "grad_norm": 0.13167640566825867, + "learning_rate": 7.774894621929026e-05, + "loss": 0.0059, + "step": 34710 + }, + { + "epoch": 5.392979186082634, + "grad_norm": 0.19908486306667328, + "learning_rate": 7.773519007533725e-05, + "loss": 0.0077, + "step": 34720 + }, + { + "epoch": 5.39453246349798, + "grad_norm": 0.14228034019470215, + "learning_rate": 7.772143089830556e-05, + "loss": 0.0109, + "step": 34730 + }, + { + "epoch": 5.396085740913327, + "grad_norm": 0.146296426653862, + "learning_rate": 7.770766868969985e-05, + "loss": 0.0079, + "step": 34740 + }, + { + "epoch": 5.3976390183286735, + "grad_norm": 0.2771199345588684, + "learning_rate": 7.769390345102518e-05, + "loss": 0.0106, + "step": 34750 + }, + { + "epoch": 5.39919229574402, + "grad_norm": 0.21245738863945007, + "learning_rate": 7.768013518378683e-05, + "loss": 0.0083, + "step": 34760 + }, + { + "epoch": 5.400745573159366, + "grad_norm": 0.34319931268692017, + "learning_rate": 7.766636388949053e-05, + "loss": 0.0067, + "step": 34770 + }, + { + "epoch": 5.402298850574713, + "grad_norm": 0.26778265833854675, + "learning_rate": 7.765258956964229e-05, + "loss": 0.0098, + "step": 34780 + }, + { + "epoch": 5.403852127990059, + "grad_norm": 0.5597488880157471, + "learning_rate": 7.76388122257484e-05, + "loss": 0.0126, + "step": 34790 + }, + { + "epoch": 5.405405405405405, + "grad_norm": 0.2935422360897064, + "learning_rate": 7.762503185931558e-05, + "loss": 0.0121, + "step": 34800 + }, + { + "epoch": 5.406958682820751, + "grad_norm": 0.1905989646911621, + "learning_rate": 7.76112484718508e-05, + "loss": 0.0111, + "step": 34810 + }, + { + "epoch": 5.4085119602360985, + "grad_norm": 0.34797704219818115, + "learning_rate": 7.75974620648614e-05, + "loss": 0.008, + "step": 34820 + }, + { + "epoch": 5.410065237651445, + "grad_norm": 0.37005579471588135, + "learning_rate": 7.758367263985503e-05, + "loss": 0.0078, + "step": 34830 + }, + { + "epoch": 5.411618515066791, + "grad_norm": 0.1591053009033203, + "learning_rate": 7.75698801983397e-05, + "loss": 0.0066, + "step": 34840 + }, + { + "epoch": 5.413171792482137, + "grad_norm": 0.24873167276382446, + "learning_rate": 7.755608474182372e-05, + "loss": 0.0069, + "step": 34850 + }, + { + "epoch": 5.414725069897484, + "grad_norm": 0.14514876902103424, + "learning_rate": 7.754228627181574e-05, + "loss": 0.0098, + "step": 34860 + }, + { + "epoch": 5.41627834731283, + "grad_norm": 0.15796658396720886, + "learning_rate": 7.752848478982476e-05, + "loss": 0.0065, + "step": 34870 + }, + { + "epoch": 5.417831624728176, + "grad_norm": 0.2233910858631134, + "learning_rate": 7.751468029736006e-05, + "loss": 0.009, + "step": 34880 + }, + { + "epoch": 5.4193849021435225, + "grad_norm": 0.16059188544750214, + "learning_rate": 7.750087279593129e-05, + "loss": 0.0067, + "step": 34890 + }, + { + "epoch": 5.42093817955887, + "grad_norm": 0.4992389678955078, + "learning_rate": 7.748706228704843e-05, + "loss": 0.0074, + "step": 34900 + }, + { + "epoch": 5.422491456974216, + "grad_norm": 0.2210060954093933, + "learning_rate": 7.747324877222176e-05, + "loss": 0.0078, + "step": 34910 + }, + { + "epoch": 5.424044734389562, + "grad_norm": 0.6979592442512512, + "learning_rate": 7.745943225296188e-05, + "loss": 0.0096, + "step": 34920 + }, + { + "epoch": 5.425598011804908, + "grad_norm": 0.6474756598472595, + "learning_rate": 7.744561273077981e-05, + "loss": 0.0094, + "step": 34930 + }, + { + "epoch": 5.427151289220255, + "grad_norm": 0.17493610084056854, + "learning_rate": 7.743179020718678e-05, + "loss": 0.0091, + "step": 34940 + }, + { + "epoch": 5.428704566635601, + "grad_norm": 0.16812968254089355, + "learning_rate": 7.741796468369443e-05, + "loss": 0.0092, + "step": 34950 + }, + { + "epoch": 5.4302578440509475, + "grad_norm": 0.4904002249240875, + "learning_rate": 7.740413616181466e-05, + "loss": 0.0092, + "step": 34960 + }, + { + "epoch": 5.431811121466294, + "grad_norm": 0.7958903312683105, + "learning_rate": 7.739030464305978e-05, + "loss": 0.0116, + "step": 34970 + }, + { + "epoch": 5.43336439888164, + "grad_norm": 0.23172925412654877, + "learning_rate": 7.737647012894235e-05, + "loss": 0.0122, + "step": 34980 + }, + { + "epoch": 5.434917676296987, + "grad_norm": 0.22203154861927032, + "learning_rate": 7.736263262097532e-05, + "loss": 0.0079, + "step": 34990 + }, + { + "epoch": 5.436470953712333, + "grad_norm": 0.26625239849090576, + "learning_rate": 7.734879212067192e-05, + "loss": 0.01, + "step": 35000 + }, + { + "epoch": 5.438024231127679, + "grad_norm": 0.28270548582077026, + "learning_rate": 7.733494862954573e-05, + "loss": 0.008, + "step": 35010 + }, + { + "epoch": 5.439577508543026, + "grad_norm": 0.37149643898010254, + "learning_rate": 7.732110214911066e-05, + "loss": 0.0084, + "step": 35020 + }, + { + "epoch": 5.441130785958372, + "grad_norm": 0.19443492591381073, + "learning_rate": 7.730725268088094e-05, + "loss": 0.0075, + "step": 35030 + }, + { + "epoch": 5.442684063373719, + "grad_norm": 0.2655029892921448, + "learning_rate": 7.729340022637112e-05, + "loss": 0.0095, + "step": 35040 + }, + { + "epoch": 5.444237340789065, + "grad_norm": 0.3472898006439209, + "learning_rate": 7.727954478709607e-05, + "loss": 0.0107, + "step": 35050 + }, + { + "epoch": 5.445790618204411, + "grad_norm": 0.30944254994392395, + "learning_rate": 7.726568636457103e-05, + "loss": 0.0094, + "step": 35060 + }, + { + "epoch": 5.447343895619758, + "grad_norm": 0.23257435858249664, + "learning_rate": 7.725182496031153e-05, + "loss": 0.0135, + "step": 35070 + }, + { + "epoch": 5.448897173035104, + "grad_norm": 0.22842183709144592, + "learning_rate": 7.723796057583341e-05, + "loss": 0.0076, + "step": 35080 + }, + { + "epoch": 5.45045045045045, + "grad_norm": 0.14797361195087433, + "learning_rate": 7.722409321265287e-05, + "loss": 0.0075, + "step": 35090 + }, + { + "epoch": 5.4520037278657965, + "grad_norm": 0.2239329218864441, + "learning_rate": 7.721022287228645e-05, + "loss": 0.0081, + "step": 35100 + }, + { + "epoch": 5.4535570052811435, + "grad_norm": 0.17532452940940857, + "learning_rate": 7.719634955625094e-05, + "loss": 0.008, + "step": 35110 + }, + { + "epoch": 5.45511028269649, + "grad_norm": 0.2374960035085678, + "learning_rate": 7.718247326606355e-05, + "loss": 0.0069, + "step": 35120 + }, + { + "epoch": 5.456663560111836, + "grad_norm": 0.18343780934810638, + "learning_rate": 7.716859400324175e-05, + "loss": 0.0103, + "step": 35130 + }, + { + "epoch": 5.458216837527182, + "grad_norm": 0.5224043726921082, + "learning_rate": 7.715471176930335e-05, + "loss": 0.0138, + "step": 35140 + }, + { + "epoch": 5.459770114942529, + "grad_norm": 0.2677820324897766, + "learning_rate": 7.714082656576651e-05, + "loss": 0.011, + "step": 35150 + }, + { + "epoch": 5.461323392357875, + "grad_norm": 0.38501250743865967, + "learning_rate": 7.712693839414968e-05, + "loss": 0.0084, + "step": 35160 + }, + { + "epoch": 5.462876669773221, + "grad_norm": 0.2707953453063965, + "learning_rate": 7.711304725597164e-05, + "loss": 0.0113, + "step": 35170 + }, + { + "epoch": 5.464429947188568, + "grad_norm": 0.16007284820079803, + "learning_rate": 7.709915315275151e-05, + "loss": 0.0105, + "step": 35180 + }, + { + "epoch": 5.465983224603915, + "grad_norm": 0.4396420121192932, + "learning_rate": 7.708525608600876e-05, + "loss": 0.0114, + "step": 35190 + }, + { + "epoch": 5.467536502019261, + "grad_norm": 0.2230774164199829, + "learning_rate": 7.707135605726311e-05, + "loss": 0.0146, + "step": 35200 + }, + { + "epoch": 5.469089779434607, + "grad_norm": 0.24589243531227112, + "learning_rate": 7.705745306803466e-05, + "loss": 0.0112, + "step": 35210 + }, + { + "epoch": 5.470643056849953, + "grad_norm": 0.2732789218425751, + "learning_rate": 7.704354711984383e-05, + "loss": 0.0074, + "step": 35220 + }, + { + "epoch": 5.4721963342653, + "grad_norm": 0.3933487832546234, + "learning_rate": 7.702963821421133e-05, + "loss": 0.0111, + "step": 35230 + }, + { + "epoch": 5.473749611680646, + "grad_norm": 0.1606975644826889, + "learning_rate": 7.701572635265826e-05, + "loss": 0.0086, + "step": 35240 + }, + { + "epoch": 5.4753028890959925, + "grad_norm": 0.20606040954589844, + "learning_rate": 7.700181153670596e-05, + "loss": 0.0077, + "step": 35250 + }, + { + "epoch": 5.476856166511339, + "grad_norm": 0.2843969464302063, + "learning_rate": 7.698789376787616e-05, + "loss": 0.009, + "step": 35260 + }, + { + "epoch": 5.478409443926686, + "grad_norm": 0.24078436195850372, + "learning_rate": 7.697397304769085e-05, + "loss": 0.01, + "step": 35270 + }, + { + "epoch": 5.479962721342032, + "grad_norm": 0.2287691980600357, + "learning_rate": 7.696004937767241e-05, + "loss": 0.0132, + "step": 35280 + }, + { + "epoch": 5.481515998757378, + "grad_norm": 0.8723113536834717, + "learning_rate": 7.694612275934352e-05, + "loss": 0.009, + "step": 35290 + }, + { + "epoch": 5.483069276172724, + "grad_norm": 0.27146583795547485, + "learning_rate": 7.693219319422714e-05, + "loss": 0.0106, + "step": 35300 + }, + { + "epoch": 5.484622553588071, + "grad_norm": 0.41068968176841736, + "learning_rate": 7.69182606838466e-05, + "loss": 0.0086, + "step": 35310 + }, + { + "epoch": 5.486175831003417, + "grad_norm": 0.3059234023094177, + "learning_rate": 7.690432522972558e-05, + "loss": 0.0088, + "step": 35320 + }, + { + "epoch": 5.487729108418764, + "grad_norm": 0.26850101351737976, + "learning_rate": 7.689038683338796e-05, + "loss": 0.0071, + "step": 35330 + }, + { + "epoch": 5.48928238583411, + "grad_norm": 0.20957981050014496, + "learning_rate": 7.687644549635808e-05, + "loss": 0.0087, + "step": 35340 + }, + { + "epoch": 5.490835663249456, + "grad_norm": 0.3214065432548523, + "learning_rate": 7.686250122016053e-05, + "loss": 0.008, + "step": 35350 + }, + { + "epoch": 5.492388940664803, + "grad_norm": 0.32882335782051086, + "learning_rate": 7.684855400632022e-05, + "loss": 0.0086, + "step": 35360 + }, + { + "epoch": 5.493942218080149, + "grad_norm": 0.2446393072605133, + "learning_rate": 7.683460385636243e-05, + "loss": 0.0107, + "step": 35370 + }, + { + "epoch": 5.495495495495495, + "grad_norm": 0.22471046447753906, + "learning_rate": 7.68206507718127e-05, + "loss": 0.0079, + "step": 35380 + }, + { + "epoch": 5.4970487729108415, + "grad_norm": 0.2275523543357849, + "learning_rate": 7.680669475419693e-05, + "loss": 0.009, + "step": 35390 + }, + { + "epoch": 5.4986020503261885, + "grad_norm": 0.16136203706264496, + "learning_rate": 7.679273580504132e-05, + "loss": 0.0067, + "step": 35400 + }, + { + "epoch": 5.500155327741535, + "grad_norm": 0.11082816123962402, + "learning_rate": 7.67787739258724e-05, + "loss": 0.008, + "step": 35410 + }, + { + "epoch": 5.501708605156881, + "grad_norm": 0.22880780696868896, + "learning_rate": 7.676480911821705e-05, + "loss": 0.0072, + "step": 35420 + }, + { + "epoch": 5.503261882572227, + "grad_norm": 0.3984946310520172, + "learning_rate": 7.675084138360238e-05, + "loss": 0.0082, + "step": 35430 + }, + { + "epoch": 5.504815159987574, + "grad_norm": 0.4033031761646271, + "learning_rate": 7.673687072355592e-05, + "loss": 0.0107, + "step": 35440 + }, + { + "epoch": 5.50636843740292, + "grad_norm": 0.2592134475708008, + "learning_rate": 7.67228971396055e-05, + "loss": 0.0111, + "step": 35450 + }, + { + "epoch": 5.507921714818266, + "grad_norm": 0.18663080036640167, + "learning_rate": 7.670892063327922e-05, + "loss": 0.0095, + "step": 35460 + }, + { + "epoch": 5.509474992233613, + "grad_norm": 0.2322949320077896, + "learning_rate": 7.669494120610552e-05, + "loss": 0.0074, + "step": 35470 + }, + { + "epoch": 5.51102826964896, + "grad_norm": 0.28032487630844116, + "learning_rate": 7.66809588596132e-05, + "loss": 0.0098, + "step": 35480 + }, + { + "epoch": 5.512581547064306, + "grad_norm": 0.15391696989536285, + "learning_rate": 7.666697359533132e-05, + "loss": 0.0082, + "step": 35490 + }, + { + "epoch": 5.514134824479652, + "grad_norm": 0.30384984612464905, + "learning_rate": 7.665298541478932e-05, + "loss": 0.0091, + "step": 35500 + }, + { + "epoch": 5.515688101894998, + "grad_norm": 0.1981964409351349, + "learning_rate": 7.66389943195169e-05, + "loss": 0.0078, + "step": 35510 + }, + { + "epoch": 5.517241379310345, + "grad_norm": 0.130614772439003, + "learning_rate": 7.66250003110441e-05, + "loss": 0.0067, + "step": 35520 + }, + { + "epoch": 5.518794656725691, + "grad_norm": 0.15902429819107056, + "learning_rate": 7.66110033909013e-05, + "loss": 0.0075, + "step": 35530 + }, + { + "epoch": 5.5203479341410375, + "grad_norm": 0.21402907371520996, + "learning_rate": 7.659700356061918e-05, + "loss": 0.0069, + "step": 35540 + }, + { + "epoch": 5.521901211556384, + "grad_norm": 0.21380165219306946, + "learning_rate": 7.658300082172875e-05, + "loss": 0.0073, + "step": 35550 + }, + { + "epoch": 5.52345448897173, + "grad_norm": 0.2539718449115753, + "learning_rate": 7.65689951757613e-05, + "loss": 0.008, + "step": 35560 + }, + { + "epoch": 5.525007766387077, + "grad_norm": 0.3414582312107086, + "learning_rate": 7.655498662424849e-05, + "loss": 0.0083, + "step": 35570 + }, + { + "epoch": 5.526561043802423, + "grad_norm": 0.32042229175567627, + "learning_rate": 7.654097516872227e-05, + "loss": 0.0085, + "step": 35580 + }, + { + "epoch": 5.528114321217769, + "grad_norm": 0.4998417794704437, + "learning_rate": 7.65269608107149e-05, + "loss": 0.0104, + "step": 35590 + }, + { + "epoch": 5.529667598633116, + "grad_norm": 0.28082138299942017, + "learning_rate": 7.6512943551759e-05, + "loss": 0.0075, + "step": 35600 + }, + { + "epoch": 5.5312208760484625, + "grad_norm": 0.16675537824630737, + "learning_rate": 7.649892339338743e-05, + "loss": 0.0083, + "step": 35610 + }, + { + "epoch": 5.532774153463809, + "grad_norm": 0.31279441714286804, + "learning_rate": 7.648490033713344e-05, + "loss": 0.0142, + "step": 35620 + }, + { + "epoch": 5.534327430879155, + "grad_norm": 0.2588692605495453, + "learning_rate": 7.647087438453058e-05, + "loss": 0.0085, + "step": 35630 + }, + { + "epoch": 5.535880708294501, + "grad_norm": 0.5030396580696106, + "learning_rate": 7.64568455371127e-05, + "loss": 0.0084, + "step": 35640 + }, + { + "epoch": 5.537433985709848, + "grad_norm": 0.4379446804523468, + "learning_rate": 7.644281379641396e-05, + "loss": 0.0115, + "step": 35650 + }, + { + "epoch": 5.538987263125194, + "grad_norm": 0.3001277446746826, + "learning_rate": 7.642877916396887e-05, + "loss": 0.0088, + "step": 35660 + }, + { + "epoch": 5.54054054054054, + "grad_norm": 0.2592935264110565, + "learning_rate": 7.641474164131221e-05, + "loss": 0.0111, + "step": 35670 + }, + { + "epoch": 5.542093817955887, + "grad_norm": 0.23163308203220367, + "learning_rate": 7.640070122997913e-05, + "loss": 0.0089, + "step": 35680 + }, + { + "epoch": 5.543647095371234, + "grad_norm": 0.18426688015460968, + "learning_rate": 7.638665793150505e-05, + "loss": 0.0082, + "step": 35690 + }, + { + "epoch": 5.54520037278658, + "grad_norm": 0.2677128314971924, + "learning_rate": 7.637261174742574e-05, + "loss": 0.0104, + "step": 35700 + }, + { + "epoch": 5.546753650201926, + "grad_norm": 0.28267332911491394, + "learning_rate": 7.635856267927725e-05, + "loss": 0.0094, + "step": 35710 + }, + { + "epoch": 5.548306927617272, + "grad_norm": 0.16255837678909302, + "learning_rate": 7.634451072859597e-05, + "loss": 0.0084, + "step": 35720 + }, + { + "epoch": 5.549860205032619, + "grad_norm": 0.41903257369995117, + "learning_rate": 7.633045589691863e-05, + "loss": 0.0099, + "step": 35730 + }, + { + "epoch": 5.551413482447965, + "grad_norm": 0.24099572002887726, + "learning_rate": 7.63163981857822e-05, + "loss": 0.0104, + "step": 35740 + }, + { + "epoch": 5.5529667598633115, + "grad_norm": 0.39385947585105896, + "learning_rate": 7.630233759672403e-05, + "loss": 0.0092, + "step": 35750 + }, + { + "epoch": 5.554520037278658, + "grad_norm": 0.20128776133060455, + "learning_rate": 7.628827413128177e-05, + "loss": 0.0098, + "step": 35760 + }, + { + "epoch": 5.556073314694005, + "grad_norm": 0.46969494223594666, + "learning_rate": 7.627420779099338e-05, + "loss": 0.0133, + "step": 35770 + }, + { + "epoch": 5.557626592109351, + "grad_norm": 0.24362744390964508, + "learning_rate": 7.626013857739711e-05, + "loss": 0.007, + "step": 35780 + }, + { + "epoch": 5.559179869524697, + "grad_norm": 0.38670238852500916, + "learning_rate": 7.624606649203158e-05, + "loss": 0.0086, + "step": 35790 + }, + { + "epoch": 5.560733146940043, + "grad_norm": 0.12210837751626968, + "learning_rate": 7.623199153643569e-05, + "loss": 0.0091, + "step": 35800 + }, + { + "epoch": 5.56228642435539, + "grad_norm": 0.2141142040491104, + "learning_rate": 7.621791371214863e-05, + "loss": 0.0089, + "step": 35810 + }, + { + "epoch": 5.563839701770736, + "grad_norm": 0.26797282695770264, + "learning_rate": 7.620383302070995e-05, + "loss": 0.0089, + "step": 35820 + }, + { + "epoch": 5.565392979186083, + "grad_norm": 0.30668532848358154, + "learning_rate": 7.61897494636595e-05, + "loss": 0.0103, + "step": 35830 + }, + { + "epoch": 5.566946256601429, + "grad_norm": 0.34426671266555786, + "learning_rate": 7.617566304253739e-05, + "loss": 0.0094, + "step": 35840 + }, + { + "epoch": 5.568499534016776, + "grad_norm": 0.3261699974536896, + "learning_rate": 7.616157375888416e-05, + "loss": 0.0088, + "step": 35850 + }, + { + "epoch": 5.570052811432122, + "grad_norm": 0.32652437686920166, + "learning_rate": 7.614748161424053e-05, + "loss": 0.0077, + "step": 35860 + }, + { + "epoch": 5.571606088847468, + "grad_norm": 0.3787689208984375, + "learning_rate": 7.613338661014763e-05, + "loss": 0.0083, + "step": 35870 + }, + { + "epoch": 5.573159366262814, + "grad_norm": 0.26056376099586487, + "learning_rate": 7.611928874814686e-05, + "loss": 0.0104, + "step": 35880 + }, + { + "epoch": 5.574712643678161, + "grad_norm": 0.36513808369636536, + "learning_rate": 7.610518802977996e-05, + "loss": 0.0085, + "step": 35890 + }, + { + "epoch": 5.5762659210935075, + "grad_norm": 0.34887996315956116, + "learning_rate": 7.609108445658893e-05, + "loss": 0.008, + "step": 35900 + }, + { + "epoch": 5.577819198508854, + "grad_norm": 0.31813594698905945, + "learning_rate": 7.607697803011612e-05, + "loss": 0.008, + "step": 35910 + }, + { + "epoch": 5.5793724759242, + "grad_norm": 0.3026588261127472, + "learning_rate": 7.606286875190421e-05, + "loss": 0.0082, + "step": 35920 + }, + { + "epoch": 5.580925753339546, + "grad_norm": 0.2792399227619171, + "learning_rate": 7.604875662349617e-05, + "loss": 0.0067, + "step": 35930 + }, + { + "epoch": 5.582479030754893, + "grad_norm": 0.22008506953716278, + "learning_rate": 7.603464164643525e-05, + "loss": 0.0103, + "step": 35940 + }, + { + "epoch": 5.584032308170239, + "grad_norm": 0.27999964356422424, + "learning_rate": 7.602052382226507e-05, + "loss": 0.0079, + "step": 35950 + }, + { + "epoch": 5.585585585585585, + "grad_norm": 0.34840238094329834, + "learning_rate": 7.600640315252954e-05, + "loss": 0.0101, + "step": 35960 + }, + { + "epoch": 5.587138863000932, + "grad_norm": 0.176471546292305, + "learning_rate": 7.599227963877284e-05, + "loss": 0.0082, + "step": 35970 + }, + { + "epoch": 5.588692140416279, + "grad_norm": 0.15439070761203766, + "learning_rate": 7.597815328253954e-05, + "loss": 0.0091, + "step": 35980 + }, + { + "epoch": 5.590245417831625, + "grad_norm": 0.2576356828212738, + "learning_rate": 7.596402408537443e-05, + "loss": 0.0086, + "step": 35990 + }, + { + "epoch": 5.591798695246971, + "grad_norm": 0.24439874291419983, + "learning_rate": 7.59498920488227e-05, + "loss": 0.0078, + "step": 36000 + }, + { + "epoch": 5.593351972662317, + "grad_norm": 0.421662300825119, + "learning_rate": 7.593575717442979e-05, + "loss": 0.0111, + "step": 36010 + }, + { + "epoch": 5.594905250077664, + "grad_norm": 0.22219504415988922, + "learning_rate": 7.592161946374147e-05, + "loss": 0.0107, + "step": 36020 + }, + { + "epoch": 5.59645852749301, + "grad_norm": 0.423248291015625, + "learning_rate": 7.590747891830381e-05, + "loss": 0.0076, + "step": 36030 + }, + { + "epoch": 5.5980118049083565, + "grad_norm": 0.2631002366542816, + "learning_rate": 7.58933355396632e-05, + "loss": 0.0104, + "step": 36040 + }, + { + "epoch": 5.599565082323703, + "grad_norm": 0.23086977005004883, + "learning_rate": 7.587918932936636e-05, + "loss": 0.0096, + "step": 36050 + }, + { + "epoch": 5.60111835973905, + "grad_norm": 0.40158578753471375, + "learning_rate": 7.586504028896028e-05, + "loss": 0.0093, + "step": 36060 + }, + { + "epoch": 5.602671637154396, + "grad_norm": 0.18272800743579865, + "learning_rate": 7.585088841999228e-05, + "loss": 0.0127, + "step": 36070 + }, + { + "epoch": 5.604224914569742, + "grad_norm": 0.5513951182365417, + "learning_rate": 7.583673372400999e-05, + "loss": 0.008, + "step": 36080 + }, + { + "epoch": 5.605778191985088, + "grad_norm": 0.3850906789302826, + "learning_rate": 7.582257620256134e-05, + "loss": 0.013, + "step": 36090 + }, + { + "epoch": 5.607331469400435, + "grad_norm": 0.20945145189762115, + "learning_rate": 7.580841585719458e-05, + "loss": 0.0113, + "step": 36100 + }, + { + "epoch": 5.608884746815781, + "grad_norm": 0.2591463625431061, + "learning_rate": 7.579425268945825e-05, + "loss": 0.0089, + "step": 36110 + }, + { + "epoch": 5.610438024231128, + "grad_norm": 0.205123171210289, + "learning_rate": 7.578008670090127e-05, + "loss": 0.0086, + "step": 36120 + }, + { + "epoch": 5.611991301646474, + "grad_norm": 0.37967449426651, + "learning_rate": 7.576591789307272e-05, + "loss": 0.0086, + "step": 36130 + }, + { + "epoch": 5.613544579061821, + "grad_norm": 0.2071877121925354, + "learning_rate": 7.575174626752216e-05, + "loss": 0.0097, + "step": 36140 + }, + { + "epoch": 5.615097856477167, + "grad_norm": 0.2309500128030777, + "learning_rate": 7.573757182579934e-05, + "loss": 0.0099, + "step": 36150 + }, + { + "epoch": 5.616651133892513, + "grad_norm": 0.3001890182495117, + "learning_rate": 7.572339456945435e-05, + "loss": 0.009, + "step": 36160 + }, + { + "epoch": 5.618204411307859, + "grad_norm": 0.2623978853225708, + "learning_rate": 7.570921450003762e-05, + "loss": 0.0068, + "step": 36170 + }, + { + "epoch": 5.619757688723206, + "grad_norm": 0.2768475413322449, + "learning_rate": 7.569503161909983e-05, + "loss": 0.0095, + "step": 36180 + }, + { + "epoch": 5.6213109661385525, + "grad_norm": 0.23106016218662262, + "learning_rate": 7.568084592819202e-05, + "loss": 0.0069, + "step": 36190 + }, + { + "epoch": 5.622864243553899, + "grad_norm": 0.2240242213010788, + "learning_rate": 7.566665742886551e-05, + "loss": 0.0078, + "step": 36200 + }, + { + "epoch": 5.624417520969245, + "grad_norm": 0.31341779232025146, + "learning_rate": 7.565246612267194e-05, + "loss": 0.0076, + "step": 36210 + }, + { + "epoch": 5.625970798384591, + "grad_norm": 0.1797037571668625, + "learning_rate": 7.563827201116325e-05, + "loss": 0.0049, + "step": 36220 + }, + { + "epoch": 5.627524075799938, + "grad_norm": 0.20742611587047577, + "learning_rate": 7.562407509589164e-05, + "loss": 0.0079, + "step": 36230 + }, + { + "epoch": 5.629077353215284, + "grad_norm": 0.1824333220720291, + "learning_rate": 7.560987537840973e-05, + "loss": 0.0088, + "step": 36240 + }, + { + "epoch": 5.63063063063063, + "grad_norm": 0.17986854910850525, + "learning_rate": 7.559567286027036e-05, + "loss": 0.0083, + "step": 36250 + }, + { + "epoch": 5.6321839080459775, + "grad_norm": 0.2945517301559448, + "learning_rate": 7.558146754302668e-05, + "loss": 0.0086, + "step": 36260 + }, + { + "epoch": 5.633737185461324, + "grad_norm": 0.29812681674957275, + "learning_rate": 7.556725942823217e-05, + "loss": 0.0094, + "step": 36270 + }, + { + "epoch": 5.63529046287667, + "grad_norm": 1.1444072723388672, + "learning_rate": 7.555304851744061e-05, + "loss": 0.0122, + "step": 36280 + }, + { + "epoch": 5.636843740292016, + "grad_norm": 0.2245127409696579, + "learning_rate": 7.553883481220608e-05, + "loss": 0.0058, + "step": 36290 + }, + { + "epoch": 5.638397017707362, + "grad_norm": 0.4592326879501343, + "learning_rate": 7.552461831408298e-05, + "loss": 0.0077, + "step": 36300 + }, + { + "epoch": 5.639950295122709, + "grad_norm": 0.32713180780410767, + "learning_rate": 7.551039902462599e-05, + "loss": 0.0083, + "step": 36310 + }, + { + "epoch": 5.641503572538055, + "grad_norm": 0.25305622816085815, + "learning_rate": 7.549617694539011e-05, + "loss": 0.0092, + "step": 36320 + }, + { + "epoch": 5.6430568499534015, + "grad_norm": 0.19172511994838715, + "learning_rate": 7.548195207793066e-05, + "loss": 0.0106, + "step": 36330 + }, + { + "epoch": 5.644610127368748, + "grad_norm": 0.2700869143009186, + "learning_rate": 7.546772442380323e-05, + "loss": 0.0112, + "step": 36340 + }, + { + "epoch": 5.646163404784095, + "grad_norm": 0.33396923542022705, + "learning_rate": 7.545349398456376e-05, + "loss": 0.0109, + "step": 36350 + }, + { + "epoch": 5.647716682199441, + "grad_norm": 0.21258936822414398, + "learning_rate": 7.543926076176845e-05, + "loss": 0.0101, + "step": 36360 + }, + { + "epoch": 5.649269959614787, + "grad_norm": 0.21759475767612457, + "learning_rate": 7.542502475697385e-05, + "loss": 0.0091, + "step": 36370 + }, + { + "epoch": 5.650823237030133, + "grad_norm": 0.25987717509269714, + "learning_rate": 7.541078597173675e-05, + "loss": 0.0094, + "step": 36380 + }, + { + "epoch": 5.65237651444548, + "grad_norm": 0.36584386229515076, + "learning_rate": 7.539654440761431e-05, + "loss": 0.0109, + "step": 36390 + }, + { + "epoch": 5.6539297918608264, + "grad_norm": 0.3174910545349121, + "learning_rate": 7.538230006616395e-05, + "loss": 0.008, + "step": 36400 + }, + { + "epoch": 5.655483069276173, + "grad_norm": 0.15846878290176392, + "learning_rate": 7.536805294894342e-05, + "loss": 0.0088, + "step": 36410 + }, + { + "epoch": 5.657036346691519, + "grad_norm": 0.14795513451099396, + "learning_rate": 7.535380305751076e-05, + "loss": 0.0107, + "step": 36420 + }, + { + "epoch": 5.658589624106866, + "grad_norm": 0.30813276767730713, + "learning_rate": 7.533955039342431e-05, + "loss": 0.0105, + "step": 36430 + }, + { + "epoch": 5.660142901522212, + "grad_norm": 0.6756739020347595, + "learning_rate": 7.532529495824274e-05, + "loss": 0.0099, + "step": 36440 + }, + { + "epoch": 5.661696178937558, + "grad_norm": 0.34494858980178833, + "learning_rate": 7.531103675352497e-05, + "loss": 0.0098, + "step": 36450 + }, + { + "epoch": 5.663249456352904, + "grad_norm": 0.21420587599277496, + "learning_rate": 7.529677578083029e-05, + "loss": 0.0086, + "step": 36460 + }, + { + "epoch": 5.664802733768251, + "grad_norm": 0.18903808295726776, + "learning_rate": 7.528251204171823e-05, + "loss": 0.0092, + "step": 36470 + }, + { + "epoch": 5.6663560111835976, + "grad_norm": 0.26773881912231445, + "learning_rate": 7.526824553774867e-05, + "loss": 0.0069, + "step": 36480 + }, + { + "epoch": 5.667909288598944, + "grad_norm": 0.26224958896636963, + "learning_rate": 7.525397627048178e-05, + "loss": 0.0078, + "step": 36490 + }, + { + "epoch": 5.66946256601429, + "grad_norm": 0.4694584608078003, + "learning_rate": 7.523970424147802e-05, + "loss": 0.0083, + "step": 36500 + }, + { + "epoch": 5.671015843429636, + "grad_norm": 0.3072797656059265, + "learning_rate": 7.522542945229813e-05, + "loss": 0.0077, + "step": 36510 + }, + { + "epoch": 5.672569120844983, + "grad_norm": 0.29702430963516235, + "learning_rate": 7.521115190450321e-05, + "loss": 0.0117, + "step": 36520 + }, + { + "epoch": 5.674122398260329, + "grad_norm": 0.15320941805839539, + "learning_rate": 7.519687159965462e-05, + "loss": 0.0075, + "step": 36530 + }, + { + "epoch": 5.675675675675675, + "grad_norm": 0.4312792122364044, + "learning_rate": 7.518258853931403e-05, + "loss": 0.0071, + "step": 36540 + }, + { + "epoch": 5.6772289530910225, + "grad_norm": 0.13176652789115906, + "learning_rate": 7.516830272504342e-05, + "loss": 0.01, + "step": 36550 + }, + { + "epoch": 5.678782230506369, + "grad_norm": 0.24509036540985107, + "learning_rate": 7.515401415840509e-05, + "loss": 0.0131, + "step": 36560 + }, + { + "epoch": 5.680335507921715, + "grad_norm": 0.2973106801509857, + "learning_rate": 7.513972284096155e-05, + "loss": 0.0095, + "step": 36570 + }, + { + "epoch": 5.681888785337061, + "grad_norm": 0.26414406299591064, + "learning_rate": 7.512542877427574e-05, + "loss": 0.0083, + "step": 36580 + }, + { + "epoch": 5.683442062752407, + "grad_norm": 0.3764958083629608, + "learning_rate": 7.51111319599108e-05, + "loss": 0.0108, + "step": 36590 + }, + { + "epoch": 5.684995340167754, + "grad_norm": 0.2086586356163025, + "learning_rate": 7.509683239943024e-05, + "loss": 0.0104, + "step": 36600 + }, + { + "epoch": 5.6865486175831, + "grad_norm": 0.2955630421638489, + "learning_rate": 7.50825300943978e-05, + "loss": 0.0107, + "step": 36610 + }, + { + "epoch": 5.6881018949984465, + "grad_norm": 0.3143620193004608, + "learning_rate": 7.50682250463776e-05, + "loss": 0.0097, + "step": 36620 + }, + { + "epoch": 5.689655172413794, + "grad_norm": 0.25253140926361084, + "learning_rate": 7.5053917256934e-05, + "loss": 0.0097, + "step": 36630 + }, + { + "epoch": 5.69120844982914, + "grad_norm": 0.21965228021144867, + "learning_rate": 7.503960672763166e-05, + "loss": 0.0094, + "step": 36640 + }, + { + "epoch": 5.692761727244486, + "grad_norm": 0.3373786509037018, + "learning_rate": 7.502529346003559e-05, + "loss": 0.0072, + "step": 36650 + }, + { + "epoch": 5.694315004659832, + "grad_norm": 0.3286905288696289, + "learning_rate": 7.501097745571106e-05, + "loss": 0.011, + "step": 36660 + }, + { + "epoch": 5.695868282075178, + "grad_norm": 0.23490869998931885, + "learning_rate": 7.499665871622361e-05, + "loss": 0.0094, + "step": 36670 + }, + { + "epoch": 5.697421559490525, + "grad_norm": 0.5730550289154053, + "learning_rate": 7.498233724313917e-05, + "loss": 0.0099, + "step": 36680 + }, + { + "epoch": 5.6989748369058715, + "grad_norm": 0.2198418527841568, + "learning_rate": 7.496801303802387e-05, + "loss": 0.0064, + "step": 36690 + }, + { + "epoch": 5.700528114321218, + "grad_norm": 0.14456093311309814, + "learning_rate": 7.495368610244423e-05, + "loss": 0.0062, + "step": 36700 + }, + { + "epoch": 5.702081391736564, + "grad_norm": 0.20709645748138428, + "learning_rate": 7.493935643796697e-05, + "loss": 0.008, + "step": 36710 + }, + { + "epoch": 5.703634669151911, + "grad_norm": 0.19869013130664825, + "learning_rate": 7.49250240461592e-05, + "loss": 0.009, + "step": 36720 + }, + { + "epoch": 5.705187946567257, + "grad_norm": 0.2773151099681854, + "learning_rate": 7.491068892858826e-05, + "loss": 0.0083, + "step": 36730 + }, + { + "epoch": 5.706741223982603, + "grad_norm": 0.18872524797916412, + "learning_rate": 7.489635108682184e-05, + "loss": 0.0079, + "step": 36740 + }, + { + "epoch": 5.708294501397949, + "grad_norm": 0.2863087058067322, + "learning_rate": 7.48820105224279e-05, + "loss": 0.0111, + "step": 36750 + }, + { + "epoch": 5.709847778813296, + "grad_norm": 0.3334999680519104, + "learning_rate": 7.48676672369747e-05, + "loss": 0.0106, + "step": 36760 + }, + { + "epoch": 5.711401056228643, + "grad_norm": 0.1106618121266365, + "learning_rate": 7.485332123203079e-05, + "loss": 0.0077, + "step": 36770 + }, + { + "epoch": 5.712954333643989, + "grad_norm": 0.22776401042938232, + "learning_rate": 7.483897250916504e-05, + "loss": 0.0098, + "step": 36780 + }, + { + "epoch": 5.714507611059335, + "grad_norm": 0.27355149388313293, + "learning_rate": 7.48246210699466e-05, + "loss": 0.0079, + "step": 36790 + }, + { + "epoch": 5.716060888474681, + "grad_norm": 0.30714482069015503, + "learning_rate": 7.481026691594492e-05, + "loss": 0.0095, + "step": 36800 + }, + { + "epoch": 5.717614165890028, + "grad_norm": 0.25470998883247375, + "learning_rate": 7.479591004872975e-05, + "loss": 0.0103, + "step": 36810 + }, + { + "epoch": 5.719167443305374, + "grad_norm": 0.26299890875816345, + "learning_rate": 7.478155046987114e-05, + "loss": 0.0111, + "step": 36820 + }, + { + "epoch": 5.7207207207207205, + "grad_norm": 0.16077369451522827, + "learning_rate": 7.476718818093944e-05, + "loss": 0.0075, + "step": 36830 + }, + { + "epoch": 5.7222739981360675, + "grad_norm": 0.32122182846069336, + "learning_rate": 7.475282318350527e-05, + "loss": 0.0109, + "step": 36840 + }, + { + "epoch": 5.723827275551414, + "grad_norm": 0.2708611786365509, + "learning_rate": 7.473845547913957e-05, + "loss": 0.0107, + "step": 36850 + }, + { + "epoch": 5.72538055296676, + "grad_norm": 0.5952991843223572, + "learning_rate": 7.472408506941357e-05, + "loss": 0.0101, + "step": 36860 + }, + { + "epoch": 5.726933830382106, + "grad_norm": 0.16238945722579956, + "learning_rate": 7.470971195589881e-05, + "loss": 0.0081, + "step": 36870 + }, + { + "epoch": 5.728487107797452, + "grad_norm": 0.4328939914703369, + "learning_rate": 7.46953361401671e-05, + "loss": 0.0089, + "step": 36880 + }, + { + "epoch": 5.730040385212799, + "grad_norm": 0.22260984778404236, + "learning_rate": 7.468095762379055e-05, + "loss": 0.0102, + "step": 36890 + }, + { + "epoch": 5.731593662628145, + "grad_norm": 0.19085003435611725, + "learning_rate": 7.466657640834158e-05, + "loss": 0.0087, + "step": 36900 + }, + { + "epoch": 5.733146940043492, + "grad_norm": 0.24263596534729004, + "learning_rate": 7.46521924953929e-05, + "loss": 0.0107, + "step": 36910 + }, + { + "epoch": 5.734700217458839, + "grad_norm": 0.2763688862323761, + "learning_rate": 7.463780588651752e-05, + "loss": 0.0104, + "step": 36920 + }, + { + "epoch": 5.736253494874185, + "grad_norm": 0.3253340423107147, + "learning_rate": 7.462341658328873e-05, + "loss": 0.0099, + "step": 36930 + }, + { + "epoch": 5.737806772289531, + "grad_norm": 0.25486475229263306, + "learning_rate": 7.460902458728012e-05, + "loss": 0.0071, + "step": 36940 + }, + { + "epoch": 5.739360049704877, + "grad_norm": 0.17271627485752106, + "learning_rate": 7.459462990006558e-05, + "loss": 0.01, + "step": 36950 + }, + { + "epoch": 5.740913327120223, + "grad_norm": 0.25206637382507324, + "learning_rate": 7.458023252321929e-05, + "loss": 0.0074, + "step": 36960 + }, + { + "epoch": 5.74246660453557, + "grad_norm": 0.23703807592391968, + "learning_rate": 7.456583245831573e-05, + "loss": 0.0078, + "step": 36970 + }, + { + "epoch": 5.7440198819509165, + "grad_norm": 0.28412553668022156, + "learning_rate": 7.455142970692968e-05, + "loss": 0.0077, + "step": 36980 + }, + { + "epoch": 5.745573159366263, + "grad_norm": 0.26491376757621765, + "learning_rate": 7.453702427063615e-05, + "loss": 0.0066, + "step": 36990 + }, + { + "epoch": 5.747126436781609, + "grad_norm": 0.14721372723579407, + "learning_rate": 7.452261615101057e-05, + "loss": 0.0083, + "step": 37000 + }, + { + "epoch": 5.748679714196956, + "grad_norm": 0.3797323405742645, + "learning_rate": 7.450820534962856e-05, + "loss": 0.0075, + "step": 37010 + }, + { + "epoch": 5.750232991612302, + "grad_norm": 0.24127665162086487, + "learning_rate": 7.449379186806604e-05, + "loss": 0.0099, + "step": 37020 + }, + { + "epoch": 5.751786269027648, + "grad_norm": 0.19041278958320618, + "learning_rate": 7.44793757078993e-05, + "loss": 0.0107, + "step": 37030 + }, + { + "epoch": 5.753339546442994, + "grad_norm": 0.37404850125312805, + "learning_rate": 7.44649568707048e-05, + "loss": 0.0099, + "step": 37040 + }, + { + "epoch": 5.754892823858341, + "grad_norm": 0.321361780166626, + "learning_rate": 7.445053535805942e-05, + "loss": 0.0071, + "step": 37050 + }, + { + "epoch": 5.756446101273688, + "grad_norm": 0.3680557906627655, + "learning_rate": 7.443611117154024e-05, + "loss": 0.0097, + "step": 37060 + }, + { + "epoch": 5.757999378689034, + "grad_norm": 0.1895262748003006, + "learning_rate": 7.442168431272469e-05, + "loss": 0.0089, + "step": 37070 + }, + { + "epoch": 5.75955265610438, + "grad_norm": 0.29353660345077515, + "learning_rate": 7.440725478319044e-05, + "loss": 0.0108, + "step": 37080 + }, + { + "epoch": 5.761105933519727, + "grad_norm": 0.2924607992172241, + "learning_rate": 7.439282258451553e-05, + "loss": 0.0115, + "step": 37090 + }, + { + "epoch": 5.762659210935073, + "grad_norm": 0.2726955711841583, + "learning_rate": 7.43783877182782e-05, + "loss": 0.009, + "step": 37100 + }, + { + "epoch": 5.764212488350419, + "grad_norm": 0.21155120432376862, + "learning_rate": 7.436395018605707e-05, + "loss": 0.0079, + "step": 37110 + }, + { + "epoch": 5.7657657657657655, + "grad_norm": 0.3696124255657196, + "learning_rate": 7.434950998943094e-05, + "loss": 0.0116, + "step": 37120 + }, + { + "epoch": 5.7673190431811125, + "grad_norm": 0.1513832062482834, + "learning_rate": 7.433506712997902e-05, + "loss": 0.0088, + "step": 37130 + }, + { + "epoch": 5.768872320596459, + "grad_norm": 0.250087708234787, + "learning_rate": 7.432062160928077e-05, + "loss": 0.0102, + "step": 37140 + }, + { + "epoch": 5.770425598011805, + "grad_norm": 0.3085533082485199, + "learning_rate": 7.430617342891588e-05, + "loss": 0.012, + "step": 37150 + }, + { + "epoch": 5.771978875427151, + "grad_norm": 0.2209763079881668, + "learning_rate": 7.429172259046442e-05, + "loss": 0.0087, + "step": 37160 + }, + { + "epoch": 5.773532152842497, + "grad_norm": 0.3953896164894104, + "learning_rate": 7.42772690955067e-05, + "loss": 0.0111, + "step": 37170 + }, + { + "epoch": 5.775085430257844, + "grad_norm": 0.3167971074581146, + "learning_rate": 7.426281294562332e-05, + "loss": 0.0117, + "step": 37180 + }, + { + "epoch": 5.77663870767319, + "grad_norm": 0.20916801691055298, + "learning_rate": 7.424835414239519e-05, + "loss": 0.0124, + "step": 37190 + }, + { + "epoch": 5.778191985088537, + "grad_norm": 0.22619812190532684, + "learning_rate": 7.423389268740352e-05, + "loss": 0.0074, + "step": 37200 + }, + { + "epoch": 5.779745262503884, + "grad_norm": 0.4035224914550781, + "learning_rate": 7.42194285822298e-05, + "loss": 0.0106, + "step": 37210 + }, + { + "epoch": 5.78129853991923, + "grad_norm": 0.17512227594852448, + "learning_rate": 7.420496182845576e-05, + "loss": 0.0058, + "step": 37220 + }, + { + "epoch": 5.782851817334576, + "grad_norm": 0.2771585285663605, + "learning_rate": 7.419049242766348e-05, + "loss": 0.0103, + "step": 37230 + }, + { + "epoch": 5.784405094749922, + "grad_norm": 0.1831742227077484, + "learning_rate": 7.417602038143533e-05, + "loss": 0.0101, + "step": 37240 + }, + { + "epoch": 5.785958372165268, + "grad_norm": 0.23132272064685822, + "learning_rate": 7.416154569135393e-05, + "loss": 0.0076, + "step": 37250 + }, + { + "epoch": 5.787511649580615, + "grad_norm": 0.2694523334503174, + "learning_rate": 7.414706835900222e-05, + "loss": 0.0069, + "step": 37260 + }, + { + "epoch": 5.7890649269959615, + "grad_norm": 0.20402033627033234, + "learning_rate": 7.413258838596345e-05, + "loss": 0.0082, + "step": 37270 + }, + { + "epoch": 5.790618204411308, + "grad_norm": 0.30495432019233704, + "learning_rate": 7.411810577382106e-05, + "loss": 0.0092, + "step": 37280 + }, + { + "epoch": 5.792171481826654, + "grad_norm": 0.18371738493442535, + "learning_rate": 7.410362052415889e-05, + "loss": 0.0058, + "step": 37290 + }, + { + "epoch": 5.793724759242001, + "grad_norm": 0.16315408051013947, + "learning_rate": 7.408913263856102e-05, + "loss": 0.0051, + "step": 37300 + }, + { + "epoch": 5.795278036657347, + "grad_norm": 0.28379425406455994, + "learning_rate": 7.407464211861183e-05, + "loss": 0.0082, + "step": 37310 + }, + { + "epoch": 5.796831314072693, + "grad_norm": 0.19847354292869568, + "learning_rate": 7.406014896589597e-05, + "loss": 0.0073, + "step": 37320 + }, + { + "epoch": 5.798384591488039, + "grad_norm": 0.2662714421749115, + "learning_rate": 7.404565318199838e-05, + "loss": 0.0091, + "step": 37330 + }, + { + "epoch": 5.7999378689033865, + "grad_norm": 0.21336495876312256, + "learning_rate": 7.403115476850432e-05, + "loss": 0.0076, + "step": 37340 + }, + { + "epoch": 5.801491146318733, + "grad_norm": 0.27351099252700806, + "learning_rate": 7.401665372699932e-05, + "loss": 0.0094, + "step": 37350 + }, + { + "epoch": 5.803044423734079, + "grad_norm": 0.20195871591567993, + "learning_rate": 7.400215005906916e-05, + "loss": 0.0085, + "step": 37360 + }, + { + "epoch": 5.804597701149425, + "grad_norm": 0.21055704355239868, + "learning_rate": 7.398764376629996e-05, + "loss": 0.0094, + "step": 37370 + }, + { + "epoch": 5.806150978564772, + "grad_norm": 0.33581238985061646, + "learning_rate": 7.39731348502781e-05, + "loss": 0.0081, + "step": 37380 + }, + { + "epoch": 5.807704255980118, + "grad_norm": 0.5155587792396545, + "learning_rate": 7.395862331259028e-05, + "loss": 0.0067, + "step": 37390 + }, + { + "epoch": 5.809257533395464, + "grad_norm": 0.2506725788116455, + "learning_rate": 7.394410915482341e-05, + "loss": 0.01, + "step": 37400 + }, + { + "epoch": 5.8108108108108105, + "grad_norm": 0.15880140662193298, + "learning_rate": 7.392959237856476e-05, + "loss": 0.0066, + "step": 37410 + }, + { + "epoch": 5.812364088226158, + "grad_norm": 0.2524281144142151, + "learning_rate": 7.391507298540187e-05, + "loss": 0.0083, + "step": 37420 + }, + { + "epoch": 5.813917365641504, + "grad_norm": 0.32336893677711487, + "learning_rate": 7.390055097692257e-05, + "loss": 0.0069, + "step": 37430 + }, + { + "epoch": 5.81547064305685, + "grad_norm": 0.4313381314277649, + "learning_rate": 7.388602635471494e-05, + "loss": 0.0095, + "step": 37440 + }, + { + "epoch": 5.817023920472196, + "grad_norm": 0.21024519205093384, + "learning_rate": 7.387149912036738e-05, + "loss": 0.0114, + "step": 37450 + }, + { + "epoch": 5.818577197887542, + "grad_norm": 0.18112744390964508, + "learning_rate": 7.385696927546854e-05, + "loss": 0.0069, + "step": 37460 + }, + { + "epoch": 5.820130475302889, + "grad_norm": 0.28076377511024475, + "learning_rate": 7.384243682160743e-05, + "loss": 0.0085, + "step": 37470 + }, + { + "epoch": 5.8216837527182355, + "grad_norm": 0.2172463834285736, + "learning_rate": 7.382790176037329e-05, + "loss": 0.0124, + "step": 37480 + }, + { + "epoch": 5.823237030133582, + "grad_norm": 0.32553502917289734, + "learning_rate": 7.381336409335562e-05, + "loss": 0.0076, + "step": 37490 + }, + { + "epoch": 5.824790307548929, + "grad_norm": 0.14253473281860352, + "learning_rate": 7.379882382214426e-05, + "loss": 0.0063, + "step": 37500 + }, + { + "epoch": 5.826343584964275, + "grad_norm": 0.22652338445186615, + "learning_rate": 7.378428094832931e-05, + "loss": 0.0066, + "step": 37510 + }, + { + "epoch": 5.827896862379621, + "grad_norm": 0.4453798234462738, + "learning_rate": 7.376973547350114e-05, + "loss": 0.0077, + "step": 37520 + }, + { + "epoch": 5.829450139794967, + "grad_norm": 0.24698950350284576, + "learning_rate": 7.375518739925046e-05, + "loss": 0.0064, + "step": 37530 + }, + { + "epoch": 5.831003417210313, + "grad_norm": 0.20238086581230164, + "learning_rate": 7.374063672716818e-05, + "loss": 0.008, + "step": 37540 + }, + { + "epoch": 5.83255669462566, + "grad_norm": 0.33805352449417114, + "learning_rate": 7.372608345884558e-05, + "loss": 0.0115, + "step": 37550 + }, + { + "epoch": 5.834109972041007, + "grad_norm": 0.28324875235557556, + "learning_rate": 7.371152759587415e-05, + "loss": 0.0087, + "step": 37560 + }, + { + "epoch": 5.835663249456353, + "grad_norm": 0.2283424586057663, + "learning_rate": 7.369696913984572e-05, + "loss": 0.0135, + "step": 37570 + }, + { + "epoch": 5.8372165268717, + "grad_norm": 0.3245650827884674, + "learning_rate": 7.368240809235237e-05, + "loss": 0.0079, + "step": 37580 + }, + { + "epoch": 5.838769804287046, + "grad_norm": 0.3912339508533478, + "learning_rate": 7.366784445498649e-05, + "loss": 0.009, + "step": 37590 + }, + { + "epoch": 5.840323081702392, + "grad_norm": 0.2572211027145386, + "learning_rate": 7.36532782293407e-05, + "loss": 0.0087, + "step": 37600 + }, + { + "epoch": 5.841876359117738, + "grad_norm": 0.41056758165359497, + "learning_rate": 7.363870941700797e-05, + "loss": 0.0118, + "step": 37610 + }, + { + "epoch": 5.8434296365330844, + "grad_norm": 0.1784379929304123, + "learning_rate": 7.362413801958152e-05, + "loss": 0.0097, + "step": 37620 + }, + { + "epoch": 5.8449829139484315, + "grad_norm": 0.23298940062522888, + "learning_rate": 7.360956403865486e-05, + "loss": 0.0077, + "step": 37630 + }, + { + "epoch": 5.846536191363778, + "grad_norm": 0.26371654868125916, + "learning_rate": 7.359498747582177e-05, + "loss": 0.0092, + "step": 37640 + }, + { + "epoch": 5.848089468779124, + "grad_norm": 0.2843253016471863, + "learning_rate": 7.358040833267634e-05, + "loss": 0.0084, + "step": 37650 + }, + { + "epoch": 5.84964274619447, + "grad_norm": 0.2819139063358307, + "learning_rate": 7.356582661081288e-05, + "loss": 0.013, + "step": 37660 + }, + { + "epoch": 5.851196023609817, + "grad_norm": 0.28890618681907654, + "learning_rate": 7.355124231182607e-05, + "loss": 0.009, + "step": 37670 + }, + { + "epoch": 5.852749301025163, + "grad_norm": 0.26072853803634644, + "learning_rate": 7.353665543731083e-05, + "loss": 0.0103, + "step": 37680 + }, + { + "epoch": 5.854302578440509, + "grad_norm": 0.19981792569160461, + "learning_rate": 7.352206598886229e-05, + "loss": 0.0109, + "step": 37690 + }, + { + "epoch": 5.8558558558558556, + "grad_norm": 0.2594221830368042, + "learning_rate": 7.350747396807601e-05, + "loss": 0.0104, + "step": 37700 + }, + { + "epoch": 5.857409133271203, + "grad_norm": 0.263247013092041, + "learning_rate": 7.349287937654772e-05, + "loss": 0.0094, + "step": 37710 + }, + { + "epoch": 5.858962410686549, + "grad_norm": 0.2497735172510147, + "learning_rate": 7.347828221587345e-05, + "loss": 0.0082, + "step": 37720 + }, + { + "epoch": 5.860515688101895, + "grad_norm": 1.0980130434036255, + "learning_rate": 7.346368248764955e-05, + "loss": 0.0083, + "step": 37730 + }, + { + "epoch": 5.862068965517241, + "grad_norm": 0.24716530740261078, + "learning_rate": 7.34490801934726e-05, + "loss": 0.008, + "step": 37740 + }, + { + "epoch": 5.863622242932587, + "grad_norm": 0.2179318070411682, + "learning_rate": 7.343447533493947e-05, + "loss": 0.0086, + "step": 37750 + }, + { + "epoch": 5.865175520347934, + "grad_norm": 0.18131127953529358, + "learning_rate": 7.341986791364739e-05, + "loss": 0.008, + "step": 37760 + }, + { + "epoch": 5.8667287977632805, + "grad_norm": 0.13518106937408447, + "learning_rate": 7.340525793119374e-05, + "loss": 0.0106, + "step": 37770 + }, + { + "epoch": 5.868282075178627, + "grad_norm": 0.2506578266620636, + "learning_rate": 7.33906453891763e-05, + "loss": 0.0069, + "step": 37780 + }, + { + "epoch": 5.869835352593974, + "grad_norm": 0.2171999216079712, + "learning_rate": 7.337603028919303e-05, + "loss": 0.0069, + "step": 37790 + }, + { + "epoch": 5.87138863000932, + "grad_norm": 0.2639518678188324, + "learning_rate": 7.336141263284225e-05, + "loss": 0.0067, + "step": 37800 + }, + { + "epoch": 5.872941907424666, + "grad_norm": 0.2016652524471283, + "learning_rate": 7.334679242172249e-05, + "loss": 0.01, + "step": 37810 + }, + { + "epoch": 5.874495184840012, + "grad_norm": 0.19300618767738342, + "learning_rate": 7.33321696574326e-05, + "loss": 0.0076, + "step": 37820 + }, + { + "epoch": 5.876048462255358, + "grad_norm": 0.2506379187107086, + "learning_rate": 7.331754434157174e-05, + "loss": 0.0076, + "step": 37830 + }, + { + "epoch": 5.877601739670705, + "grad_norm": 0.199211984872818, + "learning_rate": 7.330291647573928e-05, + "loss": 0.0084, + "step": 37840 + }, + { + "epoch": 5.879155017086052, + "grad_norm": 0.23643209040164948, + "learning_rate": 7.32882860615349e-05, + "loss": 0.0069, + "step": 37850 + }, + { + "epoch": 5.880708294501398, + "grad_norm": 0.318431556224823, + "learning_rate": 7.327365310055858e-05, + "loss": 0.0077, + "step": 37860 + }, + { + "epoch": 5.882261571916745, + "grad_norm": 0.1241612508893013, + "learning_rate": 7.325901759441055e-05, + "loss": 0.0075, + "step": 37870 + }, + { + "epoch": 5.883814849332091, + "grad_norm": 0.23242343962192535, + "learning_rate": 7.324437954469132e-05, + "loss": 0.0057, + "step": 37880 + }, + { + "epoch": 5.885368126747437, + "grad_norm": 0.21562130749225616, + "learning_rate": 7.32297389530017e-05, + "loss": 0.0107, + "step": 37890 + }, + { + "epoch": 5.886921404162783, + "grad_norm": 0.30165258049964905, + "learning_rate": 7.321509582094274e-05, + "loss": 0.0099, + "step": 37900 + }, + { + "epoch": 5.8884746815781295, + "grad_norm": 0.22422680258750916, + "learning_rate": 7.320045015011582e-05, + "loss": 0.008, + "step": 37910 + }, + { + "epoch": 5.8900279589934765, + "grad_norm": 0.15108907222747803, + "learning_rate": 7.318580194212253e-05, + "loss": 0.0083, + "step": 37920 + }, + { + "epoch": 5.891581236408823, + "grad_norm": 0.1805582344532013, + "learning_rate": 7.317115119856483e-05, + "loss": 0.0097, + "step": 37930 + }, + { + "epoch": 5.893134513824169, + "grad_norm": 0.23631064593791962, + "learning_rate": 7.315649792104485e-05, + "loss": 0.0084, + "step": 37940 + }, + { + "epoch": 5.894687791239515, + "grad_norm": 0.34445899724960327, + "learning_rate": 7.314184211116507e-05, + "loss": 0.0076, + "step": 37950 + }, + { + "epoch": 5.896241068654862, + "grad_norm": 0.10687622427940369, + "learning_rate": 7.312718377052823e-05, + "loss": 0.0112, + "step": 37960 + }, + { + "epoch": 5.897794346070208, + "grad_norm": 0.48391517996788025, + "learning_rate": 7.311252290073737e-05, + "loss": 0.0087, + "step": 37970 + }, + { + "epoch": 5.899347623485554, + "grad_norm": 0.3476260304450989, + "learning_rate": 7.309785950339572e-05, + "loss": 0.0112, + "step": 37980 + }, + { + "epoch": 5.900900900900901, + "grad_norm": 0.23324942588806152, + "learning_rate": 7.308319358010692e-05, + "loss": 0.0091, + "step": 37990 + }, + { + "epoch": 5.902454178316248, + "grad_norm": 0.46865224838256836, + "learning_rate": 7.306852513247474e-05, + "loss": 0.0086, + "step": 38000 + }, + { + "epoch": 5.904007455731594, + "grad_norm": 0.2618396282196045, + "learning_rate": 7.305385416210336e-05, + "loss": 0.0113, + "step": 38010 + }, + { + "epoch": 5.90556073314694, + "grad_norm": 0.3241732120513916, + "learning_rate": 7.303918067059711e-05, + "loss": 0.012, + "step": 38020 + }, + { + "epoch": 5.907114010562286, + "grad_norm": 0.27464500069618225, + "learning_rate": 7.302450465956075e-05, + "loss": 0.0083, + "step": 38030 + }, + { + "epoch": 5.908667287977633, + "grad_norm": 0.21682770550251007, + "learning_rate": 7.300982613059914e-05, + "loss": 0.0074, + "step": 38040 + }, + { + "epoch": 5.910220565392979, + "grad_norm": 0.36691945791244507, + "learning_rate": 7.299514508531757e-05, + "loss": 0.0071, + "step": 38050 + }, + { + "epoch": 5.9117738428083255, + "grad_norm": 0.3129473924636841, + "learning_rate": 7.298046152532148e-05, + "loss": 0.0111, + "step": 38060 + }, + { + "epoch": 5.913327120223672, + "grad_norm": 0.27516797184944153, + "learning_rate": 7.296577545221668e-05, + "loss": 0.0095, + "step": 38070 + }, + { + "epoch": 5.914880397639019, + "grad_norm": 0.25599393248558044, + "learning_rate": 7.29510868676092e-05, + "loss": 0.0081, + "step": 38080 + }, + { + "epoch": 5.916433675054365, + "grad_norm": 0.2621649205684662, + "learning_rate": 7.293639577310538e-05, + "loss": 0.0078, + "step": 38090 + }, + { + "epoch": 5.917986952469711, + "grad_norm": 0.14968141913414001, + "learning_rate": 7.292170217031179e-05, + "loss": 0.0083, + "step": 38100 + }, + { + "epoch": 5.919540229885057, + "grad_norm": 0.32401520013809204, + "learning_rate": 7.290700606083532e-05, + "loss": 0.0084, + "step": 38110 + }, + { + "epoch": 5.921093507300403, + "grad_norm": 0.16702568531036377, + "learning_rate": 7.28923074462831e-05, + "loss": 0.0076, + "step": 38120 + }, + { + "epoch": 5.9226467847157505, + "grad_norm": 0.2324889749288559, + "learning_rate": 7.287760632826255e-05, + "loss": 0.0151, + "step": 38130 + }, + { + "epoch": 5.924200062131097, + "grad_norm": 0.2080790251493454, + "learning_rate": 7.286290270838139e-05, + "loss": 0.0073, + "step": 38140 + }, + { + "epoch": 5.925753339546443, + "grad_norm": 0.26136472821235657, + "learning_rate": 7.284819658824756e-05, + "loss": 0.0088, + "step": 38150 + }, + { + "epoch": 5.92730661696179, + "grad_norm": 0.3243994414806366, + "learning_rate": 7.283348796946929e-05, + "loss": 0.0094, + "step": 38160 + }, + { + "epoch": 5.928859894377136, + "grad_norm": 0.2346004694700241, + "learning_rate": 7.28187768536551e-05, + "loss": 0.0049, + "step": 38170 + }, + { + "epoch": 5.930413171792482, + "grad_norm": 0.2540808618068695, + "learning_rate": 7.280406324241379e-05, + "loss": 0.0065, + "step": 38180 + }, + { + "epoch": 5.931966449207828, + "grad_norm": 0.16263513267040253, + "learning_rate": 7.278934713735441e-05, + "loss": 0.0094, + "step": 38190 + }, + { + "epoch": 5.9335197266231745, + "grad_norm": 0.2937818765640259, + "learning_rate": 7.277462854008629e-05, + "loss": 0.0124, + "step": 38200 + }, + { + "epoch": 5.935073004038522, + "grad_norm": 0.27789849042892456, + "learning_rate": 7.275990745221904e-05, + "loss": 0.0115, + "step": 38210 + }, + { + "epoch": 5.936626281453868, + "grad_norm": 0.15851694345474243, + "learning_rate": 7.274518387536252e-05, + "loss": 0.0105, + "step": 38220 + }, + { + "epoch": 5.938179558869214, + "grad_norm": 0.31584447622299194, + "learning_rate": 7.273045781112689e-05, + "loss": 0.0101, + "step": 38230 + }, + { + "epoch": 5.93973283628456, + "grad_norm": 0.3361538350582123, + "learning_rate": 7.271572926112257e-05, + "loss": 0.0119, + "step": 38240 + }, + { + "epoch": 5.941286113699907, + "grad_norm": 0.23376478254795074, + "learning_rate": 7.270099822696024e-05, + "loss": 0.0108, + "step": 38250 + }, + { + "epoch": 5.942839391115253, + "grad_norm": 0.37194299697875977, + "learning_rate": 7.26862647102509e-05, + "loss": 0.0094, + "step": 38260 + }, + { + "epoch": 5.944392668530599, + "grad_norm": 0.3215656578540802, + "learning_rate": 7.267152871260573e-05, + "loss": 0.0072, + "step": 38270 + }, + { + "epoch": 5.945945945945946, + "grad_norm": 0.1874818354845047, + "learning_rate": 7.265679023563629e-05, + "loss": 0.008, + "step": 38280 + }, + { + "epoch": 5.947499223361293, + "grad_norm": 0.322254478931427, + "learning_rate": 7.264204928095433e-05, + "loss": 0.0107, + "step": 38290 + }, + { + "epoch": 5.949052500776639, + "grad_norm": 0.22106774151325226, + "learning_rate": 7.262730585017188e-05, + "loss": 0.0096, + "step": 38300 + }, + { + "epoch": 5.950605778191985, + "grad_norm": 0.3189784288406372, + "learning_rate": 7.261255994490131e-05, + "loss": 0.0092, + "step": 38310 + }, + { + "epoch": 5.952159055607331, + "grad_norm": 0.5394068360328674, + "learning_rate": 7.259781156675517e-05, + "loss": 0.0075, + "step": 38320 + }, + { + "epoch": 5.953712333022678, + "grad_norm": 0.2901582717895508, + "learning_rate": 7.258306071734633e-05, + "loss": 0.009, + "step": 38330 + }, + { + "epoch": 5.955265610438024, + "grad_norm": 0.2358003705739975, + "learning_rate": 7.256830739828792e-05, + "loss": 0.01, + "step": 38340 + }, + { + "epoch": 5.9568188878533705, + "grad_norm": 0.3654966950416565, + "learning_rate": 7.255355161119336e-05, + "loss": 0.0084, + "step": 38350 + }, + { + "epoch": 5.958372165268717, + "grad_norm": 0.276597261428833, + "learning_rate": 7.253879335767628e-05, + "loss": 0.0081, + "step": 38360 + }, + { + "epoch": 5.959925442684064, + "grad_norm": 0.5349632501602173, + "learning_rate": 7.252403263935067e-05, + "loss": 0.0079, + "step": 38370 + }, + { + "epoch": 5.96147872009941, + "grad_norm": 0.28361591696739197, + "learning_rate": 7.25092694578307e-05, + "loss": 0.0102, + "step": 38380 + }, + { + "epoch": 5.963031997514756, + "grad_norm": 0.2593594491481781, + "learning_rate": 7.249450381473087e-05, + "loss": 0.0109, + "step": 38390 + }, + { + "epoch": 5.964585274930102, + "grad_norm": 0.37692564725875854, + "learning_rate": 7.247973571166593e-05, + "loss": 0.0097, + "step": 38400 + }, + { + "epoch": 5.966138552345448, + "grad_norm": 0.3443414866924286, + "learning_rate": 7.246496515025089e-05, + "loss": 0.0052, + "step": 38410 + }, + { + "epoch": 5.9676918297607955, + "grad_norm": 0.1396748274564743, + "learning_rate": 7.245019213210101e-05, + "loss": 0.0092, + "step": 38420 + }, + { + "epoch": 5.969245107176142, + "grad_norm": 0.2966932952404022, + "learning_rate": 7.24354166588319e-05, + "loss": 0.0063, + "step": 38430 + }, + { + "epoch": 5.970798384591488, + "grad_norm": 0.3237263262271881, + "learning_rate": 7.242063873205935e-05, + "loss": 0.0084, + "step": 38440 + }, + { + "epoch": 5.972351662006835, + "grad_norm": 0.3028704822063446, + "learning_rate": 7.240585835339946e-05, + "loss": 0.007, + "step": 38450 + }, + { + "epoch": 5.973904939422181, + "grad_norm": 0.2699637711048126, + "learning_rate": 7.239107552446857e-05, + "loss": 0.008, + "step": 38460 + }, + { + "epoch": 5.975458216837527, + "grad_norm": 0.33317360281944275, + "learning_rate": 7.237629024688334e-05, + "loss": 0.0077, + "step": 38470 + }, + { + "epoch": 5.977011494252873, + "grad_norm": 0.521246612071991, + "learning_rate": 7.236150252226064e-05, + "loss": 0.011, + "step": 38480 + }, + { + "epoch": 5.9785647716682195, + "grad_norm": 0.25710004568099976, + "learning_rate": 7.234671235221765e-05, + "loss": 0.0064, + "step": 38490 + }, + { + "epoch": 5.980118049083567, + "grad_norm": 0.2118135392665863, + "learning_rate": 7.233191973837179e-05, + "loss": 0.0094, + "step": 38500 + }, + { + "epoch": 5.981671326498913, + "grad_norm": 0.28118064999580383, + "learning_rate": 7.231712468234078e-05, + "loss": 0.0062, + "step": 38510 + }, + { + "epoch": 5.983224603914259, + "grad_norm": 0.19719214737415314, + "learning_rate": 7.230232718574254e-05, + "loss": 0.0093, + "step": 38520 + }, + { + "epoch": 5.984777881329605, + "grad_norm": 0.22442156076431274, + "learning_rate": 7.228752725019535e-05, + "loss": 0.0071, + "step": 38530 + }, + { + "epoch": 5.986331158744952, + "grad_norm": 0.29867470264434814, + "learning_rate": 7.227272487731769e-05, + "loss": 0.0067, + "step": 38540 + }, + { + "epoch": 5.987884436160298, + "grad_norm": 0.2515275180339813, + "learning_rate": 7.225792006872831e-05, + "loss": 0.0082, + "step": 38550 + }, + { + "epoch": 5.9894377135756445, + "grad_norm": 0.17021122574806213, + "learning_rate": 7.224311282604628e-05, + "loss": 0.0085, + "step": 38560 + }, + { + "epoch": 5.990990990990991, + "grad_norm": 0.3473406732082367, + "learning_rate": 7.222830315089085e-05, + "loss": 0.0091, + "step": 38570 + }, + { + "epoch": 5.992544268406338, + "grad_norm": 0.4410620629787445, + "learning_rate": 7.22134910448816e-05, + "loss": 0.0064, + "step": 38580 + }, + { + "epoch": 5.994097545821684, + "grad_norm": 0.3371983766555786, + "learning_rate": 7.219867650963839e-05, + "loss": 0.0104, + "step": 38590 + }, + { + "epoch": 5.99565082323703, + "grad_norm": 0.49985337257385254, + "learning_rate": 7.21838595467813e-05, + "loss": 0.0074, + "step": 38600 + }, + { + "epoch": 5.997204100652376, + "grad_norm": 0.22310973703861237, + "learning_rate": 7.216904015793067e-05, + "loss": 0.0093, + "step": 38610 + }, + { + "epoch": 5.998757378067723, + "grad_norm": 0.29040291905403137, + "learning_rate": 7.215421834470713e-05, + "loss": 0.0129, + "step": 38620 + }, + { + "epoch": 6.000310655483069, + "grad_norm": 0.3967713713645935, + "learning_rate": 7.213939410873158e-05, + "loss": 0.0089, + "step": 38630 + }, + { + "epoch": 6.001863932898416, + "grad_norm": 0.12331590056419373, + "learning_rate": 7.212456745162519e-05, + "loss": 0.0094, + "step": 38640 + }, + { + "epoch": 6.003417210313762, + "grad_norm": 0.33661800622940063, + "learning_rate": 7.210973837500937e-05, + "loss": 0.01, + "step": 38650 + }, + { + "epoch": 6.004970487729109, + "grad_norm": 0.30556684732437134, + "learning_rate": 7.209490688050578e-05, + "loss": 0.0077, + "step": 38660 + }, + { + "epoch": 6.006523765144455, + "grad_norm": 0.19469742476940155, + "learning_rate": 7.208007296973641e-05, + "loss": 0.0082, + "step": 38670 + }, + { + "epoch": 6.008077042559801, + "grad_norm": 0.41477686166763306, + "learning_rate": 7.206523664432345e-05, + "loss": 0.0129, + "step": 38680 + }, + { + "epoch": 6.009630319975147, + "grad_norm": 0.26082390546798706, + "learning_rate": 7.205039790588939e-05, + "loss": 0.0083, + "step": 38690 + }, + { + "epoch": 6.011183597390494, + "grad_norm": 0.2341446578502655, + "learning_rate": 7.203555675605697e-05, + "loss": 0.0063, + "step": 38700 + }, + { + "epoch": 6.0127368748058405, + "grad_norm": 0.33626264333724976, + "learning_rate": 7.202071319644917e-05, + "loss": 0.0053, + "step": 38710 + }, + { + "epoch": 6.014290152221187, + "grad_norm": 0.3868255615234375, + "learning_rate": 7.200586722868932e-05, + "loss": 0.0071, + "step": 38720 + }, + { + "epoch": 6.015843429636533, + "grad_norm": 0.21028025448322296, + "learning_rate": 7.199101885440088e-05, + "loss": 0.0065, + "step": 38730 + }, + { + "epoch": 6.01739670705188, + "grad_norm": 0.39213502407073975, + "learning_rate": 7.197616807520771e-05, + "loss": 0.013, + "step": 38740 + }, + { + "epoch": 6.018949984467226, + "grad_norm": 0.20422227680683136, + "learning_rate": 7.196131489273381e-05, + "loss": 0.0099, + "step": 38750 + }, + { + "epoch": 6.020503261882572, + "grad_norm": 0.33000385761260986, + "learning_rate": 7.194645930860355e-05, + "loss": 0.0079, + "step": 38760 + }, + { + "epoch": 6.022056539297918, + "grad_norm": 0.6797981858253479, + "learning_rate": 7.19316013244415e-05, + "loss": 0.0081, + "step": 38770 + }, + { + "epoch": 6.023609816713265, + "grad_norm": 0.7131702303886414, + "learning_rate": 7.191674094187248e-05, + "loss": 0.0095, + "step": 38780 + }, + { + "epoch": 6.025163094128612, + "grad_norm": 0.21595433354377747, + "learning_rate": 7.190187816252165e-05, + "loss": 0.0129, + "step": 38790 + }, + { + "epoch": 6.026716371543958, + "grad_norm": 0.2056611031293869, + "learning_rate": 7.188701298801435e-05, + "loss": 0.0071, + "step": 38800 + }, + { + "epoch": 6.028269648959304, + "grad_norm": 0.2884349822998047, + "learning_rate": 7.18721454199762e-05, + "loss": 0.008, + "step": 38810 + }, + { + "epoch": 6.02982292637465, + "grad_norm": 0.25408318638801575, + "learning_rate": 7.185727546003312e-05, + "loss": 0.0124, + "step": 38820 + }, + { + "epoch": 6.031376203789997, + "grad_norm": 0.1488032341003418, + "learning_rate": 7.184240310981126e-05, + "loss": 0.0077, + "step": 38830 + }, + { + "epoch": 6.032929481205343, + "grad_norm": 0.2790328562259674, + "learning_rate": 7.182752837093704e-05, + "loss": 0.0097, + "step": 38840 + }, + { + "epoch": 6.0344827586206895, + "grad_norm": 0.19893862307071686, + "learning_rate": 7.181265124503711e-05, + "loss": 0.0074, + "step": 38850 + }, + { + "epoch": 6.036036036036036, + "grad_norm": 0.22821606695652008, + "learning_rate": 7.179777173373847e-05, + "loss": 0.0076, + "step": 38860 + }, + { + "epoch": 6.037589313451383, + "grad_norm": 0.396765798330307, + "learning_rate": 7.178288983866826e-05, + "loss": 0.0083, + "step": 38870 + }, + { + "epoch": 6.039142590866729, + "grad_norm": 0.33871251344680786, + "learning_rate": 7.176800556145397e-05, + "loss": 0.011, + "step": 38880 + }, + { + "epoch": 6.040695868282075, + "grad_norm": 0.24878346920013428, + "learning_rate": 7.175311890372334e-05, + "loss": 0.0058, + "step": 38890 + }, + { + "epoch": 6.042249145697421, + "grad_norm": 1.169765830039978, + "learning_rate": 7.17382298671043e-05, + "loss": 0.0094, + "step": 38900 + }, + { + "epoch": 6.043802423112768, + "grad_norm": 0.3979125916957855, + "learning_rate": 7.172333845322515e-05, + "loss": 0.0091, + "step": 38910 + }, + { + "epoch": 6.045355700528114, + "grad_norm": 0.28593525290489197, + "learning_rate": 7.170844466371436e-05, + "loss": 0.007, + "step": 38920 + }, + { + "epoch": 6.046908977943461, + "grad_norm": 0.1892990916967392, + "learning_rate": 7.16935485002007e-05, + "loss": 0.0074, + "step": 38930 + }, + { + "epoch": 6.048462255358807, + "grad_norm": 0.2518410086631775, + "learning_rate": 7.167864996431319e-05, + "loss": 0.0095, + "step": 38940 + }, + { + "epoch": 6.050015532774154, + "grad_norm": 0.269004225730896, + "learning_rate": 7.166374905768111e-05, + "loss": 0.0082, + "step": 38950 + }, + { + "epoch": 6.0515688101895, + "grad_norm": 0.2578313946723938, + "learning_rate": 7.164884578193401e-05, + "loss": 0.0087, + "step": 38960 + }, + { + "epoch": 6.053122087604846, + "grad_norm": 0.3481927514076233, + "learning_rate": 7.163394013870168e-05, + "loss": 0.0092, + "step": 38970 + }, + { + "epoch": 6.054675365020192, + "grad_norm": 0.21801145374774933, + "learning_rate": 7.161903212961419e-05, + "loss": 0.0087, + "step": 38980 + }, + { + "epoch": 6.056228642435539, + "grad_norm": 0.17549005150794983, + "learning_rate": 7.160412175630183e-05, + "loss": 0.0064, + "step": 38990 + }, + { + "epoch": 6.0577819198508855, + "grad_norm": 0.34908339381217957, + "learning_rate": 7.158920902039521e-05, + "loss": 0.0088, + "step": 39000 + }, + { + "epoch": 6.059335197266232, + "grad_norm": 0.36712443828582764, + "learning_rate": 7.157429392352514e-05, + "loss": 0.0108, + "step": 39010 + }, + { + "epoch": 6.060888474681578, + "grad_norm": 0.19371671974658966, + "learning_rate": 7.155937646732275e-05, + "loss": 0.0108, + "step": 39020 + }, + { + "epoch": 6.062441752096925, + "grad_norm": 0.3309604525566101, + "learning_rate": 7.154445665341933e-05, + "loss": 0.0098, + "step": 39030 + }, + { + "epoch": 6.063995029512271, + "grad_norm": 0.28766846656799316, + "learning_rate": 7.152953448344654e-05, + "loss": 0.0093, + "step": 39040 + }, + { + "epoch": 6.065548306927617, + "grad_norm": 0.30348244309425354, + "learning_rate": 7.151460995903624e-05, + "loss": 0.0096, + "step": 39050 + }, + { + "epoch": 6.067101584342963, + "grad_norm": 0.38940590620040894, + "learning_rate": 7.149968308182052e-05, + "loss": 0.0094, + "step": 39060 + }, + { + "epoch": 6.06865486175831, + "grad_norm": 0.18919292092323303, + "learning_rate": 7.14847538534318e-05, + "loss": 0.009, + "step": 39070 + }, + { + "epoch": 6.070208139173657, + "grad_norm": 0.49613404273986816, + "learning_rate": 7.14698222755027e-05, + "loss": 0.0076, + "step": 39080 + }, + { + "epoch": 6.071761416589003, + "grad_norm": 0.2243163287639618, + "learning_rate": 7.14548883496661e-05, + "loss": 0.0067, + "step": 39090 + }, + { + "epoch": 6.073314694004349, + "grad_norm": 0.13307853043079376, + "learning_rate": 7.143995207755517e-05, + "loss": 0.0098, + "step": 39100 + }, + { + "epoch": 6.074867971419695, + "grad_norm": 0.37740668654441833, + "learning_rate": 7.142501346080333e-05, + "loss": 0.0108, + "step": 39110 + }, + { + "epoch": 6.076421248835042, + "grad_norm": 0.23919853568077087, + "learning_rate": 7.141007250104421e-05, + "loss": 0.0075, + "step": 39120 + }, + { + "epoch": 6.077974526250388, + "grad_norm": 0.15751779079437256, + "learning_rate": 7.139512919991176e-05, + "loss": 0.0064, + "step": 39130 + }, + { + "epoch": 6.0795278036657345, + "grad_norm": 0.330911248922348, + "learning_rate": 7.138018355904014e-05, + "loss": 0.0104, + "step": 39140 + }, + { + "epoch": 6.081081081081081, + "grad_norm": 0.3255827724933624, + "learning_rate": 7.13652355800638e-05, + "loss": 0.0076, + "step": 39150 + }, + { + "epoch": 6.082634358496428, + "grad_norm": 0.2120315581560135, + "learning_rate": 7.135028526461743e-05, + "loss": 0.0055, + "step": 39160 + }, + { + "epoch": 6.084187635911774, + "grad_norm": 0.5359017252922058, + "learning_rate": 7.133533261433594e-05, + "loss": 0.0089, + "step": 39170 + }, + { + "epoch": 6.08574091332712, + "grad_norm": 0.1576516181230545, + "learning_rate": 7.132037763085457e-05, + "loss": 0.0077, + "step": 39180 + }, + { + "epoch": 6.087294190742466, + "grad_norm": 0.20247609913349152, + "learning_rate": 7.130542031580875e-05, + "loss": 0.0067, + "step": 39190 + }, + { + "epoch": 6.088847468157813, + "grad_norm": 0.26227641105651855, + "learning_rate": 7.12904606708342e-05, + "loss": 0.009, + "step": 39200 + }, + { + "epoch": 6.0904007455731595, + "grad_norm": 1.4684360027313232, + "learning_rate": 7.127549869756687e-05, + "loss": 0.0096, + "step": 39210 + }, + { + "epoch": 6.091954022988506, + "grad_norm": 0.4038849472999573, + "learning_rate": 7.126053439764299e-05, + "loss": 0.0099, + "step": 39220 + }, + { + "epoch": 6.093507300403852, + "grad_norm": 0.21175597608089447, + "learning_rate": 7.124556777269904e-05, + "loss": 0.0071, + "step": 39230 + }, + { + "epoch": 6.095060577819199, + "grad_norm": 0.34824642539024353, + "learning_rate": 7.123059882437174e-05, + "loss": 0.007, + "step": 39240 + }, + { + "epoch": 6.096613855234545, + "grad_norm": 0.11603845655918121, + "learning_rate": 7.121562755429807e-05, + "loss": 0.0105, + "step": 39250 + }, + { + "epoch": 6.098167132649891, + "grad_norm": 0.28341907262802124, + "learning_rate": 7.120065396411528e-05, + "loss": 0.0082, + "step": 39260 + }, + { + "epoch": 6.099720410065237, + "grad_norm": 0.22952334582805634, + "learning_rate": 7.118567805546084e-05, + "loss": 0.0072, + "step": 39270 + }, + { + "epoch": 6.101273687480584, + "grad_norm": 0.1876501739025116, + "learning_rate": 7.117069982997248e-05, + "loss": 0.0073, + "step": 39280 + }, + { + "epoch": 6.102826964895931, + "grad_norm": 0.2428351193666458, + "learning_rate": 7.115571928928825e-05, + "loss": 0.0093, + "step": 39290 + }, + { + "epoch": 6.104380242311277, + "grad_norm": 0.196332186460495, + "learning_rate": 7.114073643504635e-05, + "loss": 0.0099, + "step": 39300 + }, + { + "epoch": 6.105933519726623, + "grad_norm": 0.17044763267040253, + "learning_rate": 7.11257512688853e-05, + "loss": 0.0069, + "step": 39310 + }, + { + "epoch": 6.10748679714197, + "grad_norm": 0.20584358274936676, + "learning_rate": 7.111076379244384e-05, + "loss": 0.0095, + "step": 39320 + }, + { + "epoch": 6.109040074557316, + "grad_norm": 0.30629590153694153, + "learning_rate": 7.109577400736101e-05, + "loss": 0.0101, + "step": 39330 + }, + { + "epoch": 6.110593351972662, + "grad_norm": 0.2754763960838318, + "learning_rate": 7.108078191527605e-05, + "loss": 0.0106, + "step": 39340 + }, + { + "epoch": 6.1121466293880085, + "grad_norm": 0.2762077748775482, + "learning_rate": 7.106578751782847e-05, + "loss": 0.0097, + "step": 39350 + }, + { + "epoch": 6.1136999068033555, + "grad_norm": 0.3858078420162201, + "learning_rate": 7.105079081665803e-05, + "loss": 0.0083, + "step": 39360 + }, + { + "epoch": 6.115253184218702, + "grad_norm": 0.2096669226884842, + "learning_rate": 7.103579181340476e-05, + "loss": 0.0066, + "step": 39370 + }, + { + "epoch": 6.116806461634048, + "grad_norm": 0.14781184494495392, + "learning_rate": 7.102079050970893e-05, + "loss": 0.0054, + "step": 39380 + }, + { + "epoch": 6.118359739049394, + "grad_norm": 0.22780385613441467, + "learning_rate": 7.100578690721105e-05, + "loss": 0.0081, + "step": 39390 + }, + { + "epoch": 6.11991301646474, + "grad_norm": 0.3038531541824341, + "learning_rate": 7.09907810075519e-05, + "loss": 0.0072, + "step": 39400 + }, + { + "epoch": 6.121466293880087, + "grad_norm": 0.20583277940750122, + "learning_rate": 7.097577281237249e-05, + "loss": 0.0064, + "step": 39410 + }, + { + "epoch": 6.123019571295433, + "grad_norm": 0.3516615033149719, + "learning_rate": 7.09607623233141e-05, + "loss": 0.0118, + "step": 39420 + }, + { + "epoch": 6.12457284871078, + "grad_norm": 0.31583985686302185, + "learning_rate": 7.094574954201827e-05, + "loss": 0.0084, + "step": 39430 + }, + { + "epoch": 6.126126126126126, + "grad_norm": 0.3441278636455536, + "learning_rate": 7.093073447012675e-05, + "loss": 0.0082, + "step": 39440 + }, + { + "epoch": 6.127679403541473, + "grad_norm": 0.3524284362792969, + "learning_rate": 7.09157171092816e-05, + "loss": 0.0092, + "step": 39450 + }, + { + "epoch": 6.129232680956819, + "grad_norm": 0.3328855633735657, + "learning_rate": 7.090069746112504e-05, + "loss": 0.0067, + "step": 39460 + }, + { + "epoch": 6.130785958372165, + "grad_norm": 0.48681357502937317, + "learning_rate": 7.088567552729965e-05, + "loss": 0.0101, + "step": 39470 + }, + { + "epoch": 6.132339235787511, + "grad_norm": 0.3940122425556183, + "learning_rate": 7.087065130944818e-05, + "loss": 0.0109, + "step": 39480 + }, + { + "epoch": 6.133892513202858, + "grad_norm": 0.2104654312133789, + "learning_rate": 7.085562480921366e-05, + "loss": 0.0099, + "step": 39490 + }, + { + "epoch": 6.1354457906182045, + "grad_norm": 0.3878689706325531, + "learning_rate": 7.084059602823937e-05, + "loss": 0.0086, + "step": 39500 + }, + { + "epoch": 6.136999068033551, + "grad_norm": 0.16748231649398804, + "learning_rate": 7.082556496816882e-05, + "loss": 0.0094, + "step": 39510 + }, + { + "epoch": 6.138552345448897, + "grad_norm": 0.3661520779132843, + "learning_rate": 7.081053163064582e-05, + "loss": 0.0137, + "step": 39520 + }, + { + "epoch": 6.140105622864244, + "grad_norm": 0.24933627247810364, + "learning_rate": 7.079549601731434e-05, + "loss": 0.0082, + "step": 39530 + }, + { + "epoch": 6.14165890027959, + "grad_norm": 0.32344114780426025, + "learning_rate": 7.07804581298187e-05, + "loss": 0.0071, + "step": 39540 + }, + { + "epoch": 6.143212177694936, + "grad_norm": 0.20949013531208038, + "learning_rate": 7.07654179698034e-05, + "loss": 0.0072, + "step": 39550 + }, + { + "epoch": 6.144765455110282, + "grad_norm": 0.3385145366191864, + "learning_rate": 7.075037553891321e-05, + "loss": 0.0098, + "step": 39560 + }, + { + "epoch": 6.146318732525629, + "grad_norm": 0.23956476151943207, + "learning_rate": 7.073533083879315e-05, + "loss": 0.0083, + "step": 39570 + }, + { + "epoch": 6.147872009940976, + "grad_norm": 0.2259061485528946, + "learning_rate": 7.072028387108849e-05, + "loss": 0.01, + "step": 39580 + }, + { + "epoch": 6.149425287356322, + "grad_norm": 0.19603826105594635, + "learning_rate": 7.070523463744472e-05, + "loss": 0.0075, + "step": 39590 + }, + { + "epoch": 6.150978564771668, + "grad_norm": 0.21002168953418732, + "learning_rate": 7.069018313950763e-05, + "loss": 0.0086, + "step": 39600 + }, + { + "epoch": 6.152531842187015, + "grad_norm": 0.292427122592926, + "learning_rate": 7.067512937892322e-05, + "loss": 0.0086, + "step": 39610 + }, + { + "epoch": 6.154085119602361, + "grad_norm": 0.3350582420825958, + "learning_rate": 7.066007335733775e-05, + "loss": 0.0078, + "step": 39620 + }, + { + "epoch": 6.155638397017707, + "grad_norm": 0.3149566948413849, + "learning_rate": 7.064501507639772e-05, + "loss": 0.0098, + "step": 39630 + }, + { + "epoch": 6.1571916744330535, + "grad_norm": 0.3557598292827606, + "learning_rate": 7.062995453774987e-05, + "loss": 0.0077, + "step": 39640 + }, + { + "epoch": 6.1587449518484005, + "grad_norm": 0.3414018452167511, + "learning_rate": 7.061489174304121e-05, + "loss": 0.0098, + "step": 39650 + }, + { + "epoch": 6.160298229263747, + "grad_norm": 0.37393003702163696, + "learning_rate": 7.0599826693919e-05, + "loss": 0.0087, + "step": 39660 + }, + { + "epoch": 6.161851506679093, + "grad_norm": 0.14590702950954437, + "learning_rate": 7.05847593920307e-05, + "loss": 0.0066, + "step": 39670 + }, + { + "epoch": 6.163404784094439, + "grad_norm": 0.2123357355594635, + "learning_rate": 7.056968983902406e-05, + "loss": 0.0089, + "step": 39680 + }, + { + "epoch": 6.164958061509786, + "grad_norm": 0.24185670912265778, + "learning_rate": 7.055461803654706e-05, + "loss": 0.0066, + "step": 39690 + }, + { + "epoch": 6.166511338925132, + "grad_norm": 0.16563405096530914, + "learning_rate": 7.053954398624794e-05, + "loss": 0.0094, + "step": 39700 + }, + { + "epoch": 6.168064616340478, + "grad_norm": 0.6271448731422424, + "learning_rate": 7.052446768977518e-05, + "loss": 0.008, + "step": 39710 + }, + { + "epoch": 6.169617893755825, + "grad_norm": 0.25981760025024414, + "learning_rate": 7.050938914877748e-05, + "loss": 0.0072, + "step": 39720 + }, + { + "epoch": 6.171171171171171, + "grad_norm": 0.3100626766681671, + "learning_rate": 7.049430836490382e-05, + "loss": 0.0098, + "step": 39730 + }, + { + "epoch": 6.172724448586518, + "grad_norm": 0.25401410460472107, + "learning_rate": 7.047922533980342e-05, + "loss": 0.0066, + "step": 39740 + }, + { + "epoch": 6.174277726001864, + "grad_norm": 0.26692691445350647, + "learning_rate": 7.046414007512571e-05, + "loss": 0.0076, + "step": 39750 + }, + { + "epoch": 6.17583100341721, + "grad_norm": 0.27836331725120544, + "learning_rate": 7.044905257252042e-05, + "loss": 0.008, + "step": 39760 + }, + { + "epoch": 6.177384280832556, + "grad_norm": 0.2614710330963135, + "learning_rate": 7.04339628336375e-05, + "loss": 0.0093, + "step": 39770 + }, + { + "epoch": 6.178937558247903, + "grad_norm": 0.2312423586845398, + "learning_rate": 7.041887086012711e-05, + "loss": 0.0092, + "step": 39780 + }, + { + "epoch": 6.1804908356632495, + "grad_norm": 0.28200066089630127, + "learning_rate": 7.040377665363969e-05, + "loss": 0.0108, + "step": 39790 + }, + { + "epoch": 6.182044113078596, + "grad_norm": 0.3243563175201416, + "learning_rate": 7.038868021582594e-05, + "loss": 0.0078, + "step": 39800 + }, + { + "epoch": 6.183597390493942, + "grad_norm": 0.47445887327194214, + "learning_rate": 7.037358154833679e-05, + "loss": 0.0103, + "step": 39810 + }, + { + "epoch": 6.185150667909289, + "grad_norm": 0.1221771165728569, + "learning_rate": 7.035848065282339e-05, + "loss": 0.007, + "step": 39820 + }, + { + "epoch": 6.186703945324635, + "grad_norm": 0.16425937414169312, + "learning_rate": 7.034337753093714e-05, + "loss": 0.0095, + "step": 39830 + }, + { + "epoch": 6.188257222739981, + "grad_norm": 0.3810938000679016, + "learning_rate": 7.032827218432972e-05, + "loss": 0.0095, + "step": 39840 + }, + { + "epoch": 6.189810500155327, + "grad_norm": 0.2689937353134155, + "learning_rate": 7.031316461465302e-05, + "loss": 0.0061, + "step": 39850 + }, + { + "epoch": 6.1913637775706745, + "grad_norm": 0.15466199815273285, + "learning_rate": 7.029805482355915e-05, + "loss": 0.011, + "step": 39860 + }, + { + "epoch": 6.192917054986021, + "grad_norm": 0.1748952567577362, + "learning_rate": 7.028294281270055e-05, + "loss": 0.0073, + "step": 39870 + }, + { + "epoch": 6.194470332401367, + "grad_norm": 0.18706844747066498, + "learning_rate": 7.026782858372982e-05, + "loss": 0.0123, + "step": 39880 + }, + { + "epoch": 6.196023609816713, + "grad_norm": 0.22885487973690033, + "learning_rate": 7.025271213829982e-05, + "loss": 0.0069, + "step": 39890 + }, + { + "epoch": 6.19757688723206, + "grad_norm": 0.2455316036939621, + "learning_rate": 7.023759347806366e-05, + "loss": 0.0103, + "step": 39900 + }, + { + "epoch": 6.199130164647406, + "grad_norm": 0.19397962093353271, + "learning_rate": 7.022247260467472e-05, + "loss": 0.0101, + "step": 39910 + }, + { + "epoch": 6.200683442062752, + "grad_norm": 0.35141342878341675, + "learning_rate": 7.020734951978658e-05, + "loss": 0.0087, + "step": 39920 + }, + { + "epoch": 6.2022367194780985, + "grad_norm": 0.2942061424255371, + "learning_rate": 7.019222422505307e-05, + "loss": 0.0116, + "step": 39930 + }, + { + "epoch": 6.203789996893446, + "grad_norm": 0.21598327159881592, + "learning_rate": 7.017709672212827e-05, + "loss": 0.0081, + "step": 39940 + }, + { + "epoch": 6.205343274308792, + "grad_norm": 0.24302925169467926, + "learning_rate": 7.016196701266652e-05, + "loss": 0.0108, + "step": 39950 + }, + { + "epoch": 6.206896551724138, + "grad_norm": 0.3312130272388458, + "learning_rate": 7.014683509832238e-05, + "loss": 0.0083, + "step": 39960 + }, + { + "epoch": 6.208449829139484, + "grad_norm": 0.17761129140853882, + "learning_rate": 7.013170098075063e-05, + "loss": 0.0066, + "step": 39970 + }, + { + "epoch": 6.210003106554831, + "grad_norm": 0.1406506896018982, + "learning_rate": 7.011656466160632e-05, + "loss": 0.0082, + "step": 39980 + }, + { + "epoch": 6.211556383970177, + "grad_norm": 0.25662466883659363, + "learning_rate": 7.010142614254475e-05, + "loss": 0.0095, + "step": 39990 + }, + { + "epoch": 6.2131096613855235, + "grad_norm": 0.47788524627685547, + "learning_rate": 7.008628542522147e-05, + "loss": 0.0072, + "step": 40000 + }, + { + "epoch": 6.21466293880087, + "grad_norm": 0.46651414036750793, + "learning_rate": 7.007114251129218e-05, + "loss": 0.0089, + "step": 40010 + }, + { + "epoch": 6.216216216216216, + "grad_norm": 0.18722522258758545, + "learning_rate": 7.005599740241293e-05, + "loss": 0.0068, + "step": 40020 + }, + { + "epoch": 6.217769493631563, + "grad_norm": 0.21562466025352478, + "learning_rate": 7.004085010023996e-05, + "loss": 0.0088, + "step": 40030 + }, + { + "epoch": 6.219322771046909, + "grad_norm": 0.18074741959571838, + "learning_rate": 7.002570060642976e-05, + "loss": 0.0069, + "step": 40040 + }, + { + "epoch": 6.220876048462255, + "grad_norm": 0.2595876455307007, + "learning_rate": 7.001054892263903e-05, + "loss": 0.0096, + "step": 40050 + }, + { + "epoch": 6.222429325877601, + "grad_norm": 0.19616860151290894, + "learning_rate": 6.999539505052477e-05, + "loss": 0.01, + "step": 40060 + }, + { + "epoch": 6.223982603292948, + "grad_norm": 0.38002026081085205, + "learning_rate": 6.998023899174418e-05, + "loss": 0.009, + "step": 40070 + }, + { + "epoch": 6.225535880708295, + "grad_norm": 0.39174196124076843, + "learning_rate": 6.996508074795467e-05, + "loss": 0.0086, + "step": 40080 + }, + { + "epoch": 6.227089158123641, + "grad_norm": 0.5276153087615967, + "learning_rate": 6.994992032081396e-05, + "loss": 0.0083, + "step": 40090 + }, + { + "epoch": 6.228642435538987, + "grad_norm": 0.3573484420776367, + "learning_rate": 6.993475771197995e-05, + "loss": 0.0074, + "step": 40100 + }, + { + "epoch": 6.230195712954334, + "grad_norm": 0.23869699239730835, + "learning_rate": 6.991959292311082e-05, + "loss": 0.0065, + "step": 40110 + }, + { + "epoch": 6.23174899036968, + "grad_norm": 0.2840666174888611, + "learning_rate": 6.990442595586495e-05, + "loss": 0.0091, + "step": 40120 + }, + { + "epoch": 6.233302267785026, + "grad_norm": 0.21197454631328583, + "learning_rate": 6.988925681190098e-05, + "loss": 0.0069, + "step": 40130 + }, + { + "epoch": 6.234855545200372, + "grad_norm": 0.3702429234981537, + "learning_rate": 6.987408549287778e-05, + "loss": 0.0083, + "step": 40140 + }, + { + "epoch": 6.2364088226157195, + "grad_norm": 0.6216975450515747, + "learning_rate": 6.985891200045449e-05, + "loss": 0.0092, + "step": 40150 + }, + { + "epoch": 6.237962100031066, + "grad_norm": 0.3049895167350769, + "learning_rate": 6.984373633629045e-05, + "loss": 0.0071, + "step": 40160 + }, + { + "epoch": 6.239515377446412, + "grad_norm": 0.4378701448440552, + "learning_rate": 6.98285585020452e-05, + "loss": 0.0106, + "step": 40170 + }, + { + "epoch": 6.241068654861758, + "grad_norm": 0.18424396216869354, + "learning_rate": 6.981337849937864e-05, + "loss": 0.0079, + "step": 40180 + }, + { + "epoch": 6.242621932277105, + "grad_norm": 0.4635373651981354, + "learning_rate": 6.979819632995078e-05, + "loss": 0.0069, + "step": 40190 + }, + { + "epoch": 6.244175209692451, + "grad_norm": 0.17195910215377808, + "learning_rate": 6.978301199542193e-05, + "loss": 0.0082, + "step": 40200 + }, + { + "epoch": 6.245728487107797, + "grad_norm": 0.29842063784599304, + "learning_rate": 6.976782549745263e-05, + "loss": 0.0075, + "step": 40210 + }, + { + "epoch": 6.2472817645231435, + "grad_norm": 0.28102460503578186, + "learning_rate": 6.975263683770367e-05, + "loss": 0.0095, + "step": 40220 + }, + { + "epoch": 6.248835041938491, + "grad_norm": 0.19571073353290558, + "learning_rate": 6.973744601783604e-05, + "loss": 0.0068, + "step": 40230 + }, + { + "epoch": 6.250388319353837, + "grad_norm": 0.20359672605991364, + "learning_rate": 6.972225303951097e-05, + "loss": 0.0062, + "step": 40240 + }, + { + "epoch": 6.251941596769183, + "grad_norm": 0.3053668439388275, + "learning_rate": 6.970705790438998e-05, + "loss": 0.0064, + "step": 40250 + }, + { + "epoch": 6.253494874184529, + "grad_norm": 0.34219369292259216, + "learning_rate": 6.969186061413477e-05, + "loss": 0.0088, + "step": 40260 + }, + { + "epoch": 6.255048151599876, + "grad_norm": 0.17378930747509003, + "learning_rate": 6.967666117040727e-05, + "loss": 0.0065, + "step": 40270 + }, + { + "epoch": 6.256601429015222, + "grad_norm": 0.22983644902706146, + "learning_rate": 6.966145957486972e-05, + "loss": 0.0093, + "step": 40280 + }, + { + "epoch": 6.2581547064305685, + "grad_norm": 0.23415115475654602, + "learning_rate": 6.964625582918449e-05, + "loss": 0.0062, + "step": 40290 + }, + { + "epoch": 6.259707983845915, + "grad_norm": 0.301581472158432, + "learning_rate": 6.963104993501425e-05, + "loss": 0.0127, + "step": 40300 + }, + { + "epoch": 6.261261261261261, + "grad_norm": 0.27423107624053955, + "learning_rate": 6.961584189402192e-05, + "loss": 0.009, + "step": 40310 + }, + { + "epoch": 6.262814538676608, + "grad_norm": 0.27320414781570435, + "learning_rate": 6.96006317078706e-05, + "loss": 0.0085, + "step": 40320 + }, + { + "epoch": 6.264367816091954, + "grad_norm": 0.21715155243873596, + "learning_rate": 6.958541937822367e-05, + "loss": 0.0085, + "step": 40330 + }, + { + "epoch": 6.2659210935073, + "grad_norm": 0.18674063682556152, + "learning_rate": 6.957020490674473e-05, + "loss": 0.0095, + "step": 40340 + }, + { + "epoch": 6.267474370922647, + "grad_norm": 0.25524985790252686, + "learning_rate": 6.95549882950976e-05, + "loss": 0.0085, + "step": 40350 + }, + { + "epoch": 6.269027648337993, + "grad_norm": 0.2608692944049835, + "learning_rate": 6.953976954494635e-05, + "loss": 0.0069, + "step": 40360 + }, + { + "epoch": 6.27058092575334, + "grad_norm": 0.2874123752117157, + "learning_rate": 6.952454865795528e-05, + "loss": 0.0076, + "step": 40370 + }, + { + "epoch": 6.272134203168686, + "grad_norm": 0.2737879455089569, + "learning_rate": 6.950932563578892e-05, + "loss": 0.0074, + "step": 40380 + }, + { + "epoch": 6.273687480584032, + "grad_norm": 0.16068531572818756, + "learning_rate": 6.949410048011206e-05, + "loss": 0.0075, + "step": 40390 + }, + { + "epoch": 6.275240757999379, + "grad_norm": 0.31769073009490967, + "learning_rate": 6.947887319258966e-05, + "loss": 0.0085, + "step": 40400 + }, + { + "epoch": 6.276794035414725, + "grad_norm": 0.1036415547132492, + "learning_rate": 6.946364377488696e-05, + "loss": 0.0065, + "step": 40410 + }, + { + "epoch": 6.278347312830071, + "grad_norm": 0.5969032645225525, + "learning_rate": 6.944841222866947e-05, + "loss": 0.0093, + "step": 40420 + }, + { + "epoch": 6.2799005902454175, + "grad_norm": 0.17978519201278687, + "learning_rate": 6.943317855560284e-05, + "loss": 0.0102, + "step": 40430 + }, + { + "epoch": 6.2814538676607645, + "grad_norm": 0.32026317715644836, + "learning_rate": 6.941794275735302e-05, + "loss": 0.008, + "step": 40440 + }, + { + "epoch": 6.283007145076111, + "grad_norm": 0.21910960972309113, + "learning_rate": 6.94027048355862e-05, + "loss": 0.008, + "step": 40450 + }, + { + "epoch": 6.284560422491457, + "grad_norm": 0.40392714738845825, + "learning_rate": 6.938746479196871e-05, + "loss": 0.0077, + "step": 40460 + }, + { + "epoch": 6.286113699906803, + "grad_norm": 0.6502711772918701, + "learning_rate": 6.937222262816724e-05, + "loss": 0.0076, + "step": 40470 + }, + { + "epoch": 6.28766697732215, + "grad_norm": 0.20242907106876373, + "learning_rate": 6.935697834584865e-05, + "loss": 0.0087, + "step": 40480 + }, + { + "epoch": 6.289220254737496, + "grad_norm": 0.24883854389190674, + "learning_rate": 6.934173194667998e-05, + "loss": 0.0099, + "step": 40490 + }, + { + "epoch": 6.290773532152842, + "grad_norm": 0.2855144143104553, + "learning_rate": 6.93264834323286e-05, + "loss": 0.0104, + "step": 40500 + }, + { + "epoch": 6.292326809568189, + "grad_norm": 0.24327728152275085, + "learning_rate": 6.931123280446206e-05, + "loss": 0.0081, + "step": 40510 + }, + { + "epoch": 6.293880086983536, + "grad_norm": 0.2800053656101227, + "learning_rate": 6.929598006474811e-05, + "loss": 0.0057, + "step": 40520 + }, + { + "epoch": 6.295433364398882, + "grad_norm": 0.36191707849502563, + "learning_rate": 6.92807252148548e-05, + "loss": 0.0086, + "step": 40530 + }, + { + "epoch": 6.296986641814228, + "grad_norm": 0.2126312404870987, + "learning_rate": 6.926546825645037e-05, + "loss": 0.0063, + "step": 40540 + }, + { + "epoch": 6.298539919229574, + "grad_norm": 0.13562379777431488, + "learning_rate": 6.92502091912033e-05, + "loss": 0.0065, + "step": 40550 + }, + { + "epoch": 6.300093196644921, + "grad_norm": 0.20860812067985535, + "learning_rate": 6.923494802078229e-05, + "loss": 0.0059, + "step": 40560 + }, + { + "epoch": 6.301646474060267, + "grad_norm": 0.24044932425022125, + "learning_rate": 6.921968474685628e-05, + "loss": 0.0076, + "step": 40570 + }, + { + "epoch": 6.3031997514756135, + "grad_norm": 0.20477715134620667, + "learning_rate": 6.920441937109445e-05, + "loss": 0.0095, + "step": 40580 + }, + { + "epoch": 6.30475302889096, + "grad_norm": 0.5618107318878174, + "learning_rate": 6.91891518951662e-05, + "loss": 0.0075, + "step": 40590 + }, + { + "epoch": 6.306306306306306, + "grad_norm": 0.21818412840366364, + "learning_rate": 6.917388232074114e-05, + "loss": 0.0104, + "step": 40600 + }, + { + "epoch": 6.307859583721653, + "grad_norm": 0.15089142322540283, + "learning_rate": 6.915861064948914e-05, + "loss": 0.0057, + "step": 40610 + }, + { + "epoch": 6.309412861136999, + "grad_norm": 0.21318180859088898, + "learning_rate": 6.914333688308025e-05, + "loss": 0.0084, + "step": 40620 + }, + { + "epoch": 6.310966138552345, + "grad_norm": 0.1737482249736786, + "learning_rate": 6.912806102318487e-05, + "loss": 0.0064, + "step": 40630 + }, + { + "epoch": 6.312519415967692, + "grad_norm": 0.14276590943336487, + "learning_rate": 6.911278307147347e-05, + "loss": 0.0068, + "step": 40640 + }, + { + "epoch": 6.3140726933830384, + "grad_norm": 0.3678019046783447, + "learning_rate": 6.909750302961684e-05, + "loss": 0.0081, + "step": 40650 + }, + { + "epoch": 6.315625970798385, + "grad_norm": 0.2325626164674759, + "learning_rate": 6.9082220899286e-05, + "loss": 0.0067, + "step": 40660 + }, + { + "epoch": 6.317179248213731, + "grad_norm": 0.18808846175670624, + "learning_rate": 6.906693668215219e-05, + "loss": 0.0088, + "step": 40670 + }, + { + "epoch": 6.318732525629077, + "grad_norm": 0.38459300994873047, + "learning_rate": 6.905165037988683e-05, + "loss": 0.0082, + "step": 40680 + }, + { + "epoch": 6.320285803044424, + "grad_norm": 0.289127379655838, + "learning_rate": 6.903636199416164e-05, + "loss": 0.0069, + "step": 40690 + }, + { + "epoch": 6.32183908045977, + "grad_norm": 0.1448410153388977, + "learning_rate": 6.902107152664851e-05, + "loss": 0.0068, + "step": 40700 + }, + { + "epoch": 6.323392357875116, + "grad_norm": 0.2889256477355957, + "learning_rate": 6.90057789790196e-05, + "loss": 0.0066, + "step": 40710 + }, + { + "epoch": 6.3249456352904625, + "grad_norm": 0.3230835199356079, + "learning_rate": 6.899048435294728e-05, + "loss": 0.0068, + "step": 40720 + }, + { + "epoch": 6.3264989127058096, + "grad_norm": 0.2115601897239685, + "learning_rate": 6.897518765010415e-05, + "loss": 0.0058, + "step": 40730 + }, + { + "epoch": 6.328052190121156, + "grad_norm": 0.2913765013217926, + "learning_rate": 6.895988887216302e-05, + "loss": 0.009, + "step": 40740 + }, + { + "epoch": 6.329605467536502, + "grad_norm": 0.29659947752952576, + "learning_rate": 6.894458802079694e-05, + "loss": 0.0096, + "step": 40750 + }, + { + "epoch": 6.331158744951848, + "grad_norm": 0.5669110417366028, + "learning_rate": 6.892928509767922e-05, + "loss": 0.0107, + "step": 40760 + }, + { + "epoch": 6.332712022367195, + "grad_norm": 0.4914481043815613, + "learning_rate": 6.891398010448333e-05, + "loss": 0.0077, + "step": 40770 + }, + { + "epoch": 6.334265299782541, + "grad_norm": 0.3330259621143341, + "learning_rate": 6.8898673042883e-05, + "loss": 0.008, + "step": 40780 + }, + { + "epoch": 6.335818577197887, + "grad_norm": 0.29875651001930237, + "learning_rate": 6.888336391455222e-05, + "loss": 0.0113, + "step": 40790 + }, + { + "epoch": 6.337371854613234, + "grad_norm": 0.32575860619544983, + "learning_rate": 6.886805272116513e-05, + "loss": 0.0058, + "step": 40800 + }, + { + "epoch": 6.338925132028581, + "grad_norm": 1.2968183755874634, + "learning_rate": 6.885273946439617e-05, + "loss": 0.0092, + "step": 40810 + }, + { + "epoch": 6.340478409443927, + "grad_norm": 0.23186823725700378, + "learning_rate": 6.883742414591998e-05, + "loss": 0.0085, + "step": 40820 + }, + { + "epoch": 6.342031686859273, + "grad_norm": 0.20135176181793213, + "learning_rate": 6.88221067674114e-05, + "loss": 0.0066, + "step": 40830 + }, + { + "epoch": 6.343584964274619, + "grad_norm": 0.31571993231773376, + "learning_rate": 6.88067873305455e-05, + "loss": 0.0087, + "step": 40840 + }, + { + "epoch": 6.345138241689966, + "grad_norm": 0.17825479805469513, + "learning_rate": 6.879146583699765e-05, + "loss": 0.0063, + "step": 40850 + }, + { + "epoch": 6.346691519105312, + "grad_norm": 0.26432669162750244, + "learning_rate": 6.877614228844334e-05, + "loss": 0.0064, + "step": 40860 + }, + { + "epoch": 6.3482447965206585, + "grad_norm": 0.17562589049339294, + "learning_rate": 6.876081668655832e-05, + "loss": 0.0085, + "step": 40870 + }, + { + "epoch": 6.349798073936005, + "grad_norm": 0.29347074031829834, + "learning_rate": 6.874548903301861e-05, + "loss": 0.0086, + "step": 40880 + }, + { + "epoch": 6.351351351351352, + "grad_norm": 0.2886922061443329, + "learning_rate": 6.873015932950039e-05, + "loss": 0.0086, + "step": 40890 + }, + { + "epoch": 6.352904628766698, + "grad_norm": 0.7020079493522644, + "learning_rate": 6.871482757768012e-05, + "loss": 0.0088, + "step": 40900 + }, + { + "epoch": 6.354457906182044, + "grad_norm": 0.3093356490135193, + "learning_rate": 6.869949377923442e-05, + "loss": 0.0078, + "step": 40910 + }, + { + "epoch": 6.35601118359739, + "grad_norm": 0.3057321608066559, + "learning_rate": 6.868415793584022e-05, + "loss": 0.0072, + "step": 40920 + }, + { + "epoch": 6.357564461012737, + "grad_norm": 0.24281452596187592, + "learning_rate": 6.86688200491746e-05, + "loss": 0.0098, + "step": 40930 + }, + { + "epoch": 6.3591177384280835, + "grad_norm": 0.13371606171131134, + "learning_rate": 6.865348012091488e-05, + "loss": 0.0088, + "step": 40940 + }, + { + "epoch": 6.36067101584343, + "grad_norm": 0.21461720764636993, + "learning_rate": 6.86381381527386e-05, + "loss": 0.0112, + "step": 40950 + }, + { + "epoch": 6.362224293258776, + "grad_norm": 0.2548106908798218, + "learning_rate": 6.862279414632359e-05, + "loss": 0.0085, + "step": 40960 + }, + { + "epoch": 6.363777570674122, + "grad_norm": 0.396268367767334, + "learning_rate": 6.860744810334778e-05, + "loss": 0.0073, + "step": 40970 + }, + { + "epoch": 6.365330848089469, + "grad_norm": 0.2555273175239563, + "learning_rate": 6.859210002548943e-05, + "loss": 0.0099, + "step": 40980 + }, + { + "epoch": 6.366884125504815, + "grad_norm": 0.24238067865371704, + "learning_rate": 6.8576749914427e-05, + "loss": 0.0058, + "step": 40990 + }, + { + "epoch": 6.368437402920161, + "grad_norm": 0.29730406403541565, + "learning_rate": 6.85613977718391e-05, + "loss": 0.0089, + "step": 41000 + }, + { + "epoch": 6.3699906803355075, + "grad_norm": 0.2214302271604538, + "learning_rate": 6.854604359940464e-05, + "loss": 0.0071, + "step": 41010 + }, + { + "epoch": 6.371543957750855, + "grad_norm": 0.164116770029068, + "learning_rate": 6.853068739880276e-05, + "loss": 0.0072, + "step": 41020 + }, + { + "epoch": 6.373097235166201, + "grad_norm": 0.2825048565864563, + "learning_rate": 6.851532917171276e-05, + "loss": 0.0065, + "step": 41030 + }, + { + "epoch": 6.374650512581547, + "grad_norm": 0.21946196258068085, + "learning_rate": 6.849996891981418e-05, + "loss": 0.0077, + "step": 41040 + }, + { + "epoch": 6.376203789996893, + "grad_norm": 0.2266911119222641, + "learning_rate": 6.848460664478684e-05, + "loss": 0.0057, + "step": 41050 + }, + { + "epoch": 6.37775706741224, + "grad_norm": 0.2654162645339966, + "learning_rate": 6.846924234831069e-05, + "loss": 0.0086, + "step": 41060 + }, + { + "epoch": 6.379310344827586, + "grad_norm": 0.2903370261192322, + "learning_rate": 6.845387603206597e-05, + "loss": 0.0077, + "step": 41070 + }, + { + "epoch": 6.3808636222429325, + "grad_norm": 0.25776562094688416, + "learning_rate": 6.843850769773312e-05, + "loss": 0.0082, + "step": 41080 + }, + { + "epoch": 6.382416899658279, + "grad_norm": 0.18460310995578766, + "learning_rate": 6.842313734699278e-05, + "loss": 0.0095, + "step": 41090 + }, + { + "epoch": 6.383970177073626, + "grad_norm": 0.2951209545135498, + "learning_rate": 6.840776498152584e-05, + "loss": 0.0077, + "step": 41100 + }, + { + "epoch": 6.385523454488972, + "grad_norm": 0.20814241468906403, + "learning_rate": 6.83923906030134e-05, + "loss": 0.006, + "step": 41110 + }, + { + "epoch": 6.387076731904318, + "grad_norm": 0.28994569182395935, + "learning_rate": 6.837701421313677e-05, + "loss": 0.007, + "step": 41120 + }, + { + "epoch": 6.388630009319664, + "grad_norm": 0.14671435952186584, + "learning_rate": 6.83616358135775e-05, + "loss": 0.0086, + "step": 41130 + }, + { + "epoch": 6.390183286735011, + "grad_norm": 0.1657404750585556, + "learning_rate": 6.834625540601734e-05, + "loss": 0.0061, + "step": 41140 + }, + { + "epoch": 6.391736564150357, + "grad_norm": 0.7385569214820862, + "learning_rate": 6.833087299213829e-05, + "loss": 0.0107, + "step": 41150 + }, + { + "epoch": 6.393289841565704, + "grad_norm": 0.17215046286582947, + "learning_rate": 6.83154885736225e-05, + "loss": 0.0072, + "step": 41160 + }, + { + "epoch": 6.39484311898105, + "grad_norm": 0.24251076579093933, + "learning_rate": 6.830010215215244e-05, + "loss": 0.0088, + "step": 41170 + }, + { + "epoch": 6.396396396396397, + "grad_norm": 0.27748754620552063, + "learning_rate": 6.828471372941072e-05, + "loss": 0.0116, + "step": 41180 + }, + { + "epoch": 6.397949673811743, + "grad_norm": 0.22673267126083374, + "learning_rate": 6.826932330708018e-05, + "loss": 0.0103, + "step": 41190 + }, + { + "epoch": 6.399502951227089, + "grad_norm": 0.33774110674858093, + "learning_rate": 6.825393088684393e-05, + "loss": 0.0062, + "step": 41200 + }, + { + "epoch": 6.401056228642435, + "grad_norm": 0.18939539790153503, + "learning_rate": 6.823853647038524e-05, + "loss": 0.0083, + "step": 41210 + }, + { + "epoch": 6.402609506057782, + "grad_norm": 0.19041737914085388, + "learning_rate": 6.822314005938762e-05, + "loss": 0.0088, + "step": 41220 + }, + { + "epoch": 6.4041627834731285, + "grad_norm": 0.12126143276691437, + "learning_rate": 6.820774165553479e-05, + "loss": 0.0084, + "step": 41230 + }, + { + "epoch": 6.405716060888475, + "grad_norm": 0.1853712946176529, + "learning_rate": 6.819234126051072e-05, + "loss": 0.0063, + "step": 41240 + }, + { + "epoch": 6.407269338303821, + "grad_norm": 0.3193918466567993, + "learning_rate": 6.817693887599956e-05, + "loss": 0.0095, + "step": 41250 + }, + { + "epoch": 6.408822615719167, + "grad_norm": 0.38374990224838257, + "learning_rate": 6.816153450368568e-05, + "loss": 0.01, + "step": 41260 + }, + { + "epoch": 6.410375893134514, + "grad_norm": 0.1915631890296936, + "learning_rate": 6.814612814525372e-05, + "loss": 0.0056, + "step": 41270 + }, + { + "epoch": 6.41192917054986, + "grad_norm": 0.18514592945575714, + "learning_rate": 6.813071980238846e-05, + "loss": 0.0076, + "step": 41280 + }, + { + "epoch": 6.413482447965206, + "grad_norm": 0.2878222167491913, + "learning_rate": 6.811530947677492e-05, + "loss": 0.0055, + "step": 41290 + }, + { + "epoch": 6.415035725380553, + "grad_norm": 0.2523154318332672, + "learning_rate": 6.809989717009839e-05, + "loss": 0.0073, + "step": 41300 + }, + { + "epoch": 6.4165890027959, + "grad_norm": 0.3569343686103821, + "learning_rate": 6.808448288404431e-05, + "loss": 0.007, + "step": 41310 + }, + { + "epoch": 6.418142280211246, + "grad_norm": 0.3206462264060974, + "learning_rate": 6.806906662029838e-05, + "loss": 0.0078, + "step": 41320 + }, + { + "epoch": 6.419695557626592, + "grad_norm": 0.5488213300704956, + "learning_rate": 6.805364838054647e-05, + "loss": 0.0073, + "step": 41330 + }, + { + "epoch": 6.421248835041938, + "grad_norm": 0.1676347702741623, + "learning_rate": 6.803822816647474e-05, + "loss": 0.007, + "step": 41340 + }, + { + "epoch": 6.422802112457285, + "grad_norm": 0.1555090844631195, + "learning_rate": 6.802280597976949e-05, + "loss": 0.0072, + "step": 41350 + }, + { + "epoch": 6.424355389872631, + "grad_norm": 0.2924245595932007, + "learning_rate": 6.800738182211727e-05, + "loss": 0.0078, + "step": 41360 + }, + { + "epoch": 6.4259086672879775, + "grad_norm": 0.32521820068359375, + "learning_rate": 6.799195569520485e-05, + "loss": 0.0076, + "step": 41370 + }, + { + "epoch": 6.427461944703324, + "grad_norm": 0.23809143900871277, + "learning_rate": 6.79765276007192e-05, + "loss": 0.0085, + "step": 41380 + }, + { + "epoch": 6.429015222118671, + "grad_norm": 0.8906163573265076, + "learning_rate": 6.796109754034753e-05, + "loss": 0.0088, + "step": 41390 + }, + { + "epoch": 6.430568499534017, + "grad_norm": 0.23675106465816498, + "learning_rate": 6.794566551577724e-05, + "loss": 0.0081, + "step": 41400 + }, + { + "epoch": 6.432121776949363, + "grad_norm": 0.8613208532333374, + "learning_rate": 6.793023152869594e-05, + "loss": 0.0096, + "step": 41410 + }, + { + "epoch": 6.433675054364709, + "grad_norm": 0.1729150414466858, + "learning_rate": 6.791479558079147e-05, + "loss": 0.0092, + "step": 41420 + }, + { + "epoch": 6.435228331780056, + "grad_norm": 0.18654818832874298, + "learning_rate": 6.789935767375191e-05, + "loss": 0.0103, + "step": 41430 + }, + { + "epoch": 6.436781609195402, + "grad_norm": 0.16405649483203888, + "learning_rate": 6.78839178092655e-05, + "loss": 0.0102, + "step": 41440 + }, + { + "epoch": 6.438334886610749, + "grad_norm": 0.16753481328487396, + "learning_rate": 6.786847598902072e-05, + "loss": 0.0102, + "step": 41450 + }, + { + "epoch": 6.439888164026095, + "grad_norm": 0.22520051896572113, + "learning_rate": 6.785303221470627e-05, + "loss": 0.0088, + "step": 41460 + }, + { + "epoch": 6.441441441441442, + "grad_norm": 0.1852322220802307, + "learning_rate": 6.783758648801108e-05, + "loss": 0.0077, + "step": 41470 + }, + { + "epoch": 6.442994718856788, + "grad_norm": 0.2049303650856018, + "learning_rate": 6.782213881062422e-05, + "loss": 0.0081, + "step": 41480 + }, + { + "epoch": 6.444547996272134, + "grad_norm": 0.25063467025756836, + "learning_rate": 6.780668918423508e-05, + "loss": 0.0093, + "step": 41490 + }, + { + "epoch": 6.44610127368748, + "grad_norm": 0.31803596019744873, + "learning_rate": 6.779123761053317e-05, + "loss": 0.0082, + "step": 41500 + }, + { + "epoch": 6.447654551102827, + "grad_norm": 0.36312010884284973, + "learning_rate": 6.777578409120826e-05, + "loss": 0.0083, + "step": 41510 + }, + { + "epoch": 6.4492078285181735, + "grad_norm": 0.1537889838218689, + "learning_rate": 6.776032862795033e-05, + "loss": 0.0059, + "step": 41520 + }, + { + "epoch": 6.45076110593352, + "grad_norm": 0.26353219151496887, + "learning_rate": 6.774487122244956e-05, + "loss": 0.0068, + "step": 41530 + }, + { + "epoch": 6.452314383348866, + "grad_norm": 0.1841905564069748, + "learning_rate": 6.772941187639637e-05, + "loss": 0.0065, + "step": 41540 + }, + { + "epoch": 6.453867660764212, + "grad_norm": 0.219265416264534, + "learning_rate": 6.771395059148134e-05, + "loss": 0.0078, + "step": 41550 + }, + { + "epoch": 6.455420938179559, + "grad_norm": 0.3510240316390991, + "learning_rate": 6.76984873693953e-05, + "loss": 0.0077, + "step": 41560 + }, + { + "epoch": 6.456974215594905, + "grad_norm": 0.5290544033050537, + "learning_rate": 6.768302221182931e-05, + "loss": 0.009, + "step": 41570 + }, + { + "epoch": 6.458527493010251, + "grad_norm": 0.2829506993293762, + "learning_rate": 6.766755512047457e-05, + "loss": 0.0072, + "step": 41580 + }, + { + "epoch": 6.4600807704255985, + "grad_norm": 0.5422075986862183, + "learning_rate": 6.765208609702259e-05, + "loss": 0.008, + "step": 41590 + }, + { + "epoch": 6.461634047840945, + "grad_norm": 0.25413277745246887, + "learning_rate": 6.763661514316499e-05, + "loss": 0.0107, + "step": 41600 + }, + { + "epoch": 6.463187325256291, + "grad_norm": 0.1643533557653427, + "learning_rate": 6.76211422605937e-05, + "loss": 0.0057, + "step": 41610 + }, + { + "epoch": 6.464740602671637, + "grad_norm": 0.22154292464256287, + "learning_rate": 6.760566745100076e-05, + "loss": 0.0088, + "step": 41620 + }, + { + "epoch": 6.466293880086983, + "grad_norm": 0.41555705666542053, + "learning_rate": 6.75901907160785e-05, + "loss": 0.0142, + "step": 41630 + }, + { + "epoch": 6.46784715750233, + "grad_norm": 0.3217713534832001, + "learning_rate": 6.757471205751943e-05, + "loss": 0.0073, + "step": 41640 + }, + { + "epoch": 6.469400434917676, + "grad_norm": 0.2175341248512268, + "learning_rate": 6.75592314770163e-05, + "loss": 0.0157, + "step": 41650 + }, + { + "epoch": 6.4709537123330225, + "grad_norm": 0.4219018816947937, + "learning_rate": 6.7543748976262e-05, + "loss": 0.0106, + "step": 41660 + }, + { + "epoch": 6.472506989748369, + "grad_norm": 0.3337368071079254, + "learning_rate": 6.752826455694968e-05, + "loss": 0.0071, + "step": 41670 + }, + { + "epoch": 6.474060267163716, + "grad_norm": 0.4645567834377289, + "learning_rate": 6.751277822077271e-05, + "loss": 0.0088, + "step": 41680 + }, + { + "epoch": 6.475613544579062, + "grad_norm": 0.21021844446659088, + "learning_rate": 6.749728996942465e-05, + "loss": 0.0121, + "step": 41690 + }, + { + "epoch": 6.477166821994408, + "grad_norm": 0.1584250032901764, + "learning_rate": 6.748179980459924e-05, + "loss": 0.0067, + "step": 41700 + }, + { + "epoch": 6.478720099409754, + "grad_norm": 0.16228994727134705, + "learning_rate": 6.746630772799052e-05, + "loss": 0.0094, + "step": 41710 + }, + { + "epoch": 6.480273376825101, + "grad_norm": 0.31698542833328247, + "learning_rate": 6.745081374129262e-05, + "loss": 0.0073, + "step": 41720 + }, + { + "epoch": 6.4818266542404475, + "grad_norm": 0.23641712963581085, + "learning_rate": 6.743531784619998e-05, + "loss": 0.0079, + "step": 41730 + }, + { + "epoch": 6.483379931655794, + "grad_norm": 0.2776964008808136, + "learning_rate": 6.741982004440719e-05, + "loss": 0.0076, + "step": 41740 + }, + { + "epoch": 6.48493320907114, + "grad_norm": 0.19963137805461884, + "learning_rate": 6.740432033760907e-05, + "loss": 0.0068, + "step": 41750 + }, + { + "epoch": 6.486486486486487, + "grad_norm": 0.1474626362323761, + "learning_rate": 6.738881872750066e-05, + "loss": 0.0051, + "step": 41760 + }, + { + "epoch": 6.488039763901833, + "grad_norm": 0.3000546395778656, + "learning_rate": 6.737331521577715e-05, + "loss": 0.0055, + "step": 41770 + }, + { + "epoch": 6.489593041317179, + "grad_norm": 0.2191782146692276, + "learning_rate": 6.735780980413403e-05, + "loss": 0.0075, + "step": 41780 + }, + { + "epoch": 6.491146318732525, + "grad_norm": 0.6637971997261047, + "learning_rate": 6.734230249426692e-05, + "loss": 0.0075, + "step": 41790 + }, + { + "epoch": 6.492699596147872, + "grad_norm": 0.49349385499954224, + "learning_rate": 6.732679328787168e-05, + "loss": 0.0075, + "step": 41800 + }, + { + "epoch": 6.494252873563219, + "grad_norm": 0.20672589540481567, + "learning_rate": 6.731128218664438e-05, + "loss": 0.0107, + "step": 41810 + }, + { + "epoch": 6.495806150978565, + "grad_norm": 0.1768510341644287, + "learning_rate": 6.729576919228129e-05, + "loss": 0.0088, + "step": 41820 + }, + { + "epoch": 6.497359428393911, + "grad_norm": 0.21165738999843597, + "learning_rate": 6.728025430647888e-05, + "loss": 0.0091, + "step": 41830 + }, + { + "epoch": 6.498912705809257, + "grad_norm": 0.1822022944688797, + "learning_rate": 6.726473753093383e-05, + "loss": 0.0088, + "step": 41840 + }, + { + "epoch": 6.500465983224604, + "grad_norm": 0.239594504237175, + "learning_rate": 6.724921886734305e-05, + "loss": 0.0079, + "step": 41850 + }, + { + "epoch": 6.50201926063995, + "grad_norm": 0.2139977067708969, + "learning_rate": 6.723369831740363e-05, + "loss": 0.0096, + "step": 41860 + }, + { + "epoch": 6.5035725380552964, + "grad_norm": 0.2688181698322296, + "learning_rate": 6.721817588281288e-05, + "loss": 0.0104, + "step": 41870 + }, + { + "epoch": 6.5051258154706435, + "grad_norm": 0.327818900346756, + "learning_rate": 6.720265156526828e-05, + "loss": 0.0073, + "step": 41880 + }, + { + "epoch": 6.50667909288599, + "grad_norm": 0.3329690396785736, + "learning_rate": 6.718712536646758e-05, + "loss": 0.0092, + "step": 41890 + }, + { + "epoch": 6.508232370301336, + "grad_norm": 0.2782459855079651, + "learning_rate": 6.71715972881087e-05, + "loss": 0.0085, + "step": 41900 + }, + { + "epoch": 6.509785647716682, + "grad_norm": 0.32273542881011963, + "learning_rate": 6.715606733188974e-05, + "loss": 0.007, + "step": 41910 + }, + { + "epoch": 6.511338925132028, + "grad_norm": 0.20098888874053955, + "learning_rate": 6.714053549950909e-05, + "loss": 0.0067, + "step": 41920 + }, + { + "epoch": 6.512892202547375, + "grad_norm": 0.3074928820133209, + "learning_rate": 6.71250017926652e-05, + "loss": 0.0112, + "step": 41930 + }, + { + "epoch": 6.514445479962721, + "grad_norm": 0.20673935115337372, + "learning_rate": 6.71094662130569e-05, + "loss": 0.0073, + "step": 41940 + }, + { + "epoch": 6.5159987573780676, + "grad_norm": 0.16904869675636292, + "learning_rate": 6.709392876238307e-05, + "loss": 0.0072, + "step": 41950 + }, + { + "epoch": 6.517552034793414, + "grad_norm": 0.31261497735977173, + "learning_rate": 6.707838944234289e-05, + "loss": 0.0097, + "step": 41960 + }, + { + "epoch": 6.519105312208761, + "grad_norm": 0.2996002733707428, + "learning_rate": 6.706284825463574e-05, + "loss": 0.0087, + "step": 41970 + }, + { + "epoch": 6.520658589624107, + "grad_norm": 0.24704314768314362, + "learning_rate": 6.704730520096114e-05, + "loss": 0.0096, + "step": 41980 + }, + { + "epoch": 6.522211867039453, + "grad_norm": 0.43126654624938965, + "learning_rate": 6.703176028301888e-05, + "loss": 0.0077, + "step": 41990 + }, + { + "epoch": 6.523765144454799, + "grad_norm": 0.3808981776237488, + "learning_rate": 6.701621350250892e-05, + "loss": 0.0088, + "step": 42000 + }, + { + "epoch": 6.525318421870146, + "grad_norm": 0.1665179580450058, + "learning_rate": 6.700066486113144e-05, + "loss": 0.0087, + "step": 42010 + }, + { + "epoch": 6.5268716992854925, + "grad_norm": 0.2742980420589447, + "learning_rate": 6.69851143605868e-05, + "loss": 0.0083, + "step": 42020 + }, + { + "epoch": 6.528424976700839, + "grad_norm": 0.20709915459156036, + "learning_rate": 6.696956200257561e-05, + "loss": 0.0101, + "step": 42030 + }, + { + "epoch": 6.529978254116185, + "grad_norm": 0.20782765746116638, + "learning_rate": 6.695400778879863e-05, + "loss": 0.008, + "step": 42040 + }, + { + "epoch": 6.531531531531532, + "grad_norm": 0.1910426765680313, + "learning_rate": 6.693845172095683e-05, + "loss": 0.007, + "step": 42050 + }, + { + "epoch": 6.533084808946878, + "grad_norm": 0.1242610439658165, + "learning_rate": 6.692289380075142e-05, + "loss": 0.0078, + "step": 42060 + }, + { + "epoch": 6.534638086362224, + "grad_norm": 0.2440660148859024, + "learning_rate": 6.690733402988379e-05, + "loss": 0.009, + "step": 42070 + }, + { + "epoch": 6.53619136377757, + "grad_norm": 0.30211517214775085, + "learning_rate": 6.689177241005553e-05, + "loss": 0.0081, + "step": 42080 + }, + { + "epoch": 6.537744641192917, + "grad_norm": 0.34693601727485657, + "learning_rate": 6.687620894296844e-05, + "loss": 0.006, + "step": 42090 + }, + { + "epoch": 6.539297918608264, + "grad_norm": 0.4385875165462494, + "learning_rate": 6.686064363032451e-05, + "loss": 0.0078, + "step": 42100 + }, + { + "epoch": 6.54085119602361, + "grad_norm": 0.26786714792251587, + "learning_rate": 6.684507647382596e-05, + "loss": 0.0094, + "step": 42110 + }, + { + "epoch": 6.542404473438956, + "grad_norm": 0.4199921190738678, + "learning_rate": 6.682950747517513e-05, + "loss": 0.0087, + "step": 42120 + }, + { + "epoch": 6.543957750854302, + "grad_norm": 0.22370867431163788, + "learning_rate": 6.68139366360747e-05, + "loss": 0.0075, + "step": 42130 + }, + { + "epoch": 6.545511028269649, + "grad_norm": 0.25220200419425964, + "learning_rate": 6.679836395822744e-05, + "loss": 0.0076, + "step": 42140 + }, + { + "epoch": 6.547064305684995, + "grad_norm": 0.45526808500289917, + "learning_rate": 6.678278944333633e-05, + "loss": 0.0114, + "step": 42150 + }, + { + "epoch": 6.5486175831003415, + "grad_norm": 0.16918189823627472, + "learning_rate": 6.676721309310462e-05, + "loss": 0.0067, + "step": 42160 + }, + { + "epoch": 6.5501708605156885, + "grad_norm": 0.21104249358177185, + "learning_rate": 6.67516349092357e-05, + "loss": 0.0085, + "step": 42170 + }, + { + "epoch": 6.551724137931035, + "grad_norm": 0.2919565737247467, + "learning_rate": 6.673605489343315e-05, + "loss": 0.0068, + "step": 42180 + }, + { + "epoch": 6.553277415346381, + "grad_norm": 0.26870256662368774, + "learning_rate": 6.672047304740081e-05, + "loss": 0.0066, + "step": 42190 + }, + { + "epoch": 6.554830692761727, + "grad_norm": 0.18606853485107422, + "learning_rate": 6.670488937284268e-05, + "loss": 0.0076, + "step": 42200 + }, + { + "epoch": 6.556383970177073, + "grad_norm": 0.36928486824035645, + "learning_rate": 6.668930387146294e-05, + "loss": 0.0092, + "step": 42210 + }, + { + "epoch": 6.55793724759242, + "grad_norm": 0.368457168340683, + "learning_rate": 6.667371654496605e-05, + "loss": 0.0095, + "step": 42220 + }, + { + "epoch": 6.559490525007766, + "grad_norm": 0.30247727036476135, + "learning_rate": 6.665812739505659e-05, + "loss": 0.0083, + "step": 42230 + }, + { + "epoch": 6.561043802423113, + "grad_norm": 0.18480481207370758, + "learning_rate": 6.664253642343935e-05, + "loss": 0.0062, + "step": 42240 + }, + { + "epoch": 6.56259707983846, + "grad_norm": 0.1750069409608841, + "learning_rate": 6.662694363181935e-05, + "loss": 0.0076, + "step": 42250 + }, + { + "epoch": 6.564150357253806, + "grad_norm": 0.35806840658187866, + "learning_rate": 6.66113490219018e-05, + "loss": 0.0077, + "step": 42260 + }, + { + "epoch": 6.565703634669152, + "grad_norm": 0.175254225730896, + "learning_rate": 6.659575259539207e-05, + "loss": 0.0059, + "step": 42270 + }, + { + "epoch": 6.567256912084498, + "grad_norm": 0.3720923364162445, + "learning_rate": 6.65801543539958e-05, + "loss": 0.0098, + "step": 42280 + }, + { + "epoch": 6.568810189499844, + "grad_norm": 0.3253500759601593, + "learning_rate": 6.65645542994188e-05, + "loss": 0.0073, + "step": 42290 + }, + { + "epoch": 6.570363466915191, + "grad_norm": 0.3075277805328369, + "learning_rate": 6.654895243336702e-05, + "loss": 0.0071, + "step": 42300 + }, + { + "epoch": 6.5719167443305375, + "grad_norm": 0.2633175551891327, + "learning_rate": 6.65333487575467e-05, + "loss": 0.0079, + "step": 42310 + }, + { + "epoch": 6.573470021745884, + "grad_norm": 0.19806928932666779, + "learning_rate": 6.651774327366422e-05, + "loss": 0.0085, + "step": 42320 + }, + { + "epoch": 6.57502329916123, + "grad_norm": 0.46415975689888, + "learning_rate": 6.650213598342613e-05, + "loss": 0.0091, + "step": 42330 + }, + { + "epoch": 6.576576576576577, + "grad_norm": 0.18172907829284668, + "learning_rate": 6.648652688853931e-05, + "loss": 0.0077, + "step": 42340 + }, + { + "epoch": 6.578129853991923, + "grad_norm": 0.16217881441116333, + "learning_rate": 6.647091599071066e-05, + "loss": 0.0069, + "step": 42350 + }, + { + "epoch": 6.579683131407269, + "grad_norm": 0.21053524315357208, + "learning_rate": 6.645530329164743e-05, + "loss": 0.0109, + "step": 42360 + }, + { + "epoch": 6.581236408822615, + "grad_norm": 0.2362072467803955, + "learning_rate": 6.643968879305694e-05, + "loss": 0.0069, + "step": 42370 + }, + { + "epoch": 6.5827896862379625, + "grad_norm": 0.40645095705986023, + "learning_rate": 6.642407249664683e-05, + "loss": 0.0076, + "step": 42380 + }, + { + "epoch": 6.584342963653309, + "grad_norm": 0.19761475920677185, + "learning_rate": 6.640845440412483e-05, + "loss": 0.0074, + "step": 42390 + }, + { + "epoch": 6.585896241068655, + "grad_norm": 0.14074921607971191, + "learning_rate": 6.639283451719893e-05, + "loss": 0.0071, + "step": 42400 + }, + { + "epoch": 6.587449518484001, + "grad_norm": 0.24026177823543549, + "learning_rate": 6.637721283757729e-05, + "loss": 0.0073, + "step": 42410 + }, + { + "epoch": 6.589002795899348, + "grad_norm": 0.28916457295417786, + "learning_rate": 6.636158936696828e-05, + "loss": 0.008, + "step": 42420 + }, + { + "epoch": 6.590556073314694, + "grad_norm": 0.17985352873802185, + "learning_rate": 6.634596410708047e-05, + "loss": 0.0089, + "step": 42430 + }, + { + "epoch": 6.59210935073004, + "grad_norm": 0.2767684757709503, + "learning_rate": 6.633033705962257e-05, + "loss": 0.0097, + "step": 42440 + }, + { + "epoch": 6.5936626281453865, + "grad_norm": 0.1512896567583084, + "learning_rate": 6.631470822630359e-05, + "loss": 0.0079, + "step": 42450 + }, + { + "epoch": 6.595215905560734, + "grad_norm": 0.10203372687101364, + "learning_rate": 6.629907760883263e-05, + "loss": 0.0068, + "step": 42460 + }, + { + "epoch": 6.59676918297608, + "grad_norm": 0.227084219455719, + "learning_rate": 6.628344520891905e-05, + "loss": 0.0082, + "step": 42470 + }, + { + "epoch": 6.598322460391426, + "grad_norm": 0.4221001863479614, + "learning_rate": 6.62678110282724e-05, + "loss": 0.0135, + "step": 42480 + }, + { + "epoch": 6.599875737806772, + "grad_norm": 0.28015264868736267, + "learning_rate": 6.625217506860237e-05, + "loss": 0.0094, + "step": 42490 + }, + { + "epoch": 6.601429015222118, + "grad_norm": 0.2596457302570343, + "learning_rate": 6.623653733161892e-05, + "loss": 0.0086, + "step": 42500 + }, + { + "epoch": 6.602982292637465, + "grad_norm": 0.2352682650089264, + "learning_rate": 6.622089781903217e-05, + "loss": 0.0094, + "step": 42510 + }, + { + "epoch": 6.604535570052811, + "grad_norm": 0.4229518473148346, + "learning_rate": 6.620525653255241e-05, + "loss": 0.0114, + "step": 42520 + }, + { + "epoch": 6.606088847468158, + "grad_norm": 0.35817816853523254, + "learning_rate": 6.618961347389015e-05, + "loss": 0.0085, + "step": 42530 + }, + { + "epoch": 6.607642124883505, + "grad_norm": 0.4679727256298065, + "learning_rate": 6.617396864475613e-05, + "loss": 0.008, + "step": 42540 + }, + { + "epoch": 6.609195402298851, + "grad_norm": 0.18740326166152954, + "learning_rate": 6.61583220468612e-05, + "loss": 0.0083, + "step": 42550 + }, + { + "epoch": 6.610748679714197, + "grad_norm": 0.2349025160074234, + "learning_rate": 6.614267368191645e-05, + "loss": 0.0075, + "step": 42560 + }, + { + "epoch": 6.612301957129543, + "grad_norm": 0.17570428550243378, + "learning_rate": 6.61270235516332e-05, + "loss": 0.0051, + "step": 42570 + }, + { + "epoch": 6.613855234544889, + "grad_norm": 0.14083634316921234, + "learning_rate": 6.611137165772287e-05, + "loss": 0.006, + "step": 42580 + }, + { + "epoch": 6.615408511960236, + "grad_norm": 0.23293106257915497, + "learning_rate": 6.60957180018972e-05, + "loss": 0.0075, + "step": 42590 + }, + { + "epoch": 6.6169617893755825, + "grad_norm": 0.18810835480690002, + "learning_rate": 6.608006258586797e-05, + "loss": 0.0064, + "step": 42600 + }, + { + "epoch": 6.618515066790929, + "grad_norm": 0.21863462030887604, + "learning_rate": 6.606440541134731e-05, + "loss": 0.0097, + "step": 42610 + }, + { + "epoch": 6.620068344206275, + "grad_norm": 0.22140170633792877, + "learning_rate": 6.60487464800474e-05, + "loss": 0.006, + "step": 42620 + }, + { + "epoch": 6.621621621621622, + "grad_norm": 0.16825026273727417, + "learning_rate": 6.603308579368071e-05, + "loss": 0.0048, + "step": 42630 + }, + { + "epoch": 6.623174899036968, + "grad_norm": 0.2997620403766632, + "learning_rate": 6.601742335395987e-05, + "loss": 0.0093, + "step": 42640 + }, + { + "epoch": 6.624728176452314, + "grad_norm": 0.2214311957359314, + "learning_rate": 6.600175916259769e-05, + "loss": 0.0064, + "step": 42650 + }, + { + "epoch": 6.62628145386766, + "grad_norm": 0.5860468745231628, + "learning_rate": 6.598609322130718e-05, + "loss": 0.0076, + "step": 42660 + }, + { + "epoch": 6.6278347312830075, + "grad_norm": 0.3202514350414276, + "learning_rate": 6.597042553180155e-05, + "loss": 0.0066, + "step": 42670 + }, + { + "epoch": 6.629388008698354, + "grad_norm": 0.2191588580608368, + "learning_rate": 6.59547560957942e-05, + "loss": 0.0067, + "step": 42680 + }, + { + "epoch": 6.6309412861137, + "grad_norm": 0.1904541701078415, + "learning_rate": 6.59390849149987e-05, + "loss": 0.0082, + "step": 42690 + }, + { + "epoch": 6.632494563529046, + "grad_norm": 0.14396189153194427, + "learning_rate": 6.592341199112886e-05, + "loss": 0.0077, + "step": 42700 + }, + { + "epoch": 6.634047840944393, + "grad_norm": 0.15084804594516754, + "learning_rate": 6.590773732589861e-05, + "loss": 0.0066, + "step": 42710 + }, + { + "epoch": 6.635601118359739, + "grad_norm": 0.32585129141807556, + "learning_rate": 6.589206092102213e-05, + "loss": 0.0089, + "step": 42720 + }, + { + "epoch": 6.637154395775085, + "grad_norm": 0.2860310673713684, + "learning_rate": 6.587638277821375e-05, + "loss": 0.0074, + "step": 42730 + }, + { + "epoch": 6.6387076731904315, + "grad_norm": 0.6986554265022278, + "learning_rate": 6.586070289918803e-05, + "loss": 0.0099, + "step": 42740 + }, + { + "epoch": 6.640260950605779, + "grad_norm": 0.2706029415130615, + "learning_rate": 6.584502128565968e-05, + "loss": 0.0063, + "step": 42750 + }, + { + "epoch": 6.641814228021125, + "grad_norm": 0.2828982472419739, + "learning_rate": 6.582933793934363e-05, + "loss": 0.0076, + "step": 42760 + }, + { + "epoch": 6.643367505436471, + "grad_norm": 0.17506767809391022, + "learning_rate": 6.581365286195498e-05, + "loss": 0.0053, + "step": 42770 + }, + { + "epoch": 6.644920782851817, + "grad_norm": 0.4393887221813202, + "learning_rate": 6.579796605520903e-05, + "loss": 0.0065, + "step": 42780 + }, + { + "epoch": 6.646474060267163, + "grad_norm": 0.14839956164360046, + "learning_rate": 6.578227752082127e-05, + "loss": 0.0069, + "step": 42790 + }, + { + "epoch": 6.64802733768251, + "grad_norm": 0.19315187633037567, + "learning_rate": 6.576658726050735e-05, + "loss": 0.0074, + "step": 42800 + }, + { + "epoch": 6.6495806150978565, + "grad_norm": 0.32658275961875916, + "learning_rate": 6.575089527598316e-05, + "loss": 0.0066, + "step": 42810 + }, + { + "epoch": 6.651133892513203, + "grad_norm": 0.3413253426551819, + "learning_rate": 6.573520156896476e-05, + "loss": 0.0078, + "step": 42820 + }, + { + "epoch": 6.65268716992855, + "grad_norm": 0.2907262146472931, + "learning_rate": 6.571950614116835e-05, + "loss": 0.0057, + "step": 42830 + }, + { + "epoch": 6.654240447343896, + "grad_norm": 0.13729390501976013, + "learning_rate": 6.570380899431039e-05, + "loss": 0.0113, + "step": 42840 + }, + { + "epoch": 6.655793724759242, + "grad_norm": 0.20574983954429626, + "learning_rate": 6.568811013010749e-05, + "loss": 0.0081, + "step": 42850 + }, + { + "epoch": 6.657347002174588, + "grad_norm": 0.2682510316371918, + "learning_rate": 6.567240955027645e-05, + "loss": 0.0106, + "step": 42860 + }, + { + "epoch": 6.658900279589934, + "grad_norm": 0.3853456377983093, + "learning_rate": 6.565670725653427e-05, + "loss": 0.0099, + "step": 42870 + }, + { + "epoch": 6.660453557005281, + "grad_norm": 0.1477942019701004, + "learning_rate": 6.564100325059812e-05, + "loss": 0.0075, + "step": 42880 + }, + { + "epoch": 6.662006834420628, + "grad_norm": 0.1579279899597168, + "learning_rate": 6.562529753418538e-05, + "loss": 0.0059, + "step": 42890 + }, + { + "epoch": 6.663560111835974, + "grad_norm": 0.17501935362815857, + "learning_rate": 6.56095901090136e-05, + "loss": 0.0077, + "step": 42900 + }, + { + "epoch": 6.66511338925132, + "grad_norm": 0.20286250114440918, + "learning_rate": 6.55938809768005e-05, + "loss": 0.0081, + "step": 42910 + }, + { + "epoch": 6.666666666666667, + "grad_norm": 0.29659730195999146, + "learning_rate": 6.557817013926402e-05, + "loss": 0.008, + "step": 42920 + }, + { + "epoch": 6.668219944082013, + "grad_norm": 0.17817477881908417, + "learning_rate": 6.556245759812227e-05, + "loss": 0.0061, + "step": 42930 + }, + { + "epoch": 6.669773221497359, + "grad_norm": 0.22212260961532593, + "learning_rate": 6.554674335509357e-05, + "loss": 0.0078, + "step": 42940 + }, + { + "epoch": 6.6713264989127055, + "grad_norm": 0.25289440155029297, + "learning_rate": 6.553102741189638e-05, + "loss": 0.0083, + "step": 42950 + }, + { + "epoch": 6.6728797763280525, + "grad_norm": 0.2557927370071411, + "learning_rate": 6.55153097702494e-05, + "loss": 0.0076, + "step": 42960 + }, + { + "epoch": 6.674433053743399, + "grad_norm": 0.1722823977470398, + "learning_rate": 6.549959043187146e-05, + "loss": 0.0064, + "step": 42970 + }, + { + "epoch": 6.675986331158745, + "grad_norm": 0.25712960958480835, + "learning_rate": 6.54838693984816e-05, + "loss": 0.0072, + "step": 42980 + }, + { + "epoch": 6.677539608574091, + "grad_norm": 0.18081742525100708, + "learning_rate": 6.546814667179909e-05, + "loss": 0.0055, + "step": 42990 + }, + { + "epoch": 6.679092885989438, + "grad_norm": 0.6677728891372681, + "learning_rate": 6.545242225354328e-05, + "loss": 0.0068, + "step": 43000 + }, + { + "epoch": 6.680646163404784, + "grad_norm": 0.22107382118701935, + "learning_rate": 6.543669614543383e-05, + "loss": 0.0076, + "step": 43010 + }, + { + "epoch": 6.68219944082013, + "grad_norm": 0.1837356984615326, + "learning_rate": 6.542096834919049e-05, + "loss": 0.0073, + "step": 43020 + }, + { + "epoch": 6.683752718235477, + "grad_norm": 0.28177785873413086, + "learning_rate": 6.540523886653324e-05, + "loss": 0.0082, + "step": 43030 + }, + { + "epoch": 6.685305995650824, + "grad_norm": 0.22667533159255981, + "learning_rate": 6.538950769918222e-05, + "loss": 0.0096, + "step": 43040 + }, + { + "epoch": 6.68685927306617, + "grad_norm": 0.16792629659175873, + "learning_rate": 6.537377484885779e-05, + "loss": 0.0122, + "step": 43050 + }, + { + "epoch": 6.688412550481516, + "grad_norm": 0.28435951471328735, + "learning_rate": 6.535804031728045e-05, + "loss": 0.0074, + "step": 43060 + }, + { + "epoch": 6.689965827896862, + "grad_norm": 0.342720091342926, + "learning_rate": 6.534230410617092e-05, + "loss": 0.0099, + "step": 43070 + }, + { + "epoch": 6.691519105312208, + "grad_norm": 0.20052389800548553, + "learning_rate": 6.532656621725007e-05, + "loss": 0.0061, + "step": 43080 + }, + { + "epoch": 6.693072382727555, + "grad_norm": 0.17431248724460602, + "learning_rate": 6.531082665223898e-05, + "loss": 0.0072, + "step": 43090 + }, + { + "epoch": 6.6946256601429015, + "grad_norm": 0.17849598824977875, + "learning_rate": 6.529508541285889e-05, + "loss": 0.0078, + "step": 43100 + }, + { + "epoch": 6.696178937558248, + "grad_norm": 0.24768488109111786, + "learning_rate": 6.527934250083128e-05, + "loss": 0.0069, + "step": 43110 + }, + { + "epoch": 6.697732214973595, + "grad_norm": 0.20498616993427277, + "learning_rate": 6.526359791787775e-05, + "loss": 0.0088, + "step": 43120 + }, + { + "epoch": 6.699285492388941, + "grad_norm": 0.29540812969207764, + "learning_rate": 6.524785166572009e-05, + "loss": 0.0072, + "step": 43130 + }, + { + "epoch": 6.700838769804287, + "grad_norm": 0.35509201884269714, + "learning_rate": 6.523210374608029e-05, + "loss": 0.0069, + "step": 43140 + }, + { + "epoch": 6.702392047219633, + "grad_norm": 0.32230427861213684, + "learning_rate": 6.521635416068054e-05, + "loss": 0.0098, + "step": 43150 + }, + { + "epoch": 6.703945324634979, + "grad_norm": 0.8685961961746216, + "learning_rate": 6.520060291124317e-05, + "loss": 0.0097, + "step": 43160 + }, + { + "epoch": 6.705498602050326, + "grad_norm": 0.3171856999397278, + "learning_rate": 6.518484999949072e-05, + "loss": 0.0081, + "step": 43170 + }, + { + "epoch": 6.707051879465673, + "grad_norm": 0.16336914896965027, + "learning_rate": 6.516909542714591e-05, + "loss": 0.007, + "step": 43180 + }, + { + "epoch": 6.708605156881019, + "grad_norm": 0.19646379351615906, + "learning_rate": 6.515333919593161e-05, + "loss": 0.0061, + "step": 43190 + }, + { + "epoch": 6.710158434296366, + "grad_norm": 0.219664067029953, + "learning_rate": 6.513758130757094e-05, + "loss": 0.0088, + "step": 43200 + }, + { + "epoch": 6.711711711711712, + "grad_norm": 0.16610091924667358, + "learning_rate": 6.512182176378713e-05, + "loss": 0.0096, + "step": 43210 + }, + { + "epoch": 6.713264989127058, + "grad_norm": 0.35652995109558105, + "learning_rate": 6.510606056630362e-05, + "loss": 0.008, + "step": 43220 + }, + { + "epoch": 6.714818266542404, + "grad_norm": 0.3615867495536804, + "learning_rate": 6.509029771684405e-05, + "loss": 0.0099, + "step": 43230 + }, + { + "epoch": 6.7163715439577505, + "grad_norm": 0.16696631908416748, + "learning_rate": 6.50745332171322e-05, + "loss": 0.007, + "step": 43240 + }, + { + "epoch": 6.7179248213730975, + "grad_norm": 0.340483695268631, + "learning_rate": 6.505876706889207e-05, + "loss": 0.0071, + "step": 43250 + }, + { + "epoch": 6.719478098788444, + "grad_norm": 0.16322201490402222, + "learning_rate": 6.504299927384781e-05, + "loss": 0.0071, + "step": 43260 + }, + { + "epoch": 6.72103137620379, + "grad_norm": 0.21279896795749664, + "learning_rate": 6.502722983372378e-05, + "loss": 0.006, + "step": 43270 + }, + { + "epoch": 6.722584653619136, + "grad_norm": 0.25543755292892456, + "learning_rate": 6.501145875024446e-05, + "loss": 0.0088, + "step": 43280 + }, + { + "epoch": 6.724137931034483, + "grad_norm": 0.2607879042625427, + "learning_rate": 6.499568602513459e-05, + "loss": 0.0084, + "step": 43290 + }, + { + "epoch": 6.725691208449829, + "grad_norm": 0.24453945457935333, + "learning_rate": 6.497991166011903e-05, + "loss": 0.0075, + "step": 43300 + }, + { + "epoch": 6.727244485865175, + "grad_norm": 0.18509964644908905, + "learning_rate": 6.49641356569229e-05, + "loss": 0.0067, + "step": 43310 + }, + { + "epoch": 6.728797763280522, + "grad_norm": 0.17002864181995392, + "learning_rate": 6.494835801727135e-05, + "loss": 0.0059, + "step": 43320 + }, + { + "epoch": 6.730351040695869, + "grad_norm": 0.21421144902706146, + "learning_rate": 6.493257874288986e-05, + "loss": 0.0065, + "step": 43330 + }, + { + "epoch": 6.731904318111215, + "grad_norm": 0.4630519449710846, + "learning_rate": 6.491679783550399e-05, + "loss": 0.0104, + "step": 43340 + }, + { + "epoch": 6.733457595526561, + "grad_norm": 0.3731827139854431, + "learning_rate": 6.490101529683957e-05, + "loss": 0.0096, + "step": 43350 + }, + { + "epoch": 6.735010872941907, + "grad_norm": 0.29731863737106323, + "learning_rate": 6.488523112862248e-05, + "loss": 0.0072, + "step": 43360 + }, + { + "epoch": 6.736564150357253, + "grad_norm": 0.30611881613731384, + "learning_rate": 6.486944533257891e-05, + "loss": 0.0069, + "step": 43370 + }, + { + "epoch": 6.7381174277726, + "grad_norm": 0.2433919906616211, + "learning_rate": 6.485365791043517e-05, + "loss": 0.0068, + "step": 43380 + }, + { + "epoch": 6.7396707051879465, + "grad_norm": 0.27975308895111084, + "learning_rate": 6.483786886391772e-05, + "loss": 0.0098, + "step": 43390 + }, + { + "epoch": 6.741223982603293, + "grad_norm": 0.22540509700775146, + "learning_rate": 6.482207819475323e-05, + "loss": 0.0104, + "step": 43400 + }, + { + "epoch": 6.74277726001864, + "grad_norm": 0.28444066643714905, + "learning_rate": 6.480628590466857e-05, + "loss": 0.01, + "step": 43410 + }, + { + "epoch": 6.744330537433986, + "grad_norm": 0.3939913511276245, + "learning_rate": 6.479049199539074e-05, + "loss": 0.0055, + "step": 43420 + }, + { + "epoch": 6.745883814849332, + "grad_norm": 0.22602471709251404, + "learning_rate": 6.477469646864693e-05, + "loss": 0.009, + "step": 43430 + }, + { + "epoch": 6.747437092264678, + "grad_norm": 0.17964260280132294, + "learning_rate": 6.475889932616454e-05, + "loss": 0.0108, + "step": 43440 + }, + { + "epoch": 6.748990369680024, + "grad_norm": 0.30708491802215576, + "learning_rate": 6.474310056967111e-05, + "loss": 0.0068, + "step": 43450 + }, + { + "epoch": 6.7505436470953715, + "grad_norm": 0.151020810008049, + "learning_rate": 6.472730020089437e-05, + "loss": 0.0065, + "step": 43460 + }, + { + "epoch": 6.752096924510718, + "grad_norm": 0.1343528777360916, + "learning_rate": 6.471149822156222e-05, + "loss": 0.0054, + "step": 43470 + }, + { + "epoch": 6.753650201926064, + "grad_norm": 0.15864969789981842, + "learning_rate": 6.469569463340275e-05, + "loss": 0.0058, + "step": 43480 + }, + { + "epoch": 6.755203479341411, + "grad_norm": 0.1541222184896469, + "learning_rate": 6.467988943814418e-05, + "loss": 0.0108, + "step": 43490 + }, + { + "epoch": 6.756756756756757, + "grad_norm": 0.2998427748680115, + "learning_rate": 6.4664082637515e-05, + "loss": 0.0089, + "step": 43500 + }, + { + "epoch": 6.758310034172103, + "grad_norm": 0.21896415948867798, + "learning_rate": 6.464827423324377e-05, + "loss": 0.0086, + "step": 43510 + }, + { + "epoch": 6.759863311587449, + "grad_norm": 0.21199767291545868, + "learning_rate": 6.463246422705931e-05, + "loss": 0.0089, + "step": 43520 + }, + { + "epoch": 6.7614165890027955, + "grad_norm": 0.4208787977695465, + "learning_rate": 6.461665262069055e-05, + "loss": 0.0075, + "step": 43530 + }, + { + "epoch": 6.762969866418143, + "grad_norm": 0.2283429354429245, + "learning_rate": 6.460083941586665e-05, + "loss": 0.0071, + "step": 43540 + }, + { + "epoch": 6.764523143833489, + "grad_norm": 0.27575525641441345, + "learning_rate": 6.45850246143169e-05, + "loss": 0.0069, + "step": 43550 + }, + { + "epoch": 6.766076421248835, + "grad_norm": 0.15802860260009766, + "learning_rate": 6.456920821777077e-05, + "loss": 0.0102, + "step": 43560 + }, + { + "epoch": 6.767629698664181, + "grad_norm": 0.31602710485458374, + "learning_rate": 6.455339022795796e-05, + "loss": 0.0078, + "step": 43570 + }, + { + "epoch": 6.769182976079528, + "grad_norm": 0.5620087385177612, + "learning_rate": 6.453757064660825e-05, + "loss": 0.0071, + "step": 43580 + }, + { + "epoch": 6.770736253494874, + "grad_norm": 0.20485340058803558, + "learning_rate": 6.452174947545169e-05, + "loss": 0.0072, + "step": 43590 + }, + { + "epoch": 6.7722895309102205, + "grad_norm": 0.2071635127067566, + "learning_rate": 6.450592671621842e-05, + "loss": 0.0076, + "step": 43600 + }, + { + "epoch": 6.773842808325567, + "grad_norm": 0.21034805476665497, + "learning_rate": 6.449010237063882e-05, + "loss": 0.0064, + "step": 43610 + }, + { + "epoch": 6.775396085740914, + "grad_norm": 0.2306043952703476, + "learning_rate": 6.447427644044343e-05, + "loss": 0.0065, + "step": 43620 + }, + { + "epoch": 6.77694936315626, + "grad_norm": 0.23089458048343658, + "learning_rate": 6.44584489273629e-05, + "loss": 0.0082, + "step": 43630 + }, + { + "epoch": 6.778502640571606, + "grad_norm": 0.4431998133659363, + "learning_rate": 6.444261983312817e-05, + "loss": 0.0062, + "step": 43640 + }, + { + "epoch": 6.780055917986952, + "grad_norm": 0.7090445756912231, + "learning_rate": 6.442678915947023e-05, + "loss": 0.0079, + "step": 43650 + }, + { + "epoch": 6.781609195402299, + "grad_norm": 0.29249587655067444, + "learning_rate": 6.441095690812032e-05, + "loss": 0.0073, + "step": 43660 + }, + { + "epoch": 6.783162472817645, + "grad_norm": 0.1509740948677063, + "learning_rate": 6.439512308080982e-05, + "loss": 0.007, + "step": 43670 + }, + { + "epoch": 6.784715750232992, + "grad_norm": 0.3094101846218109, + "learning_rate": 6.43792876792703e-05, + "loss": 0.0081, + "step": 43680 + }, + { + "epoch": 6.786269027648338, + "grad_norm": 0.3569844365119934, + "learning_rate": 6.43634507052335e-05, + "loss": 0.0085, + "step": 43690 + }, + { + "epoch": 6.787822305063685, + "grad_norm": 0.22333839535713196, + "learning_rate": 6.434761216043135e-05, + "loss": 0.0081, + "step": 43700 + }, + { + "epoch": 6.789375582479031, + "grad_norm": 0.25249722599983215, + "learning_rate": 6.433177204659589e-05, + "loss": 0.0092, + "step": 43710 + }, + { + "epoch": 6.790928859894377, + "grad_norm": 0.6781355142593384, + "learning_rate": 6.431593036545938e-05, + "loss": 0.0085, + "step": 43720 + }, + { + "epoch": 6.792482137309723, + "grad_norm": 0.26152998208999634, + "learning_rate": 6.430008711875426e-05, + "loss": 0.0112, + "step": 43730 + }, + { + "epoch": 6.794035414725069, + "grad_norm": 0.2160796821117401, + "learning_rate": 6.42842423082131e-05, + "loss": 0.0057, + "step": 43740 + }, + { + "epoch": 6.7955886921404165, + "grad_norm": 0.24803796410560608, + "learning_rate": 6.426839593556868e-05, + "loss": 0.0096, + "step": 43750 + }, + { + "epoch": 6.797141969555763, + "grad_norm": 0.22874444723129272, + "learning_rate": 6.425254800255392e-05, + "loss": 0.0081, + "step": 43760 + }, + { + "epoch": 6.798695246971109, + "grad_norm": 0.1766066551208496, + "learning_rate": 6.423669851090196e-05, + "loss": 0.0088, + "step": 43770 + }, + { + "epoch": 6.800248524386456, + "grad_norm": 0.24644456803798676, + "learning_rate": 6.422084746234604e-05, + "loss": 0.0067, + "step": 43780 + }, + { + "epoch": 6.801801801801802, + "grad_norm": 0.17411130666732788, + "learning_rate": 6.420499485861965e-05, + "loss": 0.0085, + "step": 43790 + }, + { + "epoch": 6.803355079217148, + "grad_norm": 0.17181865870952606, + "learning_rate": 6.418914070145636e-05, + "loss": 0.0056, + "step": 43800 + }, + { + "epoch": 6.804908356632494, + "grad_norm": 0.4297802150249481, + "learning_rate": 6.417328499258996e-05, + "loss": 0.0111, + "step": 43810 + }, + { + "epoch": 6.8064616340478405, + "grad_norm": 0.22092312574386597, + "learning_rate": 6.415742773375446e-05, + "loss": 0.008, + "step": 43820 + }, + { + "epoch": 6.808014911463188, + "grad_norm": 0.23837880790233612, + "learning_rate": 6.414156892668393e-05, + "loss": 0.0084, + "step": 43830 + }, + { + "epoch": 6.809568188878534, + "grad_norm": 1.2018449306488037, + "learning_rate": 6.412570857311267e-05, + "loss": 0.0122, + "step": 43840 + }, + { + "epoch": 6.81112146629388, + "grad_norm": 0.43195033073425293, + "learning_rate": 6.410984667477518e-05, + "loss": 0.0093, + "step": 43850 + }, + { + "epoch": 6.812674743709226, + "grad_norm": 0.4664139151573181, + "learning_rate": 6.409398323340607e-05, + "loss": 0.0095, + "step": 43860 + }, + { + "epoch": 6.814228021124573, + "grad_norm": 0.2560790479183197, + "learning_rate": 6.407811825074012e-05, + "loss": 0.009, + "step": 43870 + }, + { + "epoch": 6.815781298539919, + "grad_norm": 0.22524012625217438, + "learning_rate": 6.406225172851234e-05, + "loss": 0.007, + "step": 43880 + }, + { + "epoch": 6.8173345759552655, + "grad_norm": 0.16927413642406464, + "learning_rate": 6.404638366845786e-05, + "loss": 0.0093, + "step": 43890 + }, + { + "epoch": 6.818887853370612, + "grad_norm": 0.24167172610759735, + "learning_rate": 6.403051407231196e-05, + "loss": 0.0075, + "step": 43900 + }, + { + "epoch": 6.820441130785959, + "grad_norm": 0.15159372985363007, + "learning_rate": 6.401464294181016e-05, + "loss": 0.0071, + "step": 43910 + }, + { + "epoch": 6.821994408201305, + "grad_norm": 0.25685617327690125, + "learning_rate": 6.399877027868808e-05, + "loss": 0.0071, + "step": 43920 + }, + { + "epoch": 6.823547685616651, + "grad_norm": 0.2250175178050995, + "learning_rate": 6.39828960846815e-05, + "loss": 0.0095, + "step": 43930 + }, + { + "epoch": 6.825100963031997, + "grad_norm": 0.5383036732673645, + "learning_rate": 6.396702036152644e-05, + "loss": 0.0101, + "step": 43940 + }, + { + "epoch": 6.826654240447344, + "grad_norm": 0.26898160576820374, + "learning_rate": 6.395114311095905e-05, + "loss": 0.0093, + "step": 43950 + }, + { + "epoch": 6.82820751786269, + "grad_norm": 0.22481146454811096, + "learning_rate": 6.393526433471559e-05, + "loss": 0.0066, + "step": 43960 + }, + { + "epoch": 6.829760795278037, + "grad_norm": 0.3420772850513458, + "learning_rate": 6.391938403453262e-05, + "loss": 0.0088, + "step": 43970 + }, + { + "epoch": 6.831314072693383, + "grad_norm": 0.4395061135292053, + "learning_rate": 6.390350221214671e-05, + "loss": 0.0096, + "step": 43980 + }, + { + "epoch": 6.83286735010873, + "grad_norm": 0.4098595976829529, + "learning_rate": 6.38876188692947e-05, + "loss": 0.009, + "step": 43990 + }, + { + "epoch": 6.834420627524076, + "grad_norm": 0.4749778211116791, + "learning_rate": 6.387173400771358e-05, + "loss": 0.0095, + "step": 44000 + }, + { + "epoch": 6.835973904939422, + "grad_norm": 0.18496280908584595, + "learning_rate": 6.385584762914047e-05, + "loss": 0.0053, + "step": 44010 + }, + { + "epoch": 6.837527182354768, + "grad_norm": 0.31623193621635437, + "learning_rate": 6.38399597353127e-05, + "loss": 0.0105, + "step": 44020 + }, + { + "epoch": 6.8390804597701145, + "grad_norm": 0.3960568308830261, + "learning_rate": 6.382407032796775e-05, + "loss": 0.0073, + "step": 44030 + }, + { + "epoch": 6.8406337371854615, + "grad_norm": 0.1628400832414627, + "learning_rate": 6.380817940884325e-05, + "loss": 0.007, + "step": 44040 + }, + { + "epoch": 6.842187014600808, + "grad_norm": 0.26025447249412537, + "learning_rate": 6.379228697967702e-05, + "loss": 0.0077, + "step": 44050 + }, + { + "epoch": 6.843740292016154, + "grad_norm": 0.2702217698097229, + "learning_rate": 6.3776393042207e-05, + "loss": 0.0074, + "step": 44060 + }, + { + "epoch": 6.845293569431501, + "grad_norm": 0.23868729174137115, + "learning_rate": 6.376049759817137e-05, + "loss": 0.006, + "step": 44070 + }, + { + "epoch": 6.846846846846847, + "grad_norm": 0.14615929126739502, + "learning_rate": 6.37446006493084e-05, + "loss": 0.0062, + "step": 44080 + }, + { + "epoch": 6.848400124262193, + "grad_norm": 0.25579017400741577, + "learning_rate": 6.372870219735656e-05, + "loss": 0.0066, + "step": 44090 + }, + { + "epoch": 6.849953401677539, + "grad_norm": 0.24001002311706543, + "learning_rate": 6.371280224405452e-05, + "loss": 0.0075, + "step": 44100 + }, + { + "epoch": 6.851506679092886, + "grad_norm": 0.34297803044319153, + "learning_rate": 6.369690079114101e-05, + "loss": 0.0092, + "step": 44110 + }, + { + "epoch": 6.853059956508233, + "grad_norm": 0.21491600573062897, + "learning_rate": 6.368099784035504e-05, + "loss": 0.0057, + "step": 44120 + }, + { + "epoch": 6.854613233923579, + "grad_norm": 0.2866162955760956, + "learning_rate": 6.366509339343572e-05, + "loss": 0.0063, + "step": 44130 + }, + { + "epoch": 6.856166511338925, + "grad_norm": 0.2267405390739441, + "learning_rate": 6.364918745212232e-05, + "loss": 0.012, + "step": 44140 + }, + { + "epoch": 6.857719788754271, + "grad_norm": 0.6299105882644653, + "learning_rate": 6.363328001815431e-05, + "loss": 0.0084, + "step": 44150 + }, + { + "epoch": 6.859273066169618, + "grad_norm": 0.24072401225566864, + "learning_rate": 6.361737109327128e-05, + "loss": 0.0078, + "step": 44160 + }, + { + "epoch": 6.860826343584964, + "grad_norm": 0.21790984272956848, + "learning_rate": 6.360146067921301e-05, + "loss": 0.0066, + "step": 44170 + }, + { + "epoch": 6.8623796210003105, + "grad_norm": 0.2112446129322052, + "learning_rate": 6.35855487777195e-05, + "loss": 0.0107, + "step": 44180 + }, + { + "epoch": 6.863932898415657, + "grad_norm": 0.19025085866451263, + "learning_rate": 6.356963539053076e-05, + "loss": 0.0085, + "step": 44190 + }, + { + "epoch": 6.865486175831004, + "grad_norm": 0.24340879917144775, + "learning_rate": 6.355372051938711e-05, + "loss": 0.0075, + "step": 44200 + }, + { + "epoch": 6.86703945324635, + "grad_norm": 0.1737350970506668, + "learning_rate": 6.353780416602894e-05, + "loss": 0.0055, + "step": 44210 + }, + { + "epoch": 6.868592730661696, + "grad_norm": 0.20628279447555542, + "learning_rate": 6.352188633219689e-05, + "loss": 0.009, + "step": 44220 + }, + { + "epoch": 6.870146008077042, + "grad_norm": 0.1497402936220169, + "learning_rate": 6.350596701963166e-05, + "loss": 0.0055, + "step": 44230 + }, + { + "epoch": 6.871699285492389, + "grad_norm": 0.5578227639198303, + "learning_rate": 6.349004623007419e-05, + "loss": 0.0078, + "step": 44240 + }, + { + "epoch": 6.8732525629077355, + "grad_norm": 0.2465231716632843, + "learning_rate": 6.347412396526555e-05, + "loss": 0.0087, + "step": 44250 + }, + { + "epoch": 6.874805840323082, + "grad_norm": 0.4471524655818939, + "learning_rate": 6.345820022694696e-05, + "loss": 0.0124, + "step": 44260 + }, + { + "epoch": 6.876359117738428, + "grad_norm": 0.22777418792247772, + "learning_rate": 6.344227501685984e-05, + "loss": 0.0055, + "step": 44270 + }, + { + "epoch": 6.877912395153775, + "grad_norm": 0.3474384546279907, + "learning_rate": 6.342634833674572e-05, + "loss": 0.0081, + "step": 44280 + }, + { + "epoch": 6.879465672569121, + "grad_norm": 0.39094746112823486, + "learning_rate": 6.341042018834635e-05, + "loss": 0.0082, + "step": 44290 + }, + { + "epoch": 6.881018949984467, + "grad_norm": 0.3378508985042572, + "learning_rate": 6.339449057340359e-05, + "loss": 0.0093, + "step": 44300 + }, + { + "epoch": 6.882572227399813, + "grad_norm": 0.21885977685451508, + "learning_rate": 6.337855949365945e-05, + "loss": 0.0087, + "step": 44310 + }, + { + "epoch": 6.8841255048151595, + "grad_norm": 0.14768721163272858, + "learning_rate": 6.336262695085619e-05, + "loss": 0.0101, + "step": 44320 + }, + { + "epoch": 6.885678782230507, + "grad_norm": 0.2068590372800827, + "learning_rate": 6.334669294673612e-05, + "loss": 0.0071, + "step": 44330 + }, + { + "epoch": 6.887232059645853, + "grad_norm": 0.3744548261165619, + "learning_rate": 6.333075748304179e-05, + "loss": 0.0067, + "step": 44340 + }, + { + "epoch": 6.888785337061199, + "grad_norm": 0.19056546688079834, + "learning_rate": 6.331482056151585e-05, + "loss": 0.0101, + "step": 44350 + }, + { + "epoch": 6.890338614476546, + "grad_norm": 0.2730405926704407, + "learning_rate": 6.329888218390117e-05, + "loss": 0.0087, + "step": 44360 + }, + { + "epoch": 6.891891891891892, + "grad_norm": 0.422213077545166, + "learning_rate": 6.328294235194072e-05, + "loss": 0.007, + "step": 44370 + }, + { + "epoch": 6.893445169307238, + "grad_norm": 0.24543693661689758, + "learning_rate": 6.326700106737765e-05, + "loss": 0.007, + "step": 44380 + }, + { + "epoch": 6.894998446722584, + "grad_norm": 0.29643896222114563, + "learning_rate": 6.325105833195531e-05, + "loss": 0.0077, + "step": 44390 + }, + { + "epoch": 6.896551724137931, + "grad_norm": 0.16898585855960846, + "learning_rate": 6.323511414741715e-05, + "loss": 0.0087, + "step": 44400 + }, + { + "epoch": 6.898105001553278, + "grad_norm": 0.20052176713943481, + "learning_rate": 6.321916851550678e-05, + "loss": 0.007, + "step": 44410 + }, + { + "epoch": 6.899658278968624, + "grad_norm": 0.3035805821418762, + "learning_rate": 6.320322143796806e-05, + "loss": 0.0112, + "step": 44420 + }, + { + "epoch": 6.90121155638397, + "grad_norm": 0.17828136682510376, + "learning_rate": 6.318727291654488e-05, + "loss": 0.0078, + "step": 44430 + }, + { + "epoch": 6.902764833799317, + "grad_norm": 0.18253235518932343, + "learning_rate": 6.317132295298134e-05, + "loss": 0.0076, + "step": 44440 + }, + { + "epoch": 6.904318111214663, + "grad_norm": 0.17266356945037842, + "learning_rate": 6.315537154902173e-05, + "loss": 0.007, + "step": 44450 + }, + { + "epoch": 6.905871388630009, + "grad_norm": 0.29888230562210083, + "learning_rate": 6.313941870641048e-05, + "loss": 0.0066, + "step": 44460 + }, + { + "epoch": 6.9074246660453555, + "grad_norm": 0.15672901272773743, + "learning_rate": 6.312346442689214e-05, + "loss": 0.0056, + "step": 44470 + }, + { + "epoch": 6.908977943460702, + "grad_norm": 0.38121920824050903, + "learning_rate": 6.310750871221147e-05, + "loss": 0.0059, + "step": 44480 + }, + { + "epoch": 6.910531220876049, + "grad_norm": 0.5738424062728882, + "learning_rate": 6.309155156411335e-05, + "loss": 0.0067, + "step": 44490 + }, + { + "epoch": 6.912084498291395, + "grad_norm": 0.18555839359760284, + "learning_rate": 6.307559298434284e-05, + "loss": 0.0065, + "step": 44500 + }, + { + "epoch": 6.913637775706741, + "grad_norm": 0.1504029780626297, + "learning_rate": 6.305963297464512e-05, + "loss": 0.0064, + "step": 44510 + }, + { + "epoch": 6.915191053122087, + "grad_norm": 0.23435690999031067, + "learning_rate": 6.304367153676561e-05, + "loss": 0.0073, + "step": 44520 + }, + { + "epoch": 6.916744330537434, + "grad_norm": 0.26469048857688904, + "learning_rate": 6.302770867244978e-05, + "loss": 0.0063, + "step": 44530 + }, + { + "epoch": 6.9182976079527805, + "grad_norm": 0.2931421101093292, + "learning_rate": 6.301174438344328e-05, + "loss": 0.0073, + "step": 44540 + }, + { + "epoch": 6.919850885368127, + "grad_norm": 0.251475065946579, + "learning_rate": 6.299577867149202e-05, + "loss": 0.0072, + "step": 44550 + }, + { + "epoch": 6.921404162783473, + "grad_norm": 0.17620103061199188, + "learning_rate": 6.297981153834192e-05, + "loss": 0.0072, + "step": 44560 + }, + { + "epoch": 6.92295744019882, + "grad_norm": 0.17140313982963562, + "learning_rate": 6.296384298573916e-05, + "loss": 0.0061, + "step": 44570 + }, + { + "epoch": 6.924510717614166, + "grad_norm": 0.2861119210720062, + "learning_rate": 6.294787301543001e-05, + "loss": 0.0092, + "step": 44580 + }, + { + "epoch": 6.926063995029512, + "grad_norm": 0.5558478832244873, + "learning_rate": 6.293190162916095e-05, + "loss": 0.0091, + "step": 44590 + }, + { + "epoch": 6.927617272444858, + "grad_norm": 0.18377535045146942, + "learning_rate": 6.291592882867855e-05, + "loss": 0.0062, + "step": 44600 + }, + { + "epoch": 6.929170549860205, + "grad_norm": 0.18541447818279266, + "learning_rate": 6.28999546157296e-05, + "loss": 0.0069, + "step": 44610 + }, + { + "epoch": 6.930723827275552, + "grad_norm": 0.13679027557373047, + "learning_rate": 6.288397899206102e-05, + "loss": 0.0074, + "step": 44620 + }, + { + "epoch": 6.932277104690898, + "grad_norm": 0.16085505485534668, + "learning_rate": 6.286800195941984e-05, + "loss": 0.0068, + "step": 44630 + }, + { + "epoch": 6.933830382106244, + "grad_norm": 0.2707212269306183, + "learning_rate": 6.285202351955334e-05, + "loss": 0.0086, + "step": 44640 + }, + { + "epoch": 6.935383659521591, + "grad_norm": 0.4313196539878845, + "learning_rate": 6.283604367420887e-05, + "loss": 0.0116, + "step": 44650 + }, + { + "epoch": 6.936936936936937, + "grad_norm": 0.23757757246494293, + "learning_rate": 6.282006242513394e-05, + "loss": 0.0131, + "step": 44660 + }, + { + "epoch": 6.938490214352283, + "grad_norm": 0.3081974685192108, + "learning_rate": 6.280407977407628e-05, + "loss": 0.008, + "step": 44670 + }, + { + "epoch": 6.9400434917676295, + "grad_norm": 0.2106817662715912, + "learning_rate": 6.27880957227837e-05, + "loss": 0.0075, + "step": 44680 + }, + { + "epoch": 6.941596769182976, + "grad_norm": 0.25857383012771606, + "learning_rate": 6.27721102730042e-05, + "loss": 0.0081, + "step": 44690 + }, + { + "epoch": 6.943150046598323, + "grad_norm": 0.2401818037033081, + "learning_rate": 6.275612342648592e-05, + "loss": 0.0083, + "step": 44700 + }, + { + "epoch": 6.944703324013669, + "grad_norm": 0.2587735652923584, + "learning_rate": 6.274013518497716e-05, + "loss": 0.0097, + "step": 44710 + }, + { + "epoch": 6.946256601429015, + "grad_norm": 0.26423874497413635, + "learning_rate": 6.272414555022636e-05, + "loss": 0.0066, + "step": 44720 + }, + { + "epoch": 6.947809878844362, + "grad_norm": 0.2005576491355896, + "learning_rate": 6.270815452398215e-05, + "loss": 0.0063, + "step": 44730 + }, + { + "epoch": 6.949363156259708, + "grad_norm": 0.43622025847435, + "learning_rate": 6.269216210799326e-05, + "loss": 0.0059, + "step": 44740 + }, + { + "epoch": 6.950916433675054, + "grad_norm": 0.5109953284263611, + "learning_rate": 6.26761683040086e-05, + "loss": 0.0083, + "step": 44750 + }, + { + "epoch": 6.952469711090401, + "grad_norm": 0.4159678518772125, + "learning_rate": 6.266017311377723e-05, + "loss": 0.0079, + "step": 44760 + }, + { + "epoch": 6.954022988505747, + "grad_norm": 0.2560240924358368, + "learning_rate": 6.264417653904839e-05, + "loss": 0.0076, + "step": 44770 + }, + { + "epoch": 6.955576265921094, + "grad_norm": 0.3174697160720825, + "learning_rate": 6.262817858157139e-05, + "loss": 0.0106, + "step": 44780 + }, + { + "epoch": 6.95712954333644, + "grad_norm": 0.39668628573417664, + "learning_rate": 6.261217924309576e-05, + "loss": 0.0085, + "step": 44790 + }, + { + "epoch": 6.958682820751786, + "grad_norm": 0.2551380395889282, + "learning_rate": 6.259617852537118e-05, + "loss": 0.0095, + "step": 44800 + }, + { + "epoch": 6.960236098167132, + "grad_norm": 0.24508439004421234, + "learning_rate": 6.258017643014747e-05, + "loss": 0.0086, + "step": 44810 + }, + { + "epoch": 6.961789375582479, + "grad_norm": 0.2580845355987549, + "learning_rate": 6.256417295917456e-05, + "loss": 0.0127, + "step": 44820 + }, + { + "epoch": 6.9633426529978255, + "grad_norm": 0.24297897517681122, + "learning_rate": 6.254816811420258e-05, + "loss": 0.0078, + "step": 44830 + }, + { + "epoch": 6.964895930413172, + "grad_norm": 0.20604465901851654, + "learning_rate": 6.253216189698183e-05, + "loss": 0.0082, + "step": 44840 + }, + { + "epoch": 6.966449207828518, + "grad_norm": 0.33859413862228394, + "learning_rate": 6.251615430926267e-05, + "loss": 0.0066, + "step": 44850 + }, + { + "epoch": 6.968002485243865, + "grad_norm": 0.3306255638599396, + "learning_rate": 6.25001453527957e-05, + "loss": 0.0083, + "step": 44860 + }, + { + "epoch": 6.969555762659211, + "grad_norm": 0.40686938166618347, + "learning_rate": 6.248413502933164e-05, + "loss": 0.0089, + "step": 44870 + }, + { + "epoch": 6.971109040074557, + "grad_norm": 0.2039133608341217, + "learning_rate": 6.246812334062133e-05, + "loss": 0.007, + "step": 44880 + }, + { + "epoch": 6.972662317489903, + "grad_norm": 0.4514900743961334, + "learning_rate": 6.245211028841579e-05, + "loss": 0.0085, + "step": 44890 + }, + { + "epoch": 6.9742155949052504, + "grad_norm": 0.3904741108417511, + "learning_rate": 6.24360958744662e-05, + "loss": 0.0076, + "step": 44900 + }, + { + "epoch": 6.975768872320597, + "grad_norm": 0.3613899052143097, + "learning_rate": 6.242008010052387e-05, + "loss": 0.0067, + "step": 44910 + }, + { + "epoch": 6.977322149735943, + "grad_norm": 0.24538004398345947, + "learning_rate": 6.240406296834024e-05, + "loss": 0.0094, + "step": 44920 + }, + { + "epoch": 6.978875427151289, + "grad_norm": 0.23575200140476227, + "learning_rate": 6.238804447966694e-05, + "loss": 0.006, + "step": 44930 + }, + { + "epoch": 6.980428704566636, + "grad_norm": 0.31888794898986816, + "learning_rate": 6.237202463625573e-05, + "loss": 0.0059, + "step": 44940 + }, + { + "epoch": 6.981981981981982, + "grad_norm": 0.2940450608730316, + "learning_rate": 6.235600343985848e-05, + "loss": 0.0069, + "step": 44950 + }, + { + "epoch": 6.983535259397328, + "grad_norm": 0.24785774946212769, + "learning_rate": 6.233998089222729e-05, + "loss": 0.0061, + "step": 44960 + }, + { + "epoch": 6.9850885368126745, + "grad_norm": 0.18302151560783386, + "learning_rate": 6.232395699511433e-05, + "loss": 0.0052, + "step": 44970 + }, + { + "epoch": 6.986641814228021, + "grad_norm": 0.2173343002796173, + "learning_rate": 6.230793175027199e-05, + "loss": 0.0055, + "step": 44980 + }, + { + "epoch": 6.988195091643368, + "grad_norm": 0.23877212405204773, + "learning_rate": 6.22919051594527e-05, + "loss": 0.0094, + "step": 44990 + }, + { + "epoch": 6.989748369058714, + "grad_norm": 0.4571826457977295, + "learning_rate": 6.227587722440917e-05, + "loss": 0.0071, + "step": 45000 + }, + { + "epoch": 6.99130164647406, + "grad_norm": 0.1872318685054779, + "learning_rate": 6.225984794689414e-05, + "loss": 0.007, + "step": 45010 + }, + { + "epoch": 6.992854923889407, + "grad_norm": 0.5231626629829407, + "learning_rate": 6.22438173286606e-05, + "loss": 0.0072, + "step": 45020 + }, + { + "epoch": 6.994408201304753, + "grad_norm": 0.859286904335022, + "learning_rate": 6.222778537146159e-05, + "loss": 0.0102, + "step": 45030 + }, + { + "epoch": 6.995961478720099, + "grad_norm": 0.6648209691047668, + "learning_rate": 6.221175207705037e-05, + "loss": 0.0093, + "step": 45040 + }, + { + "epoch": 6.997514756135446, + "grad_norm": 0.2006026655435562, + "learning_rate": 6.219571744718027e-05, + "loss": 0.0065, + "step": 45050 + }, + { + "epoch": 6.999068033550792, + "grad_norm": 0.3147030770778656, + "learning_rate": 6.217968148360487e-05, + "loss": 0.0088, + "step": 45060 + }, + { + "epoch": 7.000621310966139, + "grad_norm": 0.22212263941764832, + "learning_rate": 6.216364418807782e-05, + "loss": 0.0071, + "step": 45070 + }, + { + "epoch": 7.002174588381485, + "grad_norm": 0.16501881182193756, + "learning_rate": 6.214760556235292e-05, + "loss": 0.0083, + "step": 45080 + }, + { + "epoch": 7.003727865796831, + "grad_norm": 0.36125558614730835, + "learning_rate": 6.213156560818414e-05, + "loss": 0.0117, + "step": 45090 + }, + { + "epoch": 7.005281143212177, + "grad_norm": 0.2945977449417114, + "learning_rate": 6.211552432732559e-05, + "loss": 0.0081, + "step": 45100 + }, + { + "epoch": 7.006834420627524, + "grad_norm": 0.44312170147895813, + "learning_rate": 6.20994817215315e-05, + "loss": 0.0074, + "step": 45110 + }, + { + "epoch": 7.0083876980428705, + "grad_norm": 0.1502169519662857, + "learning_rate": 6.208343779255628e-05, + "loss": 0.0068, + "step": 45120 + }, + { + "epoch": 7.009940975458217, + "grad_norm": 0.16974127292633057, + "learning_rate": 6.206739254215449e-05, + "loss": 0.0076, + "step": 45130 + }, + { + "epoch": 7.011494252873563, + "grad_norm": 0.2306414395570755, + "learning_rate": 6.205134597208077e-05, + "loss": 0.0094, + "step": 45140 + }, + { + "epoch": 7.01304753028891, + "grad_norm": 0.3066069781780243, + "learning_rate": 6.203529808409e-05, + "loss": 0.0065, + "step": 45150 + }, + { + "epoch": 7.014600807704256, + "grad_norm": 0.3588847815990448, + "learning_rate": 6.201924887993712e-05, + "loss": 0.0067, + "step": 45160 + }, + { + "epoch": 7.016154085119602, + "grad_norm": 0.3984798789024353, + "learning_rate": 6.200319836137725e-05, + "loss": 0.0068, + "step": 45170 + }, + { + "epoch": 7.017707362534948, + "grad_norm": 0.3337913453578949, + "learning_rate": 6.198714653016565e-05, + "loss": 0.0066, + "step": 45180 + }, + { + "epoch": 7.0192606399502955, + "grad_norm": 0.2321646809577942, + "learning_rate": 6.197109338805774e-05, + "loss": 0.0092, + "step": 45190 + }, + { + "epoch": 7.020813917365642, + "grad_norm": 0.34166476130485535, + "learning_rate": 6.195503893680903e-05, + "loss": 0.0056, + "step": 45200 + }, + { + "epoch": 7.022367194780988, + "grad_norm": 0.31729012727737427, + "learning_rate": 6.193898317817524e-05, + "loss": 0.013, + "step": 45210 + }, + { + "epoch": 7.023920472196334, + "grad_norm": 0.30252665281295776, + "learning_rate": 6.192292611391221e-05, + "loss": 0.0065, + "step": 45220 + }, + { + "epoch": 7.025473749611681, + "grad_norm": 0.3080349862575531, + "learning_rate": 6.190686774577591e-05, + "loss": 0.0086, + "step": 45230 + }, + { + "epoch": 7.027027027027027, + "grad_norm": 0.19673895835876465, + "learning_rate": 6.189080807552245e-05, + "loss": 0.008, + "step": 45240 + }, + { + "epoch": 7.028580304442373, + "grad_norm": 0.3057195246219635, + "learning_rate": 6.187474710490809e-05, + "loss": 0.006, + "step": 45250 + }, + { + "epoch": 7.0301335818577195, + "grad_norm": 0.3588107228279114, + "learning_rate": 6.185868483568926e-05, + "loss": 0.0084, + "step": 45260 + }, + { + "epoch": 7.031686859273067, + "grad_norm": 0.32037025690078735, + "learning_rate": 6.184262126962245e-05, + "loss": 0.0075, + "step": 45270 + }, + { + "epoch": 7.033240136688413, + "grad_norm": 0.5620137453079224, + "learning_rate": 6.182655640846442e-05, + "loss": 0.0069, + "step": 45280 + }, + { + "epoch": 7.034793414103759, + "grad_norm": 0.2834456264972687, + "learning_rate": 6.181049025397196e-05, + "loss": 0.0117, + "step": 45290 + }, + { + "epoch": 7.036346691519105, + "grad_norm": 0.2448885142803192, + "learning_rate": 6.179442280790202e-05, + "loss": 0.0096, + "step": 45300 + }, + { + "epoch": 7.037899968934452, + "grad_norm": 0.17780040204524994, + "learning_rate": 6.177835407201174e-05, + "loss": 0.0088, + "step": 45310 + }, + { + "epoch": 7.039453246349798, + "grad_norm": 0.2449149787425995, + "learning_rate": 6.176228404805839e-05, + "loss": 0.0068, + "step": 45320 + }, + { + "epoch": 7.0410065237651445, + "grad_norm": 0.3200204372406006, + "learning_rate": 6.174621273779932e-05, + "loss": 0.012, + "step": 45330 + }, + { + "epoch": 7.042559801180491, + "grad_norm": 0.20722927153110504, + "learning_rate": 6.17301401429921e-05, + "loss": 0.0081, + "step": 45340 + }, + { + "epoch": 7.044113078595837, + "grad_norm": 0.2388893961906433, + "learning_rate": 6.171406626539436e-05, + "loss": 0.0069, + "step": 45350 + }, + { + "epoch": 7.045666356011184, + "grad_norm": 0.21455572545528412, + "learning_rate": 6.169799110676398e-05, + "loss": 0.0082, + "step": 45360 + }, + { + "epoch": 7.04721963342653, + "grad_norm": 0.2317051738500595, + "learning_rate": 6.168191466885885e-05, + "loss": 0.0069, + "step": 45370 + }, + { + "epoch": 7.048772910841876, + "grad_norm": 0.2699611186981201, + "learning_rate": 6.166583695343711e-05, + "loss": 0.0104, + "step": 45380 + }, + { + "epoch": 7.050326188257222, + "grad_norm": 0.14288479089736938, + "learning_rate": 6.164975796225698e-05, + "loss": 0.0069, + "step": 45390 + }, + { + "epoch": 7.051879465672569, + "grad_norm": 0.1688542515039444, + "learning_rate": 6.163367769707683e-05, + "loss": 0.0089, + "step": 45400 + }, + { + "epoch": 7.053432743087916, + "grad_norm": 0.12143639475107193, + "learning_rate": 6.161759615965519e-05, + "loss": 0.0073, + "step": 45410 + }, + { + "epoch": 7.054986020503262, + "grad_norm": 0.2457956224679947, + "learning_rate": 6.16015133517507e-05, + "loss": 0.0117, + "step": 45420 + }, + { + "epoch": 7.056539297918608, + "grad_norm": 0.35843467712402344, + "learning_rate": 6.158542927512214e-05, + "loss": 0.0092, + "step": 45430 + }, + { + "epoch": 7.058092575333955, + "grad_norm": 0.1848854273557663, + "learning_rate": 6.156934393152846e-05, + "loss": 0.0063, + "step": 45440 + }, + { + "epoch": 7.059645852749301, + "grad_norm": 0.28610432147979736, + "learning_rate": 6.15532573227287e-05, + "loss": 0.0085, + "step": 45450 + }, + { + "epoch": 7.061199130164647, + "grad_norm": 0.6822926998138428, + "learning_rate": 6.153716945048212e-05, + "loss": 0.0109, + "step": 45460 + }, + { + "epoch": 7.0627524075799935, + "grad_norm": 0.3366779685020447, + "learning_rate": 6.152108031654802e-05, + "loss": 0.0073, + "step": 45470 + }, + { + "epoch": 7.0643056849953405, + "grad_norm": 0.2588805556297302, + "learning_rate": 6.15049899226859e-05, + "loss": 0.0097, + "step": 45480 + }, + { + "epoch": 7.065858962410687, + "grad_norm": 0.24690669775009155, + "learning_rate": 6.148889827065537e-05, + "loss": 0.0072, + "step": 45490 + }, + { + "epoch": 7.067412239826033, + "grad_norm": 0.24446693062782288, + "learning_rate": 6.147280536221622e-05, + "loss": 0.0068, + "step": 45500 + }, + { + "epoch": 7.068965517241379, + "grad_norm": 0.32102635502815247, + "learning_rate": 6.145671119912832e-05, + "loss": 0.0081, + "step": 45510 + }, + { + "epoch": 7.070518794656726, + "grad_norm": 0.5220139026641846, + "learning_rate": 6.144061578315169e-05, + "loss": 0.009, + "step": 45520 + }, + { + "epoch": 7.072072072072072, + "grad_norm": 0.15369437634944916, + "learning_rate": 6.142451911604654e-05, + "loss": 0.007, + "step": 45530 + }, + { + "epoch": 7.073625349487418, + "grad_norm": 0.22400866448879242, + "learning_rate": 6.140842119957315e-05, + "loss": 0.0072, + "step": 45540 + }, + { + "epoch": 7.075178626902765, + "grad_norm": 0.26696455478668213, + "learning_rate": 6.139232203549197e-05, + "loss": 0.0091, + "step": 45550 + }, + { + "epoch": 7.076731904318112, + "grad_norm": 0.14194729924201965, + "learning_rate": 6.137622162556357e-05, + "loss": 0.0068, + "step": 45560 + }, + { + "epoch": 7.078285181733458, + "grad_norm": 0.2005046308040619, + "learning_rate": 6.13601199715487e-05, + "loss": 0.0079, + "step": 45570 + }, + { + "epoch": 7.079838459148804, + "grad_norm": 0.29589685797691345, + "learning_rate": 6.134401707520816e-05, + "loss": 0.0071, + "step": 45580 + }, + { + "epoch": 7.08139173656415, + "grad_norm": 0.252824604511261, + "learning_rate": 6.132791293830298e-05, + "loss": 0.0091, + "step": 45590 + }, + { + "epoch": 7.082945013979497, + "grad_norm": 0.1434982866048813, + "learning_rate": 6.131180756259428e-05, + "loss": 0.0067, + "step": 45600 + }, + { + "epoch": 7.084498291394843, + "grad_norm": 0.31416383385658264, + "learning_rate": 6.129570094984331e-05, + "loss": 0.0073, + "step": 45610 + }, + { + "epoch": 7.0860515688101895, + "grad_norm": 0.20111845433712006, + "learning_rate": 6.127959310181145e-05, + "loss": 0.0087, + "step": 45620 + }, + { + "epoch": 7.087604846225536, + "grad_norm": 0.22523896396160126, + "learning_rate": 6.126348402026026e-05, + "loss": 0.0093, + "step": 45630 + }, + { + "epoch": 7.089158123640882, + "grad_norm": 0.371184766292572, + "learning_rate": 6.12473737069514e-05, + "loss": 0.0066, + "step": 45640 + }, + { + "epoch": 7.090711401056229, + "grad_norm": 0.42840468883514404, + "learning_rate": 6.123126216364665e-05, + "loss": 0.0069, + "step": 45650 + }, + { + "epoch": 7.092264678471575, + "grad_norm": 0.2525230348110199, + "learning_rate": 6.121514939210797e-05, + "loss": 0.0076, + "step": 45660 + }, + { + "epoch": 7.093817955886921, + "grad_norm": 0.2085445076227188, + "learning_rate": 6.119903539409741e-05, + "loss": 0.0073, + "step": 45670 + }, + { + "epoch": 7.095371233302267, + "grad_norm": 0.22731411457061768, + "learning_rate": 6.118292017137716e-05, + "loss": 0.0062, + "step": 45680 + }, + { + "epoch": 7.096924510717614, + "grad_norm": 0.19405597448349, + "learning_rate": 6.116680372570959e-05, + "loss": 0.0057, + "step": 45690 + }, + { + "epoch": 7.098477788132961, + "grad_norm": 0.2557161748409271, + "learning_rate": 6.115068605885713e-05, + "loss": 0.007, + "step": 45700 + }, + { + "epoch": 7.100031065548307, + "grad_norm": 0.3463554382324219, + "learning_rate": 6.113456717258243e-05, + "loss": 0.0074, + "step": 45710 + }, + { + "epoch": 7.101584342963653, + "grad_norm": 0.2967238426208496, + "learning_rate": 6.11184470686482e-05, + "loss": 0.0079, + "step": 45720 + }, + { + "epoch": 7.103137620379, + "grad_norm": 0.11108852922916412, + "learning_rate": 6.110232574881731e-05, + "loss": 0.0088, + "step": 45730 + }, + { + "epoch": 7.104690897794346, + "grad_norm": 0.13736490905284882, + "learning_rate": 6.108620321485277e-05, + "loss": 0.0091, + "step": 45740 + }, + { + "epoch": 7.106244175209692, + "grad_norm": 0.11928071826696396, + "learning_rate": 6.107007946851773e-05, + "loss": 0.007, + "step": 45750 + }, + { + "epoch": 7.1077974526250385, + "grad_norm": 0.21011197566986084, + "learning_rate": 6.105395451157542e-05, + "loss": 0.0067, + "step": 45760 + }, + { + "epoch": 7.1093507300403855, + "grad_norm": 0.15735988318920135, + "learning_rate": 6.103782834578928e-05, + "loss": 0.0069, + "step": 45770 + }, + { + "epoch": 7.110904007455732, + "grad_norm": 0.25999876856803894, + "learning_rate": 6.102170097292281e-05, + "loss": 0.0073, + "step": 45780 + }, + { + "epoch": 7.112457284871078, + "grad_norm": 0.2372678816318512, + "learning_rate": 6.1005572394739716e-05, + "loss": 0.0076, + "step": 45790 + }, + { + "epoch": 7.114010562286424, + "grad_norm": 0.22125621140003204, + "learning_rate": 6.0989442613003765e-05, + "loss": 0.007, + "step": 45800 + }, + { + "epoch": 7.115563839701771, + "grad_norm": 0.18331900238990784, + "learning_rate": 6.097331162947888e-05, + "loss": 0.0084, + "step": 45810 + }, + { + "epoch": 7.117117117117117, + "grad_norm": 0.3964417278766632, + "learning_rate": 6.095717944592914e-05, + "loss": 0.009, + "step": 45820 + }, + { + "epoch": 7.118670394532463, + "grad_norm": 0.3701629042625427, + "learning_rate": 6.094104606411873e-05, + "loss": 0.0068, + "step": 45830 + }, + { + "epoch": 7.12022367194781, + "grad_norm": 0.3152497708797455, + "learning_rate": 6.0924911485811966e-05, + "loss": 0.0083, + "step": 45840 + }, + { + "epoch": 7.121776949363157, + "grad_norm": 0.3358583450317383, + "learning_rate": 6.090877571277331e-05, + "loss": 0.0052, + "step": 45850 + }, + { + "epoch": 7.123330226778503, + "grad_norm": 0.20730246603488922, + "learning_rate": 6.089263874676736e-05, + "loss": 0.0117, + "step": 45860 + }, + { + "epoch": 7.124883504193849, + "grad_norm": 0.3567158877849579, + "learning_rate": 6.0876500589558796e-05, + "loss": 0.0075, + "step": 45870 + }, + { + "epoch": 7.126436781609195, + "grad_norm": 0.3449094295501709, + "learning_rate": 6.086036124291248e-05, + "loss": 0.0055, + "step": 45880 + }, + { + "epoch": 7.127990059024542, + "grad_norm": 0.40907779335975647, + "learning_rate": 6.084422070859339e-05, + "loss": 0.0076, + "step": 45890 + }, + { + "epoch": 7.129543336439888, + "grad_norm": 0.2600889801979065, + "learning_rate": 6.082807898836663e-05, + "loss": 0.0068, + "step": 45900 + }, + { + "epoch": 7.1310966138552345, + "grad_norm": 0.17113427817821503, + "learning_rate": 6.081193608399742e-05, + "loss": 0.0091, + "step": 45910 + }, + { + "epoch": 7.132649891270581, + "grad_norm": 0.4274178445339203, + "learning_rate": 6.0795791997251164e-05, + "loss": 0.0075, + "step": 45920 + }, + { + "epoch": 7.134203168685927, + "grad_norm": 0.14820823073387146, + "learning_rate": 6.0779646729893294e-05, + "loss": 0.0064, + "step": 45930 + }, + { + "epoch": 7.135756446101274, + "grad_norm": 0.2678860127925873, + "learning_rate": 6.0763500283689476e-05, + "loss": 0.0056, + "step": 45940 + }, + { + "epoch": 7.13730972351662, + "grad_norm": 0.14783459901809692, + "learning_rate": 6.0747352660405455e-05, + "loss": 0.0076, + "step": 45950 + }, + { + "epoch": 7.138863000931966, + "grad_norm": 0.261496901512146, + "learning_rate": 6.073120386180709e-05, + "loss": 0.0062, + "step": 45960 + }, + { + "epoch": 7.140416278347312, + "grad_norm": 0.24283060431480408, + "learning_rate": 6.0715053889660425e-05, + "loss": 0.0049, + "step": 45970 + }, + { + "epoch": 7.1419695557626595, + "grad_norm": 0.3219418525695801, + "learning_rate": 6.069890274573157e-05, + "loss": 0.0073, + "step": 45980 + }, + { + "epoch": 7.143522833178006, + "grad_norm": 0.23874130845069885, + "learning_rate": 6.068275043178679e-05, + "loss": 0.0105, + "step": 45990 + }, + { + "epoch": 7.145076110593352, + "grad_norm": 0.24411332607269287, + "learning_rate": 6.066659694959248e-05, + "loss": 0.0061, + "step": 46000 + }, + { + "epoch": 7.146629388008698, + "grad_norm": 0.27827003598213196, + "learning_rate": 6.0650442300915176e-05, + "loss": 0.0071, + "step": 46010 + }, + { + "epoch": 7.148182665424045, + "grad_norm": 0.3923344612121582, + "learning_rate": 6.063428648752152e-05, + "loss": 0.0072, + "step": 46020 + }, + { + "epoch": 7.149735942839391, + "grad_norm": 0.15420043468475342, + "learning_rate": 6.0618129511178266e-05, + "loss": 0.006, + "step": 46030 + }, + { + "epoch": 7.151289220254737, + "grad_norm": 0.49426761269569397, + "learning_rate": 6.060197137365234e-05, + "loss": 0.0078, + "step": 46040 + }, + { + "epoch": 7.1528424976700835, + "grad_norm": 0.32262301445007324, + "learning_rate": 6.058581207671077e-05, + "loss": 0.0077, + "step": 46050 + }, + { + "epoch": 7.154395775085431, + "grad_norm": 0.24539323151111603, + "learning_rate": 6.056965162212072e-05, + "loss": 0.0067, + "step": 46060 + }, + { + "epoch": 7.155949052500777, + "grad_norm": 0.22872942686080933, + "learning_rate": 6.055349001164943e-05, + "loss": 0.0089, + "step": 46070 + }, + { + "epoch": 7.157502329916123, + "grad_norm": 0.18355287611484528, + "learning_rate": 6.0537327247064347e-05, + "loss": 0.0067, + "step": 46080 + }, + { + "epoch": 7.159055607331469, + "grad_norm": 0.15181544423103333, + "learning_rate": 6.0521163330133014e-05, + "loss": 0.0094, + "step": 46090 + }, + { + "epoch": 7.160608884746816, + "grad_norm": 0.1113138496875763, + "learning_rate": 6.050499826262306e-05, + "loss": 0.0063, + "step": 46100 + }, + { + "epoch": 7.162162162162162, + "grad_norm": 0.16221776604652405, + "learning_rate": 6.0488832046302294e-05, + "loss": 0.0087, + "step": 46110 + }, + { + "epoch": 7.1637154395775084, + "grad_norm": 0.18174605071544647, + "learning_rate": 6.0472664682938626e-05, + "loss": 0.0076, + "step": 46120 + }, + { + "epoch": 7.165268716992855, + "grad_norm": 0.30847597122192383, + "learning_rate": 6.045649617430009e-05, + "loss": 0.0055, + "step": 46130 + }, + { + "epoch": 7.166821994408202, + "grad_norm": 0.2711980938911438, + "learning_rate": 6.0440326522154866e-05, + "loss": 0.0069, + "step": 46140 + }, + { + "epoch": 7.168375271823548, + "grad_norm": 0.28267040848731995, + "learning_rate": 6.0424155728271224e-05, + "loss": 0.0061, + "step": 46150 + }, + { + "epoch": 7.169928549238894, + "grad_norm": 0.2417801469564438, + "learning_rate": 6.040798379441758e-05, + "loss": 0.0081, + "step": 46160 + }, + { + "epoch": 7.17148182665424, + "grad_norm": 0.18769186735153198, + "learning_rate": 6.0391810722362485e-05, + "loss": 0.0076, + "step": 46170 + }, + { + "epoch": 7.173035104069587, + "grad_norm": 0.1424282342195511, + "learning_rate": 6.037563651387458e-05, + "loss": 0.0049, + "step": 46180 + }, + { + "epoch": 7.174588381484933, + "grad_norm": 0.24212872982025146, + "learning_rate": 6.0359461170722666e-05, + "loss": 0.008, + "step": 46190 + }, + { + "epoch": 7.1761416589002796, + "grad_norm": 0.19587957859039307, + "learning_rate": 6.034328469467566e-05, + "loss": 0.007, + "step": 46200 + }, + { + "epoch": 7.177694936315626, + "grad_norm": 0.2547721862792969, + "learning_rate": 6.0327107087502596e-05, + "loss": 0.0085, + "step": 46210 + }, + { + "epoch": 7.179248213730973, + "grad_norm": 0.234915629029274, + "learning_rate": 6.031092835097262e-05, + "loss": 0.0081, + "step": 46220 + }, + { + "epoch": 7.180801491146319, + "grad_norm": 0.4290907382965088, + "learning_rate": 6.0294748486855024e-05, + "loss": 0.0064, + "step": 46230 + }, + { + "epoch": 7.182354768561665, + "grad_norm": 0.2487487494945526, + "learning_rate": 6.0278567496919216e-05, + "loss": 0.0093, + "step": 46240 + }, + { + "epoch": 7.183908045977011, + "grad_norm": 0.22338636219501495, + "learning_rate": 6.026238538293472e-05, + "loss": 0.0083, + "step": 46250 + }, + { + "epoch": 7.185461323392358, + "grad_norm": 0.48567262291908264, + "learning_rate": 6.024620214667118e-05, + "loss": 0.0077, + "step": 46260 + }, + { + "epoch": 7.1870146008077045, + "grad_norm": 0.2808150053024292, + "learning_rate": 6.0230017789898384e-05, + "loss": 0.0081, + "step": 46270 + }, + { + "epoch": 7.188567878223051, + "grad_norm": 0.36323484778404236, + "learning_rate": 6.021383231438622e-05, + "loss": 0.0071, + "step": 46280 + }, + { + "epoch": 7.190121155638397, + "grad_norm": 0.31701961159706116, + "learning_rate": 6.0197645721904704e-05, + "loss": 0.007, + "step": 46290 + }, + { + "epoch": 7.191674433053743, + "grad_norm": 0.25053641200065613, + "learning_rate": 6.0181458014224e-05, + "loss": 0.0112, + "step": 46300 + }, + { + "epoch": 7.19322771046909, + "grad_norm": 0.39606696367263794, + "learning_rate": 6.016526919311435e-05, + "loss": 0.009, + "step": 46310 + }, + { + "epoch": 7.194780987884436, + "grad_norm": 0.33344846963882446, + "learning_rate": 6.014907926034613e-05, + "loss": 0.0096, + "step": 46320 + }, + { + "epoch": 7.196334265299782, + "grad_norm": 0.2274257242679596, + "learning_rate": 6.0132888217689875e-05, + "loss": 0.0072, + "step": 46330 + }, + { + "epoch": 7.1978875427151285, + "grad_norm": 0.19611701369285583, + "learning_rate": 6.01166960669162e-05, + "loss": 0.0071, + "step": 46340 + }, + { + "epoch": 7.199440820130476, + "grad_norm": 0.17017416656017303, + "learning_rate": 6.0100502809795845e-05, + "loss": 0.0084, + "step": 46350 + }, + { + "epoch": 7.200994097545822, + "grad_norm": 0.3061336576938629, + "learning_rate": 6.00843084480997e-05, + "loss": 0.0087, + "step": 46360 + }, + { + "epoch": 7.202547374961168, + "grad_norm": 0.36383000016212463, + "learning_rate": 6.0068112983598736e-05, + "loss": 0.0088, + "step": 46370 + }, + { + "epoch": 7.204100652376514, + "grad_norm": 0.26455986499786377, + "learning_rate": 6.005191641806407e-05, + "loss": 0.008, + "step": 46380 + }, + { + "epoch": 7.205653929791861, + "grad_norm": 0.3490862548351288, + "learning_rate": 6.003571875326694e-05, + "loss": 0.0087, + "step": 46390 + }, + { + "epoch": 7.207207207207207, + "grad_norm": 0.21928992867469788, + "learning_rate": 6.00195199909787e-05, + "loss": 0.0071, + "step": 46400 + }, + { + "epoch": 7.2087604846225535, + "grad_norm": 0.33684635162353516, + "learning_rate": 6.000332013297082e-05, + "loss": 0.0065, + "step": 46410 + }, + { + "epoch": 7.2103137620379, + "grad_norm": 0.3688882291316986, + "learning_rate": 5.998711918101487e-05, + "loss": 0.0053, + "step": 46420 + }, + { + "epoch": 7.211867039453247, + "grad_norm": 0.16204515099525452, + "learning_rate": 5.997091713688261e-05, + "loss": 0.007, + "step": 46430 + }, + { + "epoch": 7.213420316868593, + "grad_norm": 0.13193155825138092, + "learning_rate": 5.9954714002345836e-05, + "loss": 0.0073, + "step": 46440 + }, + { + "epoch": 7.214973594283939, + "grad_norm": 0.13960257172584534, + "learning_rate": 5.993850977917649e-05, + "loss": 0.0052, + "step": 46450 + }, + { + "epoch": 7.216526871699285, + "grad_norm": 0.13112153112888336, + "learning_rate": 5.992230446914667e-05, + "loss": 0.0062, + "step": 46460 + }, + { + "epoch": 7.218080149114632, + "grad_norm": 0.27655360102653503, + "learning_rate": 5.990609807402855e-05, + "loss": 0.0069, + "step": 46470 + }, + { + "epoch": 7.219633426529978, + "grad_norm": 0.2790256142616272, + "learning_rate": 5.988989059559443e-05, + "loss": 0.0077, + "step": 46480 + }, + { + "epoch": 7.221186703945325, + "grad_norm": 0.2476828694343567, + "learning_rate": 5.987368203561675e-05, + "loss": 0.0102, + "step": 46490 + }, + { + "epoch": 7.222739981360671, + "grad_norm": 0.5278615355491638, + "learning_rate": 5.9857472395868055e-05, + "loss": 0.0066, + "step": 46500 + }, + { + "epoch": 7.224293258776018, + "grad_norm": 0.18953868746757507, + "learning_rate": 5.9841261678120983e-05, + "loss": 0.0071, + "step": 46510 + }, + { + "epoch": 7.225846536191364, + "grad_norm": 0.14227202534675598, + "learning_rate": 5.982504988414834e-05, + "loss": 0.0052, + "step": 46520 + }, + { + "epoch": 7.22739981360671, + "grad_norm": 0.1535908728837967, + "learning_rate": 5.9808837015723015e-05, + "loss": 0.0076, + "step": 46530 + }, + { + "epoch": 7.228953091022056, + "grad_norm": 0.15204393863677979, + "learning_rate": 5.9792623074618016e-05, + "loss": 0.0055, + "step": 46540 + }, + { + "epoch": 7.230506368437403, + "grad_norm": 0.1595568209886551, + "learning_rate": 5.977640806260648e-05, + "loss": 0.0071, + "step": 46550 + }, + { + "epoch": 7.2320596458527495, + "grad_norm": 0.1910080909729004, + "learning_rate": 5.976019198146166e-05, + "loss": 0.0095, + "step": 46560 + }, + { + "epoch": 7.233612923268096, + "grad_norm": 0.1769019365310669, + "learning_rate": 5.974397483295692e-05, + "loss": 0.0082, + "step": 46570 + }, + { + "epoch": 7.235166200683442, + "grad_norm": 0.21344254910945892, + "learning_rate": 5.972775661886575e-05, + "loss": 0.0057, + "step": 46580 + }, + { + "epoch": 7.236719478098788, + "grad_norm": 0.6024601459503174, + "learning_rate": 5.971153734096173e-05, + "loss": 0.0071, + "step": 46590 + }, + { + "epoch": 7.238272755514135, + "grad_norm": 0.3088551163673401, + "learning_rate": 5.969531700101859e-05, + "loss": 0.0077, + "step": 46600 + }, + { + "epoch": 7.239826032929481, + "grad_norm": 0.24454084038734436, + "learning_rate": 5.9679095600810155e-05, + "loss": 0.0069, + "step": 46610 + }, + { + "epoch": 7.241379310344827, + "grad_norm": 0.24151912331581116, + "learning_rate": 5.9662873142110384e-05, + "loss": 0.0089, + "step": 46620 + }, + { + "epoch": 7.242932587760174, + "grad_norm": 0.19059166312217712, + "learning_rate": 5.964664962669333e-05, + "loss": 0.007, + "step": 46630 + }, + { + "epoch": 7.244485865175521, + "grad_norm": 0.27668651938438416, + "learning_rate": 5.9630425056333186e-05, + "loss": 0.0073, + "step": 46640 + }, + { + "epoch": 7.246039142590867, + "grad_norm": 0.2940583825111389, + "learning_rate": 5.961419943280422e-05, + "loss": 0.0067, + "step": 46650 + }, + { + "epoch": 7.247592420006213, + "grad_norm": 0.1638428419828415, + "learning_rate": 5.959797275788087e-05, + "loss": 0.0057, + "step": 46660 + }, + { + "epoch": 7.249145697421559, + "grad_norm": 0.24619968235492706, + "learning_rate": 5.958174503333765e-05, + "loss": 0.0066, + "step": 46670 + }, + { + "epoch": 7.250698974836906, + "grad_norm": 0.1903117448091507, + "learning_rate": 5.9565516260949195e-05, + "loss": 0.0089, + "step": 46680 + }, + { + "epoch": 7.252252252252252, + "grad_norm": 0.26548635959625244, + "learning_rate": 5.9549286442490273e-05, + "loss": 0.0066, + "step": 46690 + }, + { + "epoch": 7.2538055296675985, + "grad_norm": 0.2793957591056824, + "learning_rate": 5.953305557973572e-05, + "loss": 0.006, + "step": 46700 + }, + { + "epoch": 7.255358807082945, + "grad_norm": 0.35012418031692505, + "learning_rate": 5.9516823674460564e-05, + "loss": 0.0061, + "step": 46710 + }, + { + "epoch": 7.256912084498292, + "grad_norm": 0.9492376446723938, + "learning_rate": 5.9500590728439875e-05, + "loss": 0.0083, + "step": 46720 + }, + { + "epoch": 7.258465361913638, + "grad_norm": 0.16312392055988312, + "learning_rate": 5.948435674344886e-05, + "loss": 0.0064, + "step": 46730 + }, + { + "epoch": 7.260018639328984, + "grad_norm": 0.1357850879430771, + "learning_rate": 5.946812172126285e-05, + "loss": 0.0089, + "step": 46740 + }, + { + "epoch": 7.26157191674433, + "grad_norm": 0.8007545471191406, + "learning_rate": 5.945188566365729e-05, + "loss": 0.0061, + "step": 46750 + }, + { + "epoch": 7.263125194159677, + "grad_norm": 0.19222566485404968, + "learning_rate": 5.943564857240773e-05, + "loss": 0.0056, + "step": 46760 + }, + { + "epoch": 7.264678471575023, + "grad_norm": 0.24250945448875427, + "learning_rate": 5.941941044928981e-05, + "loss": 0.0104, + "step": 46770 + }, + { + "epoch": 7.26623174899037, + "grad_norm": 0.17558369040489197, + "learning_rate": 5.940317129607935e-05, + "loss": 0.0091, + "step": 46780 + }, + { + "epoch": 7.267785026405716, + "grad_norm": 0.18400201201438904, + "learning_rate": 5.9386931114552204e-05, + "loss": 0.0066, + "step": 46790 + }, + { + "epoch": 7.269338303821063, + "grad_norm": 0.21826672554016113, + "learning_rate": 5.937068990648438e-05, + "loss": 0.0086, + "step": 46800 + }, + { + "epoch": 7.270891581236409, + "grad_norm": 0.17443552613258362, + "learning_rate": 5.935444767365199e-05, + "loss": 0.0063, + "step": 46810 + }, + { + "epoch": 7.272444858651755, + "grad_norm": 0.18911975622177124, + "learning_rate": 5.933820441783129e-05, + "loss": 0.0067, + "step": 46820 + }, + { + "epoch": 7.273998136067101, + "grad_norm": 0.3525574505329132, + "learning_rate": 5.932196014079857e-05, + "loss": 0.0098, + "step": 46830 + }, + { + "epoch": 7.275551413482448, + "grad_norm": 0.7460734844207764, + "learning_rate": 5.930571484433032e-05, + "loss": 0.0079, + "step": 46840 + }, + { + "epoch": 7.2771046908977945, + "grad_norm": 0.36009737849235535, + "learning_rate": 5.928946853020309e-05, + "loss": 0.0095, + "step": 46850 + }, + { + "epoch": 7.278657968313141, + "grad_norm": 0.3502240478992462, + "learning_rate": 5.927322120019352e-05, + "loss": 0.0102, + "step": 46860 + }, + { + "epoch": 7.280211245728487, + "grad_norm": 0.2015589028596878, + "learning_rate": 5.925697285607844e-05, + "loss": 0.0075, + "step": 46870 + }, + { + "epoch": 7.281764523143833, + "grad_norm": 0.25773313641548157, + "learning_rate": 5.924072349963472e-05, + "loss": 0.0077, + "step": 46880 + }, + { + "epoch": 7.28331780055918, + "grad_norm": 0.43685147166252136, + "learning_rate": 5.922447313263938e-05, + "loss": 0.0086, + "step": 46890 + }, + { + "epoch": 7.284871077974526, + "grad_norm": 0.20781703293323517, + "learning_rate": 5.920822175686952e-05, + "loss": 0.0073, + "step": 46900 + }, + { + "epoch": 7.286424355389872, + "grad_norm": 0.24454118311405182, + "learning_rate": 5.919196937410239e-05, + "loss": 0.0099, + "step": 46910 + }, + { + "epoch": 7.287977632805219, + "grad_norm": 0.731916606426239, + "learning_rate": 5.917571598611529e-05, + "loss": 0.0086, + "step": 46920 + }, + { + "epoch": 7.289530910220566, + "grad_norm": 0.4981837570667267, + "learning_rate": 5.9159461594685686e-05, + "loss": 0.0125, + "step": 46930 + }, + { + "epoch": 7.291084187635912, + "grad_norm": 0.2399885356426239, + "learning_rate": 5.914320620159114e-05, + "loss": 0.0095, + "step": 46940 + }, + { + "epoch": 7.292637465051258, + "grad_norm": 0.3255470097064972, + "learning_rate": 5.912694980860931e-05, + "loss": 0.0079, + "step": 46950 + }, + { + "epoch": 7.294190742466604, + "grad_norm": 0.3730412423610687, + "learning_rate": 5.9110692417517964e-05, + "loss": 0.0068, + "step": 46960 + }, + { + "epoch": 7.295744019881951, + "grad_norm": 0.22674685716629028, + "learning_rate": 5.9094434030095e-05, + "loss": 0.0073, + "step": 46970 + }, + { + "epoch": 7.297297297297297, + "grad_norm": 0.377330482006073, + "learning_rate": 5.907817464811841e-05, + "loss": 0.0088, + "step": 46980 + }, + { + "epoch": 7.2988505747126435, + "grad_norm": 0.2723677456378937, + "learning_rate": 5.906191427336627e-05, + "loss": 0.0093, + "step": 46990 + }, + { + "epoch": 7.30040385212799, + "grad_norm": 0.32228878140449524, + "learning_rate": 5.9045652907616835e-05, + "loss": 0.01, + "step": 47000 + }, + { + "epoch": 7.301957129543337, + "grad_norm": 0.173636794090271, + "learning_rate": 5.902939055264838e-05, + "loss": 0.0079, + "step": 47010 + }, + { + "epoch": 7.303510406958683, + "grad_norm": 0.40573108196258545, + "learning_rate": 5.901312721023935e-05, + "loss": 0.0099, + "step": 47020 + }, + { + "epoch": 7.305063684374029, + "grad_norm": 0.26681214570999146, + "learning_rate": 5.8996862882168294e-05, + "loss": 0.0061, + "step": 47030 + }, + { + "epoch": 7.306616961789375, + "grad_norm": 0.2041555494070053, + "learning_rate": 5.8980597570213826e-05, + "loss": 0.0079, + "step": 47040 + }, + { + "epoch": 7.308170239204722, + "grad_norm": 0.21745537221431732, + "learning_rate": 5.896433127615471e-05, + "loss": 0.0072, + "step": 47050 + }, + { + "epoch": 7.3097235166200685, + "grad_norm": 0.280108243227005, + "learning_rate": 5.894806400176981e-05, + "loss": 0.0077, + "step": 47060 + }, + { + "epoch": 7.311276794035415, + "grad_norm": 0.28191548585891724, + "learning_rate": 5.893179574883808e-05, + "loss": 0.0079, + "step": 47070 + }, + { + "epoch": 7.312830071450761, + "grad_norm": 0.5074781775474548, + "learning_rate": 5.8915526519138585e-05, + "loss": 0.0084, + "step": 47080 + }, + { + "epoch": 7.314383348866108, + "grad_norm": 0.22991210222244263, + "learning_rate": 5.889925631445053e-05, + "loss": 0.0087, + "step": 47090 + }, + { + "epoch": 7.315936626281454, + "grad_norm": 0.28453710675239563, + "learning_rate": 5.888298513655318e-05, + "loss": 0.0082, + "step": 47100 + }, + { + "epoch": 7.3174899036968, + "grad_norm": 0.27293553948402405, + "learning_rate": 5.886671298722592e-05, + "loss": 0.0087, + "step": 47110 + }, + { + "epoch": 7.319043181112146, + "grad_norm": 0.2987730801105499, + "learning_rate": 5.885043986824828e-05, + "loss": 0.0099, + "step": 47120 + }, + { + "epoch": 7.320596458527493, + "grad_norm": 0.21957919001579285, + "learning_rate": 5.883416578139982e-05, + "loss": 0.0097, + "step": 47130 + }, + { + "epoch": 7.32214973594284, + "grad_norm": 0.2619243860244751, + "learning_rate": 5.881789072846029e-05, + "loss": 0.0075, + "step": 47140 + }, + { + "epoch": 7.323703013358186, + "grad_norm": 0.2107957899570465, + "learning_rate": 5.8801614711209474e-05, + "loss": 0.008, + "step": 47150 + }, + { + "epoch": 7.325256290773532, + "grad_norm": 0.2436595857143402, + "learning_rate": 5.878533773142732e-05, + "loss": 0.0073, + "step": 47160 + }, + { + "epoch": 7.326809568188878, + "grad_norm": 0.3029894530773163, + "learning_rate": 5.8769059790893834e-05, + "loss": 0.0083, + "step": 47170 + }, + { + "epoch": 7.328362845604225, + "grad_norm": 0.22379732131958008, + "learning_rate": 5.8752780891389145e-05, + "loss": 0.0072, + "step": 47180 + }, + { + "epoch": 7.329916123019571, + "grad_norm": 0.26571762561798096, + "learning_rate": 5.873650103469351e-05, + "loss": 0.01, + "step": 47190 + }, + { + "epoch": 7.3314694004349175, + "grad_norm": 0.3359563648700714, + "learning_rate": 5.872022022258725e-05, + "loss": 0.0059, + "step": 47200 + }, + { + "epoch": 7.3330226778502645, + "grad_norm": 0.1696869432926178, + "learning_rate": 5.8703938456850814e-05, + "loss": 0.0072, + "step": 47210 + }, + { + "epoch": 7.334575955265611, + "grad_norm": 0.21628907322883606, + "learning_rate": 5.868765573926477e-05, + "loss": 0.0061, + "step": 47220 + }, + { + "epoch": 7.336129232680957, + "grad_norm": 0.26451975107192993, + "learning_rate": 5.867137207160974e-05, + "loss": 0.0084, + "step": 47230 + }, + { + "epoch": 7.337682510096303, + "grad_norm": 0.6060254573822021, + "learning_rate": 5.86550874556665e-05, + "loss": 0.0053, + "step": 47240 + }, + { + "epoch": 7.339235787511649, + "grad_norm": 0.3174572288990021, + "learning_rate": 5.863880189321592e-05, + "loss": 0.0062, + "step": 47250 + }, + { + "epoch": 7.340789064926996, + "grad_norm": 0.6966538429260254, + "learning_rate": 5.862251538603894e-05, + "loss": 0.0065, + "step": 47260 + }, + { + "epoch": 7.342342342342342, + "grad_norm": 0.31026172637939453, + "learning_rate": 5.8606227935916656e-05, + "loss": 0.0058, + "step": 47270 + }, + { + "epoch": 7.343895619757689, + "grad_norm": 0.18543140590190887, + "learning_rate": 5.858993954463021e-05, + "loss": 0.0059, + "step": 47280 + }, + { + "epoch": 7.345448897173035, + "grad_norm": 0.20123253762722015, + "learning_rate": 5.8573650213960896e-05, + "loss": 0.0066, + "step": 47290 + }, + { + "epoch": 7.347002174588382, + "grad_norm": 0.370113730430603, + "learning_rate": 5.8557359945690084e-05, + "loss": 0.0059, + "step": 47300 + }, + { + "epoch": 7.348555452003728, + "grad_norm": 0.21075467765331268, + "learning_rate": 5.8541068741599246e-05, + "loss": 0.0072, + "step": 47310 + }, + { + "epoch": 7.350108729419074, + "grad_norm": 0.36883705854415894, + "learning_rate": 5.8524776603469985e-05, + "loss": 0.0119, + "step": 47320 + }, + { + "epoch": 7.35166200683442, + "grad_norm": 0.24943917989730835, + "learning_rate": 5.850848353308397e-05, + "loss": 0.0064, + "step": 47330 + }, + { + "epoch": 7.353215284249767, + "grad_norm": 0.4935922920703888, + "learning_rate": 5.849218953222297e-05, + "loss": 0.0073, + "step": 47340 + }, + { + "epoch": 7.3547685616651135, + "grad_norm": 0.424881249666214, + "learning_rate": 5.847589460266891e-05, + "loss": 0.0086, + "step": 47350 + }, + { + "epoch": 7.35632183908046, + "grad_norm": 0.3762344419956207, + "learning_rate": 5.845959874620375e-05, + "loss": 0.0079, + "step": 47360 + }, + { + "epoch": 7.357875116495806, + "grad_norm": 0.3681039810180664, + "learning_rate": 5.84433019646096e-05, + "loss": 0.0075, + "step": 47370 + }, + { + "epoch": 7.359428393911153, + "grad_norm": 0.19660350680351257, + "learning_rate": 5.842700425966863e-05, + "loss": 0.0074, + "step": 47380 + }, + { + "epoch": 7.360981671326499, + "grad_norm": 0.16823576390743256, + "learning_rate": 5.841070563316315e-05, + "loss": 0.0081, + "step": 47390 + }, + { + "epoch": 7.362534948741845, + "grad_norm": 0.1824968010187149, + "learning_rate": 5.839440608687554e-05, + "loss": 0.0058, + "step": 47400 + }, + { + "epoch": 7.364088226157191, + "grad_norm": 0.18529045581817627, + "learning_rate": 5.837810562258831e-05, + "loss": 0.0093, + "step": 47410 + }, + { + "epoch": 7.365641503572538, + "grad_norm": 0.21676230430603027, + "learning_rate": 5.836180424208405e-05, + "loss": 0.0074, + "step": 47420 + }, + { + "epoch": 7.367194780987885, + "grad_norm": 0.21774204075336456, + "learning_rate": 5.834550194714543e-05, + "loss": 0.0083, + "step": 47430 + }, + { + "epoch": 7.368748058403231, + "grad_norm": 0.1884126365184784, + "learning_rate": 5.832919873955528e-05, + "loss": 0.0066, + "step": 47440 + }, + { + "epoch": 7.370301335818577, + "grad_norm": 0.191366508603096, + "learning_rate": 5.831289462109648e-05, + "loss": 0.0069, + "step": 47450 + }, + { + "epoch": 7.371854613233924, + "grad_norm": 0.3936684727668762, + "learning_rate": 5.829658959355202e-05, + "loss": 0.0106, + "step": 47460 + }, + { + "epoch": 7.37340789064927, + "grad_norm": 0.15237806737422943, + "learning_rate": 5.8280283658704995e-05, + "loss": 0.0085, + "step": 47470 + }, + { + "epoch": 7.374961168064616, + "grad_norm": 0.3257547616958618, + "learning_rate": 5.826397681833859e-05, + "loss": 0.0084, + "step": 47480 + }, + { + "epoch": 7.3765144454799625, + "grad_norm": 0.5776005387306213, + "learning_rate": 5.824766907423612e-05, + "loss": 0.0108, + "step": 47490 + }, + { + "epoch": 7.3780677228953095, + "grad_norm": 0.2256627231836319, + "learning_rate": 5.8231360428180945e-05, + "loss": 0.0072, + "step": 47500 + }, + { + "epoch": 7.379621000310656, + "grad_norm": 0.20534102618694305, + "learning_rate": 5.821505088195658e-05, + "loss": 0.0101, + "step": 47510 + }, + { + "epoch": 7.381174277726002, + "grad_norm": 0.25810712575912476, + "learning_rate": 5.819874043734661e-05, + "loss": 0.0067, + "step": 47520 + }, + { + "epoch": 7.382727555141348, + "grad_norm": 0.2160571664571762, + "learning_rate": 5.8182429096134695e-05, + "loss": 0.0067, + "step": 47530 + }, + { + "epoch": 7.384280832556694, + "grad_norm": 0.27344435453414917, + "learning_rate": 5.816611686010465e-05, + "loss": 0.008, + "step": 47540 + }, + { + "epoch": 7.385834109972041, + "grad_norm": 0.26449427008628845, + "learning_rate": 5.814980373104033e-05, + "loss": 0.0092, + "step": 47550 + }, + { + "epoch": 7.387387387387387, + "grad_norm": 1.3633360862731934, + "learning_rate": 5.813348971072572e-05, + "loss": 0.0061, + "step": 47560 + }, + { + "epoch": 7.388940664802734, + "grad_norm": 0.23762886226177216, + "learning_rate": 5.811717480094492e-05, + "loss": 0.0066, + "step": 47570 + }, + { + "epoch": 7.39049394221808, + "grad_norm": 0.500587522983551, + "learning_rate": 5.810085900348209e-05, + "loss": 0.0084, + "step": 47580 + }, + { + "epoch": 7.392047219633427, + "grad_norm": 0.2950916290283203, + "learning_rate": 5.8084542320121483e-05, + "loss": 0.0081, + "step": 47590 + }, + { + "epoch": 7.393600497048773, + "grad_norm": 0.3877420425415039, + "learning_rate": 5.8068224752647497e-05, + "loss": 0.0091, + "step": 47600 + }, + { + "epoch": 7.395153774464119, + "grad_norm": 0.2682720720767975, + "learning_rate": 5.805190630284457e-05, + "loss": 0.0084, + "step": 47610 + }, + { + "epoch": 7.396707051879465, + "grad_norm": 0.25118011236190796, + "learning_rate": 5.803558697249729e-05, + "loss": 0.0094, + "step": 47620 + }, + { + "epoch": 7.398260329294812, + "grad_norm": 0.21699295938014984, + "learning_rate": 5.8019266763390276e-05, + "loss": 0.0102, + "step": 47630 + }, + { + "epoch": 7.3998136067101585, + "grad_norm": 0.3093324899673462, + "learning_rate": 5.800294567730833e-05, + "loss": 0.0096, + "step": 47640 + }, + { + "epoch": 7.401366884125505, + "grad_norm": 0.3659694790840149, + "learning_rate": 5.798662371603628e-05, + "loss": 0.0069, + "step": 47650 + }, + { + "epoch": 7.402920161540851, + "grad_norm": 0.38852259516716003, + "learning_rate": 5.797030088135904e-05, + "loss": 0.0088, + "step": 47660 + }, + { + "epoch": 7.404473438956198, + "grad_norm": 0.22355785965919495, + "learning_rate": 5.7953977175061705e-05, + "loss": 0.0068, + "step": 47670 + }, + { + "epoch": 7.406026716371544, + "grad_norm": 0.33616334199905396, + "learning_rate": 5.793765259892938e-05, + "loss": 0.0057, + "step": 47680 + }, + { + "epoch": 7.40757999378689, + "grad_norm": 0.22242726385593414, + "learning_rate": 5.792132715474729e-05, + "loss": 0.0076, + "step": 47690 + }, + { + "epoch": 7.409133271202236, + "grad_norm": 0.325411319732666, + "learning_rate": 5.790500084430078e-05, + "loss": 0.006, + "step": 47700 + }, + { + "epoch": 7.4106865486175835, + "grad_norm": 0.30817344784736633, + "learning_rate": 5.7888673669375274e-05, + "loss": 0.0069, + "step": 47710 + }, + { + "epoch": 7.41223982603293, + "grad_norm": 0.18348605930805206, + "learning_rate": 5.787234563175625e-05, + "loss": 0.0064, + "step": 47720 + }, + { + "epoch": 7.413793103448276, + "grad_norm": 0.29606321454048157, + "learning_rate": 5.7856016733229355e-05, + "loss": 0.0066, + "step": 47730 + }, + { + "epoch": 7.415346380863622, + "grad_norm": 0.22976434230804443, + "learning_rate": 5.7839686975580297e-05, + "loss": 0.0101, + "step": 47740 + }, + { + "epoch": 7.416899658278969, + "grad_norm": 0.2732445001602173, + "learning_rate": 5.782335636059484e-05, + "loss": 0.0058, + "step": 47750 + }, + { + "epoch": 7.418452935694315, + "grad_norm": 0.22344426810741425, + "learning_rate": 5.780702489005889e-05, + "loss": 0.007, + "step": 47760 + }, + { + "epoch": 7.420006213109661, + "grad_norm": 0.14310020208358765, + "learning_rate": 5.779069256575845e-05, + "loss": 0.007, + "step": 47770 + }, + { + "epoch": 7.4215594905250075, + "grad_norm": 0.4071583151817322, + "learning_rate": 5.7774359389479574e-05, + "loss": 0.0092, + "step": 47780 + }, + { + "epoch": 7.423112767940355, + "grad_norm": 0.2933697998523712, + "learning_rate": 5.775802536300845e-05, + "loss": 0.0064, + "step": 47790 + }, + { + "epoch": 7.424666045355701, + "grad_norm": 0.2124280482530594, + "learning_rate": 5.774169048813134e-05, + "loss": 0.0057, + "step": 47800 + }, + { + "epoch": 7.426219322771047, + "grad_norm": 0.23841798305511475, + "learning_rate": 5.77253547666346e-05, + "loss": 0.0072, + "step": 47810 + }, + { + "epoch": 7.427772600186393, + "grad_norm": 0.2826460599899292, + "learning_rate": 5.770901820030465e-05, + "loss": 0.0047, + "step": 47820 + }, + { + "epoch": 7.429325877601739, + "grad_norm": 0.2489573061466217, + "learning_rate": 5.769268079092809e-05, + "loss": 0.0063, + "step": 47830 + }, + { + "epoch": 7.430879155017086, + "grad_norm": 0.5974562168121338, + "learning_rate": 5.767634254029151e-05, + "loss": 0.0076, + "step": 47840 + }, + { + "epoch": 7.4324324324324325, + "grad_norm": 0.15899871289730072, + "learning_rate": 5.7660003450181655e-05, + "loss": 0.0054, + "step": 47850 + }, + { + "epoch": 7.433985709847779, + "grad_norm": 0.205241858959198, + "learning_rate": 5.764366352238534e-05, + "loss": 0.0073, + "step": 47860 + }, + { + "epoch": 7.435538987263125, + "grad_norm": 0.15894213318824768, + "learning_rate": 5.7627322758689474e-05, + "loss": 0.0074, + "step": 47870 + }, + { + "epoch": 7.437092264678472, + "grad_norm": 0.19981719553470612, + "learning_rate": 5.761098116088105e-05, + "loss": 0.0084, + "step": 47880 + }, + { + "epoch": 7.438645542093818, + "grad_norm": 0.25648802518844604, + "learning_rate": 5.759463873074717e-05, + "loss": 0.0071, + "step": 47890 + }, + { + "epoch": 7.440198819509164, + "grad_norm": 0.35556820034980774, + "learning_rate": 5.757829547007504e-05, + "loss": 0.0075, + "step": 47900 + }, + { + "epoch": 7.44175209692451, + "grad_norm": 0.2167748361825943, + "learning_rate": 5.756195138065189e-05, + "loss": 0.0076, + "step": 47910 + }, + { + "epoch": 7.443305374339857, + "grad_norm": 0.1700645387172699, + "learning_rate": 5.754560646426511e-05, + "loss": 0.0072, + "step": 47920 + }, + { + "epoch": 7.444858651755204, + "grad_norm": 0.17673645913600922, + "learning_rate": 5.752926072270216e-05, + "loss": 0.008, + "step": 47930 + }, + { + "epoch": 7.44641192917055, + "grad_norm": 0.3696078956127167, + "learning_rate": 5.7512914157750563e-05, + "loss": 0.0086, + "step": 47940 + }, + { + "epoch": 7.447965206585896, + "grad_norm": 0.24857482314109802, + "learning_rate": 5.749656677119798e-05, + "loss": 0.0068, + "step": 47950 + }, + { + "epoch": 7.449518484001243, + "grad_norm": 0.2275998890399933, + "learning_rate": 5.7480218564832125e-05, + "loss": 0.0076, + "step": 47960 + }, + { + "epoch": 7.451071761416589, + "grad_norm": 0.21559195220470428, + "learning_rate": 5.746386954044082e-05, + "loss": 0.0063, + "step": 47970 + }, + { + "epoch": 7.452625038831935, + "grad_norm": 0.27601736783981323, + "learning_rate": 5.744751969981195e-05, + "loss": 0.0104, + "step": 47980 + }, + { + "epoch": 7.454178316247281, + "grad_norm": 0.25437965989112854, + "learning_rate": 5.7431169044733526e-05, + "loss": 0.0049, + "step": 47990 + }, + { + "epoch": 7.4557315936626285, + "grad_norm": 1.1972461938858032, + "learning_rate": 5.741481757699364e-05, + "loss": 0.0074, + "step": 48000 + }, + { + "epoch": 7.457284871077975, + "grad_norm": 0.21597641706466675, + "learning_rate": 5.7398465298380434e-05, + "loss": 0.0097, + "step": 48010 + }, + { + "epoch": 7.458838148493321, + "grad_norm": 0.23838455975055695, + "learning_rate": 5.7382112210682193e-05, + "loss": 0.0063, + "step": 48020 + }, + { + "epoch": 7.460391425908667, + "grad_norm": 0.2789250314235687, + "learning_rate": 5.7365758315687266e-05, + "loss": 0.007, + "step": 48030 + }, + { + "epoch": 7.461944703324014, + "grad_norm": 0.26426437497138977, + "learning_rate": 5.734940361518407e-05, + "loss": 0.0074, + "step": 48040 + }, + { + "epoch": 7.46349798073936, + "grad_norm": 0.16477563977241516, + "learning_rate": 5.733304811096116e-05, + "loss": 0.0078, + "step": 48050 + }, + { + "epoch": 7.465051258154706, + "grad_norm": 0.29216232895851135, + "learning_rate": 5.731669180480713e-05, + "loss": 0.0044, + "step": 48060 + }, + { + "epoch": 7.4666045355700525, + "grad_norm": 0.26898616552352905, + "learning_rate": 5.730033469851067e-05, + "loss": 0.0068, + "step": 48070 + }, + { + "epoch": 7.4681578129854, + "grad_norm": 0.18740972876548767, + "learning_rate": 5.7283976793860607e-05, + "loss": 0.0048, + "step": 48080 + }, + { + "epoch": 7.469711090400746, + "grad_norm": 0.29778167605400085, + "learning_rate": 5.726761809264577e-05, + "loss": 0.0095, + "step": 48090 + }, + { + "epoch": 7.471264367816092, + "grad_norm": 0.5180748701095581, + "learning_rate": 5.7251258596655155e-05, + "loss": 0.0062, + "step": 48100 + }, + { + "epoch": 7.472817645231438, + "grad_norm": 0.16015625, + "learning_rate": 5.72348983076778e-05, + "loss": 0.005, + "step": 48110 + }, + { + "epoch": 7.474370922646784, + "grad_norm": 0.2557743787765503, + "learning_rate": 5.7218537227502854e-05, + "loss": 0.0073, + "step": 48120 + }, + { + "epoch": 7.475924200062131, + "grad_norm": 0.2774520814418793, + "learning_rate": 5.720217535791951e-05, + "loss": 0.0087, + "step": 48130 + }, + { + "epoch": 7.4774774774774775, + "grad_norm": 0.15837551653385162, + "learning_rate": 5.718581270071711e-05, + "loss": 0.0059, + "step": 48140 + }, + { + "epoch": 7.479030754892824, + "grad_norm": 0.20232850313186646, + "learning_rate": 5.716944925768505e-05, + "loss": 0.0063, + "step": 48150 + }, + { + "epoch": 7.480584032308171, + "grad_norm": 0.1742120236158371, + "learning_rate": 5.7153085030612786e-05, + "loss": 0.0077, + "step": 48160 + }, + { + "epoch": 7.482137309723517, + "grad_norm": 0.32073718309402466, + "learning_rate": 5.713672002128989e-05, + "loss": 0.007, + "step": 48170 + }, + { + "epoch": 7.483690587138863, + "grad_norm": 0.19066184759140015, + "learning_rate": 5.712035423150602e-05, + "loss": 0.006, + "step": 48180 + }, + { + "epoch": 7.485243864554209, + "grad_norm": 0.20509134232997894, + "learning_rate": 5.710398766305094e-05, + "loss": 0.0078, + "step": 48190 + }, + { + "epoch": 7.486797141969555, + "grad_norm": 0.1588825285434723, + "learning_rate": 5.708762031771442e-05, + "loss": 0.006, + "step": 48200 + }, + { + "epoch": 7.488350419384902, + "grad_norm": 0.27315303683280945, + "learning_rate": 5.707125219728642e-05, + "loss": 0.0068, + "step": 48210 + }, + { + "epoch": 7.489903696800249, + "grad_norm": 0.3831600546836853, + "learning_rate": 5.7054883303556905e-05, + "loss": 0.0102, + "step": 48220 + }, + { + "epoch": 7.491456974215595, + "grad_norm": 0.39769986271858215, + "learning_rate": 5.703851363831595e-05, + "loss": 0.0095, + "step": 48230 + }, + { + "epoch": 7.493010251630941, + "grad_norm": 0.28950273990631104, + "learning_rate": 5.702214320335374e-05, + "loss": 0.0079, + "step": 48240 + }, + { + "epoch": 7.494563529046288, + "grad_norm": 0.6819275617599487, + "learning_rate": 5.7005772000460514e-05, + "loss": 0.0075, + "step": 48250 + }, + { + "epoch": 7.496116806461634, + "grad_norm": 0.46989768743515015, + "learning_rate": 5.698940003142656e-05, + "loss": 0.0091, + "step": 48260 + }, + { + "epoch": 7.49767008387698, + "grad_norm": 0.8239760398864746, + "learning_rate": 5.697302729804236e-05, + "loss": 0.0074, + "step": 48270 + }, + { + "epoch": 7.4992233612923265, + "grad_norm": 0.4996587336063385, + "learning_rate": 5.695665380209837e-05, + "loss": 0.0093, + "step": 48280 + }, + { + "epoch": 7.5007766387076735, + "grad_norm": 0.14707542955875397, + "learning_rate": 5.6940279545385165e-05, + "loss": 0.0048, + "step": 48290 + }, + { + "epoch": 7.50232991612302, + "grad_norm": 0.5337914824485779, + "learning_rate": 5.692390452969344e-05, + "loss": 0.0072, + "step": 48300 + }, + { + "epoch": 7.503883193538366, + "grad_norm": 0.2232937514781952, + "learning_rate": 5.690752875681392e-05, + "loss": 0.0103, + "step": 48310 + }, + { + "epoch": 7.505436470953712, + "grad_norm": 0.31402361392974854, + "learning_rate": 5.6891152228537435e-05, + "loss": 0.0098, + "step": 48320 + }, + { + "epoch": 7.506989748369059, + "grad_norm": 0.2272518426179886, + "learning_rate": 5.687477494665492e-05, + "loss": 0.0073, + "step": 48330 + }, + { + "epoch": 7.508543025784405, + "grad_norm": 0.2044084072113037, + "learning_rate": 5.685839691295733e-05, + "loss": 0.0083, + "step": 48340 + }, + { + "epoch": 7.510096303199751, + "grad_norm": 0.21622301638126373, + "learning_rate": 5.6842018129235786e-05, + "loss": 0.0062, + "step": 48350 + }, + { + "epoch": 7.511649580615098, + "grad_norm": 0.20585185289382935, + "learning_rate": 5.6825638597281404e-05, + "loss": 0.0064, + "step": 48360 + }, + { + "epoch": 7.513202858030445, + "grad_norm": 0.19905877113342285, + "learning_rate": 5.680925831888546e-05, + "loss": 0.0051, + "step": 48370 + }, + { + "epoch": 7.514756135445791, + "grad_norm": 0.3333401679992676, + "learning_rate": 5.6792877295839274e-05, + "loss": 0.0068, + "step": 48380 + }, + { + "epoch": 7.516309412861137, + "grad_norm": 0.2595142722129822, + "learning_rate": 5.6776495529934224e-05, + "loss": 0.0107, + "step": 48390 + }, + { + "epoch": 7.517862690276483, + "grad_norm": 0.6170368194580078, + "learning_rate": 5.6760113022961824e-05, + "loss": 0.0108, + "step": 48400 + }, + { + "epoch": 7.519415967691829, + "grad_norm": 0.21724693477153778, + "learning_rate": 5.6743729776713617e-05, + "loss": 0.0064, + "step": 48410 + }, + { + "epoch": 7.520969245107176, + "grad_norm": 0.3179955780506134, + "learning_rate": 5.672734579298126e-05, + "loss": 0.0065, + "step": 48420 + }, + { + "epoch": 7.5225225225225225, + "grad_norm": 0.23747774958610535, + "learning_rate": 5.671096107355649e-05, + "loss": 0.0071, + "step": 48430 + }, + { + "epoch": 7.524075799937869, + "grad_norm": 0.29255959391593933, + "learning_rate": 5.66945756202311e-05, + "loss": 0.0105, + "step": 48440 + }, + { + "epoch": 7.525629077353216, + "grad_norm": 0.262992262840271, + "learning_rate": 5.667818943479699e-05, + "loss": 0.0092, + "step": 48450 + }, + { + "epoch": 7.527182354768562, + "grad_norm": 0.23993565142154694, + "learning_rate": 5.666180251904612e-05, + "loss": 0.0116, + "step": 48460 + }, + { + "epoch": 7.528735632183908, + "grad_norm": 0.38012245297431946, + "learning_rate": 5.6645414874770555e-05, + "loss": 0.0071, + "step": 48470 + }, + { + "epoch": 7.530288909599254, + "grad_norm": 0.26805880665779114, + "learning_rate": 5.66290265037624e-05, + "loss": 0.0063, + "step": 48480 + }, + { + "epoch": 7.5318421870146, + "grad_norm": 0.39218056201934814, + "learning_rate": 5.661263740781386e-05, + "loss": 0.0079, + "step": 48490 + }, + { + "epoch": 7.5333954644299475, + "grad_norm": 0.247588112950325, + "learning_rate": 5.6596247588717254e-05, + "loss": 0.0081, + "step": 48500 + }, + { + "epoch": 7.534948741845294, + "grad_norm": 0.1993151307106018, + "learning_rate": 5.6579857048264926e-05, + "loss": 0.0063, + "step": 48510 + }, + { + "epoch": 7.53650201926064, + "grad_norm": 0.2102869749069214, + "learning_rate": 5.6563465788249314e-05, + "loss": 0.0101, + "step": 48520 + }, + { + "epoch": 7.538055296675986, + "grad_norm": 0.1999887377023697, + "learning_rate": 5.6547073810462956e-05, + "loss": 0.0089, + "step": 48530 + }, + { + "epoch": 7.539608574091333, + "grad_norm": 0.20301523804664612, + "learning_rate": 5.653068111669846e-05, + "loss": 0.0067, + "step": 48540 + }, + { + "epoch": 7.541161851506679, + "grad_norm": 0.3816691040992737, + "learning_rate": 5.651428770874848e-05, + "loss": 0.0084, + "step": 48550 + }, + { + "epoch": 7.542715128922025, + "grad_norm": 0.18250644207000732, + "learning_rate": 5.64978935884058e-05, + "loss": 0.0055, + "step": 48560 + }, + { + "epoch": 7.5442684063373715, + "grad_norm": 0.16377076506614685, + "learning_rate": 5.6481498757463244e-05, + "loss": 0.0091, + "step": 48570 + }, + { + "epoch": 7.545821683752719, + "grad_norm": 0.20234356820583344, + "learning_rate": 5.646510321771373e-05, + "loss": 0.0068, + "step": 48580 + }, + { + "epoch": 7.547374961168065, + "grad_norm": 0.2164985090494156, + "learning_rate": 5.644870697095024e-05, + "loss": 0.0093, + "step": 48590 + }, + { + "epoch": 7.548928238583411, + "grad_norm": 0.2124345749616623, + "learning_rate": 5.643231001896586e-05, + "loss": 0.0089, + "step": 48600 + }, + { + "epoch": 7.550481515998757, + "grad_norm": 0.42970529198646545, + "learning_rate": 5.6415912363553726e-05, + "loss": 0.0096, + "step": 48610 + }, + { + "epoch": 7.552034793414104, + "grad_norm": 0.2507791519165039, + "learning_rate": 5.639951400650706e-05, + "loss": 0.0077, + "step": 48620 + }, + { + "epoch": 7.55358807082945, + "grad_norm": 0.4422670304775238, + "learning_rate": 5.6383114949619165e-05, + "loss": 0.0069, + "step": 48630 + }, + { + "epoch": 7.555141348244796, + "grad_norm": 0.4545094072818756, + "learning_rate": 5.636671519468342e-05, + "loss": 0.0113, + "step": 48640 + }, + { + "epoch": 7.556694625660143, + "grad_norm": 0.193319633603096, + "learning_rate": 5.635031474349327e-05, + "loss": 0.0056, + "step": 48650 + }, + { + "epoch": 7.55824790307549, + "grad_norm": 0.2346373200416565, + "learning_rate": 5.6333913597842246e-05, + "loss": 0.0085, + "step": 48660 + }, + { + "epoch": 7.559801180490836, + "grad_norm": 0.26107731461524963, + "learning_rate": 5.6317511759523955e-05, + "loss": 0.0073, + "step": 48670 + }, + { + "epoch": 7.561354457906182, + "grad_norm": 0.28542640805244446, + "learning_rate": 5.630110923033207e-05, + "loss": 0.007, + "step": 48680 + }, + { + "epoch": 7.562907735321528, + "grad_norm": 0.4480079114437103, + "learning_rate": 5.628470601206036e-05, + "loss": 0.0097, + "step": 48690 + }, + { + "epoch": 7.564461012736874, + "grad_norm": 0.4173465669155121, + "learning_rate": 5.626830210650266e-05, + "loss": 0.0074, + "step": 48700 + }, + { + "epoch": 7.566014290152221, + "grad_norm": 0.28982096910476685, + "learning_rate": 5.625189751545285e-05, + "loss": 0.0089, + "step": 48710 + }, + { + "epoch": 7.5675675675675675, + "grad_norm": 0.23827165365219116, + "learning_rate": 5.6235492240704936e-05, + "loss": 0.0074, + "step": 48720 + }, + { + "epoch": 7.569120844982914, + "grad_norm": 0.2388128787279129, + "learning_rate": 5.621908628405296e-05, + "loss": 0.0067, + "step": 48730 + }, + { + "epoch": 7.570674122398261, + "grad_norm": 0.1696355789899826, + "learning_rate": 5.620267964729106e-05, + "loss": 0.0073, + "step": 48740 + }, + { + "epoch": 7.572227399813607, + "grad_norm": 0.37102794647216797, + "learning_rate": 5.618627233221344e-05, + "loss": 0.0083, + "step": 48750 + }, + { + "epoch": 7.573780677228953, + "grad_norm": 0.25747865438461304, + "learning_rate": 5.616986434061438e-05, + "loss": 0.0061, + "step": 48760 + }, + { + "epoch": 7.575333954644299, + "grad_norm": 0.22777429223060608, + "learning_rate": 5.615345567428822e-05, + "loss": 0.006, + "step": 48770 + }, + { + "epoch": 7.576887232059645, + "grad_norm": 0.24594731628894806, + "learning_rate": 5.613704633502941e-05, + "loss": 0.0051, + "step": 48780 + }, + { + "epoch": 7.5784405094749925, + "grad_norm": 0.23993641138076782, + "learning_rate": 5.612063632463245e-05, + "loss": 0.0106, + "step": 48790 + }, + { + "epoch": 7.579993786890339, + "grad_norm": 0.2416509985923767, + "learning_rate": 5.610422564489188e-05, + "loss": 0.0078, + "step": 48800 + }, + { + "epoch": 7.581547064305685, + "grad_norm": 0.25878310203552246, + "learning_rate": 5.608781429760239e-05, + "loss": 0.0091, + "step": 48810 + }, + { + "epoch": 7.583100341721032, + "grad_norm": 0.22145168483257294, + "learning_rate": 5.607140228455866e-05, + "loss": 0.0075, + "step": 48820 + }, + { + "epoch": 7.584653619136378, + "grad_norm": 0.18961267173290253, + "learning_rate": 5.605498960755553e-05, + "loss": 0.0079, + "step": 48830 + }, + { + "epoch": 7.586206896551724, + "grad_norm": 0.1372372955083847, + "learning_rate": 5.603857626838782e-05, + "loss": 0.0063, + "step": 48840 + }, + { + "epoch": 7.58776017396707, + "grad_norm": 0.14394429326057434, + "learning_rate": 5.60221622688505e-05, + "loss": 0.0076, + "step": 48850 + }, + { + "epoch": 7.5893134513824165, + "grad_norm": 0.18128781020641327, + "learning_rate": 5.6005747610738565e-05, + "loss": 0.0058, + "step": 48860 + }, + { + "epoch": 7.590866728797764, + "grad_norm": 0.21123212575912476, + "learning_rate": 5.59893322958471e-05, + "loss": 0.0085, + "step": 48870 + }, + { + "epoch": 7.59242000621311, + "grad_norm": 0.23362566530704498, + "learning_rate": 5.5972916325971256e-05, + "loss": 0.0072, + "step": 48880 + }, + { + "epoch": 7.593973283628456, + "grad_norm": 0.23542797565460205, + "learning_rate": 5.595649970290628e-05, + "loss": 0.0068, + "step": 48890 + }, + { + "epoch": 7.595526561043802, + "grad_norm": 0.4188176393508911, + "learning_rate": 5.5940082428447426e-05, + "loss": 0.0077, + "step": 48900 + }, + { + "epoch": 7.597079838459149, + "grad_norm": 0.20529140532016754, + "learning_rate": 5.592366450439012e-05, + "loss": 0.0069, + "step": 48910 + }, + { + "epoch": 7.598633115874495, + "grad_norm": 0.17762647569179535, + "learning_rate": 5.590724593252975e-05, + "loss": 0.0052, + "step": 48920 + }, + { + "epoch": 7.6001863932898415, + "grad_norm": 0.22347351908683777, + "learning_rate": 5.589082671466184e-05, + "loss": 0.0075, + "step": 48930 + }, + { + "epoch": 7.601739670705188, + "grad_norm": 0.4105585813522339, + "learning_rate": 5.587440685258199e-05, + "loss": 0.007, + "step": 48940 + }, + { + "epoch": 7.603292948120535, + "grad_norm": 0.28544509410858154, + "learning_rate": 5.585798634808583e-05, + "loss": 0.0084, + "step": 48950 + }, + { + "epoch": 7.604846225535881, + "grad_norm": 0.4562838673591614, + "learning_rate": 5.584156520296909e-05, + "loss": 0.0063, + "step": 48960 + }, + { + "epoch": 7.606399502951227, + "grad_norm": 0.36537715792655945, + "learning_rate": 5.582514341902757e-05, + "loss": 0.0059, + "step": 48970 + }, + { + "epoch": 7.607952780366573, + "grad_norm": 0.1848907172679901, + "learning_rate": 5.580872099805713e-05, + "loss": 0.006, + "step": 48980 + }, + { + "epoch": 7.609506057781919, + "grad_norm": 0.3417721688747406, + "learning_rate": 5.5792297941853674e-05, + "loss": 0.0072, + "step": 48990 + }, + { + "epoch": 7.611059335197266, + "grad_norm": 0.2533627450466156, + "learning_rate": 5.5775874252213247e-05, + "loss": 0.0092, + "step": 49000 + }, + { + "epoch": 7.612612612612613, + "grad_norm": 0.19252334535121918, + "learning_rate": 5.575944993093189e-05, + "loss": 0.006, + "step": 49010 + }, + { + "epoch": 7.614165890027959, + "grad_norm": 0.2760136127471924, + "learning_rate": 5.574302497980574e-05, + "loss": 0.0098, + "step": 49020 + }, + { + "epoch": 7.615719167443306, + "grad_norm": 0.33339568972587585, + "learning_rate": 5.5726599400631005e-05, + "loss": 0.0085, + "step": 49030 + }, + { + "epoch": 7.617272444858652, + "grad_norm": 0.19328173995018005, + "learning_rate": 5.571017319520399e-05, + "loss": 0.0053, + "step": 49040 + }, + { + "epoch": 7.618825722273998, + "grad_norm": 0.2191249430179596, + "learning_rate": 5.5693746365320984e-05, + "loss": 0.0081, + "step": 49050 + }, + { + "epoch": 7.620378999689344, + "grad_norm": 0.1894676685333252, + "learning_rate": 5.567731891277848e-05, + "loss": 0.0063, + "step": 49060 + }, + { + "epoch": 7.6219322771046905, + "grad_norm": 0.25478753447532654, + "learning_rate": 5.566089083937288e-05, + "loss": 0.0052, + "step": 49070 + }, + { + "epoch": 7.6234855545200375, + "grad_norm": 0.2651360332965851, + "learning_rate": 5.564446214690079e-05, + "loss": 0.0068, + "step": 49080 + }, + { + "epoch": 7.625038831935384, + "grad_norm": 0.22157616913318634, + "learning_rate": 5.56280328371588e-05, + "loss": 0.0065, + "step": 49090 + }, + { + "epoch": 7.62659210935073, + "grad_norm": 0.17370983958244324, + "learning_rate": 5.5611602911943596e-05, + "loss": 0.0089, + "step": 49100 + }, + { + "epoch": 7.628145386766077, + "grad_norm": 0.16633716225624084, + "learning_rate": 5.559517237305194e-05, + "loss": 0.0084, + "step": 49110 + }, + { + "epoch": 7.629698664181423, + "grad_norm": 0.1629774123430252, + "learning_rate": 5.557874122228064e-05, + "loss": 0.0063, + "step": 49120 + }, + { + "epoch": 7.631251941596769, + "grad_norm": 0.30436283349990845, + "learning_rate": 5.55623094614266e-05, + "loss": 0.0074, + "step": 49130 + }, + { + "epoch": 7.632805219012115, + "grad_norm": 0.12693054974079132, + "learning_rate": 5.5545877092286743e-05, + "loss": 0.0069, + "step": 49140 + }, + { + "epoch": 7.634358496427462, + "grad_norm": 0.21702046692371368, + "learning_rate": 5.552944411665811e-05, + "loss": 0.0084, + "step": 49150 + }, + { + "epoch": 7.635911773842809, + "grad_norm": 0.34754621982574463, + "learning_rate": 5.5513010536337795e-05, + "loss": 0.0062, + "step": 49160 + }, + { + "epoch": 7.637465051258155, + "grad_norm": 0.16115611791610718, + "learning_rate": 5.5496576353122934e-05, + "loss": 0.0071, + "step": 49170 + }, + { + "epoch": 7.639018328673501, + "grad_norm": 0.22117017209529877, + "learning_rate": 5.548014156881074e-05, + "loss": 0.0048, + "step": 49180 + }, + { + "epoch": 7.640571606088847, + "grad_norm": 0.12749622762203217, + "learning_rate": 5.5463706185198494e-05, + "loss": 0.0065, + "step": 49190 + }, + { + "epoch": 7.642124883504194, + "grad_norm": 0.1508040875196457, + "learning_rate": 5.544727020408358e-05, + "loss": 0.0063, + "step": 49200 + }, + { + "epoch": 7.64367816091954, + "grad_norm": 0.07396609336137772, + "learning_rate": 5.5430833627263366e-05, + "loss": 0.0045, + "step": 49210 + }, + { + "epoch": 7.6452314383348865, + "grad_norm": 0.28058838844299316, + "learning_rate": 5.5414396456535354e-05, + "loss": 0.0058, + "step": 49220 + }, + { + "epoch": 7.646784715750233, + "grad_norm": 0.21938830614089966, + "learning_rate": 5.539795869369709e-05, + "loss": 0.0062, + "step": 49230 + }, + { + "epoch": 7.64833799316558, + "grad_norm": 0.22277314960956573, + "learning_rate": 5.538152034054618e-05, + "loss": 0.005, + "step": 49240 + }, + { + "epoch": 7.649891270580926, + "grad_norm": 0.1312435120344162, + "learning_rate": 5.536508139888028e-05, + "loss": 0.007, + "step": 49250 + }, + { + "epoch": 7.651444547996272, + "grad_norm": 0.29284417629241943, + "learning_rate": 5.534864187049716e-05, + "loss": 0.0091, + "step": 49260 + }, + { + "epoch": 7.652997825411618, + "grad_norm": 0.3762878477573395, + "learning_rate": 5.5332201757194615e-05, + "loss": 0.0048, + "step": 49270 + }, + { + "epoch": 7.654551102826965, + "grad_norm": 0.2153349071741104, + "learning_rate": 5.531576106077049e-05, + "loss": 0.0065, + "step": 49280 + }, + { + "epoch": 7.656104380242311, + "grad_norm": 0.26941007375717163, + "learning_rate": 5.529931978302272e-05, + "loss": 0.0051, + "step": 49290 + }, + { + "epoch": 7.657657657657658, + "grad_norm": 0.21502865850925446, + "learning_rate": 5.528287792574932e-05, + "loss": 0.0072, + "step": 49300 + }, + { + "epoch": 7.659210935073004, + "grad_norm": 0.17225806415081024, + "learning_rate": 5.5266435490748324e-05, + "loss": 0.0045, + "step": 49310 + }, + { + "epoch": 7.660764212488351, + "grad_norm": 0.3386533260345459, + "learning_rate": 5.524999247981787e-05, + "loss": 0.0068, + "step": 49320 + }, + { + "epoch": 7.662317489903697, + "grad_norm": 0.30525198578834534, + "learning_rate": 5.523354889475613e-05, + "loss": 0.0062, + "step": 49330 + }, + { + "epoch": 7.663870767319043, + "grad_norm": 0.1971990466117859, + "learning_rate": 5.521710473736134e-05, + "loss": 0.0046, + "step": 49340 + }, + { + "epoch": 7.665424044734389, + "grad_norm": 0.29584500193595886, + "learning_rate": 5.5200660009431826e-05, + "loss": 0.0071, + "step": 49350 + }, + { + "epoch": 7.6669773221497355, + "grad_norm": 0.2612174451351166, + "learning_rate": 5.518421471276596e-05, + "loss": 0.0064, + "step": 49360 + }, + { + "epoch": 7.6685305995650825, + "grad_norm": 0.3350299298763275, + "learning_rate": 5.5167768849162174e-05, + "loss": 0.0072, + "step": 49370 + }, + { + "epoch": 7.670083876980429, + "grad_norm": 0.25878170132637024, + "learning_rate": 5.515132242041893e-05, + "loss": 0.0077, + "step": 49380 + }, + { + "epoch": 7.671637154395775, + "grad_norm": 0.5495715141296387, + "learning_rate": 5.513487542833483e-05, + "loss": 0.009, + "step": 49390 + }, + { + "epoch": 7.673190431811122, + "grad_norm": 0.1282859593629837, + "learning_rate": 5.5118427874708466e-05, + "loss": 0.0077, + "step": 49400 + }, + { + "epoch": 7.674743709226468, + "grad_norm": 0.1735687106847763, + "learning_rate": 5.510197976133853e-05, + "loss": 0.0067, + "step": 49410 + }, + { + "epoch": 7.676296986641814, + "grad_norm": 0.1217457503080368, + "learning_rate": 5.508553109002376e-05, + "loss": 0.0082, + "step": 49420 + }, + { + "epoch": 7.67785026405716, + "grad_norm": 0.20877280831336975, + "learning_rate": 5.5069081862562957e-05, + "loss": 0.0077, + "step": 49430 + }, + { + "epoch": 7.679403541472507, + "grad_norm": 0.13828933238983154, + "learning_rate": 5.5052632080754965e-05, + "loss": 0.0072, + "step": 49440 + }, + { + "epoch": 7.680956818887854, + "grad_norm": 0.23203536868095398, + "learning_rate": 5.503618174639874e-05, + "loss": 0.0067, + "step": 49450 + }, + { + "epoch": 7.6825100963032, + "grad_norm": 0.2789646089076996, + "learning_rate": 5.501973086129325e-05, + "loss": 0.0077, + "step": 49460 + }, + { + "epoch": 7.684063373718546, + "grad_norm": 0.24666564166545868, + "learning_rate": 5.500327942723753e-05, + "loss": 0.0071, + "step": 49470 + }, + { + "epoch": 7.685616651133892, + "grad_norm": 0.5973349809646606, + "learning_rate": 5.498682744603071e-05, + "loss": 0.0076, + "step": 49480 + }, + { + "epoch": 7.687169928549239, + "grad_norm": 0.1652902364730835, + "learning_rate": 5.497037491947194e-05, + "loss": 0.0062, + "step": 49490 + }, + { + "epoch": 7.688723205964585, + "grad_norm": 0.14823120832443237, + "learning_rate": 5.4953921849360424e-05, + "loss": 0.0087, + "step": 49500 + }, + { + "epoch": 7.6902764833799315, + "grad_norm": 0.18330809473991394, + "learning_rate": 5.493746823749547e-05, + "loss": 0.0064, + "step": 49510 + }, + { + "epoch": 7.691829760795278, + "grad_norm": 0.36540600657463074, + "learning_rate": 5.49210140856764e-05, + "loss": 0.0072, + "step": 49520 + }, + { + "epoch": 7.693383038210625, + "grad_norm": 0.22218598425388336, + "learning_rate": 5.4904559395702647e-05, + "loss": 0.0057, + "step": 49530 + }, + { + "epoch": 7.694936315625971, + "grad_norm": 0.2531247138977051, + "learning_rate": 5.488810416937364e-05, + "loss": 0.0062, + "step": 49540 + }, + { + "epoch": 7.696489593041317, + "grad_norm": 0.15101853013038635, + "learning_rate": 5.4871648408488926e-05, + "loss": 0.0072, + "step": 49550 + }, + { + "epoch": 7.698042870456663, + "grad_norm": 0.2628510594367981, + "learning_rate": 5.485519211484807e-05, + "loss": 0.0102, + "step": 49560 + }, + { + "epoch": 7.69959614787201, + "grad_norm": 0.18988794088363647, + "learning_rate": 5.4838735290250676e-05, + "loss": 0.0073, + "step": 49570 + }, + { + "epoch": 7.7011494252873565, + "grad_norm": 0.25673073530197144, + "learning_rate": 5.482227793649649e-05, + "loss": 0.0089, + "step": 49580 + }, + { + "epoch": 7.702702702702703, + "grad_norm": 0.2550608515739441, + "learning_rate": 5.480582005538524e-05, + "loss": 0.0074, + "step": 49590 + }, + { + "epoch": 7.704255980118049, + "grad_norm": 0.20475752651691437, + "learning_rate": 5.478936164871671e-05, + "loss": 0.0081, + "step": 49600 + }, + { + "epoch": 7.705809257533396, + "grad_norm": 0.6713666915893555, + "learning_rate": 5.477290271829081e-05, + "loss": 0.0063, + "step": 49610 + }, + { + "epoch": 7.707362534948742, + "grad_norm": 0.24188877642154694, + "learning_rate": 5.475644326590744e-05, + "loss": 0.0088, + "step": 49620 + }, + { + "epoch": 7.708915812364088, + "grad_norm": 0.3375801146030426, + "learning_rate": 5.473998329336658e-05, + "loss": 0.0071, + "step": 49630 + }, + { + "epoch": 7.710469089779434, + "grad_norm": 0.3853921890258789, + "learning_rate": 5.4723522802468286e-05, + "loss": 0.0097, + "step": 49640 + }, + { + "epoch": 7.7120223671947805, + "grad_norm": 0.25889307260513306, + "learning_rate": 5.470706179501264e-05, + "loss": 0.0076, + "step": 49650 + }, + { + "epoch": 7.713575644610128, + "grad_norm": 0.27624276280403137, + "learning_rate": 5.4690600272799776e-05, + "loss": 0.005, + "step": 49660 + }, + { + "epoch": 7.715128922025474, + "grad_norm": 0.28675344586372375, + "learning_rate": 5.467413823762993e-05, + "loss": 0.0068, + "step": 49670 + }, + { + "epoch": 7.71668219944082, + "grad_norm": 0.22188253700733185, + "learning_rate": 5.465767569130335e-05, + "loss": 0.0071, + "step": 49680 + }, + { + "epoch": 7.718235476856167, + "grad_norm": 0.21233466267585754, + "learning_rate": 5.464121263562036e-05, + "loss": 0.0065, + "step": 49690 + }, + { + "epoch": 7.719788754271513, + "grad_norm": 0.2941288650035858, + "learning_rate": 5.4624749072381343e-05, + "loss": 0.0054, + "step": 49700 + }, + { + "epoch": 7.721342031686859, + "grad_norm": 0.35830745100975037, + "learning_rate": 5.460828500338672e-05, + "loss": 0.007, + "step": 49710 + }, + { + "epoch": 7.7228953091022055, + "grad_norm": 0.38687095046043396, + "learning_rate": 5.459182043043698e-05, + "loss": 0.0076, + "step": 49720 + }, + { + "epoch": 7.724448586517552, + "grad_norm": 0.17764607071876526, + "learning_rate": 5.457535535533265e-05, + "loss": 0.0063, + "step": 49730 + }, + { + "epoch": 7.726001863932899, + "grad_norm": 0.2352270931005478, + "learning_rate": 5.4558889779874334e-05, + "loss": 0.0056, + "step": 49740 + }, + { + "epoch": 7.727555141348245, + "grad_norm": 0.12511754035949707, + "learning_rate": 5.454242370586269e-05, + "loss": 0.008, + "step": 49750 + }, + { + "epoch": 7.729108418763591, + "grad_norm": 0.31455618143081665, + "learning_rate": 5.452595713509843e-05, + "loss": 0.0068, + "step": 49760 + }, + { + "epoch": 7.730661696178937, + "grad_norm": 0.42874646186828613, + "learning_rate": 5.4509490069382274e-05, + "loss": 0.0078, + "step": 49770 + }, + { + "epoch": 7.732214973594284, + "grad_norm": 0.25386562943458557, + "learning_rate": 5.449302251051509e-05, + "loss": 0.0104, + "step": 49780 + }, + { + "epoch": 7.73376825100963, + "grad_norm": 0.391970694065094, + "learning_rate": 5.44765544602977e-05, + "loss": 0.0072, + "step": 49790 + }, + { + "epoch": 7.735321528424977, + "grad_norm": 0.3459779620170593, + "learning_rate": 5.446008592053107e-05, + "loss": 0.0069, + "step": 49800 + }, + { + "epoch": 7.736874805840323, + "grad_norm": 0.30399462580680847, + "learning_rate": 5.4443616893016135e-05, + "loss": 0.0099, + "step": 49810 + }, + { + "epoch": 7.73842808325567, + "grad_norm": 0.19964586198329926, + "learning_rate": 5.4427147379553934e-05, + "loss": 0.008, + "step": 49820 + }, + { + "epoch": 7.739981360671016, + "grad_norm": 0.31813156604766846, + "learning_rate": 5.441067738194556e-05, + "loss": 0.0068, + "step": 49830 + }, + { + "epoch": 7.741534638086362, + "grad_norm": 0.1844913512468338, + "learning_rate": 5.439420690199214e-05, + "loss": 0.0063, + "step": 49840 + }, + { + "epoch": 7.743087915501708, + "grad_norm": 0.2499237060546875, + "learning_rate": 5.437773594149484e-05, + "loss": 0.0063, + "step": 49850 + }, + { + "epoch": 7.744641192917055, + "grad_norm": 0.5120940804481506, + "learning_rate": 5.436126450225495e-05, + "loss": 0.0074, + "step": 49860 + }, + { + "epoch": 7.7461944703324015, + "grad_norm": 0.491140216588974, + "learning_rate": 5.434479258607371e-05, + "loss": 0.0063, + "step": 49870 + }, + { + "epoch": 7.747747747747748, + "grad_norm": 0.2872574031352997, + "learning_rate": 5.43283201947525e-05, + "loss": 0.007, + "step": 49880 + }, + { + "epoch": 7.749301025163094, + "grad_norm": 0.33289945125579834, + "learning_rate": 5.431184733009268e-05, + "loss": 0.0082, + "step": 49890 + }, + { + "epoch": 7.750854302578441, + "grad_norm": 0.2983967065811157, + "learning_rate": 5.4295373993895736e-05, + "loss": 0.0059, + "step": 49900 + }, + { + "epoch": 7.752407579993787, + "grad_norm": 0.13793382048606873, + "learning_rate": 5.4278900187963157e-05, + "loss": 0.0095, + "step": 49910 + }, + { + "epoch": 7.753960857409133, + "grad_norm": 0.14138807356357574, + "learning_rate": 5.426242591409646e-05, + "loss": 0.0098, + "step": 49920 + }, + { + "epoch": 7.755514134824479, + "grad_norm": 0.28482064604759216, + "learning_rate": 5.42459511740973e-05, + "loss": 0.0089, + "step": 49930 + }, + { + "epoch": 7.7570674122398255, + "grad_norm": 0.21086420118808746, + "learning_rate": 5.422947596976729e-05, + "loss": 0.0063, + "step": 49940 + }, + { + "epoch": 7.758620689655173, + "grad_norm": 0.24957072734832764, + "learning_rate": 5.4213000302908134e-05, + "loss": 0.006, + "step": 49950 + }, + { + "epoch": 7.760173967070519, + "grad_norm": 0.2790999114513397, + "learning_rate": 5.419652417532162e-05, + "loss": 0.0073, + "step": 49960 + }, + { + "epoch": 7.761727244485865, + "grad_norm": 0.1313067227602005, + "learning_rate": 5.4180047588809534e-05, + "loss": 0.0073, + "step": 49970 + }, + { + "epoch": 7.763280521901212, + "grad_norm": 0.4910867214202881, + "learning_rate": 5.4163570545173704e-05, + "loss": 0.0074, + "step": 49980 + }, + { + "epoch": 7.764833799316558, + "grad_norm": 0.16055311262607574, + "learning_rate": 5.414709304621608e-05, + "loss": 0.009, + "step": 49990 + }, + { + "epoch": 7.766387076731904, + "grad_norm": 0.2896770238876343, + "learning_rate": 5.413061509373858e-05, + "loss": 0.0081, + "step": 50000 + }, + { + "epoch": 7.7679403541472505, + "grad_norm": 0.2487238496541977, + "learning_rate": 5.411413668954324e-05, + "loss": 0.0057, + "step": 50010 + }, + { + "epoch": 7.769493631562597, + "grad_norm": 0.5063707828521729, + "learning_rate": 5.4097657835432083e-05, + "loss": 0.0071, + "step": 50020 + }, + { + "epoch": 7.771046908977944, + "grad_norm": 0.2951631546020508, + "learning_rate": 5.408117853320723e-05, + "loss": 0.007, + "step": 50030 + }, + { + "epoch": 7.77260018639329, + "grad_norm": 0.17350324988365173, + "learning_rate": 5.406469878467084e-05, + "loss": 0.0055, + "step": 50040 + }, + { + "epoch": 7.774153463808636, + "grad_norm": 0.29400864243507385, + "learning_rate": 5.404821859162509e-05, + "loss": 0.0055, + "step": 50050 + }, + { + "epoch": 7.775706741223983, + "grad_norm": 0.19552329182624817, + "learning_rate": 5.403173795587225e-05, + "loss": 0.008, + "step": 50060 + }, + { + "epoch": 7.777260018639329, + "grad_norm": 0.28934717178344727, + "learning_rate": 5.4015256879214606e-05, + "loss": 0.0064, + "step": 50070 + }, + { + "epoch": 7.778813296054675, + "grad_norm": 0.20188790559768677, + "learning_rate": 5.39987753634545e-05, + "loss": 0.0061, + "step": 50080 + }, + { + "epoch": 7.780366573470022, + "grad_norm": 0.12021388113498688, + "learning_rate": 5.3982293410394335e-05, + "loss": 0.0052, + "step": 50090 + }, + { + "epoch": 7.781919850885368, + "grad_norm": 0.15532618761062622, + "learning_rate": 5.396581102183655e-05, + "loss": 0.009, + "step": 50100 + }, + { + "epoch": 7.783473128300715, + "grad_norm": 0.306369423866272, + "learning_rate": 5.3949328199583634e-05, + "loss": 0.009, + "step": 50110 + }, + { + "epoch": 7.785026405716061, + "grad_norm": 0.3048557639122009, + "learning_rate": 5.393284494543813e-05, + "loss": 0.0075, + "step": 50120 + }, + { + "epoch": 7.786579683131407, + "grad_norm": 0.15249238908290863, + "learning_rate": 5.391636126120262e-05, + "loss": 0.0063, + "step": 50130 + }, + { + "epoch": 7.788132960546753, + "grad_norm": 0.24726544320583344, + "learning_rate": 5.389987714867971e-05, + "loss": 0.0062, + "step": 50140 + }, + { + "epoch": 7.7896862379621, + "grad_norm": 0.18667268753051758, + "learning_rate": 5.3883392609672114e-05, + "loss": 0.0063, + "step": 50150 + }, + { + "epoch": 7.7912395153774465, + "grad_norm": 0.15996761620044708, + "learning_rate": 5.3866907645982545e-05, + "loss": 0.0096, + "step": 50160 + }, + { + "epoch": 7.792792792792793, + "grad_norm": 0.1735997349023819, + "learning_rate": 5.385042225941375e-05, + "loss": 0.0082, + "step": 50170 + }, + { + "epoch": 7.794346070208139, + "grad_norm": 0.1503848284482956, + "learning_rate": 5.3833936451768585e-05, + "loss": 0.0062, + "step": 50180 + }, + { + "epoch": 7.795899347623486, + "grad_norm": 0.23121577501296997, + "learning_rate": 5.3817450224849884e-05, + "loss": 0.0077, + "step": 50190 + }, + { + "epoch": 7.797452625038832, + "grad_norm": 0.2554537355899811, + "learning_rate": 5.380096358046056e-05, + "loss": 0.0077, + "step": 50200 + }, + { + "epoch": 7.799005902454178, + "grad_norm": 0.2745775580406189, + "learning_rate": 5.378447652040359e-05, + "loss": 0.0077, + "step": 50210 + }, + { + "epoch": 7.800559179869524, + "grad_norm": 0.25849586725234985, + "learning_rate": 5.376798904648195e-05, + "loss": 0.008, + "step": 50220 + }, + { + "epoch": 7.8021124572848715, + "grad_norm": 0.26920831203460693, + "learning_rate": 5.375150116049869e-05, + "loss": 0.0078, + "step": 50230 + }, + { + "epoch": 7.803665734700218, + "grad_norm": 0.2735060751438141, + "learning_rate": 5.373501286425691e-05, + "loss": 0.0065, + "step": 50240 + }, + { + "epoch": 7.805219012115564, + "grad_norm": 0.7334502339363098, + "learning_rate": 5.3718524159559726e-05, + "loss": 0.0062, + "step": 50250 + }, + { + "epoch": 7.80677228953091, + "grad_norm": 0.23680907487869263, + "learning_rate": 5.370203504821034e-05, + "loss": 0.0084, + "step": 50260 + }, + { + "epoch": 7.808325566946257, + "grad_norm": 0.1757669597864151, + "learning_rate": 5.368554553201196e-05, + "loss": 0.0073, + "step": 50270 + }, + { + "epoch": 7.809878844361603, + "grad_norm": 0.23292094469070435, + "learning_rate": 5.3669055612767874e-05, + "loss": 0.0058, + "step": 50280 + }, + { + "epoch": 7.811432121776949, + "grad_norm": 0.2624158263206482, + "learning_rate": 5.3652565292281376e-05, + "loss": 0.0096, + "step": 50290 + }, + { + "epoch": 7.8129853991922955, + "grad_norm": 0.3957735002040863, + "learning_rate": 5.3636074572355824e-05, + "loss": 0.0079, + "step": 50300 + }, + { + "epoch": 7.814538676607642, + "grad_norm": 0.20627370476722717, + "learning_rate": 5.361958345479463e-05, + "loss": 0.0061, + "step": 50310 + }, + { + "epoch": 7.816091954022989, + "grad_norm": 0.2517869770526886, + "learning_rate": 5.360309194140123e-05, + "loss": 0.0079, + "step": 50320 + }, + { + "epoch": 7.817645231438335, + "grad_norm": 0.27009695768356323, + "learning_rate": 5.3586600033979106e-05, + "loss": 0.0069, + "step": 50330 + }, + { + "epoch": 7.819198508853681, + "grad_norm": 0.26084044575691223, + "learning_rate": 5.357010773433181e-05, + "loss": 0.0066, + "step": 50340 + }, + { + "epoch": 7.820751786269028, + "grad_norm": 0.15931205451488495, + "learning_rate": 5.3553615044262895e-05, + "loss": 0.0069, + "step": 50350 + }, + { + "epoch": 7.822305063684374, + "grad_norm": 0.150386780500412, + "learning_rate": 5.3537121965575964e-05, + "loss": 0.0066, + "step": 50360 + }, + { + "epoch": 7.8238583410997204, + "grad_norm": 0.53937828540802, + "learning_rate": 5.352062850007471e-05, + "loss": 0.0078, + "step": 50370 + }, + { + "epoch": 7.825411618515067, + "grad_norm": 0.249420627951622, + "learning_rate": 5.350413464956282e-05, + "loss": 0.007, + "step": 50380 + }, + { + "epoch": 7.826964895930413, + "grad_norm": 0.7943568229675293, + "learning_rate": 5.348764041584403e-05, + "loss": 0.008, + "step": 50390 + }, + { + "epoch": 7.82851817334576, + "grad_norm": 0.11336614936590195, + "learning_rate": 5.347114580072212e-05, + "loss": 0.005, + "step": 50400 + }, + { + "epoch": 7.830071450761106, + "grad_norm": 0.3198894262313843, + "learning_rate": 5.345465080600093e-05, + "loss": 0.0064, + "step": 50410 + }, + { + "epoch": 7.831624728176452, + "grad_norm": 0.25908133387565613, + "learning_rate": 5.3438155433484326e-05, + "loss": 0.0071, + "step": 50420 + }, + { + "epoch": 7.833178005591798, + "grad_norm": 0.20173105597496033, + "learning_rate": 5.3421659684976197e-05, + "loss": 0.0082, + "step": 50430 + }, + { + "epoch": 7.834731283007145, + "grad_norm": 0.27512750029563904, + "learning_rate": 5.340516356228052e-05, + "loss": 0.0066, + "step": 50440 + }, + { + "epoch": 7.8362845604224916, + "grad_norm": 0.16988834738731384, + "learning_rate": 5.338866706720128e-05, + "loss": 0.0074, + "step": 50450 + }, + { + "epoch": 7.837837837837838, + "grad_norm": 0.34735649824142456, + "learning_rate": 5.337217020154249e-05, + "loss": 0.0062, + "step": 50460 + }, + { + "epoch": 7.839391115253184, + "grad_norm": 0.16518089175224304, + "learning_rate": 5.335567296710825e-05, + "loss": 0.006, + "step": 50470 + }, + { + "epoch": 7.840944392668531, + "grad_norm": 0.09758088737726212, + "learning_rate": 5.333917536570265e-05, + "loss": 0.0052, + "step": 50480 + }, + { + "epoch": 7.842497670083877, + "grad_norm": 0.8448899388313293, + "learning_rate": 5.332267739912986e-05, + "loss": 0.0053, + "step": 50490 + }, + { + "epoch": 7.844050947499223, + "grad_norm": 0.21678681671619415, + "learning_rate": 5.330617906919405e-05, + "loss": 0.0057, + "step": 50500 + }, + { + "epoch": 7.845604224914569, + "grad_norm": 0.7013013362884521, + "learning_rate": 5.328968037769949e-05, + "loss": 0.0069, + "step": 50510 + }, + { + "epoch": 7.8471575023299165, + "grad_norm": 0.3946565091609955, + "learning_rate": 5.327318132645042e-05, + "loss": 0.0069, + "step": 50520 + }, + { + "epoch": 7.848710779745263, + "grad_norm": 0.291164755821228, + "learning_rate": 5.325668191725116e-05, + "loss": 0.0066, + "step": 50530 + }, + { + "epoch": 7.850264057160609, + "grad_norm": 0.2446655035018921, + "learning_rate": 5.324018215190606e-05, + "loss": 0.0078, + "step": 50540 + }, + { + "epoch": 7.851817334575955, + "grad_norm": 0.17192651331424713, + "learning_rate": 5.3223682032219515e-05, + "loss": 0.006, + "step": 50550 + }, + { + "epoch": 7.853370611991302, + "grad_norm": 0.22804665565490723, + "learning_rate": 5.320718155999595e-05, + "loss": 0.0063, + "step": 50560 + }, + { + "epoch": 7.854923889406648, + "grad_norm": 0.28940480947494507, + "learning_rate": 5.3190680737039835e-05, + "loss": 0.0064, + "step": 50570 + }, + { + "epoch": 7.856477166821994, + "grad_norm": 0.3257049024105072, + "learning_rate": 5.317417956515567e-05, + "loss": 0.0069, + "step": 50580 + }, + { + "epoch": 7.8580304442373405, + "grad_norm": 0.1782526671886444, + "learning_rate": 5.315767804614798e-05, + "loss": 0.0064, + "step": 50590 + }, + { + "epoch": 7.859583721652687, + "grad_norm": 0.33351224660873413, + "learning_rate": 5.3141176181821395e-05, + "loss": 0.0059, + "step": 50600 + }, + { + "epoch": 7.861136999068034, + "grad_norm": 0.2551352083683014, + "learning_rate": 5.31246739739805e-05, + "loss": 0.0075, + "step": 50610 + }, + { + "epoch": 7.86269027648338, + "grad_norm": 0.18744733929634094, + "learning_rate": 5.310817142442994e-05, + "loss": 0.0059, + "step": 50620 + }, + { + "epoch": 7.864243553898726, + "grad_norm": 0.2447502762079239, + "learning_rate": 5.309166853497445e-05, + "loss": 0.0086, + "step": 50630 + }, + { + "epoch": 7.865796831314073, + "grad_norm": 0.25656718015670776, + "learning_rate": 5.307516530741873e-05, + "loss": 0.0071, + "step": 50640 + }, + { + "epoch": 7.867350108729419, + "grad_norm": 0.3329525887966156, + "learning_rate": 5.305866174356754e-05, + "loss": 0.0053, + "step": 50650 + }, + { + "epoch": 7.8689033861447655, + "grad_norm": 0.29315173625946045, + "learning_rate": 5.304215784522571e-05, + "loss": 0.0051, + "step": 50660 + }, + { + "epoch": 7.870456663560112, + "grad_norm": 0.18429800868034363, + "learning_rate": 5.302565361419808e-05, + "loss": 0.0037, + "step": 50670 + }, + { + "epoch": 7.872009940975458, + "grad_norm": 0.20279516279697418, + "learning_rate": 5.3009149052289507e-05, + "loss": 0.0081, + "step": 50680 + }, + { + "epoch": 7.873563218390805, + "grad_norm": 0.3126184344291687, + "learning_rate": 5.299264416130493e-05, + "loss": 0.0089, + "step": 50690 + }, + { + "epoch": 7.875116495806151, + "grad_norm": 0.2324948012828827, + "learning_rate": 5.297613894304928e-05, + "loss": 0.0073, + "step": 50700 + }, + { + "epoch": 7.876669773221497, + "grad_norm": 0.3068600594997406, + "learning_rate": 5.2959633399327534e-05, + "loss": 0.0088, + "step": 50710 + }, + { + "epoch": 7.878223050636843, + "grad_norm": 0.09955795109272003, + "learning_rate": 5.294312753194476e-05, + "loss": 0.0055, + "step": 50720 + }, + { + "epoch": 7.87977632805219, + "grad_norm": 0.18586263060569763, + "learning_rate": 5.292662134270596e-05, + "loss": 0.0076, + "step": 50730 + }, + { + "epoch": 7.881329605467537, + "grad_norm": 0.20211614668369293, + "learning_rate": 5.291011483341626e-05, + "loss": 0.0079, + "step": 50740 + }, + { + "epoch": 7.882882882882883, + "grad_norm": 0.20330381393432617, + "learning_rate": 5.2893608005880767e-05, + "loss": 0.0068, + "step": 50750 + }, + { + "epoch": 7.884436160298229, + "grad_norm": 0.21861058473587036, + "learning_rate": 5.287710086190467e-05, + "loss": 0.008, + "step": 50760 + }, + { + "epoch": 7.885989437713576, + "grad_norm": 0.3143569827079773, + "learning_rate": 5.286059340329314e-05, + "loss": 0.0057, + "step": 50770 + }, + { + "epoch": 7.887542715128922, + "grad_norm": 0.34379997849464417, + "learning_rate": 5.284408563185141e-05, + "loss": 0.0094, + "step": 50780 + }, + { + "epoch": 7.889095992544268, + "grad_norm": 0.17756900191307068, + "learning_rate": 5.282757754938476e-05, + "loss": 0.006, + "step": 50790 + }, + { + "epoch": 7.8906492699596145, + "grad_norm": 0.3176712989807129, + "learning_rate": 5.281106915769849e-05, + "loss": 0.0066, + "step": 50800 + }, + { + "epoch": 7.8922025473749615, + "grad_norm": 0.33281752467155457, + "learning_rate": 5.2794560458597897e-05, + "loss": 0.0059, + "step": 50810 + }, + { + "epoch": 7.893755824790308, + "grad_norm": 0.19338691234588623, + "learning_rate": 5.27780514538884e-05, + "loss": 0.0062, + "step": 50820 + }, + { + "epoch": 7.895309102205654, + "grad_norm": 0.37848401069641113, + "learning_rate": 5.2761542145375365e-05, + "loss": 0.0073, + "step": 50830 + }, + { + "epoch": 7.896862379621, + "grad_norm": 0.23378469049930573, + "learning_rate": 5.274503253486421e-05, + "loss": 0.0073, + "step": 50840 + }, + { + "epoch": 7.898415657036347, + "grad_norm": 0.3071150481700897, + "learning_rate": 5.272852262416046e-05, + "loss": 0.0062, + "step": 50850 + }, + { + "epoch": 7.899968934451693, + "grad_norm": 0.2065344601869583, + "learning_rate": 5.2712012415069555e-05, + "loss": 0.0069, + "step": 50860 + }, + { + "epoch": 7.901522211867039, + "grad_norm": 0.22164002060890198, + "learning_rate": 5.269550190939705e-05, + "loss": 0.0054, + "step": 50870 + }, + { + "epoch": 7.903075489282386, + "grad_norm": 0.18126074969768524, + "learning_rate": 5.267899110894852e-05, + "loss": 0.006, + "step": 50880 + }, + { + "epoch": 7.904628766697732, + "grad_norm": 0.16918915510177612, + "learning_rate": 5.266248001552955e-05, + "loss": 0.0064, + "step": 50890 + }, + { + "epoch": 7.906182044113079, + "grad_norm": 0.19816382229328156, + "learning_rate": 5.264596863094575e-05, + "loss": 0.0055, + "step": 50900 + }, + { + "epoch": 7.907735321528425, + "grad_norm": 0.16074085235595703, + "learning_rate": 5.262945695700282e-05, + "loss": 0.0068, + "step": 50910 + }, + { + "epoch": 7.909288598943771, + "grad_norm": 0.3109862208366394, + "learning_rate": 5.261294499550643e-05, + "loss": 0.0051, + "step": 50920 + }, + { + "epoch": 7.910841876359118, + "grad_norm": 0.22087301313877106, + "learning_rate": 5.25964327482623e-05, + "loss": 0.0059, + "step": 50930 + }, + { + "epoch": 7.912395153774464, + "grad_norm": 0.3777439296245575, + "learning_rate": 5.257992021707617e-05, + "loss": 0.0063, + "step": 50940 + }, + { + "epoch": 7.9139484311898105, + "grad_norm": 0.4118572473526001, + "learning_rate": 5.256340740375387e-05, + "loss": 0.0067, + "step": 50950 + }, + { + "epoch": 7.915501708605157, + "grad_norm": 0.23008926212787628, + "learning_rate": 5.254689431010117e-05, + "loss": 0.0086, + "step": 50960 + }, + { + "epoch": 7.917054986020503, + "grad_norm": 0.08658288419246674, + "learning_rate": 5.253038093792395e-05, + "loss": 0.0071, + "step": 50970 + }, + { + "epoch": 7.91860826343585, + "grad_norm": 0.30127790570259094, + "learning_rate": 5.251386728902806e-05, + "loss": 0.0103, + "step": 50980 + }, + { + "epoch": 7.920161540851196, + "grad_norm": 0.19078019261360168, + "learning_rate": 5.2497353365219446e-05, + "loss": 0.0057, + "step": 50990 + }, + { + "epoch": 7.921714818266542, + "grad_norm": 0.2882908284664154, + "learning_rate": 5.2480839168304e-05, + "loss": 0.006, + "step": 51000 + }, + { + "epoch": 7.923268095681889, + "grad_norm": 0.1723787784576416, + "learning_rate": 5.246432470008773e-05, + "loss": 0.0064, + "step": 51010 + }, + { + "epoch": 7.924821373097235, + "grad_norm": 0.24366244673728943, + "learning_rate": 5.2447809962376616e-05, + "loss": 0.0061, + "step": 51020 + }, + { + "epoch": 7.926374650512582, + "grad_norm": 0.18117576837539673, + "learning_rate": 5.243129495697666e-05, + "loss": 0.0046, + "step": 51030 + }, + { + "epoch": 7.927927927927928, + "grad_norm": 0.18932238221168518, + "learning_rate": 5.241477968569397e-05, + "loss": 0.0067, + "step": 51040 + }, + { + "epoch": 7.929481205343274, + "grad_norm": 0.2398199886083603, + "learning_rate": 5.23982641503346e-05, + "loss": 0.0085, + "step": 51050 + }, + { + "epoch": 7.931034482758621, + "grad_norm": 0.3677603602409363, + "learning_rate": 5.238174835270465e-05, + "loss": 0.0068, + "step": 51060 + }, + { + "epoch": 7.932587760173967, + "grad_norm": 0.2703368067741394, + "learning_rate": 5.23652322946103e-05, + "loss": 0.0077, + "step": 51070 + }, + { + "epoch": 7.934141037589313, + "grad_norm": 0.35787948966026306, + "learning_rate": 5.234871597785771e-05, + "loss": 0.0064, + "step": 51080 + }, + { + "epoch": 7.9356943150046595, + "grad_norm": 0.23644119501113892, + "learning_rate": 5.233219940425306e-05, + "loss": 0.0072, + "step": 51090 + }, + { + "epoch": 7.9372475924200065, + "grad_norm": 0.23822557926177979, + "learning_rate": 5.2315682575602576e-05, + "loss": 0.0069, + "step": 51100 + }, + { + "epoch": 7.938800869835353, + "grad_norm": 0.31647688150405884, + "learning_rate": 5.229916549371254e-05, + "loss": 0.0087, + "step": 51110 + }, + { + "epoch": 7.940354147250699, + "grad_norm": 0.29886242747306824, + "learning_rate": 5.228264816038924e-05, + "loss": 0.0059, + "step": 51120 + }, + { + "epoch": 7.941907424666045, + "grad_norm": 0.9451144933700562, + "learning_rate": 5.2266130577438935e-05, + "loss": 0.0067, + "step": 51130 + }, + { + "epoch": 7.943460702081392, + "grad_norm": 0.8111680746078491, + "learning_rate": 5.224961274666802e-05, + "loss": 0.007, + "step": 51140 + }, + { + "epoch": 7.945013979496738, + "grad_norm": 0.26967841386795044, + "learning_rate": 5.223309466988283e-05, + "loss": 0.0081, + "step": 51150 + }, + { + "epoch": 7.946567256912084, + "grad_norm": 0.36509689688682556, + "learning_rate": 5.221657634888976e-05, + "loss": 0.0057, + "step": 51160 + }, + { + "epoch": 7.948120534327431, + "grad_norm": 0.17376025021076202, + "learning_rate": 5.2200057785495235e-05, + "loss": 0.0068, + "step": 51170 + }, + { + "epoch": 7.949673811742777, + "grad_norm": 0.8272587060928345, + "learning_rate": 5.2183538981505706e-05, + "loss": 0.0071, + "step": 51180 + }, + { + "epoch": 7.951227089158124, + "grad_norm": 0.20619487762451172, + "learning_rate": 5.216701993872762e-05, + "loss": 0.0061, + "step": 51190 + }, + { + "epoch": 7.95278036657347, + "grad_norm": 0.16250938177108765, + "learning_rate": 5.215050065896752e-05, + "loss": 0.0077, + "step": 51200 + }, + { + "epoch": 7.954333643988816, + "grad_norm": 0.2488270103931427, + "learning_rate": 5.2133981144031875e-05, + "loss": 0.0054, + "step": 51210 + }, + { + "epoch": 7.955886921404163, + "grad_norm": 0.12287592142820358, + "learning_rate": 5.211746139572727e-05, + "loss": 0.008, + "step": 51220 + }, + { + "epoch": 7.957440198819509, + "grad_norm": 0.16766056418418884, + "learning_rate": 5.210094141586026e-05, + "loss": 0.0083, + "step": 51230 + }, + { + "epoch": 7.9589934762348555, + "grad_norm": 0.24561583995819092, + "learning_rate": 5.2084421206237476e-05, + "loss": 0.0069, + "step": 51240 + }, + { + "epoch": 7.960546753650202, + "grad_norm": 0.3412870764732361, + "learning_rate": 5.20679007686655e-05, + "loss": 0.0048, + "step": 51250 + }, + { + "epoch": 7.962100031065548, + "grad_norm": 0.32298508286476135, + "learning_rate": 5.205138010495102e-05, + "loss": 0.0051, + "step": 51260 + }, + { + "epoch": 7.963653308480895, + "grad_norm": 0.2942275404930115, + "learning_rate": 5.20348592169007e-05, + "loss": 0.0104, + "step": 51270 + }, + { + "epoch": 7.965206585896241, + "grad_norm": 0.42659348249435425, + "learning_rate": 5.201833810632123e-05, + "loss": 0.0064, + "step": 51280 + }, + { + "epoch": 7.966759863311587, + "grad_norm": 0.19609668850898743, + "learning_rate": 5.200181677501933e-05, + "loss": 0.009, + "step": 51290 + }, + { + "epoch": 7.968313140726934, + "grad_norm": 0.3481956720352173, + "learning_rate": 5.1985295224801777e-05, + "loss": 0.0051, + "step": 51300 + }, + { + "epoch": 7.9698664181422805, + "grad_norm": 0.29232707619667053, + "learning_rate": 5.196877345747531e-05, + "loss": 0.0079, + "step": 51310 + }, + { + "epoch": 7.971419695557627, + "grad_norm": 0.13202804327011108, + "learning_rate": 5.195225147484674e-05, + "loss": 0.0056, + "step": 51320 + }, + { + "epoch": 7.972972972972973, + "grad_norm": 0.20473459362983704, + "learning_rate": 5.193572927872288e-05, + "loss": 0.0077, + "step": 51330 + }, + { + "epoch": 7.974526250388319, + "grad_norm": 0.4665907621383667, + "learning_rate": 5.191920687091059e-05, + "loss": 0.0085, + "step": 51340 + }, + { + "epoch": 7.976079527803666, + "grad_norm": 0.16341309249401093, + "learning_rate": 5.1902684253216695e-05, + "loss": 0.0052, + "step": 51350 + }, + { + "epoch": 7.977632805219012, + "grad_norm": 0.23175188899040222, + "learning_rate": 5.1886161427448134e-05, + "loss": 0.0052, + "step": 51360 + }, + { + "epoch": 7.979186082634358, + "grad_norm": 0.24095767736434937, + "learning_rate": 5.186963839541179e-05, + "loss": 0.0086, + "step": 51370 + }, + { + "epoch": 7.9807393600497045, + "grad_norm": 0.1941419094800949, + "learning_rate": 5.185311515891459e-05, + "loss": 0.005, + "step": 51380 + }, + { + "epoch": 7.982292637465052, + "grad_norm": 0.2487526386976242, + "learning_rate": 5.1836591719763516e-05, + "loss": 0.0076, + "step": 51390 + }, + { + "epoch": 7.983845914880398, + "grad_norm": 0.1705748736858368, + "learning_rate": 5.182006807976554e-05, + "loss": 0.0078, + "step": 51400 + }, + { + "epoch": 7.985399192295744, + "grad_norm": 0.17035378515720367, + "learning_rate": 5.180354424072763e-05, + "loss": 0.0074, + "step": 51410 + }, + { + "epoch": 7.98695246971109, + "grad_norm": 0.22506488859653473, + "learning_rate": 5.178702020445686e-05, + "loss": 0.0064, + "step": 51420 + }, + { + "epoch": 7.988505747126437, + "grad_norm": 0.17680631577968597, + "learning_rate": 5.177049597276024e-05, + "loss": 0.0082, + "step": 51430 + }, + { + "epoch": 7.990059024541783, + "grad_norm": 0.2021997570991516, + "learning_rate": 5.175397154744482e-05, + "loss": 0.0075, + "step": 51440 + }, + { + "epoch": 7.9916123019571295, + "grad_norm": 0.2616209089756012, + "learning_rate": 5.1737446930317736e-05, + "loss": 0.0044, + "step": 51450 + }, + { + "epoch": 7.993165579372476, + "grad_norm": 0.30820432305336, + "learning_rate": 5.172092212318604e-05, + "loss": 0.004, + "step": 51460 + }, + { + "epoch": 7.994718856787823, + "grad_norm": 0.41111573576927185, + "learning_rate": 5.170439712785692e-05, + "loss": 0.0063, + "step": 51470 + }, + { + "epoch": 7.996272134203169, + "grad_norm": 0.31343141198158264, + "learning_rate": 5.1687871946137467e-05, + "loss": 0.0075, + "step": 51480 + }, + { + "epoch": 7.997825411618515, + "grad_norm": 0.2520180642604828, + "learning_rate": 5.16713465798349e-05, + "loss": 0.0083, + "step": 51490 + }, + { + "epoch": 7.999378689033861, + "grad_norm": 0.2411326766014099, + "learning_rate": 5.165482103075637e-05, + "loss": 0.007, + "step": 51500 + }, + { + "epoch": 8.000931966449208, + "grad_norm": 0.30426183342933655, + "learning_rate": 5.163829530070909e-05, + "loss": 0.0045, + "step": 51510 + }, + { + "epoch": 8.002485243864554, + "grad_norm": 0.27087295055389404, + "learning_rate": 5.162176939150032e-05, + "loss": 0.0071, + "step": 51520 + }, + { + "epoch": 8.0040385212799, + "grad_norm": 0.30350545048713684, + "learning_rate": 5.160524330493728e-05, + "loss": 0.0068, + "step": 51530 + }, + { + "epoch": 8.005591798695248, + "grad_norm": 0.28418999910354614, + "learning_rate": 5.158871704282724e-05, + "loss": 0.0054, + "step": 51540 + }, + { + "epoch": 8.007145076110593, + "grad_norm": 0.09801505506038666, + "learning_rate": 5.15721906069775e-05, + "loss": 0.0077, + "step": 51550 + }, + { + "epoch": 8.00869835352594, + "grad_norm": 0.19506396353244781, + "learning_rate": 5.155566399919536e-05, + "loss": 0.0076, + "step": 51560 + }, + { + "epoch": 8.010251630941285, + "grad_norm": 0.2719562351703644, + "learning_rate": 5.153913722128813e-05, + "loss": 0.0076, + "step": 51570 + }, + { + "epoch": 8.011804908356632, + "grad_norm": 0.35006868839263916, + "learning_rate": 5.1522610275063176e-05, + "loss": 0.0151, + "step": 51580 + }, + { + "epoch": 8.01335818577198, + "grad_norm": 0.35870596766471863, + "learning_rate": 5.150608316232786e-05, + "loss": 0.0085, + "step": 51590 + }, + { + "epoch": 8.014911463187325, + "grad_norm": 0.2003549039363861, + "learning_rate": 5.148955588488955e-05, + "loss": 0.0083, + "step": 51600 + }, + { + "epoch": 8.016464740602672, + "grad_norm": 0.33387118577957153, + "learning_rate": 5.147302844455564e-05, + "loss": 0.0074, + "step": 51610 + }, + { + "epoch": 8.018018018018019, + "grad_norm": 0.1823279857635498, + "learning_rate": 5.145650084313356e-05, + "loss": 0.0062, + "step": 51620 + }, + { + "epoch": 8.019571295433364, + "grad_norm": 0.1187206581234932, + "learning_rate": 5.1439973082430734e-05, + "loss": 0.0068, + "step": 51630 + }, + { + "epoch": 8.021124572848711, + "grad_norm": 0.4503747224807739, + "learning_rate": 5.142344516425461e-05, + "loss": 0.007, + "step": 51640 + }, + { + "epoch": 8.022677850264056, + "grad_norm": 0.22130073606967926, + "learning_rate": 5.1406917090412673e-05, + "loss": 0.0083, + "step": 51650 + }, + { + "epoch": 8.024231127679403, + "grad_norm": 0.21821369230747223, + "learning_rate": 5.1390388862712404e-05, + "loss": 0.0062, + "step": 51660 + }, + { + "epoch": 8.02578440509475, + "grad_norm": 0.2309977263212204, + "learning_rate": 5.137386048296129e-05, + "loss": 0.0065, + "step": 51670 + }, + { + "epoch": 8.027337682510096, + "grad_norm": 0.23474298417568207, + "learning_rate": 5.135733195296686e-05, + "loss": 0.0065, + "step": 51680 + }, + { + "epoch": 8.028890959925443, + "grad_norm": 0.26071974635124207, + "learning_rate": 5.134080327453664e-05, + "loss": 0.0055, + "step": 51690 + }, + { + "epoch": 8.03044423734079, + "grad_norm": 0.23012518882751465, + "learning_rate": 5.1324274449478204e-05, + "loss": 0.0072, + "step": 51700 + }, + { + "epoch": 8.031997514756135, + "grad_norm": 0.20884089171886444, + "learning_rate": 5.130774547959909e-05, + "loss": 0.0065, + "step": 51710 + }, + { + "epoch": 8.033550792171482, + "grad_norm": 0.25866788625717163, + "learning_rate": 5.129121636670692e-05, + "loss": 0.007, + "step": 51720 + }, + { + "epoch": 8.035104069586827, + "grad_norm": 0.2895497679710388, + "learning_rate": 5.1274687112609246e-05, + "loss": 0.007, + "step": 51730 + }, + { + "epoch": 8.036657347002174, + "grad_norm": 0.23748786747455597, + "learning_rate": 5.125815771911372e-05, + "loss": 0.0056, + "step": 51740 + }, + { + "epoch": 8.038210624417522, + "grad_norm": 0.15090999007225037, + "learning_rate": 5.1241628188027956e-05, + "loss": 0.0069, + "step": 51750 + }, + { + "epoch": 8.039763901832867, + "grad_norm": 0.18717539310455322, + "learning_rate": 5.1225098521159585e-05, + "loss": 0.0082, + "step": 51760 + }, + { + "epoch": 8.041317179248214, + "grad_norm": 0.10775647312402725, + "learning_rate": 5.12085687203163e-05, + "loss": 0.0053, + "step": 51770 + }, + { + "epoch": 8.042870456663561, + "grad_norm": 0.2245018482208252, + "learning_rate": 5.119203878730575e-05, + "loss": 0.0071, + "step": 51780 + }, + { + "epoch": 8.044423734078906, + "grad_norm": 0.24762900173664093, + "learning_rate": 5.117550872393564e-05, + "loss": 0.0076, + "step": 51790 + }, + { + "epoch": 8.045977011494253, + "grad_norm": 0.35855039954185486, + "learning_rate": 5.1158978532013645e-05, + "loss": 0.008, + "step": 51800 + }, + { + "epoch": 8.047530288909599, + "grad_norm": 0.10062102973461151, + "learning_rate": 5.114244821334751e-05, + "loss": 0.0057, + "step": 51810 + }, + { + "epoch": 8.049083566324946, + "grad_norm": 0.28535208106040955, + "learning_rate": 5.112591776974497e-05, + "loss": 0.0068, + "step": 51820 + }, + { + "epoch": 8.050636843740293, + "grad_norm": 0.32164284586906433, + "learning_rate": 5.110938720301374e-05, + "loss": 0.0086, + "step": 51830 + }, + { + "epoch": 8.052190121155638, + "grad_norm": 0.18999041616916656, + "learning_rate": 5.1092856514961595e-05, + "loss": 0.007, + "step": 51840 + }, + { + "epoch": 8.053743398570985, + "grad_norm": 0.21610668301582336, + "learning_rate": 5.107632570739631e-05, + "loss": 0.0055, + "step": 51850 + }, + { + "epoch": 8.055296675986332, + "grad_norm": 0.1728033423423767, + "learning_rate": 5.1059794782125645e-05, + "loss": 0.0054, + "step": 51860 + }, + { + "epoch": 8.056849953401677, + "grad_norm": 0.20129936933517456, + "learning_rate": 5.104326374095744e-05, + "loss": 0.008, + "step": 51870 + }, + { + "epoch": 8.058403230817024, + "grad_norm": 0.09041140228509903, + "learning_rate": 5.102673258569947e-05, + "loss": 0.0041, + "step": 51880 + }, + { + "epoch": 8.05995650823237, + "grad_norm": 0.29715487360954285, + "learning_rate": 5.101020131815956e-05, + "loss": 0.007, + "step": 51890 + }, + { + "epoch": 8.061509785647717, + "grad_norm": 0.19712957739830017, + "learning_rate": 5.099366994014556e-05, + "loss": 0.0053, + "step": 51900 + }, + { + "epoch": 8.063063063063064, + "grad_norm": 0.14956529438495636, + "learning_rate": 5.097713845346532e-05, + "loss": 0.0091, + "step": 51910 + }, + { + "epoch": 8.064616340478409, + "grad_norm": 0.1970095932483673, + "learning_rate": 5.096060685992668e-05, + "loss": 0.0068, + "step": 51920 + }, + { + "epoch": 8.066169617893756, + "grad_norm": 0.4036831855773926, + "learning_rate": 5.0944075161337514e-05, + "loss": 0.0106, + "step": 51930 + }, + { + "epoch": 8.067722895309101, + "grad_norm": 0.21331246197223663, + "learning_rate": 5.09275433595057e-05, + "loss": 0.0078, + "step": 51940 + }, + { + "epoch": 8.069276172724448, + "grad_norm": 0.7109797596931458, + "learning_rate": 5.0911011456239157e-05, + "loss": 0.0076, + "step": 51950 + }, + { + "epoch": 8.070829450139795, + "grad_norm": 0.2171010971069336, + "learning_rate": 5.089447945334574e-05, + "loss": 0.0096, + "step": 51960 + }, + { + "epoch": 8.07238272755514, + "grad_norm": 0.18131279945373535, + "learning_rate": 5.087794735263343e-05, + "loss": 0.0086, + "step": 51970 + }, + { + "epoch": 8.073936004970488, + "grad_norm": 0.16605402529239655, + "learning_rate": 5.086141515591011e-05, + "loss": 0.0064, + "step": 51980 + }, + { + "epoch": 8.075489282385835, + "grad_norm": 0.26812487840652466, + "learning_rate": 5.084488286498371e-05, + "loss": 0.0054, + "step": 51990 + }, + { + "epoch": 8.07704255980118, + "grad_norm": 0.16022014617919922, + "learning_rate": 5.08283504816622e-05, + "loss": 0.0055, + "step": 52000 + }, + { + "epoch": 8.078595837216527, + "grad_norm": 0.2426225244998932, + "learning_rate": 5.081181800775353e-05, + "loss": 0.0059, + "step": 52010 + }, + { + "epoch": 8.080149114631872, + "grad_norm": 0.7872242331504822, + "learning_rate": 5.079528544506565e-05, + "loss": 0.0053, + "step": 52020 + }, + { + "epoch": 8.08170239204722, + "grad_norm": 0.20110487937927246, + "learning_rate": 5.077875279540657e-05, + "loss": 0.0084, + "step": 52030 + }, + { + "epoch": 8.083255669462567, + "grad_norm": 0.19383321702480316, + "learning_rate": 5.076222006058424e-05, + "loss": 0.005, + "step": 52040 + }, + { + "epoch": 8.084808946877912, + "grad_norm": 0.23599547147750854, + "learning_rate": 5.074568724240667e-05, + "loss": 0.0064, + "step": 52050 + }, + { + "epoch": 8.086362224293259, + "grad_norm": 0.12511035799980164, + "learning_rate": 5.072915434268187e-05, + "loss": 0.0085, + "step": 52060 + }, + { + "epoch": 8.087915501708606, + "grad_norm": 0.7803983688354492, + "learning_rate": 5.071262136321785e-05, + "loss": 0.0091, + "step": 52070 + }, + { + "epoch": 8.089468779123951, + "grad_norm": 0.1697598546743393, + "learning_rate": 5.069608830582262e-05, + "loss": 0.0065, + "step": 52080 + }, + { + "epoch": 8.091022056539298, + "grad_norm": 0.21506823599338531, + "learning_rate": 5.067955517230423e-05, + "loss": 0.0065, + "step": 52090 + }, + { + "epoch": 8.092575333954644, + "grad_norm": 0.29025131464004517, + "learning_rate": 5.06630219644707e-05, + "loss": 0.0087, + "step": 52100 + }, + { + "epoch": 8.09412861136999, + "grad_norm": 0.2192917764186859, + "learning_rate": 5.064648868413008e-05, + "loss": 0.0098, + "step": 52110 + }, + { + "epoch": 8.095681888785338, + "grad_norm": 0.2532275915145874, + "learning_rate": 5.062995533309044e-05, + "loss": 0.0084, + "step": 52120 + }, + { + "epoch": 8.097235166200683, + "grad_norm": 0.2580714523792267, + "learning_rate": 5.061342191315983e-05, + "loss": 0.0077, + "step": 52130 + }, + { + "epoch": 8.09878844361603, + "grad_norm": 0.2718294858932495, + "learning_rate": 5.0596888426146325e-05, + "loss": 0.0081, + "step": 52140 + }, + { + "epoch": 8.100341721031377, + "grad_norm": 0.20272263884544373, + "learning_rate": 5.058035487385798e-05, + "loss": 0.0077, + "step": 52150 + }, + { + "epoch": 8.101894998446722, + "grad_norm": 0.17225944995880127, + "learning_rate": 5.0563821258102915e-05, + "loss": 0.0057, + "step": 52160 + }, + { + "epoch": 8.10344827586207, + "grad_norm": 0.1493672877550125, + "learning_rate": 5.05472875806892e-05, + "loss": 0.0068, + "step": 52170 + }, + { + "epoch": 8.105001553277415, + "grad_norm": 0.16467435657978058, + "learning_rate": 5.053075384342494e-05, + "loss": 0.006, + "step": 52180 + }, + { + "epoch": 8.106554830692762, + "grad_norm": 0.11285878717899323, + "learning_rate": 5.0514220048118234e-05, + "loss": 0.0047, + "step": 52190 + }, + { + "epoch": 8.108108108108109, + "grad_norm": 0.3322031497955322, + "learning_rate": 5.049768619657721e-05, + "loss": 0.0081, + "step": 52200 + }, + { + "epoch": 8.109661385523454, + "grad_norm": 0.1442246437072754, + "learning_rate": 5.048115229060996e-05, + "loss": 0.0071, + "step": 52210 + }, + { + "epoch": 8.111214662938801, + "grad_norm": 0.10306202620267868, + "learning_rate": 5.046461833202462e-05, + "loss": 0.0053, + "step": 52220 + }, + { + "epoch": 8.112767940354146, + "grad_norm": 0.14836354553699493, + "learning_rate": 5.044808432262932e-05, + "loss": 0.0045, + "step": 52230 + }, + { + "epoch": 8.114321217769493, + "grad_norm": 0.19666458666324615, + "learning_rate": 5.043155026423219e-05, + "loss": 0.0078, + "step": 52240 + }, + { + "epoch": 8.11587449518484, + "grad_norm": 0.33834749460220337, + "learning_rate": 5.041501615864137e-05, + "loss": 0.0061, + "step": 52250 + }, + { + "epoch": 8.117427772600186, + "grad_norm": 0.2900153398513794, + "learning_rate": 5.039848200766502e-05, + "loss": 0.0058, + "step": 52260 + }, + { + "epoch": 8.118981050015533, + "grad_norm": 0.8157086968421936, + "learning_rate": 5.038194781311125e-05, + "loss": 0.0063, + "step": 52270 + }, + { + "epoch": 8.12053432743088, + "grad_norm": 0.1678435206413269, + "learning_rate": 5.0365413576788264e-05, + "loss": 0.0056, + "step": 52280 + }, + { + "epoch": 8.122087604846225, + "grad_norm": 0.2618464231491089, + "learning_rate": 5.0348879300504194e-05, + "loss": 0.0049, + "step": 52290 + }, + { + "epoch": 8.123640882261572, + "grad_norm": 0.11751013249158859, + "learning_rate": 5.033234498606721e-05, + "loss": 0.0094, + "step": 52300 + }, + { + "epoch": 8.125194159676917, + "grad_norm": 0.20340868830680847, + "learning_rate": 5.0315810635285476e-05, + "loss": 0.0066, + "step": 52310 + }, + { + "epoch": 8.126747437092265, + "grad_norm": 0.1787867397069931, + "learning_rate": 5.0299276249967164e-05, + "loss": 0.0063, + "step": 52320 + }, + { + "epoch": 8.128300714507612, + "grad_norm": 0.5064620971679688, + "learning_rate": 5.0282741831920454e-05, + "loss": 0.0043, + "step": 52330 + }, + { + "epoch": 8.129853991922957, + "grad_norm": 0.14276328682899475, + "learning_rate": 5.026620738295351e-05, + "loss": 0.0048, + "step": 52340 + }, + { + "epoch": 8.131407269338304, + "grad_norm": 0.17439396679401398, + "learning_rate": 5.024967290487455e-05, + "loss": 0.0061, + "step": 52350 + }, + { + "epoch": 8.132960546753651, + "grad_norm": 0.17098048329353333, + "learning_rate": 5.023313839949172e-05, + "loss": 0.008, + "step": 52360 + }, + { + "epoch": 8.134513824168996, + "grad_norm": 0.28976741433143616, + "learning_rate": 5.0216603868613224e-05, + "loss": 0.0054, + "step": 52370 + }, + { + "epoch": 8.136067101584343, + "grad_norm": 0.2083081305027008, + "learning_rate": 5.020006931404726e-05, + "loss": 0.0053, + "step": 52380 + }, + { + "epoch": 8.137620378999689, + "grad_norm": 0.20683662593364716, + "learning_rate": 5.0183534737602014e-05, + "loss": 0.0064, + "step": 52390 + }, + { + "epoch": 8.139173656415036, + "grad_norm": 0.7104669213294983, + "learning_rate": 5.0167000141085686e-05, + "loss": 0.0072, + "step": 52400 + }, + { + "epoch": 8.140726933830383, + "grad_norm": 0.1888004094362259, + "learning_rate": 5.015046552630647e-05, + "loss": 0.0083, + "step": 52410 + }, + { + "epoch": 8.142280211245728, + "grad_norm": 0.14217513799667358, + "learning_rate": 5.013393089507258e-05, + "loss": 0.0052, + "step": 52420 + }, + { + "epoch": 8.143833488661075, + "grad_norm": 0.3606801927089691, + "learning_rate": 5.011739624919219e-05, + "loss": 0.0085, + "step": 52430 + }, + { + "epoch": 8.145386766076422, + "grad_norm": 0.31525373458862305, + "learning_rate": 5.0100861590473545e-05, + "loss": 0.006, + "step": 52440 + }, + { + "epoch": 8.146940043491767, + "grad_norm": 0.25527864694595337, + "learning_rate": 5.008432692072482e-05, + "loss": 0.007, + "step": 52450 + }, + { + "epoch": 8.148493320907114, + "grad_norm": 0.5574892163276672, + "learning_rate": 5.006779224175423e-05, + "loss": 0.0061, + "step": 52460 + }, + { + "epoch": 8.15004659832246, + "grad_norm": 0.23919498920440674, + "learning_rate": 5.005125755536998e-05, + "loss": 0.0079, + "step": 52470 + }, + { + "epoch": 8.151599875737807, + "grad_norm": 0.18220901489257812, + "learning_rate": 5.003472286338029e-05, + "loss": 0.0076, + "step": 52480 + }, + { + "epoch": 8.153153153153154, + "grad_norm": 0.1586690992116928, + "learning_rate": 5.001818816759335e-05, + "loss": 0.0052, + "step": 52490 + }, + { + "epoch": 8.154706430568499, + "grad_norm": 0.2835998237133026, + "learning_rate": 5.000165346981738e-05, + "loss": 0.0076, + "step": 52500 + }, + { + "epoch": 8.156259707983846, + "grad_norm": 0.20724058151245117, + "learning_rate": 4.9985118771860595e-05, + "loss": 0.0073, + "step": 52510 + }, + { + "epoch": 8.157812985399193, + "grad_norm": 0.5423494577407837, + "learning_rate": 4.996858407553119e-05, + "loss": 0.0106, + "step": 52520 + }, + { + "epoch": 8.159366262814538, + "grad_norm": 0.6769023537635803, + "learning_rate": 4.99520493826374e-05, + "loss": 0.006, + "step": 52530 + }, + { + "epoch": 8.160919540229886, + "grad_norm": 0.3851427733898163, + "learning_rate": 4.9935514694987384e-05, + "loss": 0.0063, + "step": 52540 + }, + { + "epoch": 8.16247281764523, + "grad_norm": 0.16343216598033905, + "learning_rate": 4.991898001438941e-05, + "loss": 0.0064, + "step": 52550 + }, + { + "epoch": 8.164026095060578, + "grad_norm": 0.36457452178001404, + "learning_rate": 4.990244534265166e-05, + "loss": 0.0056, + "step": 52560 + }, + { + "epoch": 8.165579372475925, + "grad_norm": 0.2269725352525711, + "learning_rate": 4.988591068158234e-05, + "loss": 0.0083, + "step": 52570 + }, + { + "epoch": 8.16713264989127, + "grad_norm": 0.15422044694423676, + "learning_rate": 4.986937603298965e-05, + "loss": 0.0078, + "step": 52580 + }, + { + "epoch": 8.168685927306617, + "grad_norm": 0.14657250046730042, + "learning_rate": 4.985284139868181e-05, + "loss": 0.0055, + "step": 52590 + }, + { + "epoch": 8.170239204721963, + "grad_norm": 0.27476823329925537, + "learning_rate": 4.9836306780467e-05, + "loss": 0.0056, + "step": 52600 + }, + { + "epoch": 8.17179248213731, + "grad_norm": 0.27669405937194824, + "learning_rate": 4.981977218015343e-05, + "loss": 0.006, + "step": 52610 + }, + { + "epoch": 8.173345759552657, + "grad_norm": 0.26295584440231323, + "learning_rate": 4.980323759954932e-05, + "loss": 0.01, + "step": 52620 + }, + { + "epoch": 8.174899036968002, + "grad_norm": 0.4335339367389679, + "learning_rate": 4.978670304046284e-05, + "loss": 0.0088, + "step": 52630 + }, + { + "epoch": 8.176452314383349, + "grad_norm": 0.241337850689888, + "learning_rate": 4.9770168504702195e-05, + "loss": 0.0066, + "step": 52640 + }, + { + "epoch": 8.178005591798696, + "grad_norm": 0.2558615803718567, + "learning_rate": 4.975363399407559e-05, + "loss": 0.0065, + "step": 52650 + }, + { + "epoch": 8.179558869214041, + "grad_norm": 0.19744417071342468, + "learning_rate": 4.9737099510391194e-05, + "loss": 0.007, + "step": 52660 + }, + { + "epoch": 8.181112146629388, + "grad_norm": 0.30811959505081177, + "learning_rate": 4.972056505545719e-05, + "loss": 0.0109, + "step": 52670 + }, + { + "epoch": 8.182665424044734, + "grad_norm": 0.4113149642944336, + "learning_rate": 4.9704030631081775e-05, + "loss": 0.0065, + "step": 52680 + }, + { + "epoch": 8.18421870146008, + "grad_norm": 0.8337011337280273, + "learning_rate": 4.968749623907313e-05, + "loss": 0.008, + "step": 52690 + }, + { + "epoch": 8.185771978875428, + "grad_norm": 0.20313169062137604, + "learning_rate": 4.967096188123943e-05, + "loss": 0.0066, + "step": 52700 + }, + { + "epoch": 8.187325256290773, + "grad_norm": 0.21377606689929962, + "learning_rate": 4.965442755938884e-05, + "loss": 0.0059, + "step": 52710 + }, + { + "epoch": 8.18887853370612, + "grad_norm": 0.1541278213262558, + "learning_rate": 4.963789327532954e-05, + "loss": 0.0057, + "step": 52720 + }, + { + "epoch": 8.190431811121467, + "grad_norm": 0.2962445914745331, + "learning_rate": 4.962135903086968e-05, + "loss": 0.0062, + "step": 52730 + }, + { + "epoch": 8.191985088536812, + "grad_norm": 0.31253135204315186, + "learning_rate": 4.960482482781741e-05, + "loss": 0.0061, + "step": 52740 + }, + { + "epoch": 8.19353836595216, + "grad_norm": 0.33497849106788635, + "learning_rate": 4.958829066798092e-05, + "loss": 0.0076, + "step": 52750 + }, + { + "epoch": 8.195091643367505, + "grad_norm": 0.4259152412414551, + "learning_rate": 4.957175655316835e-05, + "loss": 0.006, + "step": 52760 + }, + { + "epoch": 8.196644920782852, + "grad_norm": 0.9108617305755615, + "learning_rate": 4.9555222485187834e-05, + "loss": 0.0105, + "step": 52770 + }, + { + "epoch": 8.198198198198199, + "grad_norm": 0.38088059425354004, + "learning_rate": 4.953868846584752e-05, + "loss": 0.0081, + "step": 52780 + }, + { + "epoch": 8.199751475613544, + "grad_norm": 0.17616420984268188, + "learning_rate": 4.9522154496955544e-05, + "loss": 0.008, + "step": 52790 + }, + { + "epoch": 8.201304753028891, + "grad_norm": 0.3312690258026123, + "learning_rate": 4.950562058032001e-05, + "loss": 0.0072, + "step": 52800 + }, + { + "epoch": 8.202858030444236, + "grad_norm": 0.18395690619945526, + "learning_rate": 4.948908671774909e-05, + "loss": 0.0095, + "step": 52810 + }, + { + "epoch": 8.204411307859584, + "grad_norm": 0.15350806713104248, + "learning_rate": 4.947255291105087e-05, + "loss": 0.0071, + "step": 52820 + }, + { + "epoch": 8.20596458527493, + "grad_norm": 0.20402151346206665, + "learning_rate": 4.945601916203348e-05, + "loss": 0.0066, + "step": 52830 + }, + { + "epoch": 8.207517862690276, + "grad_norm": 0.30418795347213745, + "learning_rate": 4.9439485472505016e-05, + "loss": 0.0059, + "step": 52840 + }, + { + "epoch": 8.209071140105623, + "grad_norm": 0.3098217248916626, + "learning_rate": 4.942295184427356e-05, + "loss": 0.0081, + "step": 52850 + }, + { + "epoch": 8.21062441752097, + "grad_norm": 0.13644342124462128, + "learning_rate": 4.9406418279147234e-05, + "loss": 0.0069, + "step": 52860 + }, + { + "epoch": 8.212177694936315, + "grad_norm": 0.39916250109672546, + "learning_rate": 4.9389884778934116e-05, + "loss": 0.0058, + "step": 52870 + }, + { + "epoch": 8.213730972351662, + "grad_norm": 0.47749051451683044, + "learning_rate": 4.937335134544228e-05, + "loss": 0.0055, + "step": 52880 + }, + { + "epoch": 8.215284249767008, + "grad_norm": 0.22449038922786713, + "learning_rate": 4.935681798047979e-05, + "loss": 0.0082, + "step": 52890 + }, + { + "epoch": 8.216837527182355, + "grad_norm": 0.2942872941493988, + "learning_rate": 4.934028468585473e-05, + "loss": 0.0073, + "step": 52900 + }, + { + "epoch": 8.218390804597702, + "grad_norm": 0.2956669330596924, + "learning_rate": 4.932375146337514e-05, + "loss": 0.0054, + "step": 52910 + }, + { + "epoch": 8.219944082013047, + "grad_norm": 0.2314833104610443, + "learning_rate": 4.930721831484908e-05, + "loss": 0.0065, + "step": 52920 + }, + { + "epoch": 8.221497359428394, + "grad_norm": 0.22282254695892334, + "learning_rate": 4.929068524208456e-05, + "loss": 0.007, + "step": 52930 + }, + { + "epoch": 8.223050636843741, + "grad_norm": 0.31292465329170227, + "learning_rate": 4.927415224688965e-05, + "loss": 0.0078, + "step": 52940 + }, + { + "epoch": 8.224603914259086, + "grad_norm": 0.14388403296470642, + "learning_rate": 4.9257619331072375e-05, + "loss": 0.0067, + "step": 52950 + }, + { + "epoch": 8.226157191674433, + "grad_norm": 0.24072468280792236, + "learning_rate": 4.924108649644072e-05, + "loss": 0.0065, + "step": 52960 + }, + { + "epoch": 8.227710469089779, + "grad_norm": 0.17809221148490906, + "learning_rate": 4.922455374480271e-05, + "loss": 0.0067, + "step": 52970 + }, + { + "epoch": 8.229263746505126, + "grad_norm": 0.3096734583377838, + "learning_rate": 4.920802107796634e-05, + "loss": 0.0094, + "step": 52980 + }, + { + "epoch": 8.230817023920473, + "grad_norm": 0.3186512589454651, + "learning_rate": 4.9191488497739564e-05, + "loss": 0.0066, + "step": 52990 + }, + { + "epoch": 8.232370301335818, + "grad_norm": 0.15998488664627075, + "learning_rate": 4.917495600593042e-05, + "loss": 0.0066, + "step": 53000 + }, + { + "epoch": 8.233923578751165, + "grad_norm": 0.3445225954055786, + "learning_rate": 4.915842360434684e-05, + "loss": 0.0073, + "step": 53010 + }, + { + "epoch": 8.235476856166512, + "grad_norm": 0.14422395825386047, + "learning_rate": 4.914189129479679e-05, + "loss": 0.0046, + "step": 53020 + }, + { + "epoch": 8.237030133581857, + "grad_norm": 0.19673769176006317, + "learning_rate": 4.912535907908822e-05, + "loss": 0.0055, + "step": 53030 + }, + { + "epoch": 8.238583410997204, + "grad_norm": 0.10892754793167114, + "learning_rate": 4.910882695902907e-05, + "loss": 0.0068, + "step": 53040 + }, + { + "epoch": 8.24013668841255, + "grad_norm": 0.27467504143714905, + "learning_rate": 4.9092294936427234e-05, + "loss": 0.0096, + "step": 53050 + }, + { + "epoch": 8.241689965827897, + "grad_norm": 0.2735198736190796, + "learning_rate": 4.9075763013090685e-05, + "loss": 0.006, + "step": 53060 + }, + { + "epoch": 8.243243243243244, + "grad_norm": 0.1766369491815567, + "learning_rate": 4.9059231190827305e-05, + "loss": 0.0059, + "step": 53070 + }, + { + "epoch": 8.24479652065859, + "grad_norm": 0.10705766826868057, + "learning_rate": 4.904269947144499e-05, + "loss": 0.0076, + "step": 53080 + }, + { + "epoch": 8.246349798073936, + "grad_norm": 0.2757396399974823, + "learning_rate": 4.902616785675161e-05, + "loss": 0.0078, + "step": 53090 + }, + { + "epoch": 8.247903075489283, + "grad_norm": 0.2593548893928528, + "learning_rate": 4.900963634855505e-05, + "loss": 0.0082, + "step": 53100 + }, + { + "epoch": 8.249456352904629, + "grad_norm": 0.22853627800941467, + "learning_rate": 4.8993104948663173e-05, + "loss": 0.0074, + "step": 53110 + }, + { + "epoch": 8.251009630319976, + "grad_norm": 0.32903560996055603, + "learning_rate": 4.8976573658883823e-05, + "loss": 0.0053, + "step": 53120 + }, + { + "epoch": 8.25256290773532, + "grad_norm": 0.21417805552482605, + "learning_rate": 4.896004248102484e-05, + "loss": 0.0064, + "step": 53130 + }, + { + "epoch": 8.254116185150668, + "grad_norm": 0.2445729523897171, + "learning_rate": 4.8943511416894045e-05, + "loss": 0.0075, + "step": 53140 + }, + { + "epoch": 8.255669462566015, + "grad_norm": 0.32687851786613464, + "learning_rate": 4.892698046829926e-05, + "loss": 0.0075, + "step": 53150 + }, + { + "epoch": 8.25722273998136, + "grad_norm": 0.3291780948638916, + "learning_rate": 4.8910449637048275e-05, + "loss": 0.0072, + "step": 53160 + }, + { + "epoch": 8.258776017396707, + "grad_norm": 0.16225430369377136, + "learning_rate": 4.8893918924948884e-05, + "loss": 0.0087, + "step": 53170 + }, + { + "epoch": 8.260329294812053, + "grad_norm": 0.130655899643898, + "learning_rate": 4.8877388333808835e-05, + "loss": 0.0052, + "step": 53180 + }, + { + "epoch": 8.2618825722274, + "grad_norm": 0.34823939204216003, + "learning_rate": 4.8860857865435925e-05, + "loss": 0.0062, + "step": 53190 + }, + { + "epoch": 8.263435849642747, + "grad_norm": 0.19556890428066254, + "learning_rate": 4.884432752163789e-05, + "loss": 0.0078, + "step": 53200 + }, + { + "epoch": 8.264989127058092, + "grad_norm": 0.20764677226543427, + "learning_rate": 4.8827797304222466e-05, + "loss": 0.0056, + "step": 53210 + }, + { + "epoch": 8.266542404473439, + "grad_norm": 0.4459216296672821, + "learning_rate": 4.8811267214997366e-05, + "loss": 0.0072, + "step": 53220 + }, + { + "epoch": 8.268095681888786, + "grad_norm": 0.4180358946323395, + "learning_rate": 4.879473725577029e-05, + "loss": 0.0052, + "step": 53230 + }, + { + "epoch": 8.269648959304131, + "grad_norm": 0.23375867307186127, + "learning_rate": 4.877820742834891e-05, + "loss": 0.0069, + "step": 53240 + }, + { + "epoch": 8.271202236719478, + "grad_norm": 0.15503354370594025, + "learning_rate": 4.8761677734540956e-05, + "loss": 0.0077, + "step": 53250 + }, + { + "epoch": 8.272755514134824, + "grad_norm": 0.1638268381357193, + "learning_rate": 4.874514817615407e-05, + "loss": 0.0091, + "step": 53260 + }, + { + "epoch": 8.27430879155017, + "grad_norm": 0.9098412394523621, + "learning_rate": 4.872861875499588e-05, + "loss": 0.0086, + "step": 53270 + }, + { + "epoch": 8.275862068965518, + "grad_norm": 0.21222351491451263, + "learning_rate": 4.871208947287404e-05, + "loss": 0.008, + "step": 53280 + }, + { + "epoch": 8.277415346380863, + "grad_norm": 0.3394619822502136, + "learning_rate": 4.869556033159615e-05, + "loss": 0.0073, + "step": 53290 + }, + { + "epoch": 8.27896862379621, + "grad_norm": 0.25153863430023193, + "learning_rate": 4.8679031332969835e-05, + "loss": 0.0109, + "step": 53300 + }, + { + "epoch": 8.280521901211557, + "grad_norm": 0.19519779086112976, + "learning_rate": 4.8662502478802635e-05, + "loss": 0.0065, + "step": 53310 + }, + { + "epoch": 8.282075178626902, + "grad_norm": 0.22795788943767548, + "learning_rate": 4.8645973770902176e-05, + "loss": 0.0091, + "step": 53320 + }, + { + "epoch": 8.28362845604225, + "grad_norm": 0.2652730941772461, + "learning_rate": 4.862944521107599e-05, + "loss": 0.0065, + "step": 53330 + }, + { + "epoch": 8.285181733457595, + "grad_norm": 0.16365648806095123, + "learning_rate": 4.861291680113162e-05, + "loss": 0.0074, + "step": 53340 + }, + { + "epoch": 8.286735010872942, + "grad_norm": 0.15532907843589783, + "learning_rate": 4.8596388542876563e-05, + "loss": 0.0065, + "step": 53350 + }, + { + "epoch": 8.288288288288289, + "grad_norm": 0.49399998784065247, + "learning_rate": 4.857986043811836e-05, + "loss": 0.006, + "step": 53360 + }, + { + "epoch": 8.289841565703634, + "grad_norm": 0.17923223972320557, + "learning_rate": 4.856333248866448e-05, + "loss": 0.0077, + "step": 53370 + }, + { + "epoch": 8.291394843118981, + "grad_norm": 0.20101149380207062, + "learning_rate": 4.85468046963224e-05, + "loss": 0.0083, + "step": 53380 + }, + { + "epoch": 8.292948120534328, + "grad_norm": 0.16718056797981262, + "learning_rate": 4.853027706289957e-05, + "loss": 0.005, + "step": 53390 + }, + { + "epoch": 8.294501397949674, + "grad_norm": 0.35449835658073425, + "learning_rate": 4.851374959020344e-05, + "loss": 0.0071, + "step": 53400 + }, + { + "epoch": 8.29605467536502, + "grad_norm": 0.19529099762439728, + "learning_rate": 4.849722228004141e-05, + "loss": 0.0056, + "step": 53410 + }, + { + "epoch": 8.297607952780366, + "grad_norm": 0.34349265694618225, + "learning_rate": 4.848069513422091e-05, + "loss": 0.0074, + "step": 53420 + }, + { + "epoch": 8.299161230195713, + "grad_norm": 0.31016963720321655, + "learning_rate": 4.8464168154549294e-05, + "loss": 0.0076, + "step": 53430 + }, + { + "epoch": 8.30071450761106, + "grad_norm": 0.28016942739486694, + "learning_rate": 4.844764134283392e-05, + "loss": 0.0073, + "step": 53440 + }, + { + "epoch": 8.302267785026405, + "grad_norm": 0.2039451003074646, + "learning_rate": 4.843111470088218e-05, + "loss": 0.0097, + "step": 53450 + }, + { + "epoch": 8.303821062441752, + "grad_norm": 0.3292144536972046, + "learning_rate": 4.841458823050137e-05, + "loss": 0.0056, + "step": 53460 + }, + { + "epoch": 8.3053743398571, + "grad_norm": 0.20091469585895538, + "learning_rate": 4.8398061933498816e-05, + "loss": 0.0076, + "step": 53470 + }, + { + "epoch": 8.306927617272445, + "grad_norm": 0.14150755107402802, + "learning_rate": 4.83815358116818e-05, + "loss": 0.006, + "step": 53480 + }, + { + "epoch": 8.308480894687792, + "grad_norm": 0.16031338274478912, + "learning_rate": 4.8365009866857584e-05, + "loss": 0.0064, + "step": 53490 + }, + { + "epoch": 8.310034172103137, + "grad_norm": 0.2133699655532837, + "learning_rate": 4.834848410083342e-05, + "loss": 0.0076, + "step": 53500 + }, + { + "epoch": 8.311587449518484, + "grad_norm": 0.22288978099822998, + "learning_rate": 4.833195851541657e-05, + "loss": 0.0064, + "step": 53510 + }, + { + "epoch": 8.313140726933831, + "grad_norm": 0.1296229362487793, + "learning_rate": 4.831543311241423e-05, + "loss": 0.0056, + "step": 53520 + }, + { + "epoch": 8.314694004349176, + "grad_norm": 0.1698113977909088, + "learning_rate": 4.829890789363359e-05, + "loss": 0.0066, + "step": 53530 + }, + { + "epoch": 8.316247281764523, + "grad_norm": 0.1840553730726242, + "learning_rate": 4.828238286088182e-05, + "loss": 0.0109, + "step": 53540 + }, + { + "epoch": 8.317800559179869, + "grad_norm": 0.41518211364746094, + "learning_rate": 4.826585801596609e-05, + "loss": 0.0071, + "step": 53550 + }, + { + "epoch": 8.319353836595216, + "grad_norm": 0.08875428885221481, + "learning_rate": 4.8249333360693486e-05, + "loss": 0.0065, + "step": 53560 + }, + { + "epoch": 8.320907114010563, + "grad_norm": 0.5619421601295471, + "learning_rate": 4.8232808896871185e-05, + "loss": 0.0092, + "step": 53570 + }, + { + "epoch": 8.322460391425908, + "grad_norm": 0.1673527956008911, + "learning_rate": 4.821628462630624e-05, + "loss": 0.012, + "step": 53580 + }, + { + "epoch": 8.324013668841255, + "grad_norm": 0.1586800217628479, + "learning_rate": 4.8199760550805735e-05, + "loss": 0.0069, + "step": 53590 + }, + { + "epoch": 8.325566946256602, + "grad_norm": 0.1860741823911667, + "learning_rate": 4.818323667217669e-05, + "loss": 0.0069, + "step": 53600 + }, + { + "epoch": 8.327120223671947, + "grad_norm": 0.22531013190746307, + "learning_rate": 4.8166712992226165e-05, + "loss": 0.0066, + "step": 53610 + }, + { + "epoch": 8.328673501087295, + "grad_norm": 0.22691883146762848, + "learning_rate": 4.8150189512761154e-05, + "loss": 0.0063, + "step": 53620 + }, + { + "epoch": 8.33022677850264, + "grad_norm": 0.1853361278772354, + "learning_rate": 4.813366623558862e-05, + "loss": 0.0081, + "step": 53630 + }, + { + "epoch": 8.331780055917987, + "grad_norm": 0.31155240535736084, + "learning_rate": 4.811714316251554e-05, + "loss": 0.0061, + "step": 53640 + }, + { + "epoch": 8.333333333333334, + "grad_norm": 0.2361530363559723, + "learning_rate": 4.810062029534886e-05, + "loss": 0.0074, + "step": 53650 + }, + { + "epoch": 8.33488661074868, + "grad_norm": 0.2616337835788727, + "learning_rate": 4.8084097635895505e-05, + "loss": 0.0102, + "step": 53660 + }, + { + "epoch": 8.336439888164026, + "grad_norm": 0.29184702038764954, + "learning_rate": 4.8067575185962335e-05, + "loss": 0.0054, + "step": 53670 + }, + { + "epoch": 8.337993165579373, + "grad_norm": 0.27784356474876404, + "learning_rate": 4.805105294735625e-05, + "loss": 0.0062, + "step": 53680 + }, + { + "epoch": 8.339546442994719, + "grad_norm": 0.1235315352678299, + "learning_rate": 4.803453092188406e-05, + "loss": 0.005, + "step": 53690 + }, + { + "epoch": 8.341099720410066, + "grad_norm": 0.3121131658554077, + "learning_rate": 4.801800911135263e-05, + "loss": 0.0074, + "step": 53700 + }, + { + "epoch": 8.342652997825411, + "grad_norm": 0.24080131947994232, + "learning_rate": 4.8001487517568744e-05, + "loss": 0.0079, + "step": 53710 + }, + { + "epoch": 8.344206275240758, + "grad_norm": 0.23057889938354492, + "learning_rate": 4.798496614233918e-05, + "loss": 0.0085, + "step": 53720 + }, + { + "epoch": 8.345759552656105, + "grad_norm": 0.25244301557540894, + "learning_rate": 4.796844498747069e-05, + "loss": 0.0061, + "step": 53730 + }, + { + "epoch": 8.34731283007145, + "grad_norm": 0.5329511165618896, + "learning_rate": 4.795192405477e-05, + "loss": 0.0066, + "step": 53740 + }, + { + "epoch": 8.348866107486797, + "grad_norm": 0.14085254073143005, + "learning_rate": 4.79354033460438e-05, + "loss": 0.0064, + "step": 53750 + }, + { + "epoch": 8.350419384902143, + "grad_norm": 0.5300945043563843, + "learning_rate": 4.79188828630988e-05, + "loss": 0.0059, + "step": 53760 + }, + { + "epoch": 8.35197266231749, + "grad_norm": 0.1890885829925537, + "learning_rate": 4.790236260774165e-05, + "loss": 0.006, + "step": 53770 + }, + { + "epoch": 8.353525939732837, + "grad_norm": 0.23003651201725006, + "learning_rate": 4.788584258177896e-05, + "loss": 0.0052, + "step": 53780 + }, + { + "epoch": 8.355079217148182, + "grad_norm": 0.09650277346372604, + "learning_rate": 4.7869322787017356e-05, + "loss": 0.0052, + "step": 53790 + }, + { + "epoch": 8.356632494563529, + "grad_norm": 0.2248273342847824, + "learning_rate": 4.78528032252634e-05, + "loss": 0.0076, + "step": 53800 + }, + { + "epoch": 8.358185771978876, + "grad_norm": 0.23531296849250793, + "learning_rate": 4.783628389832366e-05, + "loss": 0.0068, + "step": 53810 + }, + { + "epoch": 8.359739049394221, + "grad_norm": 0.11047615855932236, + "learning_rate": 4.7819764808004644e-05, + "loss": 0.0052, + "step": 53820 + }, + { + "epoch": 8.361292326809568, + "grad_norm": 0.2536023259162903, + "learning_rate": 4.780324595611289e-05, + "loss": 0.0064, + "step": 53830 + }, + { + "epoch": 8.362845604224914, + "grad_norm": 0.49142560362815857, + "learning_rate": 4.7786727344454854e-05, + "loss": 0.0075, + "step": 53840 + }, + { + "epoch": 8.36439888164026, + "grad_norm": 0.22526124119758606, + "learning_rate": 4.777020897483697e-05, + "loss": 0.0053, + "step": 53850 + }, + { + "epoch": 8.365952159055608, + "grad_norm": 0.4631160497665405, + "learning_rate": 4.775369084906569e-05, + "loss": 0.0081, + "step": 53860 + }, + { + "epoch": 8.367505436470953, + "grad_norm": 0.1767514944076538, + "learning_rate": 4.773717296894742e-05, + "loss": 0.0051, + "step": 53870 + }, + { + "epoch": 8.3690587138863, + "grad_norm": 0.15114672482013702, + "learning_rate": 4.772065533628848e-05, + "loss": 0.0063, + "step": 53880 + }, + { + "epoch": 8.370611991301647, + "grad_norm": 0.2536836862564087, + "learning_rate": 4.7704137952895265e-05, + "loss": 0.0085, + "step": 53890 + }, + { + "epoch": 8.372165268716993, + "grad_norm": 0.35737329721450806, + "learning_rate": 4.768762082057407e-05, + "loss": 0.0056, + "step": 53900 + }, + { + "epoch": 8.37371854613234, + "grad_norm": 0.19514217972755432, + "learning_rate": 4.7671103941131186e-05, + "loss": 0.0055, + "step": 53910 + }, + { + "epoch": 8.375271823547685, + "grad_norm": 0.29156455397605896, + "learning_rate": 4.765458731637287e-05, + "loss": 0.0057, + "step": 53920 + }, + { + "epoch": 8.376825100963032, + "grad_norm": 0.36783918738365173, + "learning_rate": 4.763807094810537e-05, + "loss": 0.0059, + "step": 53930 + }, + { + "epoch": 8.378378378378379, + "grad_norm": 0.3829629421234131, + "learning_rate": 4.762155483813487e-05, + "loss": 0.0066, + "step": 53940 + }, + { + "epoch": 8.379931655793724, + "grad_norm": 0.2921377718448639, + "learning_rate": 4.760503898826754e-05, + "loss": 0.0082, + "step": 53950 + }, + { + "epoch": 8.381484933209071, + "grad_norm": 0.2857135534286499, + "learning_rate": 4.758852340030957e-05, + "loss": 0.0079, + "step": 53960 + }, + { + "epoch": 8.383038210624418, + "grad_norm": 0.1982695311307907, + "learning_rate": 4.757200807606705e-05, + "loss": 0.0053, + "step": 53970 + }, + { + "epoch": 8.384591488039764, + "grad_norm": 0.3025033473968506, + "learning_rate": 4.755549301734607e-05, + "loss": 0.0057, + "step": 53980 + }, + { + "epoch": 8.38614476545511, + "grad_norm": 0.20760273933410645, + "learning_rate": 4.753897822595271e-05, + "loss": 0.0067, + "step": 53990 + }, + { + "epoch": 8.387698042870456, + "grad_norm": 0.16975101828575134, + "learning_rate": 4.7522463703692994e-05, + "loss": 0.0065, + "step": 54000 + }, + { + "epoch": 8.389251320285803, + "grad_norm": 0.2700212001800537, + "learning_rate": 4.75059494523729e-05, + "loss": 0.0068, + "step": 54010 + }, + { + "epoch": 8.39080459770115, + "grad_norm": 0.15591537952423096, + "learning_rate": 4.748943547379843e-05, + "loss": 0.0063, + "step": 54020 + }, + { + "epoch": 8.392357875116495, + "grad_norm": 0.803281307220459, + "learning_rate": 4.7472921769775535e-05, + "loss": 0.0064, + "step": 54030 + }, + { + "epoch": 8.393911152531842, + "grad_norm": 0.2508612275123596, + "learning_rate": 4.745640834211011e-05, + "loss": 0.0061, + "step": 54040 + }, + { + "epoch": 8.39546442994719, + "grad_norm": 0.49277573823928833, + "learning_rate": 4.7439895192608054e-05, + "loss": 0.0072, + "step": 54050 + }, + { + "epoch": 8.397017707362535, + "grad_norm": 0.4120652675628662, + "learning_rate": 4.7423382323075186e-05, + "loss": 0.0086, + "step": 54060 + }, + { + "epoch": 8.398570984777882, + "grad_norm": 0.23105056583881378, + "learning_rate": 4.740686973531737e-05, + "loss": 0.0077, + "step": 54070 + }, + { + "epoch": 8.400124262193227, + "grad_norm": 0.19173787534236908, + "learning_rate": 4.7390357431140375e-05, + "loss": 0.0091, + "step": 54080 + }, + { + "epoch": 8.401677539608574, + "grad_norm": 0.1964334398508072, + "learning_rate": 4.737384541234998e-05, + "loss": 0.0095, + "step": 54090 + }, + { + "epoch": 8.403230817023921, + "grad_norm": 0.5184508562088013, + "learning_rate": 4.735733368075189e-05, + "loss": 0.0093, + "step": 54100 + }, + { + "epoch": 8.404784094439266, + "grad_norm": 0.2597251534461975, + "learning_rate": 4.734082223815182e-05, + "loss": 0.0073, + "step": 54110 + }, + { + "epoch": 8.406337371854613, + "grad_norm": 0.18588438630104065, + "learning_rate": 4.7324311086355436e-05, + "loss": 0.0078, + "step": 54120 + }, + { + "epoch": 8.407890649269959, + "grad_norm": 0.18860378861427307, + "learning_rate": 4.730780022716836e-05, + "loss": 0.0069, + "step": 54130 + }, + { + "epoch": 8.409443926685306, + "grad_norm": 0.3235357403755188, + "learning_rate": 4.729128966239619e-05, + "loss": 0.0067, + "step": 54140 + }, + { + "epoch": 8.410997204100653, + "grad_norm": 0.3141588270664215, + "learning_rate": 4.7274779393844536e-05, + "loss": 0.0083, + "step": 54150 + }, + { + "epoch": 8.412550481515998, + "grad_norm": 0.16951335966587067, + "learning_rate": 4.725826942331891e-05, + "loss": 0.0091, + "step": 54160 + }, + { + "epoch": 8.414103758931345, + "grad_norm": 0.167772576212883, + "learning_rate": 4.724175975262481e-05, + "loss": 0.0075, + "step": 54170 + }, + { + "epoch": 8.415657036346692, + "grad_norm": 0.3210095763206482, + "learning_rate": 4.722525038356772e-05, + "loss": 0.0064, + "step": 54180 + }, + { + "epoch": 8.417210313762038, + "grad_norm": 0.20233426988124847, + "learning_rate": 4.7208741317953075e-05, + "loss": 0.0058, + "step": 54190 + }, + { + "epoch": 8.418763591177385, + "grad_norm": 0.2702416181564331, + "learning_rate": 4.7192232557586264e-05, + "loss": 0.0061, + "step": 54200 + }, + { + "epoch": 8.42031686859273, + "grad_norm": 0.23407471179962158, + "learning_rate": 4.717572410427269e-05, + "loss": 0.006, + "step": 54210 + }, + { + "epoch": 8.421870146008077, + "grad_norm": 0.16231544315814972, + "learning_rate": 4.71592159598177e-05, + "loss": 0.0065, + "step": 54220 + }, + { + "epoch": 8.423423423423424, + "grad_norm": 0.20293031632900238, + "learning_rate": 4.714270812602657e-05, + "loss": 0.0078, + "step": 54230 + }, + { + "epoch": 8.42497670083877, + "grad_norm": 0.11430295556783676, + "learning_rate": 4.712620060470458e-05, + "loss": 0.0054, + "step": 54240 + }, + { + "epoch": 8.426529978254116, + "grad_norm": 0.20569922029972076, + "learning_rate": 4.7109693397656986e-05, + "loss": 0.0055, + "step": 54250 + }, + { + "epoch": 8.428083255669463, + "grad_norm": 0.47953858971595764, + "learning_rate": 4.709318650668894e-05, + "loss": 0.0066, + "step": 54260 + }, + { + "epoch": 8.429636533084809, + "grad_norm": 0.21953026950359344, + "learning_rate": 4.7076679933605676e-05, + "loss": 0.0073, + "step": 54270 + }, + { + "epoch": 8.431189810500156, + "grad_norm": 0.36874306201934814, + "learning_rate": 4.70601736802123e-05, + "loss": 0.0074, + "step": 54280 + }, + { + "epoch": 8.432743087915501, + "grad_norm": 0.25167548656463623, + "learning_rate": 4.704366774831391e-05, + "loss": 0.0056, + "step": 54290 + }, + { + "epoch": 8.434296365330848, + "grad_norm": 0.18190592527389526, + "learning_rate": 4.702716213971557e-05, + "loss": 0.0052, + "step": 54300 + }, + { + "epoch": 8.435849642746195, + "grad_norm": 0.1815323531627655, + "learning_rate": 4.70106568562223e-05, + "loss": 0.0066, + "step": 54310 + }, + { + "epoch": 8.43740292016154, + "grad_norm": 0.28668278455734253, + "learning_rate": 4.6994151899639105e-05, + "loss": 0.0068, + "step": 54320 + }, + { + "epoch": 8.438956197576887, + "grad_norm": 0.19932812452316284, + "learning_rate": 4.697764727177093e-05, + "loss": 0.0061, + "step": 54330 + }, + { + "epoch": 8.440509474992234, + "grad_norm": 0.1358557939529419, + "learning_rate": 4.696114297442272e-05, + "loss": 0.0046, + "step": 54340 + }, + { + "epoch": 8.44206275240758, + "grad_norm": 0.1225380003452301, + "learning_rate": 4.694463900939933e-05, + "loss": 0.009, + "step": 54350 + }, + { + "epoch": 8.443616029822927, + "grad_norm": 0.4209126830101013, + "learning_rate": 4.692813537850564e-05, + "loss": 0.0083, + "step": 54360 + }, + { + "epoch": 8.445169307238272, + "grad_norm": 0.26299038529396057, + "learning_rate": 4.6911632083546445e-05, + "loss": 0.0065, + "step": 54370 + }, + { + "epoch": 8.44672258465362, + "grad_norm": 0.28097835183143616, + "learning_rate": 4.6895129126326525e-05, + "loss": 0.0103, + "step": 54380 + }, + { + "epoch": 8.448275862068966, + "grad_norm": 0.2397453635931015, + "learning_rate": 4.687862650865059e-05, + "loss": 0.0063, + "step": 54390 + }, + { + "epoch": 8.449829139484311, + "grad_norm": 0.2646105885505676, + "learning_rate": 4.686212423232339e-05, + "loss": 0.0077, + "step": 54400 + }, + { + "epoch": 8.451382416899659, + "grad_norm": 0.21972055733203888, + "learning_rate": 4.6845622299149565e-05, + "loss": 0.0047, + "step": 54410 + }, + { + "epoch": 8.452935694315004, + "grad_norm": 0.2564970552921295, + "learning_rate": 4.682912071093374e-05, + "loss": 0.0072, + "step": 54420 + }, + { + "epoch": 8.45448897173035, + "grad_norm": 0.17958052456378937, + "learning_rate": 4.681261946948052e-05, + "loss": 0.0049, + "step": 54430 + }, + { + "epoch": 8.456042249145698, + "grad_norm": 0.3432835042476654, + "learning_rate": 4.679611857659443e-05, + "loss": 0.0094, + "step": 54440 + }, + { + "epoch": 8.457595526561043, + "grad_norm": 0.31379663944244385, + "learning_rate": 4.6779618034079975e-05, + "loss": 0.0045, + "step": 54450 + }, + { + "epoch": 8.45914880397639, + "grad_norm": 0.2095632702112198, + "learning_rate": 4.6763117843741675e-05, + "loss": 0.0054, + "step": 54460 + }, + { + "epoch": 8.460702081391737, + "grad_norm": 0.21227551996707916, + "learning_rate": 4.674661800738394e-05, + "loss": 0.006, + "step": 54470 + }, + { + "epoch": 8.462255358807083, + "grad_norm": 0.24679876863956451, + "learning_rate": 4.673011852681117e-05, + "loss": 0.0051, + "step": 54480 + }, + { + "epoch": 8.46380863622243, + "grad_norm": 0.21866920590400696, + "learning_rate": 4.671361940382772e-05, + "loss": 0.0072, + "step": 54490 + }, + { + "epoch": 8.465361913637775, + "grad_norm": 0.20750844478607178, + "learning_rate": 4.669712064023791e-05, + "loss": 0.0053, + "step": 54500 + }, + { + "epoch": 8.466915191053122, + "grad_norm": 0.6121430993080139, + "learning_rate": 4.668062223784602e-05, + "loss": 0.0066, + "step": 54510 + }, + { + "epoch": 8.468468468468469, + "grad_norm": 0.13529348373413086, + "learning_rate": 4.6664124198456274e-05, + "loss": 0.0075, + "step": 54520 + }, + { + "epoch": 8.470021745883814, + "grad_norm": 0.1671554148197174, + "learning_rate": 4.6647626523872905e-05, + "loss": 0.0053, + "step": 54530 + }, + { + "epoch": 8.471575023299161, + "grad_norm": 0.16548198461532593, + "learning_rate": 4.6631129215900066e-05, + "loss": 0.0059, + "step": 54540 + }, + { + "epoch": 8.473128300714508, + "grad_norm": 0.2874952554702759, + "learning_rate": 4.661463227634187e-05, + "loss": 0.0053, + "step": 54550 + }, + { + "epoch": 8.474681578129854, + "grad_norm": 0.17137935757637024, + "learning_rate": 4.659813570700238e-05, + "loss": 0.0057, + "step": 54560 + }, + { + "epoch": 8.4762348555452, + "grad_norm": 0.26525917649269104, + "learning_rate": 4.658163950968568e-05, + "loss": 0.0064, + "step": 54570 + }, + { + "epoch": 8.477788132960546, + "grad_norm": 0.4148707687854767, + "learning_rate": 4.6565143686195726e-05, + "loss": 0.0068, + "step": 54580 + }, + { + "epoch": 8.479341410375893, + "grad_norm": 0.18232578039169312, + "learning_rate": 4.654864823833649e-05, + "loss": 0.01, + "step": 54590 + }, + { + "epoch": 8.48089468779124, + "grad_norm": 0.1419801265001297, + "learning_rate": 4.65321531679119e-05, + "loss": 0.007, + "step": 54600 + }, + { + "epoch": 8.482447965206585, + "grad_norm": 0.30878639221191406, + "learning_rate": 4.6515658476725834e-05, + "loss": 0.0076, + "step": 54610 + }, + { + "epoch": 8.484001242621932, + "grad_norm": 0.2911578416824341, + "learning_rate": 4.649916416658212e-05, + "loss": 0.0063, + "step": 54620 + }, + { + "epoch": 8.48555452003728, + "grad_norm": 0.21470201015472412, + "learning_rate": 4.648267023928455e-05, + "loss": 0.006, + "step": 54630 + }, + { + "epoch": 8.487107797452625, + "grad_norm": 0.241508349776268, + "learning_rate": 4.646617669663687e-05, + "loss": 0.0072, + "step": 54640 + }, + { + "epoch": 8.488661074867972, + "grad_norm": 0.17475786805152893, + "learning_rate": 4.644968354044277e-05, + "loss": 0.0074, + "step": 54650 + }, + { + "epoch": 8.490214352283317, + "grad_norm": 0.10602608323097229, + "learning_rate": 4.643319077250596e-05, + "loss": 0.0049, + "step": 54660 + }, + { + "epoch": 8.491767629698664, + "grad_norm": 0.26217663288116455, + "learning_rate": 4.6416698394630055e-05, + "loss": 0.0056, + "step": 54670 + }, + { + "epoch": 8.493320907114011, + "grad_norm": 0.33748579025268555, + "learning_rate": 4.640020640861862e-05, + "loss": 0.0069, + "step": 54680 + }, + { + "epoch": 8.494874184529356, + "grad_norm": 0.2274480015039444, + "learning_rate": 4.6383714816275195e-05, + "loss": 0.0078, + "step": 54690 + }, + { + "epoch": 8.496427461944704, + "grad_norm": 0.21340057253837585, + "learning_rate": 4.636722361940329e-05, + "loss": 0.0068, + "step": 54700 + }, + { + "epoch": 8.497980739360049, + "grad_norm": 0.22977085411548615, + "learning_rate": 4.635073281980632e-05, + "loss": 0.0053, + "step": 54710 + }, + { + "epoch": 8.499534016775396, + "grad_norm": 0.28317925333976746, + "learning_rate": 4.633424241928775e-05, + "loss": 0.0039, + "step": 54720 + }, + { + "epoch": 8.501087294190743, + "grad_norm": 0.18275104463100433, + "learning_rate": 4.6317752419650914e-05, + "loss": 0.0063, + "step": 54730 + }, + { + "epoch": 8.502640571606088, + "grad_norm": 0.1527637541294098, + "learning_rate": 4.630126282269914e-05, + "loss": 0.0056, + "step": 54740 + }, + { + "epoch": 8.504193849021435, + "grad_norm": 0.23702040314674377, + "learning_rate": 4.628477363023569e-05, + "loss": 0.005, + "step": 54750 + }, + { + "epoch": 8.505747126436782, + "grad_norm": 0.13286861777305603, + "learning_rate": 4.626828484406383e-05, + "loss": 0.0054, + "step": 54760 + }, + { + "epoch": 8.507300403852128, + "grad_norm": 0.14861688017845154, + "learning_rate": 4.625179646598669e-05, + "loss": 0.0047, + "step": 54770 + }, + { + "epoch": 8.508853681267475, + "grad_norm": 0.3077995181083679, + "learning_rate": 4.6235308497807487e-05, + "loss": 0.0054, + "step": 54780 + }, + { + "epoch": 8.51040695868282, + "grad_norm": 0.17614279687404633, + "learning_rate": 4.6218820941329276e-05, + "loss": 0.0041, + "step": 54790 + }, + { + "epoch": 8.511960236098167, + "grad_norm": 0.2384728193283081, + "learning_rate": 4.620233379835513e-05, + "loss": 0.0068, + "step": 54800 + }, + { + "epoch": 8.513513513513514, + "grad_norm": 0.20600563287734985, + "learning_rate": 4.618584707068804e-05, + "loss": 0.0078, + "step": 54810 + }, + { + "epoch": 8.51506679092886, + "grad_norm": 0.269608736038208, + "learning_rate": 4.616936076013099e-05, + "loss": 0.0047, + "step": 54820 + }, + { + "epoch": 8.516620068344206, + "grad_norm": 0.24491217732429504, + "learning_rate": 4.615287486848688e-05, + "loss": 0.0043, + "step": 54830 + }, + { + "epoch": 8.518173345759553, + "grad_norm": 0.26782262325286865, + "learning_rate": 4.6136389397558584e-05, + "loss": 0.0055, + "step": 54840 + }, + { + "epoch": 8.519726623174899, + "grad_norm": 0.30399373173713684, + "learning_rate": 4.6119904349148935e-05, + "loss": 0.0056, + "step": 54850 + }, + { + "epoch": 8.521279900590246, + "grad_norm": 0.25426438450813293, + "learning_rate": 4.6103419725060726e-05, + "loss": 0.007, + "step": 54860 + }, + { + "epoch": 8.522833178005591, + "grad_norm": 0.23988290131092072, + "learning_rate": 4.608693552709667e-05, + "loss": 0.0063, + "step": 54870 + }, + { + "epoch": 8.524386455420938, + "grad_norm": 0.1609080731868744, + "learning_rate": 4.607045175705947e-05, + "loss": 0.0078, + "step": 54880 + }, + { + "epoch": 8.525939732836285, + "grad_norm": 0.1895117610692978, + "learning_rate": 4.605396841675176e-05, + "loss": 0.0103, + "step": 54890 + }, + { + "epoch": 8.52749301025163, + "grad_norm": 0.182957723736763, + "learning_rate": 4.6037485507976094e-05, + "loss": 0.0097, + "step": 54900 + }, + { + "epoch": 8.529046287666977, + "grad_norm": 0.3214780390262604, + "learning_rate": 4.602100303253509e-05, + "loss": 0.0061, + "step": 54910 + }, + { + "epoch": 8.530599565082325, + "grad_norm": 0.23340922594070435, + "learning_rate": 4.600452099223122e-05, + "loss": 0.0074, + "step": 54920 + }, + { + "epoch": 8.53215284249767, + "grad_norm": 0.23200125992298126, + "learning_rate": 4.598803938886691e-05, + "loss": 0.0074, + "step": 54930 + }, + { + "epoch": 8.533706119913017, + "grad_norm": 0.28818637132644653, + "learning_rate": 4.597155822424459e-05, + "loss": 0.0053, + "step": 54940 + }, + { + "epoch": 8.535259397328362, + "grad_norm": 0.17189039289951324, + "learning_rate": 4.59550775001666e-05, + "loss": 0.0048, + "step": 54950 + }, + { + "epoch": 8.53681267474371, + "grad_norm": 0.2674795985221863, + "learning_rate": 4.5938597218435234e-05, + "loss": 0.007, + "step": 54960 + }, + { + "epoch": 8.538365952159056, + "grad_norm": 0.5052081942558289, + "learning_rate": 4.592211738085279e-05, + "loss": 0.009, + "step": 54970 + }, + { + "epoch": 8.539919229574402, + "grad_norm": 0.3096608817577362, + "learning_rate": 4.590563798922146e-05, + "loss": 0.0068, + "step": 54980 + }, + { + "epoch": 8.541472506989749, + "grad_norm": 0.22689306735992432, + "learning_rate": 4.5889159045343404e-05, + "loss": 0.006, + "step": 54990 + }, + { + "epoch": 8.543025784405096, + "grad_norm": 0.20061905682086945, + "learning_rate": 4.5872680551020734e-05, + "loss": 0.0047, + "step": 55000 + }, + { + "epoch": 8.544579061820441, + "grad_norm": 0.3114246726036072, + "learning_rate": 4.585620250805552e-05, + "loss": 0.008, + "step": 55010 + }, + { + "epoch": 8.546132339235788, + "grad_norm": 0.217267245054245, + "learning_rate": 4.583972491824975e-05, + "loss": 0.0064, + "step": 55020 + }, + { + "epoch": 8.547685616651133, + "grad_norm": 0.19950760900974274, + "learning_rate": 4.582324778340541e-05, + "loss": 0.0079, + "step": 55030 + }, + { + "epoch": 8.54923889406648, + "grad_norm": 0.18770767748355865, + "learning_rate": 4.5806771105324424e-05, + "loss": 0.0055, + "step": 55040 + }, + { + "epoch": 8.550792171481827, + "grad_norm": 0.21583963930606842, + "learning_rate": 4.579029488580865e-05, + "loss": 0.0055, + "step": 55050 + }, + { + "epoch": 8.552345448897173, + "grad_norm": 0.25700893998146057, + "learning_rate": 4.5773819126659876e-05, + "loss": 0.0052, + "step": 55060 + }, + { + "epoch": 8.55389872631252, + "grad_norm": 0.253864049911499, + "learning_rate": 4.575734382967991e-05, + "loss": 0.006, + "step": 55070 + }, + { + "epoch": 8.555452003727865, + "grad_norm": 0.1492828130722046, + "learning_rate": 4.574086899667044e-05, + "loss": 0.0061, + "step": 55080 + }, + { + "epoch": 8.557005281143212, + "grad_norm": 0.2857794761657715, + "learning_rate": 4.5724394629433106e-05, + "loss": 0.0062, + "step": 55090 + }, + { + "epoch": 8.558558558558559, + "grad_norm": 0.2658916711807251, + "learning_rate": 4.570792072976958e-05, + "loss": 0.0082, + "step": 55100 + }, + { + "epoch": 8.560111835973904, + "grad_norm": 0.3267870843410492, + "learning_rate": 4.569144729948138e-05, + "loss": 0.0068, + "step": 55110 + }, + { + "epoch": 8.561665113389251, + "grad_norm": 0.4088593125343323, + "learning_rate": 4.5674974340370033e-05, + "loss": 0.0083, + "step": 55120 + }, + { + "epoch": 8.563218390804598, + "grad_norm": 0.27827897667884827, + "learning_rate": 4.565850185423699e-05, + "loss": 0.0069, + "step": 55130 + }, + { + "epoch": 8.564771668219944, + "grad_norm": 0.39009889960289, + "learning_rate": 4.564202984288365e-05, + "loss": 0.0101, + "step": 55140 + }, + { + "epoch": 8.56632494563529, + "grad_norm": 0.18031854927539825, + "learning_rate": 4.5625558308111356e-05, + "loss": 0.0068, + "step": 55150 + }, + { + "epoch": 8.567878223050636, + "grad_norm": 0.16561345756053925, + "learning_rate": 4.560908725172145e-05, + "loss": 0.0061, + "step": 55160 + }, + { + "epoch": 8.569431500465983, + "grad_norm": 0.33497706055641174, + "learning_rate": 4.559261667551516e-05, + "loss": 0.0055, + "step": 55170 + }, + { + "epoch": 8.57098477788133, + "grad_norm": 0.19301965832710266, + "learning_rate": 4.5576146581293685e-05, + "loss": 0.0057, + "step": 55180 + }, + { + "epoch": 8.572538055296675, + "grad_norm": 0.14372262358665466, + "learning_rate": 4.5559676970858164e-05, + "loss": 0.0078, + "step": 55190 + }, + { + "epoch": 8.574091332712023, + "grad_norm": 0.28477373719215393, + "learning_rate": 4.5543207846009705e-05, + "loss": 0.0076, + "step": 55200 + }, + { + "epoch": 8.57564461012737, + "grad_norm": 0.31533282995224, + "learning_rate": 4.552673920854933e-05, + "loss": 0.0063, + "step": 55210 + }, + { + "epoch": 8.577197887542715, + "grad_norm": 0.16613726317882538, + "learning_rate": 4.5510271060278006e-05, + "loss": 0.0053, + "step": 55220 + }, + { + "epoch": 8.578751164958062, + "grad_norm": 0.6077223420143127, + "learning_rate": 4.5493803402996704e-05, + "loss": 0.0068, + "step": 55230 + }, + { + "epoch": 8.580304442373407, + "grad_norm": 0.43858635425567627, + "learning_rate": 4.5477336238506295e-05, + "loss": 0.0101, + "step": 55240 + }, + { + "epoch": 8.581857719788754, + "grad_norm": 0.15344786643981934, + "learning_rate": 4.546086956860759e-05, + "loss": 0.0084, + "step": 55250 + }, + { + "epoch": 8.583410997204101, + "grad_norm": 0.1480504870414734, + "learning_rate": 4.5444403395101374e-05, + "loss": 0.0055, + "step": 55260 + }, + { + "epoch": 8.584964274619447, + "grad_norm": 0.16951140761375427, + "learning_rate": 4.542793771978834e-05, + "loss": 0.0082, + "step": 55270 + }, + { + "epoch": 8.586517552034794, + "grad_norm": 0.2617781162261963, + "learning_rate": 4.541147254446916e-05, + "loss": 0.0058, + "step": 55280 + }, + { + "epoch": 8.588070829450139, + "grad_norm": 0.4121676981449127, + "learning_rate": 4.539500787094445e-05, + "loss": 0.0084, + "step": 55290 + }, + { + "epoch": 8.589624106865486, + "grad_norm": 0.23759566247463226, + "learning_rate": 4.537854370101475e-05, + "loss": 0.0073, + "step": 55300 + }, + { + "epoch": 8.591177384280833, + "grad_norm": 0.30574679374694824, + "learning_rate": 4.5362080036480566e-05, + "loss": 0.0077, + "step": 55310 + }, + { + "epoch": 8.592730661696178, + "grad_norm": 0.2816096246242523, + "learning_rate": 4.534561687914235e-05, + "loss": 0.0067, + "step": 55320 + }, + { + "epoch": 8.594283939111525, + "grad_norm": 0.23870998620986938, + "learning_rate": 4.5329154230800466e-05, + "loss": 0.0055, + "step": 55330 + }, + { + "epoch": 8.595837216526872, + "grad_norm": 0.2957586944103241, + "learning_rate": 4.531269209325525e-05, + "loss": 0.0078, + "step": 55340 + }, + { + "epoch": 8.597390493942218, + "grad_norm": 0.1525478959083557, + "learning_rate": 4.529623046830696e-05, + "loss": 0.0053, + "step": 55350 + }, + { + "epoch": 8.598943771357565, + "grad_norm": 0.3713650703430176, + "learning_rate": 4.5279769357755866e-05, + "loss": 0.0093, + "step": 55360 + }, + { + "epoch": 8.600497048772912, + "grad_norm": 0.23348796367645264, + "learning_rate": 4.5263308763402084e-05, + "loss": 0.0088, + "step": 55370 + }, + { + "epoch": 8.602050326188257, + "grad_norm": 0.1648016721010208, + "learning_rate": 4.524684868704574e-05, + "loss": 0.0077, + "step": 55380 + }, + { + "epoch": 8.603603603603604, + "grad_norm": 0.2833435833454132, + "learning_rate": 4.523038913048688e-05, + "loss": 0.0079, + "step": 55390 + }, + { + "epoch": 8.60515688101895, + "grad_norm": 0.2806582450866699, + "learning_rate": 4.521393009552549e-05, + "loss": 0.0054, + "step": 55400 + }, + { + "epoch": 8.606710158434296, + "grad_norm": 0.17026187479496002, + "learning_rate": 4.5197471583961484e-05, + "loss": 0.0083, + "step": 55410 + }, + { + "epoch": 8.608263435849643, + "grad_norm": 0.183894544839859, + "learning_rate": 4.51810135975948e-05, + "loss": 0.0067, + "step": 55420 + }, + { + "epoch": 8.609816713264989, + "grad_norm": 0.20045925676822662, + "learning_rate": 4.516455613822521e-05, + "loss": 0.0114, + "step": 55430 + }, + { + "epoch": 8.611369990680336, + "grad_norm": 0.19581283628940582, + "learning_rate": 4.514809920765249e-05, + "loss": 0.0085, + "step": 55440 + }, + { + "epoch": 8.612923268095681, + "grad_norm": 0.4273311197757721, + "learning_rate": 4.5131642807676356e-05, + "loss": 0.007, + "step": 55450 + }, + { + "epoch": 8.614476545511028, + "grad_norm": 0.448529988527298, + "learning_rate": 4.511518694009644e-05, + "loss": 0.0085, + "step": 55460 + }, + { + "epoch": 8.616029822926375, + "grad_norm": 0.23427070677280426, + "learning_rate": 4.509873160671231e-05, + "loss": 0.0073, + "step": 55470 + }, + { + "epoch": 8.61758310034172, + "grad_norm": 0.3896418511867523, + "learning_rate": 4.508227680932356e-05, + "loss": 0.0076, + "step": 55480 + }, + { + "epoch": 8.619136377757068, + "grad_norm": 0.2963174283504486, + "learning_rate": 4.5065822549729605e-05, + "loss": 0.0062, + "step": 55490 + }, + { + "epoch": 8.620689655172415, + "grad_norm": 0.5088275671005249, + "learning_rate": 4.504936882972989e-05, + "loss": 0.0084, + "step": 55500 + }, + { + "epoch": 8.62224293258776, + "grad_norm": 0.19740644097328186, + "learning_rate": 4.503291565112374e-05, + "loss": 0.0058, + "step": 55510 + }, + { + "epoch": 8.623796210003107, + "grad_norm": 0.403283029794693, + "learning_rate": 4.501646301571048e-05, + "loss": 0.0078, + "step": 55520 + }, + { + "epoch": 8.625349487418452, + "grad_norm": 0.21393412351608276, + "learning_rate": 4.500001092528933e-05, + "loss": 0.0054, + "step": 55530 + }, + { + "epoch": 8.6269027648338, + "grad_norm": 0.5797279477119446, + "learning_rate": 4.4983559381659455e-05, + "loss": 0.0064, + "step": 55540 + }, + { + "epoch": 8.628456042249146, + "grad_norm": 0.2134561836719513, + "learning_rate": 4.496710838661999e-05, + "loss": 0.0051, + "step": 55550 + }, + { + "epoch": 8.630009319664492, + "grad_norm": 0.17335368692874908, + "learning_rate": 4.495065794196999e-05, + "loss": 0.0079, + "step": 55560 + }, + { + "epoch": 8.631562597079839, + "grad_norm": 0.19764338433742523, + "learning_rate": 4.4934208049508455e-05, + "loss": 0.0059, + "step": 55570 + }, + { + "epoch": 8.633115874495186, + "grad_norm": 0.21298463642597198, + "learning_rate": 4.4917758711034303e-05, + "loss": 0.0062, + "step": 55580 + }, + { + "epoch": 8.634669151910531, + "grad_norm": 0.4730924963951111, + "learning_rate": 4.490130992834642e-05, + "loss": 0.0063, + "step": 55590 + }, + { + "epoch": 8.636222429325878, + "grad_norm": 0.2021186202764511, + "learning_rate": 4.488486170324359e-05, + "loss": 0.0079, + "step": 55600 + }, + { + "epoch": 8.637775706741223, + "grad_norm": 0.2909301519393921, + "learning_rate": 4.486841403752463e-05, + "loss": 0.0058, + "step": 55610 + }, + { + "epoch": 8.63932898415657, + "grad_norm": 0.26200658082962036, + "learning_rate": 4.4851966932988174e-05, + "loss": 0.007, + "step": 55620 + }, + { + "epoch": 8.640882261571917, + "grad_norm": 0.21988344192504883, + "learning_rate": 4.4835520391432894e-05, + "loss": 0.0054, + "step": 55630 + }, + { + "epoch": 8.642435538987263, + "grad_norm": 0.2236533910036087, + "learning_rate": 4.481907441465733e-05, + "loss": 0.0055, + "step": 55640 + }, + { + "epoch": 8.64398881640261, + "grad_norm": 0.3338603675365448, + "learning_rate": 4.480262900446e-05, + "loss": 0.0059, + "step": 55650 + }, + { + "epoch": 8.645542093817955, + "grad_norm": 0.2547321319580078, + "learning_rate": 4.478618416263933e-05, + "loss": 0.0071, + "step": 55660 + }, + { + "epoch": 8.647095371233302, + "grad_norm": 0.19845078885555267, + "learning_rate": 4.476973989099373e-05, + "loss": 0.0059, + "step": 55670 + }, + { + "epoch": 8.64864864864865, + "grad_norm": 0.17970497906208038, + "learning_rate": 4.475329619132153e-05, + "loss": 0.0069, + "step": 55680 + }, + { + "epoch": 8.650201926063994, + "grad_norm": 0.256613165140152, + "learning_rate": 4.4736853065420956e-05, + "loss": 0.006, + "step": 55690 + }, + { + "epoch": 8.651755203479341, + "grad_norm": 0.17619730532169342, + "learning_rate": 4.472041051509023e-05, + "loss": 0.0051, + "step": 55700 + }, + { + "epoch": 8.653308480894689, + "grad_norm": 0.1566261649131775, + "learning_rate": 4.470396854212746e-05, + "loss": 0.0059, + "step": 55710 + }, + { + "epoch": 8.654861758310034, + "grad_norm": 0.11662936955690384, + "learning_rate": 4.468752714833072e-05, + "loss": 0.0061, + "step": 55720 + }, + { + "epoch": 8.65641503572538, + "grad_norm": 0.23173214495182037, + "learning_rate": 4.467108633549803e-05, + "loss": 0.0061, + "step": 55730 + }, + { + "epoch": 8.657968313140726, + "grad_norm": 0.17396700382232666, + "learning_rate": 4.4654646105427335e-05, + "loss": 0.0054, + "step": 55740 + }, + { + "epoch": 8.659521590556073, + "grad_norm": 0.20560355484485626, + "learning_rate": 4.463820645991651e-05, + "loss": 0.0067, + "step": 55750 + }, + { + "epoch": 8.66107486797142, + "grad_norm": 1.15131676197052, + "learning_rate": 4.462176740076334e-05, + "loss": 0.0054, + "step": 55760 + }, + { + "epoch": 8.662628145386766, + "grad_norm": 0.46452513337135315, + "learning_rate": 4.4605328929765624e-05, + "loss": 0.0087, + "step": 55770 + }, + { + "epoch": 8.664181422802113, + "grad_norm": 0.16910643875598907, + "learning_rate": 4.458889104872102e-05, + "loss": 0.0087, + "step": 55780 + }, + { + "epoch": 8.66573470021746, + "grad_norm": 0.16910357773303986, + "learning_rate": 4.457245375942715e-05, + "loss": 0.0073, + "step": 55790 + }, + { + "epoch": 8.667287977632805, + "grad_norm": 0.18822777271270752, + "learning_rate": 4.455601706368157e-05, + "loss": 0.0066, + "step": 55800 + }, + { + "epoch": 8.668841255048152, + "grad_norm": 0.6987885236740112, + "learning_rate": 4.453958096328179e-05, + "loss": 0.0055, + "step": 55810 + }, + { + "epoch": 8.670394532463497, + "grad_norm": 0.28588059544563293, + "learning_rate": 4.452314546002523e-05, + "loss": 0.0079, + "step": 55820 + }, + { + "epoch": 8.671947809878844, + "grad_norm": 0.2548387944698334, + "learning_rate": 4.4506710555709246e-05, + "loss": 0.007, + "step": 55830 + }, + { + "epoch": 8.673501087294191, + "grad_norm": 0.4296862483024597, + "learning_rate": 4.449027625213114e-05, + "loss": 0.0055, + "step": 55840 + }, + { + "epoch": 8.675054364709537, + "grad_norm": 0.3271113634109497, + "learning_rate": 4.4473842551088135e-05, + "loss": 0.0076, + "step": 55850 + }, + { + "epoch": 8.676607642124884, + "grad_norm": 0.23995527625083923, + "learning_rate": 4.445740945437738e-05, + "loss": 0.0043, + "step": 55860 + }, + { + "epoch": 8.678160919540229, + "grad_norm": 0.748514711856842, + "learning_rate": 4.444097696379602e-05, + "loss": 0.0056, + "step": 55870 + }, + { + "epoch": 8.679714196955576, + "grad_norm": 0.18538664281368256, + "learning_rate": 4.442454508114106e-05, + "loss": 0.0068, + "step": 55880 + }, + { + "epoch": 8.681267474370923, + "grad_norm": 0.2413392961025238, + "learning_rate": 4.4408113808209465e-05, + "loss": 0.0066, + "step": 55890 + }, + { + "epoch": 8.682820751786268, + "grad_norm": 0.14215068519115448, + "learning_rate": 4.439168314679813e-05, + "loss": 0.0055, + "step": 55900 + }, + { + "epoch": 8.684374029201615, + "grad_norm": 0.30329790711402893, + "learning_rate": 4.43752530987039e-05, + "loss": 0.0078, + "step": 55910 + }, + { + "epoch": 8.685927306616962, + "grad_norm": 0.43539708852767944, + "learning_rate": 4.4358823665723515e-05, + "loss": 0.0064, + "step": 55920 + }, + { + "epoch": 8.687480584032308, + "grad_norm": 0.5762438774108887, + "learning_rate": 4.434239484965371e-05, + "loss": 0.0064, + "step": 55930 + }, + { + "epoch": 8.689033861447655, + "grad_norm": 0.3900221288204193, + "learning_rate": 4.4325966652291103e-05, + "loss": 0.007, + "step": 55940 + }, + { + "epoch": 8.690587138863002, + "grad_norm": 0.18676355481147766, + "learning_rate": 4.430953907543225e-05, + "loss": 0.0086, + "step": 55950 + }, + { + "epoch": 8.692140416278347, + "grad_norm": 0.12127576023340225, + "learning_rate": 4.4293112120873645e-05, + "loss": 0.0067, + "step": 55960 + }, + { + "epoch": 8.693693693693694, + "grad_norm": 0.19136373698711395, + "learning_rate": 4.427668579041172e-05, + "loss": 0.0083, + "step": 55970 + }, + { + "epoch": 8.69524697110904, + "grad_norm": 0.22585509717464447, + "learning_rate": 4.4260260085842816e-05, + "loss": 0.0085, + "step": 55980 + }, + { + "epoch": 8.696800248524386, + "grad_norm": 0.1949659138917923, + "learning_rate": 4.4243835008963256e-05, + "loss": 0.0067, + "step": 55990 + }, + { + "epoch": 8.698353525939734, + "grad_norm": 0.23766061663627625, + "learning_rate": 4.422741056156925e-05, + "loss": 0.0051, + "step": 56000 + }, + { + "epoch": 8.699906803355079, + "grad_norm": 0.3492453098297119, + "learning_rate": 4.421098674545693e-05, + "loss": 0.007, + "step": 56010 + }, + { + "epoch": 8.701460080770426, + "grad_norm": 0.2799990475177765, + "learning_rate": 4.419456356242241e-05, + "loss": 0.0076, + "step": 56020 + }, + { + "epoch": 8.703013358185771, + "grad_norm": 0.2582622170448303, + "learning_rate": 4.4178141014261685e-05, + "loss": 0.0054, + "step": 56030 + }, + { + "epoch": 8.704566635601118, + "grad_norm": 0.2697222828865051, + "learning_rate": 4.416171910277071e-05, + "loss": 0.0054, + "step": 56040 + }, + { + "epoch": 8.706119913016465, + "grad_norm": 0.1143541932106018, + "learning_rate": 4.414529782974533e-05, + "loss": 0.0063, + "step": 56050 + }, + { + "epoch": 8.70767319043181, + "grad_norm": 0.18393227458000183, + "learning_rate": 4.41288771969814e-05, + "loss": 0.0044, + "step": 56060 + }, + { + "epoch": 8.709226467847158, + "grad_norm": 0.13601712882518768, + "learning_rate": 4.411245720627464e-05, + "loss": 0.0046, + "step": 56070 + }, + { + "epoch": 8.710779745262505, + "grad_norm": 0.40703997015953064, + "learning_rate": 4.40960378594207e-05, + "loss": 0.0076, + "step": 56080 + }, + { + "epoch": 8.71233302267785, + "grad_norm": 0.288506418466568, + "learning_rate": 4.407961915821519e-05, + "loss": 0.0056, + "step": 56090 + }, + { + "epoch": 8.713886300093197, + "grad_norm": 0.34214305877685547, + "learning_rate": 4.406320110445362e-05, + "loss": 0.005, + "step": 56100 + }, + { + "epoch": 8.715439577508542, + "grad_norm": 0.23833519220352173, + "learning_rate": 4.404678369993144e-05, + "loss": 0.006, + "step": 56110 + }, + { + "epoch": 8.71699285492389, + "grad_norm": 0.13876967132091522, + "learning_rate": 4.403036694644406e-05, + "loss": 0.0066, + "step": 56120 + }, + { + "epoch": 8.718546132339236, + "grad_norm": 0.20690500736236572, + "learning_rate": 4.4013950845786764e-05, + "loss": 0.0048, + "step": 56130 + }, + { + "epoch": 8.720099409754582, + "grad_norm": 0.4073501527309418, + "learning_rate": 4.399753539975482e-05, + "loss": 0.0067, + "step": 56140 + }, + { + "epoch": 8.721652687169929, + "grad_norm": 0.34934675693511963, + "learning_rate": 4.398112061014337e-05, + "loss": 0.0074, + "step": 56150 + }, + { + "epoch": 8.723205964585276, + "grad_norm": 0.2316003441810608, + "learning_rate": 4.396470647874753e-05, + "loss": 0.0078, + "step": 56160 + }, + { + "epoch": 8.724759242000621, + "grad_norm": 0.4112149477005005, + "learning_rate": 4.394829300736229e-05, + "loss": 0.0049, + "step": 56170 + }, + { + "epoch": 8.726312519415968, + "grad_norm": 0.2459195852279663, + "learning_rate": 4.393188019778265e-05, + "loss": 0.0069, + "step": 56180 + }, + { + "epoch": 8.727865796831313, + "grad_norm": 0.5638589262962341, + "learning_rate": 4.391546805180347e-05, + "loss": 0.0092, + "step": 56190 + }, + { + "epoch": 8.72941907424666, + "grad_norm": 0.13104484975337982, + "learning_rate": 4.389905657121955e-05, + "loss": 0.0051, + "step": 56200 + }, + { + "epoch": 8.730972351662007, + "grad_norm": 0.35849958658218384, + "learning_rate": 4.3882645757825635e-05, + "loss": 0.0053, + "step": 56210 + }, + { + "epoch": 8.732525629077353, + "grad_norm": 0.18163809180259705, + "learning_rate": 4.386623561341637e-05, + "loss": 0.0072, + "step": 56220 + }, + { + "epoch": 8.7340789064927, + "grad_norm": 0.5434969067573547, + "learning_rate": 4.384982613978637e-05, + "loss": 0.0058, + "step": 56230 + }, + { + "epoch": 8.735632183908045, + "grad_norm": 0.27854958176612854, + "learning_rate": 4.383341733873012e-05, + "loss": 0.006, + "step": 56240 + }, + { + "epoch": 8.737185461323392, + "grad_norm": 0.5229198336601257, + "learning_rate": 4.381700921204209e-05, + "loss": 0.008, + "step": 56250 + }, + { + "epoch": 8.73873873873874, + "grad_norm": 0.19081741571426392, + "learning_rate": 4.380060176151663e-05, + "loss": 0.0089, + "step": 56260 + }, + { + "epoch": 8.740292016154084, + "grad_norm": 0.41690632700920105, + "learning_rate": 4.378419498894805e-05, + "loss": 0.0075, + "step": 56270 + }, + { + "epoch": 8.741845293569432, + "grad_norm": 0.2993604242801666, + "learning_rate": 4.376778889613056e-05, + "loss": 0.0057, + "step": 56280 + }, + { + "epoch": 8.743398570984779, + "grad_norm": 0.11619120091199875, + "learning_rate": 4.375138348485831e-05, + "loss": 0.0045, + "step": 56290 + }, + { + "epoch": 8.744951848400124, + "grad_norm": 0.2343171089887619, + "learning_rate": 4.3734978756925346e-05, + "loss": 0.0069, + "step": 56300 + }, + { + "epoch": 8.746505125815471, + "grad_norm": 0.2724749445915222, + "learning_rate": 4.3718574714125706e-05, + "loss": 0.0047, + "step": 56310 + }, + { + "epoch": 8.748058403230818, + "grad_norm": 0.21381983160972595, + "learning_rate": 4.370217135825329e-05, + "loss": 0.0075, + "step": 56320 + }, + { + "epoch": 8.749611680646163, + "grad_norm": 0.19311457872390747, + "learning_rate": 4.368576869110194e-05, + "loss": 0.0069, + "step": 56330 + }, + { + "epoch": 8.75116495806151, + "grad_norm": 0.20489463210105896, + "learning_rate": 4.366936671446544e-05, + "loss": 0.0067, + "step": 56340 + }, + { + "epoch": 8.752718235476856, + "grad_norm": 0.19505465030670166, + "learning_rate": 4.3652965430137474e-05, + "loss": 0.0049, + "step": 56350 + }, + { + "epoch": 8.754271512892203, + "grad_norm": 0.38763564825057983, + "learning_rate": 4.3636564839911646e-05, + "loss": 0.0067, + "step": 56360 + }, + { + "epoch": 8.75582479030755, + "grad_norm": 0.3902176022529602, + "learning_rate": 4.3620164945581545e-05, + "loss": 0.0061, + "step": 56370 + }, + { + "epoch": 8.757378067722895, + "grad_norm": 0.29930517077445984, + "learning_rate": 4.360376574894061e-05, + "loss": 0.0069, + "step": 56380 + }, + { + "epoch": 8.758931345138242, + "grad_norm": 0.2790137231349945, + "learning_rate": 4.358736725178224e-05, + "loss": 0.0067, + "step": 56390 + }, + { + "epoch": 8.760484622553587, + "grad_norm": 0.2736736238002777, + "learning_rate": 4.357096945589974e-05, + "loss": 0.0051, + "step": 56400 + }, + { + "epoch": 8.762037899968934, + "grad_norm": 0.2294909507036209, + "learning_rate": 4.355457236308636e-05, + "loss": 0.0046, + "step": 56410 + }, + { + "epoch": 8.763591177384281, + "grad_norm": 0.2404855191707611, + "learning_rate": 4.353817597513526e-05, + "loss": 0.0084, + "step": 56420 + }, + { + "epoch": 8.765144454799627, + "grad_norm": 0.14123409986495972, + "learning_rate": 4.352178029383948e-05, + "loss": 0.0052, + "step": 56430 + }, + { + "epoch": 8.766697732214974, + "grad_norm": 0.19599711894989014, + "learning_rate": 4.3505385320992105e-05, + "loss": 0.0056, + "step": 56440 + }, + { + "epoch": 8.76825100963032, + "grad_norm": 0.23981647193431854, + "learning_rate": 4.348899105838602e-05, + "loss": 0.005, + "step": 56450 + }, + { + "epoch": 8.769804287045666, + "grad_norm": 0.3162534236907959, + "learning_rate": 4.3472597507814087e-05, + "loss": 0.0068, + "step": 56460 + }, + { + "epoch": 8.771357564461013, + "grad_norm": 0.22905337810516357, + "learning_rate": 4.3456204671069066e-05, + "loss": 0.0059, + "step": 56470 + }, + { + "epoch": 8.772910841876358, + "grad_norm": 0.1777656227350235, + "learning_rate": 4.343981254994367e-05, + "loss": 0.0063, + "step": 56480 + }, + { + "epoch": 8.774464119291705, + "grad_norm": 0.25894293189048767, + "learning_rate": 4.34234211462305e-05, + "loss": 0.0067, + "step": 56490 + }, + { + "epoch": 8.776017396707052, + "grad_norm": 0.2500094771385193, + "learning_rate": 4.340703046172213e-05, + "loss": 0.0071, + "step": 56500 + }, + { + "epoch": 8.777570674122398, + "grad_norm": 0.16239234805107117, + "learning_rate": 4.339064049821097e-05, + "loss": 0.0054, + "step": 56510 + }, + { + "epoch": 8.779123951537745, + "grad_norm": 0.3926495611667633, + "learning_rate": 4.3374251257489446e-05, + "loss": 0.0083, + "step": 56520 + }, + { + "epoch": 8.780677228953092, + "grad_norm": 0.3955729007720947, + "learning_rate": 4.335786274134984e-05, + "loss": 0.0092, + "step": 56530 + }, + { + "epoch": 8.782230506368437, + "grad_norm": 0.23745813965797424, + "learning_rate": 4.334147495158438e-05, + "loss": 0.0067, + "step": 56540 + }, + { + "epoch": 8.783783783783784, + "grad_norm": 0.2742697596549988, + "learning_rate": 4.33250878899852e-05, + "loss": 0.006, + "step": 56550 + }, + { + "epoch": 8.78533706119913, + "grad_norm": 0.09260518848896027, + "learning_rate": 4.330870155834437e-05, + "loss": 0.0061, + "step": 56560 + }, + { + "epoch": 8.786890338614477, + "grad_norm": 0.23947344720363617, + "learning_rate": 4.3292315958453884e-05, + "loss": 0.0047, + "step": 56570 + }, + { + "epoch": 8.788443616029824, + "grad_norm": 0.46439170837402344, + "learning_rate": 4.3275931092105646e-05, + "loss": 0.007, + "step": 56580 + }, + { + "epoch": 8.789996893445169, + "grad_norm": 0.2512642443180084, + "learning_rate": 4.325954696109147e-05, + "loss": 0.0063, + "step": 56590 + }, + { + "epoch": 8.791550170860516, + "grad_norm": 0.08968900889158249, + "learning_rate": 4.32431635672031e-05, + "loss": 0.0084, + "step": 56600 + }, + { + "epoch": 8.793103448275861, + "grad_norm": 0.2470143735408783, + "learning_rate": 4.3226780912232214e-05, + "loss": 0.007, + "step": 56610 + }, + { + "epoch": 8.794656725691208, + "grad_norm": 0.5473846197128296, + "learning_rate": 4.321039899797036e-05, + "loss": 0.0067, + "step": 56620 + }, + { + "epoch": 8.796210003106555, + "grad_norm": 0.4463874399662018, + "learning_rate": 4.319401782620908e-05, + "loss": 0.0073, + "step": 56630 + }, + { + "epoch": 8.7977632805219, + "grad_norm": 0.7577689290046692, + "learning_rate": 4.317763739873978e-05, + "loss": 0.0052, + "step": 56640 + }, + { + "epoch": 8.799316557937248, + "grad_norm": 0.20917940139770508, + "learning_rate": 4.31612577173538e-05, + "loss": 0.0045, + "step": 56650 + }, + { + "epoch": 8.800869835352595, + "grad_norm": 0.3934358060359955, + "learning_rate": 4.314487878384239e-05, + "loss": 0.0058, + "step": 56660 + }, + { + "epoch": 8.80242311276794, + "grad_norm": 0.1389254331588745, + "learning_rate": 4.3128500599996736e-05, + "loss": 0.0084, + "step": 56670 + }, + { + "epoch": 8.803976390183287, + "grad_norm": 0.08501488715410233, + "learning_rate": 4.31121231676079e-05, + "loss": 0.006, + "step": 56680 + }, + { + "epoch": 8.805529667598632, + "grad_norm": 0.19885671138763428, + "learning_rate": 4.309574648846694e-05, + "loss": 0.0048, + "step": 56690 + }, + { + "epoch": 8.80708294501398, + "grad_norm": 0.19477547705173492, + "learning_rate": 4.3079370564364755e-05, + "loss": 0.0056, + "step": 56700 + }, + { + "epoch": 8.808636222429326, + "grad_norm": 0.35058486461639404, + "learning_rate": 4.306299539709221e-05, + "loss": 0.0057, + "step": 56710 + }, + { + "epoch": 8.810189499844672, + "grad_norm": 0.3061440587043762, + "learning_rate": 4.3046620988440046e-05, + "loss": 0.0054, + "step": 56720 + }, + { + "epoch": 8.811742777260019, + "grad_norm": 0.20700165629386902, + "learning_rate": 4.303024734019897e-05, + "loss": 0.0069, + "step": 56730 + }, + { + "epoch": 8.813296054675366, + "grad_norm": 0.2337350994348526, + "learning_rate": 4.3013874454159565e-05, + "loss": 0.007, + "step": 56740 + }, + { + "epoch": 8.814849332090711, + "grad_norm": 0.19386787712574005, + "learning_rate": 4.299750233211233e-05, + "loss": 0.0062, + "step": 56750 + }, + { + "epoch": 8.816402609506058, + "grad_norm": 0.2655216455459595, + "learning_rate": 4.2981130975847715e-05, + "loss": 0.006, + "step": 56760 + }, + { + "epoch": 8.817955886921403, + "grad_norm": 0.45014914870262146, + "learning_rate": 4.2964760387156076e-05, + "loss": 0.0087, + "step": 56770 + }, + { + "epoch": 8.81950916433675, + "grad_norm": 0.1702646166086197, + "learning_rate": 4.294839056782767e-05, + "loss": 0.0056, + "step": 56780 + }, + { + "epoch": 8.821062441752098, + "grad_norm": 0.12641781568527222, + "learning_rate": 4.293202151965266e-05, + "loss": 0.0059, + "step": 56790 + }, + { + "epoch": 8.822615719167443, + "grad_norm": 0.20910555124282837, + "learning_rate": 4.291565324442115e-05, + "loss": 0.0035, + "step": 56800 + }, + { + "epoch": 8.82416899658279, + "grad_norm": 0.26418590545654297, + "learning_rate": 4.2899285743923135e-05, + "loss": 0.006, + "step": 56810 + }, + { + "epoch": 8.825722273998135, + "grad_norm": 0.23022300004959106, + "learning_rate": 4.288291901994857e-05, + "loss": 0.0046, + "step": 56820 + }, + { + "epoch": 8.827275551413482, + "grad_norm": 0.2857426404953003, + "learning_rate": 4.286655307428729e-05, + "loss": 0.0072, + "step": 56830 + }, + { + "epoch": 8.82882882882883, + "grad_norm": 0.9306734204292297, + "learning_rate": 4.285018790872903e-05, + "loss": 0.0071, + "step": 56840 + }, + { + "epoch": 8.830382106244175, + "grad_norm": 0.17217150330543518, + "learning_rate": 4.283382352506347e-05, + "loss": 0.006, + "step": 56850 + }, + { + "epoch": 8.831935383659522, + "grad_norm": 0.18008776009082794, + "learning_rate": 4.2817459925080205e-05, + "loss": 0.006, + "step": 56860 + }, + { + "epoch": 8.833488661074869, + "grad_norm": 0.07478263974189758, + "learning_rate": 4.28010971105687e-05, + "loss": 0.0045, + "step": 56870 + }, + { + "epoch": 8.835041938490214, + "grad_norm": 0.3804134428501129, + "learning_rate": 4.278473508331841e-05, + "loss": 0.007, + "step": 56880 + }, + { + "epoch": 8.836595215905561, + "grad_norm": 0.39959219098091125, + "learning_rate": 4.276837384511864e-05, + "loss": 0.0057, + "step": 56890 + }, + { + "epoch": 8.838148493320908, + "grad_norm": 1.1707518100738525, + "learning_rate": 4.275201339775864e-05, + "loss": 0.0076, + "step": 56900 + }, + { + "epoch": 8.839701770736253, + "grad_norm": 0.2789309322834015, + "learning_rate": 4.273565374302756e-05, + "loss": 0.008, + "step": 56910 + }, + { + "epoch": 8.8412550481516, + "grad_norm": 0.19044116139411926, + "learning_rate": 4.2719294882714454e-05, + "loss": 0.0061, + "step": 56920 + }, + { + "epoch": 8.842808325566946, + "grad_norm": 0.18053972721099854, + "learning_rate": 4.270293681860831e-05, + "loss": 0.0064, + "step": 56930 + }, + { + "epoch": 8.844361602982293, + "grad_norm": 0.19818121194839478, + "learning_rate": 4.268657955249802e-05, + "loss": 0.0084, + "step": 56940 + }, + { + "epoch": 8.84591488039764, + "grad_norm": 0.18267762660980225, + "learning_rate": 4.267022308617241e-05, + "loss": 0.006, + "step": 56950 + }, + { + "epoch": 8.847468157812985, + "grad_norm": 0.1324906051158905, + "learning_rate": 4.265386742142018e-05, + "loss": 0.0051, + "step": 56960 + }, + { + "epoch": 8.849021435228332, + "grad_norm": 0.3590368330478668, + "learning_rate": 4.263751256002996e-05, + "loss": 0.0071, + "step": 56970 + }, + { + "epoch": 8.850574712643677, + "grad_norm": 0.19497716426849365, + "learning_rate": 4.26211585037903e-05, + "loss": 0.005, + "step": 56980 + }, + { + "epoch": 8.852127990059024, + "grad_norm": 0.19232165813446045, + "learning_rate": 4.2604805254489656e-05, + "loss": 0.0067, + "step": 56990 + }, + { + "epoch": 8.853681267474371, + "grad_norm": 0.14667890965938568, + "learning_rate": 4.258845281391638e-05, + "loss": 0.0071, + "step": 57000 + }, + { + "epoch": 8.855234544889717, + "grad_norm": 0.24713169038295746, + "learning_rate": 4.257210118385876e-05, + "loss": 0.0081, + "step": 57010 + }, + { + "epoch": 8.856787822305064, + "grad_norm": 0.17021547257900238, + "learning_rate": 4.2555750366104994e-05, + "loss": 0.0066, + "step": 57020 + }, + { + "epoch": 8.85834109972041, + "grad_norm": 0.5225541591644287, + "learning_rate": 4.2539400362443184e-05, + "loss": 0.0068, + "step": 57030 + }, + { + "epoch": 8.859894377135756, + "grad_norm": 0.08993890881538391, + "learning_rate": 4.252305117466133e-05, + "loss": 0.0064, + "step": 57040 + }, + { + "epoch": 8.861447654551103, + "grad_norm": 0.192618265748024, + "learning_rate": 4.2506702804547356e-05, + "loss": 0.0054, + "step": 57050 + }, + { + "epoch": 8.863000931966448, + "grad_norm": 0.10888931155204773, + "learning_rate": 4.249035525388907e-05, + "loss": 0.0051, + "step": 57060 + }, + { + "epoch": 8.864554209381796, + "grad_norm": 0.19847451150417328, + "learning_rate": 4.247400852447427e-05, + "loss": 0.0048, + "step": 57070 + }, + { + "epoch": 8.866107486797143, + "grad_norm": 0.10742709040641785, + "learning_rate": 4.245766261809059e-05, + "loss": 0.007, + "step": 57080 + }, + { + "epoch": 8.867660764212488, + "grad_norm": 0.3639172613620758, + "learning_rate": 4.244131753652558e-05, + "loss": 0.0059, + "step": 57090 + }, + { + "epoch": 8.869214041627835, + "grad_norm": 0.723960280418396, + "learning_rate": 4.242497328156672e-05, + "loss": 0.0066, + "step": 57100 + }, + { + "epoch": 8.870767319043182, + "grad_norm": 0.22907331585884094, + "learning_rate": 4.240862985500139e-05, + "loss": 0.0058, + "step": 57110 + }, + { + "epoch": 8.872320596458527, + "grad_norm": 0.6142593622207642, + "learning_rate": 4.2392287258616894e-05, + "loss": 0.0063, + "step": 57120 + }, + { + "epoch": 8.873873873873874, + "grad_norm": 0.20245201885700226, + "learning_rate": 4.2375945494200396e-05, + "loss": 0.0081, + "step": 57130 + }, + { + "epoch": 8.87542715128922, + "grad_norm": 0.2818047106266022, + "learning_rate": 4.2359604563539054e-05, + "loss": 0.0055, + "step": 57140 + }, + { + "epoch": 8.876980428704567, + "grad_norm": 0.280661404132843, + "learning_rate": 4.2343264468419885e-05, + "loss": 0.0086, + "step": 57150 + }, + { + "epoch": 8.878533706119914, + "grad_norm": 0.2646837532520294, + "learning_rate": 4.232692521062979e-05, + "loss": 0.007, + "step": 57160 + }, + { + "epoch": 8.880086983535259, + "grad_norm": 0.4433560371398926, + "learning_rate": 4.231058679195562e-05, + "loss": 0.0066, + "step": 57170 + }, + { + "epoch": 8.881640260950606, + "grad_norm": 0.44385460019111633, + "learning_rate": 4.229424921418411e-05, + "loss": 0.0065, + "step": 57180 + }, + { + "epoch": 8.883193538365951, + "grad_norm": 0.26297080516815186, + "learning_rate": 4.2277912479101916e-05, + "loss": 0.0069, + "step": 57190 + }, + { + "epoch": 8.884746815781298, + "grad_norm": 0.23225703835487366, + "learning_rate": 4.226157658849562e-05, + "loss": 0.0084, + "step": 57200 + }, + { + "epoch": 8.886300093196645, + "grad_norm": 0.31693804264068604, + "learning_rate": 4.2245241544151664e-05, + "loss": 0.0079, + "step": 57210 + }, + { + "epoch": 8.88785337061199, + "grad_norm": 0.1711347997188568, + "learning_rate": 4.2228907347856425e-05, + "loss": 0.0059, + "step": 57220 + }, + { + "epoch": 8.889406648027338, + "grad_norm": 0.3888359069824219, + "learning_rate": 4.2212574001396214e-05, + "loss": 0.0053, + "step": 57230 + }, + { + "epoch": 8.890959925442685, + "grad_norm": 0.31779614090919495, + "learning_rate": 4.21962415065572e-05, + "loss": 0.0074, + "step": 57240 + }, + { + "epoch": 8.89251320285803, + "grad_norm": 0.3002488613128662, + "learning_rate": 4.217990986512548e-05, + "loss": 0.0059, + "step": 57250 + }, + { + "epoch": 8.894066480273377, + "grad_norm": 0.22813519835472107, + "learning_rate": 4.216357907888704e-05, + "loss": 0.0063, + "step": 57260 + }, + { + "epoch": 8.895619757688724, + "grad_norm": 0.15570443868637085, + "learning_rate": 4.2147249149627824e-05, + "loss": 0.0062, + "step": 57270 + }, + { + "epoch": 8.89717303510407, + "grad_norm": 0.21583199501037598, + "learning_rate": 4.213092007913364e-05, + "loss": 0.0068, + "step": 57280 + }, + { + "epoch": 8.898726312519416, + "grad_norm": 0.22561942040920258, + "learning_rate": 4.2114591869190205e-05, + "loss": 0.0045, + "step": 57290 + }, + { + "epoch": 8.900279589934762, + "grad_norm": 0.3107004165649414, + "learning_rate": 4.209826452158315e-05, + "loss": 0.0093, + "step": 57300 + }, + { + "epoch": 8.901832867350109, + "grad_norm": 0.16476000845432281, + "learning_rate": 4.2081938038098e-05, + "loss": 0.0055, + "step": 57310 + }, + { + "epoch": 8.903386144765456, + "grad_norm": 0.2188362330198288, + "learning_rate": 4.2065612420520185e-05, + "loss": 0.0065, + "step": 57320 + }, + { + "epoch": 8.904939422180801, + "grad_norm": 0.3151708245277405, + "learning_rate": 4.204928767063509e-05, + "loss": 0.0058, + "step": 57330 + }, + { + "epoch": 8.906492699596148, + "grad_norm": 0.534575879573822, + "learning_rate": 4.2032963790227944e-05, + "loss": 0.006, + "step": 57340 + }, + { + "epoch": 8.908045977011493, + "grad_norm": 0.2071797400712967, + "learning_rate": 4.2016640781083887e-05, + "loss": 0.0069, + "step": 57350 + }, + { + "epoch": 8.90959925442684, + "grad_norm": 0.4934163987636566, + "learning_rate": 4.2000318644988004e-05, + "loss": 0.007, + "step": 57360 + }, + { + "epoch": 8.911152531842188, + "grad_norm": 0.4255755841732025, + "learning_rate": 4.198399738372524e-05, + "loss": 0.0064, + "step": 57370 + }, + { + "epoch": 8.912705809257533, + "grad_norm": 0.1408354640007019, + "learning_rate": 4.196767699908045e-05, + "loss": 0.0061, + "step": 57380 + }, + { + "epoch": 8.91425908667288, + "grad_norm": 0.25395357608795166, + "learning_rate": 4.195135749283845e-05, + "loss": 0.0068, + "step": 57390 + }, + { + "epoch": 8.915812364088227, + "grad_norm": 0.35881614685058594, + "learning_rate": 4.1935038866783894e-05, + "loss": 0.0047, + "step": 57400 + }, + { + "epoch": 8.917365641503572, + "grad_norm": 0.26123133301734924, + "learning_rate": 4.1918721122701355e-05, + "loss": 0.0076, + "step": 57410 + }, + { + "epoch": 8.91891891891892, + "grad_norm": 0.2617660164833069, + "learning_rate": 4.190240426237533e-05, + "loss": 0.0058, + "step": 57420 + }, + { + "epoch": 8.920472196334265, + "grad_norm": 0.25788936018943787, + "learning_rate": 4.188608828759019e-05, + "loss": 0.0066, + "step": 57430 + }, + { + "epoch": 8.922025473749612, + "grad_norm": 0.3842267394065857, + "learning_rate": 4.1869773200130246e-05, + "loss": 0.0057, + "step": 57440 + }, + { + "epoch": 8.923578751164959, + "grad_norm": 0.28911271691322327, + "learning_rate": 4.185345900177967e-05, + "loss": 0.0046, + "step": 57450 + }, + { + "epoch": 8.925132028580304, + "grad_norm": 0.23289288580417633, + "learning_rate": 4.183714569432258e-05, + "loss": 0.0074, + "step": 57460 + }, + { + "epoch": 8.926685305995651, + "grad_norm": 0.48235607147216797, + "learning_rate": 4.182083327954295e-05, + "loss": 0.0075, + "step": 57470 + }, + { + "epoch": 8.928238583410998, + "grad_norm": 0.10921204090118408, + "learning_rate": 4.1804521759224714e-05, + "loss": 0.0055, + "step": 57480 + }, + { + "epoch": 8.929791860826343, + "grad_norm": 0.2596132457256317, + "learning_rate": 4.178821113515165e-05, + "loss": 0.0051, + "step": 57490 + }, + { + "epoch": 8.93134513824169, + "grad_norm": 0.4599904417991638, + "learning_rate": 4.1771901409107476e-05, + "loss": 0.0055, + "step": 57500 + }, + { + "epoch": 8.932898415657036, + "grad_norm": 0.15790221095085144, + "learning_rate": 4.175559258287577e-05, + "loss": 0.0092, + "step": 57510 + }, + { + "epoch": 8.934451693072383, + "grad_norm": 0.3994354009628296, + "learning_rate": 4.173928465824008e-05, + "loss": 0.0066, + "step": 57520 + }, + { + "epoch": 8.93600497048773, + "grad_norm": 0.13704654574394226, + "learning_rate": 4.17229776369838e-05, + "loss": 0.0061, + "step": 57530 + }, + { + "epoch": 8.937558247903075, + "grad_norm": 0.34605517983436584, + "learning_rate": 4.170667152089025e-05, + "loss": 0.0053, + "step": 57540 + }, + { + "epoch": 8.939111525318422, + "grad_norm": 0.7151920199394226, + "learning_rate": 4.1690366311742626e-05, + "loss": 0.007, + "step": 57550 + }, + { + "epoch": 8.940664802733767, + "grad_norm": 0.5261715054512024, + "learning_rate": 4.1674062011324055e-05, + "loss": 0.0081, + "step": 57560 + }, + { + "epoch": 8.942218080149114, + "grad_norm": 0.19125361740589142, + "learning_rate": 4.165775862141752e-05, + "loss": 0.0058, + "step": 57570 + }, + { + "epoch": 8.943771357564462, + "grad_norm": 0.31361672282218933, + "learning_rate": 4.164145614380598e-05, + "loss": 0.007, + "step": 57580 + }, + { + "epoch": 8.945324634979807, + "grad_norm": 0.14515502750873566, + "learning_rate": 4.162515458027223e-05, + "loss": 0.0062, + "step": 57590 + }, + { + "epoch": 8.946877912395154, + "grad_norm": 0.14303965866565704, + "learning_rate": 4.160885393259899e-05, + "loss": 0.0064, + "step": 57600 + }, + { + "epoch": 8.948431189810501, + "grad_norm": 0.36319443583488464, + "learning_rate": 4.159255420256886e-05, + "loss": 0.0051, + "step": 57610 + }, + { + "epoch": 8.949984467225846, + "grad_norm": 0.46804341673851013, + "learning_rate": 4.157625539196436e-05, + "loss": 0.0104, + "step": 57620 + }, + { + "epoch": 8.951537744641193, + "grad_norm": 0.23301951587200165, + "learning_rate": 4.155995750256791e-05, + "loss": 0.0055, + "step": 57630 + }, + { + "epoch": 8.953091022056539, + "grad_norm": 0.12308115512132645, + "learning_rate": 4.15436605361618e-05, + "loss": 0.0052, + "step": 57640 + }, + { + "epoch": 8.954644299471886, + "grad_norm": 1.1801834106445312, + "learning_rate": 4.152736449452827e-05, + "loss": 0.0055, + "step": 57650 + }, + { + "epoch": 8.956197576887233, + "grad_norm": 0.2028069645166397, + "learning_rate": 4.151106937944943e-05, + "loss": 0.0047, + "step": 57660 + }, + { + "epoch": 8.957750854302578, + "grad_norm": 0.16236189007759094, + "learning_rate": 4.149477519270727e-05, + "loss": 0.0051, + "step": 57670 + }, + { + "epoch": 8.959304131717925, + "grad_norm": 0.10609160363674164, + "learning_rate": 4.147848193608371e-05, + "loss": 0.0063, + "step": 57680 + }, + { + "epoch": 8.960857409133272, + "grad_norm": 0.3453963100910187, + "learning_rate": 4.146218961136056e-05, + "loss": 0.0082, + "step": 57690 + }, + { + "epoch": 8.962410686548617, + "grad_norm": 0.7911675572395325, + "learning_rate": 4.14458982203195e-05, + "loss": 0.007, + "step": 57700 + }, + { + "epoch": 8.963963963963964, + "grad_norm": 0.199348583817482, + "learning_rate": 4.1429607764742165e-05, + "loss": 0.0066, + "step": 57710 + }, + { + "epoch": 8.96551724137931, + "grad_norm": 0.29188597202301025, + "learning_rate": 4.141331824641003e-05, + "loss": 0.0063, + "step": 57720 + }, + { + "epoch": 8.967070518794657, + "grad_norm": 0.16968250274658203, + "learning_rate": 4.139702966710452e-05, + "loss": 0.0056, + "step": 57730 + }, + { + "epoch": 8.968623796210004, + "grad_norm": 0.17566338181495667, + "learning_rate": 4.13807420286069e-05, + "loss": 0.007, + "step": 57740 + }, + { + "epoch": 8.970177073625349, + "grad_norm": 0.17412279546260834, + "learning_rate": 4.136445533269837e-05, + "loss": 0.0103, + "step": 57750 + }, + { + "epoch": 8.971730351040696, + "grad_norm": 0.2154468446969986, + "learning_rate": 4.134816958116003e-05, + "loss": 0.0057, + "step": 57760 + }, + { + "epoch": 8.973283628456041, + "grad_norm": 0.3739718496799469, + "learning_rate": 4.133188477577283e-05, + "loss": 0.0076, + "step": 57770 + }, + { + "epoch": 8.974836905871388, + "grad_norm": 0.1853296160697937, + "learning_rate": 4.131560091831771e-05, + "loss": 0.0072, + "step": 57780 + }, + { + "epoch": 8.976390183286735, + "grad_norm": 0.1549319177865982, + "learning_rate": 4.1299318010575405e-05, + "loss": 0.005, + "step": 57790 + }, + { + "epoch": 8.97794346070208, + "grad_norm": 0.32785454392433167, + "learning_rate": 4.128303605432661e-05, + "loss": 0.0049, + "step": 57800 + }, + { + "epoch": 8.979496738117428, + "grad_norm": 0.11292579770088196, + "learning_rate": 4.126675505135188e-05, + "loss": 0.0085, + "step": 57810 + }, + { + "epoch": 8.981050015532775, + "grad_norm": 0.24792900681495667, + "learning_rate": 4.125047500343169e-05, + "loss": 0.0074, + "step": 57820 + }, + { + "epoch": 8.98260329294812, + "grad_norm": 0.3408590257167816, + "learning_rate": 4.1234195912346376e-05, + "loss": 0.008, + "step": 57830 + }, + { + "epoch": 8.984156570363467, + "grad_norm": 0.24270740151405334, + "learning_rate": 4.1217917779876235e-05, + "loss": 0.0048, + "step": 57840 + }, + { + "epoch": 8.985709847778814, + "grad_norm": 0.1763073205947876, + "learning_rate": 4.120164060780141e-05, + "loss": 0.0047, + "step": 57850 + }, + { + "epoch": 8.98726312519416, + "grad_norm": 0.6085400581359863, + "learning_rate": 4.1185364397901935e-05, + "loss": 0.0067, + "step": 57860 + }, + { + "epoch": 8.988816402609507, + "grad_norm": 0.2820707857608795, + "learning_rate": 4.116908915195776e-05, + "loss": 0.0048, + "step": 57870 + }, + { + "epoch": 8.990369680024852, + "grad_norm": 0.07247922569513321, + "learning_rate": 4.115281487174872e-05, + "loss": 0.0053, + "step": 57880 + }, + { + "epoch": 8.991922957440199, + "grad_norm": 0.11913277208805084, + "learning_rate": 4.113654155905452e-05, + "loss": 0.0066, + "step": 57890 + }, + { + "epoch": 8.993476234855546, + "grad_norm": 0.3068828284740448, + "learning_rate": 4.112026921565483e-05, + "loss": 0.0064, + "step": 57900 + }, + { + "epoch": 8.995029512270891, + "grad_norm": 0.11970120668411255, + "learning_rate": 4.1103997843329147e-05, + "loss": 0.0058, + "step": 57910 + }, + { + "epoch": 8.996582789686238, + "grad_norm": 0.2558520436286926, + "learning_rate": 4.108772744385689e-05, + "loss": 0.0073, + "step": 57920 + }, + { + "epoch": 8.998136067101584, + "grad_norm": 0.31973159313201904, + "learning_rate": 4.1071458019017346e-05, + "loss": 0.0093, + "step": 57930 + }, + { + "epoch": 8.99968934451693, + "grad_norm": 0.5280028581619263, + "learning_rate": 4.1055189570589734e-05, + "loss": 0.006, + "step": 57940 + }, + { + "epoch": 9.001242621932278, + "grad_norm": 0.20568020641803741, + "learning_rate": 4.103892210035315e-05, + "loss": 0.0071, + "step": 57950 + }, + { + "epoch": 9.002795899347623, + "grad_norm": 0.18752896785736084, + "learning_rate": 4.102265561008656e-05, + "loss": 0.0069, + "step": 57960 + }, + { + "epoch": 9.00434917676297, + "grad_norm": 0.28692662715911865, + "learning_rate": 4.100639010156885e-05, + "loss": 0.0072, + "step": 57970 + }, + { + "epoch": 9.005902454178317, + "grad_norm": 0.18581978976726532, + "learning_rate": 4.099012557657882e-05, + "loss": 0.0056, + "step": 57980 + }, + { + "epoch": 9.007455731593662, + "grad_norm": 0.1950923204421997, + "learning_rate": 4.09738620368951e-05, + "loss": 0.0062, + "step": 57990 + }, + { + "epoch": 9.00900900900901, + "grad_norm": 0.14156262576580048, + "learning_rate": 4.095759948429626e-05, + "loss": 0.0075, + "step": 58000 + }, + { + "epoch": 9.010562286424355, + "grad_norm": 0.7150994539260864, + "learning_rate": 4.094133792056075e-05, + "loss": 0.0056, + "step": 58010 + }, + { + "epoch": 9.012115563839702, + "grad_norm": 0.5688942074775696, + "learning_rate": 4.092507734746688e-05, + "loss": 0.0065, + "step": 58020 + }, + { + "epoch": 9.013668841255049, + "grad_norm": 0.27116620540618896, + "learning_rate": 4.090881776679293e-05, + "loss": 0.0053, + "step": 58030 + }, + { + "epoch": 9.015222118670394, + "grad_norm": 0.5636165738105774, + "learning_rate": 4.0892559180316994e-05, + "loss": 0.0057, + "step": 58040 + }, + { + "epoch": 9.016775396085741, + "grad_norm": 0.34808316826820374, + "learning_rate": 4.08763015898171e-05, + "loss": 0.006, + "step": 58050 + }, + { + "epoch": 9.018328673501088, + "grad_norm": 0.1433454006910324, + "learning_rate": 4.0860044997071145e-05, + "loss": 0.0064, + "step": 58060 + }, + { + "epoch": 9.019881950916433, + "grad_norm": 0.19009242951869965, + "learning_rate": 4.084378940385693e-05, + "loss": 0.0065, + "step": 58070 + }, + { + "epoch": 9.02143522833178, + "grad_norm": 0.3368106186389923, + "learning_rate": 4.082753481195211e-05, + "loss": 0.009, + "step": 58080 + }, + { + "epoch": 9.022988505747126, + "grad_norm": 0.2751627266407013, + "learning_rate": 4.0811281223134324e-05, + "loss": 0.0071, + "step": 58090 + }, + { + "epoch": 9.024541783162473, + "grad_norm": 0.1900171935558319, + "learning_rate": 4.0795028639181e-05, + "loss": 0.0057, + "step": 58100 + }, + { + "epoch": 9.02609506057782, + "grad_norm": 0.42914527654647827, + "learning_rate": 4.077877706186951e-05, + "loss": 0.0066, + "step": 58110 + }, + { + "epoch": 9.027648337993165, + "grad_norm": 0.24992026388645172, + "learning_rate": 4.076252649297709e-05, + "loss": 0.0063, + "step": 58120 + }, + { + "epoch": 9.029201615408512, + "grad_norm": 0.09784268587827682, + "learning_rate": 4.07462769342809e-05, + "loss": 0.0059, + "step": 58130 + }, + { + "epoch": 9.030754892823857, + "grad_norm": 0.2628598213195801, + "learning_rate": 4.073002838755793e-05, + "loss": 0.0049, + "step": 58140 + }, + { + "epoch": 9.032308170239205, + "grad_norm": 0.0821356326341629, + "learning_rate": 4.0713780854585115e-05, + "loss": 0.0051, + "step": 58150 + }, + { + "epoch": 9.033861447654552, + "grad_norm": 0.15578129887580872, + "learning_rate": 4.069753433713927e-05, + "loss": 0.0097, + "step": 58160 + }, + { + "epoch": 9.035414725069897, + "grad_norm": 0.15004973113536835, + "learning_rate": 4.068128883699709e-05, + "loss": 0.0039, + "step": 58170 + }, + { + "epoch": 9.036968002485244, + "grad_norm": 0.15580540895462036, + "learning_rate": 4.0665044355935155e-05, + "loss": 0.0055, + "step": 58180 + }, + { + "epoch": 9.038521279900591, + "grad_norm": 0.37097811698913574, + "learning_rate": 4.064880089572993e-05, + "loss": 0.0052, + "step": 58190 + }, + { + "epoch": 9.040074557315936, + "grad_norm": 0.5364235043525696, + "learning_rate": 4.063255845815779e-05, + "loss": 0.0074, + "step": 58200 + }, + { + "epoch": 9.041627834731283, + "grad_norm": 0.17478322982788086, + "learning_rate": 4.0616317044994945e-05, + "loss": 0.0058, + "step": 58210 + }, + { + "epoch": 9.043181112146629, + "grad_norm": 0.23577052354812622, + "learning_rate": 4.0600076658017585e-05, + "loss": 0.0055, + "step": 58220 + }, + { + "epoch": 9.044734389561976, + "grad_norm": 0.27852270007133484, + "learning_rate": 4.0583837299001706e-05, + "loss": 0.0061, + "step": 58230 + }, + { + "epoch": 9.046287666977323, + "grad_norm": 0.14753644168376923, + "learning_rate": 4.056759896972322e-05, + "loss": 0.0071, + "step": 58240 + }, + { + "epoch": 9.047840944392668, + "grad_norm": 0.19339071214199066, + "learning_rate": 4.055136167195793e-05, + "loss": 0.0046, + "step": 58250 + }, + { + "epoch": 9.049394221808015, + "grad_norm": 0.36714690923690796, + "learning_rate": 4.053512540748152e-05, + "loss": 0.0063, + "step": 58260 + }, + { + "epoch": 9.050947499223362, + "grad_norm": 0.30465298891067505, + "learning_rate": 4.051889017806955e-05, + "loss": 0.0049, + "step": 58270 + }, + { + "epoch": 9.052500776638707, + "grad_norm": 0.15577296912670135, + "learning_rate": 4.0502655985497516e-05, + "loss": 0.0054, + "step": 58280 + }, + { + "epoch": 9.054054054054054, + "grad_norm": 0.2123730182647705, + "learning_rate": 4.048642283154074e-05, + "loss": 0.0088, + "step": 58290 + }, + { + "epoch": 9.0556073314694, + "grad_norm": 0.3977429270744324, + "learning_rate": 4.0470190717974464e-05, + "loss": 0.0051, + "step": 58300 + }, + { + "epoch": 9.057160608884747, + "grad_norm": 0.3346470296382904, + "learning_rate": 4.045395964657381e-05, + "loss": 0.007, + "step": 58310 + }, + { + "epoch": 9.058713886300094, + "grad_norm": 0.3735160529613495, + "learning_rate": 4.043772961911376e-05, + "loss": 0.0069, + "step": 58320 + }, + { + "epoch": 9.060267163715439, + "grad_norm": 0.3704462945461273, + "learning_rate": 4.042150063736924e-05, + "loss": 0.0055, + "step": 58330 + }, + { + "epoch": 9.061820441130786, + "grad_norm": 0.16303551197052002, + "learning_rate": 4.0405272703114976e-05, + "loss": 0.0061, + "step": 58340 + }, + { + "epoch": 9.063373718546133, + "grad_norm": 0.2808712422847748, + "learning_rate": 4.038904581812569e-05, + "loss": 0.0057, + "step": 58350 + }, + { + "epoch": 9.064926995961478, + "grad_norm": 0.36390969157218933, + "learning_rate": 4.03728199841759e-05, + "loss": 0.0073, + "step": 58360 + }, + { + "epoch": 9.066480273376825, + "grad_norm": 0.2646663784980774, + "learning_rate": 4.0356595203040044e-05, + "loss": 0.007, + "step": 58370 + }, + { + "epoch": 9.06803355079217, + "grad_norm": 0.17063142359256744, + "learning_rate": 4.0340371476492426e-05, + "loss": 0.006, + "step": 58380 + }, + { + "epoch": 9.069586828207518, + "grad_norm": 0.3578571081161499, + "learning_rate": 4.032414880630727e-05, + "loss": 0.0051, + "step": 58390 + }, + { + "epoch": 9.071140105622865, + "grad_norm": 0.22463954985141754, + "learning_rate": 4.0307927194258646e-05, + "loss": 0.0057, + "step": 58400 + }, + { + "epoch": 9.07269338303821, + "grad_norm": 0.3034934401512146, + "learning_rate": 4.029170664212054e-05, + "loss": 0.0059, + "step": 58410 + }, + { + "epoch": 9.074246660453557, + "grad_norm": 0.24555641412734985, + "learning_rate": 4.0275487151666794e-05, + "loss": 0.0081, + "step": 58420 + }, + { + "epoch": 9.075799937868904, + "grad_norm": 0.2614927887916565, + "learning_rate": 4.025926872467116e-05, + "loss": 0.0067, + "step": 58430 + }, + { + "epoch": 9.07735321528425, + "grad_norm": 0.57460618019104, + "learning_rate": 4.024305136290725e-05, + "loss": 0.0063, + "step": 58440 + }, + { + "epoch": 9.078906492699597, + "grad_norm": 0.21106445789337158, + "learning_rate": 4.022683506814858e-05, + "loss": 0.0063, + "step": 58450 + }, + { + "epoch": 9.080459770114942, + "grad_norm": 0.24192005395889282, + "learning_rate": 4.021061984216853e-05, + "loss": 0.0058, + "step": 58460 + }, + { + "epoch": 9.082013047530289, + "grad_norm": 0.38069653511047363, + "learning_rate": 4.019440568674035e-05, + "loss": 0.0068, + "step": 58470 + }, + { + "epoch": 9.083566324945636, + "grad_norm": 0.1328102946281433, + "learning_rate": 4.017819260363724e-05, + "loss": 0.0043, + "step": 58480 + }, + { + "epoch": 9.085119602360981, + "grad_norm": 0.09910057485103607, + "learning_rate": 4.016198059463223e-05, + "loss": 0.0078, + "step": 58490 + }, + { + "epoch": 9.086672879776328, + "grad_norm": 0.10890311002731323, + "learning_rate": 4.0145769661498233e-05, + "loss": 0.0091, + "step": 58500 + }, + { + "epoch": 9.088226157191674, + "grad_norm": 0.4192606508731842, + "learning_rate": 4.012955980600804e-05, + "loss": 0.0075, + "step": 58510 + }, + { + "epoch": 9.08977943460702, + "grad_norm": 0.17907461524009705, + "learning_rate": 4.0113351029934355e-05, + "loss": 0.0066, + "step": 58520 + }, + { + "epoch": 9.091332712022368, + "grad_norm": 0.1118985116481781, + "learning_rate": 4.0097143335049714e-05, + "loss": 0.0053, + "step": 58530 + }, + { + "epoch": 9.092885989437713, + "grad_norm": 0.4614431858062744, + "learning_rate": 4.0080936723126606e-05, + "loss": 0.0058, + "step": 58540 + }, + { + "epoch": 9.09443926685306, + "grad_norm": 0.3249216377735138, + "learning_rate": 4.006473119593735e-05, + "loss": 0.0062, + "step": 58550 + }, + { + "epoch": 9.095992544268407, + "grad_norm": 0.21169596910476685, + "learning_rate": 4.004852675525415e-05, + "loss": 0.0072, + "step": 58560 + }, + { + "epoch": 9.097545821683752, + "grad_norm": 0.30312472581863403, + "learning_rate": 4.00323234028491e-05, + "loss": 0.0082, + "step": 58570 + }, + { + "epoch": 9.0990990990991, + "grad_norm": 0.18095503747463226, + "learning_rate": 4.0016121140494186e-05, + "loss": 0.0047, + "step": 58580 + }, + { + "epoch": 9.100652376514445, + "grad_norm": 0.2965857684612274, + "learning_rate": 3.999991996996122e-05, + "loss": 0.0076, + "step": 58590 + }, + { + "epoch": 9.102205653929792, + "grad_norm": 0.3353423774242401, + "learning_rate": 3.9983719893022e-05, + "loss": 0.0041, + "step": 58600 + }, + { + "epoch": 9.103758931345139, + "grad_norm": 0.3016718626022339, + "learning_rate": 3.996752091144812e-05, + "loss": 0.0074, + "step": 58610 + }, + { + "epoch": 9.105312208760484, + "grad_norm": 0.47722312808036804, + "learning_rate": 3.9951323027011055e-05, + "loss": 0.0061, + "step": 58620 + }, + { + "epoch": 9.106865486175831, + "grad_norm": 0.3575945794582367, + "learning_rate": 3.993512624148219e-05, + "loss": 0.0065, + "step": 58630 + }, + { + "epoch": 9.108418763591178, + "grad_norm": 0.7987822890281677, + "learning_rate": 3.99189305566328e-05, + "loss": 0.0074, + "step": 58640 + }, + { + "epoch": 9.109972041006523, + "grad_norm": 0.15317939221858978, + "learning_rate": 3.990273597423401e-05, + "loss": 0.006, + "step": 58650 + }, + { + "epoch": 9.11152531842187, + "grad_norm": 0.2506335973739624, + "learning_rate": 3.988654249605681e-05, + "loss": 0.0053, + "step": 58660 + }, + { + "epoch": 9.113078595837216, + "grad_norm": 0.26234281063079834, + "learning_rate": 3.987035012387213e-05, + "loss": 0.0043, + "step": 58670 + }, + { + "epoch": 9.114631873252563, + "grad_norm": 0.172758549451828, + "learning_rate": 3.9854158859450735e-05, + "loss": 0.0066, + "step": 58680 + }, + { + "epoch": 9.11618515066791, + "grad_norm": 0.1696472018957138, + "learning_rate": 3.983796870456328e-05, + "loss": 0.0092, + "step": 58690 + }, + { + "epoch": 9.117738428083255, + "grad_norm": 0.22047117352485657, + "learning_rate": 3.982177966098028e-05, + "loss": 0.0074, + "step": 58700 + }, + { + "epoch": 9.119291705498602, + "grad_norm": 0.20933914184570312, + "learning_rate": 3.980559173047216e-05, + "loss": 0.0082, + "step": 58710 + }, + { + "epoch": 9.12084498291395, + "grad_norm": 0.18006110191345215, + "learning_rate": 3.9789404914809176e-05, + "loss": 0.0039, + "step": 58720 + }, + { + "epoch": 9.122398260329295, + "grad_norm": 0.20167642831802368, + "learning_rate": 3.9773219215761544e-05, + "loss": 0.0044, + "step": 58730 + }, + { + "epoch": 9.123951537744642, + "grad_norm": 0.13628816604614258, + "learning_rate": 3.975703463509928e-05, + "loss": 0.0051, + "step": 58740 + }, + { + "epoch": 9.125504815159987, + "grad_norm": 0.4203019440174103, + "learning_rate": 3.974085117459231e-05, + "loss": 0.0071, + "step": 58750 + }, + { + "epoch": 9.127058092575334, + "grad_norm": 0.1713581383228302, + "learning_rate": 3.972466883601043e-05, + "loss": 0.0047, + "step": 58760 + }, + { + "epoch": 9.128611369990681, + "grad_norm": 0.4958839416503906, + "learning_rate": 3.9708487621123316e-05, + "loss": 0.0056, + "step": 58770 + }, + { + "epoch": 9.130164647406026, + "grad_norm": 0.22551028430461884, + "learning_rate": 3.969230753170051e-05, + "loss": 0.007, + "step": 58780 + }, + { + "epoch": 9.131717924821373, + "grad_norm": 0.32398754358291626, + "learning_rate": 3.967612856951146e-05, + "loss": 0.0055, + "step": 58790 + }, + { + "epoch": 9.133271202236719, + "grad_norm": 0.193272665143013, + "learning_rate": 3.9659950736325476e-05, + "loss": 0.0048, + "step": 58800 + }, + { + "epoch": 9.134824479652066, + "grad_norm": 0.32896921038627625, + "learning_rate": 3.964377403391174e-05, + "loss": 0.0053, + "step": 58810 + }, + { + "epoch": 9.136377757067413, + "grad_norm": 0.1335214078426361, + "learning_rate": 3.9627598464039303e-05, + "loss": 0.0063, + "step": 58820 + }, + { + "epoch": 9.137931034482758, + "grad_norm": 0.1869632601737976, + "learning_rate": 3.9611424028477096e-05, + "loss": 0.0051, + "step": 58830 + }, + { + "epoch": 9.139484311898105, + "grad_norm": 0.3253907263278961, + "learning_rate": 3.959525072899394e-05, + "loss": 0.0069, + "step": 58840 + }, + { + "epoch": 9.141037589313452, + "grad_norm": 0.21310020983219147, + "learning_rate": 3.9579078567358506e-05, + "loss": 0.0051, + "step": 58850 + }, + { + "epoch": 9.142590866728797, + "grad_norm": 0.4802098572254181, + "learning_rate": 3.956290754533939e-05, + "loss": 0.0049, + "step": 58860 + }, + { + "epoch": 9.144144144144144, + "grad_norm": 0.2129184454679489, + "learning_rate": 3.954673766470501e-05, + "loss": 0.0053, + "step": 58870 + }, + { + "epoch": 9.14569742155949, + "grad_norm": 0.20038974285125732, + "learning_rate": 3.9530568927223674e-05, + "loss": 0.005, + "step": 58880 + }, + { + "epoch": 9.147250698974837, + "grad_norm": 0.7697596549987793, + "learning_rate": 3.95144013346636e-05, + "loss": 0.0057, + "step": 58890 + }, + { + "epoch": 9.148803976390184, + "grad_norm": 0.14993880689144135, + "learning_rate": 3.949823488879282e-05, + "loss": 0.005, + "step": 58900 + }, + { + "epoch": 9.15035725380553, + "grad_norm": 0.2530815899372101, + "learning_rate": 3.9482069591379276e-05, + "loss": 0.0049, + "step": 58910 + }, + { + "epoch": 9.151910531220876, + "grad_norm": 0.33826714754104614, + "learning_rate": 3.9465905444190784e-05, + "loss": 0.0058, + "step": 58920 + }, + { + "epoch": 9.153463808636223, + "grad_norm": 0.26872071623802185, + "learning_rate": 3.944974244899505e-05, + "loss": 0.0082, + "step": 58930 + }, + { + "epoch": 9.155017086051568, + "grad_norm": 0.39338862895965576, + "learning_rate": 3.9433580607559616e-05, + "loss": 0.0066, + "step": 58940 + }, + { + "epoch": 9.156570363466916, + "grad_norm": 0.5337583422660828, + "learning_rate": 3.941741992165193e-05, + "loss": 0.0076, + "step": 58950 + }, + { + "epoch": 9.15812364088226, + "grad_norm": 0.44377845525741577, + "learning_rate": 3.940126039303928e-05, + "loss": 0.0039, + "step": 58960 + }, + { + "epoch": 9.159676918297608, + "grad_norm": 0.11951154470443726, + "learning_rate": 3.938510202348886e-05, + "loss": 0.0071, + "step": 58970 + }, + { + "epoch": 9.161230195712955, + "grad_norm": 0.24113929271697998, + "learning_rate": 3.93689448147677e-05, + "loss": 0.0063, + "step": 58980 + }, + { + "epoch": 9.1627834731283, + "grad_norm": 0.17480476200580597, + "learning_rate": 3.935278876864278e-05, + "loss": 0.0096, + "step": 58990 + }, + { + "epoch": 9.164336750543647, + "grad_norm": 0.3909773826599121, + "learning_rate": 3.933663388688087e-05, + "loss": 0.0066, + "step": 59000 + }, + { + "epoch": 9.165890027958994, + "grad_norm": 0.4046330153942108, + "learning_rate": 3.9320480171248645e-05, + "loss": 0.0064, + "step": 59010 + }, + { + "epoch": 9.16744330537434, + "grad_norm": 0.23264503479003906, + "learning_rate": 3.930432762351265e-05, + "loss": 0.0046, + "step": 59020 + }, + { + "epoch": 9.168996582789687, + "grad_norm": 0.27239713072776794, + "learning_rate": 3.928817624543931e-05, + "loss": 0.0072, + "step": 59030 + }, + { + "epoch": 9.170549860205032, + "grad_norm": 0.1891516000032425, + "learning_rate": 3.9272026038794885e-05, + "loss": 0.0073, + "step": 59040 + }, + { + "epoch": 9.172103137620379, + "grad_norm": 0.6002032160758972, + "learning_rate": 3.9255877005345586e-05, + "loss": 0.0062, + "step": 59050 + }, + { + "epoch": 9.173656415035726, + "grad_norm": 0.5490938425064087, + "learning_rate": 3.923972914685743e-05, + "loss": 0.0056, + "step": 59060 + }, + { + "epoch": 9.175209692451071, + "grad_norm": 0.13189288973808289, + "learning_rate": 3.922358246509631e-05, + "loss": 0.0047, + "step": 59070 + }, + { + "epoch": 9.176762969866418, + "grad_norm": 0.23480357229709625, + "learning_rate": 3.920743696182801e-05, + "loss": 0.0075, + "step": 59080 + }, + { + "epoch": 9.178316247281764, + "grad_norm": 0.13526515662670135, + "learning_rate": 3.9191292638818164e-05, + "loss": 0.0068, + "step": 59090 + }, + { + "epoch": 9.17986952469711, + "grad_norm": 0.1354861557483673, + "learning_rate": 3.9175149497832316e-05, + "loss": 0.0046, + "step": 59100 + }, + { + "epoch": 9.181422802112458, + "grad_norm": 0.24301743507385254, + "learning_rate": 3.915900754063584e-05, + "loss": 0.0042, + "step": 59110 + }, + { + "epoch": 9.182976079527803, + "grad_norm": 0.49956464767456055, + "learning_rate": 3.9142866768994014e-05, + "loss": 0.0048, + "step": 59120 + }, + { + "epoch": 9.18452935694315, + "grad_norm": 0.45724883675575256, + "learning_rate": 3.912672718467193e-05, + "loss": 0.0073, + "step": 59130 + }, + { + "epoch": 9.186082634358497, + "grad_norm": 0.8305158615112305, + "learning_rate": 3.911058878943463e-05, + "loss": 0.0044, + "step": 59140 + }, + { + "epoch": 9.187635911773842, + "grad_norm": 0.5242888927459717, + "learning_rate": 3.9094451585046974e-05, + "loss": 0.0065, + "step": 59150 + }, + { + "epoch": 9.18918918918919, + "grad_norm": 0.17590093612670898, + "learning_rate": 3.90783155732737e-05, + "loss": 0.0076, + "step": 59160 + }, + { + "epoch": 9.190742466604535, + "grad_norm": 0.15952394902706146, + "learning_rate": 3.9062180755879373e-05, + "loss": 0.0059, + "step": 59170 + }, + { + "epoch": 9.192295744019882, + "grad_norm": 0.1473831832408905, + "learning_rate": 3.904604713462855e-05, + "loss": 0.0046, + "step": 59180 + }, + { + "epoch": 9.193849021435229, + "grad_norm": 0.14607281982898712, + "learning_rate": 3.902991471128554e-05, + "loss": 0.0055, + "step": 59190 + }, + { + "epoch": 9.195402298850574, + "grad_norm": 0.21006909012794495, + "learning_rate": 3.901378348761456e-05, + "loss": 0.0049, + "step": 59200 + }, + { + "epoch": 9.196955576265921, + "grad_norm": 0.16535206139087677, + "learning_rate": 3.899765346537971e-05, + "loss": 0.0049, + "step": 59210 + }, + { + "epoch": 9.198508853681268, + "grad_norm": 0.4968283772468567, + "learning_rate": 3.898152464634493e-05, + "loss": 0.0058, + "step": 59220 + }, + { + "epoch": 9.200062131096614, + "grad_norm": 0.20896777510643005, + "learning_rate": 3.896539703227402e-05, + "loss": 0.0051, + "step": 59230 + }, + { + "epoch": 9.20161540851196, + "grad_norm": 0.4410974085330963, + "learning_rate": 3.894927062493073e-05, + "loss": 0.0058, + "step": 59240 + }, + { + "epoch": 9.203168685927306, + "grad_norm": 0.32966482639312744, + "learning_rate": 3.893314542607859e-05, + "loss": 0.0044, + "step": 59250 + }, + { + "epoch": 9.204721963342653, + "grad_norm": 0.20536114275455475, + "learning_rate": 3.891702143748103e-05, + "loss": 0.0052, + "step": 59260 + }, + { + "epoch": 9.206275240758, + "grad_norm": 0.44423621892929077, + "learning_rate": 3.890089866090133e-05, + "loss": 0.007, + "step": 59270 + }, + { + "epoch": 9.207828518173345, + "grad_norm": 0.35354936122894287, + "learning_rate": 3.888477709810267e-05, + "loss": 0.0069, + "step": 59280 + }, + { + "epoch": 9.209381795588692, + "grad_norm": 0.21170414984226227, + "learning_rate": 3.886865675084806e-05, + "loss": 0.0067, + "step": 59290 + }, + { + "epoch": 9.21093507300404, + "grad_norm": 0.3497121334075928, + "learning_rate": 3.885253762090043e-05, + "loss": 0.0069, + "step": 59300 + }, + { + "epoch": 9.212488350419385, + "grad_norm": 0.2348577231168747, + "learning_rate": 3.8836419710022534e-05, + "loss": 0.006, + "step": 59310 + }, + { + "epoch": 9.214041627834732, + "grad_norm": 0.15776661038398743, + "learning_rate": 3.882030301997698e-05, + "loss": 0.0033, + "step": 59320 + }, + { + "epoch": 9.215594905250077, + "grad_norm": 0.2620924115180969, + "learning_rate": 3.880418755252629e-05, + "loss": 0.006, + "step": 59330 + }, + { + "epoch": 9.217148182665424, + "grad_norm": 0.48614630103111267, + "learning_rate": 3.87880733094328e-05, + "loss": 0.0065, + "step": 59340 + }, + { + "epoch": 9.218701460080771, + "grad_norm": 0.21214351058006287, + "learning_rate": 3.877196029245877e-05, + "loss": 0.0061, + "step": 59350 + }, + { + "epoch": 9.220254737496116, + "grad_norm": 0.14682139456272125, + "learning_rate": 3.875584850336627e-05, + "loss": 0.0037, + "step": 59360 + }, + { + "epoch": 9.221808014911463, + "grad_norm": 0.1680261492729187, + "learning_rate": 3.873973794391728e-05, + "loss": 0.0056, + "step": 59370 + }, + { + "epoch": 9.22336129232681, + "grad_norm": 0.15865664184093475, + "learning_rate": 3.872362861587361e-05, + "loss": 0.0056, + "step": 59380 + }, + { + "epoch": 9.224914569742156, + "grad_norm": 0.24463409185409546, + "learning_rate": 3.870752052099698e-05, + "loss": 0.0069, + "step": 59390 + }, + { + "epoch": 9.226467847157503, + "grad_norm": 0.10801788419485092, + "learning_rate": 3.8691413661048916e-05, + "loss": 0.0048, + "step": 59400 + }, + { + "epoch": 9.228021124572848, + "grad_norm": 0.21743549406528473, + "learning_rate": 3.867530803779085e-05, + "loss": 0.0062, + "step": 59410 + }, + { + "epoch": 9.229574401988195, + "grad_norm": 0.19244712591171265, + "learning_rate": 3.8659203652984054e-05, + "loss": 0.0054, + "step": 59420 + }, + { + "epoch": 9.231127679403542, + "grad_norm": 0.20081454515457153, + "learning_rate": 3.864310050838972e-05, + "loss": 0.0052, + "step": 59430 + }, + { + "epoch": 9.232680956818887, + "grad_norm": 0.12633520364761353, + "learning_rate": 3.8626998605768835e-05, + "loss": 0.0049, + "step": 59440 + }, + { + "epoch": 9.234234234234235, + "grad_norm": 0.19643430411815643, + "learning_rate": 3.8610897946882294e-05, + "loss": 0.0057, + "step": 59450 + }, + { + "epoch": 9.23578751164958, + "grad_norm": 0.15412424504756927, + "learning_rate": 3.859479853349082e-05, + "loss": 0.007, + "step": 59460 + }, + { + "epoch": 9.237340789064927, + "grad_norm": 0.20659315586090088, + "learning_rate": 3.857870036735503e-05, + "loss": 0.0045, + "step": 59470 + }, + { + "epoch": 9.238894066480274, + "grad_norm": 0.21150140464305878, + "learning_rate": 3.856260345023538e-05, + "loss": 0.0046, + "step": 59480 + }, + { + "epoch": 9.24044734389562, + "grad_norm": 0.36859849095344543, + "learning_rate": 3.854650778389224e-05, + "loss": 0.0053, + "step": 59490 + }, + { + "epoch": 9.242000621310966, + "grad_norm": 0.5676582455635071, + "learning_rate": 3.8530413370085786e-05, + "loss": 0.0068, + "step": 59500 + }, + { + "epoch": 9.243553898726313, + "grad_norm": 0.2464291751384735, + "learning_rate": 3.851432021057608e-05, + "loss": 0.005, + "step": 59510 + }, + { + "epoch": 9.245107176141659, + "grad_norm": 0.16670097410678864, + "learning_rate": 3.849822830712306e-05, + "loss": 0.0064, + "step": 59520 + }, + { + "epoch": 9.246660453557006, + "grad_norm": 0.3657323122024536, + "learning_rate": 3.848213766148649e-05, + "loss": 0.0049, + "step": 59530 + }, + { + "epoch": 9.24821373097235, + "grad_norm": 0.1777397096157074, + "learning_rate": 3.846604827542602e-05, + "loss": 0.0054, + "step": 59540 + }, + { + "epoch": 9.249767008387698, + "grad_norm": 0.22620169818401337, + "learning_rate": 3.844996015070117e-05, + "loss": 0.0059, + "step": 59550 + }, + { + "epoch": 9.251320285803045, + "grad_norm": 0.17487825453281403, + "learning_rate": 3.843387328907132e-05, + "loss": 0.0055, + "step": 59560 + }, + { + "epoch": 9.25287356321839, + "grad_norm": 0.263471782207489, + "learning_rate": 3.8417787692295696e-05, + "loss": 0.0055, + "step": 59570 + }, + { + "epoch": 9.254426840633737, + "grad_norm": 0.15702854096889496, + "learning_rate": 3.84017033621334e-05, + "loss": 0.0062, + "step": 59580 + }, + { + "epoch": 9.255980118049084, + "grad_norm": 0.29906943440437317, + "learning_rate": 3.838562030034338e-05, + "loss": 0.0059, + "step": 59590 + }, + { + "epoch": 9.25753339546443, + "grad_norm": 0.23072534799575806, + "learning_rate": 3.836953850868447e-05, + "loss": 0.0071, + "step": 59600 + }, + { + "epoch": 9.259086672879777, + "grad_norm": 0.40087753534317017, + "learning_rate": 3.835345798891533e-05, + "loss": 0.0073, + "step": 59610 + }, + { + "epoch": 9.260639950295122, + "grad_norm": 0.19970078766345978, + "learning_rate": 3.833737874279452e-05, + "loss": 0.0086, + "step": 59620 + }, + { + "epoch": 9.262193227710469, + "grad_norm": 0.39652320742607117, + "learning_rate": 3.8321300772080433e-05, + "loss": 0.0075, + "step": 59630 + }, + { + "epoch": 9.263746505125816, + "grad_norm": 0.16404609382152557, + "learning_rate": 3.8305224078531345e-05, + "loss": 0.005, + "step": 59640 + }, + { + "epoch": 9.265299782541161, + "grad_norm": 0.21231554448604584, + "learning_rate": 3.828914866390537e-05, + "loss": 0.0068, + "step": 59650 + }, + { + "epoch": 9.266853059956508, + "grad_norm": 0.22088637948036194, + "learning_rate": 3.827307452996048e-05, + "loss": 0.0071, + "step": 59660 + }, + { + "epoch": 9.268406337371854, + "grad_norm": 0.2838064432144165, + "learning_rate": 3.825700167845454e-05, + "loss": 0.0057, + "step": 59670 + }, + { + "epoch": 9.2699596147872, + "grad_norm": 0.2138543277978897, + "learning_rate": 3.8240930111145214e-05, + "loss": 0.01, + "step": 59680 + }, + { + "epoch": 9.271512892202548, + "grad_norm": 0.19503498077392578, + "learning_rate": 3.8224859829790116e-05, + "loss": 0.0055, + "step": 59690 + }, + { + "epoch": 9.273066169617893, + "grad_norm": 0.5448906421661377, + "learning_rate": 3.820879083614664e-05, + "loss": 0.0068, + "step": 59700 + }, + { + "epoch": 9.27461944703324, + "grad_norm": 0.16513393819332123, + "learning_rate": 3.819272313197208e-05, + "loss": 0.0057, + "step": 59710 + }, + { + "epoch": 9.276172724448587, + "grad_norm": 0.1408064067363739, + "learning_rate": 3.817665671902355e-05, + "loss": 0.0048, + "step": 59720 + }, + { + "epoch": 9.277726001863932, + "grad_norm": 0.26024216413497925, + "learning_rate": 3.816059159905808e-05, + "loss": 0.0062, + "step": 59730 + }, + { + "epoch": 9.27927927927928, + "grad_norm": 0.29813650250434875, + "learning_rate": 3.8144527773832476e-05, + "loss": 0.0059, + "step": 59740 + }, + { + "epoch": 9.280832556694625, + "grad_norm": 0.17870460450649261, + "learning_rate": 3.812846524510352e-05, + "loss": 0.0065, + "step": 59750 + }, + { + "epoch": 9.282385834109972, + "grad_norm": 0.2751898467540741, + "learning_rate": 3.811240401462775e-05, + "loss": 0.0082, + "step": 59760 + }, + { + "epoch": 9.283939111525319, + "grad_norm": 0.3033530116081238, + "learning_rate": 3.809634408416162e-05, + "loss": 0.0047, + "step": 59770 + }, + { + "epoch": 9.285492388940664, + "grad_norm": 0.33602166175842285, + "learning_rate": 3.8080285455461394e-05, + "loss": 0.0057, + "step": 59780 + }, + { + "epoch": 9.287045666356011, + "grad_norm": 0.31093037128448486, + "learning_rate": 3.806422813028323e-05, + "loss": 0.0047, + "step": 59790 + }, + { + "epoch": 9.288598943771358, + "grad_norm": 0.16453754901885986, + "learning_rate": 3.8048172110383114e-05, + "loss": 0.0055, + "step": 59800 + }, + { + "epoch": 9.290152221186704, + "grad_norm": 0.1386854201555252, + "learning_rate": 3.8032117397516944e-05, + "loss": 0.0048, + "step": 59810 + }, + { + "epoch": 9.29170549860205, + "grad_norm": 0.10681294649839401, + "learning_rate": 3.801606399344042e-05, + "loss": 0.0049, + "step": 59820 + }, + { + "epoch": 9.293258776017396, + "grad_norm": 0.40088924765586853, + "learning_rate": 3.800001189990913e-05, + "loss": 0.0055, + "step": 59830 + }, + { + "epoch": 9.294812053432743, + "grad_norm": 0.33558568358421326, + "learning_rate": 3.798396111867847e-05, + "loss": 0.0071, + "step": 59840 + }, + { + "epoch": 9.29636533084809, + "grad_norm": 0.12393355369567871, + "learning_rate": 3.796791165150378e-05, + "loss": 0.0067, + "step": 59850 + }, + { + "epoch": 9.297918608263435, + "grad_norm": 0.1764371693134308, + "learning_rate": 3.795186350014018e-05, + "loss": 0.0057, + "step": 59860 + }, + { + "epoch": 9.299471885678782, + "grad_norm": 0.2099810540676117, + "learning_rate": 3.793581666634266e-05, + "loss": 0.0071, + "step": 59870 + }, + { + "epoch": 9.30102516309413, + "grad_norm": 0.30611079931259155, + "learning_rate": 3.791977115186609e-05, + "loss": 0.0064, + "step": 59880 + }, + { + "epoch": 9.302578440509475, + "grad_norm": 0.3638562858104706, + "learning_rate": 3.79037269584652e-05, + "loss": 0.0074, + "step": 59890 + }, + { + "epoch": 9.304131717924822, + "grad_norm": 1.0096733570098877, + "learning_rate": 3.7887684087894545e-05, + "loss": 0.0068, + "step": 59900 + }, + { + "epoch": 9.305684995340167, + "grad_norm": 0.23480679094791412, + "learning_rate": 3.787164254190854e-05, + "loss": 0.0053, + "step": 59910 + }, + { + "epoch": 9.307238272755514, + "grad_norm": 0.14904147386550903, + "learning_rate": 3.7855602322261473e-05, + "loss": 0.0038, + "step": 59920 + }, + { + "epoch": 9.308791550170861, + "grad_norm": 0.36297285556793213, + "learning_rate": 3.783956343070746e-05, + "loss": 0.0062, + "step": 59930 + }, + { + "epoch": 9.310344827586206, + "grad_norm": 0.6354019045829773, + "learning_rate": 3.7823525869000526e-05, + "loss": 0.0065, + "step": 59940 + }, + { + "epoch": 9.311898105001553, + "grad_norm": 0.29089415073394775, + "learning_rate": 3.78074896388945e-05, + "loss": 0.0068, + "step": 59950 + }, + { + "epoch": 9.3134513824169, + "grad_norm": 0.19902023673057556, + "learning_rate": 3.7791454742143075e-05, + "loss": 0.0053, + "step": 59960 + }, + { + "epoch": 9.315004659832246, + "grad_norm": 0.21629492938518524, + "learning_rate": 3.7775421180499795e-05, + "loss": 0.0066, + "step": 59970 + }, + { + "epoch": 9.316557937247593, + "grad_norm": 0.348696768283844, + "learning_rate": 3.7759388955718086e-05, + "loss": 0.0068, + "step": 59980 + }, + { + "epoch": 9.318111214662938, + "grad_norm": 0.12086677551269531, + "learning_rate": 3.774335806955117e-05, + "loss": 0.0038, + "step": 59990 + }, + { + "epoch": 9.319664492078285, + "grad_norm": 0.2219044715166092, + "learning_rate": 3.7727328523752215e-05, + "loss": 0.0046, + "step": 60000 + }, + { + "epoch": 9.321217769493632, + "grad_norm": 0.19239090383052826, + "learning_rate": 3.771130032007415e-05, + "loss": 0.0062, + "step": 60010 + }, + { + "epoch": 9.322771046908978, + "grad_norm": 0.32940009236335754, + "learning_rate": 3.769527346026982e-05, + "loss": 0.0055, + "step": 60020 + }, + { + "epoch": 9.324324324324325, + "grad_norm": 0.25241997838020325, + "learning_rate": 3.767924794609188e-05, + "loss": 0.0069, + "step": 60030 + }, + { + "epoch": 9.32587760173967, + "grad_norm": 0.2453344315290451, + "learning_rate": 3.7663223779292854e-05, + "loss": 0.0069, + "step": 60040 + }, + { + "epoch": 9.327430879155017, + "grad_norm": 0.20810720324516296, + "learning_rate": 3.764720096162512e-05, + "loss": 0.0073, + "step": 60050 + }, + { + "epoch": 9.328984156570364, + "grad_norm": 0.223336860537529, + "learning_rate": 3.763117949484092e-05, + "loss": 0.0087, + "step": 60060 + }, + { + "epoch": 9.33053743398571, + "grad_norm": 0.1446646898984909, + "learning_rate": 3.7615159380692344e-05, + "loss": 0.0093, + "step": 60070 + }, + { + "epoch": 9.332090711401056, + "grad_norm": 0.18440501391887665, + "learning_rate": 3.759914062093132e-05, + "loss": 0.007, + "step": 60080 + }, + { + "epoch": 9.333643988816403, + "grad_norm": 0.1809876263141632, + "learning_rate": 3.7583123217309615e-05, + "loss": 0.0067, + "step": 60090 + }, + { + "epoch": 9.335197266231749, + "grad_norm": 0.2565039098262787, + "learning_rate": 3.7567107171578904e-05, + "loss": 0.0046, + "step": 60100 + }, + { + "epoch": 9.336750543647096, + "grad_norm": 0.08837921172380447, + "learning_rate": 3.755109248549066e-05, + "loss": 0.0043, + "step": 60110 + }, + { + "epoch": 9.338303821062441, + "grad_norm": 0.11931512504816055, + "learning_rate": 3.75350791607962e-05, + "loss": 0.0082, + "step": 60120 + }, + { + "epoch": 9.339857098477788, + "grad_norm": 0.14617487788200378, + "learning_rate": 3.751906719924676e-05, + "loss": 0.0066, + "step": 60130 + }, + { + "epoch": 9.341410375893135, + "grad_norm": 0.25333520770072937, + "learning_rate": 3.750305660259337e-05, + "loss": 0.0073, + "step": 60140 + }, + { + "epoch": 9.34296365330848, + "grad_norm": 0.141086608171463, + "learning_rate": 3.748704737258693e-05, + "loss": 0.0046, + "step": 60150 + }, + { + "epoch": 9.344516930723827, + "grad_norm": 0.24732719361782074, + "learning_rate": 3.747103951097816e-05, + "loss": 0.0064, + "step": 60160 + }, + { + "epoch": 9.346070208139174, + "grad_norm": 0.21620140969753265, + "learning_rate": 3.745503301951768e-05, + "loss": 0.0041, + "step": 60170 + }, + { + "epoch": 9.34762348555452, + "grad_norm": 0.43309706449508667, + "learning_rate": 3.743902789995592e-05, + "loss": 0.0075, + "step": 60180 + }, + { + "epoch": 9.349176762969867, + "grad_norm": 0.09794929623603821, + "learning_rate": 3.7423024154043183e-05, + "loss": 0.0055, + "step": 60190 + }, + { + "epoch": 9.350730040385212, + "grad_norm": 0.2155638337135315, + "learning_rate": 3.740702178352963e-05, + "loss": 0.0074, + "step": 60200 + }, + { + "epoch": 9.352283317800559, + "grad_norm": 0.40843260288238525, + "learning_rate": 3.739102079016523e-05, + "loss": 0.0075, + "step": 60210 + }, + { + "epoch": 9.353836595215906, + "grad_norm": 0.25708869099617004, + "learning_rate": 3.7375021175699846e-05, + "loss": 0.005, + "step": 60220 + }, + { + "epoch": 9.355389872631251, + "grad_norm": 0.1626778542995453, + "learning_rate": 3.735902294188317e-05, + "loss": 0.0051, + "step": 60230 + }, + { + "epoch": 9.356943150046598, + "grad_norm": 0.3410174250602722, + "learning_rate": 3.734302609046473e-05, + "loss": 0.0063, + "step": 60240 + }, + { + "epoch": 9.358496427461946, + "grad_norm": 0.22225213050842285, + "learning_rate": 3.732703062319391e-05, + "loss": 0.0064, + "step": 60250 + }, + { + "epoch": 9.36004970487729, + "grad_norm": 0.16481083631515503, + "learning_rate": 3.731103654181999e-05, + "loss": 0.0076, + "step": 60260 + }, + { + "epoch": 9.361602982292638, + "grad_norm": 0.374654084444046, + "learning_rate": 3.729504384809203e-05, + "loss": 0.0068, + "step": 60270 + }, + { + "epoch": 9.363156259707983, + "grad_norm": 0.35513171553611755, + "learning_rate": 3.727905254375898e-05, + "loss": 0.0075, + "step": 60280 + }, + { + "epoch": 9.36470953712333, + "grad_norm": 0.19619536399841309, + "learning_rate": 3.7263062630569614e-05, + "loss": 0.006, + "step": 60290 + }, + { + "epoch": 9.366262814538677, + "grad_norm": 0.27952858805656433, + "learning_rate": 3.724707411027256e-05, + "loss": 0.0066, + "step": 60300 + }, + { + "epoch": 9.367816091954023, + "grad_norm": 0.23883825540542603, + "learning_rate": 3.723108698461631e-05, + "loss": 0.0066, + "step": 60310 + }, + { + "epoch": 9.36936936936937, + "grad_norm": 0.16593091189861298, + "learning_rate": 3.72151012553492e-05, + "loss": 0.0063, + "step": 60320 + }, + { + "epoch": 9.370922646784717, + "grad_norm": 0.23381973803043365, + "learning_rate": 3.719911692421939e-05, + "loss": 0.0048, + "step": 60330 + }, + { + "epoch": 9.372475924200062, + "grad_norm": 0.2950282692909241, + "learning_rate": 3.71831339929749e-05, + "loss": 0.0059, + "step": 60340 + }, + { + "epoch": 9.374029201615409, + "grad_norm": 0.25070077180862427, + "learning_rate": 3.716715246336362e-05, + "loss": 0.0059, + "step": 60350 + }, + { + "epoch": 9.375582479030754, + "grad_norm": 0.26806843280792236, + "learning_rate": 3.715117233713324e-05, + "loss": 0.0048, + "step": 60360 + }, + { + "epoch": 9.377135756446101, + "grad_norm": 0.2789529860019684, + "learning_rate": 3.713519361603135e-05, + "loss": 0.0057, + "step": 60370 + }, + { + "epoch": 9.378689033861448, + "grad_norm": 0.2466457188129425, + "learning_rate": 3.711921630180532e-05, + "loss": 0.0062, + "step": 60380 + }, + { + "epoch": 9.380242311276794, + "grad_norm": 0.15704083442687988, + "learning_rate": 3.710324039620245e-05, + "loss": 0.0065, + "step": 60390 + }, + { + "epoch": 9.38179558869214, + "grad_norm": 0.3434793949127197, + "learning_rate": 3.708726590096982e-05, + "loss": 0.0054, + "step": 60400 + }, + { + "epoch": 9.383348866107486, + "grad_norm": 0.4385152757167816, + "learning_rate": 3.707129281785437e-05, + "loss": 0.006, + "step": 60410 + }, + { + "epoch": 9.384902143522833, + "grad_norm": 0.28173455595970154, + "learning_rate": 3.705532114860291e-05, + "loss": 0.0061, + "step": 60420 + }, + { + "epoch": 9.38645542093818, + "grad_norm": 0.20295149087905884, + "learning_rate": 3.703935089496207e-05, + "loss": 0.0073, + "step": 60430 + }, + { + "epoch": 9.388008698353525, + "grad_norm": 0.3219221532344818, + "learning_rate": 3.702338205867829e-05, + "loss": 0.0039, + "step": 60440 + }, + { + "epoch": 9.389561975768872, + "grad_norm": 0.16673600673675537, + "learning_rate": 3.7007414641497976e-05, + "loss": 0.0059, + "step": 60450 + }, + { + "epoch": 9.39111525318422, + "grad_norm": 0.210929736495018, + "learning_rate": 3.699144864516726e-05, + "loss": 0.0082, + "step": 60460 + }, + { + "epoch": 9.392668530599565, + "grad_norm": 0.12274306267499924, + "learning_rate": 3.6975484071432153e-05, + "loss": 0.0057, + "step": 60470 + }, + { + "epoch": 9.394221808014912, + "grad_norm": 0.21718202531337738, + "learning_rate": 3.695952092203854e-05, + "loss": 0.0074, + "step": 60480 + }, + { + "epoch": 9.395775085430257, + "grad_norm": 0.3137831687927246, + "learning_rate": 3.69435591987321e-05, + "loss": 0.0056, + "step": 60490 + }, + { + "epoch": 9.397328362845604, + "grad_norm": 0.23771479725837708, + "learning_rate": 3.6927598903258374e-05, + "loss": 0.0053, + "step": 60500 + }, + { + "epoch": 9.398881640260951, + "grad_norm": 1.3928818702697754, + "learning_rate": 3.691164003736281e-05, + "loss": 0.0078, + "step": 60510 + }, + { + "epoch": 9.400434917676296, + "grad_norm": 0.290988564491272, + "learning_rate": 3.689568260279061e-05, + "loss": 0.0056, + "step": 60520 + }, + { + "epoch": 9.401988195091644, + "grad_norm": 0.24343256652355194, + "learning_rate": 3.687972660128686e-05, + "loss": 0.0056, + "step": 60530 + }, + { + "epoch": 9.40354147250699, + "grad_norm": 0.32398682832717896, + "learning_rate": 3.686377203459648e-05, + "loss": 0.0099, + "step": 60540 + }, + { + "epoch": 9.405094749922336, + "grad_norm": 0.47106438875198364, + "learning_rate": 3.6847818904464226e-05, + "loss": 0.0061, + "step": 60550 + }, + { + "epoch": 9.406648027337683, + "grad_norm": 0.10536875575780869, + "learning_rate": 3.683186721263474e-05, + "loss": 0.006, + "step": 60560 + }, + { + "epoch": 9.408201304753028, + "grad_norm": 0.4661661684513092, + "learning_rate": 3.6815916960852435e-05, + "loss": 0.0049, + "step": 60570 + }, + { + "epoch": 9.409754582168375, + "grad_norm": 0.770864725112915, + "learning_rate": 3.679996815086165e-05, + "loss": 0.0061, + "step": 60580 + }, + { + "epoch": 9.411307859583722, + "grad_norm": 0.11301708966493607, + "learning_rate": 3.678402078440649e-05, + "loss": 0.0064, + "step": 60590 + }, + { + "epoch": 9.412861136999068, + "grad_norm": 0.165828138589859, + "learning_rate": 3.676807486323096e-05, + "loss": 0.0064, + "step": 60600 + }, + { + "epoch": 9.414414414414415, + "grad_norm": 0.23441334068775177, + "learning_rate": 3.6752130389078864e-05, + "loss": 0.0069, + "step": 60610 + }, + { + "epoch": 9.41596769182976, + "grad_norm": 0.15348266065120697, + "learning_rate": 3.673618736369388e-05, + "loss": 0.0054, + "step": 60620 + }, + { + "epoch": 9.417520969245107, + "grad_norm": 0.18535393476486206, + "learning_rate": 3.6720245788819474e-05, + "loss": 0.0048, + "step": 60630 + }, + { + "epoch": 9.419074246660454, + "grad_norm": 0.16447077691555023, + "learning_rate": 3.6704305666199044e-05, + "loss": 0.0071, + "step": 60640 + }, + { + "epoch": 9.4206275240758, + "grad_norm": 0.16052569448947906, + "learning_rate": 3.668836699757576e-05, + "loss": 0.0057, + "step": 60650 + }, + { + "epoch": 9.422180801491146, + "grad_norm": 0.2559046149253845, + "learning_rate": 3.667242978469265e-05, + "loss": 0.0044, + "step": 60660 + }, + { + "epoch": 9.423734078906493, + "grad_norm": 1.0026572942733765, + "learning_rate": 3.6656494029292575e-05, + "loss": 0.0093, + "step": 60670 + }, + { + "epoch": 9.425287356321839, + "grad_norm": 0.18280933797359467, + "learning_rate": 3.664055973311825e-05, + "loss": 0.0051, + "step": 60680 + }, + { + "epoch": 9.426840633737186, + "grad_norm": 0.568939208984375, + "learning_rate": 3.662462689791221e-05, + "loss": 0.0093, + "step": 60690 + }, + { + "epoch": 9.428393911152531, + "grad_norm": 0.2912139892578125, + "learning_rate": 3.660869552541689e-05, + "loss": 0.0048, + "step": 60700 + }, + { + "epoch": 9.429947188567878, + "grad_norm": 0.43379226326942444, + "learning_rate": 3.659276561737448e-05, + "loss": 0.0071, + "step": 60710 + }, + { + "epoch": 9.431500465983225, + "grad_norm": 0.1659453958272934, + "learning_rate": 3.657683717552708e-05, + "loss": 0.0043, + "step": 60720 + }, + { + "epoch": 9.43305374339857, + "grad_norm": 0.3676670491695404, + "learning_rate": 3.6560910201616574e-05, + "loss": 0.0058, + "step": 60730 + }, + { + "epoch": 9.434607020813917, + "grad_norm": 0.29168206453323364, + "learning_rate": 3.654498469738472e-05, + "loss": 0.0056, + "step": 60740 + }, + { + "epoch": 9.436160298229264, + "grad_norm": 0.21062880754470825, + "learning_rate": 3.6529060664573126e-05, + "loss": 0.0069, + "step": 60750 + }, + { + "epoch": 9.43771357564461, + "grad_norm": 0.30596768856048584, + "learning_rate": 3.6513138104923176e-05, + "loss": 0.0065, + "step": 60760 + }, + { + "epoch": 9.439266853059957, + "grad_norm": 0.2158537060022354, + "learning_rate": 3.6497217020176176e-05, + "loss": 0.0064, + "step": 60770 + }, + { + "epoch": 9.440820130475302, + "grad_norm": 0.18099035322666168, + "learning_rate": 3.648129741207323e-05, + "loss": 0.0067, + "step": 60780 + }, + { + "epoch": 9.44237340789065, + "grad_norm": 0.14149758219718933, + "learning_rate": 3.646537928235527e-05, + "loss": 0.0066, + "step": 60790 + }, + { + "epoch": 9.443926685305996, + "grad_norm": 0.21584433317184448, + "learning_rate": 3.6449462632763075e-05, + "loss": 0.0058, + "step": 60800 + }, + { + "epoch": 9.445479962721341, + "grad_norm": 0.19466644525527954, + "learning_rate": 3.643354746503729e-05, + "loss": 0.008, + "step": 60810 + }, + { + "epoch": 9.447033240136689, + "grad_norm": 0.20962153375148773, + "learning_rate": 3.6417633780918345e-05, + "loss": 0.0087, + "step": 60820 + }, + { + "epoch": 9.448586517552036, + "grad_norm": 0.23696845769882202, + "learning_rate": 3.640172158214655e-05, + "loss": 0.0065, + "step": 60830 + }, + { + "epoch": 9.45013979496738, + "grad_norm": 0.10788781195878983, + "learning_rate": 3.6385810870462035e-05, + "loss": 0.0071, + "step": 60840 + }, + { + "epoch": 9.451693072382728, + "grad_norm": 0.2586604952812195, + "learning_rate": 3.636990164760479e-05, + "loss": 0.0043, + "step": 60850 + }, + { + "epoch": 9.453246349798073, + "grad_norm": 0.11213598400354385, + "learning_rate": 3.6353993915314596e-05, + "loss": 0.0046, + "step": 60860 + }, + { + "epoch": 9.45479962721342, + "grad_norm": 0.24669674038887024, + "learning_rate": 3.6338087675331124e-05, + "loss": 0.0058, + "step": 60870 + }, + { + "epoch": 9.456352904628767, + "grad_norm": 0.2271345555782318, + "learning_rate": 3.632218292939383e-05, + "loss": 0.004, + "step": 60880 + }, + { + "epoch": 9.457906182044113, + "grad_norm": 0.20114782452583313, + "learning_rate": 3.6306279679242033e-05, + "loss": 0.0057, + "step": 60890 + }, + { + "epoch": 9.45945945945946, + "grad_norm": 0.6563486456871033, + "learning_rate": 3.629037792661492e-05, + "loss": 0.0082, + "step": 60900 + }, + { + "epoch": 9.461012736874807, + "grad_norm": 0.1286192387342453, + "learning_rate": 3.627447767325147e-05, + "loss": 0.0045, + "step": 60910 + }, + { + "epoch": 9.462566014290152, + "grad_norm": 0.186347097158432, + "learning_rate": 3.62585789208905e-05, + "loss": 0.0047, + "step": 60920 + }, + { + "epoch": 9.464119291705499, + "grad_norm": 0.4219966530799866, + "learning_rate": 3.624268167127068e-05, + "loss": 0.0055, + "step": 60930 + }, + { + "epoch": 9.465672569120844, + "grad_norm": 0.2029234766960144, + "learning_rate": 3.622678592613051e-05, + "loss": 0.0056, + "step": 60940 + }, + { + "epoch": 9.467225846536191, + "grad_norm": 0.20935527980327606, + "learning_rate": 3.6210891687208304e-05, + "loss": 0.0071, + "step": 60950 + }, + { + "epoch": 9.468779123951538, + "grad_norm": 0.3240474760532379, + "learning_rate": 3.619499895624228e-05, + "loss": 0.0044, + "step": 60960 + }, + { + "epoch": 9.470332401366884, + "grad_norm": 0.2416113317012787, + "learning_rate": 3.61791077349704e-05, + "loss": 0.0036, + "step": 60970 + }, + { + "epoch": 9.47188567878223, + "grad_norm": 0.15636761486530304, + "learning_rate": 3.616321802513053e-05, + "loss": 0.0041, + "step": 60980 + }, + { + "epoch": 9.473438956197576, + "grad_norm": 0.13758035004138947, + "learning_rate": 3.6147329828460325e-05, + "loss": 0.0055, + "step": 60990 + }, + { + "epoch": 9.474992233612923, + "grad_norm": 0.2063911408185959, + "learning_rate": 3.6131443146697306e-05, + "loss": 0.0065, + "step": 61000 + }, + { + "epoch": 9.47654551102827, + "grad_norm": 0.34196004271507263, + "learning_rate": 3.611555798157879e-05, + "loss": 0.0053, + "step": 61010 + }, + { + "epoch": 9.478098788443615, + "grad_norm": 0.25084388256073, + "learning_rate": 3.6099674334842e-05, + "loss": 0.004, + "step": 61020 + }, + { + "epoch": 9.479652065858962, + "grad_norm": 0.17935435473918915, + "learning_rate": 3.6083792208223934e-05, + "loss": 0.0081, + "step": 61030 + }, + { + "epoch": 9.48120534327431, + "grad_norm": 0.17018897831439972, + "learning_rate": 3.606791160346142e-05, + "loss": 0.0069, + "step": 61040 + }, + { + "epoch": 9.482758620689655, + "grad_norm": 0.1905224770307541, + "learning_rate": 3.6052032522291134e-05, + "loss": 0.0078, + "step": 61050 + }, + { + "epoch": 9.484311898105002, + "grad_norm": 0.15623433887958527, + "learning_rate": 3.603615496644962e-05, + "loss": 0.0056, + "step": 61060 + }, + { + "epoch": 9.485865175520347, + "grad_norm": 0.17182354629039764, + "learning_rate": 3.60202789376732e-05, + "loss": 0.0064, + "step": 61070 + }, + { + "epoch": 9.487418452935694, + "grad_norm": 0.22580231726169586, + "learning_rate": 3.6004404437698036e-05, + "loss": 0.0074, + "step": 61080 + }, + { + "epoch": 9.488971730351041, + "grad_norm": 0.15619632601737976, + "learning_rate": 3.5988531468260184e-05, + "loss": 0.0062, + "step": 61090 + }, + { + "epoch": 9.490525007766387, + "grad_norm": 0.204032763838768, + "learning_rate": 3.597266003109546e-05, + "loss": 0.0046, + "step": 61100 + }, + { + "epoch": 9.492078285181734, + "grad_norm": 0.2199607938528061, + "learning_rate": 3.5956790127939557e-05, + "loss": 0.0081, + "step": 61110 + }, + { + "epoch": 9.49363156259708, + "grad_norm": 0.12711457908153534, + "learning_rate": 3.594092176052796e-05, + "loss": 0.0062, + "step": 61120 + }, + { + "epoch": 9.495184840012426, + "grad_norm": 0.4046298563480377, + "learning_rate": 3.592505493059603e-05, + "loss": 0.0041, + "step": 61130 + }, + { + "epoch": 9.496738117427773, + "grad_norm": 0.7692723870277405, + "learning_rate": 3.5909189639878905e-05, + "loss": 0.0058, + "step": 61140 + }, + { + "epoch": 9.498291394843118, + "grad_norm": 0.1002529114484787, + "learning_rate": 3.5893325890111626e-05, + "loss": 0.006, + "step": 61150 + }, + { + "epoch": 9.499844672258465, + "grad_norm": 0.21102212369441986, + "learning_rate": 3.5877463683029034e-05, + "loss": 0.0053, + "step": 61160 + }, + { + "epoch": 9.501397949673812, + "grad_norm": 0.3189040720462799, + "learning_rate": 3.586160302036578e-05, + "loss": 0.0058, + "step": 61170 + }, + { + "epoch": 9.502951227089158, + "grad_norm": 0.2780357301235199, + "learning_rate": 3.584574390385636e-05, + "loss": 0.0058, + "step": 61180 + }, + { + "epoch": 9.504504504504505, + "grad_norm": 0.3627569377422333, + "learning_rate": 3.582988633523511e-05, + "loss": 0.0067, + "step": 61190 + }, + { + "epoch": 9.50605778191985, + "grad_norm": 0.11697746813297272, + "learning_rate": 3.581403031623616e-05, + "loss": 0.0055, + "step": 61200 + }, + { + "epoch": 9.507611059335197, + "grad_norm": 0.4375632405281067, + "learning_rate": 3.579817584859354e-05, + "loss": 0.0056, + "step": 61210 + }, + { + "epoch": 9.509164336750544, + "grad_norm": 0.5854777693748474, + "learning_rate": 3.578232293404107e-05, + "loss": 0.0056, + "step": 61220 + }, + { + "epoch": 9.51071761416589, + "grad_norm": 0.10167963057756424, + "learning_rate": 3.5766471574312385e-05, + "loss": 0.0034, + "step": 61230 + }, + { + "epoch": 9.512270891581236, + "grad_norm": 0.15754342079162598, + "learning_rate": 3.5750621771140964e-05, + "loss": 0.0063, + "step": 61240 + }, + { + "epoch": 9.513824168996583, + "grad_norm": 0.1917184591293335, + "learning_rate": 3.5734773526260115e-05, + "loss": 0.0033, + "step": 61250 + }, + { + "epoch": 9.515377446411929, + "grad_norm": 0.20421941578388214, + "learning_rate": 3.5718926841402996e-05, + "loss": 0.0051, + "step": 61260 + }, + { + "epoch": 9.516930723827276, + "grad_norm": 0.433106392621994, + "learning_rate": 3.5703081718302555e-05, + "loss": 0.0052, + "step": 61270 + }, + { + "epoch": 9.518484001242623, + "grad_norm": 0.31076040863990784, + "learning_rate": 3.568723815869162e-05, + "loss": 0.0071, + "step": 61280 + }, + { + "epoch": 9.520037278657968, + "grad_norm": 0.31695157289505005, + "learning_rate": 3.5671396164302777e-05, + "loss": 0.0051, + "step": 61290 + }, + { + "epoch": 9.521590556073315, + "grad_norm": 0.24405471980571747, + "learning_rate": 3.565555573686853e-05, + "loss": 0.0054, + "step": 61300 + }, + { + "epoch": 9.52314383348866, + "grad_norm": 0.26999324560165405, + "learning_rate": 3.563971687812113e-05, + "loss": 0.0069, + "step": 61310 + }, + { + "epoch": 9.524697110904008, + "grad_norm": 0.13134148716926575, + "learning_rate": 3.5623879589792706e-05, + "loss": 0.0047, + "step": 61320 + }, + { + "epoch": 9.526250388319355, + "grad_norm": 0.2522430419921875, + "learning_rate": 3.5608043873615185e-05, + "loss": 0.0056, + "step": 61330 + }, + { + "epoch": 9.5278036657347, + "grad_norm": 0.1431443840265274, + "learning_rate": 3.559220973132035e-05, + "loss": 0.0038, + "step": 61340 + }, + { + "epoch": 9.529356943150047, + "grad_norm": 0.2421536147594452, + "learning_rate": 3.557637716463981e-05, + "loss": 0.0059, + "step": 61350 + }, + { + "epoch": 9.530910220565392, + "grad_norm": 0.32647472620010376, + "learning_rate": 3.556054617530497e-05, + "loss": 0.006, + "step": 61360 + }, + { + "epoch": 9.53246349798074, + "grad_norm": 0.1739676296710968, + "learning_rate": 3.5544716765047106e-05, + "loss": 0.0049, + "step": 61370 + }, + { + "epoch": 9.534016775396086, + "grad_norm": 0.2821456789970398, + "learning_rate": 3.5528888935597274e-05, + "loss": 0.0066, + "step": 61380 + }, + { + "epoch": 9.535570052811432, + "grad_norm": 0.42616790533065796, + "learning_rate": 3.551306268868636e-05, + "loss": 0.0051, + "step": 61390 + }, + { + "epoch": 9.537123330226779, + "grad_norm": 0.34798339009284973, + "learning_rate": 3.5497238026045166e-05, + "loss": 0.0061, + "step": 61400 + }, + { + "epoch": 9.538676607642126, + "grad_norm": 0.142867311835289, + "learning_rate": 3.548141494940421e-05, + "loss": 0.0065, + "step": 61410 + }, + { + "epoch": 9.540229885057471, + "grad_norm": 0.3173525631427765, + "learning_rate": 3.546559346049388e-05, + "loss": 0.0066, + "step": 61420 + }, + { + "epoch": 9.541783162472818, + "grad_norm": 0.2650151550769806, + "learning_rate": 3.544977356104441e-05, + "loss": 0.0053, + "step": 61430 + }, + { + "epoch": 9.543336439888163, + "grad_norm": 0.23999373614788055, + "learning_rate": 3.543395525278582e-05, + "loss": 0.0068, + "step": 61440 + }, + { + "epoch": 9.54488971730351, + "grad_norm": 0.10577655583620071, + "learning_rate": 3.5418138537447985e-05, + "loss": 0.006, + "step": 61450 + }, + { + "epoch": 9.546442994718857, + "grad_norm": 0.23116452991962433, + "learning_rate": 3.5402323416760574e-05, + "loss": 0.0081, + "step": 61460 + }, + { + "epoch": 9.547996272134203, + "grad_norm": 0.27436700463294983, + "learning_rate": 3.538650989245316e-05, + "loss": 0.0053, + "step": 61470 + }, + { + "epoch": 9.54954954954955, + "grad_norm": 0.4771043062210083, + "learning_rate": 3.537069796625504e-05, + "loss": 0.0058, + "step": 61480 + }, + { + "epoch": 9.551102826964897, + "grad_norm": 0.549292266368866, + "learning_rate": 3.53548876398954e-05, + "loss": 0.0056, + "step": 61490 + }, + { + "epoch": 9.552656104380242, + "grad_norm": 0.22006578743457794, + "learning_rate": 3.5339078915103214e-05, + "loss": 0.0041, + "step": 61500 + }, + { + "epoch": 9.554209381795589, + "grad_norm": 0.423580139875412, + "learning_rate": 3.532327179360733e-05, + "loss": 0.005, + "step": 61510 + }, + { + "epoch": 9.555762659210934, + "grad_norm": 0.24396921694278717, + "learning_rate": 3.530746627713636e-05, + "loss": 0.0063, + "step": 61520 + }, + { + "epoch": 9.557315936626281, + "grad_norm": 0.37055841088294983, + "learning_rate": 3.5291662367418805e-05, + "loss": 0.0044, + "step": 61530 + }, + { + "epoch": 9.558869214041628, + "grad_norm": 0.20976345241069794, + "learning_rate": 3.527586006618293e-05, + "loss": 0.0045, + "step": 61540 + }, + { + "epoch": 9.560422491456974, + "grad_norm": 0.10502319782972336, + "learning_rate": 3.526005937515687e-05, + "loss": 0.0065, + "step": 61550 + }, + { + "epoch": 9.56197576887232, + "grad_norm": 0.18493880331516266, + "learning_rate": 3.524426029606856e-05, + "loss": 0.0055, + "step": 61560 + }, + { + "epoch": 9.563529046287666, + "grad_norm": 0.362020343542099, + "learning_rate": 3.5228462830645746e-05, + "loss": 0.0069, + "step": 61570 + }, + { + "epoch": 9.565082323703013, + "grad_norm": 0.26544034481048584, + "learning_rate": 3.5212666980616036e-05, + "loss": 0.0068, + "step": 61580 + }, + { + "epoch": 9.56663560111836, + "grad_norm": 0.20050738751888275, + "learning_rate": 3.519687274770681e-05, + "loss": 0.0046, + "step": 61590 + }, + { + "epoch": 9.568188878533705, + "grad_norm": 0.16650445759296417, + "learning_rate": 3.518108013364535e-05, + "loss": 0.0048, + "step": 61600 + }, + { + "epoch": 9.569742155949053, + "grad_norm": 0.14674507081508636, + "learning_rate": 3.516528914015868e-05, + "loss": 0.0039, + "step": 61610 + }, + { + "epoch": 9.5712954333644, + "grad_norm": 0.29176533222198486, + "learning_rate": 3.5149499768973694e-05, + "loss": 0.0044, + "step": 61620 + }, + { + "epoch": 9.572848710779745, + "grad_norm": 0.1943262219429016, + "learning_rate": 3.513371202181708e-05, + "loss": 0.0061, + "step": 61630 + }, + { + "epoch": 9.574401988195092, + "grad_norm": 0.2051563411951065, + "learning_rate": 3.511792590041537e-05, + "loss": 0.0057, + "step": 61640 + }, + { + "epoch": 9.575955265610437, + "grad_norm": 0.265525221824646, + "learning_rate": 3.510214140649488e-05, + "loss": 0.0041, + "step": 61650 + }, + { + "epoch": 9.577508543025784, + "grad_norm": 0.4041307866573334, + "learning_rate": 3.508635854178184e-05, + "loss": 0.0052, + "step": 61660 + }, + { + "epoch": 9.579061820441131, + "grad_norm": 0.27201753854751587, + "learning_rate": 3.50705773080022e-05, + "loss": 0.0055, + "step": 61670 + }, + { + "epoch": 9.580615097856477, + "grad_norm": 0.19238290190696716, + "learning_rate": 3.505479770688178e-05, + "loss": 0.003, + "step": 61680 + }, + { + "epoch": 9.582168375271824, + "grad_norm": 0.1659332811832428, + "learning_rate": 3.503901974014622e-05, + "loss": 0.0077, + "step": 61690 + }, + { + "epoch": 9.58372165268717, + "grad_norm": 0.30396267771720886, + "learning_rate": 3.502324340952095e-05, + "loss": 0.0069, + "step": 61700 + }, + { + "epoch": 9.585274930102516, + "grad_norm": 0.14395874738693237, + "learning_rate": 3.500746871673126e-05, + "loss": 0.0052, + "step": 61710 + }, + { + "epoch": 9.586828207517863, + "grad_norm": 0.6128708720207214, + "learning_rate": 3.499169566350227e-05, + "loss": 0.0051, + "step": 61720 + }, + { + "epoch": 9.588381484933208, + "grad_norm": 0.5273091793060303, + "learning_rate": 3.4975924251558876e-05, + "loss": 0.0051, + "step": 61730 + }, + { + "epoch": 9.589934762348555, + "grad_norm": 0.2259529083967209, + "learning_rate": 3.496015448262582e-05, + "loss": 0.0049, + "step": 61740 + }, + { + "epoch": 9.591488039763902, + "grad_norm": 0.4460344612598419, + "learning_rate": 3.494438635842764e-05, + "loss": 0.0059, + "step": 61750 + }, + { + "epoch": 9.593041317179248, + "grad_norm": 0.30451157689094543, + "learning_rate": 3.492861988068874e-05, + "loss": 0.0047, + "step": 61760 + }, + { + "epoch": 9.594594594594595, + "grad_norm": 0.14920136332511902, + "learning_rate": 3.491285505113332e-05, + "loss": 0.0051, + "step": 61770 + }, + { + "epoch": 9.596147872009942, + "grad_norm": 0.10061533749103546, + "learning_rate": 3.4897091871485376e-05, + "loss": 0.004, + "step": 61780 + }, + { + "epoch": 9.597701149425287, + "grad_norm": 0.2887355387210846, + "learning_rate": 3.488133034346875e-05, + "loss": 0.0048, + "step": 61790 + }, + { + "epoch": 9.599254426840634, + "grad_norm": 0.20470139384269714, + "learning_rate": 3.486557046880713e-05, + "loss": 0.0064, + "step": 61800 + }, + { + "epoch": 9.60080770425598, + "grad_norm": 0.12610478699207306, + "learning_rate": 3.484981224922395e-05, + "loss": 0.0073, + "step": 61810 + }, + { + "epoch": 9.602360981671326, + "grad_norm": 0.3136866092681885, + "learning_rate": 3.4834055686442544e-05, + "loss": 0.0059, + "step": 61820 + }, + { + "epoch": 9.603914259086674, + "grad_norm": 0.15994936227798462, + "learning_rate": 3.4818300782186e-05, + "loss": 0.0053, + "step": 61830 + }, + { + "epoch": 9.605467536502019, + "grad_norm": 0.10434193909168243, + "learning_rate": 3.4802547538177236e-05, + "loss": 0.0055, + "step": 61840 + }, + { + "epoch": 9.607020813917366, + "grad_norm": 0.2637748122215271, + "learning_rate": 3.478679595613905e-05, + "loss": 0.007, + "step": 61850 + }, + { + "epoch": 9.608574091332713, + "grad_norm": 0.2813495099544525, + "learning_rate": 3.4771046037793975e-05, + "loss": 0.0062, + "step": 61860 + }, + { + "epoch": 9.610127368748058, + "grad_norm": 0.21509505808353424, + "learning_rate": 3.4755297784864416e-05, + "loss": 0.0082, + "step": 61870 + }, + { + "epoch": 9.611680646163405, + "grad_norm": 0.9313479065895081, + "learning_rate": 3.473955119907257e-05, + "loss": 0.0051, + "step": 61880 + }, + { + "epoch": 9.61323392357875, + "grad_norm": 0.5897467732429504, + "learning_rate": 3.472380628214046e-05, + "loss": 0.0075, + "step": 61890 + }, + { + "epoch": 9.614787200994098, + "grad_norm": 0.13728003203868866, + "learning_rate": 3.4708063035789906e-05, + "loss": 0.0052, + "step": 61900 + }, + { + "epoch": 9.616340478409445, + "grad_norm": 0.553530752658844, + "learning_rate": 3.4692321461742614e-05, + "loss": 0.0092, + "step": 61910 + }, + { + "epoch": 9.61789375582479, + "grad_norm": 0.40998944640159607, + "learning_rate": 3.4676581561720023e-05, + "loss": 0.0074, + "step": 61920 + }, + { + "epoch": 9.619447033240137, + "grad_norm": 0.2398047298192978, + "learning_rate": 3.466084333744344e-05, + "loss": 0.0041, + "step": 61930 + }, + { + "epoch": 9.621000310655482, + "grad_norm": 0.2833908200263977, + "learning_rate": 3.4645106790633964e-05, + "loss": 0.0037, + "step": 61940 + }, + { + "epoch": 9.62255358807083, + "grad_norm": 0.11002985388040543, + "learning_rate": 3.4629371923012526e-05, + "loss": 0.0055, + "step": 61950 + }, + { + "epoch": 9.624106865486176, + "grad_norm": 0.3455352485179901, + "learning_rate": 3.4613638736299856e-05, + "loss": 0.0054, + "step": 61960 + }, + { + "epoch": 9.625660142901522, + "grad_norm": 0.4132499694824219, + "learning_rate": 3.459790723221652e-05, + "loss": 0.0048, + "step": 61970 + }, + { + "epoch": 9.627213420316869, + "grad_norm": 0.4829326868057251, + "learning_rate": 3.45821774124829e-05, + "loss": 0.0072, + "step": 61980 + }, + { + "epoch": 9.628766697732216, + "grad_norm": 0.13648509979248047, + "learning_rate": 3.456644927881918e-05, + "loss": 0.007, + "step": 61990 + }, + { + "epoch": 9.630319975147561, + "grad_norm": 0.3619648516178131, + "learning_rate": 3.4550722832945354e-05, + "loss": 0.0068, + "step": 62000 + }, + { + "epoch": 9.631873252562908, + "grad_norm": 0.20700693130493164, + "learning_rate": 3.453499807658126e-05, + "loss": 0.0054, + "step": 62010 + }, + { + "epoch": 9.633426529978253, + "grad_norm": 0.23637183010578156, + "learning_rate": 3.451927501144653e-05, + "loss": 0.0043, + "step": 62020 + }, + { + "epoch": 9.6349798073936, + "grad_norm": 0.389933317899704, + "learning_rate": 3.450355363926059e-05, + "loss": 0.0053, + "step": 62030 + }, + { + "epoch": 9.636533084808947, + "grad_norm": 0.24668647348880768, + "learning_rate": 3.448783396174273e-05, + "loss": 0.0049, + "step": 62040 + }, + { + "epoch": 9.638086362224293, + "grad_norm": 0.25820428133010864, + "learning_rate": 3.4472115980612034e-05, + "loss": 0.0053, + "step": 62050 + }, + { + "epoch": 9.63963963963964, + "grad_norm": 0.20689880847930908, + "learning_rate": 3.4456399697587396e-05, + "loss": 0.0043, + "step": 62060 + }, + { + "epoch": 9.641192917054987, + "grad_norm": 0.3838321268558502, + "learning_rate": 3.4440685114387506e-05, + "loss": 0.0058, + "step": 62070 + }, + { + "epoch": 9.642746194470332, + "grad_norm": 0.2127143293619156, + "learning_rate": 3.442497223273091e-05, + "loss": 0.0052, + "step": 62080 + }, + { + "epoch": 9.64429947188568, + "grad_norm": 0.3746890425682068, + "learning_rate": 3.4409261054335926e-05, + "loss": 0.0097, + "step": 62090 + }, + { + "epoch": 9.645852749301024, + "grad_norm": 0.17593224346637726, + "learning_rate": 3.439355158092069e-05, + "loss": 0.0051, + "step": 62100 + }, + { + "epoch": 9.647406026716371, + "grad_norm": 0.10994379222393036, + "learning_rate": 3.437784381420322e-05, + "loss": 0.0068, + "step": 62110 + }, + { + "epoch": 9.648959304131719, + "grad_norm": 0.6738122701644897, + "learning_rate": 3.436213775590126e-05, + "loss": 0.0092, + "step": 62120 + }, + { + "epoch": 9.650512581547064, + "grad_norm": 0.1788473129272461, + "learning_rate": 3.4346433407732405e-05, + "loss": 0.0065, + "step": 62130 + }, + { + "epoch": 9.65206585896241, + "grad_norm": 0.10364004224538803, + "learning_rate": 3.433073077141405e-05, + "loss": 0.0053, + "step": 62140 + }, + { + "epoch": 9.653619136377756, + "grad_norm": 0.14882788062095642, + "learning_rate": 3.431502984866343e-05, + "loss": 0.006, + "step": 62150 + }, + { + "epoch": 9.655172413793103, + "grad_norm": 1.0218241214752197, + "learning_rate": 3.429933064119753e-05, + "loss": 0.0072, + "step": 62160 + }, + { + "epoch": 9.65672569120845, + "grad_norm": 0.26151055097579956, + "learning_rate": 3.4283633150733255e-05, + "loss": 0.0078, + "step": 62170 + }, + { + "epoch": 9.658278968623796, + "grad_norm": 0.35885289311408997, + "learning_rate": 3.4267937378987225e-05, + "loss": 0.0068, + "step": 62180 + }, + { + "epoch": 9.659832246039143, + "grad_norm": 0.20607779920101166, + "learning_rate": 3.425224332767592e-05, + "loss": 0.0061, + "step": 62190 + }, + { + "epoch": 9.66138552345449, + "grad_norm": 0.21490688621997833, + "learning_rate": 3.4236550998515595e-05, + "loss": 0.0052, + "step": 62200 + }, + { + "epoch": 9.662938800869835, + "grad_norm": 0.10499776154756546, + "learning_rate": 3.422086039322235e-05, + "loss": 0.0058, + "step": 62210 + }, + { + "epoch": 9.664492078285182, + "grad_norm": 0.11388076096773148, + "learning_rate": 3.420517151351209e-05, + "loss": 0.0055, + "step": 62220 + }, + { + "epoch": 9.666045355700529, + "grad_norm": 0.4018995761871338, + "learning_rate": 3.418948436110053e-05, + "loss": 0.0044, + "step": 62230 + }, + { + "epoch": 9.667598633115874, + "grad_norm": 0.3634881377220154, + "learning_rate": 3.417379893770319e-05, + "loss": 0.008, + "step": 62240 + }, + { + "epoch": 9.669151910531221, + "grad_norm": 0.8141986131668091, + "learning_rate": 3.41581152450354e-05, + "loss": 0.0061, + "step": 62250 + }, + { + "epoch": 9.670705187946567, + "grad_norm": 0.26999539136886597, + "learning_rate": 3.414243328481231e-05, + "loss": 0.0056, + "step": 62260 + }, + { + "epoch": 9.672258465361914, + "grad_norm": 0.797309935092926, + "learning_rate": 3.412675305874888e-05, + "loss": 0.0077, + "step": 62270 + }, + { + "epoch": 9.67381174277726, + "grad_norm": 0.2780054807662964, + "learning_rate": 3.4111074568559876e-05, + "loss": 0.0061, + "step": 62280 + }, + { + "epoch": 9.675365020192606, + "grad_norm": 0.13420440256595612, + "learning_rate": 3.4095397815959846e-05, + "loss": 0.0056, + "step": 62290 + }, + { + "epoch": 9.676918297607953, + "grad_norm": 0.1815621554851532, + "learning_rate": 3.407972280266322e-05, + "loss": 0.0061, + "step": 62300 + }, + { + "epoch": 9.678471575023298, + "grad_norm": 0.17345048487186432, + "learning_rate": 3.4064049530384166e-05, + "loss": 0.0072, + "step": 62310 + }, + { + "epoch": 9.680024852438645, + "grad_norm": 0.17373903095722198, + "learning_rate": 3.40483780008367e-05, + "loss": 0.0042, + "step": 62320 + }, + { + "epoch": 9.681578129853992, + "grad_norm": 0.15012910962104797, + "learning_rate": 3.4032708215734635e-05, + "loss": 0.0055, + "step": 62330 + }, + { + "epoch": 9.683131407269338, + "grad_norm": 0.14739836752414703, + "learning_rate": 3.4017040176791584e-05, + "loss": 0.0045, + "step": 62340 + }, + { + "epoch": 9.684684684684685, + "grad_norm": 0.13195385038852692, + "learning_rate": 3.4001373885720976e-05, + "loss": 0.0043, + "step": 62350 + }, + { + "epoch": 9.686237962100032, + "grad_norm": 0.24543964862823486, + "learning_rate": 3.398570934423608e-05, + "loss": 0.0083, + "step": 62360 + }, + { + "epoch": 9.687791239515377, + "grad_norm": 0.26534295082092285, + "learning_rate": 3.397004655404994e-05, + "loss": 0.0065, + "step": 62370 + }, + { + "epoch": 9.689344516930724, + "grad_norm": 0.33828818798065186, + "learning_rate": 3.395438551687541e-05, + "loss": 0.0069, + "step": 62380 + }, + { + "epoch": 9.69089779434607, + "grad_norm": 0.3938738703727722, + "learning_rate": 3.393872623442515e-05, + "loss": 0.006, + "step": 62390 + }, + { + "epoch": 9.692451071761417, + "grad_norm": 0.29892703890800476, + "learning_rate": 3.392306870841164e-05, + "loss": 0.0047, + "step": 62400 + }, + { + "epoch": 9.694004349176764, + "grad_norm": 0.7295840382575989, + "learning_rate": 3.390741294054715e-05, + "loss": 0.0063, + "step": 62410 + }, + { + "epoch": 9.695557626592109, + "grad_norm": 0.48743388056755066, + "learning_rate": 3.38917589325438e-05, + "loss": 0.0051, + "step": 62420 + }, + { + "epoch": 9.697110904007456, + "grad_norm": 0.2085208296775818, + "learning_rate": 3.387610668611349e-05, + "loss": 0.0067, + "step": 62430 + }, + { + "epoch": 9.698664181422803, + "grad_norm": 0.15548419952392578, + "learning_rate": 3.38604562029679e-05, + "loss": 0.0073, + "step": 62440 + }, + { + "epoch": 9.700217458838148, + "grad_norm": 0.3163524270057678, + "learning_rate": 3.384480748481855e-05, + "loss": 0.0047, + "step": 62450 + }, + { + "epoch": 9.701770736253495, + "grad_norm": 0.3003471791744232, + "learning_rate": 3.382916053337676e-05, + "loss": 0.0065, + "step": 62460 + }, + { + "epoch": 9.70332401366884, + "grad_norm": 0.2127360701560974, + "learning_rate": 3.381351535035366e-05, + "loss": 0.0036, + "step": 62470 + }, + { + "epoch": 9.704877291084188, + "grad_norm": 0.22521598637104034, + "learning_rate": 3.379787193746019e-05, + "loss": 0.0042, + "step": 62480 + }, + { + "epoch": 9.706430568499535, + "grad_norm": 0.23432978987693787, + "learning_rate": 3.378223029640709e-05, + "loss": 0.0035, + "step": 62490 + }, + { + "epoch": 9.70798384591488, + "grad_norm": 0.24968548119068146, + "learning_rate": 3.376659042890489e-05, + "loss": 0.0088, + "step": 62500 + }, + { + "epoch": 9.709537123330227, + "grad_norm": 0.2144177109003067, + "learning_rate": 3.375095233666397e-05, + "loss": 0.0046, + "step": 62510 + }, + { + "epoch": 9.711090400745572, + "grad_norm": 0.06442747265100479, + "learning_rate": 3.373531602139447e-05, + "loss": 0.0056, + "step": 62520 + }, + { + "epoch": 9.71264367816092, + "grad_norm": 0.39316433668136597, + "learning_rate": 3.371968148480635e-05, + "loss": 0.0057, + "step": 62530 + }, + { + "epoch": 9.714196955576266, + "grad_norm": 0.37051013112068176, + "learning_rate": 3.370404872860937e-05, + "loss": 0.0054, + "step": 62540 + }, + { + "epoch": 9.715750232991612, + "grad_norm": 0.31850576400756836, + "learning_rate": 3.3688417754513145e-05, + "loss": 0.0068, + "step": 62550 + }, + { + "epoch": 9.717303510406959, + "grad_norm": 0.20822834968566895, + "learning_rate": 3.367278856422703e-05, + "loss": 0.0045, + "step": 62560 + }, + { + "epoch": 9.718856787822306, + "grad_norm": 0.229004368185997, + "learning_rate": 3.365716115946021e-05, + "loss": 0.0053, + "step": 62570 + }, + { + "epoch": 9.720410065237651, + "grad_norm": 0.35547584295272827, + "learning_rate": 3.3641535541921666e-05, + "loss": 0.006, + "step": 62580 + }, + { + "epoch": 9.721963342652998, + "grad_norm": 0.2450065165758133, + "learning_rate": 3.36259117133202e-05, + "loss": 0.0037, + "step": 62590 + }, + { + "epoch": 9.723516620068343, + "grad_norm": 0.12263771891593933, + "learning_rate": 3.36102896753644e-05, + "loss": 0.0055, + "step": 62600 + }, + { + "epoch": 9.72506989748369, + "grad_norm": 0.19631096720695496, + "learning_rate": 3.35946694297627e-05, + "loss": 0.0049, + "step": 62610 + }, + { + "epoch": 9.726623174899037, + "grad_norm": 0.24166399240493774, + "learning_rate": 3.357905097822328e-05, + "loss": 0.0056, + "step": 62620 + }, + { + "epoch": 9.728176452314383, + "grad_norm": 0.2207004725933075, + "learning_rate": 3.356343432245416e-05, + "loss": 0.0056, + "step": 62630 + }, + { + "epoch": 9.72972972972973, + "grad_norm": 0.26342636346817017, + "learning_rate": 3.354781946416315e-05, + "loss": 0.0056, + "step": 62640 + }, + { + "epoch": 9.731283007145077, + "grad_norm": 0.19617900252342224, + "learning_rate": 3.3532206405057864e-05, + "loss": 0.0054, + "step": 62650 + }, + { + "epoch": 9.732836284560422, + "grad_norm": 0.14871826767921448, + "learning_rate": 3.351659514684573e-05, + "loss": 0.0099, + "step": 62660 + }, + { + "epoch": 9.73438956197577, + "grad_norm": 0.2520610988140106, + "learning_rate": 3.350098569123395e-05, + "loss": 0.0076, + "step": 62670 + }, + { + "epoch": 9.735942839391114, + "grad_norm": 0.28075554966926575, + "learning_rate": 3.348537803992958e-05, + "loss": 0.0064, + "step": 62680 + }, + { + "epoch": 9.737496116806462, + "grad_norm": 0.1685597002506256, + "learning_rate": 3.346977219463945e-05, + "loss": 0.0052, + "step": 62690 + }, + { + "epoch": 9.739049394221809, + "grad_norm": 0.24690359830856323, + "learning_rate": 3.345416815707017e-05, + "loss": 0.0067, + "step": 62700 + }, + { + "epoch": 9.740602671637154, + "grad_norm": 0.1559983193874359, + "learning_rate": 3.3438565928928185e-05, + "loss": 0.0047, + "step": 62710 + }, + { + "epoch": 9.742155949052501, + "grad_norm": 0.2671487033367157, + "learning_rate": 3.3422965511919736e-05, + "loss": 0.0056, + "step": 62720 + }, + { + "epoch": 9.743709226467848, + "grad_norm": 0.18557602167129517, + "learning_rate": 3.3407366907750846e-05, + "loss": 0.005, + "step": 62730 + }, + { + "epoch": 9.745262503883193, + "grad_norm": 0.22815559804439545, + "learning_rate": 3.339177011812737e-05, + "loss": 0.0053, + "step": 62740 + }, + { + "epoch": 9.74681578129854, + "grad_norm": 0.1916397362947464, + "learning_rate": 3.3376175144754935e-05, + "loss": 0.0061, + "step": 62750 + }, + { + "epoch": 9.748369058713886, + "grad_norm": 0.3383394479751587, + "learning_rate": 3.3360581989339004e-05, + "loss": 0.0061, + "step": 62760 + }, + { + "epoch": 9.749922336129233, + "grad_norm": 0.22458451986312866, + "learning_rate": 3.3344990653584804e-05, + "loss": 0.0062, + "step": 62770 + }, + { + "epoch": 9.75147561354458, + "grad_norm": 0.20144134759902954, + "learning_rate": 3.332940113919739e-05, + "loss": 0.0066, + "step": 62780 + }, + { + "epoch": 9.753028890959925, + "grad_norm": 0.8697532415390015, + "learning_rate": 3.331381344788161e-05, + "loss": 0.0061, + "step": 62790 + }, + { + "epoch": 9.754582168375272, + "grad_norm": 0.1958247274160385, + "learning_rate": 3.3298227581342076e-05, + "loss": 0.0054, + "step": 62800 + }, + { + "epoch": 9.756135445790619, + "grad_norm": 0.5159515738487244, + "learning_rate": 3.328264354128329e-05, + "loss": 0.006, + "step": 62810 + }, + { + "epoch": 9.757688723205964, + "grad_norm": 0.2489512711763382, + "learning_rate": 3.3267061329409465e-05, + "loss": 0.0063, + "step": 62820 + }, + { + "epoch": 9.759242000621311, + "grad_norm": 0.11764433979988098, + "learning_rate": 3.325148094742466e-05, + "loss": 0.0065, + "step": 62830 + }, + { + "epoch": 9.760795278036657, + "grad_norm": 0.2032914012670517, + "learning_rate": 3.323590239703271e-05, + "loss": 0.0056, + "step": 62840 + }, + { + "epoch": 9.762348555452004, + "grad_norm": 0.1913830041885376, + "learning_rate": 3.322032567993728e-05, + "loss": 0.0067, + "step": 62850 + }, + { + "epoch": 9.76390183286735, + "grad_norm": 0.29776161909103394, + "learning_rate": 3.3204750797841774e-05, + "loss": 0.0071, + "step": 62860 + }, + { + "epoch": 9.765455110282696, + "grad_norm": 0.2468370646238327, + "learning_rate": 3.31891777524495e-05, + "loss": 0.0056, + "step": 62870 + }, + { + "epoch": 9.767008387698043, + "grad_norm": 0.33161234855651855, + "learning_rate": 3.317360654546347e-05, + "loss": 0.0063, + "step": 62880 + }, + { + "epoch": 9.768561665113388, + "grad_norm": 0.13419964909553528, + "learning_rate": 3.3158037178586534e-05, + "loss": 0.0051, + "step": 62890 + }, + { + "epoch": 9.770114942528735, + "grad_norm": 0.4022749364376068, + "learning_rate": 3.3142469653521316e-05, + "loss": 0.0051, + "step": 62900 + }, + { + "epoch": 9.771668219944083, + "grad_norm": 0.19610999524593353, + "learning_rate": 3.3126903971970284e-05, + "loss": 0.0045, + "step": 62910 + }, + { + "epoch": 9.773221497359428, + "grad_norm": 0.40106257796287537, + "learning_rate": 3.311134013563564e-05, + "loss": 0.0056, + "step": 62920 + }, + { + "epoch": 9.774774774774775, + "grad_norm": 0.30576255917549133, + "learning_rate": 3.309577814621947e-05, + "loss": 0.0065, + "step": 62930 + }, + { + "epoch": 9.776328052190122, + "grad_norm": 0.2887972593307495, + "learning_rate": 3.308021800542359e-05, + "loss": 0.0053, + "step": 62940 + }, + { + "epoch": 9.777881329605467, + "grad_norm": 0.26482632756233215, + "learning_rate": 3.306465971494963e-05, + "loss": 0.0057, + "step": 62950 + }, + { + "epoch": 9.779434607020814, + "grad_norm": 0.2669546902179718, + "learning_rate": 3.304910327649901e-05, + "loss": 0.0043, + "step": 62960 + }, + { + "epoch": 9.78098788443616, + "grad_norm": 0.2722795605659485, + "learning_rate": 3.303354869177297e-05, + "loss": 0.0075, + "step": 62970 + }, + { + "epoch": 9.782541161851507, + "grad_norm": 0.31254732608795166, + "learning_rate": 3.301799596247255e-05, + "loss": 0.0063, + "step": 62980 + }, + { + "epoch": 9.784094439266854, + "grad_norm": 0.3725385069847107, + "learning_rate": 3.300244509029854e-05, + "loss": 0.0063, + "step": 62990 + }, + { + "epoch": 9.785647716682199, + "grad_norm": 0.2334570437669754, + "learning_rate": 3.2986896076951585e-05, + "loss": 0.0044, + "step": 63000 + }, + { + "epoch": 9.787200994097546, + "grad_norm": 0.10947565734386444, + "learning_rate": 3.297134892413211e-05, + "loss": 0.0046, + "step": 63010 + }, + { + "epoch": 9.788754271512893, + "grad_norm": 0.3787364661693573, + "learning_rate": 3.295580363354031e-05, + "loss": 0.0067, + "step": 63020 + }, + { + "epoch": 9.790307548928238, + "grad_norm": 0.15469719469547272, + "learning_rate": 3.2940260206876194e-05, + "loss": 0.0033, + "step": 63030 + }, + { + "epoch": 9.791860826343585, + "grad_norm": 0.21056126058101654, + "learning_rate": 3.292471864583958e-05, + "loss": 0.0044, + "step": 63040 + }, + { + "epoch": 9.79341410375893, + "grad_norm": 0.2184109389781952, + "learning_rate": 3.290917895213004e-05, + "loss": 0.0067, + "step": 63050 + }, + { + "epoch": 9.794967381174278, + "grad_norm": 0.21159781515598297, + "learning_rate": 3.289364112744701e-05, + "loss": 0.0044, + "step": 63060 + }, + { + "epoch": 9.796520658589625, + "grad_norm": 0.15833736956119537, + "learning_rate": 3.287810517348967e-05, + "loss": 0.0047, + "step": 63070 + }, + { + "epoch": 9.79807393600497, + "grad_norm": 0.1895364224910736, + "learning_rate": 3.2862571091957014e-05, + "loss": 0.0059, + "step": 63080 + }, + { + "epoch": 9.799627213420317, + "grad_norm": 0.2894461452960968, + "learning_rate": 3.284703888454781e-05, + "loss": 0.0074, + "step": 63090 + }, + { + "epoch": 9.801180490835662, + "grad_norm": 0.12835106253623962, + "learning_rate": 3.2831508552960646e-05, + "loss": 0.0064, + "step": 63100 + }, + { + "epoch": 9.80273376825101, + "grad_norm": 0.35981041193008423, + "learning_rate": 3.281598009889387e-05, + "loss": 0.0051, + "step": 63110 + }, + { + "epoch": 9.804287045666356, + "grad_norm": 0.1597299575805664, + "learning_rate": 3.28004535240457e-05, + "loss": 0.0039, + "step": 63120 + }, + { + "epoch": 9.805840323081702, + "grad_norm": 0.21904078125953674, + "learning_rate": 3.278492883011407e-05, + "loss": 0.0053, + "step": 63130 + }, + { + "epoch": 9.807393600497049, + "grad_norm": 0.22310051321983337, + "learning_rate": 3.276940601879675e-05, + "loss": 0.0045, + "step": 63140 + }, + { + "epoch": 9.808946877912396, + "grad_norm": 0.19639521837234497, + "learning_rate": 3.275388509179127e-05, + "loss": 0.0054, + "step": 63150 + }, + { + "epoch": 9.810500155327741, + "grad_norm": 0.27780860662460327, + "learning_rate": 3.2738366050794985e-05, + "loss": 0.0059, + "step": 63160 + }, + { + "epoch": 9.812053432743088, + "grad_norm": 0.24074628949165344, + "learning_rate": 3.2722848897505034e-05, + "loss": 0.0065, + "step": 63170 + }, + { + "epoch": 9.813606710158435, + "grad_norm": 0.342912882566452, + "learning_rate": 3.2707333633618356e-05, + "loss": 0.0078, + "step": 63180 + }, + { + "epoch": 9.81515998757378, + "grad_norm": 0.26961296796798706, + "learning_rate": 3.269182026083167e-05, + "loss": 0.0057, + "step": 63190 + }, + { + "epoch": 9.816713264989128, + "grad_norm": 0.10664170235395432, + "learning_rate": 3.2676308780841516e-05, + "loss": 0.0053, + "step": 63200 + }, + { + "epoch": 9.818266542404473, + "grad_norm": 0.2291848510503769, + "learning_rate": 3.266079919534416e-05, + "loss": 0.0053, + "step": 63210 + }, + { + "epoch": 9.81981981981982, + "grad_norm": 0.19007201492786407, + "learning_rate": 3.264529150603575e-05, + "loss": 0.004, + "step": 63220 + }, + { + "epoch": 9.821373097235167, + "grad_norm": 0.35627639293670654, + "learning_rate": 3.262978571461217e-05, + "loss": 0.0073, + "step": 63230 + }, + { + "epoch": 9.822926374650512, + "grad_norm": 0.21753764152526855, + "learning_rate": 3.26142818227691e-05, + "loss": 0.0061, + "step": 63240 + }, + { + "epoch": 9.82447965206586, + "grad_norm": 0.19293244183063507, + "learning_rate": 3.259877983220203e-05, + "loss": 0.0075, + "step": 63250 + }, + { + "epoch": 9.826032929481205, + "grad_norm": 0.2530408799648285, + "learning_rate": 3.2583279744606246e-05, + "loss": 0.0073, + "step": 63260 + }, + { + "epoch": 9.827586206896552, + "grad_norm": 0.40896347165107727, + "learning_rate": 3.25677815616768e-05, + "loss": 0.0073, + "step": 63270 + }, + { + "epoch": 9.829139484311899, + "grad_norm": 0.37677067518234253, + "learning_rate": 3.255228528510856e-05, + "loss": 0.007, + "step": 63280 + }, + { + "epoch": 9.830692761727244, + "grad_norm": 0.20600737631320953, + "learning_rate": 3.253679091659617e-05, + "loss": 0.0065, + "step": 63290 + }, + { + "epoch": 9.832246039142591, + "grad_norm": 0.3826873004436493, + "learning_rate": 3.252129845783404e-05, + "loss": 0.0043, + "step": 63300 + }, + { + "epoch": 9.833799316557938, + "grad_norm": 0.17069783806800842, + "learning_rate": 3.2505807910516456e-05, + "loss": 0.0055, + "step": 63310 + }, + { + "epoch": 9.835352593973283, + "grad_norm": 0.3790093660354614, + "learning_rate": 3.249031927633742e-05, + "loss": 0.0068, + "step": 63320 + }, + { + "epoch": 9.83690587138863, + "grad_norm": 0.1863945722579956, + "learning_rate": 3.2474832556990737e-05, + "loss": 0.0037, + "step": 63330 + }, + { + "epoch": 9.838459148803976, + "grad_norm": 0.35051649808883667, + "learning_rate": 3.245934775417002e-05, + "loss": 0.007, + "step": 63340 + }, + { + "epoch": 9.840012426219323, + "grad_norm": 0.19986602663993835, + "learning_rate": 3.244386486956866e-05, + "loss": 0.0064, + "step": 63350 + }, + { + "epoch": 9.84156570363467, + "grad_norm": 0.19038176536560059, + "learning_rate": 3.2428383904879846e-05, + "loss": 0.0061, + "step": 63360 + }, + { + "epoch": 9.843118981050015, + "grad_norm": 0.3345871567726135, + "learning_rate": 3.241290486179652e-05, + "loss": 0.0052, + "step": 63370 + }, + { + "epoch": 9.844672258465362, + "grad_norm": 0.2763960063457489, + "learning_rate": 3.239742774201151e-05, + "loss": 0.0053, + "step": 63380 + }, + { + "epoch": 9.84622553588071, + "grad_norm": 0.2834492325782776, + "learning_rate": 3.238195254721733e-05, + "loss": 0.005, + "step": 63390 + }, + { + "epoch": 9.847778813296054, + "grad_norm": 0.2573089599609375, + "learning_rate": 3.236647927910633e-05, + "loss": 0.0046, + "step": 63400 + }, + { + "epoch": 9.849332090711401, + "grad_norm": 0.21229255199432373, + "learning_rate": 3.2351007939370646e-05, + "loss": 0.0067, + "step": 63410 + }, + { + "epoch": 9.850885368126747, + "grad_norm": 0.19080516695976257, + "learning_rate": 3.2335538529702184e-05, + "loss": 0.0053, + "step": 63420 + }, + { + "epoch": 9.852438645542094, + "grad_norm": 0.13301397860050201, + "learning_rate": 3.232007105179268e-05, + "loss": 0.0055, + "step": 63430 + }, + { + "epoch": 9.85399192295744, + "grad_norm": 0.16999252140522003, + "learning_rate": 3.230460550733363e-05, + "loss": 0.0035, + "step": 63440 + }, + { + "epoch": 9.855545200372786, + "grad_norm": 0.2735725939273834, + "learning_rate": 3.2289141898016315e-05, + "loss": 0.0065, + "step": 63450 + }, + { + "epoch": 9.857098477788133, + "grad_norm": 0.19585317373275757, + "learning_rate": 3.227368022553181e-05, + "loss": 0.0057, + "step": 63460 + }, + { + "epoch": 9.858651755203478, + "grad_norm": 0.1615191102027893, + "learning_rate": 3.225822049157099e-05, + "loss": 0.0048, + "step": 63470 + }, + { + "epoch": 9.860205032618826, + "grad_norm": 0.6229748129844666, + "learning_rate": 3.2242762697824494e-05, + "loss": 0.0076, + "step": 63480 + }, + { + "epoch": 9.861758310034173, + "grad_norm": 0.1596696972846985, + "learning_rate": 3.222730684598279e-05, + "loss": 0.005, + "step": 63490 + }, + { + "epoch": 9.863311587449518, + "grad_norm": 0.271090567111969, + "learning_rate": 3.221185293773605e-05, + "loss": 0.007, + "step": 63500 + }, + { + "epoch": 9.864864864864865, + "grad_norm": 1.3806945085525513, + "learning_rate": 3.2196400974774354e-05, + "loss": 0.0073, + "step": 63510 + }, + { + "epoch": 9.866418142280212, + "grad_norm": 0.39766719937324524, + "learning_rate": 3.2180950958787485e-05, + "loss": 0.0063, + "step": 63520 + }, + { + "epoch": 9.867971419695557, + "grad_norm": 0.28880774974823, + "learning_rate": 3.2165502891465024e-05, + "loss": 0.0057, + "step": 63530 + }, + { + "epoch": 9.869524697110904, + "grad_norm": 0.2532366216182709, + "learning_rate": 3.215005677449636e-05, + "loss": 0.007, + "step": 63540 + }, + { + "epoch": 9.87107797452625, + "grad_norm": 0.41649481654167175, + "learning_rate": 3.2134612609570644e-05, + "loss": 0.0071, + "step": 63550 + }, + { + "epoch": 9.872631251941597, + "grad_norm": 0.3553480803966522, + "learning_rate": 3.211917039837682e-05, + "loss": 0.0064, + "step": 63560 + }, + { + "epoch": 9.874184529356944, + "grad_norm": 0.4112820625305176, + "learning_rate": 3.210373014260365e-05, + "loss": 0.0069, + "step": 63570 + }, + { + "epoch": 9.875737806772289, + "grad_norm": 0.15448229014873505, + "learning_rate": 3.208829184393965e-05, + "loss": 0.0065, + "step": 63580 + }, + { + "epoch": 9.877291084187636, + "grad_norm": 0.3964022099971771, + "learning_rate": 3.207285550407312e-05, + "loss": 0.0072, + "step": 63590 + }, + { + "epoch": 9.878844361602983, + "grad_norm": 0.23615224659442902, + "learning_rate": 3.205742112469217e-05, + "loss": 0.006, + "step": 63600 + }, + { + "epoch": 9.880397639018328, + "grad_norm": 0.23292593657970428, + "learning_rate": 3.204198870748465e-05, + "loss": 0.0043, + "step": 63610 + }, + { + "epoch": 9.881950916433675, + "grad_norm": 0.5391322374343872, + "learning_rate": 3.202655825413824e-05, + "loss": 0.0075, + "step": 63620 + }, + { + "epoch": 9.88350419384902, + "grad_norm": 0.35043370723724365, + "learning_rate": 3.2011129766340416e-05, + "loss": 0.0074, + "step": 63630 + }, + { + "epoch": 9.885057471264368, + "grad_norm": 0.21995866298675537, + "learning_rate": 3.1995703245778384e-05, + "loss": 0.0059, + "step": 63640 + }, + { + "epoch": 9.886610748679715, + "grad_norm": 0.36926135420799255, + "learning_rate": 3.1980278694139186e-05, + "loss": 0.0063, + "step": 63650 + }, + { + "epoch": 9.88816402609506, + "grad_norm": 0.19302406907081604, + "learning_rate": 3.196485611310962e-05, + "loss": 0.0045, + "step": 63660 + }, + { + "epoch": 9.889717303510407, + "grad_norm": 0.4754584729671478, + "learning_rate": 3.1949435504376257e-05, + "loss": 0.0062, + "step": 63670 + }, + { + "epoch": 9.891270580925752, + "grad_norm": 0.20016324520111084, + "learning_rate": 3.1934016869625505e-05, + "loss": 0.0048, + "step": 63680 + }, + { + "epoch": 9.8928238583411, + "grad_norm": 0.19377891719341278, + "learning_rate": 3.191860021054349e-05, + "loss": 0.0044, + "step": 63690 + }, + { + "epoch": 9.894377135756447, + "grad_norm": 0.2029232233762741, + "learning_rate": 3.190318552881618e-05, + "loss": 0.0063, + "step": 63700 + }, + { + "epoch": 9.895930413171792, + "grad_norm": 0.7481921315193176, + "learning_rate": 3.1887772826129285e-05, + "loss": 0.0094, + "step": 63710 + }, + { + "epoch": 9.897483690587139, + "grad_norm": 0.18317857384681702, + "learning_rate": 3.1872362104168335e-05, + "loss": 0.0044, + "step": 63720 + }, + { + "epoch": 9.899036968002486, + "grad_norm": 0.31060460209846497, + "learning_rate": 3.185695336461861e-05, + "loss": 0.0062, + "step": 63730 + }, + { + "epoch": 9.900590245417831, + "grad_norm": 0.3913334906101227, + "learning_rate": 3.184154660916519e-05, + "loss": 0.0041, + "step": 63740 + }, + { + "epoch": 9.902143522833178, + "grad_norm": 0.3926562964916229, + "learning_rate": 3.1826141839492906e-05, + "loss": 0.0063, + "step": 63750 + }, + { + "epoch": 9.903696800248525, + "grad_norm": 0.44371089339256287, + "learning_rate": 3.181073905728644e-05, + "loss": 0.0055, + "step": 63760 + }, + { + "epoch": 9.90525007766387, + "grad_norm": 0.272521048784256, + "learning_rate": 3.1795338264230214e-05, + "loss": 0.0063, + "step": 63770 + }, + { + "epoch": 9.906803355079218, + "grad_norm": 1.0125190019607544, + "learning_rate": 3.1779939462008426e-05, + "loss": 0.0068, + "step": 63780 + }, + { + "epoch": 9.908356632494563, + "grad_norm": 0.33463582396507263, + "learning_rate": 3.1764542652305066e-05, + "loss": 0.0059, + "step": 63790 + }, + { + "epoch": 9.90990990990991, + "grad_norm": 0.20710937678813934, + "learning_rate": 3.17491478368039e-05, + "loss": 0.0047, + "step": 63800 + }, + { + "epoch": 9.911463187325257, + "grad_norm": 0.2846578359603882, + "learning_rate": 3.173375501718847e-05, + "loss": 0.0057, + "step": 63810 + }, + { + "epoch": 9.913016464740602, + "grad_norm": 0.2647809386253357, + "learning_rate": 3.171836419514215e-05, + "loss": 0.0045, + "step": 63820 + }, + { + "epoch": 9.91456974215595, + "grad_norm": 0.11770126968622208, + "learning_rate": 3.170297537234803e-05, + "loss": 0.0032, + "step": 63830 + }, + { + "epoch": 9.916123019571295, + "grad_norm": 0.15648053586483002, + "learning_rate": 3.168758855048901e-05, + "loss": 0.0064, + "step": 63840 + }, + { + "epoch": 9.917676296986642, + "grad_norm": 0.19139888882637024, + "learning_rate": 3.1672203731247785e-05, + "loss": 0.0045, + "step": 63850 + }, + { + "epoch": 9.919229574401989, + "grad_norm": 0.18321824073791504, + "learning_rate": 3.16568209163068e-05, + "loss": 0.0053, + "step": 63860 + }, + { + "epoch": 9.920782851817334, + "grad_norm": 1.4593015909194946, + "learning_rate": 3.1641440107348295e-05, + "loss": 0.0092, + "step": 63870 + }, + { + "epoch": 9.922336129232681, + "grad_norm": 0.12057845294475555, + "learning_rate": 3.162606130605429e-05, + "loss": 0.0061, + "step": 63880 + }, + { + "epoch": 9.923889406648028, + "grad_norm": 0.38405758142471313, + "learning_rate": 3.16106845141066e-05, + "loss": 0.0054, + "step": 63890 + }, + { + "epoch": 9.925442684063373, + "grad_norm": 0.3609088361263275, + "learning_rate": 3.159530973318682e-05, + "loss": 0.0069, + "step": 63900 + }, + { + "epoch": 9.92699596147872, + "grad_norm": 0.2538144588470459, + "learning_rate": 3.1579936964976286e-05, + "loss": 0.004, + "step": 63910 + }, + { + "epoch": 9.928549238894066, + "grad_norm": 0.4477938115596771, + "learning_rate": 3.156456621115615e-05, + "loss": 0.0043, + "step": 63920 + }, + { + "epoch": 9.930102516309413, + "grad_norm": 0.09901205450296402, + "learning_rate": 3.154919747340734e-05, + "loss": 0.0035, + "step": 63930 + }, + { + "epoch": 9.93165579372476, + "grad_norm": 0.25617164373397827, + "learning_rate": 3.153383075341054e-05, + "loss": 0.0069, + "step": 63940 + }, + { + "epoch": 9.933209071140105, + "grad_norm": 0.13747410476207733, + "learning_rate": 3.1518466052846264e-05, + "loss": 0.0053, + "step": 63950 + }, + { + "epoch": 9.934762348555452, + "grad_norm": 0.16566334664821625, + "learning_rate": 3.150310337339474e-05, + "loss": 0.0067, + "step": 63960 + }, + { + "epoch": 9.9363156259708, + "grad_norm": 0.2850434184074402, + "learning_rate": 3.148774271673603e-05, + "loss": 0.0063, + "step": 63970 + }, + { + "epoch": 9.937868903386144, + "grad_norm": 0.8887864351272583, + "learning_rate": 3.147238408454995e-05, + "loss": 0.0061, + "step": 63980 + }, + { + "epoch": 9.939422180801492, + "grad_norm": 0.24342191219329834, + "learning_rate": 3.1457027478516086e-05, + "loss": 0.0055, + "step": 63990 + }, + { + "epoch": 9.940975458216837, + "grad_norm": 0.43935996294021606, + "learning_rate": 3.144167290031382e-05, + "loss": 0.0048, + "step": 64000 + }, + { + "epoch": 9.942528735632184, + "grad_norm": 0.18036630749702454, + "learning_rate": 3.142632035162227e-05, + "loss": 0.0074, + "step": 64010 + }, + { + "epoch": 9.944082013047531, + "grad_norm": 0.2052585780620575, + "learning_rate": 3.141096983412044e-05, + "loss": 0.0052, + "step": 64020 + }, + { + "epoch": 9.945635290462876, + "grad_norm": 0.3525688946247101, + "learning_rate": 3.139562134948699e-05, + "loss": 0.0048, + "step": 64030 + }, + { + "epoch": 9.947188567878223, + "grad_norm": 0.31829938292503357, + "learning_rate": 3.1380274899400424e-05, + "loss": 0.0092, + "step": 64040 + }, + { + "epoch": 9.948741845293569, + "grad_norm": 0.39474329352378845, + "learning_rate": 3.1364930485539e-05, + "loss": 0.0072, + "step": 64050 + }, + { + "epoch": 9.950295122708916, + "grad_norm": 0.3004770576953888, + "learning_rate": 3.134958810958076e-05, + "loss": 0.0039, + "step": 64060 + }, + { + "epoch": 9.951848400124263, + "grad_norm": 0.3836860954761505, + "learning_rate": 3.13342477732035e-05, + "loss": 0.0058, + "step": 64070 + }, + { + "epoch": 9.953401677539608, + "grad_norm": 0.26356372237205505, + "learning_rate": 3.131890947808487e-05, + "loss": 0.0057, + "step": 64080 + }, + { + "epoch": 9.954954954954955, + "grad_norm": 0.6384226083755493, + "learning_rate": 3.13035732259022e-05, + "loss": 0.0086, + "step": 64090 + }, + { + "epoch": 9.956508232370302, + "grad_norm": 0.23211228847503662, + "learning_rate": 3.1288239018332656e-05, + "loss": 0.0036, + "step": 64100 + }, + { + "epoch": 9.958061509785647, + "grad_norm": 0.19456690549850464, + "learning_rate": 3.1272906857053164e-05, + "loss": 0.0078, + "step": 64110 + }, + { + "epoch": 9.959614787200994, + "grad_norm": 0.5126039385795593, + "learning_rate": 3.125757674374042e-05, + "loss": 0.004, + "step": 64120 + }, + { + "epoch": 9.961168064616341, + "grad_norm": 0.2898824214935303, + "learning_rate": 3.124224868007088e-05, + "loss": 0.0085, + "step": 64130 + }, + { + "epoch": 9.962721342031687, + "grad_norm": 0.21889103949069977, + "learning_rate": 3.1226922667720854e-05, + "loss": 0.0058, + "step": 64140 + }, + { + "epoch": 9.964274619447034, + "grad_norm": 0.4770979881286621, + "learning_rate": 3.121159870836633e-05, + "loss": 0.0058, + "step": 64150 + }, + { + "epoch": 9.965827896862379, + "grad_norm": 0.21675921976566315, + "learning_rate": 3.119627680368312e-05, + "loss": 0.0038, + "step": 64160 + }, + { + "epoch": 9.967381174277726, + "grad_norm": 0.18871504068374634, + "learning_rate": 3.1180956955346816e-05, + "loss": 0.0049, + "step": 64170 + }, + { + "epoch": 9.968934451693073, + "grad_norm": 0.3968191146850586, + "learning_rate": 3.116563916503276e-05, + "loss": 0.0053, + "step": 64180 + }, + { + "epoch": 9.970487729108418, + "grad_norm": 0.1894296407699585, + "learning_rate": 3.1150323434416095e-05, + "loss": 0.0057, + "step": 64190 + }, + { + "epoch": 9.972041006523765, + "grad_norm": 0.18828828632831573, + "learning_rate": 3.1135009765171696e-05, + "loss": 0.0037, + "step": 64200 + }, + { + "epoch": 9.97359428393911, + "grad_norm": 1.0985952615737915, + "learning_rate": 3.111969815897429e-05, + "loss": 0.0058, + "step": 64210 + }, + { + "epoch": 9.975147561354458, + "grad_norm": 0.4730452299118042, + "learning_rate": 3.1104388617498295e-05, + "loss": 0.0048, + "step": 64220 + }, + { + "epoch": 9.976700838769805, + "grad_norm": 0.12294604629278183, + "learning_rate": 3.1089081142417966e-05, + "loss": 0.0083, + "step": 64230 + }, + { + "epoch": 9.97825411618515, + "grad_norm": 0.17243105173110962, + "learning_rate": 3.1073775735407285e-05, + "loss": 0.007, + "step": 64240 + }, + { + "epoch": 9.979807393600497, + "grad_norm": 0.34315115213394165, + "learning_rate": 3.1058472398140035e-05, + "loss": 0.0069, + "step": 64250 + }, + { + "epoch": 9.981360671015844, + "grad_norm": 0.265624076128006, + "learning_rate": 3.1043171132289736e-05, + "loss": 0.0073, + "step": 64260 + }, + { + "epoch": 9.98291394843119, + "grad_norm": 0.6477730870246887, + "learning_rate": 3.1027871939529774e-05, + "loss": 0.0043, + "step": 64270 + }, + { + "epoch": 9.984467225846537, + "grad_norm": 0.20365846157073975, + "learning_rate": 3.1012574821533214e-05, + "loss": 0.0042, + "step": 64280 + }, + { + "epoch": 9.986020503261882, + "grad_norm": 0.348341703414917, + "learning_rate": 3.099727977997291e-05, + "loss": 0.0044, + "step": 64290 + }, + { + "epoch": 9.987573780677229, + "grad_norm": 0.1975564807653427, + "learning_rate": 3.098198681652154e-05, + "loss": 0.0046, + "step": 64300 + }, + { + "epoch": 9.989127058092576, + "grad_norm": 0.11852088570594788, + "learning_rate": 3.0966695932851475e-05, + "loss": 0.0046, + "step": 64310 + }, + { + "epoch": 9.990680335507921, + "grad_norm": 0.2642635405063629, + "learning_rate": 3.0951407130634926e-05, + "loss": 0.0081, + "step": 64320 + }, + { + "epoch": 9.992233612923268, + "grad_norm": 0.28098809719085693, + "learning_rate": 3.0936120411543864e-05, + "loss": 0.0054, + "step": 64330 + }, + { + "epoch": 9.993786890338615, + "grad_norm": 0.2947198152542114, + "learning_rate": 3.0920835777250015e-05, + "loss": 0.0051, + "step": 64340 + }, + { + "epoch": 9.99534016775396, + "grad_norm": 0.3167482912540436, + "learning_rate": 3.090555322942488e-05, + "loss": 0.0038, + "step": 64350 + }, + { + "epoch": 9.996893445169308, + "grad_norm": 0.22382333874702454, + "learning_rate": 3.089027276973973e-05, + "loss": 0.0043, + "step": 64360 + }, + { + "epoch": 9.998446722584653, + "grad_norm": 0.24627867341041565, + "learning_rate": 3.0874994399865615e-05, + "loss": 0.0062, + "step": 64370 + }, + { + "epoch": 10.0, + "grad_norm": 0.15048019587993622, + "learning_rate": 3.085971812147337e-05, + "loss": 0.0046, + "step": 64380 + }, + { + "epoch": 10.001553277415347, + "grad_norm": 0.1900043487548828, + "learning_rate": 3.0844443936233557e-05, + "loss": 0.0055, + "step": 64390 + }, + { + "epoch": 10.003106554830692, + "grad_norm": 0.14249108731746674, + "learning_rate": 3.082917184581657e-05, + "loss": 0.0037, + "step": 64400 + }, + { + "epoch": 10.00465983224604, + "grad_norm": 2.0170323848724365, + "learning_rate": 3.0813901851892513e-05, + "loss": 0.0054, + "step": 64410 + }, + { + "epoch": 10.006213109661385, + "grad_norm": 0.1160987913608551, + "learning_rate": 3.0798633956131316e-05, + "loss": 0.0059, + "step": 64420 + }, + { + "epoch": 10.007766387076732, + "grad_norm": 0.16649965941905975, + "learning_rate": 3.078336816020264e-05, + "loss": 0.004, + "step": 64430 + }, + { + "epoch": 10.009319664492079, + "grad_norm": 0.4761431813240051, + "learning_rate": 3.0768104465775935e-05, + "loss": 0.0057, + "step": 64440 + }, + { + "epoch": 10.010872941907424, + "grad_norm": 0.26982617378234863, + "learning_rate": 3.0752842874520396e-05, + "loss": 0.0065, + "step": 64450 + }, + { + "epoch": 10.012426219322771, + "grad_norm": 0.2754349708557129, + "learning_rate": 3.073758338810503e-05, + "loss": 0.0054, + "step": 64460 + }, + { + "epoch": 10.013979496738118, + "grad_norm": 0.27248355746269226, + "learning_rate": 3.072232600819859e-05, + "loss": 0.0066, + "step": 64470 + }, + { + "epoch": 10.015532774153463, + "grad_norm": 0.5972883105278015, + "learning_rate": 3.070707073646959e-05, + "loss": 0.0051, + "step": 64480 + }, + { + "epoch": 10.01708605156881, + "grad_norm": 0.15091584622859955, + "learning_rate": 3.069181757458633e-05, + "loss": 0.0044, + "step": 64490 + }, + { + "epoch": 10.018639328984156, + "grad_norm": 0.23538796603679657, + "learning_rate": 3.067656652421687e-05, + "loss": 0.0068, + "step": 64500 + }, + { + "epoch": 10.020192606399503, + "grad_norm": 0.3718745708465576, + "learning_rate": 3.0661317587029025e-05, + "loss": 0.0051, + "step": 64510 + }, + { + "epoch": 10.02174588381485, + "grad_norm": 0.3590950071811676, + "learning_rate": 3.064607076469043e-05, + "loss": 0.0067, + "step": 64520 + }, + { + "epoch": 10.023299161230195, + "grad_norm": 0.1034930944442749, + "learning_rate": 3.063082605886845e-05, + "loss": 0.0046, + "step": 64530 + }, + { + "epoch": 10.024852438645542, + "grad_norm": 0.30265042185783386, + "learning_rate": 3.061558347123019e-05, + "loss": 0.0069, + "step": 64540 + }, + { + "epoch": 10.02640571606089, + "grad_norm": 0.22306384146213531, + "learning_rate": 3.060034300344259e-05, + "loss": 0.0047, + "step": 64550 + }, + { + "epoch": 10.027958993476235, + "grad_norm": 0.19355928897857666, + "learning_rate": 3.058510465717231e-05, + "loss": 0.0047, + "step": 64560 + }, + { + "epoch": 10.029512270891582, + "grad_norm": 0.2247263789176941, + "learning_rate": 3.0569868434085804e-05, + "loss": 0.0067, + "step": 64570 + }, + { + "epoch": 10.031065548306927, + "grad_norm": 0.1736288219690323, + "learning_rate": 3.055463433584924e-05, + "loss": 0.0051, + "step": 64580 + }, + { + "epoch": 10.032618825722274, + "grad_norm": 0.2288675457239151, + "learning_rate": 3.0539402364128656e-05, + "loss": 0.0043, + "step": 64590 + }, + { + "epoch": 10.034172103137621, + "grad_norm": 0.2175235152244568, + "learning_rate": 3.052417252058977e-05, + "loss": 0.0045, + "step": 64600 + }, + { + "epoch": 10.035725380552966, + "grad_norm": 0.2128462791442871, + "learning_rate": 3.0508944806898092e-05, + "loss": 0.0056, + "step": 64610 + }, + { + "epoch": 10.037278657968313, + "grad_norm": 0.32595840096473694, + "learning_rate": 3.04937192247189e-05, + "loss": 0.008, + "step": 64620 + }, + { + "epoch": 10.03883193538366, + "grad_norm": 0.35622140765190125, + "learning_rate": 3.047849577571725e-05, + "loss": 0.0065, + "step": 64630 + }, + { + "epoch": 10.040385212799006, + "grad_norm": 0.16534855961799622, + "learning_rate": 3.046327446155795e-05, + "loss": 0.0045, + "step": 64640 + }, + { + "epoch": 10.041938490214353, + "grad_norm": 0.32349008321762085, + "learning_rate": 3.0448055283905587e-05, + "loss": 0.0042, + "step": 64650 + }, + { + "epoch": 10.043491767629698, + "grad_norm": 0.20393943786621094, + "learning_rate": 3.0432838244424494e-05, + "loss": 0.0044, + "step": 64660 + }, + { + "epoch": 10.045045045045045, + "grad_norm": 0.6214424967765808, + "learning_rate": 3.04176233447788e-05, + "loss": 0.0073, + "step": 64670 + }, + { + "epoch": 10.046598322460392, + "grad_norm": 0.25512057542800903, + "learning_rate": 3.040241058663238e-05, + "loss": 0.0052, + "step": 64680 + }, + { + "epoch": 10.048151599875737, + "grad_norm": 0.12905670702457428, + "learning_rate": 3.038719997164887e-05, + "loss": 0.0066, + "step": 64690 + }, + { + "epoch": 10.049704877291084, + "grad_norm": 0.280681312084198, + "learning_rate": 3.0371991501491683e-05, + "loss": 0.0053, + "step": 64700 + }, + { + "epoch": 10.05125815470643, + "grad_norm": 0.25704699754714966, + "learning_rate": 3.0356785177823975e-05, + "loss": 0.0048, + "step": 64710 + }, + { + "epoch": 10.052811432121777, + "grad_norm": 0.17412972450256348, + "learning_rate": 3.0341581002308726e-05, + "loss": 0.0058, + "step": 64720 + }, + { + "epoch": 10.054364709537124, + "grad_norm": 0.1429041624069214, + "learning_rate": 3.0326378976608627e-05, + "loss": 0.0062, + "step": 64730 + }, + { + "epoch": 10.055917986952469, + "grad_norm": 0.812179446220398, + "learning_rate": 3.0311179102386146e-05, + "loss": 0.004, + "step": 64740 + }, + { + "epoch": 10.057471264367816, + "grad_norm": 0.2661474645137787, + "learning_rate": 3.0295981381303513e-05, + "loss": 0.0064, + "step": 64750 + }, + { + "epoch": 10.059024541783163, + "grad_norm": 0.34740912914276123, + "learning_rate": 3.0280785815022727e-05, + "loss": 0.0051, + "step": 64760 + }, + { + "epoch": 10.060577819198508, + "grad_norm": 0.2007305920124054, + "learning_rate": 3.0265592405205544e-05, + "loss": 0.0068, + "step": 64770 + }, + { + "epoch": 10.062131096613856, + "grad_norm": 0.7046357989311218, + "learning_rate": 3.0250401153513517e-05, + "loss": 0.0055, + "step": 64780 + }, + { + "epoch": 10.0636843740292, + "grad_norm": 0.10594330728054047, + "learning_rate": 3.0235212061607936e-05, + "loss": 0.0054, + "step": 64790 + }, + { + "epoch": 10.065237651444548, + "grad_norm": 0.33771345019340515, + "learning_rate": 3.0220025131149842e-05, + "loss": 0.0055, + "step": 64800 + }, + { + "epoch": 10.066790928859895, + "grad_norm": 0.9624385833740234, + "learning_rate": 3.020484036380007e-05, + "loss": 0.0067, + "step": 64810 + }, + { + "epoch": 10.06834420627524, + "grad_norm": 0.6530340313911438, + "learning_rate": 3.0189657761219188e-05, + "loss": 0.0061, + "step": 64820 + }, + { + "epoch": 10.069897483690587, + "grad_norm": 0.4229275584220886, + "learning_rate": 3.017447732506753e-05, + "loss": 0.0052, + "step": 64830 + }, + { + "epoch": 10.071450761105934, + "grad_norm": 0.1619603931903839, + "learning_rate": 3.015929905700524e-05, + "loss": 0.0074, + "step": 64840 + }, + { + "epoch": 10.07300403852128, + "grad_norm": 0.26938241720199585, + "learning_rate": 3.014412295869218e-05, + "loss": 0.0091, + "step": 64850 + }, + { + "epoch": 10.074557315936627, + "grad_norm": 0.12494578957557678, + "learning_rate": 3.0128949031787977e-05, + "loss": 0.0056, + "step": 64860 + }, + { + "epoch": 10.076110593351972, + "grad_norm": 0.14834217727184296, + "learning_rate": 3.0113777277952022e-05, + "loss": 0.0062, + "step": 64870 + }, + { + "epoch": 10.077663870767319, + "grad_norm": 0.15218645334243774, + "learning_rate": 3.0098607698843496e-05, + "loss": 0.0055, + "step": 64880 + }, + { + "epoch": 10.079217148182666, + "grad_norm": 0.3010407090187073, + "learning_rate": 3.0083440296121312e-05, + "loss": 0.0059, + "step": 64890 + }, + { + "epoch": 10.080770425598011, + "grad_norm": 0.23632334172725677, + "learning_rate": 3.0068275071444142e-05, + "loss": 0.0044, + "step": 64900 + }, + { + "epoch": 10.082323703013358, + "grad_norm": 0.2824622094631195, + "learning_rate": 3.0053112026470442e-05, + "loss": 0.0044, + "step": 64910 + }, + { + "epoch": 10.083876980428705, + "grad_norm": 0.20564697682857513, + "learning_rate": 3.0037951162858434e-05, + "loss": 0.0062, + "step": 64920 + }, + { + "epoch": 10.08543025784405, + "grad_norm": 0.17552661895751953, + "learning_rate": 3.002279248226607e-05, + "loss": 0.006, + "step": 64930 + }, + { + "epoch": 10.086983535259398, + "grad_norm": 0.2943773567676544, + "learning_rate": 3.000763598635109e-05, + "loss": 0.0042, + "step": 64940 + }, + { + "epoch": 10.088536812674743, + "grad_norm": 0.36925309896469116, + "learning_rate": 2.999248167677098e-05, + "loss": 0.005, + "step": 64950 + }, + { + "epoch": 10.09009009009009, + "grad_norm": 0.33716562390327454, + "learning_rate": 2.9977329555182976e-05, + "loss": 0.0053, + "step": 64960 + }, + { + "epoch": 10.091643367505437, + "grad_norm": 0.24941779673099518, + "learning_rate": 2.9962179623244124e-05, + "loss": 0.0052, + "step": 64970 + }, + { + "epoch": 10.093196644920782, + "grad_norm": 0.41343995928764343, + "learning_rate": 2.9947031882611175e-05, + "loss": 0.0044, + "step": 64980 + }, + { + "epoch": 10.09474992233613, + "grad_norm": 0.19811023771762848, + "learning_rate": 2.9931886334940685e-05, + "loss": 0.006, + "step": 64990 + }, + { + "epoch": 10.096303199751475, + "grad_norm": 0.2618159353733063, + "learning_rate": 2.9916742981888923e-05, + "loss": 0.0061, + "step": 65000 + }, + { + "epoch": 10.097856477166822, + "grad_norm": 0.2955986261367798, + "learning_rate": 2.990160182511196e-05, + "loss": 0.0058, + "step": 65010 + }, + { + "epoch": 10.099409754582169, + "grad_norm": 0.5842089653015137, + "learning_rate": 2.9886462866265574e-05, + "loss": 0.0043, + "step": 65020 + }, + { + "epoch": 10.100963031997514, + "grad_norm": 0.28384190797805786, + "learning_rate": 2.9871326107005392e-05, + "loss": 0.0071, + "step": 65030 + }, + { + "epoch": 10.102516309412861, + "grad_norm": 0.1482708603143692, + "learning_rate": 2.9856191548986735e-05, + "loss": 0.0047, + "step": 65040 + }, + { + "epoch": 10.104069586828208, + "grad_norm": 0.2127990573644638, + "learning_rate": 2.9841059193864672e-05, + "loss": 0.0081, + "step": 65050 + }, + { + "epoch": 10.105622864243553, + "grad_norm": 0.10007095336914062, + "learning_rate": 2.982592904329407e-05, + "loss": 0.0041, + "step": 65060 + }, + { + "epoch": 10.1071761416589, + "grad_norm": 0.27884185314178467, + "learning_rate": 2.981080109892954e-05, + "loss": 0.0045, + "step": 65070 + }, + { + "epoch": 10.108729419074246, + "grad_norm": 0.13682053983211517, + "learning_rate": 2.9795675362425434e-05, + "loss": 0.0064, + "step": 65080 + }, + { + "epoch": 10.110282696489593, + "grad_norm": 0.233717143535614, + "learning_rate": 2.97805518354359e-05, + "loss": 0.0059, + "step": 65090 + }, + { + "epoch": 10.11183597390494, + "grad_norm": 0.08991697430610657, + "learning_rate": 2.9765430519614825e-05, + "loss": 0.0051, + "step": 65100 + }, + { + "epoch": 10.113389251320285, + "grad_norm": 0.2023911029100418, + "learning_rate": 2.975031141661584e-05, + "loss": 0.0064, + "step": 65110 + }, + { + "epoch": 10.114942528735632, + "grad_norm": 0.8914273381233215, + "learning_rate": 2.9735194528092347e-05, + "loss": 0.0061, + "step": 65120 + }, + { + "epoch": 10.11649580615098, + "grad_norm": 0.2469567060470581, + "learning_rate": 2.9720079855697525e-05, + "loss": 0.0082, + "step": 65130 + }, + { + "epoch": 10.118049083566325, + "grad_norm": 0.6057780385017395, + "learning_rate": 2.970496740108427e-05, + "loss": 0.0059, + "step": 65140 + }, + { + "epoch": 10.119602360981672, + "grad_norm": 0.2685292959213257, + "learning_rate": 2.968985716590527e-05, + "loss": 0.0065, + "step": 65150 + }, + { + "epoch": 10.121155638397017, + "grad_norm": 0.1514466255903244, + "learning_rate": 2.9674749151812953e-05, + "loss": 0.0052, + "step": 65160 + }, + { + "epoch": 10.122708915812364, + "grad_norm": 0.33287203311920166, + "learning_rate": 2.965964336045951e-05, + "loss": 0.0057, + "step": 65170 + }, + { + "epoch": 10.124262193227711, + "grad_norm": 0.2903996407985687, + "learning_rate": 2.9644539793496894e-05, + "loss": 0.0038, + "step": 65180 + }, + { + "epoch": 10.125815470643056, + "grad_norm": 0.23651011288166046, + "learning_rate": 2.9629438452576807e-05, + "loss": 0.0046, + "step": 65190 + }, + { + "epoch": 10.127368748058403, + "grad_norm": 0.28176096081733704, + "learning_rate": 2.9614339339350698e-05, + "loss": 0.0055, + "step": 65200 + }, + { + "epoch": 10.12892202547375, + "grad_norm": 0.1213783323764801, + "learning_rate": 2.959924245546979e-05, + "loss": 0.0048, + "step": 65210 + }, + { + "epoch": 10.130475302889096, + "grad_norm": 0.21609874069690704, + "learning_rate": 2.9584147802585038e-05, + "loss": 0.0056, + "step": 65220 + }, + { + "epoch": 10.132028580304443, + "grad_norm": 0.3043827712535858, + "learning_rate": 2.956905538234721e-05, + "loss": 0.0062, + "step": 65230 + }, + { + "epoch": 10.133581857719788, + "grad_norm": 0.7608311772346497, + "learning_rate": 2.9553965196406764e-05, + "loss": 0.0073, + "step": 65240 + }, + { + "epoch": 10.135135135135135, + "grad_norm": 0.21910013258457184, + "learning_rate": 2.9538877246413943e-05, + "loss": 0.0061, + "step": 65250 + }, + { + "epoch": 10.136688412550482, + "grad_norm": 0.36349812150001526, + "learning_rate": 2.952379153401874e-05, + "loss": 0.0054, + "step": 65260 + }, + { + "epoch": 10.138241689965827, + "grad_norm": 0.12457887828350067, + "learning_rate": 2.9508708060870914e-05, + "loss": 0.0048, + "step": 65270 + }, + { + "epoch": 10.139794967381174, + "grad_norm": 0.2248966544866562, + "learning_rate": 2.9493626828619948e-05, + "loss": 0.0041, + "step": 65280 + }, + { + "epoch": 10.141348244796522, + "grad_norm": 0.28050607442855835, + "learning_rate": 2.9478547838915137e-05, + "loss": 0.0062, + "step": 65290 + }, + { + "epoch": 10.142901522211867, + "grad_norm": 0.27445727586746216, + "learning_rate": 2.946347109340548e-05, + "loss": 0.005, + "step": 65300 + }, + { + "epoch": 10.144454799627214, + "grad_norm": 0.21801774203777313, + "learning_rate": 2.944839659373976e-05, + "loss": 0.0066, + "step": 65310 + }, + { + "epoch": 10.14600807704256, + "grad_norm": 0.09297382831573486, + "learning_rate": 2.943332434156647e-05, + "loss": 0.004, + "step": 65320 + }, + { + "epoch": 10.147561354457906, + "grad_norm": 0.0746222734451294, + "learning_rate": 2.9418254338533912e-05, + "loss": 0.0049, + "step": 65330 + }, + { + "epoch": 10.149114631873253, + "grad_norm": 0.206834614276886, + "learning_rate": 2.940318658629011e-05, + "loss": 0.0071, + "step": 65340 + }, + { + "epoch": 10.150667909288599, + "grad_norm": 0.33276745676994324, + "learning_rate": 2.938812108648287e-05, + "loss": 0.0042, + "step": 65350 + }, + { + "epoch": 10.152221186703946, + "grad_norm": 0.30491507053375244, + "learning_rate": 2.937305784075972e-05, + "loss": 0.005, + "step": 65360 + }, + { + "epoch": 10.15377446411929, + "grad_norm": 0.37857815623283386, + "learning_rate": 2.9357996850767937e-05, + "loss": 0.0039, + "step": 65370 + }, + { + "epoch": 10.155327741534638, + "grad_norm": 0.16804121434688568, + "learning_rate": 2.9342938118154595e-05, + "loss": 0.0038, + "step": 65380 + }, + { + "epoch": 10.156881018949985, + "grad_norm": 0.30543702840805054, + "learning_rate": 2.9327881644566485e-05, + "loss": 0.0047, + "step": 65390 + }, + { + "epoch": 10.15843429636533, + "grad_norm": 0.29214951395988464, + "learning_rate": 2.9312827431650157e-05, + "loss": 0.0064, + "step": 65400 + }, + { + "epoch": 10.159987573780677, + "grad_norm": 0.17220811545848846, + "learning_rate": 2.92977754810519e-05, + "loss": 0.005, + "step": 65410 + }, + { + "epoch": 10.161540851196024, + "grad_norm": 0.1505308598279953, + "learning_rate": 2.928272579441781e-05, + "loss": 0.0055, + "step": 65420 + }, + { + "epoch": 10.16309412861137, + "grad_norm": 0.18376418948173523, + "learning_rate": 2.9267678373393676e-05, + "loss": 0.0079, + "step": 65430 + }, + { + "epoch": 10.164647406026717, + "grad_norm": 0.2876165509223938, + "learning_rate": 2.925263321962507e-05, + "loss": 0.0055, + "step": 65440 + }, + { + "epoch": 10.166200683442062, + "grad_norm": 0.2845800518989563, + "learning_rate": 2.9237590334757292e-05, + "loss": 0.0049, + "step": 65450 + }, + { + "epoch": 10.167753960857409, + "grad_norm": 0.3743513822555542, + "learning_rate": 2.922254972043543e-05, + "loss": 0.0047, + "step": 65460 + }, + { + "epoch": 10.169307238272756, + "grad_norm": 0.20058296620845795, + "learning_rate": 2.9207511378304265e-05, + "loss": 0.0053, + "step": 65470 + }, + { + "epoch": 10.170860515688101, + "grad_norm": 0.3127611577510834, + "learning_rate": 2.9192475310008428e-05, + "loss": 0.0073, + "step": 65480 + }, + { + "epoch": 10.172413793103448, + "grad_norm": 0.25923633575439453, + "learning_rate": 2.917744151719219e-05, + "loss": 0.0027, + "step": 65490 + }, + { + "epoch": 10.173967070518795, + "grad_norm": 0.28897130489349365, + "learning_rate": 2.916241000149964e-05, + "loss": 0.0048, + "step": 65500 + }, + { + "epoch": 10.17552034793414, + "grad_norm": 0.10666351765394211, + "learning_rate": 2.9147380764574622e-05, + "loss": 0.0035, + "step": 65510 + }, + { + "epoch": 10.177073625349488, + "grad_norm": 0.1554042100906372, + "learning_rate": 2.9132353808060674e-05, + "loss": 0.0046, + "step": 65520 + }, + { + "epoch": 10.178626902764833, + "grad_norm": 0.24466034770011902, + "learning_rate": 2.9117329133601134e-05, + "loss": 0.006, + "step": 65530 + }, + { + "epoch": 10.18018018018018, + "grad_norm": 0.11120536923408508, + "learning_rate": 2.9102306742839093e-05, + "loss": 0.0053, + "step": 65540 + }, + { + "epoch": 10.181733457595527, + "grad_norm": 1.2893611192703247, + "learning_rate": 2.9087286637417375e-05, + "loss": 0.0061, + "step": 65550 + }, + { + "epoch": 10.183286735010872, + "grad_norm": 0.3379560112953186, + "learning_rate": 2.9072268818978537e-05, + "loss": 0.0077, + "step": 65560 + }, + { + "epoch": 10.18484001242622, + "grad_norm": 0.22510038316249847, + "learning_rate": 2.905725328916493e-05, + "loss": 0.0064, + "step": 65570 + }, + { + "epoch": 10.186393289841567, + "grad_norm": 0.4075659215450287, + "learning_rate": 2.9042240049618608e-05, + "loss": 0.0064, + "step": 65580 + }, + { + "epoch": 10.187946567256912, + "grad_norm": 0.26629823446273804, + "learning_rate": 2.9027229101981413e-05, + "loss": 0.0061, + "step": 65590 + }, + { + "epoch": 10.189499844672259, + "grad_norm": 0.5741413831710815, + "learning_rate": 2.901222044789488e-05, + "loss": 0.0056, + "step": 65600 + }, + { + "epoch": 10.191053122087604, + "grad_norm": 0.15920403599739075, + "learning_rate": 2.89972140890004e-05, + "loss": 0.0048, + "step": 65610 + }, + { + "epoch": 10.192606399502951, + "grad_norm": 0.16541071236133575, + "learning_rate": 2.8982210026938988e-05, + "loss": 0.0063, + "step": 65620 + }, + { + "epoch": 10.194159676918298, + "grad_norm": 0.28608161211013794, + "learning_rate": 2.896720826335151e-05, + "loss": 0.0055, + "step": 65630 + }, + { + "epoch": 10.195712954333644, + "grad_norm": 0.3096913695335388, + "learning_rate": 2.8952208799878493e-05, + "loss": 0.0061, + "step": 65640 + }, + { + "epoch": 10.19726623174899, + "grad_norm": 0.24261216819286346, + "learning_rate": 2.89372116381603e-05, + "loss": 0.0052, + "step": 65650 + }, + { + "epoch": 10.198819509164336, + "grad_norm": 0.517831027507782, + "learning_rate": 2.8922216779836932e-05, + "loss": 0.0062, + "step": 65660 + }, + { + "epoch": 10.200372786579683, + "grad_norm": 0.10233043879270554, + "learning_rate": 2.8907224226548285e-05, + "loss": 0.0049, + "step": 65670 + }, + { + "epoch": 10.20192606399503, + "grad_norm": 0.5664640069007874, + "learning_rate": 2.8892233979933868e-05, + "loss": 0.0081, + "step": 65680 + }, + { + "epoch": 10.203479341410375, + "grad_norm": 0.13700859248638153, + "learning_rate": 2.8877246041633026e-05, + "loss": 0.005, + "step": 65690 + }, + { + "epoch": 10.205032618825722, + "grad_norm": 0.4377964437007904, + "learning_rate": 2.886226041328478e-05, + "loss": 0.0049, + "step": 65700 + }, + { + "epoch": 10.20658589624107, + "grad_norm": 0.16055354475975037, + "learning_rate": 2.8847277096527957e-05, + "loss": 0.006, + "step": 65710 + }, + { + "epoch": 10.208139173656415, + "grad_norm": 0.2647785544395447, + "learning_rate": 2.8832296093001105e-05, + "loss": 0.0049, + "step": 65720 + }, + { + "epoch": 10.209692451071762, + "grad_norm": 0.20727072656154633, + "learning_rate": 2.8817317404342547e-05, + "loss": 0.0053, + "step": 65730 + }, + { + "epoch": 10.211245728487107, + "grad_norm": 0.1878172904253006, + "learning_rate": 2.880234103219029e-05, + "loss": 0.0056, + "step": 65740 + }, + { + "epoch": 10.212799005902454, + "grad_norm": 0.29580581188201904, + "learning_rate": 2.878736697818215e-05, + "loss": 0.0054, + "step": 65750 + }, + { + "epoch": 10.214352283317801, + "grad_norm": 0.23773254454135895, + "learning_rate": 2.877239524395568e-05, + "loss": 0.005, + "step": 65760 + }, + { + "epoch": 10.215905560733146, + "grad_norm": 0.286472886800766, + "learning_rate": 2.8757425831148134e-05, + "loss": 0.0042, + "step": 65770 + }, + { + "epoch": 10.217458838148493, + "grad_norm": 0.374519944190979, + "learning_rate": 2.874245874139658e-05, + "loss": 0.0061, + "step": 65780 + }, + { + "epoch": 10.21901211556384, + "grad_norm": 0.15389223396778107, + "learning_rate": 2.8727493976337737e-05, + "loss": 0.0044, + "step": 65790 + }, + { + "epoch": 10.220565392979186, + "grad_norm": 0.18421761691570282, + "learning_rate": 2.8712531537608218e-05, + "loss": 0.006, + "step": 65800 + }, + { + "epoch": 10.222118670394533, + "grad_norm": 0.2558740973472595, + "learning_rate": 2.8697571426844217e-05, + "loss": 0.0052, + "step": 65810 + }, + { + "epoch": 10.223671947809878, + "grad_norm": 0.3577716052532196, + "learning_rate": 2.86826136456818e-05, + "loss": 0.0042, + "step": 65820 + }, + { + "epoch": 10.225225225225225, + "grad_norm": 0.13551293313503265, + "learning_rate": 2.8667658195756685e-05, + "loss": 0.0048, + "step": 65830 + }, + { + "epoch": 10.226778502640572, + "grad_norm": 0.14870519936084747, + "learning_rate": 2.8652705078704412e-05, + "loss": 0.0054, + "step": 65840 + }, + { + "epoch": 10.228331780055917, + "grad_norm": 0.12474199384450912, + "learning_rate": 2.863775429616018e-05, + "loss": 0.0046, + "step": 65850 + }, + { + "epoch": 10.229885057471265, + "grad_norm": 0.12301142513751984, + "learning_rate": 2.862280584975906e-05, + "loss": 0.0052, + "step": 65860 + }, + { + "epoch": 10.231438334886612, + "grad_norm": 0.15391463041305542, + "learning_rate": 2.8607859741135735e-05, + "loss": 0.0056, + "step": 65870 + }, + { + "epoch": 10.232991612301957, + "grad_norm": 0.07003511488437653, + "learning_rate": 2.8592915971924723e-05, + "loss": 0.0032, + "step": 65880 + }, + { + "epoch": 10.234544889717304, + "grad_norm": 0.3207812011241913, + "learning_rate": 2.8577974543760212e-05, + "loss": 0.0039, + "step": 65890 + }, + { + "epoch": 10.23609816713265, + "grad_norm": 0.16387788951396942, + "learning_rate": 2.8563035458276222e-05, + "loss": 0.0048, + "step": 65900 + }, + { + "epoch": 10.237651444547996, + "grad_norm": 0.2271236777305603, + "learning_rate": 2.8548098717106424e-05, + "loss": 0.0063, + "step": 65910 + }, + { + "epoch": 10.239204721963343, + "grad_norm": 0.45191702246665955, + "learning_rate": 2.85331643218843e-05, + "loss": 0.0084, + "step": 65920 + }, + { + "epoch": 10.240757999378689, + "grad_norm": 0.2711390554904938, + "learning_rate": 2.8518232274243046e-05, + "loss": 0.0063, + "step": 65930 + }, + { + "epoch": 10.242311276794036, + "grad_norm": 0.09418359398841858, + "learning_rate": 2.8503302575815638e-05, + "loss": 0.0065, + "step": 65940 + }, + { + "epoch": 10.243864554209381, + "grad_norm": 0.7103099226951599, + "learning_rate": 2.8488375228234716e-05, + "loss": 0.0092, + "step": 65950 + }, + { + "epoch": 10.245417831624728, + "grad_norm": 0.26110509037971497, + "learning_rate": 2.8473450233132738e-05, + "loss": 0.0039, + "step": 65960 + }, + { + "epoch": 10.246971109040075, + "grad_norm": 0.5014608502388, + "learning_rate": 2.8458527592141898e-05, + "loss": 0.0065, + "step": 65970 + }, + { + "epoch": 10.24852438645542, + "grad_norm": 0.5949247479438782, + "learning_rate": 2.8443607306894073e-05, + "loss": 0.0051, + "step": 65980 + }, + { + "epoch": 10.250077663870767, + "grad_norm": 0.17197874188423157, + "learning_rate": 2.8428689379020946e-05, + "loss": 0.0054, + "step": 65990 + }, + { + "epoch": 10.251630941286114, + "grad_norm": 0.5660927295684814, + "learning_rate": 2.841377381015391e-05, + "loss": 0.0039, + "step": 66000 + }, + { + "epoch": 10.25318421870146, + "grad_norm": 0.2084108144044876, + "learning_rate": 2.8398860601924144e-05, + "loss": 0.0041, + "step": 66010 + }, + { + "epoch": 10.254737496116807, + "grad_norm": 0.7993617057800293, + "learning_rate": 2.838394975596248e-05, + "loss": 0.0055, + "step": 66020 + }, + { + "epoch": 10.256290773532152, + "grad_norm": 0.15750226378440857, + "learning_rate": 2.83690412738996e-05, + "loss": 0.0047, + "step": 66030 + }, + { + "epoch": 10.257844050947499, + "grad_norm": 0.499311238527298, + "learning_rate": 2.835413515736579e-05, + "loss": 0.0063, + "step": 66040 + }, + { + "epoch": 10.259397328362846, + "grad_norm": 0.371682345867157, + "learning_rate": 2.8339231407991267e-05, + "loss": 0.006, + "step": 66050 + }, + { + "epoch": 10.260950605778191, + "grad_norm": 0.16040760278701782, + "learning_rate": 2.832433002740581e-05, + "loss": 0.0051, + "step": 66060 + }, + { + "epoch": 10.262503883193538, + "grad_norm": 0.13916246592998505, + "learning_rate": 2.8309431017239047e-05, + "loss": 0.004, + "step": 66070 + }, + { + "epoch": 10.264057160608886, + "grad_norm": 0.22225575149059296, + "learning_rate": 2.829453437912028e-05, + "loss": 0.005, + "step": 66080 + }, + { + "epoch": 10.26561043802423, + "grad_norm": 0.27782735228538513, + "learning_rate": 2.8279640114678617e-05, + "loss": 0.0057, + "step": 66090 + }, + { + "epoch": 10.267163715439578, + "grad_norm": 0.06864451617002487, + "learning_rate": 2.8264748225542836e-05, + "loss": 0.0038, + "step": 66100 + }, + { + "epoch": 10.268716992854923, + "grad_norm": 0.33262547850608826, + "learning_rate": 2.8249858713341516e-05, + "loss": 0.0049, + "step": 66110 + }, + { + "epoch": 10.27027027027027, + "grad_norm": 0.2016741931438446, + "learning_rate": 2.823497157970294e-05, + "loss": 0.0055, + "step": 66120 + }, + { + "epoch": 10.271823547685617, + "grad_norm": 0.12010609358549118, + "learning_rate": 2.8220086826255164e-05, + "loss": 0.0044, + "step": 66130 + }, + { + "epoch": 10.273376825100963, + "grad_norm": 0.14581991732120514, + "learning_rate": 2.8205204454625934e-05, + "loss": 0.0039, + "step": 66140 + }, + { + "epoch": 10.27493010251631, + "grad_norm": 0.42337116599082947, + "learning_rate": 2.8190324466442798e-05, + "loss": 0.0051, + "step": 66150 + }, + { + "epoch": 10.276483379931657, + "grad_norm": 0.32168370485305786, + "learning_rate": 2.8175446863332965e-05, + "loss": 0.0048, + "step": 66160 + }, + { + "epoch": 10.278036657347002, + "grad_norm": 0.2825300395488739, + "learning_rate": 2.816057164692345e-05, + "loss": 0.0066, + "step": 66170 + }, + { + "epoch": 10.279589934762349, + "grad_norm": 0.24760600924491882, + "learning_rate": 2.8145698818840992e-05, + "loss": 0.0064, + "step": 66180 + }, + { + "epoch": 10.281143212177694, + "grad_norm": 0.13329121470451355, + "learning_rate": 2.8130828380712065e-05, + "loss": 0.0041, + "step": 66190 + }, + { + "epoch": 10.282696489593041, + "grad_norm": 0.10546567291021347, + "learning_rate": 2.811596033416285e-05, + "loss": 0.0048, + "step": 66200 + }, + { + "epoch": 10.284249767008388, + "grad_norm": 0.3045012056827545, + "learning_rate": 2.8101094680819307e-05, + "loss": 0.0051, + "step": 66210 + }, + { + "epoch": 10.285803044423734, + "grad_norm": 0.3177582621574402, + "learning_rate": 2.808623142230715e-05, + "loss": 0.0068, + "step": 66220 + }, + { + "epoch": 10.28735632183908, + "grad_norm": 0.7662670612335205, + "learning_rate": 2.8071370560251757e-05, + "loss": 0.0054, + "step": 66230 + }, + { + "epoch": 10.288909599254428, + "grad_norm": 0.13681215047836304, + "learning_rate": 2.805651209627831e-05, + "loss": 0.0072, + "step": 66240 + }, + { + "epoch": 10.290462876669773, + "grad_norm": 0.18272511661052704, + "learning_rate": 2.804165603201171e-05, + "loss": 0.0045, + "step": 66250 + }, + { + "epoch": 10.29201615408512, + "grad_norm": 0.19691066443920135, + "learning_rate": 2.802680236907661e-05, + "loss": 0.0064, + "step": 66260 + }, + { + "epoch": 10.293569431500465, + "grad_norm": 0.17011450231075287, + "learning_rate": 2.8011951109097344e-05, + "loss": 0.0029, + "step": 66270 + }, + { + "epoch": 10.295122708915812, + "grad_norm": 0.587195098400116, + "learning_rate": 2.7997102253698072e-05, + "loss": 0.0081, + "step": 66280 + }, + { + "epoch": 10.29667598633116, + "grad_norm": 0.1776575744152069, + "learning_rate": 2.7982255804502588e-05, + "loss": 0.0039, + "step": 66290 + }, + { + "epoch": 10.298229263746505, + "grad_norm": 0.35295289754867554, + "learning_rate": 2.7967411763134516e-05, + "loss": 0.0049, + "step": 66300 + }, + { + "epoch": 10.299782541161852, + "grad_norm": 0.27765795588493347, + "learning_rate": 2.7952570131217166e-05, + "loss": 0.0036, + "step": 66310 + }, + { + "epoch": 10.301335818577197, + "grad_norm": 0.31917017698287964, + "learning_rate": 2.7937730910373618e-05, + "loss": 0.0055, + "step": 66320 + }, + { + "epoch": 10.302889095992544, + "grad_norm": 0.2506464421749115, + "learning_rate": 2.7922894102226628e-05, + "loss": 0.0045, + "step": 66330 + }, + { + "epoch": 10.304442373407891, + "grad_norm": 0.34828072786331177, + "learning_rate": 2.790805970839876e-05, + "loss": 0.0071, + "step": 66340 + }, + { + "epoch": 10.305995650823236, + "grad_norm": 0.27521464228630066, + "learning_rate": 2.789322773051225e-05, + "loss": 0.0061, + "step": 66350 + }, + { + "epoch": 10.307548928238583, + "grad_norm": 0.2529260218143463, + "learning_rate": 2.7878398170189124e-05, + "loss": 0.0062, + "step": 66360 + }, + { + "epoch": 10.30910220565393, + "grad_norm": 0.14285089075565338, + "learning_rate": 2.7863571029051112e-05, + "loss": 0.0059, + "step": 66370 + }, + { + "epoch": 10.310655483069276, + "grad_norm": 0.5401323437690735, + "learning_rate": 2.784874630871971e-05, + "loss": 0.005, + "step": 66380 + }, + { + "epoch": 10.312208760484623, + "grad_norm": 0.16723528504371643, + "learning_rate": 2.7833924010816086e-05, + "loss": 0.0051, + "step": 66390 + }, + { + "epoch": 10.313762037899968, + "grad_norm": 0.1321493685245514, + "learning_rate": 2.7819104136961226e-05, + "loss": 0.0073, + "step": 66400 + }, + { + "epoch": 10.315315315315315, + "grad_norm": 0.5096457600593567, + "learning_rate": 2.780428668877577e-05, + "loss": 0.0063, + "step": 66410 + }, + { + "epoch": 10.316868592730662, + "grad_norm": 0.18283596634864807, + "learning_rate": 2.7789471667880152e-05, + "loss": 0.0051, + "step": 66420 + }, + { + "epoch": 10.318421870146008, + "grad_norm": 0.8414400219917297, + "learning_rate": 2.7774659075894523e-05, + "loss": 0.0067, + "step": 66430 + }, + { + "epoch": 10.319975147561355, + "grad_norm": 0.3607831597328186, + "learning_rate": 2.7759848914438778e-05, + "loss": 0.0053, + "step": 66440 + }, + { + "epoch": 10.321528424976702, + "grad_norm": 0.3289492130279541, + "learning_rate": 2.7745041185132496e-05, + "loss": 0.0067, + "step": 66450 + }, + { + "epoch": 10.323081702392047, + "grad_norm": 0.3100898861885071, + "learning_rate": 2.773023588959506e-05, + "loss": 0.0053, + "step": 66460 + }, + { + "epoch": 10.324634979807394, + "grad_norm": 0.19352245330810547, + "learning_rate": 2.7715433029445555e-05, + "loss": 0.0034, + "step": 66470 + }, + { + "epoch": 10.32618825722274, + "grad_norm": 0.4773525297641754, + "learning_rate": 2.7700632606302766e-05, + "loss": 0.0055, + "step": 66480 + }, + { + "epoch": 10.327741534638086, + "grad_norm": 0.13646526634693146, + "learning_rate": 2.768583462178528e-05, + "loss": 0.0063, + "step": 66490 + }, + { + "epoch": 10.329294812053433, + "grad_norm": 0.13288235664367676, + "learning_rate": 2.767103907751136e-05, + "loss": 0.0059, + "step": 66500 + }, + { + "epoch": 10.330848089468779, + "grad_norm": 0.14999686181545258, + "learning_rate": 2.7656245975099065e-05, + "loss": 0.0065, + "step": 66510 + }, + { + "epoch": 10.332401366884126, + "grad_norm": 0.30029189586639404, + "learning_rate": 2.7641455316166093e-05, + "loss": 0.0053, + "step": 66520 + }, + { + "epoch": 10.333954644299471, + "grad_norm": 0.28949183225631714, + "learning_rate": 2.7626667102329973e-05, + "loss": 0.0036, + "step": 66530 + }, + { + "epoch": 10.335507921714818, + "grad_norm": 0.12697038054466248, + "learning_rate": 2.761188133520788e-05, + "loss": 0.0036, + "step": 66540 + }, + { + "epoch": 10.337061199130165, + "grad_norm": 0.7485699653625488, + "learning_rate": 2.7597098016416784e-05, + "loss": 0.0062, + "step": 66550 + }, + { + "epoch": 10.33861447654551, + "grad_norm": 0.31732606887817383, + "learning_rate": 2.7582317147573368e-05, + "loss": 0.0044, + "step": 66560 + }, + { + "epoch": 10.340167753960857, + "grad_norm": 0.28917354345321655, + "learning_rate": 2.7567538730294064e-05, + "loss": 0.0062, + "step": 66570 + }, + { + "epoch": 10.341721031376204, + "grad_norm": 0.2011614590883255, + "learning_rate": 2.7552762766194973e-05, + "loss": 0.0044, + "step": 66580 + }, + { + "epoch": 10.34327430879155, + "grad_norm": 0.3181355595588684, + "learning_rate": 2.753798925689202e-05, + "loss": 0.005, + "step": 66590 + }, + { + "epoch": 10.344827586206897, + "grad_norm": 0.20104752480983734, + "learning_rate": 2.7523218204000774e-05, + "loss": 0.0052, + "step": 66600 + }, + { + "epoch": 10.346380863622242, + "grad_norm": 0.29103824496269226, + "learning_rate": 2.7508449609136612e-05, + "loss": 0.0065, + "step": 66610 + }, + { + "epoch": 10.34793414103759, + "grad_norm": 0.4764154553413391, + "learning_rate": 2.749368347391454e-05, + "loss": 0.0051, + "step": 66620 + }, + { + "epoch": 10.349487418452936, + "grad_norm": 0.11493270844221115, + "learning_rate": 2.7478919799949443e-05, + "loss": 0.0055, + "step": 66630 + }, + { + "epoch": 10.351040695868281, + "grad_norm": 0.20265647768974304, + "learning_rate": 2.7464158588855803e-05, + "loss": 0.0059, + "step": 66640 + }, + { + "epoch": 10.352593973283629, + "grad_norm": 0.26151302456855774, + "learning_rate": 2.7449399842247918e-05, + "loss": 0.004, + "step": 66650 + }, + { + "epoch": 10.354147250698976, + "grad_norm": 0.42146363854408264, + "learning_rate": 2.7434643561739736e-05, + "loss": 0.0062, + "step": 66660 + }, + { + "epoch": 10.35570052811432, + "grad_norm": 0.17594602704048157, + "learning_rate": 2.7419889748945016e-05, + "loss": 0.0052, + "step": 66670 + }, + { + "epoch": 10.357253805529668, + "grad_norm": 0.3985117971897125, + "learning_rate": 2.7405138405477203e-05, + "loss": 0.0057, + "step": 66680 + }, + { + "epoch": 10.358807082945013, + "grad_norm": 0.29277920722961426, + "learning_rate": 2.73903895329495e-05, + "loss": 0.0055, + "step": 66690 + }, + { + "epoch": 10.36036036036036, + "grad_norm": 0.3036515414714813, + "learning_rate": 2.737564313297478e-05, + "loss": 0.005, + "step": 66700 + }, + { + "epoch": 10.361913637775707, + "grad_norm": 0.18209078907966614, + "learning_rate": 2.736089920716572e-05, + "loss": 0.0059, + "step": 66710 + }, + { + "epoch": 10.363466915191053, + "grad_norm": 0.20009154081344604, + "learning_rate": 2.7346157757134706e-05, + "loss": 0.0052, + "step": 66720 + }, + { + "epoch": 10.3650201926064, + "grad_norm": 0.26024970412254333, + "learning_rate": 2.733141878449379e-05, + "loss": 0.0064, + "step": 66730 + }, + { + "epoch": 10.366573470021747, + "grad_norm": 0.19367924332618713, + "learning_rate": 2.7316682290854844e-05, + "loss": 0.0053, + "step": 66740 + }, + { + "epoch": 10.368126747437092, + "grad_norm": 0.14338865876197815, + "learning_rate": 2.7301948277829414e-05, + "loss": 0.0056, + "step": 66750 + }, + { + "epoch": 10.369680024852439, + "grad_norm": 0.14995427429676056, + "learning_rate": 2.7287216747028814e-05, + "loss": 0.0041, + "step": 66760 + }, + { + "epoch": 10.371233302267784, + "grad_norm": 0.30748915672302246, + "learning_rate": 2.7272487700064025e-05, + "loss": 0.0041, + "step": 66770 + }, + { + "epoch": 10.372786579683131, + "grad_norm": 0.2541109621524811, + "learning_rate": 2.725776113854583e-05, + "loss": 0.0039, + "step": 66780 + }, + { + "epoch": 10.374339857098478, + "grad_norm": 0.28049248456954956, + "learning_rate": 2.7243037064084664e-05, + "loss": 0.0045, + "step": 66790 + }, + { + "epoch": 10.375893134513824, + "grad_norm": 0.2534404993057251, + "learning_rate": 2.7228315478290766e-05, + "loss": 0.0046, + "step": 66800 + }, + { + "epoch": 10.37744641192917, + "grad_norm": 0.12593959271907806, + "learning_rate": 2.721359638277402e-05, + "loss": 0.0049, + "step": 66810 + }, + { + "epoch": 10.378999689344518, + "grad_norm": 0.3066154420375824, + "learning_rate": 2.7198879779144148e-05, + "loss": 0.0034, + "step": 66820 + }, + { + "epoch": 10.380552966759863, + "grad_norm": 0.15268729627132416, + "learning_rate": 2.7184165669010485e-05, + "loss": 0.0044, + "step": 66830 + }, + { + "epoch": 10.38210624417521, + "grad_norm": 0.36532899737358093, + "learning_rate": 2.716945405398218e-05, + "loss": 0.0042, + "step": 66840 + }, + { + "epoch": 10.383659521590555, + "grad_norm": 0.15624333918094635, + "learning_rate": 2.715474493566803e-05, + "loss": 0.0056, + "step": 66850 + }, + { + "epoch": 10.385212799005902, + "grad_norm": 0.2573741674423218, + "learning_rate": 2.7140038315676652e-05, + "loss": 0.0083, + "step": 66860 + }, + { + "epoch": 10.38676607642125, + "grad_norm": 0.2589920163154602, + "learning_rate": 2.7125334195616275e-05, + "loss": 0.0067, + "step": 66870 + }, + { + "epoch": 10.388319353836595, + "grad_norm": 0.2269631028175354, + "learning_rate": 2.7110632577094997e-05, + "loss": 0.0068, + "step": 66880 + }, + { + "epoch": 10.389872631251942, + "grad_norm": 0.16564595699310303, + "learning_rate": 2.709593346172051e-05, + "loss": 0.0046, + "step": 66890 + }, + { + "epoch": 10.391425908667287, + "grad_norm": 0.15767455101013184, + "learning_rate": 2.708123685110032e-05, + "loss": 0.0051, + "step": 66900 + }, + { + "epoch": 10.392979186082634, + "grad_norm": 0.2578776478767395, + "learning_rate": 2.7066542746841588e-05, + "loss": 0.0051, + "step": 66910 + }, + { + "epoch": 10.394532463497981, + "grad_norm": 0.21218594908714294, + "learning_rate": 2.7051851150551267e-05, + "loss": 0.0036, + "step": 66920 + }, + { + "epoch": 10.396085740913326, + "grad_norm": 0.2708873450756073, + "learning_rate": 2.7037162063835996e-05, + "loss": 0.0084, + "step": 66930 + }, + { + "epoch": 10.397639018328674, + "grad_norm": 0.22806987166404724, + "learning_rate": 2.702247548830218e-05, + "loss": 0.0052, + "step": 66940 + }, + { + "epoch": 10.39919229574402, + "grad_norm": 0.3053355813026428, + "learning_rate": 2.7007791425555883e-05, + "loss": 0.0058, + "step": 66950 + }, + { + "epoch": 10.400745573159366, + "grad_norm": 0.5849961638450623, + "learning_rate": 2.6993109877202945e-05, + "loss": 0.0064, + "step": 66960 + }, + { + "epoch": 10.402298850574713, + "grad_norm": 0.585038959980011, + "learning_rate": 2.6978430844848935e-05, + "loss": 0.0049, + "step": 66970 + }, + { + "epoch": 10.403852127990058, + "grad_norm": 0.15306924283504486, + "learning_rate": 2.69637543300991e-05, + "loss": 0.0056, + "step": 66980 + }, + { + "epoch": 10.405405405405405, + "grad_norm": 1.3888376951217651, + "learning_rate": 2.694908033455848e-05, + "loss": 0.0045, + "step": 66990 + }, + { + "epoch": 10.406958682820752, + "grad_norm": 0.13783180713653564, + "learning_rate": 2.6934408859831733e-05, + "loss": 0.0052, + "step": 67000 + }, + { + "epoch": 10.408511960236098, + "grad_norm": 0.213615283370018, + "learning_rate": 2.69197399075234e-05, + "loss": 0.0032, + "step": 67010 + }, + { + "epoch": 10.410065237651445, + "grad_norm": 0.1479993313550949, + "learning_rate": 2.6905073479237584e-05, + "loss": 0.0058, + "step": 67020 + }, + { + "epoch": 10.411618515066792, + "grad_norm": 0.17695792019367218, + "learning_rate": 2.6890409576578235e-05, + "loss": 0.0041, + "step": 67030 + }, + { + "epoch": 10.413171792482137, + "grad_norm": 0.2726774513721466, + "learning_rate": 2.687574820114892e-05, + "loss": 0.0048, + "step": 67040 + }, + { + "epoch": 10.414725069897484, + "grad_norm": 0.1475217640399933, + "learning_rate": 2.686108935455305e-05, + "loss": 0.0038, + "step": 67050 + }, + { + "epoch": 10.41627834731283, + "grad_norm": 0.23769395053386688, + "learning_rate": 2.6846433038393603e-05, + "loss": 0.006, + "step": 67060 + }, + { + "epoch": 10.417831624728176, + "grad_norm": 0.14208035171031952, + "learning_rate": 2.6831779254273482e-05, + "loss": 0.0047, + "step": 67070 + }, + { + "epoch": 10.419384902143523, + "grad_norm": 0.3564046323299408, + "learning_rate": 2.6817128003795122e-05, + "loss": 0.0048, + "step": 67080 + }, + { + "epoch": 10.420938179558869, + "grad_norm": 0.39307382702827454, + "learning_rate": 2.680247928856081e-05, + "loss": 0.0041, + "step": 67090 + }, + { + "epoch": 10.422491456974216, + "grad_norm": 0.3574706017971039, + "learning_rate": 2.6787833110172466e-05, + "loss": 0.0061, + "step": 67100 + }, + { + "epoch": 10.424044734389563, + "grad_norm": 0.7312748432159424, + "learning_rate": 2.6773189470231807e-05, + "loss": 0.0064, + "step": 67110 + }, + { + "epoch": 10.425598011804908, + "grad_norm": 0.4445406496524811, + "learning_rate": 2.675854837034022e-05, + "loss": 0.0044, + "step": 67120 + }, + { + "epoch": 10.427151289220255, + "grad_norm": 0.18776719272136688, + "learning_rate": 2.6743909812098833e-05, + "loss": 0.0063, + "step": 67130 + }, + { + "epoch": 10.4287045666356, + "grad_norm": 0.20867684483528137, + "learning_rate": 2.67292737971085e-05, + "loss": 0.0044, + "step": 67140 + }, + { + "epoch": 10.430257844050947, + "grad_norm": 0.5187969207763672, + "learning_rate": 2.671464032696982e-05, + "loss": 0.0041, + "step": 67150 + }, + { + "epoch": 10.431811121466295, + "grad_norm": 0.12212332338094711, + "learning_rate": 2.6700009403283055e-05, + "loss": 0.0038, + "step": 67160 + }, + { + "epoch": 10.43336439888164, + "grad_norm": 0.6536049842834473, + "learning_rate": 2.6685381027648214e-05, + "loss": 0.0053, + "step": 67170 + }, + { + "epoch": 10.434917676296987, + "grad_norm": 0.11695326864719391, + "learning_rate": 2.6670755201665086e-05, + "loss": 0.0058, + "step": 67180 + }, + { + "epoch": 10.436470953712334, + "grad_norm": 1.3422852754592896, + "learning_rate": 2.6656131926933066e-05, + "loss": 0.0051, + "step": 67190 + }, + { + "epoch": 10.43802423112768, + "grad_norm": 0.23144131898880005, + "learning_rate": 2.664151120505136e-05, + "loss": 0.0055, + "step": 67200 + }, + { + "epoch": 10.439577508543026, + "grad_norm": 0.4979749023914337, + "learning_rate": 2.6626893037618873e-05, + "loss": 0.0061, + "step": 67210 + }, + { + "epoch": 10.441130785958372, + "grad_norm": 0.20595087110996246, + "learning_rate": 2.6612277426234243e-05, + "loss": 0.0049, + "step": 67220 + }, + { + "epoch": 10.442684063373719, + "grad_norm": 0.11032645404338837, + "learning_rate": 2.659766437249577e-05, + "loss": 0.0061, + "step": 67230 + }, + { + "epoch": 10.444237340789066, + "grad_norm": 0.23140110075473785, + "learning_rate": 2.6583053878001557e-05, + "loss": 0.0042, + "step": 67240 + }, + { + "epoch": 10.445790618204411, + "grad_norm": 0.4339185059070587, + "learning_rate": 2.6568445944349323e-05, + "loss": 0.0073, + "step": 67250 + }, + { + "epoch": 10.447343895619758, + "grad_norm": 0.20317348837852478, + "learning_rate": 2.6553840573136657e-05, + "loss": 0.0056, + "step": 67260 + }, + { + "epoch": 10.448897173035103, + "grad_norm": 0.3513525128364563, + "learning_rate": 2.6539237765960706e-05, + "loss": 0.0059, + "step": 67270 + }, + { + "epoch": 10.45045045045045, + "grad_norm": 0.49508875608444214, + "learning_rate": 2.6524637524418465e-05, + "loss": 0.0047, + "step": 67280 + }, + { + "epoch": 10.452003727865797, + "grad_norm": 0.2458571344614029, + "learning_rate": 2.651003985010655e-05, + "loss": 0.0095, + "step": 67290 + }, + { + "epoch": 10.453557005281143, + "grad_norm": 0.16129089891910553, + "learning_rate": 2.649544474462138e-05, + "loss": 0.0028, + "step": 67300 + }, + { + "epoch": 10.45511028269649, + "grad_norm": 0.4087439477443695, + "learning_rate": 2.6480852209559015e-05, + "loss": 0.0038, + "step": 67310 + }, + { + "epoch": 10.456663560111837, + "grad_norm": 0.0998622477054596, + "learning_rate": 2.6466262246515296e-05, + "loss": 0.0048, + "step": 67320 + }, + { + "epoch": 10.458216837527182, + "grad_norm": 0.1759682595729828, + "learning_rate": 2.6451674857085746e-05, + "loss": 0.0053, + "step": 67330 + }, + { + "epoch": 10.459770114942529, + "grad_norm": 0.19053873419761658, + "learning_rate": 2.6437090042865655e-05, + "loss": 0.005, + "step": 67340 + }, + { + "epoch": 10.461323392357874, + "grad_norm": 0.1065826416015625, + "learning_rate": 2.6422507805449947e-05, + "loss": 0.0037, + "step": 67350 + }, + { + "epoch": 10.462876669773221, + "grad_norm": 0.14576059579849243, + "learning_rate": 2.6407928146433357e-05, + "loss": 0.0054, + "step": 67360 + }, + { + "epoch": 10.464429947188568, + "grad_norm": 0.2612968683242798, + "learning_rate": 2.639335106741026e-05, + "loss": 0.0055, + "step": 67370 + }, + { + "epoch": 10.465983224603914, + "grad_norm": 0.14460988342761993, + "learning_rate": 2.6378776569974796e-05, + "loss": 0.0037, + "step": 67380 + }, + { + "epoch": 10.46753650201926, + "grad_norm": 0.24884584546089172, + "learning_rate": 2.6364204655720814e-05, + "loss": 0.0052, + "step": 67390 + }, + { + "epoch": 10.469089779434608, + "grad_norm": 0.3772946894168854, + "learning_rate": 2.6349635326241894e-05, + "loss": 0.006, + "step": 67400 + }, + { + "epoch": 10.470643056849953, + "grad_norm": 0.24948644638061523, + "learning_rate": 2.633506858313127e-05, + "loss": 0.0045, + "step": 67410 + }, + { + "epoch": 10.4721963342653, + "grad_norm": 0.12983843684196472, + "learning_rate": 2.6320504427981975e-05, + "loss": 0.0045, + "step": 67420 + }, + { + "epoch": 10.473749611680645, + "grad_norm": 0.10384825617074966, + "learning_rate": 2.6305942862386734e-05, + "loss": 0.0057, + "step": 67430 + }, + { + "epoch": 10.475302889095992, + "grad_norm": 0.24637609720230103, + "learning_rate": 2.629138388793793e-05, + "loss": 0.0053, + "step": 67440 + }, + { + "epoch": 10.47685616651134, + "grad_norm": 0.1986657679080963, + "learning_rate": 2.6276827506227743e-05, + "loss": 0.0036, + "step": 67450 + }, + { + "epoch": 10.478409443926685, + "grad_norm": 0.18366943299770355, + "learning_rate": 2.6262273718848037e-05, + "loss": 0.0039, + "step": 67460 + }, + { + "epoch": 10.479962721342032, + "grad_norm": 0.14056985080242157, + "learning_rate": 2.6247722527390407e-05, + "loss": 0.0055, + "step": 67470 + }, + { + "epoch": 10.481515998757377, + "grad_norm": 0.24850790202617645, + "learning_rate": 2.6233173933446104e-05, + "loss": 0.0054, + "step": 67480 + }, + { + "epoch": 10.483069276172724, + "grad_norm": 0.11449657380580902, + "learning_rate": 2.6218627938606188e-05, + "loss": 0.0068, + "step": 67490 + }, + { + "epoch": 10.484622553588071, + "grad_norm": 0.7842130064964294, + "learning_rate": 2.620408454446135e-05, + "loss": 0.0035, + "step": 67500 + }, + { + "epoch": 10.486175831003417, + "grad_norm": 0.2613527178764343, + "learning_rate": 2.6189543752602046e-05, + "loss": 0.0047, + "step": 67510 + }, + { + "epoch": 10.487729108418764, + "grad_norm": 0.2010316103696823, + "learning_rate": 2.6175005564618444e-05, + "loss": 0.0047, + "step": 67520 + }, + { + "epoch": 10.48928238583411, + "grad_norm": 0.24303196370601654, + "learning_rate": 2.6160469982100428e-05, + "loss": 0.0051, + "step": 67530 + }, + { + "epoch": 10.490835663249456, + "grad_norm": 0.17032557725906372, + "learning_rate": 2.6145937006637555e-05, + "loss": 0.0039, + "step": 67540 + }, + { + "epoch": 10.492388940664803, + "grad_norm": 0.15870767831802368, + "learning_rate": 2.6131406639819166e-05, + "loss": 0.0061, + "step": 67550 + }, + { + "epoch": 10.493942218080148, + "grad_norm": 0.3747347593307495, + "learning_rate": 2.611687888323424e-05, + "loss": 0.0066, + "step": 67560 + }, + { + "epoch": 10.495495495495495, + "grad_norm": 0.24972331523895264, + "learning_rate": 2.610235373847153e-05, + "loss": 0.0048, + "step": 67570 + }, + { + "epoch": 10.497048772910842, + "grad_norm": 0.12024827301502228, + "learning_rate": 2.6087831207119494e-05, + "loss": 0.0082, + "step": 67580 + }, + { + "epoch": 10.498602050326188, + "grad_norm": 0.2754182815551758, + "learning_rate": 2.60733112907663e-05, + "loss": 0.0049, + "step": 67590 + }, + { + "epoch": 10.500155327741535, + "grad_norm": 0.13661246001720428, + "learning_rate": 2.6058793990999786e-05, + "loss": 0.0055, + "step": 67600 + }, + { + "epoch": 10.501708605156882, + "grad_norm": 0.11282090097665787, + "learning_rate": 2.6044279309407593e-05, + "loss": 0.0046, + "step": 67610 + }, + { + "epoch": 10.503261882572227, + "grad_norm": 0.39383092522621155, + "learning_rate": 2.6029767247576976e-05, + "loss": 0.0036, + "step": 67620 + }, + { + "epoch": 10.504815159987574, + "grad_norm": 0.14076246321201324, + "learning_rate": 2.601525780709497e-05, + "loss": 0.0057, + "step": 67630 + }, + { + "epoch": 10.50636843740292, + "grad_norm": 0.292758971452713, + "learning_rate": 2.6000750989548318e-05, + "loss": 0.0048, + "step": 67640 + }, + { + "epoch": 10.507921714818266, + "grad_norm": 0.273561030626297, + "learning_rate": 2.5986246796523462e-05, + "loss": 0.0058, + "step": 67650 + }, + { + "epoch": 10.509474992233613, + "grad_norm": 0.48596855998039246, + "learning_rate": 2.5971745229606546e-05, + "loss": 0.0043, + "step": 67660 + }, + { + "epoch": 10.511028269648959, + "grad_norm": 0.1482127457857132, + "learning_rate": 2.5957246290383442e-05, + "loss": 0.0045, + "step": 67670 + }, + { + "epoch": 10.512581547064306, + "grad_norm": 0.09889166802167892, + "learning_rate": 2.594274998043975e-05, + "loss": 0.0054, + "step": 67680 + }, + { + "epoch": 10.514134824479653, + "grad_norm": 0.22722981870174408, + "learning_rate": 2.5928256301360733e-05, + "loss": 0.0048, + "step": 67690 + }, + { + "epoch": 10.515688101894998, + "grad_norm": 0.20110508799552917, + "learning_rate": 2.591376525473142e-05, + "loss": 0.0035, + "step": 67700 + }, + { + "epoch": 10.517241379310345, + "grad_norm": 0.15923359990119934, + "learning_rate": 2.5899276842136523e-05, + "loss": 0.0072, + "step": 67710 + }, + { + "epoch": 10.51879465672569, + "grad_norm": 0.14212282001972198, + "learning_rate": 2.5884791065160495e-05, + "loss": 0.0033, + "step": 67720 + }, + { + "epoch": 10.520347934141038, + "grad_norm": 0.18993431329727173, + "learning_rate": 2.5870307925387436e-05, + "loss": 0.0043, + "step": 67730 + }, + { + "epoch": 10.521901211556385, + "grad_norm": 0.3884390592575073, + "learning_rate": 2.585582742440125e-05, + "loss": 0.0039, + "step": 67740 + }, + { + "epoch": 10.52345448897173, + "grad_norm": 0.11328709870576859, + "learning_rate": 2.5841349563785455e-05, + "loss": 0.004, + "step": 67750 + }, + { + "epoch": 10.525007766387077, + "grad_norm": 0.2561798691749573, + "learning_rate": 2.5826874345123355e-05, + "loss": 0.0039, + "step": 67760 + }, + { + "epoch": 10.526561043802424, + "grad_norm": 0.2557944655418396, + "learning_rate": 2.581240176999792e-05, + "loss": 0.0071, + "step": 67770 + }, + { + "epoch": 10.52811432121777, + "grad_norm": 0.25791797041893005, + "learning_rate": 2.579793183999189e-05, + "loss": 0.0048, + "step": 67780 + }, + { + "epoch": 10.529667598633116, + "grad_norm": 0.33658307790756226, + "learning_rate": 2.5783464556687627e-05, + "loss": 0.0053, + "step": 67790 + }, + { + "epoch": 10.531220876048462, + "grad_norm": 0.2411465048789978, + "learning_rate": 2.576899992166729e-05, + "loss": 0.0067, + "step": 67800 + }, + { + "epoch": 10.532774153463809, + "grad_norm": 0.2467084676027298, + "learning_rate": 2.5754537936512668e-05, + "loss": 0.0042, + "step": 67810 + }, + { + "epoch": 10.534327430879156, + "grad_norm": 0.10473160445690155, + "learning_rate": 2.574007860280535e-05, + "loss": 0.0065, + "step": 67820 + }, + { + "epoch": 10.535880708294501, + "grad_norm": 0.22629843652248383, + "learning_rate": 2.5725621922126518e-05, + "loss": 0.0048, + "step": 67830 + }, + { + "epoch": 10.537433985709848, + "grad_norm": 0.09456472098827362, + "learning_rate": 2.571116789605722e-05, + "loss": 0.0037, + "step": 67840 + }, + { + "epoch": 10.538987263125193, + "grad_norm": 0.22799460589885712, + "learning_rate": 2.5696716526178054e-05, + "loss": 0.0038, + "step": 67850 + }, + { + "epoch": 10.54054054054054, + "grad_norm": 0.25493237376213074, + "learning_rate": 2.568226781406946e-05, + "loss": 0.0048, + "step": 67860 + }, + { + "epoch": 10.542093817955887, + "grad_norm": 0.5016720294952393, + "learning_rate": 2.5667821761311474e-05, + "loss": 0.0043, + "step": 67870 + }, + { + "epoch": 10.543647095371233, + "grad_norm": 0.13123422861099243, + "learning_rate": 2.5653378369483917e-05, + "loss": 0.0071, + "step": 67880 + }, + { + "epoch": 10.54520037278658, + "grad_norm": 0.22147837281227112, + "learning_rate": 2.5638937640166292e-05, + "loss": 0.0057, + "step": 67890 + }, + { + "epoch": 10.546753650201927, + "grad_norm": 0.1590048372745514, + "learning_rate": 2.5624499574937842e-05, + "loss": 0.0038, + "step": 67900 + }, + { + "epoch": 10.548306927617272, + "grad_norm": 0.14298158884048462, + "learning_rate": 2.5610064175377456e-05, + "loss": 0.0041, + "step": 67910 + }, + { + "epoch": 10.54986020503262, + "grad_norm": 0.1151859238743782, + "learning_rate": 2.5595631443063782e-05, + "loss": 0.0049, + "step": 67920 + }, + { + "epoch": 10.551413482447964, + "grad_norm": 0.2243361473083496, + "learning_rate": 2.5581201379575177e-05, + "loss": 0.0039, + "step": 67930 + }, + { + "epoch": 10.552966759863311, + "grad_norm": 0.11172261089086533, + "learning_rate": 2.5566773986489655e-05, + "loss": 0.0047, + "step": 67940 + }, + { + "epoch": 10.554520037278659, + "grad_norm": 0.33141058683395386, + "learning_rate": 2.5552349265385e-05, + "loss": 0.0046, + "step": 67950 + }, + { + "epoch": 10.556073314694004, + "grad_norm": 0.3249967396259308, + "learning_rate": 2.5537927217838675e-05, + "loss": 0.0048, + "step": 67960 + }, + { + "epoch": 10.55762659210935, + "grad_norm": 0.14301510155200958, + "learning_rate": 2.5523507845427864e-05, + "loss": 0.0044, + "step": 67970 + }, + { + "epoch": 10.559179869524698, + "grad_norm": 0.5435895919799805, + "learning_rate": 2.5509091149729413e-05, + "loss": 0.004, + "step": 67980 + }, + { + "epoch": 10.560733146940043, + "grad_norm": 0.11147929728031158, + "learning_rate": 2.5494677132319962e-05, + "loss": 0.0035, + "step": 67990 + }, + { + "epoch": 10.56228642435539, + "grad_norm": 0.42667466402053833, + "learning_rate": 2.548026579477575e-05, + "loss": 0.0043, + "step": 68000 + }, + { + "epoch": 10.563839701770736, + "grad_norm": 0.20991429686546326, + "learning_rate": 2.5465857138672834e-05, + "loss": 0.0059, + "step": 68010 + }, + { + "epoch": 10.565392979186083, + "grad_norm": 0.29275622963905334, + "learning_rate": 2.5451451165586848e-05, + "loss": 0.006, + "step": 68020 + }, + { + "epoch": 10.56694625660143, + "grad_norm": 0.25718799233436584, + "learning_rate": 2.543704787709329e-05, + "loss": 0.0044, + "step": 68030 + }, + { + "epoch": 10.568499534016775, + "grad_norm": 0.16858850419521332, + "learning_rate": 2.542264727476723e-05, + "loss": 0.0052, + "step": 68040 + }, + { + "epoch": 10.570052811432122, + "grad_norm": 0.21609915792942047, + "learning_rate": 2.5408249360183535e-05, + "loss": 0.006, + "step": 68050 + }, + { + "epoch": 10.571606088847467, + "grad_norm": 0.258887380361557, + "learning_rate": 2.5393854134916694e-05, + "loss": 0.0044, + "step": 68060 + }, + { + "epoch": 10.573159366262814, + "grad_norm": 0.37028801441192627, + "learning_rate": 2.5379461600540993e-05, + "loss": 0.0076, + "step": 68070 + }, + { + "epoch": 10.574712643678161, + "grad_norm": 0.18530789017677307, + "learning_rate": 2.5365071758630317e-05, + "loss": 0.0029, + "step": 68080 + }, + { + "epoch": 10.576265921093507, + "grad_norm": 0.22560442984104156, + "learning_rate": 2.535068461075839e-05, + "loss": 0.0047, + "step": 68090 + }, + { + "epoch": 10.577819198508854, + "grad_norm": 0.3099275827407837, + "learning_rate": 2.5336300158498516e-05, + "loss": 0.0061, + "step": 68100 + }, + { + "epoch": 10.5793724759242, + "grad_norm": 0.3806454539299011, + "learning_rate": 2.5321918403423793e-05, + "loss": 0.0055, + "step": 68110 + }, + { + "epoch": 10.580925753339546, + "grad_norm": 0.21029001474380493, + "learning_rate": 2.5307539347106957e-05, + "loss": 0.0033, + "step": 68120 + }, + { + "epoch": 10.582479030754893, + "grad_norm": 0.24656248092651367, + "learning_rate": 2.529316299112049e-05, + "loss": 0.0048, + "step": 68130 + }, + { + "epoch": 10.58403230817024, + "grad_norm": 0.25792622566223145, + "learning_rate": 2.5278789337036567e-05, + "loss": 0.0062, + "step": 68140 + }, + { + "epoch": 10.585585585585585, + "grad_norm": 0.1817338615655899, + "learning_rate": 2.52644183864271e-05, + "loss": 0.0038, + "step": 68150 + }, + { + "epoch": 10.587138863000932, + "grad_norm": 0.28489041328430176, + "learning_rate": 2.525005014086363e-05, + "loss": 0.0041, + "step": 68160 + }, + { + "epoch": 10.588692140416278, + "grad_norm": 0.32640430331230164, + "learning_rate": 2.5235684601917465e-05, + "loss": 0.0071, + "step": 68170 + }, + { + "epoch": 10.590245417831625, + "grad_norm": 0.2870074510574341, + "learning_rate": 2.5221321771159618e-05, + "loss": 0.0061, + "step": 68180 + }, + { + "epoch": 10.591798695246972, + "grad_norm": 0.2672390937805176, + "learning_rate": 2.5206961650160743e-05, + "loss": 0.0044, + "step": 68190 + }, + { + "epoch": 10.593351972662317, + "grad_norm": 0.8655813932418823, + "learning_rate": 2.519260424049129e-05, + "loss": 0.0044, + "step": 68200 + }, + { + "epoch": 10.594905250077664, + "grad_norm": 0.07937179505825043, + "learning_rate": 2.5178249543721304e-05, + "loss": 0.0064, + "step": 68210 + }, + { + "epoch": 10.59645852749301, + "grad_norm": 0.1836235523223877, + "learning_rate": 2.5163897561420668e-05, + "loss": 0.0051, + "step": 68220 + }, + { + "epoch": 10.598011804908356, + "grad_norm": 0.23776870965957642, + "learning_rate": 2.514954829515883e-05, + "loss": 0.0052, + "step": 68230 + }, + { + "epoch": 10.599565082323704, + "grad_norm": 0.14293688535690308, + "learning_rate": 2.513520174650505e-05, + "loss": 0.0063, + "step": 68240 + }, + { + "epoch": 10.601118359739049, + "grad_norm": 0.1946193277835846, + "learning_rate": 2.5120857917028207e-05, + "loss": 0.0063, + "step": 68250 + }, + { + "epoch": 10.602671637154396, + "grad_norm": 0.23792622983455658, + "learning_rate": 2.5106516808296953e-05, + "loss": 0.0047, + "step": 68260 + }, + { + "epoch": 10.604224914569743, + "grad_norm": 0.27053165435791016, + "learning_rate": 2.5092178421879552e-05, + "loss": 0.006, + "step": 68270 + }, + { + "epoch": 10.605778191985088, + "grad_norm": 0.23961801826953888, + "learning_rate": 2.5077842759344116e-05, + "loss": 0.0059, + "step": 68280 + }, + { + "epoch": 10.607331469400435, + "grad_norm": 0.1081906259059906, + "learning_rate": 2.5063509822258304e-05, + "loss": 0.0057, + "step": 68290 + }, + { + "epoch": 10.60888474681578, + "grad_norm": 0.3587450087070465, + "learning_rate": 2.5049179612189576e-05, + "loss": 0.006, + "step": 68300 + }, + { + "epoch": 10.610438024231128, + "grad_norm": 0.6584032773971558, + "learning_rate": 2.5034852130705043e-05, + "loss": 0.0069, + "step": 68310 + }, + { + "epoch": 10.611991301646475, + "grad_norm": 0.32079288363456726, + "learning_rate": 2.5020527379371572e-05, + "loss": 0.0059, + "step": 68320 + }, + { + "epoch": 10.61354457906182, + "grad_norm": 0.2974262535572052, + "learning_rate": 2.500620535975564e-05, + "loss": 0.0052, + "step": 68330 + }, + { + "epoch": 10.615097856477167, + "grad_norm": 0.4475882351398468, + "learning_rate": 2.4991886073423515e-05, + "loss": 0.0045, + "step": 68340 + }, + { + "epoch": 10.616651133892514, + "grad_norm": 0.0832873061299324, + "learning_rate": 2.4977569521941134e-05, + "loss": 0.0048, + "step": 68350 + }, + { + "epoch": 10.61820441130786, + "grad_norm": 0.1548699140548706, + "learning_rate": 2.4963255706874143e-05, + "loss": 0.0051, + "step": 68360 + }, + { + "epoch": 10.619757688723206, + "grad_norm": 0.16960348188877106, + "learning_rate": 2.4948944629787853e-05, + "loss": 0.0056, + "step": 68370 + }, + { + "epoch": 10.621310966138552, + "grad_norm": 0.6189568042755127, + "learning_rate": 2.493463629224731e-05, + "loss": 0.0058, + "step": 68380 + }, + { + "epoch": 10.622864243553899, + "grad_norm": 0.14239934086799622, + "learning_rate": 2.492033069581728e-05, + "loss": 0.0045, + "step": 68390 + }, + { + "epoch": 10.624417520969246, + "grad_norm": 0.20952017605304718, + "learning_rate": 2.4906027842062137e-05, + "loss": 0.0073, + "step": 68400 + }, + { + "epoch": 10.625970798384591, + "grad_norm": 0.2417295277118683, + "learning_rate": 2.48917277325461e-05, + "loss": 0.0057, + "step": 68410 + }, + { + "epoch": 10.627524075799938, + "grad_norm": 0.10501481592655182, + "learning_rate": 2.487743036883295e-05, + "loss": 0.0044, + "step": 68420 + }, + { + "epoch": 10.629077353215283, + "grad_norm": 0.6118320226669312, + "learning_rate": 2.486313575248626e-05, + "loss": 0.0059, + "step": 68430 + }, + { + "epoch": 10.63063063063063, + "grad_norm": 0.6355377435684204, + "learning_rate": 2.4848843885069235e-05, + "loss": 0.005, + "step": 68440 + }, + { + "epoch": 10.632183908045977, + "grad_norm": 1.1162333488464355, + "learning_rate": 2.483455476814484e-05, + "loss": 0.0042, + "step": 68450 + }, + { + "epoch": 10.633737185461323, + "grad_norm": 0.21393221616744995, + "learning_rate": 2.4820268403275672e-05, + "loss": 0.0051, + "step": 68460 + }, + { + "epoch": 10.63529046287667, + "grad_norm": 0.33863118290901184, + "learning_rate": 2.480598479202413e-05, + "loss": 0.005, + "step": 68470 + }, + { + "epoch": 10.636843740292017, + "grad_norm": 0.28251150250434875, + "learning_rate": 2.4791703935952193e-05, + "loss": 0.0042, + "step": 68480 + }, + { + "epoch": 10.638397017707362, + "grad_norm": 0.30541595816612244, + "learning_rate": 2.477742583662163e-05, + "loss": 0.0041, + "step": 68490 + }, + { + "epoch": 10.63995029512271, + "grad_norm": 0.1150239109992981, + "learning_rate": 2.476315049559384e-05, + "loss": 0.0035, + "step": 68500 + }, + { + "epoch": 10.641503572538056, + "grad_norm": 0.11303459852933884, + "learning_rate": 2.4748877914429987e-05, + "loss": 0.005, + "step": 68510 + }, + { + "epoch": 10.643056849953402, + "grad_norm": 0.32301145792007446, + "learning_rate": 2.4734608094690857e-05, + "loss": 0.0055, + "step": 68520 + }, + { + "epoch": 10.644610127368749, + "grad_norm": 0.7289570569992065, + "learning_rate": 2.4720341037937006e-05, + "loss": 0.0039, + "step": 68530 + }, + { + "epoch": 10.646163404784094, + "grad_norm": 0.7710428237915039, + "learning_rate": 2.4706076745728645e-05, + "loss": 0.006, + "step": 68540 + }, + { + "epoch": 10.647716682199441, + "grad_norm": 0.42886772751808167, + "learning_rate": 2.469181521962573e-05, + "loss": 0.0043, + "step": 68550 + }, + { + "epoch": 10.649269959614788, + "grad_norm": 0.24533730745315552, + "learning_rate": 2.467755646118783e-05, + "loss": 0.0059, + "step": 68560 + }, + { + "epoch": 10.650823237030133, + "grad_norm": 0.26400142908096313, + "learning_rate": 2.46633004719743e-05, + "loss": 0.0042, + "step": 68570 + }, + { + "epoch": 10.65237651444548, + "grad_norm": 0.10834239423274994, + "learning_rate": 2.4649047253544123e-05, + "loss": 0.0038, + "step": 68580 + }, + { + "epoch": 10.653929791860826, + "grad_norm": 0.12236666679382324, + "learning_rate": 2.4634796807456016e-05, + "loss": 0.0061, + "step": 68590 + }, + { + "epoch": 10.655483069276173, + "grad_norm": 0.32465073466300964, + "learning_rate": 2.46205491352684e-05, + "loss": 0.0067, + "step": 68600 + }, + { + "epoch": 10.65703634669152, + "grad_norm": 0.19847457110881805, + "learning_rate": 2.4606304238539375e-05, + "loss": 0.0048, + "step": 68610 + }, + { + "epoch": 10.658589624106865, + "grad_norm": 0.26424139738082886, + "learning_rate": 2.4592062118826753e-05, + "loss": 0.0037, + "step": 68620 + }, + { + "epoch": 10.660142901522212, + "grad_norm": 0.19804801046848297, + "learning_rate": 2.4577822777688004e-05, + "loss": 0.0073, + "step": 68630 + }, + { + "epoch": 10.661696178937559, + "grad_norm": 0.24481163918972015, + "learning_rate": 2.4563586216680347e-05, + "loss": 0.0075, + "step": 68640 + }, + { + "epoch": 10.663249456352904, + "grad_norm": 0.14769040048122406, + "learning_rate": 2.4549352437360622e-05, + "loss": 0.0062, + "step": 68650 + }, + { + "epoch": 10.664802733768251, + "grad_norm": 1.6859773397445679, + "learning_rate": 2.4535121441285493e-05, + "loss": 0.0046, + "step": 68660 + }, + { + "epoch": 10.666356011183597, + "grad_norm": 0.18330802023410797, + "learning_rate": 2.4520893230011172e-05, + "loss": 0.006, + "step": 68670 + }, + { + "epoch": 10.667909288598944, + "grad_norm": 0.27854540944099426, + "learning_rate": 2.450666780509368e-05, + "loss": 0.0064, + "step": 68680 + }, + { + "epoch": 10.66946256601429, + "grad_norm": 0.20625752210617065, + "learning_rate": 2.4492445168088656e-05, + "loss": 0.0048, + "step": 68690 + }, + { + "epoch": 10.671015843429636, + "grad_norm": 0.20245464146137238, + "learning_rate": 2.447822532055149e-05, + "loss": 0.0071, + "step": 68700 + }, + { + "epoch": 10.672569120844983, + "grad_norm": 0.24197864532470703, + "learning_rate": 2.4464008264037212e-05, + "loss": 0.0069, + "step": 68710 + }, + { + "epoch": 10.67412239826033, + "grad_norm": 0.28061002492904663, + "learning_rate": 2.4449794000100605e-05, + "loss": 0.006, + "step": 68720 + }, + { + "epoch": 10.675675675675675, + "grad_norm": 0.459224671125412, + "learning_rate": 2.4435582530296108e-05, + "loss": 0.0055, + "step": 68730 + }, + { + "epoch": 10.677228953091022, + "grad_norm": 0.18283210694789886, + "learning_rate": 2.4421373856177887e-05, + "loss": 0.005, + "step": 68740 + }, + { + "epoch": 10.678782230506368, + "grad_norm": 0.12719690799713135, + "learning_rate": 2.440716797929975e-05, + "loss": 0.0047, + "step": 68750 + }, + { + "epoch": 10.680335507921715, + "grad_norm": 0.11429053544998169, + "learning_rate": 2.439296490121526e-05, + "loss": 0.0062, + "step": 68760 + }, + { + "epoch": 10.681888785337062, + "grad_norm": 0.19971555471420288, + "learning_rate": 2.4378764623477608e-05, + "loss": 0.0051, + "step": 68770 + }, + { + "epoch": 10.683442062752407, + "grad_norm": 0.15146125853061676, + "learning_rate": 2.436456714763974e-05, + "loss": 0.0053, + "step": 68780 + }, + { + "epoch": 10.684995340167754, + "grad_norm": 0.27486708760261536, + "learning_rate": 2.4350372475254262e-05, + "loss": 0.006, + "step": 68790 + }, + { + "epoch": 10.6865486175831, + "grad_norm": 0.20557446777820587, + "learning_rate": 2.433618060787351e-05, + "loss": 0.0061, + "step": 68800 + }, + { + "epoch": 10.688101894998447, + "grad_norm": 0.38638731837272644, + "learning_rate": 2.432199154704944e-05, + "loss": 0.0059, + "step": 68810 + }, + { + "epoch": 10.689655172413794, + "grad_norm": 0.12537544965744019, + "learning_rate": 2.430780529433377e-05, + "loss": 0.0032, + "step": 68820 + }, + { + "epoch": 10.691208449829139, + "grad_norm": 0.1975526362657547, + "learning_rate": 2.42936218512779e-05, + "loss": 0.0058, + "step": 68830 + }, + { + "epoch": 10.692761727244486, + "grad_norm": 0.15766094624996185, + "learning_rate": 2.427944121943288e-05, + "loss": 0.0063, + "step": 68840 + }, + { + "epoch": 10.694315004659833, + "grad_norm": 0.24887123703956604, + "learning_rate": 2.4265263400349497e-05, + "loss": 0.0038, + "step": 68850 + }, + { + "epoch": 10.695868282075178, + "grad_norm": 0.1748560220003128, + "learning_rate": 2.4251088395578213e-05, + "loss": 0.0048, + "step": 68860 + }, + { + "epoch": 10.697421559490525, + "grad_norm": 0.3104209899902344, + "learning_rate": 2.4236916206669214e-05, + "loss": 0.0049, + "step": 68870 + }, + { + "epoch": 10.69897483690587, + "grad_norm": 0.23705177009105682, + "learning_rate": 2.4222746835172304e-05, + "loss": 0.005, + "step": 68880 + }, + { + "epoch": 10.700528114321218, + "grad_norm": 0.13573390245437622, + "learning_rate": 2.4208580282637065e-05, + "loss": 0.0039, + "step": 68890 + }, + { + "epoch": 10.702081391736565, + "grad_norm": 0.33230504393577576, + "learning_rate": 2.4194416550612687e-05, + "loss": 0.0052, + "step": 68900 + }, + { + "epoch": 10.70363466915191, + "grad_norm": 0.21015669405460358, + "learning_rate": 2.418025564064812e-05, + "loss": 0.006, + "step": 68910 + }, + { + "epoch": 10.705187946567257, + "grad_norm": 0.1097012311220169, + "learning_rate": 2.4166097554291978e-05, + "loss": 0.0047, + "step": 68920 + }, + { + "epoch": 10.706741223982604, + "grad_norm": 0.30342939496040344, + "learning_rate": 2.4151942293092583e-05, + "loss": 0.0041, + "step": 68930 + }, + { + "epoch": 10.70829450139795, + "grad_norm": 0.1055680587887764, + "learning_rate": 2.41377898585979e-05, + "loss": 0.0022, + "step": 68940 + }, + { + "epoch": 10.709847778813296, + "grad_norm": 0.2300688624382019, + "learning_rate": 2.4123640252355652e-05, + "loss": 0.005, + "step": 68950 + }, + { + "epoch": 10.711401056228642, + "grad_norm": 0.1868467628955841, + "learning_rate": 2.410949347591318e-05, + "loss": 0.0047, + "step": 68960 + }, + { + "epoch": 10.712954333643989, + "grad_norm": 0.2312668412923813, + "learning_rate": 2.4095349530817578e-05, + "loss": 0.0068, + "step": 68970 + }, + { + "epoch": 10.714507611059336, + "grad_norm": 0.7114292979240417, + "learning_rate": 2.4081208418615603e-05, + "loss": 0.0063, + "step": 68980 + }, + { + "epoch": 10.716060888474681, + "grad_norm": 0.149293914437294, + "learning_rate": 2.4067070140853725e-05, + "loss": 0.0058, + "step": 68990 + }, + { + "epoch": 10.717614165890028, + "grad_norm": 0.11524365097284317, + "learning_rate": 2.405293469907805e-05, + "loss": 0.0055, + "step": 69000 + }, + { + "epoch": 10.719167443305373, + "grad_norm": 0.22951211035251617, + "learning_rate": 2.4038802094834446e-05, + "loss": 0.0064, + "step": 69010 + }, + { + "epoch": 10.72072072072072, + "grad_norm": 0.28523340821266174, + "learning_rate": 2.4024672329668395e-05, + "loss": 0.0049, + "step": 69020 + }, + { + "epoch": 10.722273998136068, + "grad_norm": 0.3220877945423126, + "learning_rate": 2.4010545405125125e-05, + "loss": 0.0057, + "step": 69030 + }, + { + "epoch": 10.723827275551413, + "grad_norm": 0.6023439168930054, + "learning_rate": 2.3996421322749528e-05, + "loss": 0.0051, + "step": 69040 + }, + { + "epoch": 10.72538055296676, + "grad_norm": 0.22939594089984894, + "learning_rate": 2.3982300084086224e-05, + "loss": 0.0058, + "step": 69050 + }, + { + "epoch": 10.726933830382107, + "grad_norm": 0.26837530732154846, + "learning_rate": 2.396818169067944e-05, + "loss": 0.0058, + "step": 69060 + }, + { + "epoch": 10.728487107797452, + "grad_norm": 0.21407660841941833, + "learning_rate": 2.3954066144073178e-05, + "loss": 0.0062, + "step": 69070 + }, + { + "epoch": 10.7300403852128, + "grad_norm": 0.27377739548683167, + "learning_rate": 2.39399534458111e-05, + "loss": 0.0047, + "step": 69080 + }, + { + "epoch": 10.731593662628146, + "grad_norm": 0.31915414333343506, + "learning_rate": 2.392584359743651e-05, + "loss": 0.005, + "step": 69090 + }, + { + "epoch": 10.733146940043492, + "grad_norm": 0.17719261348247528, + "learning_rate": 2.3911736600492463e-05, + "loss": 0.0053, + "step": 69100 + }, + { + "epoch": 10.734700217458839, + "grad_norm": 0.12902255356311798, + "learning_rate": 2.3897632456521672e-05, + "loss": 0.0059, + "step": 69110 + }, + { + "epoch": 10.736253494874184, + "grad_norm": 0.1320641189813614, + "learning_rate": 2.388353116706658e-05, + "loss": 0.004, + "step": 69120 + }, + { + "epoch": 10.737806772289531, + "grad_norm": 0.18563804030418396, + "learning_rate": 2.3869432733669232e-05, + "loss": 0.0043, + "step": 69130 + }, + { + "epoch": 10.739360049704878, + "grad_norm": 0.4447570741176605, + "learning_rate": 2.385533715787145e-05, + "loss": 0.0055, + "step": 69140 + }, + { + "epoch": 10.740913327120223, + "grad_norm": 0.24679231643676758, + "learning_rate": 2.3841244441214673e-05, + "loss": 0.0067, + "step": 69150 + }, + { + "epoch": 10.74246660453557, + "grad_norm": 0.14481545984745026, + "learning_rate": 2.3827154585240073e-05, + "loss": 0.0054, + "step": 69160 + }, + { + "epoch": 10.744019881950916, + "grad_norm": 0.15879122912883759, + "learning_rate": 2.3813067591488498e-05, + "loss": 0.0047, + "step": 69170 + }, + { + "epoch": 10.745573159366263, + "grad_norm": 0.11265865713357925, + "learning_rate": 2.3798983461500494e-05, + "loss": 0.0047, + "step": 69180 + }, + { + "epoch": 10.74712643678161, + "grad_norm": 0.37954407930374146, + "learning_rate": 2.3784902196816256e-05, + "loss": 0.0042, + "step": 69190 + }, + { + "epoch": 10.748679714196955, + "grad_norm": 0.1697731912136078, + "learning_rate": 2.3770823798975716e-05, + "loss": 0.0055, + "step": 69200 + }, + { + "epoch": 10.750232991612302, + "grad_norm": 0.15007388591766357, + "learning_rate": 2.3756748269518424e-05, + "loss": 0.0055, + "step": 69210 + }, + { + "epoch": 10.75178626902765, + "grad_norm": 0.1371137648820877, + "learning_rate": 2.374267560998371e-05, + "loss": 0.003, + "step": 69220 + }, + { + "epoch": 10.753339546442994, + "grad_norm": 0.1095077395439148, + "learning_rate": 2.3728605821910477e-05, + "loss": 0.0041, + "step": 69230 + }, + { + "epoch": 10.754892823858341, + "grad_norm": 0.10120095312595367, + "learning_rate": 2.3714538906837453e-05, + "loss": 0.0048, + "step": 69240 + }, + { + "epoch": 10.756446101273687, + "grad_norm": 0.23421165347099304, + "learning_rate": 2.3700474866302913e-05, + "loss": 0.0046, + "step": 69250 + }, + { + "epoch": 10.757999378689034, + "grad_norm": 0.1542830765247345, + "learning_rate": 2.3686413701844917e-05, + "loss": 0.0039, + "step": 69260 + }, + { + "epoch": 10.75955265610438, + "grad_norm": 0.09862010926008224, + "learning_rate": 2.367235541500114e-05, + "loss": 0.0054, + "step": 69270 + }, + { + "epoch": 10.761105933519726, + "grad_norm": 0.39120516180992126, + "learning_rate": 2.3658300007308993e-05, + "loss": 0.0074, + "step": 69280 + }, + { + "epoch": 10.762659210935073, + "grad_norm": 0.2135857492685318, + "learning_rate": 2.364424748030555e-05, + "loss": 0.0059, + "step": 69290 + }, + { + "epoch": 10.76421248835042, + "grad_norm": 1.0315200090408325, + "learning_rate": 2.3630197835527595e-05, + "loss": 0.0069, + "step": 69300 + }, + { + "epoch": 10.765765765765765, + "grad_norm": 0.14575804769992828, + "learning_rate": 2.3616151074511546e-05, + "loss": 0.0073, + "step": 69310 + }, + { + "epoch": 10.767319043181113, + "grad_norm": 0.20284542441368103, + "learning_rate": 2.3602107198793545e-05, + "loss": 0.0034, + "step": 69320 + }, + { + "epoch": 10.768872320596458, + "grad_norm": 0.08699171245098114, + "learning_rate": 2.3588066209909427e-05, + "loss": 0.0051, + "step": 69330 + }, + { + "epoch": 10.770425598011805, + "grad_norm": 0.12881417572498322, + "learning_rate": 2.3574028109394662e-05, + "loss": 0.0041, + "step": 69340 + }, + { + "epoch": 10.771978875427152, + "grad_norm": 0.1866646558046341, + "learning_rate": 2.3559992898784454e-05, + "loss": 0.0068, + "step": 69350 + }, + { + "epoch": 10.773532152842497, + "grad_norm": 0.14904552698135376, + "learning_rate": 2.354596057961367e-05, + "loss": 0.0046, + "step": 69360 + }, + { + "epoch": 10.775085430257844, + "grad_norm": 0.18240559101104736, + "learning_rate": 2.353193115341688e-05, + "loss": 0.0046, + "step": 69370 + }, + { + "epoch": 10.77663870767319, + "grad_norm": 0.18174970149993896, + "learning_rate": 2.3517904621728294e-05, + "loss": 0.0044, + "step": 69380 + }, + { + "epoch": 10.778191985088537, + "grad_norm": 0.6066761612892151, + "learning_rate": 2.3503880986081855e-05, + "loss": 0.0043, + "step": 69390 + }, + { + "epoch": 10.779745262503884, + "grad_norm": 0.3147471845149994, + "learning_rate": 2.3489860248011136e-05, + "loss": 0.0069, + "step": 69400 + }, + { + "epoch": 10.781298539919229, + "grad_norm": 1.2818694114685059, + "learning_rate": 2.3475842409049464e-05, + "loss": 0.0034, + "step": 69410 + }, + { + "epoch": 10.782851817334576, + "grad_norm": 0.09747682511806488, + "learning_rate": 2.346182747072975e-05, + "loss": 0.0038, + "step": 69420 + }, + { + "epoch": 10.784405094749923, + "grad_norm": 0.1871775984764099, + "learning_rate": 2.3447815434584718e-05, + "loss": 0.0043, + "step": 69430 + }, + { + "epoch": 10.785958372165268, + "grad_norm": 0.19908364117145538, + "learning_rate": 2.3433806302146656e-05, + "loss": 0.0059, + "step": 69440 + }, + { + "epoch": 10.787511649580615, + "grad_norm": 0.08016826212406158, + "learning_rate": 2.3419800074947612e-05, + "loss": 0.0037, + "step": 69450 + }, + { + "epoch": 10.78906492699596, + "grad_norm": 0.2692437171936035, + "learning_rate": 2.3405796754519248e-05, + "loss": 0.0061, + "step": 69460 + }, + { + "epoch": 10.790618204411308, + "grad_norm": 0.19620271027088165, + "learning_rate": 2.3391796342392992e-05, + "loss": 0.0052, + "step": 69470 + }, + { + "epoch": 10.792171481826655, + "grad_norm": 0.441334992647171, + "learning_rate": 2.337779884009984e-05, + "loss": 0.0044, + "step": 69480 + }, + { + "epoch": 10.793724759242, + "grad_norm": 0.21061335504055023, + "learning_rate": 2.3363804249170624e-05, + "loss": 0.0062, + "step": 69490 + }, + { + "epoch": 10.795278036657347, + "grad_norm": 0.14966526627540588, + "learning_rate": 2.3349812571135703e-05, + "loss": 0.0054, + "step": 69500 + }, + { + "epoch": 10.796831314072694, + "grad_norm": 0.5234830379486084, + "learning_rate": 2.333582380752523e-05, + "loss": 0.0062, + "step": 69510 + }, + { + "epoch": 10.79838459148804, + "grad_norm": 0.32129496335983276, + "learning_rate": 2.3321837959868963e-05, + "loss": 0.0093, + "step": 69520 + }, + { + "epoch": 10.799937868903386, + "grad_norm": 0.16068479418754578, + "learning_rate": 2.3307855029696383e-05, + "loss": 0.0043, + "step": 69530 + }, + { + "epoch": 10.801491146318732, + "grad_norm": 0.12053877860307693, + "learning_rate": 2.3293875018536648e-05, + "loss": 0.004, + "step": 69540 + }, + { + "epoch": 10.803044423734079, + "grad_norm": 0.40888848900794983, + "learning_rate": 2.327989792791861e-05, + "loss": 0.0047, + "step": 69550 + }, + { + "epoch": 10.804597701149426, + "grad_norm": 0.6309958100318909, + "learning_rate": 2.326592375937074e-05, + "loss": 0.006, + "step": 69560 + }, + { + "epoch": 10.806150978564771, + "grad_norm": 0.11127625405788422, + "learning_rate": 2.3251952514421248e-05, + "loss": 0.0049, + "step": 69570 + }, + { + "epoch": 10.807704255980118, + "grad_norm": 0.24202698469161987, + "learning_rate": 2.323798419459804e-05, + "loss": 0.0086, + "step": 69580 + }, + { + "epoch": 10.809257533395465, + "grad_norm": 0.1694805920124054, + "learning_rate": 2.3224018801428625e-05, + "loss": 0.0063, + "step": 69590 + }, + { + "epoch": 10.81081081081081, + "grad_norm": 0.24186928570270538, + "learning_rate": 2.3210056336440273e-05, + "loss": 0.0049, + "step": 69600 + }, + { + "epoch": 10.812364088226158, + "grad_norm": 0.7393789291381836, + "learning_rate": 2.3196096801159844e-05, + "loss": 0.0046, + "step": 69610 + }, + { + "epoch": 10.813917365641503, + "grad_norm": 0.36965346336364746, + "learning_rate": 2.3182140197114012e-05, + "loss": 0.0068, + "step": 69620 + }, + { + "epoch": 10.81547064305685, + "grad_norm": 0.40148448944091797, + "learning_rate": 2.3168186525828985e-05, + "loss": 0.0049, + "step": 69630 + }, + { + "epoch": 10.817023920472197, + "grad_norm": 0.1396680772304535, + "learning_rate": 2.3154235788830757e-05, + "loss": 0.005, + "step": 69640 + }, + { + "epoch": 10.818577197887542, + "grad_norm": 0.33809152245521545, + "learning_rate": 2.3140287987644927e-05, + "loss": 0.0056, + "step": 69650 + }, + { + "epoch": 10.82013047530289, + "grad_norm": 0.18286362290382385, + "learning_rate": 2.3126343123796833e-05, + "loss": 0.0046, + "step": 69660 + }, + { + "epoch": 10.821683752718236, + "grad_norm": 0.12376571446657181, + "learning_rate": 2.3112401198811418e-05, + "loss": 0.0047, + "step": 69670 + }, + { + "epoch": 10.823237030133582, + "grad_norm": 0.1340668648481369, + "learning_rate": 2.3098462214213418e-05, + "loss": 0.0049, + "step": 69680 + }, + { + "epoch": 10.824790307548929, + "grad_norm": 0.38295578956604004, + "learning_rate": 2.308452617152712e-05, + "loss": 0.0058, + "step": 69690 + }, + { + "epoch": 10.826343584964274, + "grad_norm": 0.18612812459468842, + "learning_rate": 2.3070593072276603e-05, + "loss": 0.0056, + "step": 69700 + }, + { + "epoch": 10.827896862379621, + "grad_norm": 0.2117079794406891, + "learning_rate": 2.3056662917985518e-05, + "loss": 0.0058, + "step": 69710 + }, + { + "epoch": 10.829450139794968, + "grad_norm": 0.4562084972858429, + "learning_rate": 2.3042735710177283e-05, + "loss": 0.0052, + "step": 69720 + }, + { + "epoch": 10.831003417210313, + "grad_norm": 0.26760369539260864, + "learning_rate": 2.302881145037492e-05, + "loss": 0.0047, + "step": 69730 + }, + { + "epoch": 10.83255669462566, + "grad_norm": 0.13515892624855042, + "learning_rate": 2.3014890140101192e-05, + "loss": 0.0048, + "step": 69740 + }, + { + "epoch": 10.834109972041006, + "grad_norm": 0.12226452678442001, + "learning_rate": 2.30009717808785e-05, + "loss": 0.0035, + "step": 69750 + }, + { + "epoch": 10.835663249456353, + "grad_norm": 0.4540571868419647, + "learning_rate": 2.2987056374228967e-05, + "loss": 0.0037, + "step": 69760 + }, + { + "epoch": 10.8372165268717, + "grad_norm": 0.21110840141773224, + "learning_rate": 2.297314392167431e-05, + "loss": 0.007, + "step": 69770 + }, + { + "epoch": 10.838769804287045, + "grad_norm": 0.21625033020973206, + "learning_rate": 2.295923442473601e-05, + "loss": 0.0054, + "step": 69780 + }, + { + "epoch": 10.840323081702392, + "grad_norm": 0.24857832491397858, + "learning_rate": 2.294532788493519e-05, + "loss": 0.0049, + "step": 69790 + }, + { + "epoch": 10.84187635911774, + "grad_norm": 0.8618460893630981, + "learning_rate": 2.293142430379262e-05, + "loss": 0.0047, + "step": 69800 + }, + { + "epoch": 10.843429636533084, + "grad_norm": 0.8284844160079956, + "learning_rate": 2.291752368282879e-05, + "loss": 0.0059, + "step": 69810 + }, + { + "epoch": 10.844982913948432, + "grad_norm": 0.38928550481796265, + "learning_rate": 2.290362602356385e-05, + "loss": 0.0052, + "step": 69820 + }, + { + "epoch": 10.846536191363777, + "grad_norm": 0.0829816684126854, + "learning_rate": 2.288973132751765e-05, + "loss": 0.0044, + "step": 69830 + }, + { + "epoch": 10.848089468779124, + "grad_norm": 0.3602513074874878, + "learning_rate": 2.287583959620965e-05, + "loss": 0.0057, + "step": 69840 + }, + { + "epoch": 10.84964274619447, + "grad_norm": 0.11917679756879807, + "learning_rate": 2.2861950831159074e-05, + "loss": 0.0045, + "step": 69850 + }, + { + "epoch": 10.851196023609816, + "grad_norm": 0.13360849022865295, + "learning_rate": 2.284806503388472e-05, + "loss": 0.0052, + "step": 69860 + }, + { + "epoch": 10.852749301025163, + "grad_norm": 0.20119445025920868, + "learning_rate": 2.283418220590518e-05, + "loss": 0.0041, + "step": 69870 + }, + { + "epoch": 10.85430257844051, + "grad_norm": 0.204827681183815, + "learning_rate": 2.282030234873862e-05, + "loss": 0.0045, + "step": 69880 + }, + { + "epoch": 10.855855855855856, + "grad_norm": 0.17241765558719635, + "learning_rate": 2.2806425463902948e-05, + "loss": 0.0055, + "step": 69890 + }, + { + "epoch": 10.857409133271203, + "grad_norm": 0.1961793750524521, + "learning_rate": 2.279255155291568e-05, + "loss": 0.0043, + "step": 69900 + }, + { + "epoch": 10.858962410686548, + "grad_norm": 0.3753833770751953, + "learning_rate": 2.277868061729409e-05, + "loss": 0.0043, + "step": 69910 + }, + { + "epoch": 10.860515688101895, + "grad_norm": 0.2112334817647934, + "learning_rate": 2.276481265855504e-05, + "loss": 0.0059, + "step": 69920 + }, + { + "epoch": 10.862068965517242, + "grad_norm": 0.16202257573604584, + "learning_rate": 2.2750947678215128e-05, + "loss": 0.0038, + "step": 69930 + }, + { + "epoch": 10.863622242932587, + "grad_norm": 0.1442611962556839, + "learning_rate": 2.2737085677790616e-05, + "loss": 0.0061, + "step": 69940 + }, + { + "epoch": 10.865175520347934, + "grad_norm": 0.13996922969818115, + "learning_rate": 2.2723226658797443e-05, + "loss": 0.0042, + "step": 69950 + }, + { + "epoch": 10.86672879776328, + "grad_norm": 0.18813107907772064, + "learning_rate": 2.270937062275117e-05, + "loss": 0.0046, + "step": 69960 + }, + { + "epoch": 10.868282075178627, + "grad_norm": 0.1460079699754715, + "learning_rate": 2.2695517571167113e-05, + "loss": 0.0043, + "step": 69970 + }, + { + "epoch": 10.869835352593974, + "grad_norm": 0.16750331223011017, + "learning_rate": 2.268166750556019e-05, + "loss": 0.0049, + "step": 69980 + }, + { + "epoch": 10.871388630009319, + "grad_norm": 0.4110412895679474, + "learning_rate": 2.266782042744503e-05, + "loss": 0.0054, + "step": 69990 + }, + { + "epoch": 10.872941907424666, + "grad_norm": 0.2396266907453537, + "learning_rate": 2.2653976338335935e-05, + "loss": 0.0063, + "step": 70000 + }, + { + "epoch": 10.874495184840013, + "grad_norm": 0.44221416115760803, + "learning_rate": 2.2640135239746895e-05, + "loss": 0.0038, + "step": 70010 + }, + { + "epoch": 10.876048462255358, + "grad_norm": 0.1013026162981987, + "learning_rate": 2.2626297133191515e-05, + "loss": 0.0059, + "step": 70020 + }, + { + "epoch": 10.877601739670705, + "grad_norm": 0.6626214385032654, + "learning_rate": 2.261246202018312e-05, + "loss": 0.0058, + "step": 70030 + }, + { + "epoch": 10.879155017086052, + "grad_norm": 0.7639119029045105, + "learning_rate": 2.2598629902234718e-05, + "loss": 0.0048, + "step": 70040 + }, + { + "epoch": 10.880708294501398, + "grad_norm": 0.38143542408943176, + "learning_rate": 2.258480078085894e-05, + "loss": 0.0063, + "step": 70050 + }, + { + "epoch": 10.882261571916745, + "grad_norm": 0.16215695440769196, + "learning_rate": 2.2570974657568126e-05, + "loss": 0.0042, + "step": 70060 + }, + { + "epoch": 10.88381484933209, + "grad_norm": 0.1859569102525711, + "learning_rate": 2.2557151533874286e-05, + "loss": 0.0058, + "step": 70070 + }, + { + "epoch": 10.885368126747437, + "grad_norm": 0.3264278471469879, + "learning_rate": 2.2543331411289116e-05, + "loss": 0.0036, + "step": 70080 + }, + { + "epoch": 10.886921404162784, + "grad_norm": 0.33245810866355896, + "learning_rate": 2.252951429132392e-05, + "loss": 0.0058, + "step": 70090 + }, + { + "epoch": 10.88847468157813, + "grad_norm": 0.13819588720798492, + "learning_rate": 2.2515700175489762e-05, + "loss": 0.004, + "step": 70100 + }, + { + "epoch": 10.890027958993477, + "grad_norm": 0.16021256148815155, + "learning_rate": 2.2501889065297288e-05, + "loss": 0.0033, + "step": 70110 + }, + { + "epoch": 10.891581236408822, + "grad_norm": 0.23097549378871918, + "learning_rate": 2.2488080962256887e-05, + "loss": 0.0052, + "step": 70120 + }, + { + "epoch": 10.893134513824169, + "grad_norm": 0.3124566972255707, + "learning_rate": 2.2474275867878585e-05, + "loss": 0.0049, + "step": 70130 + }, + { + "epoch": 10.894687791239516, + "grad_norm": 0.33655762672424316, + "learning_rate": 2.246047378367211e-05, + "loss": 0.004, + "step": 70140 + }, + { + "epoch": 10.896241068654861, + "grad_norm": 0.47074243426322937, + "learning_rate": 2.2446674711146798e-05, + "loss": 0.0053, + "step": 70150 + }, + { + "epoch": 10.897794346070208, + "grad_norm": 0.15718583762645721, + "learning_rate": 2.2432878651811734e-05, + "loss": 0.005, + "step": 70160 + }, + { + "epoch": 10.899347623485555, + "grad_norm": 0.2074878215789795, + "learning_rate": 2.2419085607175594e-05, + "loss": 0.0051, + "step": 70170 + }, + { + "epoch": 10.9009009009009, + "grad_norm": 0.12998268008232117, + "learning_rate": 2.240529557874678e-05, + "loss": 0.005, + "step": 70180 + }, + { + "epoch": 10.902454178316248, + "grad_norm": 0.1674163043498993, + "learning_rate": 2.2391508568033358e-05, + "loss": 0.0049, + "step": 70190 + }, + { + "epoch": 10.904007455731593, + "grad_norm": 0.21410290896892548, + "learning_rate": 2.237772457654307e-05, + "loss": 0.005, + "step": 70200 + }, + { + "epoch": 10.90556073314694, + "grad_norm": 0.12989558279514313, + "learning_rate": 2.236394360578327e-05, + "loss": 0.0053, + "step": 70210 + }, + { + "epoch": 10.907114010562287, + "grad_norm": 0.5572404265403748, + "learning_rate": 2.2350165657261073e-05, + "loss": 0.0046, + "step": 70220 + }, + { + "epoch": 10.908667287977632, + "grad_norm": 0.6321815848350525, + "learning_rate": 2.2336390732483164e-05, + "loss": 0.0052, + "step": 70230 + }, + { + "epoch": 10.91022056539298, + "grad_norm": 0.22971075773239136, + "learning_rate": 2.2322618832955976e-05, + "loss": 0.0048, + "step": 70240 + }, + { + "epoch": 10.911773842808326, + "grad_norm": 0.8600208759307861, + "learning_rate": 2.2308849960185578e-05, + "loss": 0.0056, + "step": 70250 + }, + { + "epoch": 10.913327120223672, + "grad_norm": 0.20969878137111664, + "learning_rate": 2.229508411567774e-05, + "loss": 0.0046, + "step": 70260 + }, + { + "epoch": 10.914880397639019, + "grad_norm": 0.22071640193462372, + "learning_rate": 2.2281321300937823e-05, + "loss": 0.0046, + "step": 70270 + }, + { + "epoch": 10.916433675054364, + "grad_norm": 0.23678788542747498, + "learning_rate": 2.2267561517470937e-05, + "loss": 0.0053, + "step": 70280 + }, + { + "epoch": 10.917986952469711, + "grad_norm": 0.2872283160686493, + "learning_rate": 2.2253804766781845e-05, + "loss": 0.0051, + "step": 70290 + }, + { + "epoch": 10.919540229885058, + "grad_norm": 0.16718436777591705, + "learning_rate": 2.2240051050374933e-05, + "loss": 0.0039, + "step": 70300 + }, + { + "epoch": 10.921093507300403, + "grad_norm": 0.2526150643825531, + "learning_rate": 2.22263003697543e-05, + "loss": 0.0048, + "step": 70310 + }, + { + "epoch": 10.92264678471575, + "grad_norm": 0.36128363013267517, + "learning_rate": 2.2212552726423702e-05, + "loss": 0.0064, + "step": 70320 + }, + { + "epoch": 10.924200062131096, + "grad_norm": 0.12339125573635101, + "learning_rate": 2.2198808121886578e-05, + "loss": 0.0041, + "step": 70330 + }, + { + "epoch": 10.925753339546443, + "grad_norm": 0.7820106148719788, + "learning_rate": 2.218506655764598e-05, + "loss": 0.0038, + "step": 70340 + }, + { + "epoch": 10.92730661696179, + "grad_norm": 0.3083457350730896, + "learning_rate": 2.2171328035204704e-05, + "loss": 0.0058, + "step": 70350 + }, + { + "epoch": 10.928859894377135, + "grad_norm": 0.19419291615486145, + "learning_rate": 2.2157592556065133e-05, + "loss": 0.0039, + "step": 70360 + }, + { + "epoch": 10.930413171792482, + "grad_norm": 0.31140807271003723, + "learning_rate": 2.2143860121729382e-05, + "loss": 0.0058, + "step": 70370 + }, + { + "epoch": 10.93196644920783, + "grad_norm": 0.13551349937915802, + "learning_rate": 2.2130130733699206e-05, + "loss": 0.0032, + "step": 70380 + }, + { + "epoch": 10.933519726623175, + "grad_norm": 0.8859524726867676, + "learning_rate": 2.2116404393476052e-05, + "loss": 0.0033, + "step": 70390 + }, + { + "epoch": 10.935073004038522, + "grad_norm": 0.2122979313135147, + "learning_rate": 2.2102681102560967e-05, + "loss": 0.0045, + "step": 70400 + }, + { + "epoch": 10.936626281453867, + "grad_norm": 0.16752228140830994, + "learning_rate": 2.208896086245476e-05, + "loss": 0.0052, + "step": 70410 + }, + { + "epoch": 10.938179558869214, + "grad_norm": 0.19504328072071075, + "learning_rate": 2.2075243674657808e-05, + "loss": 0.0045, + "step": 70420 + }, + { + "epoch": 10.939732836284561, + "grad_norm": 0.1750888228416443, + "learning_rate": 2.206152954067024e-05, + "loss": 0.0054, + "step": 70430 + }, + { + "epoch": 10.941286113699906, + "grad_norm": 0.21516506373882294, + "learning_rate": 2.2047818461991758e-05, + "loss": 0.0046, + "step": 70440 + }, + { + "epoch": 10.942839391115253, + "grad_norm": 0.25411465764045715, + "learning_rate": 2.2034110440121864e-05, + "loss": 0.0047, + "step": 70450 + }, + { + "epoch": 10.9443926685306, + "grad_norm": 0.1013675183057785, + "learning_rate": 2.2020405476559585e-05, + "loss": 0.0045, + "step": 70460 + }, + { + "epoch": 10.945945945945946, + "grad_norm": 0.3838209807872772, + "learning_rate": 2.2006703572803722e-05, + "loss": 0.0051, + "step": 70470 + }, + { + "epoch": 10.947499223361293, + "grad_norm": 0.23189589381217957, + "learning_rate": 2.1993004730352646e-05, + "loss": 0.006, + "step": 70480 + }, + { + "epoch": 10.949052500776638, + "grad_norm": 1.7529661655426025, + "learning_rate": 2.1979308950704464e-05, + "loss": 0.0047, + "step": 70490 + }, + { + "epoch": 10.950605778191985, + "grad_norm": 0.2026868611574173, + "learning_rate": 2.1965616235356923e-05, + "loss": 0.0047, + "step": 70500 + }, + { + "epoch": 10.952159055607332, + "grad_norm": 0.1745285987854004, + "learning_rate": 2.1951926585807465e-05, + "loss": 0.0017, + "step": 70510 + }, + { + "epoch": 10.953712333022677, + "grad_norm": 0.22273515164852142, + "learning_rate": 2.193824000355313e-05, + "loss": 0.0047, + "step": 70520 + }, + { + "epoch": 10.955265610438024, + "grad_norm": 0.13795118033885956, + "learning_rate": 2.192455649009067e-05, + "loss": 0.0036, + "step": 70530 + }, + { + "epoch": 10.956818887853371, + "grad_norm": 0.2607761025428772, + "learning_rate": 2.1910876046916523e-05, + "loss": 0.0036, + "step": 70540 + }, + { + "epoch": 10.958372165268717, + "grad_norm": 0.4351838231086731, + "learning_rate": 2.1897198675526725e-05, + "loss": 0.0046, + "step": 70550 + }, + { + "epoch": 10.959925442684064, + "grad_norm": 0.26950111985206604, + "learning_rate": 2.1883524377417024e-05, + "loss": 0.0064, + "step": 70560 + }, + { + "epoch": 10.961478720099409, + "grad_norm": 0.09679341316223145, + "learning_rate": 2.1869853154082825e-05, + "loss": 0.0051, + "step": 70570 + }, + { + "epoch": 10.963031997514756, + "grad_norm": 0.12202561646699905, + "learning_rate": 2.1856185007019215e-05, + "loss": 0.0037, + "step": 70580 + }, + { + "epoch": 10.964585274930103, + "grad_norm": 0.4150641858577728, + "learning_rate": 2.1842519937720874e-05, + "loss": 0.0049, + "step": 70590 + }, + { + "epoch": 10.966138552345448, + "grad_norm": 0.36483368277549744, + "learning_rate": 2.182885794768224e-05, + "loss": 0.0058, + "step": 70600 + }, + { + "epoch": 10.967691829760795, + "grad_norm": 0.2409229576587677, + "learning_rate": 2.1815199038397326e-05, + "loss": 0.0049, + "step": 70610 + }, + { + "epoch": 10.969245107176143, + "grad_norm": 0.40512946248054504, + "learning_rate": 2.1801543211359893e-05, + "loss": 0.0043, + "step": 70620 + }, + { + "epoch": 10.970798384591488, + "grad_norm": 0.2146773338317871, + "learning_rate": 2.178789046806326e-05, + "loss": 0.0042, + "step": 70630 + }, + { + "epoch": 10.972351662006835, + "grad_norm": 0.1723087728023529, + "learning_rate": 2.1774240810000547e-05, + "loss": 0.0063, + "step": 70640 + }, + { + "epoch": 10.97390493942218, + "grad_norm": 0.1933303326368332, + "learning_rate": 2.17605942386644e-05, + "loss": 0.0044, + "step": 70650 + }, + { + "epoch": 10.975458216837527, + "grad_norm": 0.6868093013763428, + "learning_rate": 2.174695075554723e-05, + "loss": 0.0056, + "step": 70660 + }, + { + "epoch": 10.977011494252874, + "grad_norm": 0.24986408650875092, + "learning_rate": 2.173331036214103e-05, + "loss": 0.0064, + "step": 70670 + }, + { + "epoch": 10.97856477166822, + "grad_norm": 0.26087817549705505, + "learning_rate": 2.171967305993753e-05, + "loss": 0.004, + "step": 70680 + }, + { + "epoch": 10.980118049083567, + "grad_norm": 0.1654045581817627, + "learning_rate": 2.1706038850428024e-05, + "loss": 0.0038, + "step": 70690 + }, + { + "epoch": 10.981671326498912, + "grad_norm": 0.3086773455142975, + "learning_rate": 2.169240773510361e-05, + "loss": 0.0071, + "step": 70700 + }, + { + "epoch": 10.983224603914259, + "grad_norm": 0.1881350576877594, + "learning_rate": 2.1678779715454906e-05, + "loss": 0.0057, + "step": 70710 + }, + { + "epoch": 10.984777881329606, + "grad_norm": 0.15209995210170746, + "learning_rate": 2.166515479297229e-05, + "loss": 0.006, + "step": 70720 + }, + { + "epoch": 10.986331158744951, + "grad_norm": 0.2437264621257782, + "learning_rate": 2.1651532969145732e-05, + "loss": 0.0062, + "step": 70730 + }, + { + "epoch": 10.987884436160298, + "grad_norm": 0.13338075578212738, + "learning_rate": 2.1637914245464903e-05, + "loss": 0.0062, + "step": 70740 + }, + { + "epoch": 10.989437713575645, + "grad_norm": 0.31522297859191895, + "learning_rate": 2.1624298623419133e-05, + "loss": 0.0057, + "step": 70750 + }, + { + "epoch": 10.99099099099099, + "grad_norm": 0.114352747797966, + "learning_rate": 2.1610686104497413e-05, + "loss": 0.0053, + "step": 70760 + }, + { + "epoch": 10.992544268406338, + "grad_norm": 0.3262905180454254, + "learning_rate": 2.1597076690188366e-05, + "loss": 0.0061, + "step": 70770 + }, + { + "epoch": 10.994097545821683, + "grad_norm": 0.17193379998207092, + "learning_rate": 2.1583470381980315e-05, + "loss": 0.0038, + "step": 70780 + }, + { + "epoch": 10.99565082323703, + "grad_norm": 0.15734851360321045, + "learning_rate": 2.156986718136123e-05, + "loss": 0.005, + "step": 70790 + }, + { + "epoch": 10.997204100652377, + "grad_norm": 0.3851509392261505, + "learning_rate": 2.155626708981871e-05, + "loss": 0.0083, + "step": 70800 + }, + { + "epoch": 10.998757378067722, + "grad_norm": 0.08077817410230637, + "learning_rate": 2.1542670108840075e-05, + "loss": 0.0047, + "step": 70810 + }, + { + "epoch": 11.00031065548307, + "grad_norm": 1.1493887901306152, + "learning_rate": 2.152907623991221e-05, + "loss": 0.0051, + "step": 70820 + }, + { + "epoch": 11.001863932898416, + "grad_norm": 0.3941313624382019, + "learning_rate": 2.1515485484521802e-05, + "loss": 0.0056, + "step": 70830 + }, + { + "epoch": 11.003417210313762, + "grad_norm": 0.1941499263048172, + "learning_rate": 2.150189784415506e-05, + "loss": 0.0065, + "step": 70840 + }, + { + "epoch": 11.004970487729109, + "grad_norm": 0.15321417152881622, + "learning_rate": 2.148831332029794e-05, + "loss": 0.0055, + "step": 70850 + }, + { + "epoch": 11.006523765144454, + "grad_norm": 0.11669991165399551, + "learning_rate": 2.147473191443599e-05, + "loss": 0.0035, + "step": 70860 + }, + { + "epoch": 11.008077042559801, + "grad_norm": 0.1931895762681961, + "learning_rate": 2.1461153628054492e-05, + "loss": 0.0031, + "step": 70870 + }, + { + "epoch": 11.009630319975148, + "grad_norm": 0.37454771995544434, + "learning_rate": 2.1447578462638284e-05, + "loss": 0.0051, + "step": 70880 + }, + { + "epoch": 11.011183597390493, + "grad_norm": 0.1975000649690628, + "learning_rate": 2.143400641967201e-05, + "loss": 0.0032, + "step": 70890 + }, + { + "epoch": 11.01273687480584, + "grad_norm": 0.13747546076774597, + "learning_rate": 2.1420437500639823e-05, + "loss": 0.0051, + "step": 70900 + }, + { + "epoch": 11.014290152221188, + "grad_norm": 0.13758593797683716, + "learning_rate": 2.1406871707025646e-05, + "loss": 0.0057, + "step": 70910 + }, + { + "epoch": 11.015843429636533, + "grad_norm": 0.1560998558998108, + "learning_rate": 2.1393309040312963e-05, + "loss": 0.0036, + "step": 70920 + }, + { + "epoch": 11.01739670705188, + "grad_norm": 0.22731991112232208, + "learning_rate": 2.137974950198502e-05, + "loss": 0.0061, + "step": 70930 + }, + { + "epoch": 11.018949984467225, + "grad_norm": 0.15172980725765228, + "learning_rate": 2.1366193093524617e-05, + "loss": 0.0039, + "step": 70940 + }, + { + "epoch": 11.020503261882572, + "grad_norm": 0.17130140960216522, + "learning_rate": 2.1352639816414288e-05, + "loss": 0.0049, + "step": 70950 + }, + { + "epoch": 11.02205653929792, + "grad_norm": 0.17928123474121094, + "learning_rate": 2.1339089672136202e-05, + "loss": 0.0037, + "step": 70960 + }, + { + "epoch": 11.023609816713265, + "grad_norm": 0.35975930094718933, + "learning_rate": 2.1325542662172193e-05, + "loss": 0.0053, + "step": 70970 + }, + { + "epoch": 11.025163094128612, + "grad_norm": 0.185054749250412, + "learning_rate": 2.1311998788003712e-05, + "loss": 0.0045, + "step": 70980 + }, + { + "epoch": 11.026716371543957, + "grad_norm": 0.2383609563112259, + "learning_rate": 2.1298458051111907e-05, + "loss": 0.0041, + "step": 70990 + }, + { + "epoch": 11.028269648959304, + "grad_norm": 0.16152220964431763, + "learning_rate": 2.1284920452977602e-05, + "loss": 0.0049, + "step": 71000 + }, + { + "epoch": 11.029822926374651, + "grad_norm": 0.1142956018447876, + "learning_rate": 2.12713859950812e-05, + "loss": 0.0052, + "step": 71010 + }, + { + "epoch": 11.031376203789996, + "grad_norm": 0.24865463376045227, + "learning_rate": 2.1257854678902832e-05, + "loss": 0.0045, + "step": 71020 + }, + { + "epoch": 11.032929481205343, + "grad_norm": 0.22322514653205872, + "learning_rate": 2.124432650592227e-05, + "loss": 0.0076, + "step": 71030 + }, + { + "epoch": 11.03448275862069, + "grad_norm": 0.17944207787513733, + "learning_rate": 2.1230801477618937e-05, + "loss": 0.0043, + "step": 71040 + }, + { + "epoch": 11.036036036036036, + "grad_norm": 0.22995510697364807, + "learning_rate": 2.121727959547189e-05, + "loss": 0.0081, + "step": 71050 + }, + { + "epoch": 11.037589313451383, + "grad_norm": 0.15338961780071259, + "learning_rate": 2.1203760860959886e-05, + "loss": 0.0039, + "step": 71060 + }, + { + "epoch": 11.039142590866728, + "grad_norm": 0.10466353595256805, + "learning_rate": 2.119024527556127e-05, + "loss": 0.0076, + "step": 71070 + }, + { + "epoch": 11.040695868282075, + "grad_norm": 0.15386329591274261, + "learning_rate": 2.117673284075415e-05, + "loss": 0.0038, + "step": 71080 + }, + { + "epoch": 11.042249145697422, + "grad_norm": 0.1813739538192749, + "learning_rate": 2.1163223558016165e-05, + "loss": 0.0055, + "step": 71090 + }, + { + "epoch": 11.043802423112767, + "grad_norm": 0.20983678102493286, + "learning_rate": 2.1149717428824722e-05, + "loss": 0.0035, + "step": 71100 + }, + { + "epoch": 11.045355700528114, + "grad_norm": 0.13092073798179626, + "learning_rate": 2.1136214454656784e-05, + "loss": 0.0044, + "step": 71110 + }, + { + "epoch": 11.046908977943461, + "grad_norm": 0.253780722618103, + "learning_rate": 2.1122714636989056e-05, + "loss": 0.0052, + "step": 71120 + }, + { + "epoch": 11.048462255358807, + "grad_norm": 0.18078631162643433, + "learning_rate": 2.1109217977297823e-05, + "loss": 0.0036, + "step": 71130 + }, + { + "epoch": 11.050015532774154, + "grad_norm": 0.37760207056999207, + "learning_rate": 2.1095724477059077e-05, + "loss": 0.0051, + "step": 71140 + }, + { + "epoch": 11.051568810189499, + "grad_norm": 0.29003408551216125, + "learning_rate": 2.1082234137748434e-05, + "loss": 0.0049, + "step": 71150 + }, + { + "epoch": 11.053122087604846, + "grad_norm": 0.3939213454723358, + "learning_rate": 2.106874696084122e-05, + "loss": 0.0038, + "step": 71160 + }, + { + "epoch": 11.054675365020193, + "grad_norm": 0.24258936941623688, + "learning_rate": 2.1055262947812315e-05, + "loss": 0.0053, + "step": 71170 + }, + { + "epoch": 11.056228642435538, + "grad_norm": 0.19285640120506287, + "learning_rate": 2.104178210013636e-05, + "loss": 0.0048, + "step": 71180 + }, + { + "epoch": 11.057781919850886, + "grad_norm": 0.14733253419399261, + "learning_rate": 2.1028304419287554e-05, + "loss": 0.0062, + "step": 71190 + }, + { + "epoch": 11.059335197266233, + "grad_norm": 0.2484322190284729, + "learning_rate": 2.101482990673982e-05, + "loss": 0.0071, + "step": 71200 + }, + { + "epoch": 11.060888474681578, + "grad_norm": 0.1858920305967331, + "learning_rate": 2.1001358563966707e-05, + "loss": 0.0041, + "step": 71210 + }, + { + "epoch": 11.062441752096925, + "grad_norm": 0.4510752856731415, + "learning_rate": 2.098789039244144e-05, + "loss": 0.005, + "step": 71220 + }, + { + "epoch": 11.06399502951227, + "grad_norm": 0.17595535516738892, + "learning_rate": 2.097442539363685e-05, + "loss": 0.0054, + "step": 71230 + }, + { + "epoch": 11.065548306927617, + "grad_norm": 0.3926459550857544, + "learning_rate": 2.0960963569025448e-05, + "loss": 0.0043, + "step": 71240 + }, + { + "epoch": 11.067101584342964, + "grad_norm": 0.3354703485965729, + "learning_rate": 2.0947504920079435e-05, + "loss": 0.0045, + "step": 71250 + }, + { + "epoch": 11.06865486175831, + "grad_norm": 0.8135310411453247, + "learning_rate": 2.093404944827059e-05, + "loss": 0.0035, + "step": 71260 + }, + { + "epoch": 11.070208139173657, + "grad_norm": 0.4338677227497101, + "learning_rate": 2.0920597155070394e-05, + "loss": 0.0054, + "step": 71270 + }, + { + "epoch": 11.071761416589002, + "grad_norm": 0.28985095024108887, + "learning_rate": 2.090714804194997e-05, + "loss": 0.0036, + "step": 71280 + }, + { + "epoch": 11.073314694004349, + "grad_norm": 0.13348977267742157, + "learning_rate": 2.0893702110380114e-05, + "loss": 0.0045, + "step": 71290 + }, + { + "epoch": 11.074867971419696, + "grad_norm": 0.6459881663322449, + "learning_rate": 2.0880259361831216e-05, + "loss": 0.0072, + "step": 71300 + }, + { + "epoch": 11.076421248835041, + "grad_norm": 0.2711358964443207, + "learning_rate": 2.0866819797773397e-05, + "loss": 0.0047, + "step": 71310 + }, + { + "epoch": 11.077974526250388, + "grad_norm": 0.24167034029960632, + "learning_rate": 2.085338341967634e-05, + "loss": 0.0034, + "step": 71320 + }, + { + "epoch": 11.079527803665735, + "grad_norm": 0.6666051745414734, + "learning_rate": 2.083995022900946e-05, + "loss": 0.0042, + "step": 71330 + }, + { + "epoch": 11.08108108108108, + "grad_norm": 0.24606357514858246, + "learning_rate": 2.0826520227241768e-05, + "loss": 0.0057, + "step": 71340 + }, + { + "epoch": 11.082634358496428, + "grad_norm": 0.32502588629722595, + "learning_rate": 2.081309341584199e-05, + "loss": 0.0049, + "step": 71350 + }, + { + "epoch": 11.084187635911773, + "grad_norm": 0.2589203417301178, + "learning_rate": 2.0799669796278413e-05, + "loss": 0.0038, + "step": 71360 + }, + { + "epoch": 11.08574091332712, + "grad_norm": 0.11738696694374084, + "learning_rate": 2.078624937001906e-05, + "loss": 0.005, + "step": 71370 + }, + { + "epoch": 11.087294190742467, + "grad_norm": 0.27427080273628235, + "learning_rate": 2.0772832138531544e-05, + "loss": 0.0049, + "step": 71380 + }, + { + "epoch": 11.088847468157812, + "grad_norm": 0.15168704092502594, + "learning_rate": 2.0759418103283158e-05, + "loss": 0.0045, + "step": 71390 + }, + { + "epoch": 11.09040074557316, + "grad_norm": 0.2505611479282379, + "learning_rate": 2.0746007265740846e-05, + "loss": 0.0071, + "step": 71400 + }, + { + "epoch": 11.091954022988507, + "grad_norm": 0.4690091907978058, + "learning_rate": 2.0732599627371215e-05, + "loss": 0.0068, + "step": 71410 + }, + { + "epoch": 11.093507300403852, + "grad_norm": 0.276019424200058, + "learning_rate": 2.0719195189640468e-05, + "loss": 0.0061, + "step": 71420 + }, + { + "epoch": 11.095060577819199, + "grad_norm": 0.12974144518375397, + "learning_rate": 2.0705793954014528e-05, + "loss": 0.005, + "step": 71430 + }, + { + "epoch": 11.096613855234544, + "grad_norm": 0.6676236391067505, + "learning_rate": 2.06923959219589e-05, + "loss": 0.0031, + "step": 71440 + }, + { + "epoch": 11.098167132649891, + "grad_norm": 0.11593053489923477, + "learning_rate": 2.0679001094938793e-05, + "loss": 0.0035, + "step": 71450 + }, + { + "epoch": 11.099720410065238, + "grad_norm": 0.13423849642276764, + "learning_rate": 2.066560947441904e-05, + "loss": 0.0041, + "step": 71460 + }, + { + "epoch": 11.101273687480584, + "grad_norm": 0.19772930443286896, + "learning_rate": 2.0652221061864152e-05, + "loss": 0.003, + "step": 71470 + }, + { + "epoch": 11.10282696489593, + "grad_norm": 0.42647936940193176, + "learning_rate": 2.0638835858738226e-05, + "loss": 0.0057, + "step": 71480 + }, + { + "epoch": 11.104380242311278, + "grad_norm": 0.25464391708374023, + "learning_rate": 2.0625453866505068e-05, + "loss": 0.0047, + "step": 71490 + }, + { + "epoch": 11.105933519726623, + "grad_norm": 0.26872873306274414, + "learning_rate": 2.0612075086628123e-05, + "loss": 0.0051, + "step": 71500 + }, + { + "epoch": 11.10748679714197, + "grad_norm": 0.24042092263698578, + "learning_rate": 2.059869952057045e-05, + "loss": 0.0067, + "step": 71510 + }, + { + "epoch": 11.109040074557315, + "grad_norm": 0.1844133883714676, + "learning_rate": 2.0585327169794793e-05, + "loss": 0.0058, + "step": 71520 + }, + { + "epoch": 11.110593351972662, + "grad_norm": 0.2359984815120697, + "learning_rate": 2.0571958035763534e-05, + "loss": 0.0055, + "step": 71530 + }, + { + "epoch": 11.11214662938801, + "grad_norm": 0.21592886745929718, + "learning_rate": 2.0558592119938715e-05, + "loss": 0.0042, + "step": 71540 + }, + { + "epoch": 11.113699906803355, + "grad_norm": 0.31756868958473206, + "learning_rate": 2.054522942378198e-05, + "loss": 0.0031, + "step": 71550 + }, + { + "epoch": 11.115253184218702, + "grad_norm": 0.32961133122444153, + "learning_rate": 2.053186994875469e-05, + "loss": 0.0042, + "step": 71560 + }, + { + "epoch": 11.116806461634049, + "grad_norm": 0.34911590814590454, + "learning_rate": 2.0518513696317775e-05, + "loss": 0.0048, + "step": 71570 + }, + { + "epoch": 11.118359739049394, + "grad_norm": 0.16567832231521606, + "learning_rate": 2.050516066793188e-05, + "loss": 0.0051, + "step": 71580 + }, + { + "epoch": 11.119913016464741, + "grad_norm": 0.5268526077270508, + "learning_rate": 2.0491810865057276e-05, + "loss": 0.0058, + "step": 71590 + }, + { + "epoch": 11.121466293880086, + "grad_norm": 0.18885429203510284, + "learning_rate": 2.0478464289153882e-05, + "loss": 0.0063, + "step": 71600 + }, + { + "epoch": 11.123019571295433, + "grad_norm": 0.31495848298072815, + "learning_rate": 2.0465120941681232e-05, + "loss": 0.0051, + "step": 71610 + }, + { + "epoch": 11.12457284871078, + "grad_norm": 0.10716014355421066, + "learning_rate": 2.0451780824098566e-05, + "loss": 0.0066, + "step": 71620 + }, + { + "epoch": 11.126126126126126, + "grad_norm": 0.1613614410161972, + "learning_rate": 2.043844393786471e-05, + "loss": 0.006, + "step": 71630 + }, + { + "epoch": 11.127679403541473, + "grad_norm": 0.14106778800487518, + "learning_rate": 2.0425110284438193e-05, + "loss": 0.0031, + "step": 71640 + }, + { + "epoch": 11.129232680956818, + "grad_norm": 0.14979293942451477, + "learning_rate": 2.041177986527712e-05, + "loss": 0.0037, + "step": 71650 + }, + { + "epoch": 11.130785958372165, + "grad_norm": 0.1960051953792572, + "learning_rate": 2.0398452681839347e-05, + "loss": 0.0052, + "step": 71660 + }, + { + "epoch": 11.132339235787512, + "grad_norm": 0.2301163524389267, + "learning_rate": 2.0385128735582272e-05, + "loss": 0.0034, + "step": 71670 + }, + { + "epoch": 11.133892513202857, + "grad_norm": 0.40294554829597473, + "learning_rate": 2.0371808027963003e-05, + "loss": 0.0043, + "step": 71680 + }, + { + "epoch": 11.135445790618204, + "grad_norm": 0.23776167631149292, + "learning_rate": 2.035849056043825e-05, + "loss": 0.0051, + "step": 71690 + }, + { + "epoch": 11.136999068033552, + "grad_norm": 0.2329411804676056, + "learning_rate": 2.03451763344644e-05, + "loss": 0.0058, + "step": 71700 + }, + { + "epoch": 11.138552345448897, + "grad_norm": 0.11418352276086807, + "learning_rate": 2.0331865351497482e-05, + "loss": 0.0035, + "step": 71710 + }, + { + "epoch": 11.140105622864244, + "grad_norm": 0.12014733999967575, + "learning_rate": 2.0318557612993185e-05, + "loss": 0.0043, + "step": 71720 + }, + { + "epoch": 11.14165890027959, + "grad_norm": 0.136143758893013, + "learning_rate": 2.0305253120406782e-05, + "loss": 0.0043, + "step": 71730 + }, + { + "epoch": 11.143212177694936, + "grad_norm": 0.3555728495121002, + "learning_rate": 2.029195187519325e-05, + "loss": 0.0043, + "step": 71740 + }, + { + "epoch": 11.144765455110283, + "grad_norm": 0.48968058824539185, + "learning_rate": 2.0278653878807217e-05, + "loss": 0.0054, + "step": 71750 + }, + { + "epoch": 11.146318732525629, + "grad_norm": 0.2529805302619934, + "learning_rate": 2.0265359132702892e-05, + "loss": 0.0031, + "step": 71760 + }, + { + "epoch": 11.147872009940976, + "grad_norm": 0.0941871777176857, + "learning_rate": 2.0252067638334182e-05, + "loss": 0.0051, + "step": 71770 + }, + { + "epoch": 11.149425287356323, + "grad_norm": 0.3353314697742462, + "learning_rate": 2.023877939715464e-05, + "loss": 0.004, + "step": 71780 + }, + { + "epoch": 11.150978564771668, + "grad_norm": 0.16586321592330933, + "learning_rate": 2.0225494410617448e-05, + "loss": 0.0045, + "step": 71790 + }, + { + "epoch": 11.152531842187015, + "grad_norm": 0.15890271961688995, + "learning_rate": 2.02122126801754e-05, + "loss": 0.005, + "step": 71800 + }, + { + "epoch": 11.15408511960236, + "grad_norm": 0.20330360531806946, + "learning_rate": 2.0198934207281018e-05, + "loss": 0.0034, + "step": 71810 + }, + { + "epoch": 11.155638397017707, + "grad_norm": 0.20051635801792145, + "learning_rate": 2.0185658993386357e-05, + "loss": 0.0052, + "step": 71820 + }, + { + "epoch": 11.157191674433054, + "grad_norm": 0.24551253020763397, + "learning_rate": 2.0172387039943226e-05, + "loss": 0.0064, + "step": 71830 + }, + { + "epoch": 11.1587449518484, + "grad_norm": 0.24536091089248657, + "learning_rate": 2.0159118348402968e-05, + "loss": 0.0039, + "step": 71840 + }, + { + "epoch": 11.160298229263747, + "grad_norm": 0.30936747789382935, + "learning_rate": 2.0145852920216697e-05, + "loss": 0.0059, + "step": 71850 + }, + { + "epoch": 11.161851506679094, + "grad_norm": 0.4175112843513489, + "learning_rate": 2.0132590756835045e-05, + "loss": 0.0065, + "step": 71860 + }, + { + "epoch": 11.163404784094439, + "grad_norm": 0.1888560801744461, + "learning_rate": 2.0119331859708386e-05, + "loss": 0.0045, + "step": 71870 + }, + { + "epoch": 11.164958061509786, + "grad_norm": 0.26300349831581116, + "learning_rate": 2.0106076230286654e-05, + "loss": 0.0028, + "step": 71880 + }, + { + "epoch": 11.166511338925131, + "grad_norm": 0.4095747768878937, + "learning_rate": 2.0092823870019488e-05, + "loss": 0.0067, + "step": 71890 + }, + { + "epoch": 11.168064616340478, + "grad_norm": 0.17434152960777283, + "learning_rate": 2.0079574780356115e-05, + "loss": 0.009, + "step": 71900 + }, + { + "epoch": 11.169617893755825, + "grad_norm": 0.38022565841674805, + "learning_rate": 2.0066328962745486e-05, + "loss": 0.0056, + "step": 71910 + }, + { + "epoch": 11.17117117117117, + "grad_norm": 0.310102254152298, + "learning_rate": 2.00530864186361e-05, + "loss": 0.0048, + "step": 71920 + }, + { + "epoch": 11.172724448586518, + "grad_norm": 0.2775309085845947, + "learning_rate": 2.003984714947618e-05, + "loss": 0.0049, + "step": 71930 + }, + { + "epoch": 11.174277726001863, + "grad_norm": 0.19196107983589172, + "learning_rate": 2.0026611156713515e-05, + "loss": 0.0048, + "step": 71940 + }, + { + "epoch": 11.17583100341721, + "grad_norm": 0.36948299407958984, + "learning_rate": 2.0013378441795584e-05, + "loss": 0.0051, + "step": 71950 + }, + { + "epoch": 11.177384280832557, + "grad_norm": 0.42509961128234863, + "learning_rate": 2.0000149006169506e-05, + "loss": 0.006, + "step": 71960 + }, + { + "epoch": 11.178937558247902, + "grad_norm": 0.28975698351860046, + "learning_rate": 1.9986922851282048e-05, + "loss": 0.0029, + "step": 71970 + }, + { + "epoch": 11.18049083566325, + "grad_norm": 0.11436290293931961, + "learning_rate": 1.997369997857956e-05, + "loss": 0.0044, + "step": 71980 + }, + { + "epoch": 11.182044113078597, + "grad_norm": 0.12665167450904846, + "learning_rate": 1.9960480389508103e-05, + "loss": 0.0029, + "step": 71990 + }, + { + "epoch": 11.183597390493942, + "grad_norm": 0.23395222425460815, + "learning_rate": 1.9947264085513363e-05, + "loss": 0.0091, + "step": 72000 + }, + { + "epoch": 11.185150667909289, + "grad_norm": 0.0892309695482254, + "learning_rate": 1.993405106804062e-05, + "loss": 0.0066, + "step": 72010 + }, + { + "epoch": 11.186703945324634, + "grad_norm": 0.10203956812620163, + "learning_rate": 1.9920841338534867e-05, + "loss": 0.0059, + "step": 72020 + }, + { + "epoch": 11.188257222739981, + "grad_norm": 0.20119620859622955, + "learning_rate": 1.9907634898440646e-05, + "loss": 0.0039, + "step": 72030 + }, + { + "epoch": 11.189810500155328, + "grad_norm": 0.201751247048378, + "learning_rate": 1.989443174920227e-05, + "loss": 0.0025, + "step": 72040 + }, + { + "epoch": 11.191363777570674, + "grad_norm": 0.12900903820991516, + "learning_rate": 1.988123189226355e-05, + "loss": 0.0048, + "step": 72050 + }, + { + "epoch": 11.19291705498602, + "grad_norm": 0.2406466156244278, + "learning_rate": 1.986803532906805e-05, + "loss": 0.0062, + "step": 72060 + }, + { + "epoch": 11.194470332401368, + "grad_norm": 0.1678350418806076, + "learning_rate": 1.985484206105889e-05, + "loss": 0.0052, + "step": 72070 + }, + { + "epoch": 11.196023609816713, + "grad_norm": 0.5716090202331543, + "learning_rate": 1.9841652089678887e-05, + "loss": 0.0042, + "step": 72080 + }, + { + "epoch": 11.19757688723206, + "grad_norm": 0.1362645924091339, + "learning_rate": 1.9828465416370435e-05, + "loss": 0.0051, + "step": 72090 + }, + { + "epoch": 11.199130164647405, + "grad_norm": 0.15531697869300842, + "learning_rate": 1.9815282042575685e-05, + "loss": 0.0046, + "step": 72100 + }, + { + "epoch": 11.200683442062752, + "grad_norm": 0.13970038294792175, + "learning_rate": 1.9802101969736285e-05, + "loss": 0.0037, + "step": 72110 + }, + { + "epoch": 11.2022367194781, + "grad_norm": 0.2703031003475189, + "learning_rate": 1.9788925199293633e-05, + "loss": 0.0039, + "step": 72120 + }, + { + "epoch": 11.203789996893445, + "grad_norm": 2.8649463653564453, + "learning_rate": 1.977575173268868e-05, + "loss": 0.0045, + "step": 72130 + }, + { + "epoch": 11.205343274308792, + "grad_norm": 0.3563433885574341, + "learning_rate": 1.9762581571362094e-05, + "loss": 0.0075, + "step": 72140 + }, + { + "epoch": 11.206896551724139, + "grad_norm": 0.1817062646150589, + "learning_rate": 1.9749414716754112e-05, + "loss": 0.0039, + "step": 72150 + }, + { + "epoch": 11.208449829139484, + "grad_norm": 0.24427232146263123, + "learning_rate": 1.973625117030465e-05, + "loss": 0.0069, + "step": 72160 + }, + { + "epoch": 11.210003106554831, + "grad_norm": 1.0237871408462524, + "learning_rate": 1.972309093345326e-05, + "loss": 0.0052, + "step": 72170 + }, + { + "epoch": 11.211556383970176, + "grad_norm": 0.41570988297462463, + "learning_rate": 1.970993400763915e-05, + "loss": 0.0048, + "step": 72180 + }, + { + "epoch": 11.213109661385523, + "grad_norm": 0.31468772888183594, + "learning_rate": 1.969678039430109e-05, + "loss": 0.0049, + "step": 72190 + }, + { + "epoch": 11.21466293880087, + "grad_norm": 0.17340224981307983, + "learning_rate": 1.9683630094877564e-05, + "loss": 0.0052, + "step": 72200 + }, + { + "epoch": 11.216216216216216, + "grad_norm": 0.4300670027732849, + "learning_rate": 1.9670483110806696e-05, + "loss": 0.0041, + "step": 72210 + }, + { + "epoch": 11.217769493631563, + "grad_norm": 0.2825261950492859, + "learning_rate": 1.9657339443526172e-05, + "loss": 0.0058, + "step": 72220 + }, + { + "epoch": 11.219322771046908, + "grad_norm": 0.341659814119339, + "learning_rate": 1.964419909447339e-05, + "loss": 0.0042, + "step": 72230 + }, + { + "epoch": 11.220876048462255, + "grad_norm": 0.11549928784370422, + "learning_rate": 1.963106206508536e-05, + "loss": 0.0047, + "step": 72240 + }, + { + "epoch": 11.222429325877602, + "grad_norm": 0.4247821271419525, + "learning_rate": 1.9617928356798738e-05, + "loss": 0.0055, + "step": 72250 + }, + { + "epoch": 11.223982603292948, + "grad_norm": 0.334469199180603, + "learning_rate": 1.960479797104977e-05, + "loss": 0.004, + "step": 72260 + }, + { + "epoch": 11.225535880708295, + "grad_norm": 0.10604409873485565, + "learning_rate": 1.9591670909274418e-05, + "loss": 0.0042, + "step": 72270 + }, + { + "epoch": 11.227089158123642, + "grad_norm": 0.2030370682477951, + "learning_rate": 1.9578547172908184e-05, + "loss": 0.0044, + "step": 72280 + }, + { + "epoch": 11.228642435538987, + "grad_norm": 0.44812721014022827, + "learning_rate": 1.9565426763386325e-05, + "loss": 0.0054, + "step": 72290 + }, + { + "epoch": 11.230195712954334, + "grad_norm": 0.15990184247493744, + "learning_rate": 1.9552309682143622e-05, + "loss": 0.004, + "step": 72300 + }, + { + "epoch": 11.23174899036968, + "grad_norm": 0.5360035300254822, + "learning_rate": 1.9539195930614574e-05, + "loss": 0.006, + "step": 72310 + }, + { + "epoch": 11.233302267785026, + "grad_norm": 0.5669746398925781, + "learning_rate": 1.952608551023324e-05, + "loss": 0.0045, + "step": 72320 + }, + { + "epoch": 11.234855545200373, + "grad_norm": 0.11231599748134613, + "learning_rate": 1.9512978422433397e-05, + "loss": 0.0045, + "step": 72330 + }, + { + "epoch": 11.236408822615719, + "grad_norm": 0.13078445196151733, + "learning_rate": 1.9499874668648376e-05, + "loss": 0.0038, + "step": 72340 + }, + { + "epoch": 11.237962100031066, + "grad_norm": 0.5992356538772583, + "learning_rate": 1.9486774250311213e-05, + "loss": 0.0042, + "step": 72350 + }, + { + "epoch": 11.239515377446413, + "grad_norm": 0.1848103404045105, + "learning_rate": 1.9473677168854537e-05, + "loss": 0.0051, + "step": 72360 + }, + { + "epoch": 11.241068654861758, + "grad_norm": 0.19147251546382904, + "learning_rate": 1.946058342571065e-05, + "loss": 0.0062, + "step": 72370 + }, + { + "epoch": 11.242621932277105, + "grad_norm": 0.23384517431259155, + "learning_rate": 1.9447493022311424e-05, + "loss": 0.0049, + "step": 72380 + }, + { + "epoch": 11.24417520969245, + "grad_norm": 0.1697472333908081, + "learning_rate": 1.9434405960088442e-05, + "loss": 0.0032, + "step": 72390 + }, + { + "epoch": 11.245728487107797, + "grad_norm": 0.1486765593290329, + "learning_rate": 1.9421322240472857e-05, + "loss": 0.006, + "step": 72400 + }, + { + "epoch": 11.247281764523144, + "grad_norm": 0.29063937067985535, + "learning_rate": 1.9408241864895494e-05, + "loss": 0.0041, + "step": 72410 + }, + { + "epoch": 11.24883504193849, + "grad_norm": 0.3313412368297577, + "learning_rate": 1.939516483478681e-05, + "loss": 0.0039, + "step": 72420 + }, + { + "epoch": 11.250388319353837, + "grad_norm": 0.13005489110946655, + "learning_rate": 1.9382091151576902e-05, + "loss": 0.005, + "step": 72430 + }, + { + "epoch": 11.251941596769184, + "grad_norm": 0.16149769723415375, + "learning_rate": 1.9369020816695454e-05, + "loss": 0.0044, + "step": 72440 + }, + { + "epoch": 11.253494874184529, + "grad_norm": 0.3652762174606323, + "learning_rate": 1.935595383157184e-05, + "loss": 0.007, + "step": 72450 + }, + { + "epoch": 11.255048151599876, + "grad_norm": 0.10736048966646194, + "learning_rate": 1.9342890197635057e-05, + "loss": 0.0054, + "step": 72460 + }, + { + "epoch": 11.256601429015221, + "grad_norm": 0.2647077143192291, + "learning_rate": 1.9329829916313685e-05, + "loss": 0.0041, + "step": 72470 + }, + { + "epoch": 11.258154706430568, + "grad_norm": 0.23333996534347534, + "learning_rate": 1.9316772989036013e-05, + "loss": 0.0043, + "step": 72480 + }, + { + "epoch": 11.259707983845916, + "grad_norm": 0.26964327692985535, + "learning_rate": 1.930371941722991e-05, + "loss": 0.0062, + "step": 72490 + }, + { + "epoch": 11.26126126126126, + "grad_norm": 0.33128491044044495, + "learning_rate": 1.929066920232292e-05, + "loss": 0.0048, + "step": 72500 + }, + { + "epoch": 11.262814538676608, + "grad_norm": 0.21283192932605743, + "learning_rate": 1.9277622345742152e-05, + "loss": 0.0039, + "step": 72510 + }, + { + "epoch": 11.264367816091955, + "grad_norm": 0.11359206587076187, + "learning_rate": 1.926457884891444e-05, + "loss": 0.0045, + "step": 72520 + }, + { + "epoch": 11.2659210935073, + "grad_norm": 0.2165306657552719, + "learning_rate": 1.9251538713266147e-05, + "loss": 0.0039, + "step": 72530 + }, + { + "epoch": 11.267474370922647, + "grad_norm": 0.1087597981095314, + "learning_rate": 1.9238501940223358e-05, + "loss": 0.0051, + "step": 72540 + }, + { + "epoch": 11.269027648337993, + "grad_norm": 0.12469678372144699, + "learning_rate": 1.922546853121175e-05, + "loss": 0.0041, + "step": 72550 + }, + { + "epoch": 11.27058092575334, + "grad_norm": 0.1556176394224167, + "learning_rate": 1.9212438487656647e-05, + "loss": 0.0049, + "step": 72560 + }, + { + "epoch": 11.272134203168687, + "grad_norm": 0.23082755506038666, + "learning_rate": 1.9199411810982965e-05, + "loss": 0.0049, + "step": 72570 + }, + { + "epoch": 11.273687480584032, + "grad_norm": 0.1898658275604248, + "learning_rate": 1.918638850261532e-05, + "loss": 0.0057, + "step": 72580 + }, + { + "epoch": 11.275240757999379, + "grad_norm": 0.29646703600883484, + "learning_rate": 1.9173368563977885e-05, + "loss": 0.0039, + "step": 72590 + }, + { + "epoch": 11.276794035414724, + "grad_norm": 0.31475207209587097, + "learning_rate": 1.9160351996494513e-05, + "loss": 0.0054, + "step": 72600 + }, + { + "epoch": 11.278347312830071, + "grad_norm": 0.3417920768260956, + "learning_rate": 1.9147338801588677e-05, + "loss": 0.0039, + "step": 72610 + }, + { + "epoch": 11.279900590245418, + "grad_norm": 0.08862005919218063, + "learning_rate": 1.913432898068351e-05, + "loss": 0.004, + "step": 72620 + }, + { + "epoch": 11.281453867660764, + "grad_norm": 0.3165111243724823, + "learning_rate": 1.912132253520169e-05, + "loss": 0.0049, + "step": 72630 + }, + { + "epoch": 11.28300714507611, + "grad_norm": 0.33439281582832336, + "learning_rate": 1.910831946656564e-05, + "loss": 0.0049, + "step": 72640 + }, + { + "epoch": 11.284560422491458, + "grad_norm": 0.44204944372177124, + "learning_rate": 1.909531977619731e-05, + "loss": 0.0043, + "step": 72650 + }, + { + "epoch": 11.286113699906803, + "grad_norm": 0.1872558444738388, + "learning_rate": 1.908232346551834e-05, + "loss": 0.0047, + "step": 72660 + }, + { + "epoch": 11.28766697732215, + "grad_norm": 0.16098690032958984, + "learning_rate": 1.9069330535949993e-05, + "loss": 0.005, + "step": 72670 + }, + { + "epoch": 11.289220254737495, + "grad_norm": 0.15850567817687988, + "learning_rate": 1.9056340988913167e-05, + "loss": 0.0041, + "step": 72680 + }, + { + "epoch": 11.290773532152842, + "grad_norm": 0.21558475494384766, + "learning_rate": 1.9043354825828357e-05, + "loss": 0.0042, + "step": 72690 + }, + { + "epoch": 11.29232680956819, + "grad_norm": 0.28967955708503723, + "learning_rate": 1.9030372048115724e-05, + "loss": 0.0072, + "step": 72700 + }, + { + "epoch": 11.293880086983535, + "grad_norm": 0.27605611085891724, + "learning_rate": 1.9017392657195056e-05, + "loss": 0.0033, + "step": 72710 + }, + { + "epoch": 11.295433364398882, + "grad_norm": 0.44577884674072266, + "learning_rate": 1.9004416654485724e-05, + "loss": 0.0033, + "step": 72720 + }, + { + "epoch": 11.296986641814229, + "grad_norm": 0.43744999170303345, + "learning_rate": 1.8991444041406785e-05, + "loss": 0.0044, + "step": 72730 + }, + { + "epoch": 11.298539919229574, + "grad_norm": 0.28598272800445557, + "learning_rate": 1.8978474819376912e-05, + "loss": 0.0058, + "step": 72740 + }, + { + "epoch": 11.300093196644921, + "grad_norm": 0.1399775743484497, + "learning_rate": 1.8965508989814406e-05, + "loss": 0.0057, + "step": 72750 + }, + { + "epoch": 11.301646474060266, + "grad_norm": 0.3167654275894165, + "learning_rate": 1.8952546554137158e-05, + "loss": 0.0046, + "step": 72760 + }, + { + "epoch": 11.303199751475614, + "grad_norm": 0.1806868463754654, + "learning_rate": 1.8939587513762757e-05, + "loss": 0.0039, + "step": 72770 + }, + { + "epoch": 11.30475302889096, + "grad_norm": 0.12687325477600098, + "learning_rate": 1.892663187010835e-05, + "loss": 0.0043, + "step": 72780 + }, + { + "epoch": 11.306306306306306, + "grad_norm": 0.198631152510643, + "learning_rate": 1.891367962459077e-05, + "loss": 0.0076, + "step": 72790 + }, + { + "epoch": 11.307859583721653, + "grad_norm": 0.23182813823223114, + "learning_rate": 1.890073077862644e-05, + "loss": 0.0056, + "step": 72800 + }, + { + "epoch": 11.309412861136998, + "grad_norm": 0.12706835567951202, + "learning_rate": 1.888778533363146e-05, + "loss": 0.0062, + "step": 72810 + }, + { + "epoch": 11.310966138552345, + "grad_norm": 0.3429762125015259, + "learning_rate": 1.8874843291021478e-05, + "loss": 0.0041, + "step": 72820 + }, + { + "epoch": 11.312519415967692, + "grad_norm": 0.14406192302703857, + "learning_rate": 1.8861904652211853e-05, + "loss": 0.005, + "step": 72830 + }, + { + "epoch": 11.314072693383038, + "grad_norm": 0.261538028717041, + "learning_rate": 1.8848969418617507e-05, + "loss": 0.0042, + "step": 72840 + }, + { + "epoch": 11.315625970798385, + "grad_norm": 0.10867989808320999, + "learning_rate": 1.8836037591653044e-05, + "loss": 0.0051, + "step": 72850 + }, + { + "epoch": 11.317179248213732, + "grad_norm": 0.28836527466773987, + "learning_rate": 1.882310917273262e-05, + "loss": 0.0033, + "step": 72860 + }, + { + "epoch": 11.318732525629077, + "grad_norm": 0.30055347084999084, + "learning_rate": 1.881018416327014e-05, + "loss": 0.0043, + "step": 72870 + }, + { + "epoch": 11.320285803044424, + "grad_norm": 0.23452240228652954, + "learning_rate": 1.8797262564679014e-05, + "loss": 0.0057, + "step": 72880 + }, + { + "epoch": 11.32183908045977, + "grad_norm": 0.11522645503282547, + "learning_rate": 1.8784344378372355e-05, + "loss": 0.0031, + "step": 72890 + }, + { + "epoch": 11.323392357875116, + "grad_norm": 0.7637314200401306, + "learning_rate": 1.8771429605762843e-05, + "loss": 0.0046, + "step": 72900 + }, + { + "epoch": 11.324945635290463, + "grad_norm": 0.23002588748931885, + "learning_rate": 1.8758518248262835e-05, + "loss": 0.0032, + "step": 72910 + }, + { + "epoch": 11.326498912705809, + "grad_norm": 0.26681095361709595, + "learning_rate": 1.8745610307284305e-05, + "loss": 0.0041, + "step": 72920 + }, + { + "epoch": 11.328052190121156, + "grad_norm": 0.20481616258621216, + "learning_rate": 1.8732705784238862e-05, + "loss": 0.0043, + "step": 72930 + }, + { + "epoch": 11.329605467536503, + "grad_norm": 0.1255795657634735, + "learning_rate": 1.8719804680537682e-05, + "loss": 0.0036, + "step": 72940 + }, + { + "epoch": 11.331158744951848, + "grad_norm": 0.41720035672187805, + "learning_rate": 1.870690699759164e-05, + "loss": 0.0038, + "step": 72950 + }, + { + "epoch": 11.332712022367195, + "grad_norm": 0.12255550175905228, + "learning_rate": 1.8694012736811213e-05, + "loss": 0.0054, + "step": 72960 + }, + { + "epoch": 11.33426529978254, + "grad_norm": 0.6751704216003418, + "learning_rate": 1.8681121899606463e-05, + "loss": 0.0053, + "step": 72970 + }, + { + "epoch": 11.335818577197887, + "grad_norm": 0.40464237332344055, + "learning_rate": 1.8668234487387144e-05, + "loss": 0.0044, + "step": 72980 + }, + { + "epoch": 11.337371854613234, + "grad_norm": 0.23228217661380768, + "learning_rate": 1.8655350501562592e-05, + "loss": 0.0052, + "step": 72990 + }, + { + "epoch": 11.33892513202858, + "grad_norm": 0.1365201622247696, + "learning_rate": 1.8642469943541796e-05, + "loss": 0.0049, + "step": 73000 + }, + { + "epoch": 11.340478409443927, + "grad_norm": 0.3359537124633789, + "learning_rate": 1.8629592814733328e-05, + "loss": 0.0041, + "step": 73010 + }, + { + "epoch": 11.342031686859274, + "grad_norm": 0.18348577618598938, + "learning_rate": 1.8616719116545438e-05, + "loss": 0.0044, + "step": 73020 + }, + { + "epoch": 11.34358496427462, + "grad_norm": 0.2682133615016937, + "learning_rate": 1.8603848850385952e-05, + "loss": 0.0041, + "step": 73030 + }, + { + "epoch": 11.345138241689966, + "grad_norm": 0.26517176628112793, + "learning_rate": 1.859098201766236e-05, + "loss": 0.0052, + "step": 73040 + }, + { + "epoch": 11.346691519105311, + "grad_norm": 0.5137731432914734, + "learning_rate": 1.8578118619781726e-05, + "loss": 0.0063, + "step": 73050 + }, + { + "epoch": 11.348244796520659, + "grad_norm": 0.19090388715267181, + "learning_rate": 1.856525865815083e-05, + "loss": 0.0044, + "step": 73060 + }, + { + "epoch": 11.349798073936006, + "grad_norm": 0.17831991612911224, + "learning_rate": 1.8552402134175966e-05, + "loss": 0.0045, + "step": 73070 + }, + { + "epoch": 11.35135135135135, + "grad_norm": 0.20100019872188568, + "learning_rate": 1.8539549049263143e-05, + "loss": 0.0032, + "step": 73080 + }, + { + "epoch": 11.352904628766698, + "grad_norm": 0.23844829201698303, + "learning_rate": 1.8526699404817917e-05, + "loss": 0.0048, + "step": 73090 + }, + { + "epoch": 11.354457906182045, + "grad_norm": 0.0971560850739479, + "learning_rate": 1.8513853202245546e-05, + "loss": 0.0046, + "step": 73100 + }, + { + "epoch": 11.35601118359739, + "grad_norm": 0.19305390119552612, + "learning_rate": 1.850101044295081e-05, + "loss": 0.005, + "step": 73110 + }, + { + "epoch": 11.357564461012737, + "grad_norm": 0.33015328645706177, + "learning_rate": 1.848817112833825e-05, + "loss": 0.004, + "step": 73120 + }, + { + "epoch": 11.359117738428083, + "grad_norm": 0.2997412085533142, + "learning_rate": 1.8475335259811905e-05, + "loss": 0.0043, + "step": 73130 + }, + { + "epoch": 11.36067101584343, + "grad_norm": 0.15519596636295319, + "learning_rate": 1.8462502838775508e-05, + "loss": 0.0067, + "step": 73140 + }, + { + "epoch": 11.362224293258777, + "grad_norm": 0.13406716287136078, + "learning_rate": 1.8449673866632378e-05, + "loss": 0.005, + "step": 73150 + }, + { + "epoch": 11.363777570674122, + "grad_norm": 0.29688552021980286, + "learning_rate": 1.8436848344785473e-05, + "loss": 0.0057, + "step": 73160 + }, + { + "epoch": 11.365330848089469, + "grad_norm": 0.4843170642852783, + "learning_rate": 1.8424026274637378e-05, + "loss": 0.0045, + "step": 73170 + }, + { + "epoch": 11.366884125504814, + "grad_norm": 0.4275299310684204, + "learning_rate": 1.8411207657590312e-05, + "loss": 0.0046, + "step": 73180 + }, + { + "epoch": 11.368437402920161, + "grad_norm": 0.14719398319721222, + "learning_rate": 1.839839249504607e-05, + "loss": 0.0046, + "step": 73190 + }, + { + "epoch": 11.369990680335508, + "grad_norm": 0.23104864358901978, + "learning_rate": 1.8385580788406108e-05, + "loss": 0.0046, + "step": 73200 + }, + { + "epoch": 11.371543957750854, + "grad_norm": 0.1740625947713852, + "learning_rate": 1.837277253907152e-05, + "loss": 0.0046, + "step": 73210 + }, + { + "epoch": 11.3730972351662, + "grad_norm": 0.09106577187776566, + "learning_rate": 1.8359967748442957e-05, + "loss": 0.0045, + "step": 73220 + }, + { + "epoch": 11.374650512581548, + "grad_norm": 0.26277267932891846, + "learning_rate": 1.8347166417920774e-05, + "loss": 0.0048, + "step": 73230 + }, + { + "epoch": 11.376203789996893, + "grad_norm": 0.24771888554096222, + "learning_rate": 1.833436854890484e-05, + "loss": 0.0035, + "step": 73240 + }, + { + "epoch": 11.37775706741224, + "grad_norm": 0.6390925049781799, + "learning_rate": 1.8321574142794796e-05, + "loss": 0.0037, + "step": 73250 + }, + { + "epoch": 11.379310344827585, + "grad_norm": 0.23233217000961304, + "learning_rate": 1.8308783200989748e-05, + "loss": 0.0088, + "step": 73260 + }, + { + "epoch": 11.380863622242932, + "grad_norm": 0.1372080147266388, + "learning_rate": 1.829599572488855e-05, + "loss": 0.0036, + "step": 73270 + }, + { + "epoch": 11.38241689965828, + "grad_norm": 0.2934480309486389, + "learning_rate": 1.828321171588957e-05, + "loss": 0.0058, + "step": 73280 + }, + { + "epoch": 11.383970177073625, + "grad_norm": 0.2041151076555252, + "learning_rate": 1.827043117539089e-05, + "loss": 0.0044, + "step": 73290 + }, + { + "epoch": 11.385523454488972, + "grad_norm": 0.1564798653125763, + "learning_rate": 1.8257654104790116e-05, + "loss": 0.0055, + "step": 73300 + }, + { + "epoch": 11.387076731904319, + "grad_norm": 0.3760044574737549, + "learning_rate": 1.82448805054846e-05, + "loss": 0.0052, + "step": 73310 + }, + { + "epoch": 11.388630009319664, + "grad_norm": 0.1845661699771881, + "learning_rate": 1.823211037887119e-05, + "loss": 0.0048, + "step": 73320 + }, + { + "epoch": 11.390183286735011, + "grad_norm": 0.12114574015140533, + "learning_rate": 1.821934372634645e-05, + "loss": 0.0033, + "step": 73330 + }, + { + "epoch": 11.391736564150357, + "grad_norm": 0.6352759003639221, + "learning_rate": 1.8206580549306472e-05, + "loss": 0.0029, + "step": 73340 + }, + { + "epoch": 11.393289841565704, + "grad_norm": 0.29950642585754395, + "learning_rate": 1.819382084914707e-05, + "loss": 0.0049, + "step": 73350 + }, + { + "epoch": 11.39484311898105, + "grad_norm": 0.5295136570930481, + "learning_rate": 1.8181064627263577e-05, + "loss": 0.0036, + "step": 73360 + }, + { + "epoch": 11.396396396396396, + "grad_norm": 0.16019456088542938, + "learning_rate": 1.8168311885051022e-05, + "loss": 0.004, + "step": 73370 + }, + { + "epoch": 11.397949673811743, + "grad_norm": 0.13711974024772644, + "learning_rate": 1.815556262390402e-05, + "loss": 0.0047, + "step": 73380 + }, + { + "epoch": 11.39950295122709, + "grad_norm": 0.17328760027885437, + "learning_rate": 1.8142816845216826e-05, + "loss": 0.0042, + "step": 73390 + }, + { + "epoch": 11.401056228642435, + "grad_norm": 0.11445876210927963, + "learning_rate": 1.8130074550383276e-05, + "loss": 0.0033, + "step": 73400 + }, + { + "epoch": 11.402609506057782, + "grad_norm": 0.18623945116996765, + "learning_rate": 1.8117335740796848e-05, + "loss": 0.0046, + "step": 73410 + }, + { + "epoch": 11.404162783473128, + "grad_norm": 0.06368239223957062, + "learning_rate": 1.810460041785067e-05, + "loss": 0.0048, + "step": 73420 + }, + { + "epoch": 11.405716060888475, + "grad_norm": 0.44462400674819946, + "learning_rate": 1.809186858293742e-05, + "loss": 0.005, + "step": 73430 + }, + { + "epoch": 11.407269338303822, + "grad_norm": 0.2894054055213928, + "learning_rate": 1.8079140237449448e-05, + "loss": 0.0033, + "step": 73440 + }, + { + "epoch": 11.408822615719167, + "grad_norm": 0.14234477281570435, + "learning_rate": 1.8066415382778707e-05, + "loss": 0.0027, + "step": 73450 + }, + { + "epoch": 11.410375893134514, + "grad_norm": 0.42816612124443054, + "learning_rate": 1.8053694020316786e-05, + "loss": 0.0018, + "step": 73460 + }, + { + "epoch": 11.411929170549861, + "grad_norm": 0.1555679589509964, + "learning_rate": 1.8040976151454837e-05, + "loss": 0.0034, + "step": 73470 + }, + { + "epoch": 11.413482447965206, + "grad_norm": 0.1773136556148529, + "learning_rate": 1.8028261777583704e-05, + "loss": 0.0042, + "step": 73480 + }, + { + "epoch": 11.415035725380553, + "grad_norm": 0.11830470710992813, + "learning_rate": 1.8015550900093758e-05, + "loss": 0.0039, + "step": 73490 + }, + { + "epoch": 11.416589002795899, + "grad_norm": 0.2658129930496216, + "learning_rate": 1.800284352037512e-05, + "loss": 0.0025, + "step": 73500 + }, + { + "epoch": 11.418142280211246, + "grad_norm": 0.632256269454956, + "learning_rate": 1.799013963981738e-05, + "loss": 0.0064, + "step": 73510 + }, + { + "epoch": 11.419695557626593, + "grad_norm": 0.1637689173221588, + "learning_rate": 1.797743925980987e-05, + "loss": 0.004, + "step": 73520 + }, + { + "epoch": 11.421248835041938, + "grad_norm": 0.17495176196098328, + "learning_rate": 1.7964742381741435e-05, + "loss": 0.0033, + "step": 73530 + }, + { + "epoch": 11.422802112457285, + "grad_norm": 0.35167884826660156, + "learning_rate": 1.7952049007000627e-05, + "loss": 0.0059, + "step": 73540 + }, + { + "epoch": 11.42435538987263, + "grad_norm": 0.12261158227920532, + "learning_rate": 1.7939359136975535e-05, + "loss": 0.0042, + "step": 73550 + }, + { + "epoch": 11.425908667287977, + "grad_norm": 0.09664774686098099, + "learning_rate": 1.7926672773053932e-05, + "loss": 0.0043, + "step": 73560 + }, + { + "epoch": 11.427461944703325, + "grad_norm": 0.15273790061473846, + "learning_rate": 1.791398991662317e-05, + "loss": 0.0042, + "step": 73570 + }, + { + "epoch": 11.42901522211867, + "grad_norm": 0.4280132055282593, + "learning_rate": 1.7901310569070244e-05, + "loss": 0.0064, + "step": 73580 + }, + { + "epoch": 11.430568499534017, + "grad_norm": 0.13872560858726501, + "learning_rate": 1.7888634731781724e-05, + "loss": 0.0056, + "step": 73590 + }, + { + "epoch": 11.432121776949364, + "grad_norm": 0.1355285793542862, + "learning_rate": 1.7875962406143844e-05, + "loss": 0.0034, + "step": 73600 + }, + { + "epoch": 11.43367505436471, + "grad_norm": 0.17798998951911926, + "learning_rate": 1.7863293593542403e-05, + "loss": 0.0034, + "step": 73610 + }, + { + "epoch": 11.435228331780056, + "grad_norm": 0.2638545632362366, + "learning_rate": 1.7850628295362853e-05, + "loss": 0.0102, + "step": 73620 + }, + { + "epoch": 11.436781609195402, + "grad_norm": 0.22721704840660095, + "learning_rate": 1.7837966512990263e-05, + "loss": 0.0035, + "step": 73630 + }, + { + "epoch": 11.438334886610749, + "grad_norm": 0.19230683147907257, + "learning_rate": 1.782530824780931e-05, + "loss": 0.0042, + "step": 73640 + }, + { + "epoch": 11.439888164026096, + "grad_norm": 0.1881578415632248, + "learning_rate": 1.7812653501204263e-05, + "loss": 0.0036, + "step": 73650 + }, + { + "epoch": 11.441441441441441, + "grad_norm": 0.14847950637340546, + "learning_rate": 1.780000227455903e-05, + "loss": 0.0042, + "step": 73660 + }, + { + "epoch": 11.442994718856788, + "grad_norm": 0.3985734283924103, + "learning_rate": 1.778735456925716e-05, + "loss": 0.0049, + "step": 73670 + }, + { + "epoch": 11.444547996272135, + "grad_norm": 0.2796265780925751, + "learning_rate": 1.7774710386681737e-05, + "loss": 0.0048, + "step": 73680 + }, + { + "epoch": 11.44610127368748, + "grad_norm": 0.1485593467950821, + "learning_rate": 1.7762069728215545e-05, + "loss": 0.0046, + "step": 73690 + }, + { + "epoch": 11.447654551102827, + "grad_norm": 0.3008168935775757, + "learning_rate": 1.7749432595240932e-05, + "loss": 0.0041, + "step": 73700 + }, + { + "epoch": 11.449207828518173, + "grad_norm": 0.16287758946418762, + "learning_rate": 1.773679898913991e-05, + "loss": 0.0049, + "step": 73710 + }, + { + "epoch": 11.45076110593352, + "grad_norm": 0.3698389232158661, + "learning_rate": 1.7724168911294015e-05, + "loss": 0.0047, + "step": 73720 + }, + { + "epoch": 11.452314383348867, + "grad_norm": 0.3004859387874603, + "learning_rate": 1.7711542363084504e-05, + "loss": 0.0037, + "step": 73730 + }, + { + "epoch": 11.453867660764212, + "grad_norm": 0.1965809017419815, + "learning_rate": 1.7698919345892158e-05, + "loss": 0.0027, + "step": 73740 + }, + { + "epoch": 11.455420938179559, + "grad_norm": 0.3329622447490692, + "learning_rate": 1.768629986109743e-05, + "loss": 0.0063, + "step": 73750 + }, + { + "epoch": 11.456974215594904, + "grad_norm": 0.13114970922470093, + "learning_rate": 1.7673683910080357e-05, + "loss": 0.0061, + "step": 73760 + }, + { + "epoch": 11.458527493010251, + "grad_norm": 0.17881810665130615, + "learning_rate": 1.7661071494220638e-05, + "loss": 0.0032, + "step": 73770 + }, + { + "epoch": 11.460080770425598, + "grad_norm": 0.19534064829349518, + "learning_rate": 1.7648462614897498e-05, + "loss": 0.0048, + "step": 73780 + }, + { + "epoch": 11.461634047840944, + "grad_norm": 0.35680142045021057, + "learning_rate": 1.7635857273489863e-05, + "loss": 0.0055, + "step": 73790 + }, + { + "epoch": 11.46318732525629, + "grad_norm": 0.23321524262428284, + "learning_rate": 1.7623255471376198e-05, + "loss": 0.0081, + "step": 73800 + }, + { + "epoch": 11.464740602671638, + "grad_norm": 0.4657316207885742, + "learning_rate": 1.7610657209934646e-05, + "loss": 0.0038, + "step": 73810 + }, + { + "epoch": 11.466293880086983, + "grad_norm": 0.370973140001297, + "learning_rate": 1.7598062490542915e-05, + "loss": 0.0043, + "step": 73820 + }, + { + "epoch": 11.46784715750233, + "grad_norm": 0.3705332577228546, + "learning_rate": 1.7585471314578377e-05, + "loss": 0.0068, + "step": 73830 + }, + { + "epoch": 11.469400434917675, + "grad_norm": 0.315127432346344, + "learning_rate": 1.7572883683417945e-05, + "loss": 0.0035, + "step": 73840 + }, + { + "epoch": 11.470953712333023, + "grad_norm": 0.15776076912879944, + "learning_rate": 1.7560299598438212e-05, + "loss": 0.005, + "step": 73850 + }, + { + "epoch": 11.47250698974837, + "grad_norm": 0.3678083121776581, + "learning_rate": 1.7547719061015334e-05, + "loss": 0.0054, + "step": 73860 + }, + { + "epoch": 11.474060267163715, + "grad_norm": 0.21004535257816315, + "learning_rate": 1.75351420725251e-05, + "loss": 0.0037, + "step": 73870 + }, + { + "epoch": 11.475613544579062, + "grad_norm": 0.25257426500320435, + "learning_rate": 1.7522568634342923e-05, + "loss": 0.0081, + "step": 73880 + }, + { + "epoch": 11.477166821994409, + "grad_norm": 0.34639766812324524, + "learning_rate": 1.7509998747843825e-05, + "loss": 0.0041, + "step": 73890 + }, + { + "epoch": 11.478720099409754, + "grad_norm": 0.41317880153656006, + "learning_rate": 1.7497432414402403e-05, + "loss": 0.0041, + "step": 73900 + }, + { + "epoch": 11.480273376825101, + "grad_norm": 0.43203142285346985, + "learning_rate": 1.7484869635392903e-05, + "loss": 0.0069, + "step": 73910 + }, + { + "epoch": 11.481826654240447, + "grad_norm": 0.12962737679481506, + "learning_rate": 1.747231041218919e-05, + "loss": 0.004, + "step": 73920 + }, + { + "epoch": 11.483379931655794, + "grad_norm": 0.49958792328834534, + "learning_rate": 1.7459754746164692e-05, + "loss": 0.0046, + "step": 73930 + }, + { + "epoch": 11.48493320907114, + "grad_norm": 0.2589349150657654, + "learning_rate": 1.744720263869249e-05, + "loss": 0.0038, + "step": 73940 + }, + { + "epoch": 11.486486486486486, + "grad_norm": 0.21121074259281158, + "learning_rate": 1.7434654091145263e-05, + "loss": 0.0048, + "step": 73950 + }, + { + "epoch": 11.488039763901833, + "grad_norm": 0.19288291037082672, + "learning_rate": 1.742210910489532e-05, + "loss": 0.0035, + "step": 73960 + }, + { + "epoch": 11.48959304131718, + "grad_norm": 0.13749268651008606, + "learning_rate": 1.7409567681314532e-05, + "loss": 0.0042, + "step": 73970 + }, + { + "epoch": 11.491146318732525, + "grad_norm": 0.19980476796627045, + "learning_rate": 1.7397029821774434e-05, + "loss": 0.0042, + "step": 73980 + }, + { + "epoch": 11.492699596147872, + "grad_norm": 0.09497810900211334, + "learning_rate": 1.7384495527646126e-05, + "loss": 0.0035, + "step": 73990 + }, + { + "epoch": 11.494252873563218, + "grad_norm": 0.09212589263916016, + "learning_rate": 1.737196480030035e-05, + "loss": 0.0038, + "step": 74000 + }, + { + "epoch": 11.495806150978565, + "grad_norm": 0.09571229666471481, + "learning_rate": 1.7359437641107447e-05, + "loss": 0.004, + "step": 74010 + }, + { + "epoch": 11.497359428393912, + "grad_norm": 0.17024169862270355, + "learning_rate": 1.7346914051437387e-05, + "loss": 0.0045, + "step": 74020 + }, + { + "epoch": 11.498912705809257, + "grad_norm": 0.4549461305141449, + "learning_rate": 1.733439403265969e-05, + "loss": 0.004, + "step": 74030 + }, + { + "epoch": 11.500465983224604, + "grad_norm": 0.18171873688697815, + "learning_rate": 1.732187758614357e-05, + "loss": 0.0042, + "step": 74040 + }, + { + "epoch": 11.502019260639951, + "grad_norm": 0.45419731736183167, + "learning_rate": 1.730936471325777e-05, + "loss": 0.0039, + "step": 74050 + }, + { + "epoch": 11.503572538055296, + "grad_norm": 0.12535686790943146, + "learning_rate": 1.7296855415370718e-05, + "loss": 0.0038, + "step": 74060 + }, + { + "epoch": 11.505125815470644, + "grad_norm": 0.20028667151927948, + "learning_rate": 1.728434969385035e-05, + "loss": 0.007, + "step": 74070 + }, + { + "epoch": 11.506679092885989, + "grad_norm": 0.358511745929718, + "learning_rate": 1.7271847550064346e-05, + "loss": 0.005, + "step": 74080 + }, + { + "epoch": 11.508232370301336, + "grad_norm": 0.1635730117559433, + "learning_rate": 1.725934898537987e-05, + "loss": 0.0047, + "step": 74090 + }, + { + "epoch": 11.509785647716683, + "grad_norm": 0.33254191279411316, + "learning_rate": 1.7246854001163783e-05, + "loss": 0.006, + "step": 74100 + }, + { + "epoch": 11.511338925132028, + "grad_norm": 0.20789508521556854, + "learning_rate": 1.7234362598782476e-05, + "loss": 0.0049, + "step": 74110 + }, + { + "epoch": 11.512892202547375, + "grad_norm": 1.3917343616485596, + "learning_rate": 1.722187477960202e-05, + "loss": 0.0041, + "step": 74120 + }, + { + "epoch": 11.51444547996272, + "grad_norm": 0.13188384473323822, + "learning_rate": 1.7209390544988057e-05, + "loss": 0.0058, + "step": 74130 + }, + { + "epoch": 11.515998757378068, + "grad_norm": 0.11215704679489136, + "learning_rate": 1.7196909896305856e-05, + "loss": 0.003, + "step": 74140 + }, + { + "epoch": 11.517552034793415, + "grad_norm": 0.3712516725063324, + "learning_rate": 1.718443283492026e-05, + "loss": 0.0043, + "step": 74150 + }, + { + "epoch": 11.51910531220876, + "grad_norm": 0.3049639165401459, + "learning_rate": 1.7171959362195746e-05, + "loss": 0.0037, + "step": 74160 + }, + { + "epoch": 11.520658589624107, + "grad_norm": 0.2701740264892578, + "learning_rate": 1.715948947949642e-05, + "loss": 0.0031, + "step": 74170 + }, + { + "epoch": 11.522211867039454, + "grad_norm": 0.6254023909568787, + "learning_rate": 1.714702318818593e-05, + "loss": 0.005, + "step": 74180 + }, + { + "epoch": 11.5237651444548, + "grad_norm": 0.13579094409942627, + "learning_rate": 1.7134560489627592e-05, + "loss": 0.004, + "step": 74190 + }, + { + "epoch": 11.525318421870146, + "grad_norm": 0.1914711892604828, + "learning_rate": 1.712210138518431e-05, + "loss": 0.0037, + "step": 74200 + }, + { + "epoch": 11.526871699285492, + "grad_norm": 0.20286762714385986, + "learning_rate": 1.71096458762186e-05, + "loss": 0.0052, + "step": 74210 + }, + { + "epoch": 11.528424976700839, + "grad_norm": 0.21560563147068024, + "learning_rate": 1.7097193964092552e-05, + "loss": 0.0042, + "step": 74220 + }, + { + "epoch": 11.529978254116186, + "grad_norm": 0.5334694981575012, + "learning_rate": 1.708474565016792e-05, + "loss": 0.0042, + "step": 74230 + }, + { + "epoch": 11.531531531531531, + "grad_norm": 0.1257220208644867, + "learning_rate": 1.7072300935806e-05, + "loss": 0.005, + "step": 74240 + }, + { + "epoch": 11.533084808946878, + "grad_norm": 0.13631340861320496, + "learning_rate": 1.705985982236776e-05, + "loss": 0.004, + "step": 74250 + }, + { + "epoch": 11.534638086362225, + "grad_norm": 0.27722859382629395, + "learning_rate": 1.7047422311213697e-05, + "loss": 0.0055, + "step": 74260 + }, + { + "epoch": 11.53619136377757, + "grad_norm": 0.5288522243499756, + "learning_rate": 1.703498840370401e-05, + "loss": 0.0033, + "step": 74270 + }, + { + "epoch": 11.537744641192917, + "grad_norm": 0.0894496813416481, + "learning_rate": 1.702255810119842e-05, + "loss": 0.0046, + "step": 74280 + }, + { + "epoch": 11.539297918608263, + "grad_norm": 0.1595582216978073, + "learning_rate": 1.7010131405056305e-05, + "loss": 0.0047, + "step": 74290 + }, + { + "epoch": 11.54085119602361, + "grad_norm": 0.11169987171888351, + "learning_rate": 1.6997708316636613e-05, + "loss": 0.0023, + "step": 74300 + }, + { + "epoch": 11.542404473438957, + "grad_norm": 0.16192877292633057, + "learning_rate": 1.6985288837297932e-05, + "loss": 0.0045, + "step": 74310 + }, + { + "epoch": 11.543957750854302, + "grad_norm": 0.1618555784225464, + "learning_rate": 1.6972872968398394e-05, + "loss": 0.0041, + "step": 74320 + }, + { + "epoch": 11.54551102826965, + "grad_norm": 0.19167865812778473, + "learning_rate": 1.6960460711295845e-05, + "loss": 0.003, + "step": 74330 + }, + { + "epoch": 11.547064305684994, + "grad_norm": 0.4183284044265747, + "learning_rate": 1.6948052067347623e-05, + "loss": 0.0072, + "step": 74340 + }, + { + "epoch": 11.548617583100341, + "grad_norm": 0.24957889318466187, + "learning_rate": 1.693564703791075e-05, + "loss": 0.0048, + "step": 74350 + }, + { + "epoch": 11.550170860515689, + "grad_norm": 0.36554983258247375, + "learning_rate": 1.6923245624341793e-05, + "loss": 0.0064, + "step": 74360 + }, + { + "epoch": 11.551724137931034, + "grad_norm": 0.26543280482292175, + "learning_rate": 1.6910847827996962e-05, + "loss": 0.0057, + "step": 74370 + }, + { + "epoch": 11.55327741534638, + "grad_norm": 0.25921398401260376, + "learning_rate": 1.689845365023206e-05, + "loss": 0.006, + "step": 74380 + }, + { + "epoch": 11.554830692761728, + "grad_norm": 0.17807960510253906, + "learning_rate": 1.6886063092402505e-05, + "loss": 0.0074, + "step": 74390 + }, + { + "epoch": 11.556383970177073, + "grad_norm": 0.06897468864917755, + "learning_rate": 1.6873676155863317e-05, + "loss": 0.0044, + "step": 74400 + }, + { + "epoch": 11.55793724759242, + "grad_norm": 0.32574495673179626, + "learning_rate": 1.6861292841969085e-05, + "loss": 0.0062, + "step": 74410 + }, + { + "epoch": 11.559490525007767, + "grad_norm": 0.1361086517572403, + "learning_rate": 1.684891315207406e-05, + "loss": 0.0043, + "step": 74420 + }, + { + "epoch": 11.561043802423113, + "grad_norm": 0.2107788324356079, + "learning_rate": 1.6836537087532034e-05, + "loss": 0.0046, + "step": 74430 + }, + { + "epoch": 11.56259707983846, + "grad_norm": 0.12787558138370514, + "learning_rate": 1.6824164649696465e-05, + "loss": 0.0032, + "step": 74440 + }, + { + "epoch": 11.564150357253805, + "grad_norm": 0.1236582100391388, + "learning_rate": 1.681179583992034e-05, + "loss": 0.0034, + "step": 74450 + }, + { + "epoch": 11.565703634669152, + "grad_norm": 0.10485135763883591, + "learning_rate": 1.679943065955636e-05, + "loss": 0.0046, + "step": 74460 + }, + { + "epoch": 11.567256912084499, + "grad_norm": 0.5111619830131531, + "learning_rate": 1.678706910995671e-05, + "loss": 0.004, + "step": 74470 + }, + { + "epoch": 11.568810189499844, + "grad_norm": 0.2690525949001312, + "learning_rate": 1.6774711192473262e-05, + "loss": 0.0056, + "step": 74480 + }, + { + "epoch": 11.570363466915191, + "grad_norm": 0.17831097543239594, + "learning_rate": 1.676235690845743e-05, + "loss": 0.0044, + "step": 74490 + }, + { + "epoch": 11.571916744330537, + "grad_norm": 0.23010654747486115, + "learning_rate": 1.6750006259260285e-05, + "loss": 0.0039, + "step": 74500 + }, + { + "epoch": 11.573470021745884, + "grad_norm": 0.37772515416145325, + "learning_rate": 1.673765924623244e-05, + "loss": 0.0074, + "step": 74510 + }, + { + "epoch": 11.57502329916123, + "grad_norm": 0.2669723331928253, + "learning_rate": 1.67253158707242e-05, + "loss": 0.004, + "step": 74520 + }, + { + "epoch": 11.576576576576576, + "grad_norm": 0.429057240486145, + "learning_rate": 1.671297613408537e-05, + "loss": 0.004, + "step": 74530 + }, + { + "epoch": 11.578129853991923, + "grad_norm": 0.19798628985881805, + "learning_rate": 1.6700640037665444e-05, + "loss": 0.008, + "step": 74540 + }, + { + "epoch": 11.57968313140727, + "grad_norm": 0.06918162852525711, + "learning_rate": 1.668830758281344e-05, + "loss": 0.0031, + "step": 74550 + }, + { + "epoch": 11.581236408822615, + "grad_norm": 0.20075955986976624, + "learning_rate": 1.6675978770878052e-05, + "loss": 0.0063, + "step": 74560 + }, + { + "epoch": 11.582789686237962, + "grad_norm": 0.19539715349674225, + "learning_rate": 1.6663653603207506e-05, + "loss": 0.004, + "step": 74570 + }, + { + "epoch": 11.584342963653308, + "grad_norm": 0.10380958020687103, + "learning_rate": 1.6651332081149685e-05, + "loss": 0.0047, + "step": 74580 + }, + { + "epoch": 11.585896241068655, + "grad_norm": 0.09930895268917084, + "learning_rate": 1.6639014206052046e-05, + "loss": 0.0048, + "step": 74590 + }, + { + "epoch": 11.587449518484002, + "grad_norm": 0.22042377293109894, + "learning_rate": 1.6626699979261657e-05, + "loss": 0.0043, + "step": 74600 + }, + { + "epoch": 11.589002795899347, + "grad_norm": 0.6099016666412354, + "learning_rate": 1.6614389402125197e-05, + "loss": 0.0052, + "step": 74610 + }, + { + "epoch": 11.590556073314694, + "grad_norm": 0.16572853922843933, + "learning_rate": 1.66020824759889e-05, + "loss": 0.0033, + "step": 74620 + }, + { + "epoch": 11.592109350730041, + "grad_norm": 0.0856999084353447, + "learning_rate": 1.6589779202198675e-05, + "loss": 0.0039, + "step": 74630 + }, + { + "epoch": 11.593662628145387, + "grad_norm": 0.2956998944282532, + "learning_rate": 1.6577479582099924e-05, + "loss": 0.0043, + "step": 74640 + }, + { + "epoch": 11.595215905560734, + "grad_norm": 0.12462200224399567, + "learning_rate": 1.6565183617037788e-05, + "loss": 0.0023, + "step": 74650 + }, + { + "epoch": 11.596769182976079, + "grad_norm": 0.31053683161735535, + "learning_rate": 1.6552891308356894e-05, + "loss": 0.0044, + "step": 74660 + }, + { + "epoch": 11.598322460391426, + "grad_norm": 0.2551661729812622, + "learning_rate": 1.6540602657401526e-05, + "loss": 0.0039, + "step": 74670 + }, + { + "epoch": 11.599875737806773, + "grad_norm": 0.19993790984153748, + "learning_rate": 1.6528317665515537e-05, + "loss": 0.0039, + "step": 74680 + }, + { + "epoch": 11.601429015222118, + "grad_norm": 0.21580494940280914, + "learning_rate": 1.6516036334042416e-05, + "loss": 0.004, + "step": 74690 + }, + { + "epoch": 11.602982292637465, + "grad_norm": 0.1896001547574997, + "learning_rate": 1.650375866432519e-05, + "loss": 0.0051, + "step": 74700 + }, + { + "epoch": 11.60453557005281, + "grad_norm": 0.19104723632335663, + "learning_rate": 1.649148465770659e-05, + "loss": 0.0055, + "step": 74710 + }, + { + "epoch": 11.606088847468158, + "grad_norm": 0.6296045184135437, + "learning_rate": 1.647921431552883e-05, + "loss": 0.006, + "step": 74720 + }, + { + "epoch": 11.607642124883505, + "grad_norm": 0.1215696632862091, + "learning_rate": 1.646694763913381e-05, + "loss": 0.0033, + "step": 74730 + }, + { + "epoch": 11.60919540229885, + "grad_norm": 0.13551682233810425, + "learning_rate": 1.6454684629862966e-05, + "loss": 0.004, + "step": 74740 + }, + { + "epoch": 11.610748679714197, + "grad_norm": 0.14106279611587524, + "learning_rate": 1.6442425289057388e-05, + "loss": 0.0034, + "step": 74750 + }, + { + "epoch": 11.612301957129544, + "grad_norm": 0.47432050108909607, + "learning_rate": 1.6430169618057718e-05, + "loss": 0.0047, + "step": 74760 + }, + { + "epoch": 11.61385523454489, + "grad_norm": 0.8467065691947937, + "learning_rate": 1.641791761820422e-05, + "loss": 0.0054, + "step": 74770 + }, + { + "epoch": 11.615408511960236, + "grad_norm": 0.42309674620628357, + "learning_rate": 1.640566929083676e-05, + "loss": 0.0047, + "step": 74780 + }, + { + "epoch": 11.616961789375582, + "grad_norm": 0.3723576068878174, + "learning_rate": 1.6393424637294823e-05, + "loss": 0.0043, + "step": 74790 + }, + { + "epoch": 11.618515066790929, + "grad_norm": 0.29161337018013, + "learning_rate": 1.638118365891742e-05, + "loss": 0.0067, + "step": 74800 + }, + { + "epoch": 11.620068344206276, + "grad_norm": 0.4829865097999573, + "learning_rate": 1.636894635704323e-05, + "loss": 0.0034, + "step": 74810 + }, + { + "epoch": 11.621621621621621, + "grad_norm": 0.2939799427986145, + "learning_rate": 1.6356712733010517e-05, + "loss": 0.0032, + "step": 74820 + }, + { + "epoch": 11.623174899036968, + "grad_norm": 0.166342630982399, + "learning_rate": 1.6344482788157096e-05, + "loss": 0.004, + "step": 74830 + }, + { + "epoch": 11.624728176452315, + "grad_norm": 0.3029206097126007, + "learning_rate": 1.6332256523820444e-05, + "loss": 0.0069, + "step": 74840 + }, + { + "epoch": 11.62628145386766, + "grad_norm": 0.1551598608493805, + "learning_rate": 1.6320033941337597e-05, + "loss": 0.004, + "step": 74850 + }, + { + "epoch": 11.627834731283007, + "grad_norm": 0.2505780756473541, + "learning_rate": 1.6307815042045216e-05, + "loss": 0.0029, + "step": 74860 + }, + { + "epoch": 11.629388008698353, + "grad_norm": 0.2594831585884094, + "learning_rate": 1.6295599827279513e-05, + "loss": 0.0038, + "step": 74870 + }, + { + "epoch": 11.6309412861137, + "grad_norm": 0.6741961240768433, + "learning_rate": 1.6283388298376356e-05, + "loss": 0.0043, + "step": 74880 + }, + { + "epoch": 11.632494563529047, + "grad_norm": 0.280573308467865, + "learning_rate": 1.6271180456671125e-05, + "loss": 0.0049, + "step": 74890 + }, + { + "epoch": 11.634047840944392, + "grad_norm": 0.4041886329650879, + "learning_rate": 1.625897630349893e-05, + "loss": 0.0057, + "step": 74900 + }, + { + "epoch": 11.63560111835974, + "grad_norm": 0.47988471388816833, + "learning_rate": 1.6246775840194335e-05, + "loss": 0.0035, + "step": 74910 + }, + { + "epoch": 11.637154395775086, + "grad_norm": 0.7368084788322449, + "learning_rate": 1.6234579068091604e-05, + "loss": 0.0048, + "step": 74920 + }, + { + "epoch": 11.638707673190432, + "grad_norm": 0.2834901809692383, + "learning_rate": 1.6222385988524524e-05, + "loss": 0.0043, + "step": 74930 + }, + { + "epoch": 11.640260950605779, + "grad_norm": 0.2194076031446457, + "learning_rate": 1.6210196602826544e-05, + "loss": 0.0033, + "step": 74940 + }, + { + "epoch": 11.641814228021124, + "grad_norm": 0.11370894312858582, + "learning_rate": 1.6198010912330646e-05, + "loss": 0.0033, + "step": 74950 + }, + { + "epoch": 11.643367505436471, + "grad_norm": 0.2915503680706024, + "learning_rate": 1.6185828918369456e-05, + "loss": 0.0073, + "step": 74960 + }, + { + "epoch": 11.644920782851818, + "grad_norm": 0.22823157906532288, + "learning_rate": 1.617365062227517e-05, + "loss": 0.0035, + "step": 74970 + }, + { + "epoch": 11.646474060267163, + "grad_norm": 0.37448179721832275, + "learning_rate": 1.616147602537962e-05, + "loss": 0.0044, + "step": 74980 + }, + { + "epoch": 11.64802733768251, + "grad_norm": 0.25801557302474976, + "learning_rate": 1.614930512901415e-05, + "loss": 0.0042, + "step": 74990 + }, + { + "epoch": 11.649580615097857, + "grad_norm": 0.7388191819190979, + "learning_rate": 1.6137137934509804e-05, + "loss": 0.0037, + "step": 75000 + }, + { + "epoch": 11.651133892513203, + "grad_norm": 0.26795265078544617, + "learning_rate": 1.6124974443197123e-05, + "loss": 0.0043, + "step": 75010 + }, + { + "epoch": 11.65268716992855, + "grad_norm": 0.31340667605400085, + "learning_rate": 1.6112814656406306e-05, + "loss": 0.0081, + "step": 75020 + }, + { + "epoch": 11.654240447343895, + "grad_norm": 0.40782150626182556, + "learning_rate": 1.6100658575467125e-05, + "loss": 0.0051, + "step": 75030 + }, + { + "epoch": 11.655793724759242, + "grad_norm": 0.19027116894721985, + "learning_rate": 1.6088506201708974e-05, + "loss": 0.0053, + "step": 75040 + }, + { + "epoch": 11.657347002174589, + "grad_norm": 0.2598519027233124, + "learning_rate": 1.6076357536460785e-05, + "loss": 0.0046, + "step": 75050 + }, + { + "epoch": 11.658900279589934, + "grad_norm": 0.12864838540554047, + "learning_rate": 1.606421258105113e-05, + "loss": 0.0041, + "step": 75060 + }, + { + "epoch": 11.660453557005281, + "grad_norm": 1.4381283521652222, + "learning_rate": 1.6052071336808183e-05, + "loss": 0.0036, + "step": 75070 + }, + { + "epoch": 11.662006834420627, + "grad_norm": 0.3433666527271271, + "learning_rate": 1.6039933805059658e-05, + "loss": 0.0036, + "step": 75080 + }, + { + "epoch": 11.663560111835974, + "grad_norm": 0.3247523307800293, + "learning_rate": 1.6027799987132915e-05, + "loss": 0.0029, + "step": 75090 + }, + { + "epoch": 11.66511338925132, + "grad_norm": 0.3659552037715912, + "learning_rate": 1.6015669884354888e-05, + "loss": 0.0039, + "step": 75100 + }, + { + "epoch": 11.666666666666666, + "grad_norm": 0.2268746942281723, + "learning_rate": 1.6003543498052125e-05, + "loss": 0.0038, + "step": 75110 + }, + { + "epoch": 11.668219944082013, + "grad_norm": 0.15137672424316406, + "learning_rate": 1.5991420829550715e-05, + "loss": 0.0039, + "step": 75120 + }, + { + "epoch": 11.66977322149736, + "grad_norm": 0.23350109159946442, + "learning_rate": 1.5979301880176405e-05, + "loss": 0.0033, + "step": 75130 + }, + { + "epoch": 11.671326498912705, + "grad_norm": 0.25215116143226624, + "learning_rate": 1.5967186651254485e-05, + "loss": 0.0051, + "step": 75140 + }, + { + "epoch": 11.672879776328053, + "grad_norm": 0.12982913851737976, + "learning_rate": 1.5955075144109853e-05, + "loss": 0.0056, + "step": 75150 + }, + { + "epoch": 11.674433053743398, + "grad_norm": 1.1451942920684814, + "learning_rate": 1.594296736006703e-05, + "loss": 0.0052, + "step": 75160 + }, + { + "epoch": 11.675986331158745, + "grad_norm": 0.3517308235168457, + "learning_rate": 1.59308633004501e-05, + "loss": 0.0038, + "step": 75170 + }, + { + "epoch": 11.677539608574092, + "grad_norm": 0.2228683978319168, + "learning_rate": 1.5918762966582724e-05, + "loss": 0.005, + "step": 75180 + }, + { + "epoch": 11.679092885989437, + "grad_norm": 0.10888398438692093, + "learning_rate": 1.5906666359788203e-05, + "loss": 0.0031, + "step": 75190 + }, + { + "epoch": 11.680646163404784, + "grad_norm": 0.19458694756031036, + "learning_rate": 1.589457348138938e-05, + "loss": 0.0046, + "step": 75200 + }, + { + "epoch": 11.682199440820131, + "grad_norm": 0.09127356857061386, + "learning_rate": 1.588248433270872e-05, + "loss": 0.0047, + "step": 75210 + }, + { + "epoch": 11.683752718235477, + "grad_norm": 0.07170148938894272, + "learning_rate": 1.587039891506828e-05, + "loss": 0.0049, + "step": 75220 + }, + { + "epoch": 11.685305995650824, + "grad_norm": 0.24806813895702362, + "learning_rate": 1.5858317229789715e-05, + "loss": 0.0045, + "step": 75230 + }, + { + "epoch": 11.686859273066169, + "grad_norm": 0.47831296920776367, + "learning_rate": 1.584623927819423e-05, + "loss": 0.0047, + "step": 75240 + }, + { + "epoch": 11.688412550481516, + "grad_norm": 0.42540407180786133, + "learning_rate": 1.583416506160269e-05, + "loss": 0.0038, + "step": 75250 + }, + { + "epoch": 11.689965827896863, + "grad_norm": 0.11788169294595718, + "learning_rate": 1.5822094581335472e-05, + "loss": 0.0051, + "step": 75260 + }, + { + "epoch": 11.691519105312208, + "grad_norm": 0.20639924705028534, + "learning_rate": 1.5810027838712604e-05, + "loss": 0.0052, + "step": 75270 + }, + { + "epoch": 11.693072382727555, + "grad_norm": 0.2156325727701187, + "learning_rate": 1.5797964835053687e-05, + "loss": 0.0081, + "step": 75280 + }, + { + "epoch": 11.6946256601429, + "grad_norm": 0.2416425198316574, + "learning_rate": 1.5785905571677933e-05, + "loss": 0.0039, + "step": 75290 + }, + { + "epoch": 11.696178937558248, + "grad_norm": 0.3804059624671936, + "learning_rate": 1.5773850049904086e-05, + "loss": 0.0032, + "step": 75300 + }, + { + "epoch": 11.697732214973595, + "grad_norm": 0.13453301787376404, + "learning_rate": 1.5761798271050543e-05, + "loss": 0.0055, + "step": 75310 + }, + { + "epoch": 11.69928549238894, + "grad_norm": 0.272578626871109, + "learning_rate": 1.5749750236435277e-05, + "loss": 0.004, + "step": 75320 + }, + { + "epoch": 11.700838769804287, + "grad_norm": 0.22831137478351593, + "learning_rate": 1.5737705947375815e-05, + "loss": 0.0044, + "step": 75330 + }, + { + "epoch": 11.702392047219634, + "grad_norm": 0.13941723108291626, + "learning_rate": 1.5725665405189322e-05, + "loss": 0.0031, + "step": 75340 + }, + { + "epoch": 11.70394532463498, + "grad_norm": 0.35188913345336914, + "learning_rate": 1.571362861119253e-05, + "loss": 0.0043, + "step": 75350 + }, + { + "epoch": 11.705498602050326, + "grad_norm": 0.5102052092552185, + "learning_rate": 1.570159556670177e-05, + "loss": 0.0033, + "step": 75360 + }, + { + "epoch": 11.707051879465673, + "grad_norm": 0.0995287224650383, + "learning_rate": 1.5689566273032946e-05, + "loss": 0.0045, + "step": 75370 + }, + { + "epoch": 11.708605156881019, + "grad_norm": 0.2403194159269333, + "learning_rate": 1.5677540731501583e-05, + "loss": 0.0042, + "step": 75380 + }, + { + "epoch": 11.710158434296366, + "grad_norm": 0.44569700956344604, + "learning_rate": 1.5665518943422746e-05, + "loss": 0.0057, + "step": 75390 + }, + { + "epoch": 11.711711711711711, + "grad_norm": 0.9587117433547974, + "learning_rate": 1.5653500910111134e-05, + "loss": 0.0041, + "step": 75400 + }, + { + "epoch": 11.713264989127058, + "grad_norm": 0.23965202271938324, + "learning_rate": 1.5641486632881023e-05, + "loss": 0.007, + "step": 75410 + }, + { + "epoch": 11.714818266542405, + "grad_norm": 0.3666289448738098, + "learning_rate": 1.56294761130463e-05, + "loss": 0.0046, + "step": 75420 + }, + { + "epoch": 11.71637154395775, + "grad_norm": 0.16492338478565216, + "learning_rate": 1.5617469351920373e-05, + "loss": 0.0046, + "step": 75430 + }, + { + "epoch": 11.717924821373098, + "grad_norm": 0.18527373671531677, + "learning_rate": 1.5605466350816315e-05, + "loss": 0.0032, + "step": 75440 + }, + { + "epoch": 11.719478098788443, + "grad_norm": 0.3112706243991852, + "learning_rate": 1.559346711104674e-05, + "loss": 0.0047, + "step": 75450 + }, + { + "epoch": 11.72103137620379, + "grad_norm": 0.20836864411830902, + "learning_rate": 1.5581471633923883e-05, + "loss": 0.0038, + "step": 75460 + }, + { + "epoch": 11.722584653619137, + "grad_norm": 0.25435125827789307, + "learning_rate": 1.5569479920759505e-05, + "loss": 0.0055, + "step": 75470 + }, + { + "epoch": 11.724137931034482, + "grad_norm": 0.2512362599372864, + "learning_rate": 1.5557491972865072e-05, + "loss": 0.0048, + "step": 75480 + }, + { + "epoch": 11.72569120844983, + "grad_norm": 0.3812212347984314, + "learning_rate": 1.554550779155152e-05, + "loss": 0.0049, + "step": 75490 + }, + { + "epoch": 11.727244485865176, + "grad_norm": 0.23097531497478485, + "learning_rate": 1.5533527378129453e-05, + "loss": 0.0046, + "step": 75500 + }, + { + "epoch": 11.728797763280522, + "grad_norm": 0.24438098073005676, + "learning_rate": 1.552155073390899e-05, + "loss": 0.0046, + "step": 75510 + }, + { + "epoch": 11.730351040695869, + "grad_norm": 0.17648304998874664, + "learning_rate": 1.5509577860199903e-05, + "loss": 0.0056, + "step": 75520 + }, + { + "epoch": 11.731904318111214, + "grad_norm": 0.1980321705341339, + "learning_rate": 1.5497608758311525e-05, + "loss": 0.0033, + "step": 75530 + }, + { + "epoch": 11.733457595526561, + "grad_norm": 0.08889583498239517, + "learning_rate": 1.5485643429552797e-05, + "loss": 0.0047, + "step": 75540 + }, + { + "epoch": 11.735010872941908, + "grad_norm": 0.08770729601383209, + "learning_rate": 1.5473681875232197e-05, + "loss": 0.005, + "step": 75550 + }, + { + "epoch": 11.736564150357253, + "grad_norm": 0.20905126631259918, + "learning_rate": 1.5461724096657837e-05, + "loss": 0.0035, + "step": 75560 + }, + { + "epoch": 11.7381174277726, + "grad_norm": 0.11722727119922638, + "learning_rate": 1.5449770095137416e-05, + "loss": 0.0035, + "step": 75570 + }, + { + "epoch": 11.739670705187947, + "grad_norm": 0.2711830735206604, + "learning_rate": 1.5437819871978178e-05, + "loss": 0.0052, + "step": 75580 + }, + { + "epoch": 11.741223982603293, + "grad_norm": 0.3722388744354248, + "learning_rate": 1.5425873428486992e-05, + "loss": 0.004, + "step": 75590 + }, + { + "epoch": 11.74277726001864, + "grad_norm": 0.3474970757961273, + "learning_rate": 1.5413930765970312e-05, + "loss": 0.0031, + "step": 75600 + }, + { + "epoch": 11.744330537433985, + "grad_norm": 0.2510335445404053, + "learning_rate": 1.5401991885734175e-05, + "loss": 0.0061, + "step": 75610 + }, + { + "epoch": 11.745883814849332, + "grad_norm": 0.1863049566745758, + "learning_rate": 1.5390056789084174e-05, + "loss": 0.0059, + "step": 75620 + }, + { + "epoch": 11.74743709226468, + "grad_norm": 0.37798142433166504, + "learning_rate": 1.5378125477325545e-05, + "loss": 0.0034, + "step": 75630 + }, + { + "epoch": 11.748990369680024, + "grad_norm": 0.40424859523773193, + "learning_rate": 1.5366197951763046e-05, + "loss": 0.0046, + "step": 75640 + }, + { + "epoch": 11.750543647095371, + "grad_norm": 0.6067039966583252, + "learning_rate": 1.5354274213701075e-05, + "loss": 0.0043, + "step": 75650 + }, + { + "epoch": 11.752096924510717, + "grad_norm": 0.2766215205192566, + "learning_rate": 1.5342354264443566e-05, + "loss": 0.0057, + "step": 75660 + }, + { + "epoch": 11.753650201926064, + "grad_norm": 0.14039798080921173, + "learning_rate": 1.533043810529411e-05, + "loss": 0.0053, + "step": 75670 + }, + { + "epoch": 11.75520347934141, + "grad_norm": 0.2652338743209839, + "learning_rate": 1.5318525737555807e-05, + "loss": 0.0048, + "step": 75680 + }, + { + "epoch": 11.756756756756756, + "grad_norm": 0.14568477869033813, + "learning_rate": 1.53066171625314e-05, + "loss": 0.004, + "step": 75690 + }, + { + "epoch": 11.758310034172103, + "grad_norm": 0.2074168622493744, + "learning_rate": 1.529471238152317e-05, + "loss": 0.0044, + "step": 75700 + }, + { + "epoch": 11.75986331158745, + "grad_norm": 0.2782326936721802, + "learning_rate": 1.5282811395833023e-05, + "loss": 0.0029, + "step": 75710 + }, + { + "epoch": 11.761416589002796, + "grad_norm": 0.21005305647850037, + "learning_rate": 1.5270914206762397e-05, + "loss": 0.0052, + "step": 75720 + }, + { + "epoch": 11.762969866418143, + "grad_norm": 0.1319836527109146, + "learning_rate": 1.525902081561241e-05, + "loss": 0.0042, + "step": 75730 + }, + { + "epoch": 11.764523143833488, + "grad_norm": 0.14956261217594147, + "learning_rate": 1.5247131223683664e-05, + "loss": 0.0042, + "step": 75740 + }, + { + "epoch": 11.766076421248835, + "grad_norm": 0.4180019795894623, + "learning_rate": 1.5235245432276408e-05, + "loss": 0.0055, + "step": 75750 + }, + { + "epoch": 11.767629698664182, + "grad_norm": 0.190344899892807, + "learning_rate": 1.5223363442690424e-05, + "loss": 0.0064, + "step": 75760 + }, + { + "epoch": 11.769182976079527, + "grad_norm": 0.34421294927597046, + "learning_rate": 1.5211485256225133e-05, + "loss": 0.0051, + "step": 75770 + }, + { + "epoch": 11.770736253494874, + "grad_norm": 0.15143387019634247, + "learning_rate": 1.519961087417951e-05, + "loss": 0.004, + "step": 75780 + }, + { + "epoch": 11.772289530910221, + "grad_norm": 0.17496532201766968, + "learning_rate": 1.5187740297852137e-05, + "loss": 0.0041, + "step": 75790 + }, + { + "epoch": 11.773842808325567, + "grad_norm": 0.1624952107667923, + "learning_rate": 1.517587352854113e-05, + "loss": 0.0052, + "step": 75800 + }, + { + "epoch": 11.775396085740914, + "grad_norm": 0.13546589016914368, + "learning_rate": 1.516401056754423e-05, + "loss": 0.0059, + "step": 75810 + }, + { + "epoch": 11.776949363156259, + "grad_norm": 0.5717527270317078, + "learning_rate": 1.5152151416158783e-05, + "loss": 0.0043, + "step": 75820 + }, + { + "epoch": 11.778502640571606, + "grad_norm": 0.16718055307865143, + "learning_rate": 1.5140296075681642e-05, + "loss": 0.0048, + "step": 75830 + }, + { + "epoch": 11.780055917986953, + "grad_norm": 0.27752700448036194, + "learning_rate": 1.5128444547409331e-05, + "loss": 0.0043, + "step": 75840 + }, + { + "epoch": 11.781609195402298, + "grad_norm": 0.20376677811145782, + "learning_rate": 1.5116596832637864e-05, + "loss": 0.0059, + "step": 75850 + }, + { + "epoch": 11.783162472817645, + "grad_norm": 0.3429269790649414, + "learning_rate": 1.5104752932662952e-05, + "loss": 0.0041, + "step": 75860 + }, + { + "epoch": 11.78471575023299, + "grad_norm": 0.22615770995616913, + "learning_rate": 1.5092912848779778e-05, + "loss": 0.0038, + "step": 75870 + }, + { + "epoch": 11.786269027648338, + "grad_norm": 1.2615430355072021, + "learning_rate": 1.5081076582283194e-05, + "loss": 0.0061, + "step": 75880 + }, + { + "epoch": 11.787822305063685, + "grad_norm": 1.7853779792785645, + "learning_rate": 1.5069244134467553e-05, + "loss": 0.0056, + "step": 75890 + }, + { + "epoch": 11.78937558247903, + "grad_norm": 0.16905711591243744, + "learning_rate": 1.5057415506626876e-05, + "loss": 0.0033, + "step": 75900 + }, + { + "epoch": 11.790928859894377, + "grad_norm": 0.2550378441810608, + "learning_rate": 1.5045590700054668e-05, + "loss": 0.005, + "step": 75910 + }, + { + "epoch": 11.792482137309724, + "grad_norm": 0.1955009549856186, + "learning_rate": 1.5033769716044144e-05, + "loss": 0.0041, + "step": 75920 + }, + { + "epoch": 11.79403541472507, + "grad_norm": 0.3789423406124115, + "learning_rate": 1.5021952555887975e-05, + "loss": 0.0055, + "step": 75930 + }, + { + "epoch": 11.795588692140416, + "grad_norm": 0.20212110877037048, + "learning_rate": 1.50101392208785e-05, + "loss": 0.0048, + "step": 75940 + }, + { + "epoch": 11.797141969555764, + "grad_norm": 0.18743537366390228, + "learning_rate": 1.4998329712307579e-05, + "loss": 0.0041, + "step": 75950 + }, + { + "epoch": 11.798695246971109, + "grad_norm": 0.29571375250816345, + "learning_rate": 1.4986524031466708e-05, + "loss": 0.0061, + "step": 75960 + }, + { + "epoch": 11.800248524386456, + "grad_norm": 0.15539495646953583, + "learning_rate": 1.4974722179646916e-05, + "loss": 0.0043, + "step": 75970 + }, + { + "epoch": 11.801801801801801, + "grad_norm": 0.04440300166606903, + "learning_rate": 1.4962924158138847e-05, + "loss": 0.0037, + "step": 75980 + }, + { + "epoch": 11.803355079217148, + "grad_norm": 0.16103781759738922, + "learning_rate": 1.4951129968232714e-05, + "loss": 0.004, + "step": 75990 + }, + { + "epoch": 11.804908356632495, + "grad_norm": 0.17656248807907104, + "learning_rate": 1.4939339611218329e-05, + "loss": 0.0037, + "step": 76000 + }, + { + "epoch": 11.80646163404784, + "grad_norm": 0.16073431074619293, + "learning_rate": 1.4927553088385032e-05, + "loss": 0.0033, + "step": 76010 + }, + { + "epoch": 11.808014911463188, + "grad_norm": 0.2417776882648468, + "learning_rate": 1.4915770401021801e-05, + "loss": 0.0045, + "step": 76020 + }, + { + "epoch": 11.809568188878533, + "grad_norm": 0.13424444198608398, + "learning_rate": 1.4903991550417185e-05, + "loss": 0.0031, + "step": 76030 + }, + { + "epoch": 11.81112146629388, + "grad_norm": 0.07972048968076706, + "learning_rate": 1.4892216537859277e-05, + "loss": 0.004, + "step": 76040 + }, + { + "epoch": 11.812674743709227, + "grad_norm": 0.15216319262981415, + "learning_rate": 1.4880445364635775e-05, + "loss": 0.004, + "step": 76050 + }, + { + "epoch": 11.814228021124572, + "grad_norm": 0.3022521734237671, + "learning_rate": 1.4868678032033973e-05, + "loss": 0.006, + "step": 76060 + }, + { + "epoch": 11.81578129853992, + "grad_norm": 0.3961845934391022, + "learning_rate": 1.4856914541340739e-05, + "loss": 0.0036, + "step": 76070 + }, + { + "epoch": 11.817334575955266, + "grad_norm": 0.36575692892074585, + "learning_rate": 1.4845154893842472e-05, + "loss": 0.0038, + "step": 76080 + }, + { + "epoch": 11.818887853370612, + "grad_norm": 0.19221234321594238, + "learning_rate": 1.4833399090825234e-05, + "loss": 0.004, + "step": 76090 + }, + { + "epoch": 11.820441130785959, + "grad_norm": 0.24157226085662842, + "learning_rate": 1.4821647133574562e-05, + "loss": 0.0035, + "step": 76100 + }, + { + "epoch": 11.821994408201304, + "grad_norm": 0.14494694769382477, + "learning_rate": 1.4809899023375706e-05, + "loss": 0.0039, + "step": 76110 + }, + { + "epoch": 11.823547685616651, + "grad_norm": 0.3030936121940613, + "learning_rate": 1.4798154761513366e-05, + "loss": 0.0049, + "step": 76120 + }, + { + "epoch": 11.825100963031998, + "grad_norm": 0.1611282378435135, + "learning_rate": 1.4786414349271916e-05, + "loss": 0.0038, + "step": 76130 + }, + { + "epoch": 11.826654240447343, + "grad_norm": 0.27155569195747375, + "learning_rate": 1.4774677787935226e-05, + "loss": 0.0034, + "step": 76140 + }, + { + "epoch": 11.82820751786269, + "grad_norm": 0.139035165309906, + "learning_rate": 1.4762945078786838e-05, + "loss": 0.0045, + "step": 76150 + }, + { + "epoch": 11.829760795278037, + "grad_norm": 0.18332821130752563, + "learning_rate": 1.4751216223109777e-05, + "loss": 0.0049, + "step": 76160 + }, + { + "epoch": 11.831314072693383, + "grad_norm": 0.17584367096424103, + "learning_rate": 1.4739491222186714e-05, + "loss": 0.0049, + "step": 76170 + }, + { + "epoch": 11.83286735010873, + "grad_norm": 0.1610264927148819, + "learning_rate": 1.4727770077299879e-05, + "loss": 0.0059, + "step": 76180 + }, + { + "epoch": 11.834420627524075, + "grad_norm": 0.2570303976535797, + "learning_rate": 1.4716052789731094e-05, + "loss": 0.0041, + "step": 76190 + }, + { + "epoch": 11.835973904939422, + "grad_norm": 0.3748077154159546, + "learning_rate": 1.470433936076171e-05, + "loss": 0.0069, + "step": 76200 + }, + { + "epoch": 11.83752718235477, + "grad_norm": 0.09772200137376785, + "learning_rate": 1.4692629791672719e-05, + "loss": 0.0025, + "step": 76210 + }, + { + "epoch": 11.839080459770114, + "grad_norm": 0.24643005430698395, + "learning_rate": 1.468092408374464e-05, + "loss": 0.005, + "step": 76220 + }, + { + "epoch": 11.840633737185462, + "grad_norm": 0.5915942192077637, + "learning_rate": 1.4669222238257602e-05, + "loss": 0.0058, + "step": 76230 + }, + { + "epoch": 11.842187014600807, + "grad_norm": 0.5092863440513611, + "learning_rate": 1.4657524256491301e-05, + "loss": 0.0048, + "step": 76240 + }, + { + "epoch": 11.843740292016154, + "grad_norm": 0.5239884257316589, + "learning_rate": 1.464583013972502e-05, + "loss": 0.0039, + "step": 76250 + }, + { + "epoch": 11.845293569431501, + "grad_norm": 0.05703571066260338, + "learning_rate": 1.4634139889237591e-05, + "loss": 0.0036, + "step": 76260 + }, + { + "epoch": 11.846846846846846, + "grad_norm": 0.13615411520004272, + "learning_rate": 1.4622453506307448e-05, + "loss": 0.0041, + "step": 76270 + }, + { + "epoch": 11.848400124262193, + "grad_norm": 0.2978256642818451, + "learning_rate": 1.461077099221262e-05, + "loss": 0.0032, + "step": 76280 + }, + { + "epoch": 11.84995340167754, + "grad_norm": 0.17444437742233276, + "learning_rate": 1.4599092348230652e-05, + "loss": 0.0044, + "step": 76290 + }, + { + "epoch": 11.851506679092886, + "grad_norm": 0.07487655431032181, + "learning_rate": 1.4587417575638724e-05, + "loss": 0.0033, + "step": 76300 + }, + { + "epoch": 11.853059956508233, + "grad_norm": 1.1299433708190918, + "learning_rate": 1.4575746675713563e-05, + "loss": 0.0044, + "step": 76310 + }, + { + "epoch": 11.85461323392358, + "grad_norm": 0.3350653350353241, + "learning_rate": 1.4564079649731504e-05, + "loss": 0.0049, + "step": 76320 + }, + { + "epoch": 11.856166511338925, + "grad_norm": 0.22803446650505066, + "learning_rate": 1.4552416498968402e-05, + "loss": 0.0075, + "step": 76330 + }, + { + "epoch": 11.857719788754272, + "grad_norm": 0.1188831627368927, + "learning_rate": 1.4540757224699747e-05, + "loss": 0.0043, + "step": 76340 + }, + { + "epoch": 11.859273066169617, + "grad_norm": 0.8070045113563538, + "learning_rate": 1.452910182820056e-05, + "loss": 0.0052, + "step": 76350 + }, + { + "epoch": 11.860826343584964, + "grad_norm": 0.29104411602020264, + "learning_rate": 1.4517450310745468e-05, + "loss": 0.0048, + "step": 76360 + }, + { + "epoch": 11.862379621000311, + "grad_norm": 0.20006173849105835, + "learning_rate": 1.450580267360866e-05, + "loss": 0.0053, + "step": 76370 + }, + { + "epoch": 11.863932898415657, + "grad_norm": 0.394898384809494, + "learning_rate": 1.4494158918063922e-05, + "loss": 0.0051, + "step": 76380 + }, + { + "epoch": 11.865486175831004, + "grad_norm": 0.7172906994819641, + "learning_rate": 1.4482519045384568e-05, + "loss": 0.0046, + "step": 76390 + }, + { + "epoch": 11.867039453246349, + "grad_norm": 0.6026968955993652, + "learning_rate": 1.4470883056843548e-05, + "loss": 0.0039, + "step": 76400 + }, + { + "epoch": 11.868592730661696, + "grad_norm": 0.23199278116226196, + "learning_rate": 1.4459250953713322e-05, + "loss": 0.0052, + "step": 76410 + }, + { + "epoch": 11.870146008077043, + "grad_norm": 0.30503422021865845, + "learning_rate": 1.444762273726598e-05, + "loss": 0.0052, + "step": 76420 + }, + { + "epoch": 11.871699285492388, + "grad_norm": 0.34928134083747864, + "learning_rate": 1.443599840877316e-05, + "loss": 0.0047, + "step": 76430 + }, + { + "epoch": 11.873252562907735, + "grad_norm": 2.8535666465759277, + "learning_rate": 1.4424377969506103e-05, + "loss": 0.0045, + "step": 76440 + }, + { + "epoch": 11.874805840323083, + "grad_norm": 0.09410843253135681, + "learning_rate": 1.4412761420735565e-05, + "loss": 0.0044, + "step": 76450 + }, + { + "epoch": 11.876359117738428, + "grad_norm": 0.23456904292106628, + "learning_rate": 1.4401148763731953e-05, + "loss": 0.005, + "step": 76460 + }, + { + "epoch": 11.877912395153775, + "grad_norm": 0.07265105843544006, + "learning_rate": 1.4389539999765173e-05, + "loss": 0.0036, + "step": 76470 + }, + { + "epoch": 11.87946567256912, + "grad_norm": 0.23868833482265472, + "learning_rate": 1.4377935130104758e-05, + "loss": 0.0051, + "step": 76480 + }, + { + "epoch": 11.881018949984467, + "grad_norm": 0.4161061942577362, + "learning_rate": 1.4366334156019801e-05, + "loss": 0.0031, + "step": 76490 + }, + { + "epoch": 11.882572227399814, + "grad_norm": 0.17573218047618866, + "learning_rate": 1.435473707877898e-05, + "loss": 0.005, + "step": 76500 + }, + { + "epoch": 11.88412550481516, + "grad_norm": 0.31579503417015076, + "learning_rate": 1.4343143899650508e-05, + "loss": 0.0049, + "step": 76510 + }, + { + "epoch": 11.885678782230507, + "grad_norm": 0.264680415391922, + "learning_rate": 1.4331554619902204e-05, + "loss": 0.0038, + "step": 76520 + }, + { + "epoch": 11.887232059645854, + "grad_norm": 0.1299149990081787, + "learning_rate": 1.4319969240801484e-05, + "loss": 0.0046, + "step": 76530 + }, + { + "epoch": 11.888785337061199, + "grad_norm": 0.5804929733276367, + "learning_rate": 1.430838776361526e-05, + "loss": 0.0058, + "step": 76540 + }, + { + "epoch": 11.890338614476546, + "grad_norm": 0.17165936529636383, + "learning_rate": 1.4296810189610093e-05, + "loss": 0.0042, + "step": 76550 + }, + { + "epoch": 11.891891891891891, + "grad_norm": 0.08596237748861313, + "learning_rate": 1.4285236520052087e-05, + "loss": 0.0056, + "step": 76560 + }, + { + "epoch": 11.893445169307238, + "grad_norm": 0.6201119422912598, + "learning_rate": 1.4273666756206933e-05, + "loss": 0.0036, + "step": 76570 + }, + { + "epoch": 11.894998446722585, + "grad_norm": 0.08677850663661957, + "learning_rate": 1.4262100899339848e-05, + "loss": 0.0037, + "step": 76580 + }, + { + "epoch": 11.89655172413793, + "grad_norm": 0.33054161071777344, + "learning_rate": 1.4250538950715696e-05, + "loss": 0.0032, + "step": 76590 + }, + { + "epoch": 11.898105001553278, + "grad_norm": 0.3053969442844391, + "learning_rate": 1.4238980911598842e-05, + "loss": 0.0074, + "step": 76600 + }, + { + "epoch": 11.899658278968623, + "grad_norm": 0.11404471844434738, + "learning_rate": 1.4227426783253267e-05, + "loss": 0.0032, + "step": 76610 + }, + { + "epoch": 11.90121155638397, + "grad_norm": 0.23580385744571686, + "learning_rate": 1.4215876566942521e-05, + "loss": 0.0043, + "step": 76620 + }, + { + "epoch": 11.902764833799317, + "grad_norm": 0.1181488037109375, + "learning_rate": 1.4204330263929721e-05, + "loss": 0.0037, + "step": 76630 + }, + { + "epoch": 11.904318111214662, + "grad_norm": 0.4906100630760193, + "learning_rate": 1.4192787875477537e-05, + "loss": 0.0037, + "step": 76640 + }, + { + "epoch": 11.90587138863001, + "grad_norm": 0.11276430636644363, + "learning_rate": 1.4181249402848246e-05, + "loss": 0.0043, + "step": 76650 + }, + { + "epoch": 11.907424666045356, + "grad_norm": 0.4872598946094513, + "learning_rate": 1.4169714847303656e-05, + "loss": 0.0043, + "step": 76660 + }, + { + "epoch": 11.908977943460702, + "grad_norm": 0.31405752897262573, + "learning_rate": 1.4158184210105197e-05, + "loss": 0.005, + "step": 76670 + }, + { + "epoch": 11.910531220876049, + "grad_norm": 0.14548005163669586, + "learning_rate": 1.4146657492513787e-05, + "loss": 0.0046, + "step": 76680 + }, + { + "epoch": 11.912084498291394, + "grad_norm": 0.23648902773857117, + "learning_rate": 1.4135134695790047e-05, + "loss": 0.0043, + "step": 76690 + }, + { + "epoch": 11.913637775706741, + "grad_norm": 0.5251891016960144, + "learning_rate": 1.4123615821194037e-05, + "loss": 0.003, + "step": 76700 + }, + { + "epoch": 11.915191053122088, + "grad_norm": 0.19846421480178833, + "learning_rate": 1.4112100869985473e-05, + "loss": 0.0048, + "step": 76710 + }, + { + "epoch": 11.916744330537433, + "grad_norm": 0.5799883008003235, + "learning_rate": 1.4100589843423589e-05, + "loss": 0.0038, + "step": 76720 + }, + { + "epoch": 11.91829760795278, + "grad_norm": 0.17670230567455292, + "learning_rate": 1.408908274276723e-05, + "loss": 0.0036, + "step": 76730 + }, + { + "epoch": 11.919850885368128, + "grad_norm": 0.20518241822719574, + "learning_rate": 1.4077579569274786e-05, + "loss": 0.0035, + "step": 76740 + }, + { + "epoch": 11.921404162783473, + "grad_norm": 0.37139472365379333, + "learning_rate": 1.4066080324204245e-05, + "loss": 0.0043, + "step": 76750 + }, + { + "epoch": 11.92295744019882, + "grad_norm": 0.10707151889801025, + "learning_rate": 1.405458500881312e-05, + "loss": 0.0045, + "step": 76760 + }, + { + "epoch": 11.924510717614165, + "grad_norm": 0.1691114753484726, + "learning_rate": 1.4043093624358538e-05, + "loss": 0.004, + "step": 76770 + }, + { + "epoch": 11.926063995029512, + "grad_norm": 0.12006210535764694, + "learning_rate": 1.4031606172097189e-05, + "loss": 0.0054, + "step": 76780 + }, + { + "epoch": 11.92761727244486, + "grad_norm": 0.08501261472702026, + "learning_rate": 1.40201226532853e-05, + "loss": 0.0063, + "step": 76790 + }, + { + "epoch": 11.929170549860205, + "grad_norm": 0.12931473553180695, + "learning_rate": 1.4008643069178701e-05, + "loss": 0.0034, + "step": 76800 + }, + { + "epoch": 11.930723827275552, + "grad_norm": 0.11968027800321579, + "learning_rate": 1.399716742103278e-05, + "loss": 0.0042, + "step": 76810 + }, + { + "epoch": 11.932277104690897, + "grad_norm": 0.15338371694087982, + "learning_rate": 1.398569571010252e-05, + "loss": 0.0046, + "step": 76820 + }, + { + "epoch": 11.933830382106244, + "grad_norm": 0.2727174460887909, + "learning_rate": 1.3974227937642414e-05, + "loss": 0.0052, + "step": 76830 + }, + { + "epoch": 11.935383659521591, + "grad_norm": 0.13766175508499146, + "learning_rate": 1.3962764104906595e-05, + "loss": 0.0035, + "step": 76840 + }, + { + "epoch": 11.936936936936936, + "grad_norm": 1.2140073776245117, + "learning_rate": 1.3951304213148696e-05, + "loss": 0.0047, + "step": 76850 + }, + { + "epoch": 11.938490214352283, + "grad_norm": 0.32242488861083984, + "learning_rate": 1.3939848263621979e-05, + "loss": 0.0046, + "step": 76860 + }, + { + "epoch": 11.94004349176763, + "grad_norm": 0.2505887746810913, + "learning_rate": 1.3928396257579219e-05, + "loss": 0.0048, + "step": 76870 + }, + { + "epoch": 11.941596769182976, + "grad_norm": 0.1428765505552292, + "learning_rate": 1.3916948196272833e-05, + "loss": 0.0037, + "step": 76880 + }, + { + "epoch": 11.943150046598323, + "grad_norm": 0.44658878445625305, + "learning_rate": 1.3905504080954734e-05, + "loss": 0.005, + "step": 76890 + }, + { + "epoch": 11.94470332401367, + "grad_norm": 0.2467750757932663, + "learning_rate": 1.3894063912876454e-05, + "loss": 0.0047, + "step": 76900 + }, + { + "epoch": 11.946256601429015, + "grad_norm": 0.18945686519145966, + "learning_rate": 1.3882627693289046e-05, + "loss": 0.0047, + "step": 76910 + }, + { + "epoch": 11.947809878844362, + "grad_norm": 0.1200399398803711, + "learning_rate": 1.3871195423443183e-05, + "loss": 0.003, + "step": 76920 + }, + { + "epoch": 11.949363156259707, + "grad_norm": 0.13856743276119232, + "learning_rate": 1.3859767104589044e-05, + "loss": 0.0037, + "step": 76930 + }, + { + "epoch": 11.950916433675054, + "grad_norm": 0.18500274419784546, + "learning_rate": 1.3848342737976467e-05, + "loss": 0.0036, + "step": 76940 + }, + { + "epoch": 11.952469711090401, + "grad_norm": 0.25238633155822754, + "learning_rate": 1.3836922324854757e-05, + "loss": 0.0044, + "step": 76950 + }, + { + "epoch": 11.954022988505747, + "grad_norm": 0.2979259788990021, + "learning_rate": 1.382550586647287e-05, + "loss": 0.0035, + "step": 76960 + }, + { + "epoch": 11.955576265921094, + "grad_norm": 0.07887810468673706, + "learning_rate": 1.3814093364079257e-05, + "loss": 0.003, + "step": 76970 + }, + { + "epoch": 11.957129543336439, + "grad_norm": 0.23205405473709106, + "learning_rate": 1.3802684818921991e-05, + "loss": 0.0046, + "step": 76980 + }, + { + "epoch": 11.958682820751786, + "grad_norm": 0.4680868685245514, + "learning_rate": 1.3791280232248687e-05, + "loss": 0.0043, + "step": 76990 + }, + { + "epoch": 11.960236098167133, + "grad_norm": 0.099068284034729, + "learning_rate": 1.3779879605306555e-05, + "loss": 0.0039, + "step": 77000 + }, + { + "epoch": 11.961789375582478, + "grad_norm": 0.3186691403388977, + "learning_rate": 1.376848293934232e-05, + "loss": 0.0044, + "step": 77010 + }, + { + "epoch": 11.963342652997826, + "grad_norm": 0.2178763598203659, + "learning_rate": 1.3757090235602316e-05, + "loss": 0.0034, + "step": 77020 + }, + { + "epoch": 11.964895930413173, + "grad_norm": 0.20145967602729797, + "learning_rate": 1.3745701495332447e-05, + "loss": 0.0036, + "step": 77030 + }, + { + "epoch": 11.966449207828518, + "grad_norm": 0.19817830622196198, + "learning_rate": 1.3734316719778145e-05, + "loss": 0.0048, + "step": 77040 + }, + { + "epoch": 11.968002485243865, + "grad_norm": 0.3588070869445801, + "learning_rate": 1.3722935910184454e-05, + "loss": 0.0028, + "step": 77050 + }, + { + "epoch": 11.96955576265921, + "grad_norm": 0.3032948076725006, + "learning_rate": 1.3711559067795921e-05, + "loss": 0.0026, + "step": 77060 + }, + { + "epoch": 11.971109040074557, + "grad_norm": 0.3744406998157501, + "learning_rate": 1.3700186193856756e-05, + "loss": 0.0033, + "step": 77070 + }, + { + "epoch": 11.972662317489904, + "grad_norm": 0.34825366735458374, + "learning_rate": 1.368881728961064e-05, + "loss": 0.0031, + "step": 77080 + }, + { + "epoch": 11.97421559490525, + "grad_norm": 0.6610490083694458, + "learning_rate": 1.3677452356300885e-05, + "loss": 0.0051, + "step": 77090 + }, + { + "epoch": 11.975768872320597, + "grad_norm": 0.284860759973526, + "learning_rate": 1.3666091395170321e-05, + "loss": 0.0045, + "step": 77100 + }, + { + "epoch": 11.977322149735944, + "grad_norm": 0.11856839805841446, + "learning_rate": 1.3654734407461384e-05, + "loss": 0.0053, + "step": 77110 + }, + { + "epoch": 11.978875427151289, + "grad_norm": 0.1154608502984047, + "learning_rate": 1.3643381394416022e-05, + "loss": 0.0045, + "step": 77120 + }, + { + "epoch": 11.980428704566636, + "grad_norm": 0.3405645191669464, + "learning_rate": 1.3632032357275838e-05, + "loss": 0.0065, + "step": 77130 + }, + { + "epoch": 11.981981981981981, + "grad_norm": 0.2897831201553345, + "learning_rate": 1.3620687297281898e-05, + "loss": 0.0041, + "step": 77140 + }, + { + "epoch": 11.983535259397328, + "grad_norm": 0.16267867386341095, + "learning_rate": 1.3609346215674917e-05, + "loss": 0.0075, + "step": 77150 + }, + { + "epoch": 11.985088536812675, + "grad_norm": 0.2923293113708496, + "learning_rate": 1.3598009113695109e-05, + "loss": 0.005, + "step": 77160 + }, + { + "epoch": 11.98664181422802, + "grad_norm": 0.7428278923034668, + "learning_rate": 1.358667599258231e-05, + "loss": 0.0059, + "step": 77170 + }, + { + "epoch": 11.988195091643368, + "grad_norm": 0.24885357916355133, + "learning_rate": 1.357534685357586e-05, + "loss": 0.0033, + "step": 77180 + }, + { + "epoch": 11.989748369058713, + "grad_norm": 0.7442656755447388, + "learning_rate": 1.3564021697914715e-05, + "loss": 0.0039, + "step": 77190 + }, + { + "epoch": 11.99130164647406, + "grad_norm": 0.20888791978359222, + "learning_rate": 1.3552700526837381e-05, + "loss": 0.0045, + "step": 77200 + }, + { + "epoch": 11.992854923889407, + "grad_norm": 0.32704678177833557, + "learning_rate": 1.3541383341581925e-05, + "loss": 0.0055, + "step": 77210 + }, + { + "epoch": 11.994408201304752, + "grad_norm": 0.14185300469398499, + "learning_rate": 1.3530070143385965e-05, + "loss": 0.0029, + "step": 77220 + }, + { + "epoch": 11.9959614787201, + "grad_norm": 0.28803178668022156, + "learning_rate": 1.3518760933486702e-05, + "loss": 0.0052, + "step": 77230 + }, + { + "epoch": 11.997514756135446, + "grad_norm": 0.3333883285522461, + "learning_rate": 1.3507455713120908e-05, + "loss": 0.003, + "step": 77240 + }, + { + "epoch": 11.999068033550792, + "grad_norm": 0.16366247832775116, + "learning_rate": 1.3496154483524876e-05, + "loss": 0.0055, + "step": 77250 + }, + { + "epoch": 12.000621310966139, + "grad_norm": 0.18396633863449097, + "learning_rate": 1.3484857245934506e-05, + "loss": 0.0059, + "step": 77260 + }, + { + "epoch": 12.002174588381484, + "grad_norm": 0.48095056414604187, + "learning_rate": 1.3473564001585248e-05, + "loss": 0.0045, + "step": 77270 + }, + { + "epoch": 12.003727865796831, + "grad_norm": 0.17973534762859344, + "learning_rate": 1.3462274751712129e-05, + "loss": 0.005, + "step": 77280 + }, + { + "epoch": 12.005281143212178, + "grad_norm": 0.18013662099838257, + "learning_rate": 1.3450989497549698e-05, + "loss": 0.0033, + "step": 77290 + }, + { + "epoch": 12.006834420627523, + "grad_norm": 0.38646531105041504, + "learning_rate": 1.3439708240332115e-05, + "loss": 0.005, + "step": 77300 + }, + { + "epoch": 12.00838769804287, + "grad_norm": 0.8711829781532288, + "learning_rate": 1.3428430981293043e-05, + "loss": 0.0043, + "step": 77310 + }, + { + "epoch": 12.009940975458218, + "grad_norm": 0.13591091334819794, + "learning_rate": 1.3417157721665808e-05, + "loss": 0.0056, + "step": 77320 + }, + { + "epoch": 12.011494252873563, + "grad_norm": 0.06980961561203003, + "learning_rate": 1.3405888462683186e-05, + "loss": 0.0029, + "step": 77330 + }, + { + "epoch": 12.01304753028891, + "grad_norm": 0.17116664350032806, + "learning_rate": 1.3394623205577595e-05, + "loss": 0.0051, + "step": 77340 + }, + { + "epoch": 12.014600807704255, + "grad_norm": 0.25034645199775696, + "learning_rate": 1.3383361951580963e-05, + "loss": 0.0036, + "step": 77350 + }, + { + "epoch": 12.016154085119602, + "grad_norm": 0.33781561255455017, + "learning_rate": 1.337210470192483e-05, + "loss": 0.0038, + "step": 77360 + }, + { + "epoch": 12.01770736253495, + "grad_norm": 0.10370053350925446, + "learning_rate": 1.3360851457840245e-05, + "loss": 0.0059, + "step": 77370 + }, + { + "epoch": 12.019260639950295, + "grad_norm": 0.684610903263092, + "learning_rate": 1.3349602220557856e-05, + "loss": 0.0041, + "step": 77380 + }, + { + "epoch": 12.020813917365642, + "grad_norm": 0.11621585488319397, + "learning_rate": 1.3338356991307865e-05, + "loss": 0.005, + "step": 77390 + }, + { + "epoch": 12.022367194780989, + "grad_norm": 0.330899178981781, + "learning_rate": 1.3327115771320048e-05, + "loss": 0.003, + "step": 77400 + }, + { + "epoch": 12.023920472196334, + "grad_norm": 0.3539591133594513, + "learning_rate": 1.3315878561823697e-05, + "loss": 0.004, + "step": 77410 + }, + { + "epoch": 12.025473749611681, + "grad_norm": 0.35185739398002625, + "learning_rate": 1.330464536404773e-05, + "loss": 0.0044, + "step": 77420 + }, + { + "epoch": 12.027027027027026, + "grad_norm": 0.6061408519744873, + "learning_rate": 1.3293416179220563e-05, + "loss": 0.0034, + "step": 77430 + }, + { + "epoch": 12.028580304442373, + "grad_norm": 0.21204113960266113, + "learning_rate": 1.328219100857021e-05, + "loss": 0.0036, + "step": 77440 + }, + { + "epoch": 12.03013358185772, + "grad_norm": 0.5672768950462341, + "learning_rate": 1.327096985332425e-05, + "loss": 0.0042, + "step": 77450 + }, + { + "epoch": 12.031686859273066, + "grad_norm": 0.5347579121589661, + "learning_rate": 1.3259752714709816e-05, + "loss": 0.0034, + "step": 77460 + }, + { + "epoch": 12.033240136688413, + "grad_norm": 0.6831604242324829, + "learning_rate": 1.3248539593953569e-05, + "loss": 0.0051, + "step": 77470 + }, + { + "epoch": 12.03479341410376, + "grad_norm": 0.3954754173755646, + "learning_rate": 1.323733049228178e-05, + "loss": 0.0058, + "step": 77480 + }, + { + "epoch": 12.036346691519105, + "grad_norm": 0.12270206958055496, + "learning_rate": 1.3226125410920276e-05, + "loss": 0.0042, + "step": 77490 + }, + { + "epoch": 12.037899968934452, + "grad_norm": 0.07927706837654114, + "learning_rate": 1.3214924351094388e-05, + "loss": 0.0044, + "step": 77500 + }, + { + "epoch": 12.039453246349797, + "grad_norm": 0.2537188231945038, + "learning_rate": 1.3203727314029069e-05, + "loss": 0.0057, + "step": 77510 + }, + { + "epoch": 12.041006523765144, + "grad_norm": 0.23453141748905182, + "learning_rate": 1.3192534300948811e-05, + "loss": 0.0058, + "step": 77520 + }, + { + "epoch": 12.042559801180492, + "grad_norm": 0.5634287595748901, + "learning_rate": 1.3181345313077675e-05, + "loss": 0.0037, + "step": 77530 + }, + { + "epoch": 12.044113078595837, + "grad_norm": 0.2796516418457031, + "learning_rate": 1.3170160351639243e-05, + "loss": 0.0039, + "step": 77540 + }, + { + "epoch": 12.045666356011184, + "grad_norm": 0.09099318087100983, + "learning_rate": 1.3158979417856721e-05, + "loss": 0.004, + "step": 77550 + }, + { + "epoch": 12.04721963342653, + "grad_norm": 0.18005093932151794, + "learning_rate": 1.31478025129528e-05, + "loss": 0.0052, + "step": 77560 + }, + { + "epoch": 12.048772910841876, + "grad_norm": 0.18383654952049255, + "learning_rate": 1.3136629638149794e-05, + "loss": 0.0033, + "step": 77570 + }, + { + "epoch": 12.050326188257223, + "grad_norm": 0.12913690507411957, + "learning_rate": 1.3125460794669547e-05, + "loss": 0.0036, + "step": 77580 + }, + { + "epoch": 12.051879465672569, + "grad_norm": 0.13480162620544434, + "learning_rate": 1.3114295983733483e-05, + "loss": 0.004, + "step": 77590 + }, + { + "epoch": 12.053432743087916, + "grad_norm": 0.15540148317813873, + "learning_rate": 1.3103135206562534e-05, + "loss": 0.0035, + "step": 77600 + }, + { + "epoch": 12.054986020503263, + "grad_norm": 0.31260862946510315, + "learning_rate": 1.3091978464377264e-05, + "loss": 0.005, + "step": 77610 + }, + { + "epoch": 12.056539297918608, + "grad_norm": 0.14944113790988922, + "learning_rate": 1.3080825758397724e-05, + "loss": 0.0042, + "step": 77620 + }, + { + "epoch": 12.058092575333955, + "grad_norm": 0.25068435072898865, + "learning_rate": 1.3069677089843568e-05, + "loss": 0.0063, + "step": 77630 + }, + { + "epoch": 12.0596458527493, + "grad_norm": 0.31365612149238586, + "learning_rate": 1.305853245993401e-05, + "loss": 0.0036, + "step": 77640 + }, + { + "epoch": 12.061199130164647, + "grad_norm": 0.42459508776664734, + "learning_rate": 1.3047391869887809e-05, + "loss": 0.0039, + "step": 77650 + }, + { + "epoch": 12.062752407579994, + "grad_norm": 0.18057531118392944, + "learning_rate": 1.3036255320923263e-05, + "loss": 0.0047, + "step": 77660 + }, + { + "epoch": 12.06430568499534, + "grad_norm": 0.487621009349823, + "learning_rate": 1.3025122814258273e-05, + "loss": 0.0076, + "step": 77670 + }, + { + "epoch": 12.065858962410687, + "grad_norm": 0.5909806489944458, + "learning_rate": 1.3013994351110249e-05, + "loss": 0.0054, + "step": 77680 + }, + { + "epoch": 12.067412239826034, + "grad_norm": 0.08182959258556366, + "learning_rate": 1.3002869932696194e-05, + "loss": 0.0039, + "step": 77690 + }, + { + "epoch": 12.068965517241379, + "grad_norm": 0.22234217822551727, + "learning_rate": 1.2991749560232664e-05, + "loss": 0.0035, + "step": 77700 + }, + { + "epoch": 12.070518794656726, + "grad_norm": 0.32212138175964355, + "learning_rate": 1.2980633234935768e-05, + "loss": 0.0042, + "step": 77710 + }, + { + "epoch": 12.072072072072071, + "grad_norm": 0.382282555103302, + "learning_rate": 1.2969520958021153e-05, + "loss": 0.0037, + "step": 77720 + }, + { + "epoch": 12.073625349487418, + "grad_norm": 0.13392215967178345, + "learning_rate": 1.295841273070405e-05, + "loss": 0.0052, + "step": 77730 + }, + { + "epoch": 12.075178626902765, + "grad_norm": 0.5163207054138184, + "learning_rate": 1.2947308554199255e-05, + "loss": 0.0061, + "step": 77740 + }, + { + "epoch": 12.07673190431811, + "grad_norm": 0.11470451205968857, + "learning_rate": 1.2936208429721075e-05, + "loss": 0.004, + "step": 77750 + }, + { + "epoch": 12.078285181733458, + "grad_norm": 0.4785136282444, + "learning_rate": 1.2925112358483416e-05, + "loss": 0.0054, + "step": 77760 + }, + { + "epoch": 12.079838459148805, + "grad_norm": 0.29743170738220215, + "learning_rate": 1.2914020341699728e-05, + "loss": 0.0038, + "step": 77770 + }, + { + "epoch": 12.08139173656415, + "grad_norm": 0.548364520072937, + "learning_rate": 1.2902932380583032e-05, + "loss": 0.0045, + "step": 77780 + }, + { + "epoch": 12.082945013979497, + "grad_norm": 0.1740257292985916, + "learning_rate": 1.2891848476345863e-05, + "loss": 0.0045, + "step": 77790 + }, + { + "epoch": 12.084498291394842, + "grad_norm": 0.3370230197906494, + "learning_rate": 1.2880768630200368e-05, + "loss": 0.0043, + "step": 77800 + }, + { + "epoch": 12.08605156881019, + "grad_norm": 0.17287980020046234, + "learning_rate": 1.286969284335819e-05, + "loss": 0.0037, + "step": 77810 + }, + { + "epoch": 12.087604846225537, + "grad_norm": 0.5146299004554749, + "learning_rate": 1.2858621117030578e-05, + "loss": 0.0038, + "step": 77820 + }, + { + "epoch": 12.089158123640882, + "grad_norm": 0.8917320370674133, + "learning_rate": 1.2847553452428324e-05, + "loss": 0.0048, + "step": 77830 + }, + { + "epoch": 12.090711401056229, + "grad_norm": 1.010943055152893, + "learning_rate": 1.283648985076178e-05, + "loss": 0.0046, + "step": 77840 + }, + { + "epoch": 12.092264678471574, + "grad_norm": 0.12104608118534088, + "learning_rate": 1.2825430313240817e-05, + "loss": 0.0042, + "step": 77850 + }, + { + "epoch": 12.093817955886921, + "grad_norm": 0.32957497239112854, + "learning_rate": 1.281437484107491e-05, + "loss": 0.0038, + "step": 77860 + }, + { + "epoch": 12.095371233302268, + "grad_norm": 0.2901570498943329, + "learning_rate": 1.2803323435473058e-05, + "loss": 0.0035, + "step": 77870 + }, + { + "epoch": 12.096924510717614, + "grad_norm": 0.13612069189548492, + "learning_rate": 1.2792276097643841e-05, + "loss": 0.0075, + "step": 77880 + }, + { + "epoch": 12.09847778813296, + "grad_norm": 0.1601485162973404, + "learning_rate": 1.2781232828795342e-05, + "loss": 0.003, + "step": 77890 + }, + { + "epoch": 12.100031065548308, + "grad_norm": 0.14188452064990997, + "learning_rate": 1.2770193630135292e-05, + "loss": 0.0044, + "step": 77900 + }, + { + "epoch": 12.101584342963653, + "grad_norm": 0.410351425409317, + "learning_rate": 1.2759158502870877e-05, + "loss": 0.0045, + "step": 77910 + }, + { + "epoch": 12.103137620379, + "grad_norm": 0.14369778335094452, + "learning_rate": 1.2748127448208913e-05, + "loss": 0.0035, + "step": 77920 + }, + { + "epoch": 12.104690897794345, + "grad_norm": 0.16805362701416016, + "learning_rate": 1.2737100467355706e-05, + "loss": 0.0037, + "step": 77930 + }, + { + "epoch": 12.106244175209692, + "grad_norm": 0.27911385893821716, + "learning_rate": 1.2726077561517169e-05, + "loss": 0.0044, + "step": 77940 + }, + { + "epoch": 12.10779745262504, + "grad_norm": 0.16378678381443024, + "learning_rate": 1.2715058731898755e-05, + "loss": 0.0033, + "step": 77950 + }, + { + "epoch": 12.109350730040385, + "grad_norm": 0.1360207200050354, + "learning_rate": 1.2704043979705471e-05, + "loss": 0.0034, + "step": 77960 + }, + { + "epoch": 12.110904007455732, + "grad_norm": 0.15228615701198578, + "learning_rate": 1.269303330614185e-05, + "loss": 0.0028, + "step": 77970 + }, + { + "epoch": 12.112457284871079, + "grad_norm": 0.07530476897954941, + "learning_rate": 1.2682026712412015e-05, + "loss": 0.0051, + "step": 77980 + }, + { + "epoch": 12.114010562286424, + "grad_norm": 0.19186876714229584, + "learning_rate": 1.2671024199719644e-05, + "loss": 0.0033, + "step": 77990 + }, + { + "epoch": 12.115563839701771, + "grad_norm": 0.34914258122444153, + "learning_rate": 1.2660025769267936e-05, + "loss": 0.0056, + "step": 78000 + }, + { + "epoch": 12.117117117117116, + "grad_norm": 0.6887285709381104, + "learning_rate": 1.2649031422259667e-05, + "loss": 0.0036, + "step": 78010 + }, + { + "epoch": 12.118670394532463, + "grad_norm": 0.1285756379365921, + "learning_rate": 1.2638041159897163e-05, + "loss": 0.0026, + "step": 78020 + }, + { + "epoch": 12.12022367194781, + "grad_norm": 0.16734078526496887, + "learning_rate": 1.262705498338232e-05, + "loss": 0.0038, + "step": 78030 + }, + { + "epoch": 12.121776949363156, + "grad_norm": 0.17484529316425323, + "learning_rate": 1.2616072893916542e-05, + "loss": 0.0036, + "step": 78040 + }, + { + "epoch": 12.123330226778503, + "grad_norm": 0.2493404597043991, + "learning_rate": 1.2605094892700836e-05, + "loss": 0.0039, + "step": 78050 + }, + { + "epoch": 12.12488350419385, + "grad_norm": 0.1671963632106781, + "learning_rate": 1.259412098093572e-05, + "loss": 0.0041, + "step": 78060 + }, + { + "epoch": 12.126436781609195, + "grad_norm": 0.18962766230106354, + "learning_rate": 1.258315115982131e-05, + "loss": 0.0031, + "step": 78070 + }, + { + "epoch": 12.127990059024542, + "grad_norm": 0.9481022357940674, + "learning_rate": 1.2572185430557203e-05, + "loss": 0.0052, + "step": 78080 + }, + { + "epoch": 12.129543336439887, + "grad_norm": 0.10896912962198257, + "learning_rate": 1.2561223794342659e-05, + "loss": 0.0054, + "step": 78090 + }, + { + "epoch": 12.131096613855235, + "grad_norm": 0.2305329591035843, + "learning_rate": 1.255026625237637e-05, + "loss": 0.0045, + "step": 78100 + }, + { + "epoch": 12.132649891270582, + "grad_norm": 0.08958141505718231, + "learning_rate": 1.2539312805856678e-05, + "loss": 0.0028, + "step": 78110 + }, + { + "epoch": 12.134203168685927, + "grad_norm": 0.3263246417045593, + "learning_rate": 1.25283634559814e-05, + "loss": 0.0039, + "step": 78120 + }, + { + "epoch": 12.135756446101274, + "grad_norm": 0.09749916940927505, + "learning_rate": 1.2517418203947972e-05, + "loss": 0.0041, + "step": 78130 + }, + { + "epoch": 12.13730972351662, + "grad_norm": 0.3151092231273651, + "learning_rate": 1.2506477050953308e-05, + "loss": 0.0043, + "step": 78140 + }, + { + "epoch": 12.138863000931966, + "grad_norm": 0.1451532393693924, + "learning_rate": 1.2495539998193972e-05, + "loss": 0.0028, + "step": 78150 + }, + { + "epoch": 12.140416278347313, + "grad_norm": 0.10375065356492996, + "learning_rate": 1.2484607046865975e-05, + "loss": 0.0022, + "step": 78160 + }, + { + "epoch": 12.141969555762659, + "grad_norm": 0.3287896513938904, + "learning_rate": 1.2473678198164968e-05, + "loss": 0.0033, + "step": 78170 + }, + { + "epoch": 12.143522833178006, + "grad_norm": 1.1925761699676514, + "learning_rate": 1.2462753453286075e-05, + "loss": 0.005, + "step": 78180 + }, + { + "epoch": 12.145076110593353, + "grad_norm": 0.33533722162246704, + "learning_rate": 1.2451832813424031e-05, + "loss": 0.0053, + "step": 78190 + }, + { + "epoch": 12.146629388008698, + "grad_norm": 0.6112080812454224, + "learning_rate": 1.2440916279773095e-05, + "loss": 0.0036, + "step": 78200 + }, + { + "epoch": 12.148182665424045, + "grad_norm": 0.16411544382572174, + "learning_rate": 1.2430003853527101e-05, + "loss": 0.0029, + "step": 78210 + }, + { + "epoch": 12.14973594283939, + "grad_norm": 0.4970741271972656, + "learning_rate": 1.2419095535879383e-05, + "loss": 0.0047, + "step": 78220 + }, + { + "epoch": 12.151289220254737, + "grad_norm": 0.5054155588150024, + "learning_rate": 1.2408191328022878e-05, + "loss": 0.0048, + "step": 78230 + }, + { + "epoch": 12.152842497670084, + "grad_norm": 0.10078468918800354, + "learning_rate": 1.2397291231150066e-05, + "loss": 0.0029, + "step": 78240 + }, + { + "epoch": 12.15439577508543, + "grad_norm": 0.41438576579093933, + "learning_rate": 1.2386395246452937e-05, + "loss": 0.0039, + "step": 78250 + }, + { + "epoch": 12.155949052500777, + "grad_norm": 0.9380391836166382, + "learning_rate": 1.2375503375123083e-05, + "loss": 0.0024, + "step": 78260 + }, + { + "epoch": 12.157502329916124, + "grad_norm": 0.1445508450269699, + "learning_rate": 1.2364615618351583e-05, + "loss": 0.0034, + "step": 78270 + }, + { + "epoch": 12.159055607331469, + "grad_norm": 0.43993479013442993, + "learning_rate": 1.2353731977329169e-05, + "loss": 0.005, + "step": 78280 + }, + { + "epoch": 12.160608884746816, + "grad_norm": 3.7594282627105713, + "learning_rate": 1.2342852453246007e-05, + "loss": 0.0034, + "step": 78290 + }, + { + "epoch": 12.162162162162161, + "grad_norm": 0.544767439365387, + "learning_rate": 1.2331977047291898e-05, + "loss": 0.0057, + "step": 78300 + }, + { + "epoch": 12.163715439577508, + "grad_norm": 0.46340304613113403, + "learning_rate": 1.2321105760656137e-05, + "loss": 0.0044, + "step": 78310 + }, + { + "epoch": 12.165268716992856, + "grad_norm": 0.16296479105949402, + "learning_rate": 1.2310238594527613e-05, + "loss": 0.0049, + "step": 78320 + }, + { + "epoch": 12.1668219944082, + "grad_norm": 0.29941707849502563, + "learning_rate": 1.2299375550094693e-05, + "loss": 0.0046, + "step": 78330 + }, + { + "epoch": 12.168375271823548, + "grad_norm": 0.43808990716934204, + "learning_rate": 1.2288516628545421e-05, + "loss": 0.0041, + "step": 78340 + }, + { + "epoch": 12.169928549238895, + "grad_norm": 0.2280614972114563, + "learning_rate": 1.2277661831067255e-05, + "loss": 0.0061, + "step": 78350 + }, + { + "epoch": 12.17148182665424, + "grad_norm": 0.7215897440910339, + "learning_rate": 1.2266811158847286e-05, + "loss": 0.0033, + "step": 78360 + }, + { + "epoch": 12.173035104069587, + "grad_norm": 0.3558289408683777, + "learning_rate": 1.2255964613072107e-05, + "loss": 0.0055, + "step": 78370 + }, + { + "epoch": 12.174588381484932, + "grad_norm": 0.2490091621875763, + "learning_rate": 1.2245122194927905e-05, + "loss": 0.004, + "step": 78380 + }, + { + "epoch": 12.17614165890028, + "grad_norm": 0.569469153881073, + "learning_rate": 1.2234283905600364e-05, + "loss": 0.0042, + "step": 78390 + }, + { + "epoch": 12.177694936315627, + "grad_norm": 0.27763524651527405, + "learning_rate": 1.222344974627475e-05, + "loss": 0.0042, + "step": 78400 + }, + { + "epoch": 12.179248213730972, + "grad_norm": 0.22214628756046295, + "learning_rate": 1.2212619718135875e-05, + "loss": 0.0057, + "step": 78410 + }, + { + "epoch": 12.180801491146319, + "grad_norm": 2.6674716472625732, + "learning_rate": 1.2201793822368118e-05, + "loss": 0.0065, + "step": 78420 + }, + { + "epoch": 12.182354768561666, + "grad_norm": 0.37902700901031494, + "learning_rate": 1.2190972060155337e-05, + "loss": 0.004, + "step": 78430 + }, + { + "epoch": 12.183908045977011, + "grad_norm": 0.13427917659282684, + "learning_rate": 1.2180154432681013e-05, + "loss": 0.0041, + "step": 78440 + }, + { + "epoch": 12.185461323392358, + "grad_norm": 0.1944209188222885, + "learning_rate": 1.2169340941128148e-05, + "loss": 0.0047, + "step": 78450 + }, + { + "epoch": 12.187014600807704, + "grad_norm": 0.14587756991386414, + "learning_rate": 1.2158531586679267e-05, + "loss": 0.0036, + "step": 78460 + }, + { + "epoch": 12.18856787822305, + "grad_norm": 0.13657325506210327, + "learning_rate": 1.2147726370516477e-05, + "loss": 0.0033, + "step": 78470 + }, + { + "epoch": 12.190121155638398, + "grad_norm": 0.06811736524105072, + "learning_rate": 1.2136925293821422e-05, + "loss": 0.0042, + "step": 78480 + }, + { + "epoch": 12.191674433053743, + "grad_norm": 0.21304544806480408, + "learning_rate": 1.2126128357775306e-05, + "loss": 0.0036, + "step": 78490 + }, + { + "epoch": 12.19322771046909, + "grad_norm": 0.3418938219547272, + "learning_rate": 1.2115335563558838e-05, + "loss": 0.0044, + "step": 78500 + }, + { + "epoch": 12.194780987884435, + "grad_norm": 0.26399293541908264, + "learning_rate": 1.2104546912352327e-05, + "loss": 0.002, + "step": 78510 + }, + { + "epoch": 12.196334265299782, + "grad_norm": 0.08083785325288773, + "learning_rate": 1.2093762405335557e-05, + "loss": 0.0048, + "step": 78520 + }, + { + "epoch": 12.19788754271513, + "grad_norm": 0.10419393330812454, + "learning_rate": 1.2082982043687974e-05, + "loss": 0.0043, + "step": 78530 + }, + { + "epoch": 12.199440820130475, + "grad_norm": 0.3295285105705261, + "learning_rate": 1.207220582858845e-05, + "loss": 0.0039, + "step": 78540 + }, + { + "epoch": 12.200994097545822, + "grad_norm": 0.6079289317131042, + "learning_rate": 1.206143376121549e-05, + "loss": 0.0039, + "step": 78550 + }, + { + "epoch": 12.202547374961169, + "grad_norm": 0.378658264875412, + "learning_rate": 1.2050665842747078e-05, + "loss": 0.0034, + "step": 78560 + }, + { + "epoch": 12.204100652376514, + "grad_norm": 0.2511140704154968, + "learning_rate": 1.2039902074360804e-05, + "loss": 0.0033, + "step": 78570 + }, + { + "epoch": 12.205653929791861, + "grad_norm": 0.13704650104045868, + "learning_rate": 1.2029142457233755e-05, + "loss": 0.0041, + "step": 78580 + }, + { + "epoch": 12.207207207207206, + "grad_norm": 0.5229604244232178, + "learning_rate": 1.2018386992542601e-05, + "loss": 0.0041, + "step": 78590 + }, + { + "epoch": 12.208760484622553, + "grad_norm": 0.09368748217821121, + "learning_rate": 1.2007635681463542e-05, + "loss": 0.0029, + "step": 78600 + }, + { + "epoch": 12.2103137620379, + "grad_norm": 0.7345999479293823, + "learning_rate": 1.1996888525172334e-05, + "loss": 0.0045, + "step": 78610 + }, + { + "epoch": 12.211867039453246, + "grad_norm": 0.34581494331359863, + "learning_rate": 1.198614552484425e-05, + "loss": 0.0064, + "step": 78620 + }, + { + "epoch": 12.213420316868593, + "grad_norm": 0.14371748268604279, + "learning_rate": 1.197540668165415e-05, + "loss": 0.0035, + "step": 78630 + }, + { + "epoch": 12.21497359428394, + "grad_norm": 0.5162265300750732, + "learning_rate": 1.1964671996776395e-05, + "loss": 0.0053, + "step": 78640 + }, + { + "epoch": 12.216526871699285, + "grad_norm": 0.1764303296804428, + "learning_rate": 1.1953941471384922e-05, + "loss": 0.0036, + "step": 78650 + }, + { + "epoch": 12.218080149114632, + "grad_norm": 0.2127513885498047, + "learning_rate": 1.1943215106653205e-05, + "loss": 0.0039, + "step": 78660 + }, + { + "epoch": 12.219633426529978, + "grad_norm": 0.3410782814025879, + "learning_rate": 1.193249290375429e-05, + "loss": 0.0037, + "step": 78670 + }, + { + "epoch": 12.221186703945325, + "grad_norm": 0.12479238957166672, + "learning_rate": 1.1921774863860696e-05, + "loss": 0.0035, + "step": 78680 + }, + { + "epoch": 12.222739981360672, + "grad_norm": 0.33516597747802734, + "learning_rate": 1.1911060988144556e-05, + "loss": 0.0039, + "step": 78690 + }, + { + "epoch": 12.224293258776017, + "grad_norm": 0.12185104191303253, + "learning_rate": 1.1900351277777533e-05, + "loss": 0.0054, + "step": 78700 + }, + { + "epoch": 12.225846536191364, + "grad_norm": 1.2568539381027222, + "learning_rate": 1.1889645733930798e-05, + "loss": 0.0056, + "step": 78710 + }, + { + "epoch": 12.227399813606711, + "grad_norm": 1.612746238708496, + "learning_rate": 1.1878944357775107e-05, + "loss": 0.0051, + "step": 78720 + }, + { + "epoch": 12.228953091022056, + "grad_norm": 0.4671415090560913, + "learning_rate": 1.1868247150480749e-05, + "loss": 0.0038, + "step": 78730 + }, + { + "epoch": 12.230506368437403, + "grad_norm": 0.17981234192848206, + "learning_rate": 1.1857554113217567e-05, + "loss": 0.0043, + "step": 78740 + }, + { + "epoch": 12.232059645852749, + "grad_norm": 0.20569391548633575, + "learning_rate": 1.1846865247154903e-05, + "loss": 0.0043, + "step": 78750 + }, + { + "epoch": 12.233612923268096, + "grad_norm": 0.18210692703723907, + "learning_rate": 1.1836180553461711e-05, + "loss": 0.0049, + "step": 78760 + }, + { + "epoch": 12.235166200683443, + "grad_norm": 0.20364651083946228, + "learning_rate": 1.1825500033306419e-05, + "loss": 0.0026, + "step": 78770 + }, + { + "epoch": 12.236719478098788, + "grad_norm": 0.18137578666210175, + "learning_rate": 1.1814823687857052e-05, + "loss": 0.0038, + "step": 78780 + }, + { + "epoch": 12.238272755514135, + "grad_norm": 0.3038828372955322, + "learning_rate": 1.1804151518281153e-05, + "loss": 0.0036, + "step": 78790 + }, + { + "epoch": 12.23982603292948, + "grad_norm": 0.25246158242225647, + "learning_rate": 1.179348352574583e-05, + "loss": 0.0042, + "step": 78800 + }, + { + "epoch": 12.241379310344827, + "grad_norm": 0.07785843312740326, + "learning_rate": 1.1782819711417697e-05, + "loss": 0.0024, + "step": 78810 + }, + { + "epoch": 12.242932587760174, + "grad_norm": 0.29932764172554016, + "learning_rate": 1.1772160076462951e-05, + "loss": 0.0042, + "step": 78820 + }, + { + "epoch": 12.24448586517552, + "grad_norm": 0.8854097723960876, + "learning_rate": 1.1761504622047293e-05, + "loss": 0.0046, + "step": 78830 + }, + { + "epoch": 12.246039142590867, + "grad_norm": 0.2305775284767151, + "learning_rate": 1.1750853349335999e-05, + "loss": 0.0036, + "step": 78840 + }, + { + "epoch": 12.247592420006214, + "grad_norm": 0.10806585848331451, + "learning_rate": 1.1740206259493874e-05, + "loss": 0.0034, + "step": 78850 + }, + { + "epoch": 12.24914569742156, + "grad_norm": 0.16914096474647522, + "learning_rate": 1.1729563353685286e-05, + "loss": 0.0048, + "step": 78860 + }, + { + "epoch": 12.250698974836906, + "grad_norm": 0.2945624887943268, + "learning_rate": 1.1718924633074091e-05, + "loss": 0.0045, + "step": 78870 + }, + { + "epoch": 12.252252252252251, + "grad_norm": 0.1935996413230896, + "learning_rate": 1.1708290098823765e-05, + "loss": 0.0032, + "step": 78880 + }, + { + "epoch": 12.253805529667599, + "grad_norm": 0.23934762179851532, + "learning_rate": 1.1697659752097245e-05, + "loss": 0.0065, + "step": 78890 + }, + { + "epoch": 12.255358807082946, + "grad_norm": 0.1630963832139969, + "learning_rate": 1.1687033594057068e-05, + "loss": 0.0042, + "step": 78900 + }, + { + "epoch": 12.25691208449829, + "grad_norm": 0.1984458565711975, + "learning_rate": 1.1676411625865297e-05, + "loss": 0.0043, + "step": 78910 + }, + { + "epoch": 12.258465361913638, + "grad_norm": 0.054070644080638885, + "learning_rate": 1.1665793848683542e-05, + "loss": 0.0043, + "step": 78920 + }, + { + "epoch": 12.260018639328985, + "grad_norm": 0.15010108053684235, + "learning_rate": 1.1655180263672927e-05, + "loss": 0.0043, + "step": 78930 + }, + { + "epoch": 12.26157191674433, + "grad_norm": 0.13768000900745392, + "learning_rate": 1.1644570871994142e-05, + "loss": 0.0041, + "step": 78940 + }, + { + "epoch": 12.263125194159677, + "grad_norm": 0.17079894244670868, + "learning_rate": 1.163396567480744e-05, + "loss": 0.0035, + "step": 78950 + }, + { + "epoch": 12.264678471575023, + "grad_norm": 0.18334375321865082, + "learning_rate": 1.1623364673272552e-05, + "loss": 0.0045, + "step": 78960 + }, + { + "epoch": 12.26623174899037, + "grad_norm": 0.15374967455863953, + "learning_rate": 1.1612767868548807e-05, + "loss": 0.0065, + "step": 78970 + }, + { + "epoch": 12.267785026405717, + "grad_norm": 0.08750604093074799, + "learning_rate": 1.1602175261795057e-05, + "loss": 0.0039, + "step": 78980 + }, + { + "epoch": 12.269338303821062, + "grad_norm": 0.5052042603492737, + "learning_rate": 1.1591586854169696e-05, + "loss": 0.0028, + "step": 78990 + }, + { + "epoch": 12.270891581236409, + "grad_norm": 0.29076752066612244, + "learning_rate": 1.158100264683064e-05, + "loss": 0.0042, + "step": 79000 + }, + { + "epoch": 12.272444858651756, + "grad_norm": 0.1293010413646698, + "learning_rate": 1.1570422640935386e-05, + "loss": 0.0033, + "step": 79010 + }, + { + "epoch": 12.273998136067101, + "grad_norm": 0.24824471771717072, + "learning_rate": 1.155984683764092e-05, + "loss": 0.0043, + "step": 79020 + }, + { + "epoch": 12.275551413482448, + "grad_norm": 0.27812460064888, + "learning_rate": 1.1549275238103812e-05, + "loss": 0.0031, + "step": 79030 + }, + { + "epoch": 12.277104690897794, + "grad_norm": 0.11151633411645889, + "learning_rate": 1.1538707843480152e-05, + "loss": 0.0033, + "step": 79040 + }, + { + "epoch": 12.27865796831314, + "grad_norm": 0.33399006724357605, + "learning_rate": 1.1528144654925593e-05, + "loss": 0.0036, + "step": 79050 + }, + { + "epoch": 12.280211245728488, + "grad_norm": 0.5123697519302368, + "learning_rate": 1.1517585673595277e-05, + "loss": 0.0038, + "step": 79060 + }, + { + "epoch": 12.281764523143833, + "grad_norm": 0.23763863742351532, + "learning_rate": 1.150703090064395e-05, + "loss": 0.0067, + "step": 79070 + }, + { + "epoch": 12.28331780055918, + "grad_norm": 0.4083828628063202, + "learning_rate": 1.1496480337225835e-05, + "loss": 0.004, + "step": 79080 + }, + { + "epoch": 12.284871077974525, + "grad_norm": 0.3059656620025635, + "learning_rate": 1.1485933984494751e-05, + "loss": 0.0032, + "step": 79090 + }, + { + "epoch": 12.286424355389872, + "grad_norm": 0.9014805555343628, + "learning_rate": 1.1475391843604e-05, + "loss": 0.0043, + "step": 79100 + }, + { + "epoch": 12.28797763280522, + "grad_norm": 0.32963940501213074, + "learning_rate": 1.1464853915706503e-05, + "loss": 0.0047, + "step": 79110 + }, + { + "epoch": 12.289530910220565, + "grad_norm": 0.5835508704185486, + "learning_rate": 1.1454320201954626e-05, + "loss": 0.0046, + "step": 79120 + }, + { + "epoch": 12.291084187635912, + "grad_norm": 0.20761841535568237, + "learning_rate": 1.1443790703500356e-05, + "loss": 0.0029, + "step": 79130 + }, + { + "epoch": 12.292637465051259, + "grad_norm": 0.2624588906764984, + "learning_rate": 1.1433265421495154e-05, + "loss": 0.008, + "step": 79140 + }, + { + "epoch": 12.294190742466604, + "grad_norm": 0.2722976505756378, + "learning_rate": 1.1422744357090065e-05, + "loss": 0.0042, + "step": 79150 + }, + { + "epoch": 12.295744019881951, + "grad_norm": 0.2815559506416321, + "learning_rate": 1.1412227511435647e-05, + "loss": 0.004, + "step": 79160 + }, + { + "epoch": 12.297297297297296, + "grad_norm": 0.17027828097343445, + "learning_rate": 1.1401714885682025e-05, + "loss": 0.0037, + "step": 79170 + }, + { + "epoch": 12.298850574712644, + "grad_norm": 0.09916683286428452, + "learning_rate": 1.1391206480978823e-05, + "loss": 0.003, + "step": 79180 + }, + { + "epoch": 12.30040385212799, + "grad_norm": 0.2813432812690735, + "learning_rate": 1.1380702298475227e-05, + "loss": 0.0033, + "step": 79190 + }, + { + "epoch": 12.301957129543336, + "grad_norm": 0.21600034832954407, + "learning_rate": 1.1370202339319985e-05, + "loss": 0.0045, + "step": 79200 + }, + { + "epoch": 12.303510406958683, + "grad_norm": 0.4156380891799927, + "learning_rate": 1.1359706604661319e-05, + "loss": 0.0081, + "step": 79210 + }, + { + "epoch": 12.30506368437403, + "grad_norm": 0.3126453459262848, + "learning_rate": 1.1349215095647042e-05, + "loss": 0.0041, + "step": 79220 + }, + { + "epoch": 12.306616961789375, + "grad_norm": 0.37335526943206787, + "learning_rate": 1.1338727813424493e-05, + "loss": 0.0041, + "step": 79230 + }, + { + "epoch": 12.308170239204722, + "grad_norm": 0.28682276606559753, + "learning_rate": 1.132824475914056e-05, + "loss": 0.0065, + "step": 79240 + }, + { + "epoch": 12.309723516620068, + "grad_norm": 0.11741720885038376, + "learning_rate": 1.131776593394162e-05, + "loss": 0.0035, + "step": 79250 + }, + { + "epoch": 12.311276794035415, + "grad_norm": 0.3440779447555542, + "learning_rate": 1.1307291338973657e-05, + "loss": 0.004, + "step": 79260 + }, + { + "epoch": 12.312830071450762, + "grad_norm": 0.34480467438697815, + "learning_rate": 1.1296820975382121e-05, + "loss": 0.0052, + "step": 79270 + }, + { + "epoch": 12.314383348866107, + "grad_norm": 0.3920366168022156, + "learning_rate": 1.1286354844312074e-05, + "loss": 0.006, + "step": 79280 + }, + { + "epoch": 12.315936626281454, + "grad_norm": 0.15882955491542816, + "learning_rate": 1.1275892946908023e-05, + "loss": 0.0044, + "step": 79290 + }, + { + "epoch": 12.317489903696801, + "grad_norm": 1.4192404747009277, + "learning_rate": 1.126543528431413e-05, + "loss": 0.0058, + "step": 79300 + }, + { + "epoch": 12.319043181112146, + "grad_norm": 0.10156641155481339, + "learning_rate": 1.125498185767398e-05, + "loss": 0.0054, + "step": 79310 + }, + { + "epoch": 12.320596458527493, + "grad_norm": 0.12252563238143921, + "learning_rate": 1.1244532668130781e-05, + "loss": 0.0041, + "step": 79320 + }, + { + "epoch": 12.322149735942839, + "grad_norm": 0.27931734919548035, + "learning_rate": 1.1234087716827203e-05, + "loss": 0.004, + "step": 79330 + }, + { + "epoch": 12.323703013358186, + "grad_norm": 0.3414384722709656, + "learning_rate": 1.1223647004905524e-05, + "loss": 0.0047, + "step": 79340 + }, + { + "epoch": 12.325256290773533, + "grad_norm": 0.27032414078712463, + "learning_rate": 1.1213210533507485e-05, + "loss": 0.0029, + "step": 79350 + }, + { + "epoch": 12.326809568188878, + "grad_norm": 0.11407127231359482, + "learning_rate": 1.1202778303774447e-05, + "loss": 0.0024, + "step": 79360 + }, + { + "epoch": 12.328362845604225, + "grad_norm": 1.3398696184158325, + "learning_rate": 1.1192350316847234e-05, + "loss": 0.0064, + "step": 79370 + }, + { + "epoch": 12.329916123019572, + "grad_norm": 0.1400047391653061, + "learning_rate": 1.1181926573866258e-05, + "loss": 0.0044, + "step": 79380 + }, + { + "epoch": 12.331469400434917, + "grad_norm": 0.3658885657787323, + "learning_rate": 1.1171507075971416e-05, + "loss": 0.0068, + "step": 79390 + }, + { + "epoch": 12.333022677850265, + "grad_norm": 0.4557456374168396, + "learning_rate": 1.1161091824302177e-05, + "loss": 0.0039, + "step": 79400 + }, + { + "epoch": 12.33457595526561, + "grad_norm": 0.18775241076946259, + "learning_rate": 1.115068081999755e-05, + "loss": 0.0033, + "step": 79410 + }, + { + "epoch": 12.336129232680957, + "grad_norm": 0.15419647097587585, + "learning_rate": 1.1140274064196071e-05, + "loss": 0.0038, + "step": 79420 + }, + { + "epoch": 12.337682510096304, + "grad_norm": 0.3909624516963959, + "learning_rate": 1.112987155803578e-05, + "loss": 0.0041, + "step": 79430 + }, + { + "epoch": 12.33923578751165, + "grad_norm": 0.562582790851593, + "learning_rate": 1.1119473302654298e-05, + "loss": 0.0035, + "step": 79440 + }, + { + "epoch": 12.340789064926996, + "grad_norm": 0.5799680948257446, + "learning_rate": 1.1109079299188769e-05, + "loss": 0.0045, + "step": 79450 + }, + { + "epoch": 12.342342342342342, + "grad_norm": 0.1903868168592453, + "learning_rate": 1.1098689548775847e-05, + "loss": 0.003, + "step": 79460 + }, + { + "epoch": 12.343895619757689, + "grad_norm": 0.17065396904945374, + "learning_rate": 1.1088304052551762e-05, + "loss": 0.0033, + "step": 79470 + }, + { + "epoch": 12.345448897173036, + "grad_norm": 0.79046231508255, + "learning_rate": 1.1077922811652215e-05, + "loss": 0.0047, + "step": 79480 + }, + { + "epoch": 12.347002174588381, + "grad_norm": 0.16020414233207703, + "learning_rate": 1.1067545827212534e-05, + "loss": 0.0036, + "step": 79490 + }, + { + "epoch": 12.348555452003728, + "grad_norm": 0.1127830371260643, + "learning_rate": 1.1057173100367496e-05, + "loss": 0.0034, + "step": 79500 + }, + { + "epoch": 12.350108729419075, + "grad_norm": 0.28807175159454346, + "learning_rate": 1.1046804632251474e-05, + "loss": 0.0027, + "step": 79510 + }, + { + "epoch": 12.35166200683442, + "grad_norm": 0.21550726890563965, + "learning_rate": 1.1036440423998318e-05, + "loss": 0.0046, + "step": 79520 + }, + { + "epoch": 12.353215284249767, + "grad_norm": 0.5161415934562683, + "learning_rate": 1.1026080476741469e-05, + "loss": 0.0048, + "step": 79530 + }, + { + "epoch": 12.354768561665113, + "grad_norm": 0.1503225564956665, + "learning_rate": 1.1015724791613836e-05, + "loss": 0.0048, + "step": 79540 + }, + { + "epoch": 12.35632183908046, + "grad_norm": 0.1668093204498291, + "learning_rate": 1.1005373369747951e-05, + "loss": 0.0033, + "step": 79550 + }, + { + "epoch": 12.357875116495807, + "grad_norm": 0.7498464584350586, + "learning_rate": 1.09950262122758e-05, + "loss": 0.0053, + "step": 79560 + }, + { + "epoch": 12.359428393911152, + "grad_norm": 0.19230768084526062, + "learning_rate": 1.0984683320328948e-05, + "loss": 0.0042, + "step": 79570 + }, + { + "epoch": 12.360981671326499, + "grad_norm": 0.14803723990917206, + "learning_rate": 1.0974344695038458e-05, + "loss": 0.0036, + "step": 79580 + }, + { + "epoch": 12.362534948741846, + "grad_norm": 0.28626981377601624, + "learning_rate": 1.0964010337534968e-05, + "loss": 0.0048, + "step": 79590 + }, + { + "epoch": 12.364088226157191, + "grad_norm": 0.08798684179782867, + "learning_rate": 1.0953680248948611e-05, + "loss": 0.0041, + "step": 79600 + }, + { + "epoch": 12.365641503572538, + "grad_norm": 0.27176302671432495, + "learning_rate": 1.0943354430409069e-05, + "loss": 0.0037, + "step": 79610 + }, + { + "epoch": 12.367194780987884, + "grad_norm": 0.09420944005250931, + "learning_rate": 1.0933032883045574e-05, + "loss": 0.0036, + "step": 79620 + }, + { + "epoch": 12.36874805840323, + "grad_norm": 0.3272762596607208, + "learning_rate": 1.0922715607986872e-05, + "loss": 0.0053, + "step": 79630 + }, + { + "epoch": 12.370301335818578, + "grad_norm": 0.15078094601631165, + "learning_rate": 1.0912402606361234e-05, + "loss": 0.0056, + "step": 79640 + }, + { + "epoch": 12.371854613233923, + "grad_norm": 0.18852677941322327, + "learning_rate": 1.090209387929647e-05, + "loss": 0.0042, + "step": 79650 + }, + { + "epoch": 12.37340789064927, + "grad_norm": 0.2440940886735916, + "learning_rate": 1.0891789427919957e-05, + "loss": 0.0036, + "step": 79660 + }, + { + "epoch": 12.374961168064615, + "grad_norm": 0.2330942004919052, + "learning_rate": 1.088148925335854e-05, + "loss": 0.0039, + "step": 79670 + }, + { + "epoch": 12.376514445479962, + "grad_norm": 0.13217197358608246, + "learning_rate": 1.087119335673864e-05, + "loss": 0.002, + "step": 79680 + }, + { + "epoch": 12.37806772289531, + "grad_norm": 0.1422892063856125, + "learning_rate": 1.0860901739186208e-05, + "loss": 0.0035, + "step": 79690 + }, + { + "epoch": 12.379621000310655, + "grad_norm": 0.24951499700546265, + "learning_rate": 1.085061440182673e-05, + "loss": 0.0047, + "step": 79700 + }, + { + "epoch": 12.381174277726002, + "grad_norm": 1.4916070699691772, + "learning_rate": 1.0840331345785193e-05, + "loss": 0.0044, + "step": 79710 + }, + { + "epoch": 12.382727555141349, + "grad_norm": 0.1739381104707718, + "learning_rate": 1.083005257218615e-05, + "loss": 0.0033, + "step": 79720 + }, + { + "epoch": 12.384280832556694, + "grad_norm": 0.09785740822553635, + "learning_rate": 1.0819778082153642e-05, + "loss": 0.004, + "step": 79730 + }, + { + "epoch": 12.385834109972041, + "grad_norm": 0.6749999523162842, + "learning_rate": 1.0809507876811325e-05, + "loss": 0.0031, + "step": 79740 + }, + { + "epoch": 12.387387387387387, + "grad_norm": 0.2201736867427826, + "learning_rate": 1.0799241957282292e-05, + "loss": 0.0046, + "step": 79750 + }, + { + "epoch": 12.388940664802734, + "grad_norm": 0.09742862731218338, + "learning_rate": 1.078898032468923e-05, + "loss": 0.0039, + "step": 79760 + }, + { + "epoch": 12.39049394221808, + "grad_norm": 0.2138320654630661, + "learning_rate": 1.0778722980154315e-05, + "loss": 0.0039, + "step": 79770 + }, + { + "epoch": 12.392047219633426, + "grad_norm": 0.06832782179117203, + "learning_rate": 1.0768469924799302e-05, + "loss": 0.0031, + "step": 79780 + }, + { + "epoch": 12.393600497048773, + "grad_norm": 0.2110033482313156, + "learning_rate": 1.0758221159745419e-05, + "loss": 0.0037, + "step": 79790 + }, + { + "epoch": 12.39515377446412, + "grad_norm": 0.16761961579322815, + "learning_rate": 1.074797668611347e-05, + "loss": 0.0042, + "step": 79800 + }, + { + "epoch": 12.396707051879465, + "grad_norm": 0.16050386428833008, + "learning_rate": 1.0737736505023777e-05, + "loss": 0.0042, + "step": 79810 + }, + { + "epoch": 12.398260329294812, + "grad_norm": 0.3170331120491028, + "learning_rate": 1.0727500617596203e-05, + "loss": 0.0054, + "step": 79820 + }, + { + "epoch": 12.399813606710158, + "grad_norm": 1.0585968494415283, + "learning_rate": 1.0717269024950104e-05, + "loss": 0.0053, + "step": 79830 + }, + { + "epoch": 12.401366884125505, + "grad_norm": 0.16701184213161469, + "learning_rate": 1.070704172820442e-05, + "loss": 0.0031, + "step": 79840 + }, + { + "epoch": 12.402920161540852, + "grad_norm": 0.23623734712600708, + "learning_rate": 1.0696818728477559e-05, + "loss": 0.0039, + "step": 79850 + }, + { + "epoch": 12.404473438956197, + "grad_norm": 0.0929407849907875, + "learning_rate": 1.0686600026887511e-05, + "loss": 0.0043, + "step": 79860 + }, + { + "epoch": 12.406026716371544, + "grad_norm": 1.3641343116760254, + "learning_rate": 1.0676385624551777e-05, + "loss": 0.0047, + "step": 79870 + }, + { + "epoch": 12.407579993786891, + "grad_norm": 0.48621639609336853, + "learning_rate": 1.0666175522587402e-05, + "loss": 0.0045, + "step": 79880 + }, + { + "epoch": 12.409133271202236, + "grad_norm": 0.24501267075538635, + "learning_rate": 1.0655969722110925e-05, + "loss": 0.005, + "step": 79890 + }, + { + "epoch": 12.410686548617583, + "grad_norm": 0.22941908240318298, + "learning_rate": 1.0645768224238439e-05, + "loss": 0.0046, + "step": 79900 + }, + { + "epoch": 12.412239826032929, + "grad_norm": 0.15804466605186462, + "learning_rate": 1.063557103008559e-05, + "loss": 0.0024, + "step": 79910 + }, + { + "epoch": 12.413793103448276, + "grad_norm": 0.3708431124687195, + "learning_rate": 1.0625378140767494e-05, + "loss": 0.0041, + "step": 79920 + }, + { + "epoch": 12.415346380863623, + "grad_norm": 0.36681699752807617, + "learning_rate": 1.0615189557398846e-05, + "loss": 0.0052, + "step": 79930 + }, + { + "epoch": 12.416899658278968, + "grad_norm": 0.26231226325035095, + "learning_rate": 1.060500528109386e-05, + "loss": 0.0035, + "step": 79940 + }, + { + "epoch": 12.418452935694315, + "grad_norm": 0.19140270352363586, + "learning_rate": 1.0594825312966272e-05, + "loss": 0.0061, + "step": 79950 + }, + { + "epoch": 12.420006213109662, + "grad_norm": 0.3749616742134094, + "learning_rate": 1.0584649654129336e-05, + "loss": 0.005, + "step": 79960 + }, + { + "epoch": 12.421559490525008, + "grad_norm": 0.9667952060699463, + "learning_rate": 1.0574478305695863e-05, + "loss": 0.0031, + "step": 79970 + }, + { + "epoch": 12.423112767940355, + "grad_norm": 0.1778305172920227, + "learning_rate": 1.0564311268778154e-05, + "loss": 0.005, + "step": 79980 + }, + { + "epoch": 12.4246660453557, + "grad_norm": 0.2128196805715561, + "learning_rate": 1.0554148544488073e-05, + "loss": 0.0041, + "step": 79990 + }, + { + "epoch": 12.426219322771047, + "grad_norm": 0.6734607815742493, + "learning_rate": 1.0543990133937004e-05, + "loss": 0.0045, + "step": 80000 + }, + { + "epoch": 12.427772600186394, + "grad_norm": 0.16819126904010773, + "learning_rate": 1.0533836038235868e-05, + "loss": 0.0061, + "step": 80010 + }, + { + "epoch": 12.42932587760174, + "grad_norm": 0.12744417786598206, + "learning_rate": 1.052368625849507e-05, + "loss": 0.0048, + "step": 80020 + }, + { + "epoch": 12.430879155017086, + "grad_norm": 0.1797970086336136, + "learning_rate": 1.05135407958246e-05, + "loss": 0.0038, + "step": 80030 + }, + { + "epoch": 12.432432432432432, + "grad_norm": 0.1499725878238678, + "learning_rate": 1.0503399651333934e-05, + "loss": 0.0039, + "step": 80040 + }, + { + "epoch": 12.433985709847779, + "grad_norm": 1.0291261672973633, + "learning_rate": 1.04932628261321e-05, + "loss": 0.0044, + "step": 80050 + }, + { + "epoch": 12.435538987263126, + "grad_norm": 0.48554369807243347, + "learning_rate": 1.0483130321327644e-05, + "loss": 0.0045, + "step": 80060 + }, + { + "epoch": 12.437092264678471, + "grad_norm": 0.19484476745128632, + "learning_rate": 1.0473002138028654e-05, + "loss": 0.0048, + "step": 80070 + }, + { + "epoch": 12.438645542093818, + "grad_norm": 0.4275241494178772, + "learning_rate": 1.0462878277342713e-05, + "loss": 0.0057, + "step": 80080 + }, + { + "epoch": 12.440198819509165, + "grad_norm": 0.33713939785957336, + "learning_rate": 1.0452758740376972e-05, + "loss": 0.0042, + "step": 80090 + }, + { + "epoch": 12.44175209692451, + "grad_norm": 0.17741887271404266, + "learning_rate": 1.0442643528238067e-05, + "loss": 0.0036, + "step": 80100 + }, + { + "epoch": 12.443305374339857, + "grad_norm": 0.17576710879802704, + "learning_rate": 1.0432532642032188e-05, + "loss": 0.0044, + "step": 80110 + }, + { + "epoch": 12.444858651755203, + "grad_norm": 0.1766822189092636, + "learning_rate": 1.042242608286505e-05, + "loss": 0.0033, + "step": 80120 + }, + { + "epoch": 12.44641192917055, + "grad_norm": 0.1474056839942932, + "learning_rate": 1.041232385184191e-05, + "loss": 0.0031, + "step": 80130 + }, + { + "epoch": 12.447965206585897, + "grad_norm": 0.22057217359542847, + "learning_rate": 1.0402225950067495e-05, + "loss": 0.0028, + "step": 80140 + }, + { + "epoch": 12.449518484001242, + "grad_norm": 0.19510585069656372, + "learning_rate": 1.0392132378646119e-05, + "loss": 0.0029, + "step": 80150 + }, + { + "epoch": 12.45107176141659, + "grad_norm": 0.22813084721565247, + "learning_rate": 1.0382043138681607e-05, + "loss": 0.0055, + "step": 80160 + }, + { + "epoch": 12.452625038831936, + "grad_norm": 0.23139621317386627, + "learning_rate": 1.0371958231277284e-05, + "loss": 0.0051, + "step": 80170 + }, + { + "epoch": 12.454178316247281, + "grad_norm": 0.0933229848742485, + "learning_rate": 1.0361877657536024e-05, + "loss": 0.0031, + "step": 80180 + }, + { + "epoch": 12.455731593662628, + "grad_norm": 0.2239259034395218, + "learning_rate": 1.035180141856023e-05, + "loss": 0.0035, + "step": 80190 + }, + { + "epoch": 12.457284871077974, + "grad_norm": 0.16343817114830017, + "learning_rate": 1.0341729515451836e-05, + "loss": 0.0051, + "step": 80200 + }, + { + "epoch": 12.45883814849332, + "grad_norm": 0.16189774870872498, + "learning_rate": 1.0331661949312265e-05, + "loss": 0.0048, + "step": 80210 + }, + { + "epoch": 12.460391425908668, + "grad_norm": 0.10367555916309357, + "learning_rate": 1.0321598721242514e-05, + "loss": 0.0034, + "step": 80220 + }, + { + "epoch": 12.461944703324013, + "grad_norm": 0.30899521708488464, + "learning_rate": 1.0311539832343064e-05, + "loss": 0.0046, + "step": 80230 + }, + { + "epoch": 12.46349798073936, + "grad_norm": 0.22596609592437744, + "learning_rate": 1.0301485283713942e-05, + "loss": 0.0037, + "step": 80240 + }, + { + "epoch": 12.465051258154707, + "grad_norm": 0.11009097099304199, + "learning_rate": 1.0291435076454703e-05, + "loss": 0.0055, + "step": 80250 + }, + { + "epoch": 12.466604535570053, + "grad_norm": 0.18918876349925995, + "learning_rate": 1.028138921166444e-05, + "loss": 0.0034, + "step": 80260 + }, + { + "epoch": 12.4681578129854, + "grad_norm": 0.24155712127685547, + "learning_rate": 1.0271347690441719e-05, + "loss": 0.0034, + "step": 80270 + }, + { + "epoch": 12.469711090400745, + "grad_norm": 0.12538443505764008, + "learning_rate": 1.0261310513884698e-05, + "loss": 0.0026, + "step": 80280 + }, + { + "epoch": 12.471264367816092, + "grad_norm": 0.29843851923942566, + "learning_rate": 1.0251277683090998e-05, + "loss": 0.0044, + "step": 80290 + }, + { + "epoch": 12.472817645231439, + "grad_norm": 0.20516182482242584, + "learning_rate": 1.0241249199157826e-05, + "loss": 0.0046, + "step": 80300 + }, + { + "epoch": 12.474370922646784, + "grad_norm": 0.15631398558616638, + "learning_rate": 1.0231225063181837e-05, + "loss": 0.0049, + "step": 80310 + }, + { + "epoch": 12.475924200062131, + "grad_norm": 0.16399741172790527, + "learning_rate": 1.0221205276259316e-05, + "loss": 0.0041, + "step": 80320 + }, + { + "epoch": 12.477477477477478, + "grad_norm": 0.1669779121875763, + "learning_rate": 1.0211189839485958e-05, + "loss": 0.0051, + "step": 80330 + }, + { + "epoch": 12.479030754892824, + "grad_norm": 0.29420357942581177, + "learning_rate": 1.0201178753957075e-05, + "loss": 0.0042, + "step": 80340 + }, + { + "epoch": 12.48058403230817, + "grad_norm": 0.103866346180439, + "learning_rate": 1.0191172020767432e-05, + "loss": 0.0031, + "step": 80350 + }, + { + "epoch": 12.482137309723516, + "grad_norm": 0.23001334071159363, + "learning_rate": 1.0181169641011362e-05, + "loss": 0.0046, + "step": 80360 + }, + { + "epoch": 12.483690587138863, + "grad_norm": 0.25927868485450745, + "learning_rate": 1.0171171615782721e-05, + "loss": 0.0053, + "step": 80370 + }, + { + "epoch": 12.48524386455421, + "grad_norm": 0.12981988489627838, + "learning_rate": 1.0161177946174866e-05, + "loss": 0.0057, + "step": 80380 + }, + { + "epoch": 12.486797141969555, + "grad_norm": 0.24648067355155945, + "learning_rate": 1.0151188633280706e-05, + "loss": 0.0049, + "step": 80390 + }, + { + "epoch": 12.488350419384902, + "grad_norm": 0.39955222606658936, + "learning_rate": 1.0141203678192635e-05, + "loss": 0.0037, + "step": 80400 + }, + { + "epoch": 12.489903696800248, + "grad_norm": 0.13287420570850372, + "learning_rate": 1.0131223082002612e-05, + "loss": 0.0021, + "step": 80410 + }, + { + "epoch": 12.491456974215595, + "grad_norm": 0.12279435992240906, + "learning_rate": 1.012124684580208e-05, + "loss": 0.0036, + "step": 80420 + }, + { + "epoch": 12.493010251630942, + "grad_norm": 0.11773012578487396, + "learning_rate": 1.0111274970682039e-05, + "loss": 0.0038, + "step": 80430 + }, + { + "epoch": 12.494563529046287, + "grad_norm": 0.22958284616470337, + "learning_rate": 1.010130745773299e-05, + "loss": 0.0049, + "step": 80440 + }, + { + "epoch": 12.496116806461634, + "grad_norm": 0.17428602278232574, + "learning_rate": 1.0091344308044986e-05, + "loss": 0.0029, + "step": 80450 + }, + { + "epoch": 12.497670083876981, + "grad_norm": 0.052539125084877014, + "learning_rate": 1.0081385522707554e-05, + "loss": 0.0037, + "step": 80460 + }, + { + "epoch": 12.499223361292326, + "grad_norm": 0.2100350707769394, + "learning_rate": 1.0071431102809803e-05, + "loss": 0.0039, + "step": 80470 + }, + { + "epoch": 12.500776638707674, + "grad_norm": 0.18955965340137482, + "learning_rate": 1.0061481049440297e-05, + "loss": 0.0052, + "step": 80480 + }, + { + "epoch": 12.502329916123019, + "grad_norm": 0.2765100598335266, + "learning_rate": 1.0051535363687187e-05, + "loss": 0.0038, + "step": 80490 + }, + { + "epoch": 12.503883193538366, + "grad_norm": 0.22831302881240845, + "learning_rate": 1.0041594046638087e-05, + "loss": 0.0049, + "step": 80500 + }, + { + "epoch": 12.505436470953713, + "grad_norm": 0.09662459045648575, + "learning_rate": 1.0031657099380204e-05, + "loss": 0.0043, + "step": 80510 + }, + { + "epoch": 12.506989748369058, + "grad_norm": 0.5143348574638367, + "learning_rate": 1.0021724523000204e-05, + "loss": 0.0045, + "step": 80520 + }, + { + "epoch": 12.508543025784405, + "grad_norm": 0.11005505174398422, + "learning_rate": 1.001179631858431e-05, + "loss": 0.003, + "step": 80530 + }, + { + "epoch": 12.510096303199752, + "grad_norm": 0.3201066255569458, + "learning_rate": 1.0001872487218245e-05, + "loss": 0.0024, + "step": 80540 + }, + { + "epoch": 12.511649580615098, + "grad_norm": 0.4066184163093567, + "learning_rate": 9.991953029987272e-06, + "loss": 0.0029, + "step": 80550 + }, + { + "epoch": 12.513202858030445, + "grad_norm": 0.15240159630775452, + "learning_rate": 9.982037947976147e-06, + "loss": 0.0048, + "step": 80560 + }, + { + "epoch": 12.51475613544579, + "grad_norm": 0.189146488904953, + "learning_rate": 9.972127242269208e-06, + "loss": 0.0027, + "step": 80570 + }, + { + "epoch": 12.516309412861137, + "grad_norm": 0.30953457951545715, + "learning_rate": 9.962220913950238e-06, + "loss": 0.0032, + "step": 80580 + }, + { + "epoch": 12.517862690276484, + "grad_norm": 1.376413106918335, + "learning_rate": 9.952318964102591e-06, + "loss": 0.0046, + "step": 80590 + }, + { + "epoch": 12.51941596769183, + "grad_norm": 0.060818299651145935, + "learning_rate": 9.942421393809148e-06, + "loss": 0.0054, + "step": 80600 + }, + { + "epoch": 12.520969245107176, + "grad_norm": 0.26405826210975647, + "learning_rate": 9.932528204152264e-06, + "loss": 0.0043, + "step": 80610 + }, + { + "epoch": 12.522522522522522, + "grad_norm": 0.2826443314552307, + "learning_rate": 9.922639396213856e-06, + "loss": 0.0052, + "step": 80620 + }, + { + "epoch": 12.524075799937869, + "grad_norm": 0.24404245615005493, + "learning_rate": 9.912754971075344e-06, + "loss": 0.0042, + "step": 80630 + }, + { + "epoch": 12.525629077353216, + "grad_norm": 0.23170463740825653, + "learning_rate": 9.902874929817696e-06, + "loss": 0.0037, + "step": 80640 + }, + { + "epoch": 12.527182354768561, + "grad_norm": 0.10151856392621994, + "learning_rate": 9.892999273521342e-06, + "loss": 0.0037, + "step": 80650 + }, + { + "epoch": 12.528735632183908, + "grad_norm": 0.19725744426250458, + "learning_rate": 9.883128003266302e-06, + "loss": 0.0037, + "step": 80660 + }, + { + "epoch": 12.530288909599255, + "grad_norm": 0.11943849921226501, + "learning_rate": 9.87326112013206e-06, + "loss": 0.0038, + "step": 80670 + }, + { + "epoch": 12.5318421870146, + "grad_norm": 0.27894899249076843, + "learning_rate": 9.863398625197662e-06, + "loss": 0.0036, + "step": 80680 + }, + { + "epoch": 12.533395464429947, + "grad_norm": 0.36788278818130493, + "learning_rate": 9.853540519541615e-06, + "loss": 0.0039, + "step": 80690 + }, + { + "epoch": 12.534948741845295, + "grad_norm": 0.10588627308607101, + "learning_rate": 9.843686804242053e-06, + "loss": 0.004, + "step": 80700 + }, + { + "epoch": 12.53650201926064, + "grad_norm": 0.4532286524772644, + "learning_rate": 9.833837480376506e-06, + "loss": 0.0047, + "step": 80710 + }, + { + "epoch": 12.538055296675987, + "grad_norm": 0.24882885813713074, + "learning_rate": 9.82399254902212e-06, + "loss": 0.0042, + "step": 80720 + }, + { + "epoch": 12.539608574091332, + "grad_norm": 0.3570880591869354, + "learning_rate": 9.814152011255494e-06, + "loss": 0.0065, + "step": 80730 + }, + { + "epoch": 12.54116185150668, + "grad_norm": 0.2899796664714813, + "learning_rate": 9.804315868152797e-06, + "loss": 0.0036, + "step": 80740 + }, + { + "epoch": 12.542715128922026, + "grad_norm": 0.2829747796058655, + "learning_rate": 9.79448412078966e-06, + "loss": 0.0048, + "step": 80750 + }, + { + "epoch": 12.544268406337372, + "grad_norm": 0.10144510120153427, + "learning_rate": 9.784656770241312e-06, + "loss": 0.0032, + "step": 80760 + }, + { + "epoch": 12.545821683752719, + "grad_norm": 0.22101078927516937, + "learning_rate": 9.77483381758243e-06, + "loss": 0.0035, + "step": 80770 + }, + { + "epoch": 12.547374961168064, + "grad_norm": 0.04665708169341087, + "learning_rate": 9.76501526388725e-06, + "loss": 0.0051, + "step": 80780 + }, + { + "epoch": 12.54892823858341, + "grad_norm": 0.10124723613262177, + "learning_rate": 9.7552011102295e-06, + "loss": 0.0037, + "step": 80790 + }, + { + "epoch": 12.550481515998758, + "grad_norm": 0.36596646904945374, + "learning_rate": 9.745391357682449e-06, + "loss": 0.0034, + "step": 80800 + }, + { + "epoch": 12.552034793414103, + "grad_norm": 0.4143047630786896, + "learning_rate": 9.735586007318875e-06, + "loss": 0.0038, + "step": 80810 + }, + { + "epoch": 12.55358807082945, + "grad_norm": 0.2808021903038025, + "learning_rate": 9.725785060211096e-06, + "loss": 0.0031, + "step": 80820 + }, + { + "epoch": 12.555141348244797, + "grad_norm": 0.22911788523197174, + "learning_rate": 9.715988517430896e-06, + "loss": 0.0062, + "step": 80830 + }, + { + "epoch": 12.556694625660143, + "grad_norm": 0.7380132675170898, + "learning_rate": 9.706196380049625e-06, + "loss": 0.0032, + "step": 80840 + }, + { + "epoch": 12.55824790307549, + "grad_norm": 0.11324888467788696, + "learning_rate": 9.69640864913815e-06, + "loss": 0.0059, + "step": 80850 + }, + { + "epoch": 12.559801180490835, + "grad_norm": 0.1778561770915985, + "learning_rate": 9.686625325766818e-06, + "loss": 0.0061, + "step": 80860 + }, + { + "epoch": 12.561354457906182, + "grad_norm": 0.168364480137825, + "learning_rate": 9.67684641100554e-06, + "loss": 0.0039, + "step": 80870 + }, + { + "epoch": 12.562907735321529, + "grad_norm": 0.20636776089668274, + "learning_rate": 9.667071905923692e-06, + "loss": 0.0048, + "step": 80880 + }, + { + "epoch": 12.564461012736874, + "grad_norm": 0.28533047437667847, + "learning_rate": 9.657301811590242e-06, + "loss": 0.0032, + "step": 80890 + }, + { + "epoch": 12.566014290152221, + "grad_norm": 0.09532660990953445, + "learning_rate": 9.647536129073603e-06, + "loss": 0.0025, + "step": 80900 + }, + { + "epoch": 12.567567567567568, + "grad_norm": 0.25796160101890564, + "learning_rate": 9.637774859441751e-06, + "loss": 0.0039, + "step": 80910 + }, + { + "epoch": 12.569120844982914, + "grad_norm": 0.3583016097545624, + "learning_rate": 9.628018003762152e-06, + "loss": 0.0039, + "step": 80920 + }, + { + "epoch": 12.57067412239826, + "grad_norm": 0.37327200174331665, + "learning_rate": 9.618265563101813e-06, + "loss": 0.0061, + "step": 80930 + }, + { + "epoch": 12.572227399813606, + "grad_norm": 0.11694610118865967, + "learning_rate": 9.608517538527211e-06, + "loss": 0.0034, + "step": 80940 + }, + { + "epoch": 12.573780677228953, + "grad_norm": 0.12285171449184418, + "learning_rate": 9.598773931104431e-06, + "loss": 0.0034, + "step": 80950 + }, + { + "epoch": 12.5753339546443, + "grad_norm": 0.3961528539657593, + "learning_rate": 9.589034741898983e-06, + "loss": 0.004, + "step": 80960 + }, + { + "epoch": 12.576887232059645, + "grad_norm": 0.1282815933227539, + "learning_rate": 9.579299971975947e-06, + "loss": 0.0051, + "step": 80970 + }, + { + "epoch": 12.578440509474992, + "grad_norm": 0.2542479634284973, + "learning_rate": 9.569569622399882e-06, + "loss": 0.0029, + "step": 80980 + }, + { + "epoch": 12.579993786890338, + "grad_norm": 0.29332205653190613, + "learning_rate": 9.559843694234916e-06, + "loss": 0.0038, + "step": 80990 + }, + { + "epoch": 12.581547064305685, + "grad_norm": 0.5207366347312927, + "learning_rate": 9.550122188544625e-06, + "loss": 0.0029, + "step": 81000 + }, + { + "epoch": 12.583100341721032, + "grad_norm": 1.5312436819076538, + "learning_rate": 9.540405106392152e-06, + "loss": 0.0043, + "step": 81010 + }, + { + "epoch": 12.584653619136377, + "grad_norm": 0.38605740666389465, + "learning_rate": 9.53069244884015e-06, + "loss": 0.0046, + "step": 81020 + }, + { + "epoch": 12.586206896551724, + "grad_norm": 0.16500040888786316, + "learning_rate": 9.52098421695079e-06, + "loss": 0.0032, + "step": 81030 + }, + { + "epoch": 12.587760173967071, + "grad_norm": 0.3219698369503021, + "learning_rate": 9.511280411785728e-06, + "loss": 0.0033, + "step": 81040 + }, + { + "epoch": 12.589313451382417, + "grad_norm": 0.2427758276462555, + "learning_rate": 9.50158103440616e-06, + "loss": 0.004, + "step": 81050 + }, + { + "epoch": 12.590866728797764, + "grad_norm": 0.29887816309928894, + "learning_rate": 9.491886085872815e-06, + "loss": 0.0037, + "step": 81060 + }, + { + "epoch": 12.592420006213109, + "grad_norm": 0.09263424575328827, + "learning_rate": 9.482195567245895e-06, + "loss": 0.0036, + "step": 81070 + }, + { + "epoch": 12.593973283628456, + "grad_norm": 0.25626808404922485, + "learning_rate": 9.472509479585151e-06, + "loss": 0.0034, + "step": 81080 + }, + { + "epoch": 12.595526561043803, + "grad_norm": 0.1005820706486702, + "learning_rate": 9.462827823949833e-06, + "loss": 0.0029, + "step": 81090 + }, + { + "epoch": 12.597079838459148, + "grad_norm": 0.18049882352352142, + "learning_rate": 9.453150601398735e-06, + "loss": 0.0038, + "step": 81100 + }, + { + "epoch": 12.598633115874495, + "grad_norm": 0.13312825560569763, + "learning_rate": 9.44347781299011e-06, + "loss": 0.0029, + "step": 81110 + }, + { + "epoch": 12.600186393289842, + "grad_norm": 0.2213100790977478, + "learning_rate": 9.43380945978179e-06, + "loss": 0.0037, + "step": 81120 + }, + { + "epoch": 12.601739670705188, + "grad_norm": 0.13008303940296173, + "learning_rate": 9.424145542831047e-06, + "loss": 0.0048, + "step": 81130 + }, + { + "epoch": 12.603292948120535, + "grad_norm": 0.4579942524433136, + "learning_rate": 9.414486063194778e-06, + "loss": 0.0064, + "step": 81140 + }, + { + "epoch": 12.60484622553588, + "grad_norm": 0.13693274557590485, + "learning_rate": 9.404831021929273e-06, + "loss": 0.0043, + "step": 81150 + }, + { + "epoch": 12.606399502951227, + "grad_norm": 0.3876551389694214, + "learning_rate": 9.395180420090427e-06, + "loss": 0.0047, + "step": 81160 + }, + { + "epoch": 12.607952780366574, + "grad_norm": 0.1515849232673645, + "learning_rate": 9.385534258733591e-06, + "loss": 0.0034, + "step": 81170 + }, + { + "epoch": 12.60950605778192, + "grad_norm": 0.16496218740940094, + "learning_rate": 9.375892538913677e-06, + "loss": 0.004, + "step": 81180 + }, + { + "epoch": 12.611059335197266, + "grad_norm": 0.3511877954006195, + "learning_rate": 9.36625526168507e-06, + "loss": 0.0035, + "step": 81190 + }, + { + "epoch": 12.612612612612612, + "grad_norm": 0.45246225595474243, + "learning_rate": 9.356622428101686e-06, + "loss": 0.0057, + "step": 81200 + }, + { + "epoch": 12.614165890027959, + "grad_norm": 0.22876331210136414, + "learning_rate": 9.346994039216972e-06, + "loss": 0.003, + "step": 81210 + }, + { + "epoch": 12.615719167443306, + "grad_norm": 0.2662716507911682, + "learning_rate": 9.337370096083886e-06, + "loss": 0.0051, + "step": 81220 + }, + { + "epoch": 12.617272444858651, + "grad_norm": 0.16868484020233154, + "learning_rate": 9.327750599754854e-06, + "loss": 0.003, + "step": 81230 + }, + { + "epoch": 12.618825722273998, + "grad_norm": 0.17296317219734192, + "learning_rate": 9.318135551281875e-06, + "loss": 0.0034, + "step": 81240 + }, + { + "epoch": 12.620378999689345, + "grad_norm": 0.33284977078437805, + "learning_rate": 9.308524951716424e-06, + "loss": 0.0035, + "step": 81250 + }, + { + "epoch": 12.62193227710469, + "grad_norm": 0.7371264100074768, + "learning_rate": 9.298918802109497e-06, + "loss": 0.003, + "step": 81260 + }, + { + "epoch": 12.623485554520038, + "grad_norm": 0.4510236382484436, + "learning_rate": 9.289317103511625e-06, + "loss": 0.0039, + "step": 81270 + }, + { + "epoch": 12.625038831935385, + "grad_norm": 0.1819225251674652, + "learning_rate": 9.279719856972835e-06, + "loss": 0.0042, + "step": 81280 + }, + { + "epoch": 12.62659210935073, + "grad_norm": 0.6615320444107056, + "learning_rate": 9.270127063542645e-06, + "loss": 0.0038, + "step": 81290 + }, + { + "epoch": 12.628145386766077, + "grad_norm": 0.0955703929066658, + "learning_rate": 9.260538724270124e-06, + "loss": 0.0042, + "step": 81300 + }, + { + "epoch": 12.629698664181422, + "grad_norm": 0.20109017193317413, + "learning_rate": 9.250954840203852e-06, + "loss": 0.0039, + "step": 81310 + }, + { + "epoch": 12.63125194159677, + "grad_norm": 0.12807200849056244, + "learning_rate": 9.241375412391878e-06, + "loss": 0.0059, + "step": 81320 + }, + { + "epoch": 12.632805219012116, + "grad_norm": 0.5822877287864685, + "learning_rate": 9.23180044188181e-06, + "loss": 0.0042, + "step": 81330 + }, + { + "epoch": 12.634358496427462, + "grad_norm": 0.36046576499938965, + "learning_rate": 9.222229929720755e-06, + "loss": 0.0042, + "step": 81340 + }, + { + "epoch": 12.635911773842809, + "grad_norm": 0.44564905762672424, + "learning_rate": 9.212663876955336e-06, + "loss": 0.0033, + "step": 81350 + }, + { + "epoch": 12.637465051258154, + "grad_norm": 0.2584969401359558, + "learning_rate": 9.20310228463166e-06, + "loss": 0.0044, + "step": 81360 + }, + { + "epoch": 12.639018328673501, + "grad_norm": 0.33148878812789917, + "learning_rate": 9.19354515379539e-06, + "loss": 0.0059, + "step": 81370 + }, + { + "epoch": 12.640571606088848, + "grad_norm": 0.27569422125816345, + "learning_rate": 9.183992485491666e-06, + "loss": 0.0051, + "step": 81380 + }, + { + "epoch": 12.642124883504193, + "grad_norm": 0.16118374466896057, + "learning_rate": 9.174444280765154e-06, + "loss": 0.0034, + "step": 81390 + }, + { + "epoch": 12.64367816091954, + "grad_norm": 0.22387544810771942, + "learning_rate": 9.164900540660032e-06, + "loss": 0.007, + "step": 81400 + }, + { + "epoch": 12.645231438334887, + "grad_norm": 0.5019624829292297, + "learning_rate": 9.155361266220008e-06, + "loss": 0.0042, + "step": 81410 + }, + { + "epoch": 12.646784715750233, + "grad_norm": 0.35379743576049805, + "learning_rate": 9.145826458488254e-06, + "loss": 0.003, + "step": 81420 + }, + { + "epoch": 12.64833799316558, + "grad_norm": 0.1634192019701004, + "learning_rate": 9.136296118507509e-06, + "loss": 0.0045, + "step": 81430 + }, + { + "epoch": 12.649891270580925, + "grad_norm": 0.1324181705713272, + "learning_rate": 9.126770247319966e-06, + "loss": 0.0037, + "step": 81440 + }, + { + "epoch": 12.651444547996272, + "grad_norm": 0.21997645497322083, + "learning_rate": 9.11724884596738e-06, + "loss": 0.0052, + "step": 81450 + }, + { + "epoch": 12.652997825411619, + "grad_norm": 0.12160667032003403, + "learning_rate": 9.107731915490996e-06, + "loss": 0.0046, + "step": 81460 + }, + { + "epoch": 12.654551102826964, + "grad_norm": 0.3785523474216461, + "learning_rate": 9.098219456931579e-06, + "loss": 0.0046, + "step": 81470 + }, + { + "epoch": 12.656104380242311, + "grad_norm": 0.2634195387363434, + "learning_rate": 9.088711471329376e-06, + "loss": 0.0044, + "step": 81480 + }, + { + "epoch": 12.657657657657658, + "grad_norm": 0.22803688049316406, + "learning_rate": 9.079207959724184e-06, + "loss": 0.0044, + "step": 81490 + }, + { + "epoch": 12.659210935073004, + "grad_norm": 0.05760055407881737, + "learning_rate": 9.069708923155279e-06, + "loss": 0.0032, + "step": 81500 + }, + { + "epoch": 12.66076421248835, + "grad_norm": 0.42182090878486633, + "learning_rate": 9.060214362661468e-06, + "loss": 0.0052, + "step": 81510 + }, + { + "epoch": 12.662317489903696, + "grad_norm": 0.07868874073028564, + "learning_rate": 9.050724279281064e-06, + "loss": 0.0047, + "step": 81520 + }, + { + "epoch": 12.663870767319043, + "grad_norm": 0.8612228035926819, + "learning_rate": 9.041238674051894e-06, + "loss": 0.0057, + "step": 81530 + }, + { + "epoch": 12.66542404473439, + "grad_norm": 0.12269312143325806, + "learning_rate": 9.031757548011277e-06, + "loss": 0.0039, + "step": 81540 + }, + { + "epoch": 12.666977322149735, + "grad_norm": 0.3630163073539734, + "learning_rate": 9.022280902196051e-06, + "loss": 0.0041, + "step": 81550 + }, + { + "epoch": 12.668530599565083, + "grad_norm": 0.1554444134235382, + "learning_rate": 9.012808737642592e-06, + "loss": 0.0046, + "step": 81560 + }, + { + "epoch": 12.670083876980428, + "grad_norm": 0.11248339712619781, + "learning_rate": 9.003341055386739e-06, + "loss": 0.0046, + "step": 81570 + }, + { + "epoch": 12.671637154395775, + "grad_norm": 0.07514215260744095, + "learning_rate": 8.993877856463862e-06, + "loss": 0.0038, + "step": 81580 + }, + { + "epoch": 12.673190431811122, + "grad_norm": 0.12195941060781479, + "learning_rate": 8.984419141908856e-06, + "loss": 0.0049, + "step": 81590 + }, + { + "epoch": 12.674743709226467, + "grad_norm": 0.2239120900630951, + "learning_rate": 8.974964912756117e-06, + "loss": 0.0022, + "step": 81600 + }, + { + "epoch": 12.676296986641814, + "grad_norm": 0.2882470488548279, + "learning_rate": 8.965515170039523e-06, + "loss": 0.0044, + "step": 81610 + }, + { + "epoch": 12.677850264057161, + "grad_norm": 0.5245731472969055, + "learning_rate": 8.956069914792514e-06, + "loss": 0.0045, + "step": 81620 + }, + { + "epoch": 12.679403541472507, + "grad_norm": 0.1872268170118332, + "learning_rate": 8.946629148047969e-06, + "loss": 0.0046, + "step": 81630 + }, + { + "epoch": 12.680956818887854, + "grad_norm": 0.12625490128993988, + "learning_rate": 8.93719287083834e-06, + "loss": 0.0043, + "step": 81640 + }, + { + "epoch": 12.682510096303199, + "grad_norm": 1.1842910051345825, + "learning_rate": 8.927761084195569e-06, + "loss": 0.0042, + "step": 81650 + }, + { + "epoch": 12.684063373718546, + "grad_norm": 0.34945687651634216, + "learning_rate": 8.918333789151096e-06, + "loss": 0.0026, + "step": 81660 + }, + { + "epoch": 12.685616651133893, + "grad_norm": 0.25722554326057434, + "learning_rate": 8.90891098673587e-06, + "loss": 0.0037, + "step": 81670 + }, + { + "epoch": 12.687169928549238, + "grad_norm": 0.3793644309043884, + "learning_rate": 8.89949267798037e-06, + "loss": 0.0051, + "step": 81680 + }, + { + "epoch": 12.688723205964585, + "grad_norm": 0.4883086681365967, + "learning_rate": 8.890078863914542e-06, + "loss": 0.0037, + "step": 81690 + }, + { + "epoch": 12.690276483379932, + "grad_norm": 0.16387592256069183, + "learning_rate": 8.880669545567894e-06, + "loss": 0.0039, + "step": 81700 + }, + { + "epoch": 12.691829760795278, + "grad_norm": 0.11742222309112549, + "learning_rate": 8.871264723969381e-06, + "loss": 0.0032, + "step": 81710 + }, + { + "epoch": 12.693383038210625, + "grad_norm": 0.4048082232475281, + "learning_rate": 8.86186440014754e-06, + "loss": 0.0043, + "step": 81720 + }, + { + "epoch": 12.69493631562597, + "grad_norm": 0.6104451417922974, + "learning_rate": 8.852468575130351e-06, + "loss": 0.0041, + "step": 81730 + }, + { + "epoch": 12.696489593041317, + "grad_norm": 0.22791723906993866, + "learning_rate": 8.843077249945347e-06, + "loss": 0.0042, + "step": 81740 + }, + { + "epoch": 12.698042870456664, + "grad_norm": 0.2746424078941345, + "learning_rate": 8.833690425619523e-06, + "loss": 0.0057, + "step": 81750 + }, + { + "epoch": 12.69959614787201, + "grad_norm": 0.1959468424320221, + "learning_rate": 8.82430810317942e-06, + "loss": 0.0037, + "step": 81760 + }, + { + "epoch": 12.701149425287356, + "grad_norm": 0.13304157555103302, + "learning_rate": 8.814930283651074e-06, + "loss": 0.0034, + "step": 81770 + }, + { + "epoch": 12.702702702702704, + "grad_norm": 0.35258227586746216, + "learning_rate": 8.805556968060048e-06, + "loss": 0.0037, + "step": 81780 + }, + { + "epoch": 12.704255980118049, + "grad_norm": 0.7531948089599609, + "learning_rate": 8.796188157431362e-06, + "loss": 0.0056, + "step": 81790 + }, + { + "epoch": 12.705809257533396, + "grad_norm": 0.09685437381267548, + "learning_rate": 8.786823852789595e-06, + "loss": 0.0032, + "step": 81800 + }, + { + "epoch": 12.707362534948741, + "grad_norm": 0.3118429481983185, + "learning_rate": 8.777464055158813e-06, + "loss": 0.0048, + "step": 81810 + }, + { + "epoch": 12.708915812364088, + "grad_norm": 0.7653037905693054, + "learning_rate": 8.768108765562578e-06, + "loss": 0.0055, + "step": 81820 + }, + { + "epoch": 12.710469089779435, + "grad_norm": 0.44344398379325867, + "learning_rate": 8.758757985023975e-06, + "loss": 0.0032, + "step": 81830 + }, + { + "epoch": 12.71202236719478, + "grad_norm": 0.09401684999465942, + "learning_rate": 8.749411714565598e-06, + "loss": 0.0042, + "step": 81840 + }, + { + "epoch": 12.713575644610128, + "grad_norm": 0.24950715899467468, + "learning_rate": 8.74006995520954e-06, + "loss": 0.0032, + "step": 81850 + }, + { + "epoch": 12.715128922025475, + "grad_norm": 0.5442221164703369, + "learning_rate": 8.730732707977397e-06, + "loss": 0.004, + "step": 81860 + }, + { + "epoch": 12.71668219944082, + "grad_norm": 0.42536044120788574, + "learning_rate": 8.721399973890281e-06, + "loss": 0.0054, + "step": 81870 + }, + { + "epoch": 12.718235476856167, + "grad_norm": 0.11779452860355377, + "learning_rate": 8.712071753968792e-06, + "loss": 0.0036, + "step": 81880 + }, + { + "epoch": 12.719788754271512, + "grad_norm": 0.19944651424884796, + "learning_rate": 8.702748049233072e-06, + "loss": 0.0045, + "step": 81890 + }, + { + "epoch": 12.72134203168686, + "grad_norm": 0.07717528194189072, + "learning_rate": 8.693428860702718e-06, + "loss": 0.0038, + "step": 81900 + }, + { + "epoch": 12.722895309102206, + "grad_norm": 0.4408000707626343, + "learning_rate": 8.684114189396898e-06, + "loss": 0.0031, + "step": 81910 + }, + { + "epoch": 12.724448586517552, + "grad_norm": 0.4566461443901062, + "learning_rate": 8.674804036334222e-06, + "loss": 0.0046, + "step": 81920 + }, + { + "epoch": 12.726001863932899, + "grad_norm": 0.21899062395095825, + "learning_rate": 8.665498402532857e-06, + "loss": 0.0038, + "step": 81930 + }, + { + "epoch": 12.727555141348244, + "grad_norm": 0.9216101169586182, + "learning_rate": 8.656197289010432e-06, + "loss": 0.0048, + "step": 81940 + }, + { + "epoch": 12.729108418763591, + "grad_norm": 0.09144724905490875, + "learning_rate": 8.646900696784122e-06, + "loss": 0.0063, + "step": 81950 + }, + { + "epoch": 12.730661696178938, + "grad_norm": 0.14278079569339752, + "learning_rate": 8.637608626870553e-06, + "loss": 0.0038, + "step": 81960 + }, + { + "epoch": 12.732214973594283, + "grad_norm": 0.13467729091644287, + "learning_rate": 8.628321080285944e-06, + "loss": 0.0032, + "step": 81970 + }, + { + "epoch": 12.73376825100963, + "grad_norm": 0.12920992076396942, + "learning_rate": 8.619038058045925e-06, + "loss": 0.002, + "step": 81980 + }, + { + "epoch": 12.735321528424977, + "grad_norm": 0.4170820415019989, + "learning_rate": 8.609759561165704e-06, + "loss": 0.0045, + "step": 81990 + }, + { + "epoch": 12.736874805840323, + "grad_norm": 0.2401798814535141, + "learning_rate": 8.600485590659934e-06, + "loss": 0.0056, + "step": 82000 + }, + { + "epoch": 12.73842808325567, + "grad_norm": 0.11637873202562332, + "learning_rate": 8.591216147542813e-06, + "loss": 0.0043, + "step": 82010 + }, + { + "epoch": 12.739981360671015, + "grad_norm": 0.06723124533891678, + "learning_rate": 8.581951232828045e-06, + "loss": 0.0038, + "step": 82020 + }, + { + "epoch": 12.741534638086362, + "grad_norm": 0.43779516220092773, + "learning_rate": 8.572690847528825e-06, + "loss": 0.0022, + "step": 82030 + }, + { + "epoch": 12.74308791550171, + "grad_norm": 1.171702265739441, + "learning_rate": 8.563434992657838e-06, + "loss": 0.0036, + "step": 82040 + }, + { + "epoch": 12.744641192917054, + "grad_norm": 0.6184371113777161, + "learning_rate": 8.554183669227307e-06, + "loss": 0.0036, + "step": 82050 + }, + { + "epoch": 12.746194470332401, + "grad_norm": 0.1928822249174118, + "learning_rate": 8.544936878248937e-06, + "loss": 0.0042, + "step": 82060 + }, + { + "epoch": 12.747747747747749, + "grad_norm": 0.23988239467144012, + "learning_rate": 8.535694620733942e-06, + "loss": 0.0037, + "step": 82070 + }, + { + "epoch": 12.749301025163094, + "grad_norm": 0.13795387744903564, + "learning_rate": 8.526456897693053e-06, + "loss": 0.0035, + "step": 82080 + }, + { + "epoch": 12.75085430257844, + "grad_norm": 0.21072643995285034, + "learning_rate": 8.517223710136456e-06, + "loss": 0.0034, + "step": 82090 + }, + { + "epoch": 12.752407579993786, + "grad_norm": 0.43753746151924133, + "learning_rate": 8.507995059073931e-06, + "loss": 0.0059, + "step": 82100 + }, + { + "epoch": 12.753960857409133, + "grad_norm": 0.4226934015750885, + "learning_rate": 8.498770945514673e-06, + "loss": 0.0036, + "step": 82110 + }, + { + "epoch": 12.75551413482448, + "grad_norm": 0.045280300080776215, + "learning_rate": 8.489551370467436e-06, + "loss": 0.0048, + "step": 82120 + }, + { + "epoch": 12.757067412239826, + "grad_norm": 0.08775725215673447, + "learning_rate": 8.480336334940442e-06, + "loss": 0.0043, + "step": 82130 + }, + { + "epoch": 12.758620689655173, + "grad_norm": 0.3521093726158142, + "learning_rate": 8.471125839941458e-06, + "loss": 0.0036, + "step": 82140 + }, + { + "epoch": 12.760173967070518, + "grad_norm": 0.10539062321186066, + "learning_rate": 8.461919886477682e-06, + "loss": 0.003, + "step": 82150 + }, + { + "epoch": 12.761727244485865, + "grad_norm": 0.5820005536079407, + "learning_rate": 8.452718475555927e-06, + "loss": 0.0043, + "step": 82160 + }, + { + "epoch": 12.763280521901212, + "grad_norm": 0.12781348824501038, + "learning_rate": 8.443521608182398e-06, + "loss": 0.0032, + "step": 82170 + }, + { + "epoch": 12.764833799316557, + "grad_norm": 0.364327073097229, + "learning_rate": 8.43432928536288e-06, + "loss": 0.0031, + "step": 82180 + }, + { + "epoch": 12.766387076731904, + "grad_norm": 0.13962316513061523, + "learning_rate": 8.425141508102607e-06, + "loss": 0.0044, + "step": 82190 + }, + { + "epoch": 12.767940354147251, + "grad_norm": 0.6696566343307495, + "learning_rate": 8.415958277406365e-06, + "loss": 0.0034, + "step": 82200 + }, + { + "epoch": 12.769493631562597, + "grad_norm": 0.05308643728494644, + "learning_rate": 8.406779594278392e-06, + "loss": 0.0029, + "step": 82210 + }, + { + "epoch": 12.771046908977944, + "grad_norm": 0.22938638925552368, + "learning_rate": 8.397605459722468e-06, + "loss": 0.0083, + "step": 82220 + }, + { + "epoch": 12.77260018639329, + "grad_norm": 0.34891659021377563, + "learning_rate": 8.388435874741863e-06, + "loss": 0.0039, + "step": 82230 + }, + { + "epoch": 12.774153463808636, + "grad_norm": 0.10194548964500427, + "learning_rate": 8.37927084033936e-06, + "loss": 0.0038, + "step": 82240 + }, + { + "epoch": 12.775706741223983, + "grad_norm": 0.15809740126132965, + "learning_rate": 8.370110357517208e-06, + "loss": 0.0047, + "step": 82250 + }, + { + "epoch": 12.777260018639328, + "grad_norm": 0.33011704683303833, + "learning_rate": 8.360954427277201e-06, + "loss": 0.0027, + "step": 82260 + }, + { + "epoch": 12.778813296054675, + "grad_norm": 0.11875507980585098, + "learning_rate": 8.351803050620627e-06, + "loss": 0.0065, + "step": 82270 + }, + { + "epoch": 12.780366573470022, + "grad_norm": 0.2551243305206299, + "learning_rate": 8.34265622854824e-06, + "loss": 0.0042, + "step": 82280 + }, + { + "epoch": 12.781919850885368, + "grad_norm": 0.1472456008195877, + "learning_rate": 8.333513962060335e-06, + "loss": 0.0048, + "step": 82290 + }, + { + "epoch": 12.783473128300715, + "grad_norm": 0.3975915014743805, + "learning_rate": 8.324376252156701e-06, + "loss": 0.005, + "step": 82300 + }, + { + "epoch": 12.78502640571606, + "grad_norm": 0.07974831759929657, + "learning_rate": 8.315243099836633e-06, + "loss": 0.0041, + "step": 82310 + }, + { + "epoch": 12.786579683131407, + "grad_norm": 0.19762147963047028, + "learning_rate": 8.3061145060989e-06, + "loss": 0.0042, + "step": 82320 + }, + { + "epoch": 12.788132960546754, + "grad_norm": 0.19135107100009918, + "learning_rate": 8.296990471941806e-06, + "loss": 0.0033, + "step": 82330 + }, + { + "epoch": 12.7896862379621, + "grad_norm": 0.3219304382801056, + "learning_rate": 8.287870998363112e-06, + "loss": 0.0038, + "step": 82340 + }, + { + "epoch": 12.791239515377447, + "grad_norm": 0.09616968035697937, + "learning_rate": 8.278756086360156e-06, + "loss": 0.0035, + "step": 82350 + }, + { + "epoch": 12.792792792792794, + "grad_norm": 0.2062777578830719, + "learning_rate": 8.269645736929693e-06, + "loss": 0.0048, + "step": 82360 + }, + { + "epoch": 12.794346070208139, + "grad_norm": 0.512552797794342, + "learning_rate": 8.260539951068048e-06, + "loss": 0.004, + "step": 82370 + }, + { + "epoch": 12.795899347623486, + "grad_norm": 0.31414735317230225, + "learning_rate": 8.251438729770983e-06, + "loss": 0.0045, + "step": 82380 + }, + { + "epoch": 12.797452625038831, + "grad_norm": 0.1922992318868637, + "learning_rate": 8.242342074033826e-06, + "loss": 0.003, + "step": 82390 + }, + { + "epoch": 12.799005902454178, + "grad_norm": 0.6264548301696777, + "learning_rate": 8.233249984851344e-06, + "loss": 0.0028, + "step": 82400 + }, + { + "epoch": 12.800559179869525, + "grad_norm": 0.07289901375770569, + "learning_rate": 8.224162463217849e-06, + "loss": 0.0033, + "step": 82410 + }, + { + "epoch": 12.80211245728487, + "grad_norm": 0.10248501598834991, + "learning_rate": 8.215079510127139e-06, + "loss": 0.0048, + "step": 82420 + }, + { + "epoch": 12.803665734700218, + "grad_norm": 0.4257126748561859, + "learning_rate": 8.206001126572521e-06, + "loss": 0.0024, + "step": 82430 + }, + { + "epoch": 12.805219012115565, + "grad_norm": 0.1459687054157257, + "learning_rate": 8.196927313546776e-06, + "loss": 0.0028, + "step": 82440 + }, + { + "epoch": 12.80677228953091, + "grad_norm": 0.3457234799861908, + "learning_rate": 8.18785807204222e-06, + "loss": 0.0041, + "step": 82450 + }, + { + "epoch": 12.808325566946257, + "grad_norm": 0.08983021974563599, + "learning_rate": 8.178793403050627e-06, + "loss": 0.0027, + "step": 82460 + }, + { + "epoch": 12.809878844361602, + "grad_norm": 0.2562587857246399, + "learning_rate": 8.16973330756331e-06, + "loss": 0.0028, + "step": 82470 + }, + { + "epoch": 12.81143212177695, + "grad_norm": 0.16973650455474854, + "learning_rate": 8.16067778657107e-06, + "loss": 0.0026, + "step": 82480 + }, + { + "epoch": 12.812985399192296, + "grad_norm": 0.17103278636932373, + "learning_rate": 8.151626841064214e-06, + "loss": 0.0026, + "step": 82490 + }, + { + "epoch": 12.814538676607642, + "grad_norm": 0.15589189529418945, + "learning_rate": 8.142580472032518e-06, + "loss": 0.0028, + "step": 82500 + }, + { + "epoch": 12.816091954022989, + "grad_norm": 0.12582644820213318, + "learning_rate": 8.133538680465285e-06, + "loss": 0.0031, + "step": 82510 + }, + { + "epoch": 12.817645231438334, + "grad_norm": 0.4532320201396942, + "learning_rate": 8.124501467351336e-06, + "loss": 0.0059, + "step": 82520 + }, + { + "epoch": 12.819198508853681, + "grad_norm": 0.9928760528564453, + "learning_rate": 8.11546883367893e-06, + "loss": 0.0037, + "step": 82530 + }, + { + "epoch": 12.820751786269028, + "grad_norm": 0.6244738698005676, + "learning_rate": 8.106440780435882e-06, + "loss": 0.0049, + "step": 82540 + }, + { + "epoch": 12.822305063684373, + "grad_norm": 0.18397517502307892, + "learning_rate": 8.097417308609485e-06, + "loss": 0.0034, + "step": 82550 + }, + { + "epoch": 12.82385834109972, + "grad_norm": 0.10098672658205032, + "learning_rate": 8.088398419186544e-06, + "loss": 0.0027, + "step": 82560 + }, + { + "epoch": 12.825411618515068, + "grad_norm": 0.37561482191085815, + "learning_rate": 8.079384113153321e-06, + "loss": 0.0043, + "step": 82570 + }, + { + "epoch": 12.826964895930413, + "grad_norm": 0.15450751781463623, + "learning_rate": 8.070374391495644e-06, + "loss": 0.0031, + "step": 82580 + }, + { + "epoch": 12.82851817334576, + "grad_norm": 0.43880364298820496, + "learning_rate": 8.061369255198764e-06, + "loss": 0.0034, + "step": 82590 + }, + { + "epoch": 12.830071450761105, + "grad_norm": 0.18151704967021942, + "learning_rate": 8.052368705247493e-06, + "loss": 0.0027, + "step": 82600 + }, + { + "epoch": 12.831624728176452, + "grad_norm": 0.672444760799408, + "learning_rate": 8.043372742626114e-06, + "loss": 0.006, + "step": 82610 + }, + { + "epoch": 12.8331780055918, + "grad_norm": 0.11764443665742874, + "learning_rate": 8.034381368318422e-06, + "loss": 0.0041, + "step": 82620 + }, + { + "epoch": 12.834731283007144, + "grad_norm": 0.15469752252101898, + "learning_rate": 8.025394583307677e-06, + "loss": 0.0033, + "step": 82630 + }, + { + "epoch": 12.836284560422492, + "grad_norm": 0.2588527202606201, + "learning_rate": 8.016412388576683e-06, + "loss": 0.0043, + "step": 82640 + }, + { + "epoch": 12.837837837837839, + "grad_norm": 0.44309762120246887, + "learning_rate": 8.007434785107699e-06, + "loss": 0.0067, + "step": 82650 + }, + { + "epoch": 12.839391115253184, + "grad_norm": 0.46401819586753845, + "learning_rate": 7.998461773882509e-06, + "loss": 0.0038, + "step": 82660 + }, + { + "epoch": 12.840944392668531, + "grad_norm": 0.232129767537117, + "learning_rate": 7.989493355882393e-06, + "loss": 0.0036, + "step": 82670 + }, + { + "epoch": 12.842497670083876, + "grad_norm": 0.23098869621753693, + "learning_rate": 7.980529532088137e-06, + "loss": 0.0042, + "step": 82680 + }, + { + "epoch": 12.844050947499223, + "grad_norm": 0.08347819745540619, + "learning_rate": 7.97157030347998e-06, + "loss": 0.0051, + "step": 82690 + }, + { + "epoch": 12.84560422491457, + "grad_norm": 0.117672860622406, + "learning_rate": 7.96261567103772e-06, + "loss": 0.0038, + "step": 82700 + }, + { + "epoch": 12.847157502329916, + "grad_norm": 0.06288520246744156, + "learning_rate": 7.953665635740593e-06, + "loss": 0.0034, + "step": 82710 + }, + { + "epoch": 12.848710779745263, + "grad_norm": 0.250168114900589, + "learning_rate": 7.94472019856738e-06, + "loss": 0.0039, + "step": 82720 + }, + { + "epoch": 12.85026405716061, + "grad_norm": 0.9462501406669617, + "learning_rate": 7.935779360496337e-06, + "loss": 0.0055, + "step": 82730 + }, + { + "epoch": 12.851817334575955, + "grad_norm": 0.19866903126239777, + "learning_rate": 7.92684312250523e-06, + "loss": 0.005, + "step": 82740 + }, + { + "epoch": 12.853370611991302, + "grad_norm": 0.13777995109558105, + "learning_rate": 7.917911485571289e-06, + "loss": 0.0034, + "step": 82750 + }, + { + "epoch": 12.854923889406647, + "grad_norm": 1.3950836658477783, + "learning_rate": 7.90898445067128e-06, + "loss": 0.0041, + "step": 82760 + }, + { + "epoch": 12.856477166821994, + "grad_norm": 0.46852996945381165, + "learning_rate": 7.900062018781462e-06, + "loss": 0.0044, + "step": 82770 + }, + { + "epoch": 12.858030444237341, + "grad_norm": 0.3158811926841736, + "learning_rate": 7.891144190877553e-06, + "loss": 0.0038, + "step": 82780 + }, + { + "epoch": 12.859583721652687, + "grad_norm": 0.09631819278001785, + "learning_rate": 7.882230967934806e-06, + "loss": 0.0034, + "step": 82790 + }, + { + "epoch": 12.861136999068034, + "grad_norm": 0.17732705175876617, + "learning_rate": 7.873322350927959e-06, + "loss": 0.0032, + "step": 82800 + }, + { + "epoch": 12.86269027648338, + "grad_norm": 0.16101977229118347, + "learning_rate": 7.864418340831253e-06, + "loss": 0.0046, + "step": 82810 + }, + { + "epoch": 12.864243553898726, + "grad_norm": 0.1492070108652115, + "learning_rate": 7.855518938618395e-06, + "loss": 0.0046, + "step": 82820 + }, + { + "epoch": 12.865796831314073, + "grad_norm": 0.18597276508808136, + "learning_rate": 7.846624145262633e-06, + "loss": 0.004, + "step": 82830 + }, + { + "epoch": 12.867350108729418, + "grad_norm": 0.10329832136631012, + "learning_rate": 7.837733961736672e-06, + "loss": 0.0031, + "step": 82840 + }, + { + "epoch": 12.868903386144765, + "grad_norm": 0.25918978452682495, + "learning_rate": 7.828848389012727e-06, + "loss": 0.0027, + "step": 82850 + }, + { + "epoch": 12.870456663560113, + "grad_norm": 0.07301712036132812, + "learning_rate": 7.819967428062524e-06, + "loss": 0.0037, + "step": 82860 + }, + { + "epoch": 12.872009940975458, + "grad_norm": 0.21195076406002045, + "learning_rate": 7.811091079857274e-06, + "loss": 0.0041, + "step": 82870 + }, + { + "epoch": 12.873563218390805, + "grad_norm": 0.17382894456386566, + "learning_rate": 7.802219345367667e-06, + "loss": 0.0043, + "step": 82880 + }, + { + "epoch": 12.87511649580615, + "grad_norm": 0.18206988275051117, + "learning_rate": 7.793352225563916e-06, + "loss": 0.0039, + "step": 82890 + }, + { + "epoch": 12.876669773221497, + "grad_norm": 0.16357138752937317, + "learning_rate": 7.7844897214157e-06, + "loss": 0.0035, + "step": 82900 + }, + { + "epoch": 12.878223050636844, + "grad_norm": 0.17445045709609985, + "learning_rate": 7.775631833892228e-06, + "loss": 0.0031, + "step": 82910 + }, + { + "epoch": 12.87977632805219, + "grad_norm": 0.12665802240371704, + "learning_rate": 7.76677856396215e-06, + "loss": 0.0058, + "step": 82920 + }, + { + "epoch": 12.881329605467537, + "grad_norm": 0.7514060139656067, + "learning_rate": 7.757929912593697e-06, + "loss": 0.0046, + "step": 82930 + }, + { + "epoch": 12.882882882882884, + "grad_norm": 0.25269395112991333, + "learning_rate": 7.749085880754509e-06, + "loss": 0.0045, + "step": 82940 + }, + { + "epoch": 12.884436160298229, + "grad_norm": 0.1471550464630127, + "learning_rate": 7.740246469411777e-06, + "loss": 0.0045, + "step": 82950 + }, + { + "epoch": 12.885989437713576, + "grad_norm": 0.38535982370376587, + "learning_rate": 7.731411679532142e-06, + "loss": 0.0049, + "step": 82960 + }, + { + "epoch": 12.887542715128921, + "grad_norm": 0.08167625963687897, + "learning_rate": 7.722581512081779e-06, + "loss": 0.0034, + "step": 82970 + }, + { + "epoch": 12.889095992544268, + "grad_norm": 0.35000720620155334, + "learning_rate": 7.713755968026331e-06, + "loss": 0.0029, + "step": 82980 + }, + { + "epoch": 12.890649269959615, + "grad_norm": 0.08507560193538666, + "learning_rate": 7.704935048330969e-06, + "loss": 0.0029, + "step": 82990 + }, + { + "epoch": 12.89220254737496, + "grad_norm": 0.3145081698894501, + "learning_rate": 7.696118753960313e-06, + "loss": 0.0039, + "step": 83000 + }, + { + "epoch": 12.893755824790308, + "grad_norm": 0.26186245679855347, + "learning_rate": 7.687307085878503e-06, + "loss": 0.0034, + "step": 83010 + }, + { + "epoch": 12.895309102205655, + "grad_norm": 0.33008450269699097, + "learning_rate": 7.678500045049186e-06, + "loss": 0.0033, + "step": 83020 + }, + { + "epoch": 12.896862379621, + "grad_norm": 0.1627390831708908, + "learning_rate": 7.669697632435463e-06, + "loss": 0.0038, + "step": 83030 + }, + { + "epoch": 12.898415657036347, + "grad_norm": 0.14363990724086761, + "learning_rate": 7.660899848999963e-06, + "loss": 0.0031, + "step": 83040 + }, + { + "epoch": 12.899968934451692, + "grad_norm": 0.2892288863658905, + "learning_rate": 7.652106695704803e-06, + "loss": 0.0035, + "step": 83050 + }, + { + "epoch": 12.90152221186704, + "grad_norm": 0.20862330496311188, + "learning_rate": 7.643318173511598e-06, + "loss": 0.0036, + "step": 83060 + }, + { + "epoch": 12.903075489282386, + "grad_norm": 0.4528431296348572, + "learning_rate": 7.634534283381423e-06, + "loss": 0.0038, + "step": 83070 + }, + { + "epoch": 12.904628766697732, + "grad_norm": 0.07736127823591232, + "learning_rate": 7.625755026274889e-06, + "loss": 0.0023, + "step": 83080 + }, + { + "epoch": 12.906182044113079, + "grad_norm": 0.12006108462810516, + "learning_rate": 7.616980403152069e-06, + "loss": 0.0039, + "step": 83090 + }, + { + "epoch": 12.907735321528424, + "grad_norm": 0.8563117980957031, + "learning_rate": 7.608210414972561e-06, + "loss": 0.005, + "step": 83100 + }, + { + "epoch": 12.909288598943771, + "grad_norm": 0.6033213138580322, + "learning_rate": 7.599445062695404e-06, + "loss": 0.004, + "step": 83110 + }, + { + "epoch": 12.910841876359118, + "grad_norm": 0.0921410620212555, + "learning_rate": 7.590684347279214e-06, + "loss": 0.0044, + "step": 83120 + }, + { + "epoch": 12.912395153774463, + "grad_norm": 0.33481356501579285, + "learning_rate": 7.581928269682004e-06, + "loss": 0.0028, + "step": 83130 + }, + { + "epoch": 12.91394843118981, + "grad_norm": 0.14977772533893585, + "learning_rate": 7.573176830861361e-06, + "loss": 0.0044, + "step": 83140 + }, + { + "epoch": 12.915501708605158, + "grad_norm": 0.6457715034484863, + "learning_rate": 7.5644300317743e-06, + "loss": 0.0039, + "step": 83150 + }, + { + "epoch": 12.917054986020503, + "grad_norm": 0.16480426490306854, + "learning_rate": 7.555687873377376e-06, + "loss": 0.0046, + "step": 83160 + }, + { + "epoch": 12.91860826343585, + "grad_norm": 0.36464643478393555, + "learning_rate": 7.546950356626592e-06, + "loss": 0.0029, + "step": 83170 + }, + { + "epoch": 12.920161540851197, + "grad_norm": 0.4235709607601166, + "learning_rate": 7.538217482477516e-06, + "loss": 0.0032, + "step": 83180 + }, + { + "epoch": 12.921714818266542, + "grad_norm": 0.10891039669513702, + "learning_rate": 7.529489251885119e-06, + "loss": 0.0034, + "step": 83190 + }, + { + "epoch": 12.92326809568189, + "grad_norm": 0.2601395547389984, + "learning_rate": 7.520765665803941e-06, + "loss": 0.0022, + "step": 83200 + }, + { + "epoch": 12.924821373097235, + "grad_norm": 0.1755746752023697, + "learning_rate": 7.512046725187949e-06, + "loss": 0.0042, + "step": 83210 + }, + { + "epoch": 12.926374650512582, + "grad_norm": 0.3065468668937683, + "learning_rate": 7.503332430990645e-06, + "loss": 0.0025, + "step": 83220 + }, + { + "epoch": 12.927927927927929, + "grad_norm": 0.10238602012395859, + "learning_rate": 7.494622784165017e-06, + "loss": 0.0033, + "step": 83230 + }, + { + "epoch": 12.929481205343274, + "grad_norm": 0.5703092813491821, + "learning_rate": 7.485917785663543e-06, + "loss": 0.0028, + "step": 83240 + }, + { + "epoch": 12.931034482758621, + "grad_norm": 0.27455979585647583, + "learning_rate": 7.477217436438172e-06, + "loss": 0.0055, + "step": 83250 + }, + { + "epoch": 12.932587760173966, + "grad_norm": 0.1379951685667038, + "learning_rate": 7.46852173744037e-06, + "loss": 0.0043, + "step": 83260 + }, + { + "epoch": 12.934141037589313, + "grad_norm": 0.21303115785121918, + "learning_rate": 7.4598306896211004e-06, + "loss": 0.0028, + "step": 83270 + }, + { + "epoch": 12.93569431500466, + "grad_norm": 0.1416223645210266, + "learning_rate": 7.451144293930773e-06, + "loss": 0.0027, + "step": 83280 + }, + { + "epoch": 12.937247592420006, + "grad_norm": 0.28897231817245483, + "learning_rate": 7.4424625513193455e-06, + "loss": 0.003, + "step": 83290 + }, + { + "epoch": 12.938800869835353, + "grad_norm": 0.17686545848846436, + "learning_rate": 7.433785462736209e-06, + "loss": 0.0029, + "step": 83300 + }, + { + "epoch": 12.9403541472507, + "grad_norm": 0.5459927916526794, + "learning_rate": 7.425113029130315e-06, + "loss": 0.0035, + "step": 83310 + }, + { + "epoch": 12.941907424666045, + "grad_norm": 0.14050625264644623, + "learning_rate": 7.416445251450044e-06, + "loss": 0.0065, + "step": 83320 + }, + { + "epoch": 12.943460702081392, + "grad_norm": 0.6761507391929626, + "learning_rate": 7.4077821306433035e-06, + "loss": 0.0032, + "step": 83330 + }, + { + "epoch": 12.945013979496737, + "grad_norm": 0.65687096118927, + "learning_rate": 7.399123667657465e-06, + "loss": 0.0052, + "step": 83340 + }, + { + "epoch": 12.946567256912084, + "grad_norm": 0.3422023355960846, + "learning_rate": 7.390469863439426e-06, + "loss": 0.0029, + "step": 83350 + }, + { + "epoch": 12.948120534327431, + "grad_norm": 0.1456567347049713, + "learning_rate": 7.381820718935512e-06, + "loss": 0.0043, + "step": 83360 + }, + { + "epoch": 12.949673811742777, + "grad_norm": 0.15881691873073578, + "learning_rate": 7.373176235091633e-06, + "loss": 0.0032, + "step": 83370 + }, + { + "epoch": 12.951227089158124, + "grad_norm": 0.24140207469463348, + "learning_rate": 7.364536412853101e-06, + "loss": 0.0032, + "step": 83380 + }, + { + "epoch": 12.95278036657347, + "grad_norm": 0.13913024961948395, + "learning_rate": 7.355901253164776e-06, + "loss": 0.0059, + "step": 83390 + }, + { + "epoch": 12.954333643988816, + "grad_norm": 0.16691753268241882, + "learning_rate": 7.34727075697097e-06, + "loss": 0.0026, + "step": 83400 + }, + { + "epoch": 12.955886921404163, + "grad_norm": 0.6538392305374146, + "learning_rate": 7.3386449252155095e-06, + "loss": 0.0041, + "step": 83410 + }, + { + "epoch": 12.957440198819508, + "grad_norm": 0.2280614823102951, + "learning_rate": 7.330023758841692e-06, + "loss": 0.0024, + "step": 83420 + }, + { + "epoch": 12.958993476234856, + "grad_norm": 0.2863432765007019, + "learning_rate": 7.321407258792323e-06, + "loss": 0.0032, + "step": 83430 + }, + { + "epoch": 12.960546753650203, + "grad_norm": 0.44679689407348633, + "learning_rate": 7.312795426009694e-06, + "loss": 0.0032, + "step": 83440 + }, + { + "epoch": 12.962100031065548, + "grad_norm": 0.15974675118923187, + "learning_rate": 7.304188261435585e-06, + "loss": 0.0034, + "step": 83450 + }, + { + "epoch": 12.963653308480895, + "grad_norm": 0.15404212474822998, + "learning_rate": 7.295585766011242e-06, + "loss": 0.0034, + "step": 83460 + }, + { + "epoch": 12.96520658589624, + "grad_norm": 0.7833622097969055, + "learning_rate": 7.286987940677436e-06, + "loss": 0.0044, + "step": 83470 + }, + { + "epoch": 12.966759863311587, + "grad_norm": 0.3861064016819, + "learning_rate": 7.278394786374426e-06, + "loss": 0.0051, + "step": 83480 + }, + { + "epoch": 12.968313140726934, + "grad_norm": 0.10153409093618393, + "learning_rate": 7.269806304041915e-06, + "loss": 0.0027, + "step": 83490 + }, + { + "epoch": 12.96986641814228, + "grad_norm": 0.2689985930919647, + "learning_rate": 7.261222494619147e-06, + "loss": 0.0046, + "step": 83500 + }, + { + "epoch": 12.971419695557627, + "grad_norm": 0.158586323261261, + "learning_rate": 7.252643359044825e-06, + "loss": 0.0037, + "step": 83510 + }, + { + "epoch": 12.972972972972974, + "grad_norm": 0.20886142551898956, + "learning_rate": 7.244068898257172e-06, + "loss": 0.0029, + "step": 83520 + }, + { + "epoch": 12.974526250388319, + "grad_norm": 0.08037198334932327, + "learning_rate": 7.2354991131938495e-06, + "loss": 0.0028, + "step": 83530 + }, + { + "epoch": 12.976079527803666, + "grad_norm": 0.43314671516418457, + "learning_rate": 7.2269340047920574e-06, + "loss": 0.007, + "step": 83540 + }, + { + "epoch": 12.977632805219011, + "grad_norm": 0.23820307850837708, + "learning_rate": 7.218373573988435e-06, + "loss": 0.0051, + "step": 83550 + }, + { + "epoch": 12.979186082634358, + "grad_norm": 0.20327678322792053, + "learning_rate": 7.209817821719178e-06, + "loss": 0.0032, + "step": 83560 + }, + { + "epoch": 12.980739360049705, + "grad_norm": 0.21261455118656158, + "learning_rate": 7.201266748919894e-06, + "loss": 0.0044, + "step": 83570 + }, + { + "epoch": 12.98229263746505, + "grad_norm": 0.30334556102752686, + "learning_rate": 7.19272035652575e-06, + "loss": 0.0034, + "step": 83580 + }, + { + "epoch": 12.983845914880398, + "grad_norm": 0.23061031103134155, + "learning_rate": 7.184178645471334e-06, + "loss": 0.0027, + "step": 83590 + }, + { + "epoch": 12.985399192295745, + "grad_norm": 0.9671372771263123, + "learning_rate": 7.175641616690775e-06, + "loss": 0.0045, + "step": 83600 + }, + { + "epoch": 12.98695246971109, + "grad_norm": 0.05658929422497749, + "learning_rate": 7.167109271117661e-06, + "loss": 0.0025, + "step": 83610 + }, + { + "epoch": 12.988505747126437, + "grad_norm": 0.09229499846696854, + "learning_rate": 7.158581609685072e-06, + "loss": 0.0029, + "step": 83620 + }, + { + "epoch": 12.990059024541782, + "grad_norm": 0.11052844673395157, + "learning_rate": 7.150058633325591e-06, + "loss": 0.0032, + "step": 83630 + }, + { + "epoch": 12.99161230195713, + "grad_norm": 0.10396695882081985, + "learning_rate": 7.141540342971287e-06, + "loss": 0.004, + "step": 83640 + }, + { + "epoch": 12.993165579372477, + "grad_norm": 0.483384370803833, + "learning_rate": 7.133026739553677e-06, + "loss": 0.0031, + "step": 83650 + }, + { + "epoch": 12.994718856787822, + "grad_norm": 0.21623723208904266, + "learning_rate": 7.124517824003829e-06, + "loss": 0.0043, + "step": 83660 + }, + { + "epoch": 12.996272134203169, + "grad_norm": 0.47476252913475037, + "learning_rate": 7.116013597252236e-06, + "loss": 0.0037, + "step": 83670 + }, + { + "epoch": 12.997825411618514, + "grad_norm": 0.19735200703144073, + "learning_rate": 7.107514060228921e-06, + "loss": 0.0029, + "step": 83680 + }, + { + "epoch": 12.999378689033861, + "grad_norm": 0.24340078234672546, + "learning_rate": 7.099019213863378e-06, + "loss": 0.0061, + "step": 83690 + }, + { + "epoch": 13.000931966449208, + "grad_norm": 0.0944773256778717, + "learning_rate": 7.090529059084611e-06, + "loss": 0.0039, + "step": 83700 + }, + { + "epoch": 13.002485243864554, + "grad_norm": 0.07545460760593414, + "learning_rate": 7.082043596821058e-06, + "loss": 0.0052, + "step": 83710 + }, + { + "epoch": 13.0040385212799, + "grad_norm": 0.25489428639411926, + "learning_rate": 7.07356282800069e-06, + "loss": 0.0045, + "step": 83720 + }, + { + "epoch": 13.005591798695248, + "grad_norm": 0.955245316028595, + "learning_rate": 7.0650867535509645e-06, + "loss": 0.0032, + "step": 83730 + }, + { + "epoch": 13.007145076110593, + "grad_norm": 0.3211326003074646, + "learning_rate": 7.0566153743987895e-06, + "loss": 0.0026, + "step": 83740 + }, + { + "epoch": 13.00869835352594, + "grad_norm": 0.10911476612091064, + "learning_rate": 7.04814869147059e-06, + "loss": 0.0044, + "step": 83750 + }, + { + "epoch": 13.010251630941285, + "grad_norm": 0.11355632543563843, + "learning_rate": 7.0396867056922725e-06, + "loss": 0.0044, + "step": 83760 + }, + { + "epoch": 13.011804908356632, + "grad_norm": 0.1815163642168045, + "learning_rate": 7.031229417989244e-06, + "loss": 0.003, + "step": 83770 + }, + { + "epoch": 13.01335818577198, + "grad_norm": 0.11823105067014694, + "learning_rate": 7.022776829286348e-06, + "loss": 0.002, + "step": 83780 + }, + { + "epoch": 13.014911463187325, + "grad_norm": 0.30388525128364563, + "learning_rate": 7.014328940507975e-06, + "loss": 0.0033, + "step": 83790 + }, + { + "epoch": 13.016464740602672, + "grad_norm": 0.3212294280529022, + "learning_rate": 7.005885752577945e-06, + "loss": 0.0038, + "step": 83800 + }, + { + "epoch": 13.018018018018019, + "grad_norm": 0.3180811405181885, + "learning_rate": 6.997447266419615e-06, + "loss": 0.0033, + "step": 83810 + }, + { + "epoch": 13.019571295433364, + "grad_norm": 0.563592255115509, + "learning_rate": 6.989013482955797e-06, + "loss": 0.0043, + "step": 83820 + }, + { + "epoch": 13.021124572848711, + "grad_norm": 0.9720098972320557, + "learning_rate": 6.980584403108808e-06, + "loss": 0.0025, + "step": 83830 + }, + { + "epoch": 13.022677850264056, + "grad_norm": 0.1183229386806488, + "learning_rate": 6.972160027800417e-06, + "loss": 0.0031, + "step": 83840 + }, + { + "epoch": 13.024231127679403, + "grad_norm": 0.35100093483924866, + "learning_rate": 6.963740357951931e-06, + "loss": 0.0034, + "step": 83850 + }, + { + "epoch": 13.02578440509475, + "grad_norm": 0.22198784351348877, + "learning_rate": 6.95532539448408e-06, + "loss": 0.0033, + "step": 83860 + }, + { + "epoch": 13.027337682510096, + "grad_norm": 0.1390935182571411, + "learning_rate": 6.946915138317128e-06, + "loss": 0.0029, + "step": 83870 + }, + { + "epoch": 13.028890959925443, + "grad_norm": 0.3466652035713196, + "learning_rate": 6.9385095903708065e-06, + "loss": 0.0058, + "step": 83880 + }, + { + "epoch": 13.03044423734079, + "grad_norm": 0.16672228276729584, + "learning_rate": 6.930108751564346e-06, + "loss": 0.004, + "step": 83890 + }, + { + "epoch": 13.031997514756135, + "grad_norm": 0.22346943616867065, + "learning_rate": 6.921712622816428e-06, + "loss": 0.005, + "step": 83900 + }, + { + "epoch": 13.033550792171482, + "grad_norm": 0.27110812067985535, + "learning_rate": 6.913321205045259e-06, + "loss": 0.0039, + "step": 83910 + }, + { + "epoch": 13.035104069586827, + "grad_norm": 0.21346008777618408, + "learning_rate": 6.904934499168491e-06, + "loss": 0.0052, + "step": 83920 + }, + { + "epoch": 13.036657347002174, + "grad_norm": 0.07129185646772385, + "learning_rate": 6.896552506103299e-06, + "loss": 0.0031, + "step": 83930 + }, + { + "epoch": 13.038210624417522, + "grad_norm": 0.40501677989959717, + "learning_rate": 6.888175226766313e-06, + "loss": 0.0077, + "step": 83940 + }, + { + "epoch": 13.039763901832867, + "grad_norm": 0.09281077980995178, + "learning_rate": 6.879802662073686e-06, + "loss": 0.0032, + "step": 83950 + }, + { + "epoch": 13.041317179248214, + "grad_norm": 0.21126431226730347, + "learning_rate": 6.871434812940991e-06, + "loss": 0.0025, + "step": 83960 + }, + { + "epoch": 13.042870456663561, + "grad_norm": 0.25379133224487305, + "learning_rate": 6.863071680283345e-06, + "loss": 0.0039, + "step": 83970 + }, + { + "epoch": 13.044423734078906, + "grad_norm": 0.21685895323753357, + "learning_rate": 6.854713265015339e-06, + "loss": 0.004, + "step": 83980 + }, + { + "epoch": 13.045977011494253, + "grad_norm": 0.2892158627510071, + "learning_rate": 6.8463595680510025e-06, + "loss": 0.0032, + "step": 83990 + }, + { + "epoch": 13.047530288909599, + "grad_norm": 0.1402921825647354, + "learning_rate": 6.838010590303911e-06, + "loss": 0.0043, + "step": 84000 + }, + { + "epoch": 13.049083566324946, + "grad_norm": 0.10956777632236481, + "learning_rate": 6.829666332687085e-06, + "loss": 0.0036, + "step": 84010 + }, + { + "epoch": 13.050636843740293, + "grad_norm": 0.1354426145553589, + "learning_rate": 6.821326796113059e-06, + "loss": 0.0048, + "step": 84020 + }, + { + "epoch": 13.052190121155638, + "grad_norm": 1.3232089281082153, + "learning_rate": 6.812991981493799e-06, + "loss": 0.0036, + "step": 84030 + }, + { + "epoch": 13.053743398570985, + "grad_norm": 0.5538829565048218, + "learning_rate": 6.804661889740821e-06, + "loss": 0.0036, + "step": 84040 + }, + { + "epoch": 13.055296675986332, + "grad_norm": 0.8820099234580994, + "learning_rate": 6.7963365217650625e-06, + "loss": 0.0027, + "step": 84050 + }, + { + "epoch": 13.056849953401677, + "grad_norm": 0.10856667906045914, + "learning_rate": 6.788015878476983e-06, + "loss": 0.0037, + "step": 84060 + }, + { + "epoch": 13.058403230817024, + "grad_norm": 0.8264832496643066, + "learning_rate": 6.779699960786523e-06, + "loss": 0.0048, + "step": 84070 + }, + { + "epoch": 13.05995650823237, + "grad_norm": 0.4333375096321106, + "learning_rate": 6.771388769603104e-06, + "loss": 0.005, + "step": 84080 + }, + { + "epoch": 13.061509785647717, + "grad_norm": 0.14531154930591583, + "learning_rate": 6.763082305835605e-06, + "loss": 0.0032, + "step": 84090 + }, + { + "epoch": 13.063063063063064, + "grad_norm": 0.10985038429498672, + "learning_rate": 6.754780570392433e-06, + "loss": 0.0022, + "step": 84100 + }, + { + "epoch": 13.064616340478409, + "grad_norm": 0.18631774187088013, + "learning_rate": 6.7464835641814215e-06, + "loss": 0.0039, + "step": 84110 + }, + { + "epoch": 13.066169617893756, + "grad_norm": 0.3639563024044037, + "learning_rate": 6.7381912881099515e-06, + "loss": 0.0043, + "step": 84120 + }, + { + "epoch": 13.067722895309101, + "grad_norm": 0.22480610013008118, + "learning_rate": 6.72990374308482e-06, + "loss": 0.003, + "step": 84130 + }, + { + "epoch": 13.069276172724448, + "grad_norm": 1.2447949647903442, + "learning_rate": 6.721620930012379e-06, + "loss": 0.0043, + "step": 84140 + }, + { + "epoch": 13.070829450139795, + "grad_norm": 0.21138887107372284, + "learning_rate": 6.713342849798393e-06, + "loss": 0.0055, + "step": 84150 + }, + { + "epoch": 13.07238272755514, + "grad_norm": 0.36446234583854675, + "learning_rate": 6.705069503348166e-06, + "loss": 0.0052, + "step": 84160 + }, + { + "epoch": 13.073936004970488, + "grad_norm": 0.3073952794075012, + "learning_rate": 6.696800891566435e-06, + "loss": 0.003, + "step": 84170 + }, + { + "epoch": 13.075489282385835, + "grad_norm": 0.4412481188774109, + "learning_rate": 6.688537015357449e-06, + "loss": 0.0035, + "step": 84180 + }, + { + "epoch": 13.07704255980118, + "grad_norm": 0.06483007222414017, + "learning_rate": 6.680277875624941e-06, + "loss": 0.0035, + "step": 84190 + }, + { + "epoch": 13.078595837216527, + "grad_norm": 0.2841991186141968, + "learning_rate": 6.672023473272121e-06, + "loss": 0.0024, + "step": 84200 + }, + { + "epoch": 13.080149114631872, + "grad_norm": 0.08100730925798416, + "learning_rate": 6.663773809201667e-06, + "loss": 0.0031, + "step": 84210 + }, + { + "epoch": 13.08170239204722, + "grad_norm": 0.22499114274978638, + "learning_rate": 6.655528884315748e-06, + "loss": 0.0029, + "step": 84220 + }, + { + "epoch": 13.083255669462567, + "grad_norm": 0.3963281214237213, + "learning_rate": 6.647288699516036e-06, + "loss": 0.0025, + "step": 84230 + }, + { + "epoch": 13.084808946877912, + "grad_norm": 0.07727702707052231, + "learning_rate": 6.639053255703642e-06, + "loss": 0.0035, + "step": 84240 + }, + { + "epoch": 13.086362224293259, + "grad_norm": 0.0744708701968193, + "learning_rate": 6.6308225537791925e-06, + "loss": 0.0023, + "step": 84250 + }, + { + "epoch": 13.087915501708606, + "grad_norm": 0.6965486407279968, + "learning_rate": 6.622596594642777e-06, + "loss": 0.0035, + "step": 84260 + }, + { + "epoch": 13.089468779123951, + "grad_norm": 0.3036384582519531, + "learning_rate": 6.614375379193999e-06, + "loss": 0.0028, + "step": 84270 + }, + { + "epoch": 13.091022056539298, + "grad_norm": 0.10725299268960953, + "learning_rate": 6.606158908331889e-06, + "loss": 0.0034, + "step": 84280 + }, + { + "epoch": 13.092575333954644, + "grad_norm": 0.13260656595230103, + "learning_rate": 6.597947182955005e-06, + "loss": 0.0038, + "step": 84290 + }, + { + "epoch": 13.09412861136999, + "grad_norm": 0.2850256860256195, + "learning_rate": 6.589740203961359e-06, + "loss": 0.003, + "step": 84300 + }, + { + "epoch": 13.095681888785338, + "grad_norm": 0.09179051220417023, + "learning_rate": 6.581537972248464e-06, + "loss": 0.0036, + "step": 84310 + }, + { + "epoch": 13.097235166200683, + "grad_norm": 0.20097309350967407, + "learning_rate": 6.573340488713275e-06, + "loss": 0.0038, + "step": 84320 + }, + { + "epoch": 13.09878844361603, + "grad_norm": 0.08633670210838318, + "learning_rate": 6.565147754252304e-06, + "loss": 0.0036, + "step": 84330 + }, + { + "epoch": 13.100341721031377, + "grad_norm": 0.733285665512085, + "learning_rate": 6.556959769761462e-06, + "loss": 0.0049, + "step": 84340 + }, + { + "epoch": 13.101894998446722, + "grad_norm": 0.26854196190834045, + "learning_rate": 6.548776536136192e-06, + "loss": 0.0047, + "step": 84350 + }, + { + "epoch": 13.10344827586207, + "grad_norm": 0.38128000497817993, + "learning_rate": 6.54059805427138e-06, + "loss": 0.0037, + "step": 84360 + }, + { + "epoch": 13.105001553277415, + "grad_norm": 0.07515225559473038, + "learning_rate": 6.532424325061437e-06, + "loss": 0.0037, + "step": 84370 + }, + { + "epoch": 13.106554830692762, + "grad_norm": 0.11748513579368591, + "learning_rate": 6.5242553494001935e-06, + "loss": 0.0038, + "step": 84380 + }, + { + "epoch": 13.108108108108109, + "grad_norm": 0.4407905638217926, + "learning_rate": 6.516091128181046e-06, + "loss": 0.0028, + "step": 84390 + }, + { + "epoch": 13.109661385523454, + "grad_norm": 0.15374131500720978, + "learning_rate": 6.507931662296785e-06, + "loss": 0.002, + "step": 84400 + }, + { + "epoch": 13.111214662938801, + "grad_norm": 0.20332986116409302, + "learning_rate": 6.4997769526397354e-06, + "loss": 0.0033, + "step": 84410 + }, + { + "epoch": 13.112767940354146, + "grad_norm": 0.2199648916721344, + "learning_rate": 6.491627000101663e-06, + "loss": 0.003, + "step": 84420 + }, + { + "epoch": 13.114321217769493, + "grad_norm": 0.6175654530525208, + "learning_rate": 6.4834818055738545e-06, + "loss": 0.0035, + "step": 84430 + }, + { + "epoch": 13.11587449518484, + "grad_norm": 0.5094794631004333, + "learning_rate": 6.4753413699470465e-06, + "loss": 0.0044, + "step": 84440 + }, + { + "epoch": 13.117427772600186, + "grad_norm": 0.5454263687133789, + "learning_rate": 6.467205694111478e-06, + "loss": 0.0021, + "step": 84450 + }, + { + "epoch": 13.118981050015533, + "grad_norm": 0.11880569159984589, + "learning_rate": 6.459074778956836e-06, + "loss": 0.0016, + "step": 84460 + }, + { + "epoch": 13.12053432743088, + "grad_norm": 0.17585380375385284, + "learning_rate": 6.450948625372311e-06, + "loss": 0.0046, + "step": 84470 + }, + { + "epoch": 13.122087604846225, + "grad_norm": 0.7426710724830627, + "learning_rate": 6.442827234246585e-06, + "loss": 0.0025, + "step": 84480 + }, + { + "epoch": 13.123640882261572, + "grad_norm": 0.9535205364227295, + "learning_rate": 6.43471060646777e-06, + "loss": 0.0045, + "step": 84490 + }, + { + "epoch": 13.125194159676917, + "grad_norm": 0.5982680320739746, + "learning_rate": 6.426598742923512e-06, + "loss": 0.0034, + "step": 84500 + }, + { + "epoch": 13.126747437092265, + "grad_norm": 0.11809143424034119, + "learning_rate": 6.418491644500885e-06, + "loss": 0.0046, + "step": 84510 + }, + { + "epoch": 13.128300714507612, + "grad_norm": 0.2204229086637497, + "learning_rate": 6.410389312086512e-06, + "loss": 0.0025, + "step": 84520 + }, + { + "epoch": 13.129853991922957, + "grad_norm": 0.9471418857574463, + "learning_rate": 6.4022917465664135e-06, + "loss": 0.0033, + "step": 84530 + }, + { + "epoch": 13.131407269338304, + "grad_norm": 0.5173757672309875, + "learning_rate": 6.3941989488261575e-06, + "loss": 0.0034, + "step": 84540 + }, + { + "epoch": 13.132960546753651, + "grad_norm": 0.34051838517189026, + "learning_rate": 6.386110919750726e-06, + "loss": 0.0042, + "step": 84550 + }, + { + "epoch": 13.134513824168996, + "grad_norm": 0.1569308042526245, + "learning_rate": 6.378027660224651e-06, + "loss": 0.0037, + "step": 84560 + }, + { + "epoch": 13.136067101584343, + "grad_norm": 0.09186914563179016, + "learning_rate": 6.369949171131856e-06, + "loss": 0.0023, + "step": 84570 + }, + { + "epoch": 13.137620378999689, + "grad_norm": 0.32118475437164307, + "learning_rate": 6.361875453355848e-06, + "loss": 0.0052, + "step": 84580 + }, + { + "epoch": 13.139173656415036, + "grad_norm": 0.45316067337989807, + "learning_rate": 6.353806507779525e-06, + "loss": 0.0041, + "step": 84590 + }, + { + "epoch": 13.140726933830383, + "grad_norm": 0.20160260796546936, + "learning_rate": 6.3457423352853044e-06, + "loss": 0.0041, + "step": 84600 + }, + { + "epoch": 13.142280211245728, + "grad_norm": 1.089206576347351, + "learning_rate": 6.337682936755062e-06, + "loss": 0.0028, + "step": 84610 + }, + { + "epoch": 13.143833488661075, + "grad_norm": 0.08994436264038086, + "learning_rate": 6.329628313070179e-06, + "loss": 0.0026, + "step": 84620 + }, + { + "epoch": 13.145386766076422, + "grad_norm": 0.12921936810016632, + "learning_rate": 6.321578465111477e-06, + "loss": 0.0042, + "step": 84630 + }, + { + "epoch": 13.146940043491767, + "grad_norm": 0.1998635232448578, + "learning_rate": 6.31353339375928e-06, + "loss": 0.0039, + "step": 84640 + }, + { + "epoch": 13.148493320907114, + "grad_norm": 0.4444003999233246, + "learning_rate": 6.305493099893395e-06, + "loss": 0.0034, + "step": 84650 + }, + { + "epoch": 13.15004659832246, + "grad_norm": 0.1444067806005478, + "learning_rate": 6.297457584393096e-06, + "loss": 0.0041, + "step": 84660 + }, + { + "epoch": 13.151599875737807, + "grad_norm": 0.23666463792324066, + "learning_rate": 6.289426848137126e-06, + "loss": 0.005, + "step": 84670 + }, + { + "epoch": 13.153153153153154, + "grad_norm": 0.3109318017959595, + "learning_rate": 6.281400892003713e-06, + "loss": 0.0037, + "step": 84680 + }, + { + "epoch": 13.154706430568499, + "grad_norm": 0.09504018723964691, + "learning_rate": 6.273379716870581e-06, + "loss": 0.0022, + "step": 84690 + }, + { + "epoch": 13.156259707983846, + "grad_norm": 0.07789898663759232, + "learning_rate": 6.265363323614887e-06, + "loss": 0.0042, + "step": 84700 + }, + { + "epoch": 13.157812985399193, + "grad_norm": 0.252877801656723, + "learning_rate": 6.257351713113308e-06, + "loss": 0.0044, + "step": 84710 + }, + { + "epoch": 13.159366262814538, + "grad_norm": 0.18291419744491577, + "learning_rate": 6.249344886241981e-06, + "loss": 0.0027, + "step": 84720 + }, + { + "epoch": 13.160919540229886, + "grad_norm": 0.2703639566898346, + "learning_rate": 6.2413428438765295e-06, + "loss": 0.0049, + "step": 84730 + }, + { + "epoch": 13.16247281764523, + "grad_norm": 0.4255599081516266, + "learning_rate": 6.233345586892026e-06, + "loss": 0.0043, + "step": 84740 + }, + { + "epoch": 13.164026095060578, + "grad_norm": 0.23633955419063568, + "learning_rate": 6.225353116163052e-06, + "loss": 0.004, + "step": 84750 + }, + { + "epoch": 13.165579372475925, + "grad_norm": 0.6814929246902466, + "learning_rate": 6.217365432563627e-06, + "loss": 0.0056, + "step": 84760 + }, + { + "epoch": 13.16713264989127, + "grad_norm": 0.22051525115966797, + "learning_rate": 6.209382536967318e-06, + "loss": 0.0039, + "step": 84770 + }, + { + "epoch": 13.168685927306617, + "grad_norm": 0.13152168691158295, + "learning_rate": 6.201404430247082e-06, + "loss": 0.0026, + "step": 84780 + }, + { + "epoch": 13.170239204721963, + "grad_norm": 0.18647998571395874, + "learning_rate": 6.1934311132754165e-06, + "loss": 0.0046, + "step": 84790 + }, + { + "epoch": 13.17179248213731, + "grad_norm": 0.06063205376267433, + "learning_rate": 6.185462586924257e-06, + "loss": 0.0025, + "step": 84800 + }, + { + "epoch": 13.173345759552657, + "grad_norm": 0.13453581929206848, + "learning_rate": 6.177498852065039e-06, + "loss": 0.0055, + "step": 84810 + }, + { + "epoch": 13.174899036968002, + "grad_norm": 0.1758534014225006, + "learning_rate": 6.169539909568655e-06, + "loss": 0.0034, + "step": 84820 + }, + { + "epoch": 13.176452314383349, + "grad_norm": 0.5986335873603821, + "learning_rate": 6.161585760305488e-06, + "loss": 0.003, + "step": 84830 + }, + { + "epoch": 13.178005591798696, + "grad_norm": 0.2640782594680786, + "learning_rate": 6.153636405145391e-06, + "loss": 0.0046, + "step": 84840 + }, + { + "epoch": 13.179558869214041, + "grad_norm": 0.3563140034675598, + "learning_rate": 6.145691844957702e-06, + "loss": 0.0042, + "step": 84850 + }, + { + "epoch": 13.181112146629388, + "grad_norm": 0.2718847393989563, + "learning_rate": 6.1377520806112145e-06, + "loss": 0.0036, + "step": 84860 + }, + { + "epoch": 13.182665424044734, + "grad_norm": 0.13210555911064148, + "learning_rate": 6.129817112974218e-06, + "loss": 0.0032, + "step": 84870 + }, + { + "epoch": 13.18421870146008, + "grad_norm": 0.22894302010536194, + "learning_rate": 6.121886942914457e-06, + "loss": 0.0042, + "step": 84880 + }, + { + "epoch": 13.185771978875428, + "grad_norm": 0.6171071529388428, + "learning_rate": 6.11396157129917e-06, + "loss": 0.0037, + "step": 84890 + }, + { + "epoch": 13.187325256290773, + "grad_norm": 0.4459966719150543, + "learning_rate": 6.106040998995066e-06, + "loss": 0.0038, + "step": 84900 + }, + { + "epoch": 13.18887853370612, + "grad_norm": 0.14908136427402496, + "learning_rate": 6.098125226868329e-06, + "loss": 0.004, + "step": 84910 + }, + { + "epoch": 13.190431811121467, + "grad_norm": 0.2971910536289215, + "learning_rate": 6.090214255784604e-06, + "loss": 0.0031, + "step": 84920 + }, + { + "epoch": 13.191985088536812, + "grad_norm": 0.1256866753101349, + "learning_rate": 6.0823080866090346e-06, + "loss": 0.0027, + "step": 84930 + }, + { + "epoch": 13.19353836595216, + "grad_norm": 0.3141131103038788, + "learning_rate": 6.074406720206233e-06, + "loss": 0.0032, + "step": 84940 + }, + { + "epoch": 13.195091643367505, + "grad_norm": 0.3348129689693451, + "learning_rate": 6.066510157440259e-06, + "loss": 0.0032, + "step": 84950 + }, + { + "epoch": 13.196644920782852, + "grad_norm": 0.41858217120170593, + "learning_rate": 6.058618399174676e-06, + "loss": 0.0028, + "step": 84960 + }, + { + "epoch": 13.198198198198199, + "grad_norm": 0.07764916121959686, + "learning_rate": 6.050731446272528e-06, + "loss": 0.0026, + "step": 84970 + }, + { + "epoch": 13.199751475613544, + "grad_norm": 0.1624676138162613, + "learning_rate": 6.042849299596321e-06, + "loss": 0.0036, + "step": 84980 + }, + { + "epoch": 13.201304753028891, + "grad_norm": 0.9255379438400269, + "learning_rate": 6.034971960008013e-06, + "loss": 0.0044, + "step": 84990 + }, + { + "epoch": 13.202858030444236, + "grad_norm": 0.1593611240386963, + "learning_rate": 6.027099428369082e-06, + "loss": 0.0028, + "step": 85000 + }, + { + "epoch": 13.204411307859584, + "grad_norm": 0.8360854983329773, + "learning_rate": 6.019231705540435e-06, + "loss": 0.0032, + "step": 85010 + }, + { + "epoch": 13.20596458527493, + "grad_norm": 0.1634587049484253, + "learning_rate": 6.011368792382488e-06, + "loss": 0.0034, + "step": 85020 + }, + { + "epoch": 13.207517862690276, + "grad_norm": 0.05609651282429695, + "learning_rate": 6.00351068975511e-06, + "loss": 0.004, + "step": 85030 + }, + { + "epoch": 13.209071140105623, + "grad_norm": 0.12037114053964615, + "learning_rate": 5.9956573985176674e-06, + "loss": 0.0037, + "step": 85040 + }, + { + "epoch": 13.21062441752097, + "grad_norm": 0.14173585176467896, + "learning_rate": 5.987808919528959e-06, + "loss": 0.0054, + "step": 85050 + }, + { + "epoch": 13.212177694936315, + "grad_norm": 0.23739658296108246, + "learning_rate": 5.979965253647307e-06, + "loss": 0.0027, + "step": 85060 + }, + { + "epoch": 13.213730972351662, + "grad_norm": 0.10539974272251129, + "learning_rate": 5.9721264017304655e-06, + "loss": 0.0036, + "step": 85070 + }, + { + "epoch": 13.215284249767008, + "grad_norm": 0.15515649318695068, + "learning_rate": 5.964292364635682e-06, + "loss": 0.0024, + "step": 85080 + }, + { + "epoch": 13.216837527182355, + "grad_norm": 0.1802852749824524, + "learning_rate": 5.956463143219682e-06, + "loss": 0.0031, + "step": 85090 + }, + { + "epoch": 13.218390804597702, + "grad_norm": 0.2389083057641983, + "learning_rate": 5.948638738338663e-06, + "loss": 0.0032, + "step": 85100 + }, + { + "epoch": 13.219944082013047, + "grad_norm": 0.5299113988876343, + "learning_rate": 5.940819150848271e-06, + "loss": 0.0042, + "step": 85110 + }, + { + "epoch": 13.221497359428394, + "grad_norm": 0.2523755729198456, + "learning_rate": 5.933004381603663e-06, + "loss": 0.0037, + "step": 85120 + }, + { + "epoch": 13.223050636843741, + "grad_norm": 0.9535553455352783, + "learning_rate": 5.92519443145943e-06, + "loss": 0.003, + "step": 85130 + }, + { + "epoch": 13.224603914259086, + "grad_norm": 0.14961746335029602, + "learning_rate": 5.917389301269671e-06, + "loss": 0.0044, + "step": 85140 + }, + { + "epoch": 13.226157191674433, + "grad_norm": 0.19852180778980255, + "learning_rate": 5.9095889918879435e-06, + "loss": 0.0034, + "step": 85150 + }, + { + "epoch": 13.227710469089779, + "grad_norm": 0.2567481994628906, + "learning_rate": 5.901793504167286e-06, + "loss": 0.0033, + "step": 85160 + }, + { + "epoch": 13.229263746505126, + "grad_norm": 0.19048595428466797, + "learning_rate": 5.894002838960172e-06, + "loss": 0.0046, + "step": 85170 + }, + { + "epoch": 13.230817023920473, + "grad_norm": 0.24323654174804688, + "learning_rate": 5.886216997118604e-06, + "loss": 0.003, + "step": 85180 + }, + { + "epoch": 13.232370301335818, + "grad_norm": 0.10905881971120834, + "learning_rate": 5.878435979494023e-06, + "loss": 0.0022, + "step": 85190 + }, + { + "epoch": 13.233923578751165, + "grad_norm": 0.3815138638019562, + "learning_rate": 5.870659786937344e-06, + "loss": 0.0058, + "step": 85200 + }, + { + "epoch": 13.235476856166512, + "grad_norm": 0.1507059931755066, + "learning_rate": 5.862888420298962e-06, + "loss": 0.0037, + "step": 85210 + }, + { + "epoch": 13.237030133581857, + "grad_norm": 0.06095986068248749, + "learning_rate": 5.855121880428743e-06, + "loss": 0.0031, + "step": 85220 + }, + { + "epoch": 13.238583410997204, + "grad_norm": 0.2041425108909607, + "learning_rate": 5.84736016817603e-06, + "loss": 0.003, + "step": 85230 + }, + { + "epoch": 13.24013668841255, + "grad_norm": 0.05491364374756813, + "learning_rate": 5.839603284389622e-06, + "loss": 0.0027, + "step": 85240 + }, + { + "epoch": 13.241689965827897, + "grad_norm": 0.313003271818161, + "learning_rate": 5.831851229917812e-06, + "loss": 0.0036, + "step": 85250 + }, + { + "epoch": 13.243243243243244, + "grad_norm": 0.16525554656982422, + "learning_rate": 5.824104005608338e-06, + "loss": 0.0037, + "step": 85260 + }, + { + "epoch": 13.24479652065859, + "grad_norm": 0.5405550003051758, + "learning_rate": 5.81636161230843e-06, + "loss": 0.0042, + "step": 85270 + }, + { + "epoch": 13.246349798073936, + "grad_norm": 0.05564845725893974, + "learning_rate": 5.808624050864791e-06, + "loss": 0.0031, + "step": 85280 + }, + { + "epoch": 13.247903075489283, + "grad_norm": 0.1573687493801117, + "learning_rate": 5.800891322123592e-06, + "loss": 0.0043, + "step": 85290 + }, + { + "epoch": 13.249456352904629, + "grad_norm": 0.28969067335128784, + "learning_rate": 5.793163426930454e-06, + "loss": 0.0035, + "step": 85300 + }, + { + "epoch": 13.251009630319976, + "grad_norm": 0.348210871219635, + "learning_rate": 5.785440366130512e-06, + "loss": 0.0027, + "step": 85310 + }, + { + "epoch": 13.25256290773532, + "grad_norm": 0.09776909649372101, + "learning_rate": 5.777722140568326e-06, + "loss": 0.0076, + "step": 85320 + }, + { + "epoch": 13.254116185150668, + "grad_norm": 0.48790544271469116, + "learning_rate": 5.7700087510879685e-06, + "loss": 0.0042, + "step": 85330 + }, + { + "epoch": 13.255669462566015, + "grad_norm": 0.20497293770313263, + "learning_rate": 5.762300198532933e-06, + "loss": 0.0029, + "step": 85340 + }, + { + "epoch": 13.25722273998136, + "grad_norm": 0.4769412577152252, + "learning_rate": 5.7545964837462606e-06, + "loss": 0.0038, + "step": 85350 + }, + { + "epoch": 13.258776017396707, + "grad_norm": 0.134464830160141, + "learning_rate": 5.746897607570384e-06, + "loss": 0.0027, + "step": 85360 + }, + { + "epoch": 13.260329294812053, + "grad_norm": 0.19741976261138916, + "learning_rate": 5.7392035708472626e-06, + "loss": 0.0043, + "step": 85370 + }, + { + "epoch": 13.2618825722274, + "grad_norm": 0.9018554091453552, + "learning_rate": 5.731514374418278e-06, + "loss": 0.0043, + "step": 85380 + }, + { + "epoch": 13.263435849642747, + "grad_norm": 0.37566596269607544, + "learning_rate": 5.7238300191243325e-06, + "loss": 0.005, + "step": 85390 + }, + { + "epoch": 13.264989127058092, + "grad_norm": 0.8097374439239502, + "learning_rate": 5.716150505805762e-06, + "loss": 0.0032, + "step": 85400 + }, + { + "epoch": 13.266542404473439, + "grad_norm": 0.07495655864477158, + "learning_rate": 5.708475835302407e-06, + "loss": 0.0032, + "step": 85410 + }, + { + "epoch": 13.268095681888786, + "grad_norm": 1.5846704244613647, + "learning_rate": 5.70080600845353e-06, + "loss": 0.0033, + "step": 85420 + }, + { + "epoch": 13.269648959304131, + "grad_norm": 0.18753856420516968, + "learning_rate": 5.693141026097909e-06, + "loss": 0.003, + "step": 85430 + }, + { + "epoch": 13.271202236719478, + "grad_norm": 0.30306270718574524, + "learning_rate": 5.6854808890737745e-06, + "loss": 0.0038, + "step": 85440 + }, + { + "epoch": 13.272755514134824, + "grad_norm": 0.2767413258552551, + "learning_rate": 5.6778255982188225e-06, + "loss": 0.004, + "step": 85450 + }, + { + "epoch": 13.27430879155017, + "grad_norm": 0.09737075865268707, + "learning_rate": 5.670175154370217e-06, + "loss": 0.004, + "step": 85460 + }, + { + "epoch": 13.275862068965518, + "grad_norm": 0.6474468111991882, + "learning_rate": 5.662529558364616e-06, + "loss": 0.0028, + "step": 85470 + }, + { + "epoch": 13.277415346380863, + "grad_norm": 0.09166873246431351, + "learning_rate": 5.654888811038128e-06, + "loss": 0.0029, + "step": 85480 + }, + { + "epoch": 13.27896862379621, + "grad_norm": 0.185664102435112, + "learning_rate": 5.647252913226314e-06, + "loss": 0.0038, + "step": 85490 + }, + { + "epoch": 13.280521901211557, + "grad_norm": 0.200959712266922, + "learning_rate": 5.639621865764255e-06, + "loss": 0.005, + "step": 85500 + }, + { + "epoch": 13.282075178626902, + "grad_norm": 0.24699072539806366, + "learning_rate": 5.63199566948644e-06, + "loss": 0.0029, + "step": 85510 + }, + { + "epoch": 13.28362845604225, + "grad_norm": 0.8784429430961609, + "learning_rate": 5.6243743252268835e-06, + "loss": 0.0026, + "step": 85520 + }, + { + "epoch": 13.285181733457595, + "grad_norm": 0.09656670689582825, + "learning_rate": 5.616757833819009e-06, + "loss": 0.003, + "step": 85530 + }, + { + "epoch": 13.286735010872942, + "grad_norm": 0.2463567852973938, + "learning_rate": 5.609146196095794e-06, + "loss": 0.0033, + "step": 85540 + }, + { + "epoch": 13.288288288288289, + "grad_norm": 0.03867262974381447, + "learning_rate": 5.6015394128896005e-06, + "loss": 0.0033, + "step": 85550 + }, + { + "epoch": 13.289841565703634, + "grad_norm": 0.3726813495159149, + "learning_rate": 5.593937485032308e-06, + "loss": 0.0049, + "step": 85560 + }, + { + "epoch": 13.291394843118981, + "grad_norm": 0.4562947154045105, + "learning_rate": 5.586340413355246e-06, + "loss": 0.004, + "step": 85570 + }, + { + "epoch": 13.292948120534328, + "grad_norm": 0.24369145929813385, + "learning_rate": 5.578748198689227e-06, + "loss": 0.0027, + "step": 85580 + }, + { + "epoch": 13.294501397949674, + "grad_norm": 0.13106031715869904, + "learning_rate": 5.571160841864504e-06, + "loss": 0.0043, + "step": 85590 + }, + { + "epoch": 13.29605467536502, + "grad_norm": 0.23895679414272308, + "learning_rate": 5.563578343710846e-06, + "loss": 0.0033, + "step": 85600 + }, + { + "epoch": 13.297607952780366, + "grad_norm": 0.22872132062911987, + "learning_rate": 5.556000705057446e-06, + "loss": 0.0025, + "step": 85610 + }, + { + "epoch": 13.299161230195713, + "grad_norm": 0.5548662543296814, + "learning_rate": 5.548427926732996e-06, + "loss": 0.0043, + "step": 85620 + }, + { + "epoch": 13.30071450761106, + "grad_norm": 0.36192724108695984, + "learning_rate": 5.540860009565629e-06, + "loss": 0.005, + "step": 85630 + }, + { + "epoch": 13.302267785026405, + "grad_norm": 0.3020627796649933, + "learning_rate": 5.533296954382966e-06, + "loss": 0.004, + "step": 85640 + }, + { + "epoch": 13.303821062441752, + "grad_norm": 0.20324598252773285, + "learning_rate": 5.525738762012095e-06, + "loss": 0.0025, + "step": 85650 + }, + { + "epoch": 13.3053743398571, + "grad_norm": 0.09150432795286179, + "learning_rate": 5.518185433279577e-06, + "loss": 0.0069, + "step": 85660 + }, + { + "epoch": 13.306927617272445, + "grad_norm": 0.3104618787765503, + "learning_rate": 5.510636969011407e-06, + "loss": 0.0045, + "step": 85670 + }, + { + "epoch": 13.308480894687792, + "grad_norm": 0.3947426378726959, + "learning_rate": 5.503093370033096e-06, + "loss": 0.0035, + "step": 85680 + }, + { + "epoch": 13.310034172103137, + "grad_norm": 0.20932133495807648, + "learning_rate": 5.495554637169603e-06, + "loss": 0.0033, + "step": 85690 + }, + { + "epoch": 13.311587449518484, + "grad_norm": 0.4200102686882019, + "learning_rate": 5.488020771245328e-06, + "loss": 0.0043, + "step": 85700 + }, + { + "epoch": 13.313140726933831, + "grad_norm": 0.11065024137496948, + "learning_rate": 5.480491773084195e-06, + "loss": 0.0035, + "step": 85710 + }, + { + "epoch": 13.314694004349176, + "grad_norm": 0.8678122758865356, + "learning_rate": 5.472967643509525e-06, + "loss": 0.0033, + "step": 85720 + }, + { + "epoch": 13.316247281764523, + "grad_norm": 0.3614940047264099, + "learning_rate": 5.46544838334419e-06, + "loss": 0.0046, + "step": 85730 + }, + { + "epoch": 13.317800559179869, + "grad_norm": 0.3895358741283417, + "learning_rate": 5.4579339934104525e-06, + "loss": 0.0029, + "step": 85740 + }, + { + "epoch": 13.319353836595216, + "grad_norm": 0.08776775747537613, + "learning_rate": 5.450424474530091e-06, + "loss": 0.0031, + "step": 85750 + }, + { + "epoch": 13.320907114010563, + "grad_norm": 0.20701579749584198, + "learning_rate": 5.442919827524323e-06, + "loss": 0.0046, + "step": 85760 + }, + { + "epoch": 13.322460391425908, + "grad_norm": 0.27942219376564026, + "learning_rate": 5.435420053213863e-06, + "loss": 0.0022, + "step": 85770 + }, + { + "epoch": 13.324013668841255, + "grad_norm": 0.10542292892932892, + "learning_rate": 5.427925152418845e-06, + "loss": 0.0036, + "step": 85780 + }, + { + "epoch": 13.325566946256602, + "grad_norm": 0.35287514328956604, + "learning_rate": 5.420435125958934e-06, + "loss": 0.0052, + "step": 85790 + }, + { + "epoch": 13.327120223671947, + "grad_norm": 0.21757741272449493, + "learning_rate": 5.412949974653209e-06, + "loss": 0.004, + "step": 85800 + }, + { + "epoch": 13.328673501087295, + "grad_norm": 0.16911427676677704, + "learning_rate": 5.405469699320248e-06, + "loss": 0.0036, + "step": 85810 + }, + { + "epoch": 13.33022677850264, + "grad_norm": 0.16022755205631256, + "learning_rate": 5.397994300778064e-06, + "loss": 0.0039, + "step": 85820 + }, + { + "epoch": 13.331780055917987, + "grad_norm": 0.11138159036636353, + "learning_rate": 5.39052377984417e-06, + "loss": 0.0053, + "step": 85830 + }, + { + "epoch": 13.333333333333334, + "grad_norm": 0.63075852394104, + "learning_rate": 5.383058137335523e-06, + "loss": 0.008, + "step": 85840 + }, + { + "epoch": 13.33488661074868, + "grad_norm": 0.3659897744655609, + "learning_rate": 5.375597374068553e-06, + "loss": 0.0033, + "step": 85850 + }, + { + "epoch": 13.336439888164026, + "grad_norm": 0.22725212574005127, + "learning_rate": 5.368141490859163e-06, + "loss": 0.0037, + "step": 85860 + }, + { + "epoch": 13.337993165579373, + "grad_norm": 0.4975545406341553, + "learning_rate": 5.360690488522729e-06, + "loss": 0.0038, + "step": 85870 + }, + { + "epoch": 13.339546442994719, + "grad_norm": 0.20353446900844574, + "learning_rate": 5.353244367874055e-06, + "loss": 0.0041, + "step": 85880 + }, + { + "epoch": 13.341099720410066, + "grad_norm": 0.11852649599313736, + "learning_rate": 5.345803129727456e-06, + "loss": 0.0032, + "step": 85890 + }, + { + "epoch": 13.342652997825411, + "grad_norm": 0.22124741971492767, + "learning_rate": 5.338366774896697e-06, + "loss": 0.0038, + "step": 85900 + }, + { + "epoch": 13.344206275240758, + "grad_norm": 0.22222386300563812, + "learning_rate": 5.330935304194995e-06, + "loss": 0.0031, + "step": 85910 + }, + { + "epoch": 13.345759552656105, + "grad_norm": 0.7393425107002258, + "learning_rate": 5.323508718435044e-06, + "loss": 0.0035, + "step": 85920 + }, + { + "epoch": 13.34731283007145, + "grad_norm": 0.2736526429653168, + "learning_rate": 5.316087018429011e-06, + "loss": 0.0044, + "step": 85930 + }, + { + "epoch": 13.348866107486797, + "grad_norm": 0.31413009762763977, + "learning_rate": 5.30867020498853e-06, + "loss": 0.0042, + "step": 85940 + }, + { + "epoch": 13.350419384902143, + "grad_norm": 0.33409833908081055, + "learning_rate": 5.30125827892467e-06, + "loss": 0.0028, + "step": 85950 + }, + { + "epoch": 13.35197266231749, + "grad_norm": 0.2234392762184143, + "learning_rate": 5.293851241048015e-06, + "loss": 0.0028, + "step": 85960 + }, + { + "epoch": 13.353525939732837, + "grad_norm": 0.1733299195766449, + "learning_rate": 5.286449092168555e-06, + "loss": 0.0037, + "step": 85970 + }, + { + "epoch": 13.355079217148182, + "grad_norm": 0.22162818908691406, + "learning_rate": 5.2790518330958105e-06, + "loss": 0.0031, + "step": 85980 + }, + { + "epoch": 13.356632494563529, + "grad_norm": 0.2693232595920563, + "learning_rate": 5.271659464638717e-06, + "loss": 0.0028, + "step": 85990 + }, + { + "epoch": 13.358185771978876, + "grad_norm": 0.0901266485452652, + "learning_rate": 5.264271987605701e-06, + "loss": 0.0051, + "step": 86000 + }, + { + "epoch": 13.359739049394221, + "grad_norm": 0.13924753665924072, + "learning_rate": 5.256889402804632e-06, + "loss": 0.0029, + "step": 86010 + }, + { + "epoch": 13.361292326809568, + "grad_norm": 0.2082546204328537, + "learning_rate": 5.249511711042881e-06, + "loss": 0.0036, + "step": 86020 + }, + { + "epoch": 13.362845604224914, + "grad_norm": 0.1784372627735138, + "learning_rate": 5.242138913127237e-06, + "loss": 0.0046, + "step": 86030 + }, + { + "epoch": 13.36439888164026, + "grad_norm": 0.18187890946865082, + "learning_rate": 5.234771009863987e-06, + "loss": 0.0031, + "step": 86040 + }, + { + "epoch": 13.365952159055608, + "grad_norm": 0.08973561972379684, + "learning_rate": 5.227408002058881e-06, + "loss": 0.0054, + "step": 86050 + }, + { + "epoch": 13.367505436470953, + "grad_norm": 0.2463931441307068, + "learning_rate": 5.220049890517126e-06, + "loss": 0.0044, + "step": 86060 + }, + { + "epoch": 13.3690587138863, + "grad_norm": 0.1807875782251358, + "learning_rate": 5.212696676043377e-06, + "loss": 0.0035, + "step": 86070 + }, + { + "epoch": 13.370611991301647, + "grad_norm": 0.3166561424732208, + "learning_rate": 5.205348359441798e-06, + "loss": 0.0035, + "step": 86080 + }, + { + "epoch": 13.372165268716993, + "grad_norm": 0.4769025444984436, + "learning_rate": 5.198004941515966e-06, + "loss": 0.0034, + "step": 86090 + }, + { + "epoch": 13.37371854613234, + "grad_norm": 0.5248228311538696, + "learning_rate": 5.190666423068952e-06, + "loss": 0.0042, + "step": 86100 + }, + { + "epoch": 13.375271823547685, + "grad_norm": 0.09291534870862961, + "learning_rate": 5.183332804903284e-06, + "loss": 0.0046, + "step": 86110 + }, + { + "epoch": 13.376825100963032, + "grad_norm": 0.5599253177642822, + "learning_rate": 5.176004087820979e-06, + "loss": 0.0034, + "step": 86120 + }, + { + "epoch": 13.378378378378379, + "grad_norm": 0.34279918670654297, + "learning_rate": 5.168680272623455e-06, + "loss": 0.0036, + "step": 86130 + }, + { + "epoch": 13.379931655793724, + "grad_norm": 0.2553570866584778, + "learning_rate": 5.161361360111661e-06, + "loss": 0.005, + "step": 86140 + }, + { + "epoch": 13.381484933209071, + "grad_norm": 0.4024793803691864, + "learning_rate": 5.154047351085984e-06, + "loss": 0.0039, + "step": 86150 + }, + { + "epoch": 13.383038210624418, + "grad_norm": 0.5390143990516663, + "learning_rate": 5.1467382463462385e-06, + "loss": 0.0043, + "step": 86160 + }, + { + "epoch": 13.384591488039764, + "grad_norm": 0.20850706100463867, + "learning_rate": 5.139434046691788e-06, + "loss": 0.0041, + "step": 86170 + }, + { + "epoch": 13.38614476545511, + "grad_norm": 0.20248188078403473, + "learning_rate": 5.13213475292137e-06, + "loss": 0.0053, + "step": 86180 + }, + { + "epoch": 13.387698042870456, + "grad_norm": 0.3273236155509949, + "learning_rate": 5.12484036583325e-06, + "loss": 0.0022, + "step": 86190 + }, + { + "epoch": 13.389251320285803, + "grad_norm": 0.24570295214653015, + "learning_rate": 5.117550886225109e-06, + "loss": 0.0034, + "step": 86200 + }, + { + "epoch": 13.39080459770115, + "grad_norm": 0.2546813189983368, + "learning_rate": 5.11026631489413e-06, + "loss": 0.0043, + "step": 86210 + }, + { + "epoch": 13.392357875116495, + "grad_norm": 0.2856784760951996, + "learning_rate": 5.102986652636926e-06, + "loss": 0.0029, + "step": 86220 + }, + { + "epoch": 13.393911152531842, + "grad_norm": 1.2382997274398804, + "learning_rate": 5.095711900249606e-06, + "loss": 0.0034, + "step": 86230 + }, + { + "epoch": 13.39546442994719, + "grad_norm": 0.2571060061454773, + "learning_rate": 5.088442058527715e-06, + "loss": 0.0026, + "step": 86240 + }, + { + "epoch": 13.397017707362535, + "grad_norm": 0.05561377480626106, + "learning_rate": 5.08117712826629e-06, + "loss": 0.0066, + "step": 86250 + }, + { + "epoch": 13.398570984777882, + "grad_norm": 0.19753730297088623, + "learning_rate": 5.073917110259785e-06, + "loss": 0.0054, + "step": 86260 + }, + { + "epoch": 13.400124262193227, + "grad_norm": 0.21613925695419312, + "learning_rate": 5.066662005302175e-06, + "loss": 0.0031, + "step": 86270 + }, + { + "epoch": 13.401677539608574, + "grad_norm": 0.15313854813575745, + "learning_rate": 5.059411814186838e-06, + "loss": 0.0041, + "step": 86280 + }, + { + "epoch": 13.403230817023921, + "grad_norm": 0.1318517029285431, + "learning_rate": 5.05216653770666e-06, + "loss": 0.0023, + "step": 86290 + }, + { + "epoch": 13.404784094439266, + "grad_norm": 0.22457122802734375, + "learning_rate": 5.044926176653969e-06, + "loss": 0.0052, + "step": 86300 + }, + { + "epoch": 13.406337371854613, + "grad_norm": 0.2315739095211029, + "learning_rate": 5.037690731820571e-06, + "loss": 0.0047, + "step": 86310 + }, + { + "epoch": 13.407890649269959, + "grad_norm": 0.49495500326156616, + "learning_rate": 5.030460203997711e-06, + "loss": 0.003, + "step": 86320 + }, + { + "epoch": 13.409443926685306, + "grad_norm": 0.2236814796924591, + "learning_rate": 5.0232345939761174e-06, + "loss": 0.0029, + "step": 86330 + }, + { + "epoch": 13.410997204100653, + "grad_norm": 0.4712522029876709, + "learning_rate": 5.016013902545957e-06, + "loss": 0.0047, + "step": 86340 + }, + { + "epoch": 13.412550481515998, + "grad_norm": 0.1562870293855667, + "learning_rate": 5.008798130496889e-06, + "loss": 0.0051, + "step": 86350 + }, + { + "epoch": 13.414103758931345, + "grad_norm": 0.26402756571769714, + "learning_rate": 5.001587278618008e-06, + "loss": 0.0048, + "step": 86360 + }, + { + "epoch": 13.415657036346692, + "grad_norm": 0.24963366985321045, + "learning_rate": 4.994381347697891e-06, + "loss": 0.0026, + "step": 86370 + }, + { + "epoch": 13.417210313762038, + "grad_norm": 0.7043657302856445, + "learning_rate": 4.987180338524572e-06, + "loss": 0.0042, + "step": 86380 + }, + { + "epoch": 13.418763591177385, + "grad_norm": 0.17758852243423462, + "learning_rate": 4.9799842518855275e-06, + "loss": 0.0036, + "step": 86390 + }, + { + "epoch": 13.42031686859273, + "grad_norm": 0.5735266804695129, + "learning_rate": 4.9727930885677285e-06, + "loss": 0.0038, + "step": 86400 + }, + { + "epoch": 13.421870146008077, + "grad_norm": 0.4054911732673645, + "learning_rate": 4.965606849357573e-06, + "loss": 0.0047, + "step": 86410 + }, + { + "epoch": 13.423423423423424, + "grad_norm": 0.0789642259478569, + "learning_rate": 4.9584255350409445e-06, + "loss": 0.0019, + "step": 86420 + }, + { + "epoch": 13.42497670083877, + "grad_norm": 0.3032204508781433, + "learning_rate": 4.951249146403175e-06, + "loss": 0.005, + "step": 86430 + }, + { + "epoch": 13.426529978254116, + "grad_norm": 0.14546078443527222, + "learning_rate": 4.9440776842290815e-06, + "loss": 0.0049, + "step": 86440 + }, + { + "epoch": 13.428083255669463, + "grad_norm": 0.3175618052482605, + "learning_rate": 4.936911149302903e-06, + "loss": 0.0032, + "step": 86450 + }, + { + "epoch": 13.429636533084809, + "grad_norm": 0.15905947983264923, + "learning_rate": 4.929749542408379e-06, + "loss": 0.0033, + "step": 86460 + }, + { + "epoch": 13.431189810500156, + "grad_norm": 0.09931192547082901, + "learning_rate": 4.922592864328673e-06, + "loss": 0.0063, + "step": 86470 + }, + { + "epoch": 13.432743087915501, + "grad_norm": 0.3739932179450989, + "learning_rate": 4.915441115846436e-06, + "loss": 0.0053, + "step": 86480 + }, + { + "epoch": 13.434296365330848, + "grad_norm": 0.3091528117656708, + "learning_rate": 4.908294297743782e-06, + "loss": 0.0033, + "step": 86490 + }, + { + "epoch": 13.435849642746195, + "grad_norm": 0.09555879980325699, + "learning_rate": 4.901152410802273e-06, + "loss": 0.0045, + "step": 86500 + }, + { + "epoch": 13.43740292016154, + "grad_norm": 0.1865515559911728, + "learning_rate": 4.894015455802919e-06, + "loss": 0.0025, + "step": 86510 + }, + { + "epoch": 13.438956197576887, + "grad_norm": 0.2074289172887802, + "learning_rate": 4.8868834335262346e-06, + "loss": 0.0022, + "step": 86520 + }, + { + "epoch": 13.440509474992234, + "grad_norm": 0.06997258216142654, + "learning_rate": 4.87975634475214e-06, + "loss": 0.0029, + "step": 86530 + }, + { + "epoch": 13.44206275240758, + "grad_norm": 0.3091980516910553, + "learning_rate": 4.8726341902600614e-06, + "loss": 0.0025, + "step": 86540 + }, + { + "epoch": 13.443616029822927, + "grad_norm": 0.2289896011352539, + "learning_rate": 4.865516970828843e-06, + "loss": 0.0041, + "step": 86550 + }, + { + "epoch": 13.445169307238272, + "grad_norm": 0.39504745602607727, + "learning_rate": 4.85840468723685e-06, + "loss": 0.0029, + "step": 86560 + }, + { + "epoch": 13.44672258465362, + "grad_norm": 0.2731008231639862, + "learning_rate": 4.85129734026184e-06, + "loss": 0.0054, + "step": 86570 + }, + { + "epoch": 13.448275862068966, + "grad_norm": 0.06921585649251938, + "learning_rate": 4.844194930681073e-06, + "loss": 0.003, + "step": 86580 + }, + { + "epoch": 13.449829139484311, + "grad_norm": 0.12551920115947723, + "learning_rate": 4.837097459271267e-06, + "loss": 0.0035, + "step": 86590 + }, + { + "epoch": 13.451382416899659, + "grad_norm": 0.05900665745139122, + "learning_rate": 4.830004926808573e-06, + "loss": 0.0041, + "step": 86600 + }, + { + "epoch": 13.452935694315004, + "grad_norm": 0.220416858792305, + "learning_rate": 4.822917334068627e-06, + "loss": 0.0056, + "step": 86610 + }, + { + "epoch": 13.45448897173035, + "grad_norm": 0.06911660730838776, + "learning_rate": 4.81583468182652e-06, + "loss": 0.0031, + "step": 86620 + }, + { + "epoch": 13.456042249145698, + "grad_norm": 0.2186310887336731, + "learning_rate": 4.808756970856809e-06, + "loss": 0.0032, + "step": 86630 + }, + { + "epoch": 13.457595526561043, + "grad_norm": 1.1679641008377075, + "learning_rate": 4.801684201933476e-06, + "loss": 0.0032, + "step": 86640 + }, + { + "epoch": 13.45914880397639, + "grad_norm": 0.2635987102985382, + "learning_rate": 4.794616375830019e-06, + "loss": 0.0046, + "step": 86650 + }, + { + "epoch": 13.460702081391737, + "grad_norm": 0.21714159846305847, + "learning_rate": 4.787553493319341e-06, + "loss": 0.0042, + "step": 86660 + }, + { + "epoch": 13.462255358807083, + "grad_norm": 0.14743193984031677, + "learning_rate": 4.780495555173836e-06, + "loss": 0.0039, + "step": 86670 + }, + { + "epoch": 13.46380863622243, + "grad_norm": 0.1753527969121933, + "learning_rate": 4.7734425621653475e-06, + "loss": 0.0032, + "step": 86680 + }, + { + "epoch": 13.465361913637775, + "grad_norm": 0.1050160601735115, + "learning_rate": 4.766394515065198e-06, + "loss": 0.0024, + "step": 86690 + }, + { + "epoch": 13.466915191053122, + "grad_norm": 0.1586662381887436, + "learning_rate": 4.75935141464412e-06, + "loss": 0.003, + "step": 86700 + }, + { + "epoch": 13.468468468468469, + "grad_norm": 0.4634625017642975, + "learning_rate": 4.75231326167237e-06, + "loss": 0.004, + "step": 86710 + }, + { + "epoch": 13.470021745883814, + "grad_norm": 0.2046770602464676, + "learning_rate": 4.745280056919599e-06, + "loss": 0.005, + "step": 86720 + }, + { + "epoch": 13.471575023299161, + "grad_norm": 0.19072002172470093, + "learning_rate": 4.738251801154975e-06, + "loss": 0.004, + "step": 86730 + }, + { + "epoch": 13.473128300714508, + "grad_norm": 0.042329899966716766, + "learning_rate": 4.731228495147061e-06, + "loss": 0.0036, + "step": 86740 + }, + { + "epoch": 13.474681578129854, + "grad_norm": 0.1475701630115509, + "learning_rate": 4.7242101396639595e-06, + "loss": 0.005, + "step": 86750 + }, + { + "epoch": 13.4762348555452, + "grad_norm": 0.4968946874141693, + "learning_rate": 4.717196735473151e-06, + "loss": 0.0037, + "step": 86760 + }, + { + "epoch": 13.477788132960546, + "grad_norm": 0.19476300477981567, + "learning_rate": 4.710188283341643e-06, + "loss": 0.0035, + "step": 86770 + }, + { + "epoch": 13.479341410375893, + "grad_norm": 0.31400012969970703, + "learning_rate": 4.703184784035835e-06, + "loss": 0.0052, + "step": 86780 + }, + { + "epoch": 13.48089468779124, + "grad_norm": 0.18480552732944489, + "learning_rate": 4.696186238321642e-06, + "loss": 0.0029, + "step": 86790 + }, + { + "epoch": 13.482447965206585, + "grad_norm": 0.11069396883249283, + "learning_rate": 4.689192646964408e-06, + "loss": 0.0036, + "step": 86800 + }, + { + "epoch": 13.484001242621932, + "grad_norm": 0.5640435218811035, + "learning_rate": 4.682204010728952e-06, + "loss": 0.0045, + "step": 86810 + }, + { + "epoch": 13.48555452003728, + "grad_norm": 0.2091747671365738, + "learning_rate": 4.67522033037952e-06, + "loss": 0.004, + "step": 86820 + }, + { + "epoch": 13.487107797452625, + "grad_norm": 0.1543406844139099, + "learning_rate": 4.668241606679846e-06, + "loss": 0.0048, + "step": 86830 + }, + { + "epoch": 13.488661074867972, + "grad_norm": 0.15302690863609314, + "learning_rate": 4.66126784039313e-06, + "loss": 0.003, + "step": 86840 + }, + { + "epoch": 13.490214352283317, + "grad_norm": 0.12793605029582977, + "learning_rate": 4.6542990322819836e-06, + "loss": 0.0026, + "step": 86850 + }, + { + "epoch": 13.491767629698664, + "grad_norm": 0.17897532880306244, + "learning_rate": 4.64733518310852e-06, + "loss": 0.0037, + "step": 86860 + }, + { + "epoch": 13.493320907114011, + "grad_norm": 0.3025084137916565, + "learning_rate": 4.640376293634291e-06, + "loss": 0.0044, + "step": 86870 + }, + { + "epoch": 13.494874184529356, + "grad_norm": 0.2762679159641266, + "learning_rate": 4.633422364620327e-06, + "loss": 0.0026, + "step": 86880 + }, + { + "epoch": 13.496427461944704, + "grad_norm": 0.4421159625053406, + "learning_rate": 4.6264733968270745e-06, + "loss": 0.0042, + "step": 86890 + }, + { + "epoch": 13.497980739360049, + "grad_norm": 0.21795612573623657, + "learning_rate": 4.619529391014482e-06, + "loss": 0.0037, + "step": 86900 + }, + { + "epoch": 13.499534016775396, + "grad_norm": 0.13573138415813446, + "learning_rate": 4.61259034794192e-06, + "loss": 0.0024, + "step": 86910 + }, + { + "epoch": 13.501087294190743, + "grad_norm": 0.11788051575422287, + "learning_rate": 4.605656268368247e-06, + "loss": 0.0035, + "step": 86920 + }, + { + "epoch": 13.502640571606088, + "grad_norm": 0.9034392237663269, + "learning_rate": 4.598727153051735e-06, + "loss": 0.0032, + "step": 86930 + }, + { + "epoch": 13.504193849021435, + "grad_norm": 0.1092156320810318, + "learning_rate": 4.591803002750178e-06, + "loss": 0.002, + "step": 86940 + }, + { + "epoch": 13.505747126436782, + "grad_norm": 0.19382718205451965, + "learning_rate": 4.58488381822077e-06, + "loss": 0.0038, + "step": 86950 + }, + { + "epoch": 13.507300403852128, + "grad_norm": 0.12508442997932434, + "learning_rate": 4.5779696002201946e-06, + "loss": 0.0059, + "step": 86960 + }, + { + "epoch": 13.508853681267475, + "grad_norm": 0.19464626908302307, + "learning_rate": 4.571060349504563e-06, + "loss": 0.0039, + "step": 86970 + }, + { + "epoch": 13.51040695868282, + "grad_norm": 0.08803722262382507, + "learning_rate": 4.564156066829478e-06, + "loss": 0.003, + "step": 86980 + }, + { + "epoch": 13.511960236098167, + "grad_norm": 0.0974809005856514, + "learning_rate": 4.557256752949957e-06, + "loss": 0.0039, + "step": 86990 + }, + { + "epoch": 13.513513513513514, + "grad_norm": 0.15018446743488312, + "learning_rate": 4.550362408620534e-06, + "loss": 0.0057, + "step": 87000 + }, + { + "epoch": 13.51506679092886, + "grad_norm": 0.4553622901439667, + "learning_rate": 4.543473034595136e-06, + "loss": 0.0029, + "step": 87010 + }, + { + "epoch": 13.516620068344206, + "grad_norm": 0.18487612903118134, + "learning_rate": 4.536588631627192e-06, + "loss": 0.0037, + "step": 87020 + }, + { + "epoch": 13.518173345759553, + "grad_norm": 0.15042009949684143, + "learning_rate": 4.5297092004695565e-06, + "loss": 0.0043, + "step": 87030 + }, + { + "epoch": 13.519726623174899, + "grad_norm": 0.07782559841871262, + "learning_rate": 4.522834741874555e-06, + "loss": 0.0038, + "step": 87040 + }, + { + "epoch": 13.521279900590246, + "grad_norm": 0.5551584362983704, + "learning_rate": 4.515965256593974e-06, + "loss": 0.0041, + "step": 87050 + }, + { + "epoch": 13.522833178005591, + "grad_norm": 0.19030429422855377, + "learning_rate": 4.509100745379058e-06, + "loss": 0.0041, + "step": 87060 + }, + { + "epoch": 13.524386455420938, + "grad_norm": 0.4502018988132477, + "learning_rate": 4.502241208980479e-06, + "loss": 0.002, + "step": 87070 + }, + { + "epoch": 13.525939732836285, + "grad_norm": 1.1100070476531982, + "learning_rate": 4.495386648148398e-06, + "loss": 0.0045, + "step": 87080 + }, + { + "epoch": 13.52749301025163, + "grad_norm": 0.46909892559051514, + "learning_rate": 4.488537063632431e-06, + "loss": 0.003, + "step": 87090 + }, + { + "epoch": 13.529046287666977, + "grad_norm": 0.08670759946107864, + "learning_rate": 4.481692456181608e-06, + "loss": 0.0024, + "step": 87100 + }, + { + "epoch": 13.530599565082325, + "grad_norm": 0.33636394143104553, + "learning_rate": 4.474852826544479e-06, + "loss": 0.0039, + "step": 87110 + }, + { + "epoch": 13.53215284249767, + "grad_norm": 0.25192371010780334, + "learning_rate": 4.468018175468974e-06, + "loss": 0.0076, + "step": 87120 + }, + { + "epoch": 13.533706119913017, + "grad_norm": 0.07054472714662552, + "learning_rate": 4.461188503702568e-06, + "loss": 0.0056, + "step": 87130 + }, + { + "epoch": 13.535259397328362, + "grad_norm": 0.5166176557540894, + "learning_rate": 4.454363811992107e-06, + "loss": 0.0035, + "step": 87140 + }, + { + "epoch": 13.53681267474371, + "grad_norm": 0.10283929854631424, + "learning_rate": 4.447544101083956e-06, + "loss": 0.0031, + "step": 87150 + }, + { + "epoch": 13.538365952159056, + "grad_norm": 0.23031163215637207, + "learning_rate": 4.440729371723884e-06, + "loss": 0.0033, + "step": 87160 + }, + { + "epoch": 13.539919229574402, + "grad_norm": 0.19513508677482605, + "learning_rate": 4.433919624657162e-06, + "loss": 0.0043, + "step": 87170 + }, + { + "epoch": 13.541472506989749, + "grad_norm": 0.3014228940010071, + "learning_rate": 4.427114860628462e-06, + "loss": 0.0042, + "step": 87180 + }, + { + "epoch": 13.543025784405096, + "grad_norm": 0.3646093010902405, + "learning_rate": 4.420315080381982e-06, + "loss": 0.0038, + "step": 87190 + }, + { + "epoch": 13.544579061820441, + "grad_norm": 0.2623579502105713, + "learning_rate": 4.4135202846613045e-06, + "loss": 0.0056, + "step": 87200 + }, + { + "epoch": 13.546132339235788, + "grad_norm": 0.6740797162055969, + "learning_rate": 4.406730474209525e-06, + "loss": 0.0047, + "step": 87210 + }, + { + "epoch": 13.547685616651133, + "grad_norm": 0.20154674351215363, + "learning_rate": 4.399945649769138e-06, + "loss": 0.004, + "step": 87220 + }, + { + "epoch": 13.54923889406648, + "grad_norm": 0.5500326752662659, + "learning_rate": 4.393165812082151e-06, + "loss": 0.0035, + "step": 87230 + }, + { + "epoch": 13.550792171481827, + "grad_norm": 0.5115874409675598, + "learning_rate": 4.386390961889969e-06, + "loss": 0.0036, + "step": 87240 + }, + { + "epoch": 13.552345448897173, + "grad_norm": 0.15449915826320648, + "learning_rate": 4.3796210999334895e-06, + "loss": 0.0033, + "step": 87250 + }, + { + "epoch": 13.55389872631252, + "grad_norm": 0.073911152780056, + "learning_rate": 4.372856226953065e-06, + "loss": 0.0037, + "step": 87260 + }, + { + "epoch": 13.555452003727865, + "grad_norm": 0.5708444118499756, + "learning_rate": 4.366096343688486e-06, + "loss": 0.0035, + "step": 87270 + }, + { + "epoch": 13.557005281143212, + "grad_norm": 0.08646012097597122, + "learning_rate": 4.359341450878995e-06, + "loss": 0.0027, + "step": 87280 + }, + { + "epoch": 13.558558558558559, + "grad_norm": 0.373253732919693, + "learning_rate": 4.352591549263302e-06, + "loss": 0.0044, + "step": 87290 + }, + { + "epoch": 13.560111835973904, + "grad_norm": 0.915134847164154, + "learning_rate": 4.345846639579582e-06, + "loss": 0.005, + "step": 87300 + }, + { + "epoch": 13.561665113389251, + "grad_norm": 0.0849207416176796, + "learning_rate": 4.339106722565417e-06, + "loss": 0.0032, + "step": 87310 + }, + { + "epoch": 13.563218390804598, + "grad_norm": 0.267776221036911, + "learning_rate": 4.332371798957896e-06, + "loss": 0.003, + "step": 87320 + }, + { + "epoch": 13.564771668219944, + "grad_norm": 0.2947346568107605, + "learning_rate": 4.3256418694935295e-06, + "loss": 0.0032, + "step": 87330 + }, + { + "epoch": 13.56632494563529, + "grad_norm": 0.09385289996862411, + "learning_rate": 4.3189169349083125e-06, + "loss": 0.0031, + "step": 87340 + }, + { + "epoch": 13.567878223050636, + "grad_norm": 0.10976707935333252, + "learning_rate": 4.312196995937645e-06, + "loss": 0.002, + "step": 87350 + }, + { + "epoch": 13.569431500465983, + "grad_norm": 0.08056145161390305, + "learning_rate": 4.305482053316434e-06, + "loss": 0.0029, + "step": 87360 + }, + { + "epoch": 13.57098477788133, + "grad_norm": 0.18706154823303223, + "learning_rate": 4.2987721077789925e-06, + "loss": 0.0027, + "step": 87370 + }, + { + "epoch": 13.572538055296675, + "grad_norm": 0.09103485941886902, + "learning_rate": 4.2920671600591325e-06, + "loss": 0.0023, + "step": 87380 + }, + { + "epoch": 13.574091332712023, + "grad_norm": 0.3740396797657013, + "learning_rate": 4.285367210890084e-06, + "loss": 0.0025, + "step": 87390 + }, + { + "epoch": 13.57564461012737, + "grad_norm": 0.20846863090991974, + "learning_rate": 4.278672261004551e-06, + "loss": 0.0028, + "step": 87400 + }, + { + "epoch": 13.577197887542715, + "grad_norm": 0.3705672025680542, + "learning_rate": 4.2719823111346704e-06, + "loss": 0.0041, + "step": 87410 + }, + { + "epoch": 13.578751164958062, + "grad_norm": 0.07937968522310257, + "learning_rate": 4.265297362012061e-06, + "loss": 0.002, + "step": 87420 + }, + { + "epoch": 13.580304442373407, + "grad_norm": 0.3343628942966461, + "learning_rate": 4.258617414367766e-06, + "loss": 0.0028, + "step": 87430 + }, + { + "epoch": 13.581857719788754, + "grad_norm": 0.24601741135120392, + "learning_rate": 4.251942468932296e-06, + "loss": 0.0045, + "step": 87440 + }, + { + "epoch": 13.583410997204101, + "grad_norm": 0.09946048259735107, + "learning_rate": 4.2452725264356175e-06, + "loss": 0.0035, + "step": 87450 + }, + { + "epoch": 13.584964274619447, + "grad_norm": 0.19896206259727478, + "learning_rate": 4.2386075876071575e-06, + "loss": 0.0032, + "step": 87460 + }, + { + "epoch": 13.586517552034794, + "grad_norm": 0.3042018413543701, + "learning_rate": 4.231947653175755e-06, + "loss": 0.0029, + "step": 87470 + }, + { + "epoch": 13.588070829450139, + "grad_norm": 0.057627517729997635, + "learning_rate": 4.225292723869761e-06, + "loss": 0.0024, + "step": 87480 + }, + { + "epoch": 13.589624106865486, + "grad_norm": 0.8915314078330994, + "learning_rate": 4.218642800416916e-06, + "loss": 0.0027, + "step": 87490 + }, + { + "epoch": 13.591177384280833, + "grad_norm": 0.16437098383903503, + "learning_rate": 4.211997883544472e-06, + "loss": 0.0043, + "step": 87500 + }, + { + "epoch": 13.592730661696178, + "grad_norm": 0.18966887891292572, + "learning_rate": 4.205357973979096e-06, + "loss": 0.004, + "step": 87510 + }, + { + "epoch": 13.594283939111525, + "grad_norm": 0.3045698404312134, + "learning_rate": 4.198723072446925e-06, + "loss": 0.0029, + "step": 87520 + }, + { + "epoch": 13.595837216526872, + "grad_norm": 0.4881443977355957, + "learning_rate": 4.192093179673534e-06, + "loss": 0.0033, + "step": 87530 + }, + { + "epoch": 13.597390493942218, + "grad_norm": 0.12585827708244324, + "learning_rate": 4.185468296383959e-06, + "loss": 0.0041, + "step": 87540 + }, + { + "epoch": 13.598943771357565, + "grad_norm": 0.0723823755979538, + "learning_rate": 4.1788484233027005e-06, + "loss": 0.0031, + "step": 87550 + }, + { + "epoch": 13.600497048772912, + "grad_norm": 0.1982218474149704, + "learning_rate": 4.172233561153677e-06, + "loss": 0.0034, + "step": 87560 + }, + { + "epoch": 13.602050326188257, + "grad_norm": 0.15829181671142578, + "learning_rate": 4.165623710660288e-06, + "loss": 0.0026, + "step": 87570 + }, + { + "epoch": 13.603603603603604, + "grad_norm": 0.25385627150535583, + "learning_rate": 4.1590188725453835e-06, + "loss": 0.0056, + "step": 87580 + }, + { + "epoch": 13.60515688101895, + "grad_norm": 0.161699578166008, + "learning_rate": 4.152419047531259e-06, + "loss": 0.0031, + "step": 87590 + }, + { + "epoch": 13.606710158434296, + "grad_norm": 0.5502298474311829, + "learning_rate": 4.145824236339651e-06, + "loss": 0.0061, + "step": 87600 + }, + { + "epoch": 13.608263435849643, + "grad_norm": 0.4371795952320099, + "learning_rate": 4.1392344396917695e-06, + "loss": 0.0029, + "step": 87610 + }, + { + "epoch": 13.609816713264989, + "grad_norm": 0.1868765950202942, + "learning_rate": 4.132649658308252e-06, + "loss": 0.0035, + "step": 87620 + }, + { + "epoch": 13.611369990680336, + "grad_norm": 0.16483396291732788, + "learning_rate": 4.126069892909213e-06, + "loss": 0.0042, + "step": 87630 + }, + { + "epoch": 13.612923268095681, + "grad_norm": 0.04825844243168831, + "learning_rate": 4.119495144214191e-06, + "loss": 0.0037, + "step": 87640 + }, + { + "epoch": 13.614476545511028, + "grad_norm": 0.2614704966545105, + "learning_rate": 4.112925412942214e-06, + "loss": 0.0035, + "step": 87650 + }, + { + "epoch": 13.616029822926375, + "grad_norm": 0.10133355110883713, + "learning_rate": 4.106360699811712e-06, + "loss": 0.0038, + "step": 87660 + }, + { + "epoch": 13.61758310034172, + "grad_norm": 0.1049819141626358, + "learning_rate": 4.099801005540616e-06, + "loss": 0.0033, + "step": 87670 + }, + { + "epoch": 13.619136377757068, + "grad_norm": 0.10557018220424652, + "learning_rate": 4.093246330846256e-06, + "loss": 0.0039, + "step": 87680 + }, + { + "epoch": 13.620689655172415, + "grad_norm": 0.44169631600379944, + "learning_rate": 4.086696676445467e-06, + "loss": 0.006, + "step": 87690 + }, + { + "epoch": 13.62224293258776, + "grad_norm": 0.09764260053634644, + "learning_rate": 4.0801520430544914e-06, + "loss": 0.0035, + "step": 87700 + }, + { + "epoch": 13.623796210003107, + "grad_norm": 0.2884490489959717, + "learning_rate": 4.073612431389062e-06, + "loss": 0.0034, + "step": 87710 + }, + { + "epoch": 13.625349487418452, + "grad_norm": 0.14353835582733154, + "learning_rate": 4.067077842164319e-06, + "loss": 0.0038, + "step": 87720 + }, + { + "epoch": 13.6269027648338, + "grad_norm": 0.5429804921150208, + "learning_rate": 4.0605482760948945e-06, + "loss": 0.0036, + "step": 87730 + }, + { + "epoch": 13.628456042249146, + "grad_norm": 0.19629134237766266, + "learning_rate": 4.054023733894829e-06, + "loss": 0.0032, + "step": 87740 + }, + { + "epoch": 13.630009319664492, + "grad_norm": 0.33434250950813293, + "learning_rate": 4.047504216277653e-06, + "loss": 0.0027, + "step": 87750 + }, + { + "epoch": 13.631562597079839, + "grad_norm": 0.35173752903938293, + "learning_rate": 4.040989723956323e-06, + "loss": 0.0033, + "step": 87760 + }, + { + "epoch": 13.633115874495186, + "grad_norm": 0.22189168632030487, + "learning_rate": 4.03448025764327e-06, + "loss": 0.0037, + "step": 87770 + }, + { + "epoch": 13.634669151910531, + "grad_norm": 0.5592881441116333, + "learning_rate": 4.027975818050339e-06, + "loss": 0.0032, + "step": 87780 + }, + { + "epoch": 13.636222429325878, + "grad_norm": 0.2908085882663727, + "learning_rate": 4.021476405888858e-06, + "loss": 0.003, + "step": 87790 + }, + { + "epoch": 13.637775706741223, + "grad_norm": 0.6151463389396667, + "learning_rate": 4.01498202186959e-06, + "loss": 0.0036, + "step": 87800 + }, + { + "epoch": 13.63932898415657, + "grad_norm": 0.18141019344329834, + "learning_rate": 4.008492666702751e-06, + "loss": 0.0046, + "step": 87810 + }, + { + "epoch": 13.640882261571917, + "grad_norm": 0.19580282270908356, + "learning_rate": 4.002008341098001e-06, + "loss": 0.0066, + "step": 87820 + }, + { + "epoch": 13.642435538987263, + "grad_norm": 0.31354549527168274, + "learning_rate": 3.995529045764462e-06, + "loss": 0.0029, + "step": 87830 + }, + { + "epoch": 13.64398881640261, + "grad_norm": 0.1803373545408249, + "learning_rate": 3.989054781410712e-06, + "loss": 0.0041, + "step": 87840 + }, + { + "epoch": 13.645542093817955, + "grad_norm": 0.194765105843544, + "learning_rate": 3.982585548744744e-06, + "loss": 0.0039, + "step": 87850 + }, + { + "epoch": 13.647095371233302, + "grad_norm": 0.248404398560524, + "learning_rate": 3.9761213484740435e-06, + "loss": 0.0029, + "step": 87860 + }, + { + "epoch": 13.64864864864865, + "grad_norm": 0.3472294807434082, + "learning_rate": 3.9696621813055055e-06, + "loss": 0.0042, + "step": 87870 + }, + { + "epoch": 13.650201926063994, + "grad_norm": 0.0774853304028511, + "learning_rate": 3.9632080479455044e-06, + "loss": 0.0023, + "step": 87880 + }, + { + "epoch": 13.651755203479341, + "grad_norm": 0.19039373099803925, + "learning_rate": 3.956758949099859e-06, + "loss": 0.0034, + "step": 87890 + }, + { + "epoch": 13.653308480894689, + "grad_norm": 0.42094117403030396, + "learning_rate": 3.9503148854738385e-06, + "loss": 0.0027, + "step": 87900 + }, + { + "epoch": 13.654861758310034, + "grad_norm": 0.12223581969738007, + "learning_rate": 3.943875857772139e-06, + "loss": 0.0045, + "step": 87910 + }, + { + "epoch": 13.65641503572538, + "grad_norm": 0.5023924112319946, + "learning_rate": 3.9374418666989375e-06, + "loss": 0.0041, + "step": 87920 + }, + { + "epoch": 13.657968313140726, + "grad_norm": 1.2205561399459839, + "learning_rate": 3.931012912957838e-06, + "loss": 0.0025, + "step": 87930 + }, + { + "epoch": 13.659521590556073, + "grad_norm": 0.42152366042137146, + "learning_rate": 3.9245889972519045e-06, + "loss": 0.0034, + "step": 87940 + }, + { + "epoch": 13.66107486797142, + "grad_norm": 0.2735656499862671, + "learning_rate": 3.9181701202836265e-06, + "loss": 0.0046, + "step": 87950 + }, + { + "epoch": 13.662628145386766, + "grad_norm": 0.47650110721588135, + "learning_rate": 3.911756282755003e-06, + "loss": 0.0049, + "step": 87960 + }, + { + "epoch": 13.664181422802113, + "grad_norm": 0.083956278860569, + "learning_rate": 3.905347485367411e-06, + "loss": 0.0032, + "step": 87970 + }, + { + "epoch": 13.66573470021746, + "grad_norm": 0.045832447707653046, + "learning_rate": 3.898943728821725e-06, + "loss": 0.0035, + "step": 87980 + }, + { + "epoch": 13.667287977632805, + "grad_norm": 0.1290157288312912, + "learning_rate": 3.892545013818227e-06, + "loss": 0.0032, + "step": 87990 + }, + { + "epoch": 13.668841255048152, + "grad_norm": 0.21693530678749084, + "learning_rate": 3.886151341056687e-06, + "loss": 0.0035, + "step": 88000 + }, + { + "epoch": 13.670394532463497, + "grad_norm": 0.051996827125549316, + "learning_rate": 3.8797627112363066e-06, + "loss": 0.0025, + "step": 88010 + }, + { + "epoch": 13.671947809878844, + "grad_norm": 0.1265115737915039, + "learning_rate": 3.873379125055743e-06, + "loss": 0.0031, + "step": 88020 + }, + { + "epoch": 13.673501087294191, + "grad_norm": 0.3381035625934601, + "learning_rate": 3.867000583213082e-06, + "loss": 0.0033, + "step": 88030 + }, + { + "epoch": 13.675054364709537, + "grad_norm": 0.12935805320739746, + "learning_rate": 3.860627086405871e-06, + "loss": 0.0034, + "step": 88040 + }, + { + "epoch": 13.676607642124884, + "grad_norm": 0.2807576358318329, + "learning_rate": 3.854258635331126e-06, + "loss": 0.0036, + "step": 88050 + }, + { + "epoch": 13.678160919540229, + "grad_norm": 0.5597702264785767, + "learning_rate": 3.8478952306852665e-06, + "loss": 0.0051, + "step": 88060 + }, + { + "epoch": 13.679714196955576, + "grad_norm": 0.23214593529701233, + "learning_rate": 3.841536873164198e-06, + "loss": 0.0016, + "step": 88070 + }, + { + "epoch": 13.681267474370923, + "grad_norm": 0.22349166870117188, + "learning_rate": 3.8351835634632626e-06, + "loss": 0.0033, + "step": 88080 + }, + { + "epoch": 13.682820751786268, + "grad_norm": 0.955020010471344, + "learning_rate": 3.828835302277245e-06, + "loss": 0.0039, + "step": 88090 + }, + { + "epoch": 13.684374029201615, + "grad_norm": 0.22438108921051025, + "learning_rate": 3.822492090300378e-06, + "loss": 0.0031, + "step": 88100 + }, + { + "epoch": 13.685927306616962, + "grad_norm": 0.3905150890350342, + "learning_rate": 3.8161539282263566e-06, + "loss": 0.0033, + "step": 88110 + }, + { + "epoch": 13.687480584032308, + "grad_norm": 0.15421147644519806, + "learning_rate": 3.8098208167482995e-06, + "loss": 0.0041, + "step": 88120 + }, + { + "epoch": 13.689033861447655, + "grad_norm": 0.16890253126621246, + "learning_rate": 3.803492756558796e-06, + "loss": 0.0028, + "step": 88130 + }, + { + "epoch": 13.690587138863002, + "grad_norm": 0.09071087092161179, + "learning_rate": 3.797169748349855e-06, + "loss": 0.0043, + "step": 88140 + }, + { + "epoch": 13.692140416278347, + "grad_norm": 0.16316579282283783, + "learning_rate": 3.790851792812977e-06, + "loss": 0.0034, + "step": 88150 + }, + { + "epoch": 13.693693693693694, + "grad_norm": 0.1267879158258438, + "learning_rate": 3.784538890639072e-06, + "loss": 0.0031, + "step": 88160 + }, + { + "epoch": 13.69524697110904, + "grad_norm": 0.24049299955368042, + "learning_rate": 3.7782310425185153e-06, + "loss": 0.0039, + "step": 88170 + }, + { + "epoch": 13.696800248524386, + "grad_norm": 0.10318229347467422, + "learning_rate": 3.77192824914111e-06, + "loss": 0.0031, + "step": 88180 + }, + { + "epoch": 13.698353525939734, + "grad_norm": 0.2152523398399353, + "learning_rate": 3.7656305111961333e-06, + "loss": 0.0036, + "step": 88190 + }, + { + "epoch": 13.699906803355079, + "grad_norm": 0.2375965267419815, + "learning_rate": 3.7593378293722782e-06, + "loss": 0.0034, + "step": 88200 + }, + { + "epoch": 13.701460080770426, + "grad_norm": 0.14525476098060608, + "learning_rate": 3.7530502043577333e-06, + "loss": 0.0037, + "step": 88210 + }, + { + "epoch": 13.703013358185771, + "grad_norm": 0.24679166078567505, + "learning_rate": 3.746767636840076e-06, + "loss": 0.0045, + "step": 88220 + }, + { + "epoch": 13.704566635601118, + "grad_norm": 0.10503613203763962, + "learning_rate": 3.74049012750638e-06, + "loss": 0.0047, + "step": 88230 + }, + { + "epoch": 13.706119913016465, + "grad_norm": 0.2290170043706894, + "learning_rate": 3.734217677043128e-06, + "loss": 0.0038, + "step": 88240 + }, + { + "epoch": 13.70767319043181, + "grad_norm": 0.1554582566022873, + "learning_rate": 3.727950286136267e-06, + "loss": 0.0034, + "step": 88250 + }, + { + "epoch": 13.709226467847158, + "grad_norm": 0.6119568943977356, + "learning_rate": 3.7216879554711925e-06, + "loss": 0.0025, + "step": 88260 + }, + { + "epoch": 13.710779745262505, + "grad_norm": 0.6791837215423584, + "learning_rate": 3.715430685732757e-06, + "loss": 0.0032, + "step": 88270 + }, + { + "epoch": 13.71233302267785, + "grad_norm": 0.14196796715259552, + "learning_rate": 3.7091784776052243e-06, + "loss": 0.0034, + "step": 88280 + }, + { + "epoch": 13.713886300093197, + "grad_norm": 0.2795102298259735, + "learning_rate": 3.7029313317723356e-06, + "loss": 0.0046, + "step": 88290 + }, + { + "epoch": 13.715439577508542, + "grad_norm": 0.06883997470140457, + "learning_rate": 3.6966892489172787e-06, + "loss": 0.0052, + "step": 88300 + }, + { + "epoch": 13.71699285492389, + "grad_norm": 0.10720191150903702, + "learning_rate": 3.6904522297226575e-06, + "loss": 0.0028, + "step": 88310 + }, + { + "epoch": 13.718546132339236, + "grad_norm": 0.16641828417778015, + "learning_rate": 3.6842202748705647e-06, + "loss": 0.0043, + "step": 88320 + }, + { + "epoch": 13.720099409754582, + "grad_norm": 0.07274191826581955, + "learning_rate": 3.677993385042494e-06, + "loss": 0.0035, + "step": 88330 + }, + { + "epoch": 13.721652687169929, + "grad_norm": 0.5394167900085449, + "learning_rate": 3.67177156091944e-06, + "loss": 0.0048, + "step": 88340 + }, + { + "epoch": 13.723205964585276, + "grad_norm": 0.1118447482585907, + "learning_rate": 3.6655548031817866e-06, + "loss": 0.0031, + "step": 88350 + }, + { + "epoch": 13.724759242000621, + "grad_norm": 0.5324036478996277, + "learning_rate": 3.6593431125094e-06, + "loss": 0.0036, + "step": 88360 + }, + { + "epoch": 13.726312519415968, + "grad_norm": 0.11192566156387329, + "learning_rate": 3.6531364895815766e-06, + "loss": 0.0058, + "step": 88370 + }, + { + "epoch": 13.727865796831313, + "grad_norm": 0.12104995548725128, + "learning_rate": 3.6469349350770677e-06, + "loss": 0.0034, + "step": 88380 + }, + { + "epoch": 13.72941907424666, + "grad_norm": 0.1697828322649002, + "learning_rate": 3.6407384496740525e-06, + "loss": 0.0038, + "step": 88390 + }, + { + "epoch": 13.730972351662007, + "grad_norm": 0.2897087037563324, + "learning_rate": 3.6345470340501954e-06, + "loss": 0.0037, + "step": 88400 + }, + { + "epoch": 13.732525629077353, + "grad_norm": 0.6425383687019348, + "learning_rate": 3.6283606888825537e-06, + "loss": 0.0034, + "step": 88410 + }, + { + "epoch": 13.7340789064927, + "grad_norm": 0.1612258404493332, + "learning_rate": 3.622179414847676e-06, + "loss": 0.0024, + "step": 88420 + }, + { + "epoch": 13.735632183908045, + "grad_norm": 0.3883056640625, + "learning_rate": 3.616003212621527e-06, + "loss": 0.0045, + "step": 88430 + }, + { + "epoch": 13.737185461323392, + "grad_norm": 0.1393423229455948, + "learning_rate": 3.609832082879533e-06, + "loss": 0.0037, + "step": 88440 + }, + { + "epoch": 13.73873873873874, + "grad_norm": 0.046419911086559296, + "learning_rate": 3.6036660262965485e-06, + "loss": 0.0042, + "step": 88450 + }, + { + "epoch": 13.740292016154084, + "grad_norm": 0.3471551537513733, + "learning_rate": 3.597505043546895e-06, + "loss": 0.0026, + "step": 88460 + }, + { + "epoch": 13.741845293569432, + "grad_norm": 0.2640274465084076, + "learning_rate": 3.5913491353043227e-06, + "loss": 0.0037, + "step": 88470 + }, + { + "epoch": 13.743398570984779, + "grad_norm": 0.8933520913124084, + "learning_rate": 3.5851983022420375e-06, + "loss": 0.0053, + "step": 88480 + }, + { + "epoch": 13.744951848400124, + "grad_norm": 0.9311317801475525, + "learning_rate": 3.5790525450326797e-06, + "loss": 0.0051, + "step": 88490 + }, + { + "epoch": 13.746505125815471, + "grad_norm": 0.13117995858192444, + "learning_rate": 3.572911864348344e-06, + "loss": 0.0028, + "step": 88500 + }, + { + "epoch": 13.748058403230818, + "grad_norm": 0.14919666945934296, + "learning_rate": 3.5667762608605717e-06, + "loss": 0.0042, + "step": 88510 + }, + { + "epoch": 13.749611680646163, + "grad_norm": 0.2050829529762268, + "learning_rate": 3.5606457352403255e-06, + "loss": 0.0036, + "step": 88520 + }, + { + "epoch": 13.75116495806151, + "grad_norm": 0.2705601751804352, + "learning_rate": 3.554520288158042e-06, + "loss": 0.0035, + "step": 88530 + }, + { + "epoch": 13.752718235476856, + "grad_norm": 0.49211806058883667, + "learning_rate": 3.5483999202835903e-06, + "loss": 0.0043, + "step": 88540 + }, + { + "epoch": 13.754271512892203, + "grad_norm": 0.2823273837566376, + "learning_rate": 3.5422846322862913e-06, + "loss": 0.0037, + "step": 88550 + }, + { + "epoch": 13.75582479030755, + "grad_norm": 0.12851683795452118, + "learning_rate": 3.5361744248348925e-06, + "loss": 0.006, + "step": 88560 + }, + { + "epoch": 13.757378067722895, + "grad_norm": 0.17918206751346588, + "learning_rate": 3.5300692985976047e-06, + "loss": 0.0032, + "step": 88570 + }, + { + "epoch": 13.758931345138242, + "grad_norm": 0.363151490688324, + "learning_rate": 3.52396925424206e-06, + "loss": 0.0041, + "step": 88580 + }, + { + "epoch": 13.760484622553587, + "grad_norm": 0.21334844827651978, + "learning_rate": 3.5178742924353747e-06, + "loss": 0.003, + "step": 88590 + }, + { + "epoch": 13.762037899968934, + "grad_norm": 0.11311841011047363, + "learning_rate": 3.511784413844066e-06, + "loss": 0.0036, + "step": 88600 + }, + { + "epoch": 13.763591177384281, + "grad_norm": 0.035220153629779816, + "learning_rate": 3.5056996191341287e-06, + "loss": 0.002, + "step": 88610 + }, + { + "epoch": 13.765144454799627, + "grad_norm": 0.36711862683296204, + "learning_rate": 3.4996199089709692e-06, + "loss": 0.0031, + "step": 88620 + }, + { + "epoch": 13.766697732214974, + "grad_norm": 0.12106525152921677, + "learning_rate": 3.4935452840194725e-06, + "loss": 0.0032, + "step": 88630 + }, + { + "epoch": 13.76825100963032, + "grad_norm": 0.44627219438552856, + "learning_rate": 3.48747574494393e-06, + "loss": 0.0039, + "step": 88640 + }, + { + "epoch": 13.769804287045666, + "grad_norm": 1.725696086883545, + "learning_rate": 3.4814112924081154e-06, + "loss": 0.004, + "step": 88650 + }, + { + "epoch": 13.771357564461013, + "grad_norm": 0.7218125462532043, + "learning_rate": 3.475351927075221e-06, + "loss": 0.0052, + "step": 88660 + }, + { + "epoch": 13.772910841876358, + "grad_norm": 0.13351763784885406, + "learning_rate": 3.469297649607894e-06, + "loss": 0.0023, + "step": 88670 + }, + { + "epoch": 13.774464119291705, + "grad_norm": 0.14965569972991943, + "learning_rate": 3.463248460668217e-06, + "loss": 0.0028, + "step": 88680 + }, + { + "epoch": 13.776017396707052, + "grad_norm": 0.6745101809501648, + "learning_rate": 3.457204360917726e-06, + "loss": 0.0031, + "step": 88690 + }, + { + "epoch": 13.777570674122398, + "grad_norm": 0.45734184980392456, + "learning_rate": 3.4511653510173827e-06, + "loss": 0.0034, + "step": 88700 + }, + { + "epoch": 13.779123951537745, + "grad_norm": 0.3139828145503998, + "learning_rate": 3.4451314316276074e-06, + "loss": 0.0034, + "step": 88710 + }, + { + "epoch": 13.780677228953092, + "grad_norm": 0.11098187416791916, + "learning_rate": 3.4391026034082674e-06, + "loss": 0.0038, + "step": 88720 + }, + { + "epoch": 13.782230506368437, + "grad_norm": 0.29114776849746704, + "learning_rate": 3.433078867018674e-06, + "loss": 0.0028, + "step": 88730 + }, + { + "epoch": 13.783783783783784, + "grad_norm": 0.7237936854362488, + "learning_rate": 3.427060223117551e-06, + "loss": 0.0035, + "step": 88740 + }, + { + "epoch": 13.78533706119913, + "grad_norm": 0.1328185349702835, + "learning_rate": 3.4210466723631042e-06, + "loss": 0.0031, + "step": 88750 + }, + { + "epoch": 13.786890338614477, + "grad_norm": 0.23847387731075287, + "learning_rate": 3.415038215412969e-06, + "loss": 0.0032, + "step": 88760 + }, + { + "epoch": 13.788443616029824, + "grad_norm": 0.0951489731669426, + "learning_rate": 3.4090348529242088e-06, + "loss": 0.0045, + "step": 88770 + }, + { + "epoch": 13.789996893445169, + "grad_norm": 0.17070607841014862, + "learning_rate": 3.4030365855533487e-06, + "loss": 0.0027, + "step": 88780 + }, + { + "epoch": 13.791550170860516, + "grad_norm": 0.8070406317710876, + "learning_rate": 3.397043413956347e-06, + "loss": 0.0034, + "step": 88790 + }, + { + "epoch": 13.793103448275861, + "grad_norm": 0.1718774139881134, + "learning_rate": 3.3910553387886237e-06, + "loss": 0.0036, + "step": 88800 + }, + { + "epoch": 13.794656725691208, + "grad_norm": 0.3343864977359772, + "learning_rate": 3.3850723607049996e-06, + "loss": 0.0044, + "step": 88810 + }, + { + "epoch": 13.796210003106555, + "grad_norm": 0.8078945875167847, + "learning_rate": 3.3790944803597902e-06, + "loss": 0.0032, + "step": 88820 + }, + { + "epoch": 13.7977632805219, + "grad_norm": 0.29552945494651794, + "learning_rate": 3.373121698406706e-06, + "loss": 0.004, + "step": 88830 + }, + { + "epoch": 13.799316557937248, + "grad_norm": 0.36647504568099976, + "learning_rate": 3.3671540154989302e-06, + "loss": 0.004, + "step": 88840 + }, + { + "epoch": 13.800869835352595, + "grad_norm": 0.5518717169761658, + "learning_rate": 3.361191432289079e-06, + "loss": 0.0051, + "step": 88850 + }, + { + "epoch": 13.80242311276794, + "grad_norm": 0.09171056747436523, + "learning_rate": 3.3552339494292207e-06, + "loss": 0.0026, + "step": 88860 + }, + { + "epoch": 13.803976390183287, + "grad_norm": 0.11210417747497559, + "learning_rate": 3.349281567570839e-06, + "loss": 0.0037, + "step": 88870 + }, + { + "epoch": 13.805529667598632, + "grad_norm": 0.15773271024227142, + "learning_rate": 3.3433342873648965e-06, + "loss": 0.0037, + "step": 88880 + }, + { + "epoch": 13.80708294501398, + "grad_norm": 0.3121143579483032, + "learning_rate": 3.337392109461762e-06, + "loss": 0.0042, + "step": 88890 + }, + { + "epoch": 13.808636222429326, + "grad_norm": 0.7699894309043884, + "learning_rate": 3.331455034511266e-06, + "loss": 0.0044, + "step": 88900 + }, + { + "epoch": 13.810189499844672, + "grad_norm": 0.22690169513225555, + "learning_rate": 3.325523063162689e-06, + "loss": 0.0038, + "step": 88910 + }, + { + "epoch": 13.811742777260019, + "grad_norm": 0.1503302901983261, + "learning_rate": 3.3195961960647393e-06, + "loss": 0.0029, + "step": 88920 + }, + { + "epoch": 13.813296054675366, + "grad_norm": 1.1321834325790405, + "learning_rate": 3.31367443386556e-06, + "loss": 0.0024, + "step": 88930 + }, + { + "epoch": 13.814849332090711, + "grad_norm": 0.144097700715065, + "learning_rate": 3.3077577772127556e-06, + "loss": 0.0033, + "step": 88940 + }, + { + "epoch": 13.816402609506058, + "grad_norm": 0.14860202372074127, + "learning_rate": 3.301846226753358e-06, + "loss": 0.0041, + "step": 88950 + }, + { + "epoch": 13.817955886921403, + "grad_norm": 0.18116436898708344, + "learning_rate": 3.2959397831338444e-06, + "loss": 0.0023, + "step": 88960 + }, + { + "epoch": 13.81950916433675, + "grad_norm": 0.1279270499944687, + "learning_rate": 3.2900384470001432e-06, + "loss": 0.0048, + "step": 88970 + }, + { + "epoch": 13.821062441752098, + "grad_norm": 0.6328348517417908, + "learning_rate": 3.28414221899761e-06, + "loss": 0.0029, + "step": 88980 + }, + { + "epoch": 13.822615719167443, + "grad_norm": 0.420532763004303, + "learning_rate": 3.278251099771046e-06, + "loss": 0.0034, + "step": 88990 + }, + { + "epoch": 13.82416899658279, + "grad_norm": 0.20449036359786987, + "learning_rate": 3.2723650899646906e-06, + "loss": 0.0028, + "step": 89000 + }, + { + "epoch": 13.825722273998135, + "grad_norm": 0.08761309087276459, + "learning_rate": 3.2664841902222466e-06, + "loss": 0.0052, + "step": 89010 + }, + { + "epoch": 13.827275551413482, + "grad_norm": 0.17461633682250977, + "learning_rate": 3.2606084011868208e-06, + "loss": 0.004, + "step": 89020 + }, + { + "epoch": 13.82882882882883, + "grad_norm": 0.9157598614692688, + "learning_rate": 3.254737723500989e-06, + "loss": 0.0039, + "step": 89030 + }, + { + "epoch": 13.830382106244175, + "grad_norm": 0.3647926449775696, + "learning_rate": 3.2488721578067595e-06, + "loss": 0.0032, + "step": 89040 + }, + { + "epoch": 13.831935383659522, + "grad_norm": 0.18567298352718353, + "learning_rate": 3.243011704745591e-06, + "loss": 0.004, + "step": 89050 + }, + { + "epoch": 13.833488661074869, + "grad_norm": 0.28039801120758057, + "learning_rate": 3.237156364958355e-06, + "loss": 0.003, + "step": 89060 + }, + { + "epoch": 13.835041938490214, + "grad_norm": 0.11133961379528046, + "learning_rate": 3.231306139085394e-06, + "loss": 0.0024, + "step": 89070 + }, + { + "epoch": 13.836595215905561, + "grad_norm": 0.40444415807724, + "learning_rate": 3.2254610277664797e-06, + "loss": 0.0029, + "step": 89080 + }, + { + "epoch": 13.838148493320908, + "grad_norm": 1.851792812347412, + "learning_rate": 3.2196210316408183e-06, + "loss": 0.0044, + "step": 89090 + }, + { + "epoch": 13.839701770736253, + "grad_norm": 0.10286524891853333, + "learning_rate": 3.2137861513470644e-06, + "loss": 0.0018, + "step": 89100 + }, + { + "epoch": 13.8412550481516, + "grad_norm": 0.05271201208233833, + "learning_rate": 3.2079563875233252e-06, + "loss": 0.004, + "step": 89110 + }, + { + "epoch": 13.842808325566946, + "grad_norm": 0.7301967144012451, + "learning_rate": 3.2021317408071184e-06, + "loss": 0.0032, + "step": 89120 + }, + { + "epoch": 13.844361602982293, + "grad_norm": 0.2859667241573334, + "learning_rate": 3.196312211835434e-06, + "loss": 0.0045, + "step": 89130 + }, + { + "epoch": 13.84591488039764, + "grad_norm": 0.23793168365955353, + "learning_rate": 3.1904978012446686e-06, + "loss": 0.0048, + "step": 89140 + }, + { + "epoch": 13.847468157812985, + "grad_norm": 0.5807203650474548, + "learning_rate": 3.184688509670691e-06, + "loss": 0.0041, + "step": 89150 + }, + { + "epoch": 13.849021435228332, + "grad_norm": 0.30760347843170166, + "learning_rate": 3.178884337748783e-06, + "loss": 0.0039, + "step": 89160 + }, + { + "epoch": 13.850574712643677, + "grad_norm": 0.04239450767636299, + "learning_rate": 3.173085286113697e-06, + "loss": 0.0028, + "step": 89170 + }, + { + "epoch": 13.852127990059024, + "grad_norm": 0.11316999047994614, + "learning_rate": 3.1672913553995986e-06, + "loss": 0.0037, + "step": 89180 + }, + { + "epoch": 13.853681267474371, + "grad_norm": 0.11629611253738403, + "learning_rate": 3.1615025462401138e-06, + "loss": 0.0027, + "step": 89190 + }, + { + "epoch": 13.855234544889717, + "grad_norm": 0.24929815530776978, + "learning_rate": 3.1557188592682808e-06, + "loss": 0.0037, + "step": 89200 + }, + { + "epoch": 13.856787822305064, + "grad_norm": 0.2192678153514862, + "learning_rate": 3.1499402951166045e-06, + "loss": 0.0045, + "step": 89210 + }, + { + "epoch": 13.85834109972041, + "grad_norm": 0.4190911054611206, + "learning_rate": 3.144166854417019e-06, + "loss": 0.005, + "step": 89220 + }, + { + "epoch": 13.859894377135756, + "grad_norm": 0.2506757378578186, + "learning_rate": 3.1383985378009073e-06, + "loss": 0.0029, + "step": 89230 + }, + { + "epoch": 13.861447654551103, + "grad_norm": 0.3192744553089142, + "learning_rate": 3.1326353458990654e-06, + "loss": 0.0027, + "step": 89240 + }, + { + "epoch": 13.863000931966448, + "grad_norm": 0.38812386989593506, + "learning_rate": 3.126877279341761e-06, + "loss": 0.0026, + "step": 89250 + }, + { + "epoch": 13.864554209381796, + "grad_norm": 0.4275188446044922, + "learning_rate": 3.1211243387586964e-06, + "loss": 0.0031, + "step": 89260 + }, + { + "epoch": 13.866107486797143, + "grad_norm": 0.34856027364730835, + "learning_rate": 3.115376524778979e-06, + "loss": 0.0042, + "step": 89270 + }, + { + "epoch": 13.867660764212488, + "grad_norm": 0.1508437693119049, + "learning_rate": 3.109633838031195e-06, + "loss": 0.0043, + "step": 89280 + }, + { + "epoch": 13.869214041627835, + "grad_norm": 0.1050061509013176, + "learning_rate": 3.103896279143359e-06, + "loss": 0.0039, + "step": 89290 + }, + { + "epoch": 13.870767319043182, + "grad_norm": 0.22897568345069885, + "learning_rate": 3.098163848742924e-06, + "loss": 0.0037, + "step": 89300 + }, + { + "epoch": 13.872320596458527, + "grad_norm": 0.0788344293832779, + "learning_rate": 3.0924365474567675e-06, + "loss": 0.0036, + "step": 89310 + }, + { + "epoch": 13.873873873873874, + "grad_norm": 0.3792574405670166, + "learning_rate": 3.086714375911237e-06, + "loss": 0.0024, + "step": 89320 + }, + { + "epoch": 13.87542715128922, + "grad_norm": 0.2632588744163513, + "learning_rate": 3.0809973347320775e-06, + "loss": 0.0038, + "step": 89330 + }, + { + "epoch": 13.876980428704567, + "grad_norm": 0.14205658435821533, + "learning_rate": 3.075285424544516e-06, + "loss": 0.0038, + "step": 89340 + }, + { + "epoch": 13.878533706119914, + "grad_norm": 0.1449604034423828, + "learning_rate": 3.069578645973176e-06, + "loss": 0.0055, + "step": 89350 + }, + { + "epoch": 13.880086983535259, + "grad_norm": 0.1359230875968933, + "learning_rate": 3.0638769996421735e-06, + "loss": 0.004, + "step": 89360 + }, + { + "epoch": 13.881640260950606, + "grad_norm": 0.08433335274457932, + "learning_rate": 3.058180486175005e-06, + "loss": 0.0038, + "step": 89370 + }, + { + "epoch": 13.883193538365951, + "grad_norm": 0.169738307595253, + "learning_rate": 3.052489106194645e-06, + "loss": 0.0036, + "step": 89380 + }, + { + "epoch": 13.884746815781298, + "grad_norm": 0.05708065256476402, + "learning_rate": 3.0468028603234946e-06, + "loss": 0.0034, + "step": 89390 + }, + { + "epoch": 13.886300093196645, + "grad_norm": 0.3576609790325165, + "learning_rate": 3.04112174918339e-06, + "loss": 0.0031, + "step": 89400 + }, + { + "epoch": 13.88785337061199, + "grad_norm": 0.5501489639282227, + "learning_rate": 3.0354457733955954e-06, + "loss": 0.0026, + "step": 89410 + }, + { + "epoch": 13.889406648027338, + "grad_norm": 0.45197567343711853, + "learning_rate": 3.029774933580859e-06, + "loss": 0.0041, + "step": 89420 + }, + { + "epoch": 13.890959925442685, + "grad_norm": 0.4135373532772064, + "learning_rate": 3.0241092303593065e-06, + "loss": 0.0036, + "step": 89430 + }, + { + "epoch": 13.89251320285803, + "grad_norm": 0.11111979186534882, + "learning_rate": 3.018448664350554e-06, + "loss": 0.003, + "step": 89440 + }, + { + "epoch": 13.894066480273377, + "grad_norm": 0.34034931659698486, + "learning_rate": 3.0127932361736055e-06, + "loss": 0.0034, + "step": 89450 + }, + { + "epoch": 13.895619757688724, + "grad_norm": 0.153697669506073, + "learning_rate": 3.00714294644695e-06, + "loss": 0.0027, + "step": 89460 + }, + { + "epoch": 13.89717303510407, + "grad_norm": 0.13158175349235535, + "learning_rate": 3.0014977957884828e-06, + "loss": 0.0045, + "step": 89470 + }, + { + "epoch": 13.898726312519416, + "grad_norm": 0.19552159309387207, + "learning_rate": 2.99585778481557e-06, + "loss": 0.0038, + "step": 89480 + }, + { + "epoch": 13.900279589934762, + "grad_norm": 0.15241944789886475, + "learning_rate": 2.990222914144969e-06, + "loss": 0.0032, + "step": 89490 + }, + { + "epoch": 13.901832867350109, + "grad_norm": 0.11928153783082962, + "learning_rate": 2.984593184392914e-06, + "loss": 0.0025, + "step": 89500 + }, + { + "epoch": 13.903386144765456, + "grad_norm": 0.11409655958414078, + "learning_rate": 2.978968596175069e-06, + "loss": 0.0026, + "step": 89510 + }, + { + "epoch": 13.904939422180801, + "grad_norm": 0.6128997206687927, + "learning_rate": 2.9733491501065137e-06, + "loss": 0.0026, + "step": 89520 + }, + { + "epoch": 13.906492699596148, + "grad_norm": 0.22433312237262726, + "learning_rate": 2.9677348468018005e-06, + "loss": 0.004, + "step": 89530 + }, + { + "epoch": 13.908045977011493, + "grad_norm": 0.19554275274276733, + "learning_rate": 2.9621256868748726e-06, + "loss": 0.0043, + "step": 89540 + }, + { + "epoch": 13.90959925442684, + "grad_norm": 0.22590504586696625, + "learning_rate": 2.956521670939183e-06, + "loss": 0.0034, + "step": 89550 + }, + { + "epoch": 13.911152531842188, + "grad_norm": 0.1197437196969986, + "learning_rate": 2.950922799607542e-06, + "loss": 0.0024, + "step": 89560 + }, + { + "epoch": 13.912705809257533, + "grad_norm": 0.10079610347747803, + "learning_rate": 2.9453290734922535e-06, + "loss": 0.0039, + "step": 89570 + }, + { + "epoch": 13.91425908667288, + "grad_norm": 0.22850394248962402, + "learning_rate": 2.939740493205029e-06, + "loss": 0.0027, + "step": 89580 + }, + { + "epoch": 13.915812364088227, + "grad_norm": 0.13470280170440674, + "learning_rate": 2.9341570593570345e-06, + "loss": 0.0044, + "step": 89590 + }, + { + "epoch": 13.917365641503572, + "grad_norm": 0.19004958868026733, + "learning_rate": 2.928578772558843e-06, + "loss": 0.003, + "step": 89600 + }, + { + "epoch": 13.91891891891892, + "grad_norm": 0.13745911419391632, + "learning_rate": 2.923005633420528e-06, + "loss": 0.0031, + "step": 89610 + }, + { + "epoch": 13.920472196334265, + "grad_norm": 0.45252326130867004, + "learning_rate": 2.9174376425515237e-06, + "loss": 0.0021, + "step": 89620 + }, + { + "epoch": 13.922025473749612, + "grad_norm": 0.2301137000322342, + "learning_rate": 2.91187480056076e-06, + "loss": 0.004, + "step": 89630 + }, + { + "epoch": 13.923578751164959, + "grad_norm": 0.20227555930614471, + "learning_rate": 2.9063171080565664e-06, + "loss": 0.0028, + "step": 89640 + }, + { + "epoch": 13.925132028580304, + "grad_norm": 0.05528193712234497, + "learning_rate": 2.900764565646741e-06, + "loss": 0.0023, + "step": 89650 + }, + { + "epoch": 13.926685305995651, + "grad_norm": 0.1690881997346878, + "learning_rate": 2.895217173938475e-06, + "loss": 0.0032, + "step": 89660 + }, + { + "epoch": 13.928238583410998, + "grad_norm": 0.46335646510124207, + "learning_rate": 2.8896749335384445e-06, + "loss": 0.0035, + "step": 89670 + }, + { + "epoch": 13.929791860826343, + "grad_norm": 0.14134445786476135, + "learning_rate": 2.884137845052731e-06, + "loss": 0.0041, + "step": 89680 + }, + { + "epoch": 13.93134513824169, + "grad_norm": 0.23313169181346893, + "learning_rate": 2.8786059090868734e-06, + "loss": 0.003, + "step": 89690 + }, + { + "epoch": 13.932898415657036, + "grad_norm": 0.3080216944217682, + "learning_rate": 2.87307912624582e-06, + "loss": 0.0041, + "step": 89700 + }, + { + "epoch": 13.934451693072383, + "grad_norm": 0.17297746241092682, + "learning_rate": 2.867557497133977e-06, + "loss": 0.0032, + "step": 89710 + }, + { + "epoch": 13.93600497048773, + "grad_norm": 0.23316746950149536, + "learning_rate": 2.862041022355194e-06, + "loss": 0.0032, + "step": 89720 + }, + { + "epoch": 13.937558247903075, + "grad_norm": 0.6083120107650757, + "learning_rate": 2.8565297025127226e-06, + "loss": 0.0037, + "step": 89730 + }, + { + "epoch": 13.939111525318422, + "grad_norm": 0.07108212262392044, + "learning_rate": 2.8510235382092853e-06, + "loss": 0.0029, + "step": 89740 + }, + { + "epoch": 13.940664802733767, + "grad_norm": 0.19390130043029785, + "learning_rate": 2.845522530047029e-06, + "loss": 0.0035, + "step": 89750 + }, + { + "epoch": 13.942218080149114, + "grad_norm": 0.788049578666687, + "learning_rate": 2.8400266786275387e-06, + "loss": 0.0027, + "step": 89760 + }, + { + "epoch": 13.943771357564462, + "grad_norm": 0.08155955374240875, + "learning_rate": 2.834535984551817e-06, + "loss": 0.0034, + "step": 89770 + }, + { + "epoch": 13.945324634979807, + "grad_norm": 0.11044920235872269, + "learning_rate": 2.8290504484203397e-06, + "loss": 0.0032, + "step": 89780 + }, + { + "epoch": 13.946877912395154, + "grad_norm": 0.41046279668807983, + "learning_rate": 2.8235700708329705e-06, + "loss": 0.0034, + "step": 89790 + }, + { + "epoch": 13.948431189810501, + "grad_norm": 0.09953475743532181, + "learning_rate": 2.818094852389058e-06, + "loss": 0.0045, + "step": 89800 + }, + { + "epoch": 13.949984467225846, + "grad_norm": 0.5034199953079224, + "learning_rate": 2.8126247936873516e-06, + "loss": 0.0032, + "step": 89810 + }, + { + "epoch": 13.951537744641193, + "grad_norm": 0.4652537405490875, + "learning_rate": 2.8071598953260614e-06, + "loss": 0.0048, + "step": 89820 + }, + { + "epoch": 13.953091022056539, + "grad_norm": 0.17770463228225708, + "learning_rate": 2.8017001579028035e-06, + "loss": 0.0026, + "step": 89830 + }, + { + "epoch": 13.954644299471886, + "grad_norm": 0.444657564163208, + "learning_rate": 2.7962455820146617e-06, + "loss": 0.0037, + "step": 89840 + }, + { + "epoch": 13.956197576887233, + "grad_norm": 0.1532718986272812, + "learning_rate": 2.7907961682581253e-06, + "loss": 0.0046, + "step": 89850 + }, + { + "epoch": 13.957750854302578, + "grad_norm": 0.23314808309078217, + "learning_rate": 2.7853519172291453e-06, + "loss": 0.0032, + "step": 89860 + }, + { + "epoch": 13.959304131717925, + "grad_norm": 0.6055729389190674, + "learning_rate": 2.7799128295230904e-06, + "loss": 0.0027, + "step": 89870 + }, + { + "epoch": 13.960857409133272, + "grad_norm": 0.20974920690059662, + "learning_rate": 2.774478905734784e-06, + "loss": 0.0044, + "step": 89880 + }, + { + "epoch": 13.962410686548617, + "grad_norm": 0.5194052457809448, + "learning_rate": 2.7690501464584516e-06, + "loss": 0.0039, + "step": 89890 + }, + { + "epoch": 13.963963963963964, + "grad_norm": 0.20329153537750244, + "learning_rate": 2.763626552287796e-06, + "loss": 0.0043, + "step": 89900 + }, + { + "epoch": 13.96551724137931, + "grad_norm": 0.2295408695936203, + "learning_rate": 2.7582081238159083e-06, + "loss": 0.004, + "step": 89910 + }, + { + "epoch": 13.967070518794657, + "grad_norm": 0.14818179607391357, + "learning_rate": 2.7527948616353548e-06, + "loss": 0.0026, + "step": 89920 + }, + { + "epoch": 13.968623796210004, + "grad_norm": 0.07662753015756607, + "learning_rate": 2.7473867663381226e-06, + "loss": 0.0038, + "step": 89930 + }, + { + "epoch": 13.970177073625349, + "grad_norm": 0.15371553599834442, + "learning_rate": 2.7419838385156383e-06, + "loss": 0.0027, + "step": 89940 + }, + { + "epoch": 13.971730351040696, + "grad_norm": 0.15733231604099274, + "learning_rate": 2.7365860787587407e-06, + "loss": 0.0034, + "step": 89950 + }, + { + "epoch": 13.973283628456041, + "grad_norm": 0.16436854004859924, + "learning_rate": 2.7311934876577295e-06, + "loss": 0.0039, + "step": 89960 + }, + { + "epoch": 13.974836905871388, + "grad_norm": 0.1119225025177002, + "learning_rate": 2.725806065802339e-06, + "loss": 0.0018, + "step": 89970 + }, + { + "epoch": 13.976390183286735, + "grad_norm": 0.7870364785194397, + "learning_rate": 2.720423813781714e-06, + "loss": 0.0051, + "step": 89980 + }, + { + "epoch": 13.97794346070208, + "grad_norm": 0.3911462724208832, + "learning_rate": 2.715046732184462e-06, + "loss": 0.0031, + "step": 89990 + }, + { + "epoch": 13.979496738117428, + "grad_norm": 0.3522031605243683, + "learning_rate": 2.7096748215986013e-06, + "loss": 0.0028, + "step": 90000 + }, + { + "epoch": 13.981050015532775, + "grad_norm": 0.46878349781036377, + "learning_rate": 2.704308082611612e-06, + "loss": 0.0026, + "step": 90010 + }, + { + "epoch": 13.98260329294812, + "grad_norm": 0.13459397852420807, + "learning_rate": 2.698946515810374e-06, + "loss": 0.0042, + "step": 90020 + }, + { + "epoch": 13.984156570363467, + "grad_norm": 0.1008145660161972, + "learning_rate": 2.6935901217812367e-06, + "loss": 0.0076, + "step": 90030 + }, + { + "epoch": 13.985709847778814, + "grad_norm": 0.4363640546798706, + "learning_rate": 2.688238901109952e-06, + "loss": 0.0048, + "step": 90040 + }, + { + "epoch": 13.98726312519416, + "grad_norm": 0.347302109003067, + "learning_rate": 2.68289285438173e-06, + "loss": 0.0016, + "step": 90050 + }, + { + "epoch": 13.988816402609507, + "grad_norm": 0.22911742329597473, + "learning_rate": 2.6775519821812033e-06, + "loss": 0.0032, + "step": 90060 + }, + { + "epoch": 13.990369680024852, + "grad_norm": 0.21454884111881256, + "learning_rate": 2.672216285092449e-06, + "loss": 0.0036, + "step": 90070 + }, + { + "epoch": 13.991922957440199, + "grad_norm": 0.22365540266036987, + "learning_rate": 2.666885763698962e-06, + "loss": 0.0037, + "step": 90080 + }, + { + "epoch": 13.993476234855546, + "grad_norm": 0.1796766221523285, + "learning_rate": 2.661560418583692e-06, + "loss": 0.0077, + "step": 90090 + }, + { + "epoch": 13.995029512270891, + "grad_norm": 0.09989383816719055, + "learning_rate": 2.656240250328995e-06, + "loss": 0.0035, + "step": 90100 + }, + { + "epoch": 13.996582789686238, + "grad_norm": 0.18798315525054932, + "learning_rate": 2.650925259516679e-06, + "loss": 0.0026, + "step": 90110 + }, + { + "epoch": 13.998136067101584, + "grad_norm": 0.10085374116897583, + "learning_rate": 2.645615446727995e-06, + "loss": 0.004, + "step": 90120 + }, + { + "epoch": 13.99968934451693, + "grad_norm": 0.1291380077600479, + "learning_rate": 2.6403108125436115e-06, + "loss": 0.0028, + "step": 90130 + }, + { + "epoch": 14.001242621932278, + "grad_norm": 0.5200474858283997, + "learning_rate": 2.6350113575436264e-06, + "loss": 0.0038, + "step": 90140 + }, + { + "epoch": 14.002795899347623, + "grad_norm": 0.05166244134306908, + "learning_rate": 2.629717082307598e-06, + "loss": 0.0032, + "step": 90150 + }, + { + "epoch": 14.00434917676297, + "grad_norm": 0.11524758487939835, + "learning_rate": 2.6244279874144793e-06, + "loss": 0.0027, + "step": 90160 + }, + { + "epoch": 14.005902454178317, + "grad_norm": 0.11080348491668701, + "learning_rate": 2.6191440734426862e-06, + "loss": 0.0037, + "step": 90170 + }, + { + "epoch": 14.007455731593662, + "grad_norm": 0.3510868549346924, + "learning_rate": 2.613865340970062e-06, + "loss": 0.0036, + "step": 90180 + }, + { + "epoch": 14.00900900900901, + "grad_norm": 0.11258184164762497, + "learning_rate": 2.608591790573889e-06, + "loss": 0.0039, + "step": 90190 + }, + { + "epoch": 14.010562286424355, + "grad_norm": 0.10905561596155167, + "learning_rate": 2.603323422830861e-06, + "loss": 0.0021, + "step": 90200 + }, + { + "epoch": 14.012115563839702, + "grad_norm": 0.17594024538993835, + "learning_rate": 2.5980602383171183e-06, + "loss": 0.0032, + "step": 90210 + }, + { + "epoch": 14.013668841255049, + "grad_norm": 0.08994173258543015, + "learning_rate": 2.5928022376082496e-06, + "loss": 0.0023, + "step": 90220 + }, + { + "epoch": 14.015222118670394, + "grad_norm": 0.1694878488779068, + "learning_rate": 2.5875494212792505e-06, + "loss": 0.0031, + "step": 90230 + }, + { + "epoch": 14.016775396085741, + "grad_norm": 0.24232858419418335, + "learning_rate": 2.5823017899045565e-06, + "loss": 0.002, + "step": 90240 + }, + { + "epoch": 14.018328673501088, + "grad_norm": 0.4176792800426483, + "learning_rate": 2.577059344058047e-06, + "loss": 0.0046, + "step": 90250 + }, + { + "epoch": 14.019881950916433, + "grad_norm": 0.07444755733013153, + "learning_rate": 2.5718220843130415e-06, + "loss": 0.0022, + "step": 90260 + }, + { + "epoch": 14.02143522833178, + "grad_norm": 0.27528390288352966, + "learning_rate": 2.5665900112422535e-06, + "loss": 0.0034, + "step": 90270 + }, + { + "epoch": 14.022988505747126, + "grad_norm": 0.2299262285232544, + "learning_rate": 2.561363125417876e-06, + "loss": 0.0046, + "step": 90280 + }, + { + "epoch": 14.024541783162473, + "grad_norm": 0.3418178856372833, + "learning_rate": 2.556141427411496e-06, + "loss": 0.0029, + "step": 90290 + }, + { + "epoch": 14.02609506057782, + "grad_norm": 0.6128387451171875, + "learning_rate": 2.550924917794156e-06, + "loss": 0.0035, + "step": 90300 + }, + { + "epoch": 14.027648337993165, + "grad_norm": 0.20506887137889862, + "learning_rate": 2.545713597136329e-06, + "loss": 0.0054, + "step": 90310 + }, + { + "epoch": 14.029201615408512, + "grad_norm": 0.11417394131422043, + "learning_rate": 2.5405074660079298e-06, + "loss": 0.0041, + "step": 90320 + }, + { + "epoch": 14.030754892823857, + "grad_norm": 0.438774973154068, + "learning_rate": 2.5353065249782647e-06, + "loss": 0.0028, + "step": 90330 + }, + { + "epoch": 14.032308170239205, + "grad_norm": 0.7027615904808044, + "learning_rate": 2.530110774616129e-06, + "loss": 0.0021, + "step": 90340 + }, + { + "epoch": 14.033861447654552, + "grad_norm": 0.11464565247297287, + "learning_rate": 2.524920215489701e-06, + "loss": 0.0023, + "step": 90350 + }, + { + "epoch": 14.035414725069897, + "grad_norm": 0.2154930830001831, + "learning_rate": 2.519734848166627e-06, + "loss": 0.0025, + "step": 90360 + }, + { + "epoch": 14.036968002485244, + "grad_norm": 0.14035643637180328, + "learning_rate": 2.514554673213948e-06, + "loss": 0.0027, + "step": 90370 + }, + { + "epoch": 14.038521279900591, + "grad_norm": 0.3553699553012848, + "learning_rate": 2.5093796911981936e-06, + "loss": 0.0034, + "step": 90380 + }, + { + "epoch": 14.040074557315936, + "grad_norm": 0.15443110466003418, + "learning_rate": 2.5042099026852672e-06, + "loss": 0.0044, + "step": 90390 + }, + { + "epoch": 14.041627834731283, + "grad_norm": 0.2139890193939209, + "learning_rate": 2.499045308240544e-06, + "loss": 0.0046, + "step": 90400 + }, + { + "epoch": 14.043181112146629, + "grad_norm": 0.2583225965499878, + "learning_rate": 2.4938859084288003e-06, + "loss": 0.0029, + "step": 90410 + }, + { + "epoch": 14.044734389561976, + "grad_norm": 0.36456382274627686, + "learning_rate": 2.4887317038142734e-06, + "loss": 0.0042, + "step": 90420 + }, + { + "epoch": 14.046287666977323, + "grad_norm": 0.4770275354385376, + "learning_rate": 2.4835826949606124e-06, + "loss": 0.0033, + "step": 90430 + }, + { + "epoch": 14.047840944392668, + "grad_norm": 0.06281634420156479, + "learning_rate": 2.4784388824309167e-06, + "loss": 0.0037, + "step": 90440 + }, + { + "epoch": 14.049394221808015, + "grad_norm": 0.08406046032905579, + "learning_rate": 2.473300266787687e-06, + "loss": 0.004, + "step": 90450 + }, + { + "epoch": 14.050947499223362, + "grad_norm": 3.6081202030181885, + "learning_rate": 2.46816684859289e-06, + "loss": 0.0037, + "step": 90460 + }, + { + "epoch": 14.052500776638707, + "grad_norm": 0.09084528684616089, + "learning_rate": 2.4630386284079043e-06, + "loss": 0.0029, + "step": 90470 + }, + { + "epoch": 14.054054054054054, + "grad_norm": 0.1207643523812294, + "learning_rate": 2.4579156067935427e-06, + "loss": 0.005, + "step": 90480 + }, + { + "epoch": 14.0556073314694, + "grad_norm": 0.11403881013393402, + "learning_rate": 2.452797784310046e-06, + "loss": 0.0033, + "step": 90490 + }, + { + "epoch": 14.057160608884747, + "grad_norm": 0.24790742993354797, + "learning_rate": 2.447685161517105e-06, + "loss": 0.0049, + "step": 90500 + }, + { + "epoch": 14.058713886300094, + "grad_norm": 0.04220765456557274, + "learning_rate": 2.4425777389738224e-06, + "loss": 0.0021, + "step": 90510 + }, + { + "epoch": 14.060267163715439, + "grad_norm": 0.3594664931297302, + "learning_rate": 2.43747551723873e-06, + "loss": 0.0038, + "step": 90520 + }, + { + "epoch": 14.061820441130786, + "grad_norm": 0.2452695667743683, + "learning_rate": 2.4323784968698136e-06, + "loss": 0.003, + "step": 90530 + }, + { + "epoch": 14.063373718546133, + "grad_norm": 0.11778180301189423, + "learning_rate": 2.4272866784244565e-06, + "loss": 0.0035, + "step": 90540 + }, + { + "epoch": 14.064926995961478, + "grad_norm": 0.44343477487564087, + "learning_rate": 2.422200062459518e-06, + "loss": 0.0041, + "step": 90550 + }, + { + "epoch": 14.066480273376825, + "grad_norm": 1.0508512258529663, + "learning_rate": 2.41711864953123e-06, + "loss": 0.0038, + "step": 90560 + }, + { + "epoch": 14.06803355079217, + "grad_norm": 0.9995453357696533, + "learning_rate": 2.412042440195322e-06, + "loss": 0.0031, + "step": 90570 + }, + { + "epoch": 14.069586828207518, + "grad_norm": 0.7044494152069092, + "learning_rate": 2.4069714350068993e-06, + "loss": 0.0035, + "step": 90580 + }, + { + "epoch": 14.071140105622865, + "grad_norm": 0.25877049565315247, + "learning_rate": 2.4019056345205294e-06, + "loss": 0.0054, + "step": 90590 + }, + { + "epoch": 14.07269338303821, + "grad_norm": 0.08247864246368408, + "learning_rate": 2.396845039290191e-06, + "loss": 0.0028, + "step": 90600 + }, + { + "epoch": 14.074246660453557, + "grad_norm": 0.06677687168121338, + "learning_rate": 2.3917896498693194e-06, + "loss": 0.0055, + "step": 90610 + }, + { + "epoch": 14.075799937868904, + "grad_norm": 0.45538216829299927, + "learning_rate": 2.3867394668107335e-06, + "loss": 0.0044, + "step": 90620 + }, + { + "epoch": 14.07735321528425, + "grad_norm": 0.1393604427576065, + "learning_rate": 2.3816944906667527e-06, + "loss": 0.0047, + "step": 90630 + }, + { + "epoch": 14.078906492699597, + "grad_norm": 0.16864806413650513, + "learning_rate": 2.3766547219890634e-06, + "loss": 0.0035, + "step": 90640 + }, + { + "epoch": 14.080459770114942, + "grad_norm": 0.11926407366991043, + "learning_rate": 2.371620161328819e-06, + "loss": 0.0023, + "step": 90650 + }, + { + "epoch": 14.082013047530289, + "grad_norm": 0.31043919920921326, + "learning_rate": 2.366590809236574e-06, + "loss": 0.0047, + "step": 90660 + }, + { + "epoch": 14.083566324945636, + "grad_norm": 0.24878060817718506, + "learning_rate": 2.3615666662623494e-06, + "loss": 0.0031, + "step": 90670 + }, + { + "epoch": 14.085119602360981, + "grad_norm": 0.169907346367836, + "learning_rate": 2.3565477329555664e-06, + "loss": 0.0031, + "step": 90680 + }, + { + "epoch": 14.086672879776328, + "grad_norm": 0.28480851650238037, + "learning_rate": 2.3515340098650975e-06, + "loss": 0.0035, + "step": 90690 + }, + { + "epoch": 14.088226157191674, + "grad_norm": 0.2344961166381836, + "learning_rate": 2.3465254975392258e-06, + "loss": 0.0039, + "step": 90700 + }, + { + "epoch": 14.08977943460702, + "grad_norm": 0.3268965780735016, + "learning_rate": 2.3415221965256805e-06, + "loss": 0.0037, + "step": 90710 + }, + { + "epoch": 14.091332712022368, + "grad_norm": 0.2480953335762024, + "learning_rate": 2.3365241073716237e-06, + "loss": 0.0029, + "step": 90720 + }, + { + "epoch": 14.092885989437713, + "grad_norm": 0.10231126844882965, + "learning_rate": 2.3315312306236183e-06, + "loss": 0.0022, + "step": 90730 + }, + { + "epoch": 14.09443926685306, + "grad_norm": 0.10824563354253769, + "learning_rate": 2.326543566827699e-06, + "loss": 0.0025, + "step": 90740 + }, + { + "epoch": 14.095992544268407, + "grad_norm": 0.4669142961502075, + "learning_rate": 2.3215611165292862e-06, + "loss": 0.0041, + "step": 90750 + }, + { + "epoch": 14.097545821683752, + "grad_norm": 1.5773779153823853, + "learning_rate": 2.3165838802732765e-06, + "loss": 0.0025, + "step": 90760 + }, + { + "epoch": 14.0990990990991, + "grad_norm": 0.09401612728834152, + "learning_rate": 2.311611858603957e-06, + "loss": 0.0025, + "step": 90770 + }, + { + "epoch": 14.100652376514445, + "grad_norm": 0.2717852294445038, + "learning_rate": 2.3066450520650763e-06, + "loss": 0.0046, + "step": 90780 + }, + { + "epoch": 14.102205653929792, + "grad_norm": 0.24356484413146973, + "learning_rate": 2.3016834611997773e-06, + "loss": 0.0037, + "step": 90790 + }, + { + "epoch": 14.103758931345139, + "grad_norm": 0.08236187696456909, + "learning_rate": 2.296727086550665e-06, + "loss": 0.0023, + "step": 90800 + }, + { + "epoch": 14.105312208760484, + "grad_norm": 0.14308448135852814, + "learning_rate": 2.2917759286597507e-06, + "loss": 0.0025, + "step": 90810 + }, + { + "epoch": 14.106865486175831, + "grad_norm": 0.14165861904621124, + "learning_rate": 2.2868299880685006e-06, + "loss": 0.0039, + "step": 90820 + }, + { + "epoch": 14.108418763591178, + "grad_norm": 0.16027972102165222, + "learning_rate": 2.2818892653177827e-06, + "loss": 0.0026, + "step": 90830 + }, + { + "epoch": 14.109972041006523, + "grad_norm": 0.4920848608016968, + "learning_rate": 2.2769537609479196e-06, + "loss": 0.0033, + "step": 90840 + }, + { + "epoch": 14.11152531842187, + "grad_norm": 0.14828403294086456, + "learning_rate": 2.2720234754986357e-06, + "loss": 0.0032, + "step": 90850 + }, + { + "epoch": 14.113078595837216, + "grad_norm": 0.7444202899932861, + "learning_rate": 2.2670984095091108e-06, + "loss": 0.0026, + "step": 90860 + }, + { + "epoch": 14.114631873252563, + "grad_norm": 0.12725554406642914, + "learning_rate": 2.262178563517936e-06, + "loss": 0.0035, + "step": 90870 + }, + { + "epoch": 14.11618515066791, + "grad_norm": 0.2033882737159729, + "learning_rate": 2.257263938063137e-06, + "loss": 0.0042, + "step": 90880 + }, + { + "epoch": 14.117738428083255, + "grad_norm": 0.442676305770874, + "learning_rate": 2.2523545336821715e-06, + "loss": 0.0051, + "step": 90890 + }, + { + "epoch": 14.119291705498602, + "grad_norm": 0.16419175267219543, + "learning_rate": 2.247450350911939e-06, + "loss": 0.0041, + "step": 90900 + }, + { + "epoch": 14.12084498291395, + "grad_norm": 0.3280315697193146, + "learning_rate": 2.2425513902887328e-06, + "loss": 0.0033, + "step": 90910 + }, + { + "epoch": 14.122398260329295, + "grad_norm": 0.2475491464138031, + "learning_rate": 2.2376576523483015e-06, + "loss": 0.0046, + "step": 90920 + }, + { + "epoch": 14.123951537744642, + "grad_norm": 0.730791449546814, + "learning_rate": 2.232769137625829e-06, + "loss": 0.0054, + "step": 90930 + }, + { + "epoch": 14.125504815159987, + "grad_norm": 0.2435154765844345, + "learning_rate": 2.227885846655903e-06, + "loss": 0.0055, + "step": 90940 + }, + { + "epoch": 14.127058092575334, + "grad_norm": 0.1261776089668274, + "learning_rate": 2.2230077799725523e-06, + "loss": 0.0044, + "step": 90950 + }, + { + "epoch": 14.128611369990681, + "grad_norm": 0.054443854838609695, + "learning_rate": 2.2181349381092396e-06, + "loss": 0.0033, + "step": 90960 + }, + { + "epoch": 14.130164647406026, + "grad_norm": 0.12088106572628021, + "learning_rate": 2.2132673215988552e-06, + "loss": 0.003, + "step": 90970 + }, + { + "epoch": 14.131717924821373, + "grad_norm": 0.297175794839859, + "learning_rate": 2.2084049309737065e-06, + "loss": 0.0026, + "step": 90980 + }, + { + "epoch": 14.133271202236719, + "grad_norm": 0.5040339827537537, + "learning_rate": 2.2035477667655513e-06, + "loss": 0.0029, + "step": 90990 + }, + { + "epoch": 14.134824479652066, + "grad_norm": 0.28665557503700256, + "learning_rate": 2.1986958295055315e-06, + "loss": 0.0028, + "step": 91000 + }, + { + "epoch": 14.136377757067413, + "grad_norm": 0.20217610895633698, + "learning_rate": 2.193849119724284e-06, + "loss": 0.0037, + "step": 91010 + }, + { + "epoch": 14.137931034482758, + "grad_norm": 0.19204576313495636, + "learning_rate": 2.1890076379518175e-06, + "loss": 0.005, + "step": 91020 + }, + { + "epoch": 14.139484311898105, + "grad_norm": 0.4534927010536194, + "learning_rate": 2.1841713847175925e-06, + "loss": 0.0031, + "step": 91030 + }, + { + "epoch": 14.141037589313452, + "grad_norm": 0.3605010509490967, + "learning_rate": 2.1793403605504968e-06, + "loss": 0.0029, + "step": 91040 + }, + { + "epoch": 14.142590866728797, + "grad_norm": 0.19428303837776184, + "learning_rate": 2.1745145659788413e-06, + "loss": 0.0033, + "step": 91050 + }, + { + "epoch": 14.144144144144144, + "grad_norm": 0.5432263612747192, + "learning_rate": 2.1696940015303645e-06, + "loss": 0.0028, + "step": 91060 + }, + { + "epoch": 14.14569742155949, + "grad_norm": 0.3040321171283722, + "learning_rate": 2.1648786677322397e-06, + "loss": 0.0031, + "step": 91070 + }, + { + "epoch": 14.147250698974837, + "grad_norm": 0.6169268488883972, + "learning_rate": 2.1600685651110673e-06, + "loss": 0.0036, + "step": 91080 + }, + { + "epoch": 14.148803976390184, + "grad_norm": 0.21184030175209045, + "learning_rate": 2.1552636941928717e-06, + "loss": 0.0026, + "step": 91090 + }, + { + "epoch": 14.15035725380553, + "grad_norm": 0.3605688512325287, + "learning_rate": 2.1504640555031043e-06, + "loss": 0.0047, + "step": 91100 + }, + { + "epoch": 14.151910531220876, + "grad_norm": 0.20297560095787048, + "learning_rate": 2.1456696495666506e-06, + "loss": 0.0026, + "step": 91110 + }, + { + "epoch": 14.153463808636223, + "grad_norm": 0.22465308010578156, + "learning_rate": 2.1408804769078084e-06, + "loss": 0.0051, + "step": 91120 + }, + { + "epoch": 14.155017086051568, + "grad_norm": 0.13957533240318298, + "learning_rate": 2.136096538050325e-06, + "loss": 0.0038, + "step": 91130 + }, + { + "epoch": 14.156570363466916, + "grad_norm": 0.06585568189620972, + "learning_rate": 2.13131783351736e-06, + "loss": 0.0046, + "step": 91140 + }, + { + "epoch": 14.15812364088226, + "grad_norm": 0.25310173630714417, + "learning_rate": 2.126544363831512e-06, + "loss": 0.0022, + "step": 91150 + }, + { + "epoch": 14.159676918297608, + "grad_norm": 0.1747933179140091, + "learning_rate": 2.121776129514791e-06, + "loss": 0.0023, + "step": 91160 + }, + { + "epoch": 14.161230195712955, + "grad_norm": 0.17898041009902954, + "learning_rate": 2.1170131310886523e-06, + "loss": 0.0029, + "step": 91170 + }, + { + "epoch": 14.1627834731283, + "grad_norm": 0.3210783898830414, + "learning_rate": 2.112255369073968e-06, + "loss": 0.003, + "step": 91180 + }, + { + "epoch": 14.164336750543647, + "grad_norm": 0.16324898600578308, + "learning_rate": 2.1075028439910328e-06, + "loss": 0.0035, + "step": 91190 + }, + { + "epoch": 14.165890027958994, + "grad_norm": 0.406013160943985, + "learning_rate": 2.1027555563595813e-06, + "loss": 0.0031, + "step": 91200 + }, + { + "epoch": 14.16744330537434, + "grad_norm": 1.2392057180404663, + "learning_rate": 2.0980135066987706e-06, + "loss": 0.0027, + "step": 91210 + }, + { + "epoch": 14.168996582789687, + "grad_norm": 0.6025571227073669, + "learning_rate": 2.093276695527191e-06, + "loss": 0.0043, + "step": 91220 + }, + { + "epoch": 14.170549860205032, + "grad_norm": 0.08293837308883667, + "learning_rate": 2.0885451233628396e-06, + "loss": 0.0039, + "step": 91230 + }, + { + "epoch": 14.172103137620379, + "grad_norm": 0.11715065687894821, + "learning_rate": 2.0838187907231633e-06, + "loss": 0.0035, + "step": 91240 + }, + { + "epoch": 14.173656415035726, + "grad_norm": 0.2028699666261673, + "learning_rate": 2.0790976981250153e-06, + "loss": 0.0022, + "step": 91250 + }, + { + "epoch": 14.175209692451071, + "grad_norm": 0.03976826369762421, + "learning_rate": 2.0743818460847e-06, + "loss": 0.0029, + "step": 91260 + }, + { + "epoch": 14.176762969866418, + "grad_norm": 0.06714428216218948, + "learning_rate": 2.069671235117926e-06, + "loss": 0.0022, + "step": 91270 + }, + { + "epoch": 14.178316247281764, + "grad_norm": 0.3402387797832489, + "learning_rate": 2.0649658657398484e-06, + "loss": 0.0034, + "step": 91280 + }, + { + "epoch": 14.17986952469711, + "grad_norm": 0.4913824498653412, + "learning_rate": 2.060265738465034e-06, + "loss": 0.0044, + "step": 91290 + }, + { + "epoch": 14.181422802112458, + "grad_norm": 0.12130588293075562, + "learning_rate": 2.055570853807487e-06, + "loss": 0.0034, + "step": 91300 + }, + { + "epoch": 14.182976079527803, + "grad_norm": 0.511326014995575, + "learning_rate": 2.0508812122806196e-06, + "loss": 0.0042, + "step": 91310 + }, + { + "epoch": 14.18452935694315, + "grad_norm": 0.2501344084739685, + "learning_rate": 2.046196814397294e-06, + "loss": 0.0029, + "step": 91320 + }, + { + "epoch": 14.186082634358497, + "grad_norm": 0.18996314704418182, + "learning_rate": 2.0415176606697896e-06, + "loss": 0.0027, + "step": 91330 + }, + { + "epoch": 14.187635911773842, + "grad_norm": 0.07607777416706085, + "learning_rate": 2.0368437516098126e-06, + "loss": 0.0017, + "step": 91340 + }, + { + "epoch": 14.18918918918919, + "grad_norm": 0.38875165581703186, + "learning_rate": 2.032175087728483e-06, + "loss": 0.0035, + "step": 91350 + }, + { + "epoch": 14.190742466604535, + "grad_norm": 0.11259061098098755, + "learning_rate": 2.027511669536375e-06, + "loss": 0.004, + "step": 91360 + }, + { + "epoch": 14.192295744019882, + "grad_norm": 0.13668511807918549, + "learning_rate": 2.0228534975434643e-06, + "loss": 0.0031, + "step": 91370 + }, + { + "epoch": 14.193849021435229, + "grad_norm": 0.2657930254936218, + "learning_rate": 2.0182005722591548e-06, + "loss": 0.0039, + "step": 91380 + }, + { + "epoch": 14.195402298850574, + "grad_norm": 0.11134219169616699, + "learning_rate": 2.0135528941922944e-06, + "loss": 0.0016, + "step": 91390 + }, + { + "epoch": 14.196955576265921, + "grad_norm": 0.20009711384773254, + "learning_rate": 2.008910463851149e-06, + "loss": 0.0047, + "step": 91400 + }, + { + "epoch": 14.198508853681268, + "grad_norm": 0.38144636154174805, + "learning_rate": 2.0042732817433963e-06, + "loss": 0.0023, + "step": 91410 + }, + { + "epoch": 14.200062131096614, + "grad_norm": 0.11669957637786865, + "learning_rate": 1.999641348376158e-06, + "loss": 0.0031, + "step": 91420 + }, + { + "epoch": 14.20161540851196, + "grad_norm": 0.1066427230834961, + "learning_rate": 1.9950146642559787e-06, + "loss": 0.0028, + "step": 91430 + }, + { + "epoch": 14.203168685927306, + "grad_norm": 0.07551544159650803, + "learning_rate": 1.990393229888815e-06, + "loss": 0.0042, + "step": 91440 + }, + { + "epoch": 14.204721963342653, + "grad_norm": 0.4613092541694641, + "learning_rate": 1.9857770457800684e-06, + "loss": 0.004, + "step": 91450 + }, + { + "epoch": 14.206275240758, + "grad_norm": 0.0992426946759224, + "learning_rate": 1.9811661124345517e-06, + "loss": 0.0028, + "step": 91460 + }, + { + "epoch": 14.207828518173345, + "grad_norm": 0.11777812987565994, + "learning_rate": 1.976560430356522e-06, + "loss": 0.0033, + "step": 91470 + }, + { + "epoch": 14.209381795588692, + "grad_norm": 1.129764437675476, + "learning_rate": 1.9719600000496383e-06, + "loss": 0.0046, + "step": 91480 + }, + { + "epoch": 14.21093507300404, + "grad_norm": 0.11792176216840744, + "learning_rate": 1.967364822017004e-06, + "loss": 0.003, + "step": 91490 + }, + { + "epoch": 14.212488350419385, + "grad_norm": 0.250274121761322, + "learning_rate": 1.962774896761127e-06, + "loss": 0.0029, + "step": 91500 + }, + { + "epoch": 14.214041627834732, + "grad_norm": 0.40933772921562195, + "learning_rate": 1.958190224783968e-06, + "loss": 0.0042, + "step": 91510 + }, + { + "epoch": 14.215594905250077, + "grad_norm": 0.3606562614440918, + "learning_rate": 1.9536108065868975e-06, + "loss": 0.0052, + "step": 91520 + }, + { + "epoch": 14.217148182665424, + "grad_norm": 0.04512186348438263, + "learning_rate": 1.949036642670715e-06, + "loss": 0.0024, + "step": 91530 + }, + { + "epoch": 14.218701460080771, + "grad_norm": 0.7846224904060364, + "learning_rate": 1.9444677335356377e-06, + "loss": 0.0038, + "step": 91540 + }, + { + "epoch": 14.220254737496116, + "grad_norm": 0.11735707521438599, + "learning_rate": 1.939904079681326e-06, + "loss": 0.0037, + "step": 91550 + }, + { + "epoch": 14.221808014911463, + "grad_norm": 0.04224707558751106, + "learning_rate": 1.935345681606837e-06, + "loss": 0.0027, + "step": 91560 + }, + { + "epoch": 14.22336129232681, + "grad_norm": 0.08978762477636337, + "learning_rate": 1.930792539810694e-06, + "loss": 0.0031, + "step": 91570 + }, + { + "epoch": 14.224914569742156, + "grad_norm": 0.26540035009384155, + "learning_rate": 1.9262446547907877e-06, + "loss": 0.0039, + "step": 91580 + }, + { + "epoch": 14.226467847157503, + "grad_norm": 0.34115687012672424, + "learning_rate": 1.921702027044503e-06, + "loss": 0.004, + "step": 91590 + }, + { + "epoch": 14.228021124572848, + "grad_norm": 0.15561890602111816, + "learning_rate": 1.917164657068593e-06, + "loss": 0.0038, + "step": 91600 + }, + { + "epoch": 14.229574401988195, + "grad_norm": 0.3738061189651489, + "learning_rate": 1.912632545359272e-06, + "loss": 0.0027, + "step": 91610 + }, + { + "epoch": 14.231127679403542, + "grad_norm": 0.05707774683833122, + "learning_rate": 1.9081056924121488e-06, + "loss": 0.0028, + "step": 91620 + }, + { + "epoch": 14.232680956818887, + "grad_norm": 0.05332459509372711, + "learning_rate": 1.9035840987222887e-06, + "loss": 0.0028, + "step": 91630 + }, + { + "epoch": 14.234234234234235, + "grad_norm": 0.3472473621368408, + "learning_rate": 1.8990677647841515e-06, + "loss": 0.0036, + "step": 91640 + }, + { + "epoch": 14.23578751164958, + "grad_norm": 0.49260714650154114, + "learning_rate": 1.8945566910916534e-06, + "loss": 0.0034, + "step": 91650 + }, + { + "epoch": 14.237340789064927, + "grad_norm": 0.4434695243835449, + "learning_rate": 1.8900508781381054e-06, + "loss": 0.0024, + "step": 91660 + }, + { + "epoch": 14.238894066480274, + "grad_norm": 0.1985630989074707, + "learning_rate": 1.8855503264162633e-06, + "loss": 0.0028, + "step": 91670 + }, + { + "epoch": 14.24044734389562, + "grad_norm": 0.807144284248352, + "learning_rate": 1.8810550364182998e-06, + "loss": 0.004, + "step": 91680 + }, + { + "epoch": 14.242000621310966, + "grad_norm": 0.09434196352958679, + "learning_rate": 1.8765650086358112e-06, + "loss": 0.0017, + "step": 91690 + }, + { + "epoch": 14.243553898726313, + "grad_norm": 0.14351309835910797, + "learning_rate": 1.8720802435598151e-06, + "loss": 0.0031, + "step": 91700 + }, + { + "epoch": 14.245107176141659, + "grad_norm": 0.8181309103965759, + "learning_rate": 1.8676007416807694e-06, + "loss": 0.0044, + "step": 91710 + }, + { + "epoch": 14.246660453557006, + "grad_norm": 0.1343361884355545, + "learning_rate": 1.8631265034885437e-06, + "loss": 0.0028, + "step": 91720 + }, + { + "epoch": 14.24821373097235, + "grad_norm": 0.10979611426591873, + "learning_rate": 1.858657529472424e-06, + "loss": 0.0021, + "step": 91730 + }, + { + "epoch": 14.249767008387698, + "grad_norm": 0.31210342049598694, + "learning_rate": 1.8541938201211473e-06, + "loss": 0.003, + "step": 91740 + }, + { + "epoch": 14.251320285803045, + "grad_norm": 0.6171484589576721, + "learning_rate": 1.8497353759228398e-06, + "loss": 0.004, + "step": 91750 + }, + { + "epoch": 14.25287356321839, + "grad_norm": 0.0835859403014183, + "learning_rate": 1.8452821973650835e-06, + "loss": 0.0035, + "step": 91760 + }, + { + "epoch": 14.254426840633737, + "grad_norm": 0.46234485507011414, + "learning_rate": 1.840834284934856e-06, + "loss": 0.006, + "step": 91770 + }, + { + "epoch": 14.255980118049084, + "grad_norm": 0.09649311006069183, + "learning_rate": 1.8363916391185954e-06, + "loss": 0.004, + "step": 91780 + }, + { + "epoch": 14.25753339546443, + "grad_norm": 0.5553662180900574, + "learning_rate": 1.8319542604021245e-06, + "loss": 0.0032, + "step": 91790 + }, + { + "epoch": 14.259086672879777, + "grad_norm": 0.1819329708814621, + "learning_rate": 1.8275221492707273e-06, + "loss": 0.0034, + "step": 91800 + }, + { + "epoch": 14.260639950295122, + "grad_norm": 0.24206550419330597, + "learning_rate": 1.8230953062090717e-06, + "loss": 0.0039, + "step": 91810 + }, + { + "epoch": 14.262193227710469, + "grad_norm": 0.2337588369846344, + "learning_rate": 1.8186737317012926e-06, + "loss": 0.0029, + "step": 91820 + }, + { + "epoch": 14.263746505125816, + "grad_norm": 0.4547994136810303, + "learning_rate": 1.8142574262308977e-06, + "loss": 0.0023, + "step": 91830 + }, + { + "epoch": 14.265299782541161, + "grad_norm": 0.1502959430217743, + "learning_rate": 1.8098463902808727e-06, + "loss": 0.0037, + "step": 91840 + }, + { + "epoch": 14.266853059956508, + "grad_norm": 0.10427621752023697, + "learning_rate": 1.805440624333593e-06, + "loss": 0.004, + "step": 91850 + }, + { + "epoch": 14.268406337371854, + "grad_norm": 0.12540756165981293, + "learning_rate": 1.8010401288708734e-06, + "loss": 0.0018, + "step": 91860 + }, + { + "epoch": 14.2699596147872, + "grad_norm": 0.9183646440505981, + "learning_rate": 1.7966449043739343e-06, + "loss": 0.0033, + "step": 91870 + }, + { + "epoch": 14.271512892202548, + "grad_norm": 0.07948744297027588, + "learning_rate": 1.79225495132343e-06, + "loss": 0.0038, + "step": 91880 + }, + { + "epoch": 14.273066169617893, + "grad_norm": 0.25637078285217285, + "learning_rate": 1.7878702701994487e-06, + "loss": 0.0039, + "step": 91890 + }, + { + "epoch": 14.27461944703324, + "grad_norm": 0.2638019025325775, + "learning_rate": 1.783490861481496e-06, + "loss": 0.0031, + "step": 91900 + }, + { + "epoch": 14.276172724448587, + "grad_norm": 0.18753954768180847, + "learning_rate": 1.7791167256484832e-06, + "loss": 0.0026, + "step": 91910 + }, + { + "epoch": 14.277726001863932, + "grad_norm": 0.1039358526468277, + "learning_rate": 1.774747863178766e-06, + "loss": 0.0048, + "step": 91920 + }, + { + "epoch": 14.27927927927928, + "grad_norm": 0.1665470004081726, + "learning_rate": 1.7703842745501186e-06, + "loss": 0.0028, + "step": 91930 + }, + { + "epoch": 14.280832556694625, + "grad_norm": 0.16331252455711365, + "learning_rate": 1.7660259602397366e-06, + "loss": 0.004, + "step": 91940 + }, + { + "epoch": 14.282385834109972, + "grad_norm": 0.4834240972995758, + "learning_rate": 1.7616729207242388e-06, + "loss": 0.0033, + "step": 91950 + }, + { + "epoch": 14.283939111525319, + "grad_norm": 0.22618812322616577, + "learning_rate": 1.7573251564796556e-06, + "loss": 0.0051, + "step": 91960 + }, + { + "epoch": 14.285492388940664, + "grad_norm": 0.16684648394584656, + "learning_rate": 1.7529826679814676e-06, + "loss": 0.0046, + "step": 91970 + }, + { + "epoch": 14.287045666356011, + "grad_norm": 0.26372596621513367, + "learning_rate": 1.7486454557045562e-06, + "loss": 0.0026, + "step": 91980 + }, + { + "epoch": 14.288598943771358, + "grad_norm": 0.20257554948329926, + "learning_rate": 1.7443135201232364e-06, + "loss": 0.0036, + "step": 91990 + }, + { + "epoch": 14.290152221186704, + "grad_norm": 0.16222569346427917, + "learning_rate": 1.7399868617112347e-06, + "loss": 0.0026, + "step": 92000 + }, + { + "epoch": 14.29170549860205, + "grad_norm": 0.09057555347681046, + "learning_rate": 1.7356654809417171e-06, + "loss": 0.0037, + "step": 92010 + }, + { + "epoch": 14.293258776017396, + "grad_norm": 0.2608858346939087, + "learning_rate": 1.7313493782872448e-06, + "loss": 0.0031, + "step": 92020 + }, + { + "epoch": 14.294812053432743, + "grad_norm": 0.6543021202087402, + "learning_rate": 1.7270385542198453e-06, + "loss": 0.0055, + "step": 92030 + }, + { + "epoch": 14.29636533084809, + "grad_norm": 0.6830952167510986, + "learning_rate": 1.7227330092109307e-06, + "loss": 0.0035, + "step": 92040 + }, + { + "epoch": 14.297918608263435, + "grad_norm": 0.1429004967212677, + "learning_rate": 1.718432743731352e-06, + "loss": 0.0023, + "step": 92050 + }, + { + "epoch": 14.299471885678782, + "grad_norm": 0.28473928570747375, + "learning_rate": 1.7141377582513718e-06, + "loss": 0.0045, + "step": 92060 + }, + { + "epoch": 14.30102516309413, + "grad_norm": 0.41992324590682983, + "learning_rate": 1.709848053240698e-06, + "loss": 0.0041, + "step": 92070 + }, + { + "epoch": 14.302578440509475, + "grad_norm": 0.5239316821098328, + "learning_rate": 1.7055636291684328e-06, + "loss": 0.0033, + "step": 92080 + }, + { + "epoch": 14.304131717924822, + "grad_norm": 0.22833184897899628, + "learning_rate": 1.7012844865031185e-06, + "loss": 0.0034, + "step": 92090 + }, + { + "epoch": 14.305684995340167, + "grad_norm": 0.14552493393421173, + "learning_rate": 1.6970106257127138e-06, + "loss": 0.0062, + "step": 92100 + }, + { + "epoch": 14.307238272755514, + "grad_norm": 0.16043522953987122, + "learning_rate": 1.6927420472646117e-06, + "loss": 0.0042, + "step": 92110 + }, + { + "epoch": 14.308791550170861, + "grad_norm": 0.12063106894493103, + "learning_rate": 1.6884787516256106e-06, + "loss": 0.003, + "step": 92120 + }, + { + "epoch": 14.310344827586206, + "grad_norm": 0.31977832317352295, + "learning_rate": 1.6842207392619325e-06, + "loss": 0.0025, + "step": 92130 + }, + { + "epoch": 14.311898105001553, + "grad_norm": 0.06833496689796448, + "learning_rate": 1.679968010639238e-06, + "loss": 0.0035, + "step": 92140 + }, + { + "epoch": 14.3134513824169, + "grad_norm": 0.03046376071870327, + "learning_rate": 1.6757205662225829e-06, + "loss": 0.0023, + "step": 92150 + }, + { + "epoch": 14.315004659832246, + "grad_norm": 0.13571806252002716, + "learning_rate": 1.6714784064764844e-06, + "loss": 0.0038, + "step": 92160 + }, + { + "epoch": 14.316557937247593, + "grad_norm": 0.38591253757476807, + "learning_rate": 1.6672415318648383e-06, + "loss": 0.0031, + "step": 92170 + }, + { + "epoch": 14.318111214662938, + "grad_norm": 0.059365250170230865, + "learning_rate": 1.6630099428510016e-06, + "loss": 0.0028, + "step": 92180 + }, + { + "epoch": 14.319664492078285, + "grad_norm": 1.3428471088409424, + "learning_rate": 1.6587836398977153e-06, + "loss": 0.0037, + "step": 92190 + }, + { + "epoch": 14.321217769493632, + "grad_norm": 0.10730842500925064, + "learning_rate": 1.6545626234671762e-06, + "loss": 0.0024, + "step": 92200 + }, + { + "epoch": 14.322771046908978, + "grad_norm": 0.041098203510046005, + "learning_rate": 1.650346894020971e-06, + "loss": 0.0032, + "step": 92210 + }, + { + "epoch": 14.324324324324325, + "grad_norm": 0.9751054644584656, + "learning_rate": 1.6461364520201472e-06, + "loss": 0.0036, + "step": 92220 + }, + { + "epoch": 14.32587760173967, + "grad_norm": 1.0681086778640747, + "learning_rate": 1.6419312979251366e-06, + "loss": 0.005, + "step": 92230 + }, + { + "epoch": 14.327430879155017, + "grad_norm": 0.283186674118042, + "learning_rate": 1.6377314321958214e-06, + "loss": 0.0035, + "step": 92240 + }, + { + "epoch": 14.328984156570364, + "grad_norm": 0.1665160208940506, + "learning_rate": 1.6335368552914787e-06, + "loss": 0.0031, + "step": 92250 + }, + { + "epoch": 14.33053743398571, + "grad_norm": 0.083781898021698, + "learning_rate": 1.6293475676708358e-06, + "loss": 0.004, + "step": 92260 + }, + { + "epoch": 14.332090711401056, + "grad_norm": 0.1566161811351776, + "learning_rate": 1.6251635697920097e-06, + "loss": 0.0023, + "step": 92270 + }, + { + "epoch": 14.333643988816403, + "grad_norm": 0.12010283768177032, + "learning_rate": 1.6209848621125624e-06, + "loss": 0.0029, + "step": 92280 + }, + { + "epoch": 14.335197266231749, + "grad_norm": 0.05562485009431839, + "learning_rate": 1.6168114450894778e-06, + "loss": 0.0028, + "step": 92290 + }, + { + "epoch": 14.336750543647096, + "grad_norm": 0.30963483452796936, + "learning_rate": 1.612643319179158e-06, + "loss": 0.0028, + "step": 92300 + }, + { + "epoch": 14.338303821062441, + "grad_norm": 0.16249899566173553, + "learning_rate": 1.6084804848374046e-06, + "loss": 0.0033, + "step": 92310 + }, + { + "epoch": 14.339857098477788, + "grad_norm": 0.8386635184288025, + "learning_rate": 1.6043229425194817e-06, + "loss": 0.0044, + "step": 92320 + }, + { + "epoch": 14.341410375893135, + "grad_norm": 0.2416711002588272, + "learning_rate": 1.6001706926800308e-06, + "loss": 0.0039, + "step": 92330 + }, + { + "epoch": 14.34296365330848, + "grad_norm": 0.23965957760810852, + "learning_rate": 1.5960237357731444e-06, + "loss": 0.0021, + "step": 92340 + }, + { + "epoch": 14.344516930723827, + "grad_norm": 0.25460970401763916, + "learning_rate": 1.5918820722523264e-06, + "loss": 0.0061, + "step": 92350 + }, + { + "epoch": 14.346070208139174, + "grad_norm": 0.24390161037445068, + "learning_rate": 1.587745702570509e-06, + "loss": 0.003, + "step": 92360 + }, + { + "epoch": 14.34762348555452, + "grad_norm": 0.7802774310112, + "learning_rate": 1.5836146271800356e-06, + "loss": 0.0034, + "step": 92370 + }, + { + "epoch": 14.349176762969867, + "grad_norm": 0.0894826203584671, + "learning_rate": 1.5794888465326728e-06, + "loss": 0.0035, + "step": 92380 + }, + { + "epoch": 14.350730040385212, + "grad_norm": 0.09277123212814331, + "learning_rate": 1.5753683610796155e-06, + "loss": 0.0052, + "step": 92390 + }, + { + "epoch": 14.352283317800559, + "grad_norm": 0.21947155892848969, + "learning_rate": 1.5712531712714528e-06, + "loss": 0.0034, + "step": 92400 + }, + { + "epoch": 14.353836595215906, + "grad_norm": 0.1713954508304596, + "learning_rate": 1.5671432775582474e-06, + "loss": 0.0028, + "step": 92410 + }, + { + "epoch": 14.355389872631251, + "grad_norm": 0.16034550964832306, + "learning_rate": 1.563038680389428e-06, + "loss": 0.0027, + "step": 92420 + }, + { + "epoch": 14.356943150046598, + "grad_norm": 0.14182987809181213, + "learning_rate": 1.5589393802138808e-06, + "loss": 0.0016, + "step": 92430 + }, + { + "epoch": 14.358496427461946, + "grad_norm": 0.12273739278316498, + "learning_rate": 1.554845377479891e-06, + "loss": 0.0045, + "step": 92440 + }, + { + "epoch": 14.36004970487729, + "grad_norm": 0.3009955585002899, + "learning_rate": 1.550756672635173e-06, + "loss": 0.0034, + "step": 92450 + }, + { + "epoch": 14.361602982292638, + "grad_norm": 0.47055962681770325, + "learning_rate": 1.5466732661268635e-06, + "loss": 0.0027, + "step": 92460 + }, + { + "epoch": 14.363156259707983, + "grad_norm": 0.43086957931518555, + "learning_rate": 1.5425951584015163e-06, + "loss": 0.0039, + "step": 92470 + }, + { + "epoch": 14.36470953712333, + "grad_norm": 0.12431784719228745, + "learning_rate": 1.5385223499051026e-06, + "loss": 0.0039, + "step": 92480 + }, + { + "epoch": 14.366262814538677, + "grad_norm": 0.2080470472574234, + "learning_rate": 1.5344548410830328e-06, + "loss": 0.0044, + "step": 92490 + }, + { + "epoch": 14.367816091954023, + "grad_norm": 0.10351238399744034, + "learning_rate": 1.5303926323801121e-06, + "loss": 0.0031, + "step": 92500 + }, + { + "epoch": 14.36936936936937, + "grad_norm": 0.07880205661058426, + "learning_rate": 1.5263357242405795e-06, + "loss": 0.0034, + "step": 92510 + }, + { + "epoch": 14.370922646784717, + "grad_norm": 0.23707622289657593, + "learning_rate": 1.5222841171080915e-06, + "loss": 0.0039, + "step": 92520 + }, + { + "epoch": 14.372475924200062, + "grad_norm": 0.10677875578403473, + "learning_rate": 1.5182378114257268e-06, + "loss": 0.0025, + "step": 92530 + }, + { + "epoch": 14.374029201615409, + "grad_norm": 0.4365364611148834, + "learning_rate": 1.5141968076359869e-06, + "loss": 0.0065, + "step": 92540 + }, + { + "epoch": 14.375582479030754, + "grad_norm": 0.118776336312294, + "learning_rate": 1.5101611061807908e-06, + "loss": 0.0041, + "step": 92550 + }, + { + "epoch": 14.377135756446101, + "grad_norm": 0.0427917055785656, + "learning_rate": 1.506130707501463e-06, + "loss": 0.0025, + "step": 92560 + }, + { + "epoch": 14.378689033861448, + "grad_norm": 0.2871929109096527, + "learning_rate": 1.5021056120387734e-06, + "loss": 0.0046, + "step": 92570 + }, + { + "epoch": 14.380242311276794, + "grad_norm": 0.43306007981300354, + "learning_rate": 1.4980858202329084e-06, + "loss": 0.0037, + "step": 92580 + }, + { + "epoch": 14.38179558869214, + "grad_norm": 0.04329737275838852, + "learning_rate": 1.4940713325234445e-06, + "loss": 0.0032, + "step": 92590 + }, + { + "epoch": 14.383348866107486, + "grad_norm": 0.7617989182472229, + "learning_rate": 1.4900621493494137e-06, + "loss": 0.0046, + "step": 92600 + }, + { + "epoch": 14.384902143522833, + "grad_norm": 0.1183992475271225, + "learning_rate": 1.4860582711492544e-06, + "loss": 0.0036, + "step": 92610 + }, + { + "epoch": 14.38645542093818, + "grad_norm": 0.2145618349313736, + "learning_rate": 1.4820596983608325e-06, + "loss": 0.0041, + "step": 92620 + }, + { + "epoch": 14.388008698353525, + "grad_norm": 0.24066893756389618, + "learning_rate": 1.4780664314214044e-06, + "loss": 0.0023, + "step": 92630 + }, + { + "epoch": 14.389561975768872, + "grad_norm": 0.28641360998153687, + "learning_rate": 1.474078470767687e-06, + "loss": 0.0043, + "step": 92640 + }, + { + "epoch": 14.39111525318422, + "grad_norm": 0.15636591613292694, + "learning_rate": 1.470095816835787e-06, + "loss": 0.003, + "step": 92650 + }, + { + "epoch": 14.392668530599565, + "grad_norm": 0.10641726851463318, + "learning_rate": 1.4661184700612395e-06, + "loss": 0.0021, + "step": 92660 + }, + { + "epoch": 14.394221808014912, + "grad_norm": 0.3236183524131775, + "learning_rate": 1.4621464308790133e-06, + "loss": 0.0023, + "step": 92670 + }, + { + "epoch": 14.395775085430257, + "grad_norm": 0.259550005197525, + "learning_rate": 1.4581796997234775e-06, + "loss": 0.005, + "step": 92680 + }, + { + "epoch": 14.397328362845604, + "grad_norm": 0.0679916962981224, + "learning_rate": 1.4542182770284297e-06, + "loss": 0.0032, + "step": 92690 + }, + { + "epoch": 14.398881640260951, + "grad_norm": 0.10248277336359024, + "learning_rate": 1.450262163227084e-06, + "loss": 0.0017, + "step": 92700 + }, + { + "epoch": 14.400434917676296, + "grad_norm": 0.1624564826488495, + "learning_rate": 1.4463113587520671e-06, + "loss": 0.0055, + "step": 92710 + }, + { + "epoch": 14.401988195091644, + "grad_norm": 0.2134920358657837, + "learning_rate": 1.4423658640354443e-06, + "loss": 0.0036, + "step": 92720 + }, + { + "epoch": 14.40354147250699, + "grad_norm": 0.45713090896606445, + "learning_rate": 1.438425679508687e-06, + "loss": 0.0049, + "step": 92730 + }, + { + "epoch": 14.405094749922336, + "grad_norm": 0.1405794620513916, + "learning_rate": 1.4344908056026895e-06, + "loss": 0.0027, + "step": 92740 + }, + { + "epoch": 14.406648027337683, + "grad_norm": 0.3766534924507141, + "learning_rate": 1.4305612427477521e-06, + "loss": 0.004, + "step": 92750 + }, + { + "epoch": 14.408201304753028, + "grad_norm": 0.2005361169576645, + "learning_rate": 1.42663699137362e-06, + "loss": 0.0027, + "step": 92760 + }, + { + "epoch": 14.409754582168375, + "grad_norm": 0.7615877985954285, + "learning_rate": 1.4227180519094386e-06, + "loss": 0.0034, + "step": 92770 + }, + { + "epoch": 14.411307859583722, + "grad_norm": 0.16785037517547607, + "learning_rate": 1.4188044247837706e-06, + "loss": 0.0035, + "step": 92780 + }, + { + "epoch": 14.412861136999068, + "grad_norm": 0.2186514139175415, + "learning_rate": 1.4148961104246128e-06, + "loss": 0.003, + "step": 92790 + }, + { + "epoch": 14.414414414414415, + "grad_norm": 0.22753112018108368, + "learning_rate": 1.4109931092593731e-06, + "loss": 0.0031, + "step": 92800 + }, + { + "epoch": 14.41596769182976, + "grad_norm": 0.47788485884666443, + "learning_rate": 1.4070954217148658e-06, + "loss": 0.0055, + "step": 92810 + }, + { + "epoch": 14.417520969245107, + "grad_norm": 0.5139647722244263, + "learning_rate": 1.4032030482173497e-06, + "loss": 0.0034, + "step": 92820 + }, + { + "epoch": 14.419074246660454, + "grad_norm": 0.3819708824157715, + "learning_rate": 1.3993159891924901e-06, + "loss": 0.0027, + "step": 92830 + }, + { + "epoch": 14.4206275240758, + "grad_norm": 0.6320515871047974, + "learning_rate": 1.3954342450653524e-06, + "loss": 0.0034, + "step": 92840 + }, + { + "epoch": 14.422180801491146, + "grad_norm": 0.07546205073595047, + "learning_rate": 1.3915578162604582e-06, + "loss": 0.0063, + "step": 92850 + }, + { + "epoch": 14.423734078906493, + "grad_norm": 0.0414867140352726, + "learning_rate": 1.3876867032017126e-06, + "loss": 0.0019, + "step": 92860 + }, + { + "epoch": 14.425287356321839, + "grad_norm": 0.26807889342308044, + "learning_rate": 1.383820906312472e-06, + "loss": 0.004, + "step": 92870 + }, + { + "epoch": 14.426840633737186, + "grad_norm": 0.8063940405845642, + "learning_rate": 1.3799604260154752e-06, + "loss": 0.0041, + "step": 92880 + }, + { + "epoch": 14.428393911152531, + "grad_norm": 0.10023944079875946, + "learning_rate": 1.376105262732913e-06, + "loss": 0.0035, + "step": 92890 + }, + { + "epoch": 14.429947188567878, + "grad_norm": 0.3308759331703186, + "learning_rate": 1.37225541688637e-06, + "loss": 0.0045, + "step": 92900 + }, + { + "epoch": 14.431500465983225, + "grad_norm": 0.0575830303132534, + "learning_rate": 1.3684108888968594e-06, + "loss": 0.0039, + "step": 92910 + }, + { + "epoch": 14.43305374339857, + "grad_norm": 0.19535300135612488, + "learning_rate": 1.364571679184823e-06, + "loss": 0.0039, + "step": 92920 + }, + { + "epoch": 14.434607020813917, + "grad_norm": 0.19826531410217285, + "learning_rate": 1.3607377881701077e-06, + "loss": 0.0037, + "step": 92930 + }, + { + "epoch": 14.436160298229264, + "grad_norm": 0.27734124660491943, + "learning_rate": 1.3569092162719787e-06, + "loss": 0.0036, + "step": 92940 + }, + { + "epoch": 14.43771357564461, + "grad_norm": 0.1350114345550537, + "learning_rate": 1.353085963909123e-06, + "loss": 0.0042, + "step": 92950 + }, + { + "epoch": 14.439266853059957, + "grad_norm": 0.40919607877731323, + "learning_rate": 1.3492680314996454e-06, + "loss": 0.0033, + "step": 92960 + }, + { + "epoch": 14.440820130475302, + "grad_norm": 0.4932903051376343, + "learning_rate": 1.345455419461078e-06, + "loss": 0.0034, + "step": 92970 + }, + { + "epoch": 14.44237340789065, + "grad_norm": 0.7995103597640991, + "learning_rate": 1.3416481282103431e-06, + "loss": 0.0027, + "step": 92980 + }, + { + "epoch": 14.443926685305996, + "grad_norm": 0.13485755026340485, + "learning_rate": 1.337846158163819e-06, + "loss": 0.0041, + "step": 92990 + }, + { + "epoch": 14.445479962721341, + "grad_norm": 0.47379714250564575, + "learning_rate": 1.334049509737273e-06, + "loss": 0.0026, + "step": 93000 + }, + { + "epoch": 14.447033240136689, + "grad_norm": 0.36495232582092285, + "learning_rate": 1.3302581833459116e-06, + "loss": 0.0029, + "step": 93010 + }, + { + "epoch": 14.448586517552036, + "grad_norm": 0.26625269651412964, + "learning_rate": 1.3264721794043366e-06, + "loss": 0.0043, + "step": 93020 + }, + { + "epoch": 14.45013979496738, + "grad_norm": 0.10671861469745636, + "learning_rate": 1.3226914983265836e-06, + "loss": 0.004, + "step": 93030 + }, + { + "epoch": 14.451693072382728, + "grad_norm": 0.12438871711492538, + "learning_rate": 1.3189161405260997e-06, + "loss": 0.0027, + "step": 93040 + }, + { + "epoch": 14.453246349798073, + "grad_norm": 0.07643439620733261, + "learning_rate": 1.3151461064157656e-06, + "loss": 0.0038, + "step": 93050 + }, + { + "epoch": 14.45479962721342, + "grad_norm": 0.2385583221912384, + "learning_rate": 1.3113813964078459e-06, + "loss": 0.0031, + "step": 93060 + }, + { + "epoch": 14.456352904628767, + "grad_norm": 0.1750241219997406, + "learning_rate": 1.3076220109140614e-06, + "loss": 0.0046, + "step": 93070 + }, + { + "epoch": 14.457906182044113, + "grad_norm": 0.2743850350379944, + "learning_rate": 1.3038679503455275e-06, + "loss": 0.0056, + "step": 93080 + }, + { + "epoch": 14.45945945945946, + "grad_norm": 0.0626261904835701, + "learning_rate": 1.3001192151127773e-06, + "loss": 0.0028, + "step": 93090 + }, + { + "epoch": 14.461012736874807, + "grad_norm": 0.04343130812048912, + "learning_rate": 1.2963758056257657e-06, + "loss": 0.0014, + "step": 93100 + }, + { + "epoch": 14.462566014290152, + "grad_norm": 0.056662045419216156, + "learning_rate": 1.2926377222938768e-06, + "loss": 0.0022, + "step": 93110 + }, + { + "epoch": 14.464119291705499, + "grad_norm": 0.14083699882030487, + "learning_rate": 1.2889049655259e-06, + "loss": 0.0028, + "step": 93120 + }, + { + "epoch": 14.465672569120844, + "grad_norm": 0.05304040387272835, + "learning_rate": 1.2851775357300312e-06, + "loss": 0.0031, + "step": 93130 + }, + { + "epoch": 14.467225846536191, + "grad_norm": 0.16328316926956177, + "learning_rate": 1.281455433313916e-06, + "loss": 0.0021, + "step": 93140 + }, + { + "epoch": 14.468779123951538, + "grad_norm": 0.05895676836371422, + "learning_rate": 1.2777386586845796e-06, + "loss": 0.0028, + "step": 93150 + }, + { + "epoch": 14.470332401366884, + "grad_norm": 0.18596282601356506, + "learning_rate": 1.2740272122484964e-06, + "loss": 0.0036, + "step": 93160 + }, + { + "epoch": 14.47188567878223, + "grad_norm": 0.07940446585416794, + "learning_rate": 1.2703210944115308e-06, + "loss": 0.0023, + "step": 93170 + }, + { + "epoch": 14.473438956197576, + "grad_norm": 0.521403431892395, + "learning_rate": 1.2666203055789916e-06, + "loss": 0.0031, + "step": 93180 + }, + { + "epoch": 14.474992233612923, + "grad_norm": 0.09104963392019272, + "learning_rate": 1.2629248461555832e-06, + "loss": 0.0028, + "step": 93190 + }, + { + "epoch": 14.47654551102827, + "grad_norm": 0.0817810595035553, + "learning_rate": 1.2592347165454431e-06, + "loss": 0.0027, + "step": 93200 + }, + { + "epoch": 14.478098788443615, + "grad_norm": 0.2569970488548279, + "learning_rate": 1.2555499171521102e-06, + "loss": 0.0037, + "step": 93210 + }, + { + "epoch": 14.479652065858962, + "grad_norm": 0.07211538404226303, + "learning_rate": 1.251870448378556e-06, + "loss": 0.0026, + "step": 93220 + }, + { + "epoch": 14.48120534327431, + "grad_norm": 0.17158930003643036, + "learning_rate": 1.2481963106271478e-06, + "loss": 0.0031, + "step": 93230 + }, + { + "epoch": 14.482758620689655, + "grad_norm": 0.43794718384742737, + "learning_rate": 1.2445275042997084e-06, + "loss": 0.0037, + "step": 93240 + }, + { + "epoch": 14.484311898105002, + "grad_norm": 0.551372230052948, + "learning_rate": 1.2408640297974283e-06, + "loss": 0.0034, + "step": 93250 + }, + { + "epoch": 14.485865175520347, + "grad_norm": 0.1267453134059906, + "learning_rate": 1.2372058875209535e-06, + "loss": 0.0038, + "step": 93260 + }, + { + "epoch": 14.487418452935694, + "grad_norm": 0.25864359736442566, + "learning_rate": 1.2335530778703252e-06, + "loss": 0.0042, + "step": 93270 + }, + { + "epoch": 14.488971730351041, + "grad_norm": 0.22118264436721802, + "learning_rate": 1.2299056012450182e-06, + "loss": 0.0036, + "step": 93280 + }, + { + "epoch": 14.490525007766387, + "grad_norm": 0.19702066481113434, + "learning_rate": 1.2262634580439026e-06, + "loss": 0.0033, + "step": 93290 + }, + { + "epoch": 14.492078285181734, + "grad_norm": 0.22182497382164001, + "learning_rate": 1.2226266486652926e-06, + "loss": 0.0036, + "step": 93300 + }, + { + "epoch": 14.49363156259708, + "grad_norm": 0.27838173508644104, + "learning_rate": 1.2189951735068927e-06, + "loss": 0.003, + "step": 93310 + }, + { + "epoch": 14.495184840012426, + "grad_norm": 0.11119594424962997, + "learning_rate": 1.2153690329658407e-06, + "loss": 0.005, + "step": 93320 + }, + { + "epoch": 14.496738117427773, + "grad_norm": 0.17172332108020782, + "learning_rate": 1.2117482274386859e-06, + "loss": 0.0035, + "step": 93330 + }, + { + "epoch": 14.498291394843118, + "grad_norm": 0.12326064705848694, + "learning_rate": 1.2081327573213953e-06, + "loss": 0.0035, + "step": 93340 + }, + { + "epoch": 14.499844672258465, + "grad_norm": 0.32694995403289795, + "learning_rate": 1.2045226230093465e-06, + "loss": 0.0027, + "step": 93350 + }, + { + "epoch": 14.501397949673812, + "grad_norm": 1.4429112672805786, + "learning_rate": 1.2009178248973296e-06, + "loss": 0.0055, + "step": 93360 + }, + { + "epoch": 14.502951227089158, + "grad_norm": 0.1650204211473465, + "learning_rate": 1.1973183633795849e-06, + "loss": 0.0027, + "step": 93370 + }, + { + "epoch": 14.504504504504505, + "grad_norm": 0.15001438558101654, + "learning_rate": 1.193724238849725e-06, + "loss": 0.0035, + "step": 93380 + }, + { + "epoch": 14.50605778191985, + "grad_norm": 0.11662241071462631, + "learning_rate": 1.190135451700808e-06, + "loss": 0.0035, + "step": 93390 + }, + { + "epoch": 14.507611059335197, + "grad_norm": 0.5895871520042419, + "learning_rate": 1.186552002325292e-06, + "loss": 0.0034, + "step": 93400 + }, + { + "epoch": 14.509164336750544, + "grad_norm": 0.05182207003235817, + "learning_rate": 1.1829738911150578e-06, + "loss": 0.0022, + "step": 93410 + }, + { + "epoch": 14.51071761416589, + "grad_norm": 0.290671169757843, + "learning_rate": 1.1794011184613985e-06, + "loss": 0.0036, + "step": 93420 + }, + { + "epoch": 14.512270891581236, + "grad_norm": 0.2881949245929718, + "learning_rate": 1.17583368475504e-06, + "loss": 0.0046, + "step": 93430 + }, + { + "epoch": 14.513824168996583, + "grad_norm": 0.17964833974838257, + "learning_rate": 1.1722715903860982e-06, + "loss": 0.0026, + "step": 93440 + }, + { + "epoch": 14.515377446411929, + "grad_norm": 0.2503201961517334, + "learning_rate": 1.168714835744128e-06, + "loss": 0.0029, + "step": 93450 + }, + { + "epoch": 14.516930723827276, + "grad_norm": 0.16299766302108765, + "learning_rate": 1.165163421218085e-06, + "loss": 0.0025, + "step": 93460 + }, + { + "epoch": 14.518484001242623, + "grad_norm": 0.194589301943779, + "learning_rate": 1.1616173471963477e-06, + "loss": 0.0034, + "step": 93470 + }, + { + "epoch": 14.520037278657968, + "grad_norm": 0.30955755710601807, + "learning_rate": 1.1580766140667109e-06, + "loss": 0.0025, + "step": 93480 + }, + { + "epoch": 14.521590556073315, + "grad_norm": 0.0639658197760582, + "learning_rate": 1.1545412222163764e-06, + "loss": 0.001, + "step": 93490 + }, + { + "epoch": 14.52314383348866, + "grad_norm": 0.1920512318611145, + "learning_rate": 1.1510111720319795e-06, + "loss": 0.0034, + "step": 93500 + }, + { + "epoch": 14.524697110904008, + "grad_norm": 0.21205708384513855, + "learning_rate": 1.1474864638995663e-06, + "loss": 0.0039, + "step": 93510 + }, + { + "epoch": 14.526250388319355, + "grad_norm": 0.16106338798999786, + "learning_rate": 1.1439670982045735e-06, + "loss": 0.0024, + "step": 93520 + }, + { + "epoch": 14.5278036657347, + "grad_norm": 0.48325657844543457, + "learning_rate": 1.1404530753318877e-06, + "loss": 0.004, + "step": 93530 + }, + { + "epoch": 14.529356943150047, + "grad_norm": 0.08626876026391983, + "learning_rate": 1.136944395665801e-06, + "loss": 0.0029, + "step": 93540 + }, + { + "epoch": 14.530910220565392, + "grad_norm": 1.3848204612731934, + "learning_rate": 1.1334410595900014e-06, + "loss": 0.0025, + "step": 93550 + }, + { + "epoch": 14.53246349798074, + "grad_norm": 0.10604795813560486, + "learning_rate": 1.129943067487621e-06, + "loss": 0.0025, + "step": 93560 + }, + { + "epoch": 14.534016775396086, + "grad_norm": 0.15872348845005035, + "learning_rate": 1.126450419741193e-06, + "loss": 0.0051, + "step": 93570 + }, + { + "epoch": 14.535570052811432, + "grad_norm": 0.06881314516067505, + "learning_rate": 1.1229631167326671e-06, + "loss": 0.003, + "step": 93580 + }, + { + "epoch": 14.537123330226779, + "grad_norm": 0.20707492530345917, + "learning_rate": 1.1194811588434052e-06, + "loss": 0.0048, + "step": 93590 + }, + { + "epoch": 14.538676607642126, + "grad_norm": 0.08164869993925095, + "learning_rate": 1.1160045464542023e-06, + "loss": 0.002, + "step": 93600 + }, + { + "epoch": 14.540229885057471, + "grad_norm": 0.12302481383085251, + "learning_rate": 1.1125332799452325e-06, + "loss": 0.0035, + "step": 93610 + }, + { + "epoch": 14.541783162472818, + "grad_norm": 0.15914078056812286, + "learning_rate": 1.109067359696131e-06, + "loss": 0.0044, + "step": 93620 + }, + { + "epoch": 14.543336439888163, + "grad_norm": 0.21462535858154297, + "learning_rate": 1.1056067860859166e-06, + "loss": 0.0035, + "step": 93630 + }, + { + "epoch": 14.54488971730351, + "grad_norm": 0.12384826689958572, + "learning_rate": 1.102151559493031e-06, + "loss": 0.0038, + "step": 93640 + }, + { + "epoch": 14.546442994718857, + "grad_norm": 0.2087160348892212, + "learning_rate": 1.0987016802953332e-06, + "loss": 0.0051, + "step": 93650 + }, + { + "epoch": 14.547996272134203, + "grad_norm": 2.3288955688476562, + "learning_rate": 1.0952571488701048e-06, + "loss": 0.0037, + "step": 93660 + }, + { + "epoch": 14.54954954954955, + "grad_norm": 0.6301618814468384, + "learning_rate": 1.0918179655940163e-06, + "loss": 0.0023, + "step": 93670 + }, + { + "epoch": 14.551102826964897, + "grad_norm": 0.13886190950870514, + "learning_rate": 1.088384130843184e-06, + "loss": 0.0035, + "step": 93680 + }, + { + "epoch": 14.552656104380242, + "grad_norm": 0.033146120607852936, + "learning_rate": 1.0849556449931296e-06, + "loss": 0.0037, + "step": 93690 + }, + { + "epoch": 14.554209381795589, + "grad_norm": 0.06948021799325943, + "learning_rate": 1.0815325084187866e-06, + "loss": 0.0033, + "step": 93700 + }, + { + "epoch": 14.555762659210934, + "grad_norm": 0.09368661046028137, + "learning_rate": 1.0781147214944943e-06, + "loss": 0.0036, + "step": 93710 + }, + { + "epoch": 14.557315936626281, + "grad_norm": 0.1927870213985443, + "learning_rate": 1.0747022845940313e-06, + "loss": 0.0055, + "step": 93720 + }, + { + "epoch": 14.558869214041628, + "grad_norm": 0.04995107278227806, + "learning_rate": 1.0712951980905605e-06, + "loss": 0.0036, + "step": 93730 + }, + { + "epoch": 14.560422491456974, + "grad_norm": 0.16948357224464417, + "learning_rate": 1.0678934623566839e-06, + "loss": 0.0033, + "step": 93740 + }, + { + "epoch": 14.56197576887232, + "grad_norm": 1.1557799577713013, + "learning_rate": 1.0644970777644091e-06, + "loss": 0.0033, + "step": 93750 + }, + { + "epoch": 14.563529046287666, + "grad_norm": 0.027370654046535492, + "learning_rate": 1.0611060446851673e-06, + "loss": 0.0028, + "step": 93760 + }, + { + "epoch": 14.565082323703013, + "grad_norm": 0.47818195819854736, + "learning_rate": 1.0577203634897836e-06, + "loss": 0.0037, + "step": 93770 + }, + { + "epoch": 14.56663560111836, + "grad_norm": 1.0862984657287598, + "learning_rate": 1.0543400345485178e-06, + "loss": 0.0041, + "step": 93780 + }, + { + "epoch": 14.568188878533705, + "grad_norm": 1.0523468255996704, + "learning_rate": 1.0509650582310404e-06, + "loss": 0.0038, + "step": 93790 + }, + { + "epoch": 14.569742155949053, + "grad_norm": 0.391824334859848, + "learning_rate": 1.0475954349064287e-06, + "loss": 0.0051, + "step": 93800 + }, + { + "epoch": 14.5712954333644, + "grad_norm": 0.34986522793769836, + "learning_rate": 1.0442311649431825e-06, + "loss": 0.0032, + "step": 93810 + }, + { + "epoch": 14.572848710779745, + "grad_norm": 0.07869528979063034, + "learning_rate": 1.0408722487092071e-06, + "loss": 0.0034, + "step": 93820 + }, + { + "epoch": 14.574401988195092, + "grad_norm": 0.2995017170906067, + "learning_rate": 1.0375186865718424e-06, + "loss": 0.0029, + "step": 93830 + }, + { + "epoch": 14.575955265610437, + "grad_norm": 0.18891437351703644, + "learning_rate": 1.0341704788978113e-06, + "loss": 0.0021, + "step": 93840 + }, + { + "epoch": 14.577508543025784, + "grad_norm": 0.09220641851425171, + "learning_rate": 1.0308276260532879e-06, + "loss": 0.0049, + "step": 93850 + }, + { + "epoch": 14.579061820441131, + "grad_norm": 0.23448140919208527, + "learning_rate": 1.0274901284038241e-06, + "loss": 0.0043, + "step": 93860 + }, + { + "epoch": 14.580615097856477, + "grad_norm": 0.15682746469974518, + "learning_rate": 1.0241579863144114e-06, + "loss": 0.0031, + "step": 93870 + }, + { + "epoch": 14.582168375271824, + "grad_norm": 0.14418847858905792, + "learning_rate": 1.0208312001494413e-06, + "loss": 0.0032, + "step": 93880 + }, + { + "epoch": 14.58372165268717, + "grad_norm": 0.3479169011116028, + "learning_rate": 1.0175097702727453e-06, + "loss": 0.0027, + "step": 93890 + }, + { + "epoch": 14.585274930102516, + "grad_norm": 0.05447511002421379, + "learning_rate": 1.0141936970475275e-06, + "loss": 0.0022, + "step": 93900 + }, + { + "epoch": 14.586828207517863, + "grad_norm": 0.2745799720287323, + "learning_rate": 1.0108829808364417e-06, + "loss": 0.0026, + "step": 93910 + }, + { + "epoch": 14.588381484933208, + "grad_norm": 0.10664288699626923, + "learning_rate": 1.0075776220015376e-06, + "loss": 0.0028, + "step": 93920 + }, + { + "epoch": 14.589934762348555, + "grad_norm": 0.2479051649570465, + "learning_rate": 1.0042776209042816e-06, + "loss": 0.0024, + "step": 93930 + }, + { + "epoch": 14.591488039763902, + "grad_norm": 0.20577864348888397, + "learning_rate": 1.000982977905568e-06, + "loss": 0.0042, + "step": 93940 + }, + { + "epoch": 14.593041317179248, + "grad_norm": 0.8580231070518494, + "learning_rate": 9.976936933656867e-07, + "loss": 0.0035, + "step": 93950 + }, + { + "epoch": 14.594594594594595, + "grad_norm": 0.24296297132968903, + "learning_rate": 9.944097676443442e-07, + "loss": 0.0025, + "step": 93960 + }, + { + "epoch": 14.596147872009942, + "grad_norm": 0.2110612690448761, + "learning_rate": 9.911312011006757e-07, + "loss": 0.0041, + "step": 93970 + }, + { + "epoch": 14.597701149425287, + "grad_norm": 0.0656420886516571, + "learning_rate": 9.878579940932109e-07, + "loss": 0.0059, + "step": 93980 + }, + { + "epoch": 14.599254426840634, + "grad_norm": 0.1384367048740387, + "learning_rate": 9.84590146979908e-07, + "loss": 0.0041, + "step": 93990 + }, + { + "epoch": 14.60080770425598, + "grad_norm": 0.06685636192560196, + "learning_rate": 9.813276601181365e-07, + "loss": 0.0025, + "step": 94000 + }, + { + "epoch": 14.602360981671326, + "grad_norm": 0.23247118294239044, + "learning_rate": 9.780705338646724e-07, + "loss": 0.0033, + "step": 94010 + }, + { + "epoch": 14.603914259086674, + "grad_norm": 0.3435850143432617, + "learning_rate": 9.748187685757083e-07, + "loss": 0.0027, + "step": 94020 + }, + { + "epoch": 14.605467536502019, + "grad_norm": 0.04300348833203316, + "learning_rate": 9.715723646068543e-07, + "loss": 0.0027, + "step": 94030 + }, + { + "epoch": 14.607020813917366, + "grad_norm": 0.06960182636976242, + "learning_rate": 9.68331322313143e-07, + "loss": 0.0038, + "step": 94040 + }, + { + "epoch": 14.608574091332713, + "grad_norm": 0.11104118078947067, + "learning_rate": 9.650956420489909e-07, + "loss": 0.0025, + "step": 94050 + }, + { + "epoch": 14.610127368748058, + "grad_norm": 0.36521098017692566, + "learning_rate": 9.618653241682595e-07, + "loss": 0.0034, + "step": 94060 + }, + { + "epoch": 14.611680646163405, + "grad_norm": 0.14326788485050201, + "learning_rate": 9.586403690242053e-07, + "loss": 0.0042, + "step": 94070 + }, + { + "epoch": 14.61323392357875, + "grad_norm": 0.09567058086395264, + "learning_rate": 9.554207769695122e-07, + "loss": 0.0027, + "step": 94080 + }, + { + "epoch": 14.614787200994098, + "grad_norm": 0.11584049463272095, + "learning_rate": 9.522065483562659e-07, + "loss": 0.0028, + "step": 94090 + }, + { + "epoch": 14.616340478409445, + "grad_norm": 0.062368933111429214, + "learning_rate": 9.489976835359681e-07, + "loss": 0.0042, + "step": 94100 + }, + { + "epoch": 14.61789375582479, + "grad_norm": 0.12218926846981049, + "learning_rate": 9.457941828595329e-07, + "loss": 0.0033, + "step": 94110 + }, + { + "epoch": 14.619447033240137, + "grad_norm": 0.4115569293498993, + "learning_rate": 9.425960466772965e-07, + "loss": 0.003, + "step": 94120 + }, + { + "epoch": 14.621000310655482, + "grad_norm": 0.08021320402622223, + "learning_rate": 9.394032753390014e-07, + "loss": 0.0026, + "step": 94130 + }, + { + "epoch": 14.62255358807083, + "grad_norm": 0.12094592303037643, + "learning_rate": 9.362158691938017e-07, + "loss": 0.0035, + "step": 94140 + }, + { + "epoch": 14.624106865486176, + "grad_norm": 0.5250458717346191, + "learning_rate": 9.330338285902685e-07, + "loss": 0.0035, + "step": 94150 + }, + { + "epoch": 14.625660142901522, + "grad_norm": 0.30520468950271606, + "learning_rate": 9.298571538763845e-07, + "loss": 0.0034, + "step": 94160 + }, + { + "epoch": 14.627213420316869, + "grad_norm": 0.8125352263450623, + "learning_rate": 9.266858453995441e-07, + "loss": 0.0033, + "step": 94170 + }, + { + "epoch": 14.628766697732216, + "grad_norm": 0.14838337898254395, + "learning_rate": 9.235199035065645e-07, + "loss": 0.0026, + "step": 94180 + }, + { + "epoch": 14.630319975147561, + "grad_norm": 0.5893619656562805, + "learning_rate": 9.203593285436574e-07, + "loss": 0.0026, + "step": 94190 + }, + { + "epoch": 14.631873252562908, + "grad_norm": 0.14083927869796753, + "learning_rate": 9.172041208564686e-07, + "loss": 0.003, + "step": 94200 + }, + { + "epoch": 14.633426529978253, + "grad_norm": 1.115826964378357, + "learning_rate": 9.140542807900332e-07, + "loss": 0.0041, + "step": 94210 + }, + { + "epoch": 14.6349798073936, + "grad_norm": 0.2355314940214157, + "learning_rate": 9.109098086888313e-07, + "loss": 0.0023, + "step": 94220 + }, + { + "epoch": 14.636533084808947, + "grad_norm": 0.1243186816573143, + "learning_rate": 9.077707048967265e-07, + "loss": 0.0031, + "step": 94230 + }, + { + "epoch": 14.638086362224293, + "grad_norm": 0.4373056888580322, + "learning_rate": 9.046369697570057e-07, + "loss": 0.0038, + "step": 94240 + }, + { + "epoch": 14.63963963963964, + "grad_norm": 1.573752999305725, + "learning_rate": 9.015086036123721e-07, + "loss": 0.0033, + "step": 94250 + }, + { + "epoch": 14.641192917054987, + "grad_norm": 0.12404125928878784, + "learning_rate": 8.983856068049413e-07, + "loss": 0.0023, + "step": 94260 + }, + { + "epoch": 14.642746194470332, + "grad_norm": 0.3431504964828491, + "learning_rate": 8.952679796762398e-07, + "loss": 0.0049, + "step": 94270 + }, + { + "epoch": 14.64429947188568, + "grad_norm": 2.183439016342163, + "learning_rate": 8.921557225671951e-07, + "loss": 0.0029, + "step": 94280 + }, + { + "epoch": 14.645852749301024, + "grad_norm": 0.0669480711221695, + "learning_rate": 8.890488358181792e-07, + "loss": 0.004, + "step": 94290 + }, + { + "epoch": 14.647406026716371, + "grad_norm": 0.0904403030872345, + "learning_rate": 8.859473197689372e-07, + "loss": 0.0022, + "step": 94300 + }, + { + "epoch": 14.648959304131719, + "grad_norm": 0.09622298181056976, + "learning_rate": 8.828511747586588e-07, + "loss": 0.0053, + "step": 94310 + }, + { + "epoch": 14.650512581547064, + "grad_norm": 0.4502916932106018, + "learning_rate": 8.797604011259287e-07, + "loss": 0.0025, + "step": 94320 + }, + { + "epoch": 14.65206585896241, + "grad_norm": 0.5580009818077087, + "learning_rate": 8.766749992087487e-07, + "loss": 0.0023, + "step": 94330 + }, + { + "epoch": 14.653619136377756, + "grad_norm": 0.26705148816108704, + "learning_rate": 8.735949693445378e-07, + "loss": 0.0038, + "step": 94340 + }, + { + "epoch": 14.655172413793103, + "grad_norm": 0.45974960923194885, + "learning_rate": 8.705203118701211e-07, + "loss": 0.004, + "step": 94350 + }, + { + "epoch": 14.65672569120845, + "grad_norm": 0.19873034954071045, + "learning_rate": 8.674510271217295e-07, + "loss": 0.0048, + "step": 94360 + }, + { + "epoch": 14.658278968623796, + "grad_norm": 0.43046560883522034, + "learning_rate": 8.643871154350336e-07, + "loss": 0.0031, + "step": 94370 + }, + { + "epoch": 14.659832246039143, + "grad_norm": 0.5174643993377686, + "learning_rate": 8.613285771450818e-07, + "loss": 0.0041, + "step": 94380 + }, + { + "epoch": 14.66138552345449, + "grad_norm": 0.3279443085193634, + "learning_rate": 8.582754125863623e-07, + "loss": 0.0049, + "step": 94390 + }, + { + "epoch": 14.662938800869835, + "grad_norm": 0.07669593393802643, + "learning_rate": 8.55227622092758e-07, + "loss": 0.0028, + "step": 94400 + }, + { + "epoch": 14.664492078285182, + "grad_norm": 0.1971748024225235, + "learning_rate": 8.521852059975743e-07, + "loss": 0.0037, + "step": 94410 + }, + { + "epoch": 14.666045355700529, + "grad_norm": 0.10234641283750534, + "learning_rate": 8.491481646335231e-07, + "loss": 0.0033, + "step": 94420 + }, + { + "epoch": 14.667598633115874, + "grad_norm": 0.09123992919921875, + "learning_rate": 8.46116498332733e-07, + "loss": 0.0036, + "step": 94430 + }, + { + "epoch": 14.669151910531221, + "grad_norm": 0.26922234892845154, + "learning_rate": 8.430902074267333e-07, + "loss": 0.0028, + "step": 94440 + }, + { + "epoch": 14.670705187946567, + "grad_norm": 0.14418771862983704, + "learning_rate": 8.400692922464925e-07, + "loss": 0.0031, + "step": 94450 + }, + { + "epoch": 14.672258465361914, + "grad_norm": 0.2571566104888916, + "learning_rate": 8.370537531223577e-07, + "loss": 0.0024, + "step": 94460 + }, + { + "epoch": 14.67381174277726, + "grad_norm": 0.10254274308681488, + "learning_rate": 8.340435903841092e-07, + "loss": 0.0033, + "step": 94470 + }, + { + "epoch": 14.675365020192606, + "grad_norm": 0.5138433575630188, + "learning_rate": 8.310388043609341e-07, + "loss": 0.0042, + "step": 94480 + }, + { + "epoch": 14.676918297607953, + "grad_norm": 0.4586487114429474, + "learning_rate": 8.280393953814303e-07, + "loss": 0.0035, + "step": 94490 + }, + { + "epoch": 14.678471575023298, + "grad_norm": 0.15043804049491882, + "learning_rate": 8.250453637736077e-07, + "loss": 0.0033, + "step": 94500 + }, + { + "epoch": 14.680024852438645, + "grad_norm": 0.18551892042160034, + "learning_rate": 8.22056709864899e-07, + "loss": 0.0045, + "step": 94510 + }, + { + "epoch": 14.681578129853992, + "grad_norm": 0.19584649801254272, + "learning_rate": 8.190734339821204e-07, + "loss": 0.0038, + "step": 94520 + }, + { + "epoch": 14.683131407269338, + "grad_norm": 0.22894242405891418, + "learning_rate": 8.160955364515333e-07, + "loss": 0.0034, + "step": 94530 + }, + { + "epoch": 14.684684684684685, + "grad_norm": 0.6059014201164246, + "learning_rate": 8.131230175987936e-07, + "loss": 0.0037, + "step": 94540 + }, + { + "epoch": 14.686237962100032, + "grad_norm": 0.3421262204647064, + "learning_rate": 8.101558777489693e-07, + "loss": 0.0048, + "step": 94550 + }, + { + "epoch": 14.687791239515377, + "grad_norm": 0.10174999386072159, + "learning_rate": 8.071941172265396e-07, + "loss": 0.0029, + "step": 94560 + }, + { + "epoch": 14.689344516930724, + "grad_norm": 0.191126748919487, + "learning_rate": 8.042377363553955e-07, + "loss": 0.005, + "step": 94570 + }, + { + "epoch": 14.69089779434607, + "grad_norm": 0.13786940276622772, + "learning_rate": 8.01286735458856e-07, + "loss": 0.0026, + "step": 94580 + }, + { + "epoch": 14.692451071761417, + "grad_norm": 0.17839092016220093, + "learning_rate": 7.983411148596298e-07, + "loss": 0.0053, + "step": 94590 + }, + { + "epoch": 14.694004349176764, + "grad_norm": 0.17346739768981934, + "learning_rate": 7.954008748798425e-07, + "loss": 0.0037, + "step": 94600 + }, + { + "epoch": 14.695557626592109, + "grad_norm": 0.11964564770460129, + "learning_rate": 7.924660158410369e-07, + "loss": 0.0033, + "step": 94610 + }, + { + "epoch": 14.697110904007456, + "grad_norm": 0.2165236473083496, + "learning_rate": 7.895365380641728e-07, + "loss": 0.0037, + "step": 94620 + }, + { + "epoch": 14.698664181422803, + "grad_norm": 0.2990745007991791, + "learning_rate": 7.866124418695941e-07, + "loss": 0.0032, + "step": 94630 + }, + { + "epoch": 14.700217458838148, + "grad_norm": 0.1918642371892929, + "learning_rate": 7.83693727577095e-07, + "loss": 0.0032, + "step": 94640 + }, + { + "epoch": 14.701770736253495, + "grad_norm": 0.2194504588842392, + "learning_rate": 7.807803955058535e-07, + "loss": 0.0024, + "step": 94650 + }, + { + "epoch": 14.70332401366884, + "grad_norm": 0.1771080493927002, + "learning_rate": 7.778724459744702e-07, + "loss": 0.0033, + "step": 94660 + }, + { + "epoch": 14.704877291084188, + "grad_norm": 0.06507997959852219, + "learning_rate": 7.749698793009464e-07, + "loss": 0.0045, + "step": 94670 + }, + { + "epoch": 14.706430568499535, + "grad_norm": 0.6297133564949036, + "learning_rate": 7.720726958027169e-07, + "loss": 0.0031, + "step": 94680 + }, + { + "epoch": 14.70798384591488, + "grad_norm": 1.4674731492996216, + "learning_rate": 7.691808957965951e-07, + "loss": 0.0031, + "step": 94690 + }, + { + "epoch": 14.709537123330227, + "grad_norm": 0.10591240227222443, + "learning_rate": 7.662944795988336e-07, + "loss": 0.0033, + "step": 94700 + }, + { + "epoch": 14.711090400745572, + "grad_norm": 0.6599385142326355, + "learning_rate": 7.634134475250854e-07, + "loss": 0.0032, + "step": 94710 + }, + { + "epoch": 14.71264367816092, + "grad_norm": 0.2901135981082916, + "learning_rate": 7.605377998904262e-07, + "loss": 0.0034, + "step": 94720 + }, + { + "epoch": 14.714196955576266, + "grad_norm": 0.2270892858505249, + "learning_rate": 7.5766753700931e-07, + "loss": 0.0025, + "step": 94730 + }, + { + "epoch": 14.715750232991612, + "grad_norm": 0.22557108104228973, + "learning_rate": 7.54802659195647e-07, + "loss": 0.0025, + "step": 94740 + }, + { + "epoch": 14.717303510406959, + "grad_norm": 0.2525418698787689, + "learning_rate": 7.519431667627253e-07, + "loss": 0.0031, + "step": 94750 + }, + { + "epoch": 14.718856787822306, + "grad_norm": 0.9214674234390259, + "learning_rate": 7.490890600232503e-07, + "loss": 0.0023, + "step": 94760 + }, + { + "epoch": 14.720410065237651, + "grad_norm": 0.09449414908885956, + "learning_rate": 7.462403392893502e-07, + "loss": 0.0037, + "step": 94770 + }, + { + "epoch": 14.721963342652998, + "grad_norm": 0.26442763209342957, + "learning_rate": 7.433970048725536e-07, + "loss": 0.0045, + "step": 94780 + }, + { + "epoch": 14.723516620068343, + "grad_norm": 0.06376110762357712, + "learning_rate": 7.405590570838062e-07, + "loss": 0.003, + "step": 94790 + }, + { + "epoch": 14.72506989748369, + "grad_norm": 0.3369109034538269, + "learning_rate": 7.377264962334596e-07, + "loss": 0.0043, + "step": 94800 + }, + { + "epoch": 14.726623174899037, + "grad_norm": 0.2724217474460602, + "learning_rate": 7.348993226312773e-07, + "loss": 0.0027, + "step": 94810 + }, + { + "epoch": 14.728176452314383, + "grad_norm": 0.23557865619659424, + "learning_rate": 7.32077536586434e-07, + "loss": 0.0039, + "step": 94820 + }, + { + "epoch": 14.72972972972973, + "grad_norm": 0.07396728545427322, + "learning_rate": 7.292611384075221e-07, + "loss": 0.0025, + "step": 94830 + }, + { + "epoch": 14.731283007145077, + "grad_norm": 0.9402545690536499, + "learning_rate": 7.264501284025338e-07, + "loss": 0.0035, + "step": 94840 + }, + { + "epoch": 14.732836284560422, + "grad_norm": 0.27160006761550903, + "learning_rate": 7.236445068788789e-07, + "loss": 0.0032, + "step": 94850 + }, + { + "epoch": 14.73438956197577, + "grad_norm": 0.10096425563097, + "learning_rate": 7.20844274143373e-07, + "loss": 0.0029, + "step": 94860 + }, + { + "epoch": 14.735942839391114, + "grad_norm": 0.6042953729629517, + "learning_rate": 7.180494305022489e-07, + "loss": 0.0052, + "step": 94870 + }, + { + "epoch": 14.737496116806462, + "grad_norm": 0.06103172153234482, + "learning_rate": 7.152599762611456e-07, + "loss": 0.0069, + "step": 94880 + }, + { + "epoch": 14.739049394221809, + "grad_norm": 0.11410584300756454, + "learning_rate": 7.124759117251079e-07, + "loss": 0.0032, + "step": 94890 + }, + { + "epoch": 14.740602671637154, + "grad_norm": 0.030128000304102898, + "learning_rate": 7.096972371986033e-07, + "loss": 0.0025, + "step": 94900 + }, + { + "epoch": 14.742155949052501, + "grad_norm": 0.25605064630508423, + "learning_rate": 7.069239529855054e-07, + "loss": 0.0025, + "step": 94910 + }, + { + "epoch": 14.743709226467848, + "grad_norm": 0.2203761786222458, + "learning_rate": 7.041560593890939e-07, + "loss": 0.0026, + "step": 94920 + }, + { + "epoch": 14.745262503883193, + "grad_norm": 0.20852969586849213, + "learning_rate": 7.013935567120599e-07, + "loss": 0.0032, + "step": 94930 + }, + { + "epoch": 14.74681578129854, + "grad_norm": 0.16755419969558716, + "learning_rate": 6.986364452565009e-07, + "loss": 0.0039, + "step": 94940 + }, + { + "epoch": 14.748369058713886, + "grad_norm": 0.038310352712869644, + "learning_rate": 6.95884725323942e-07, + "loss": 0.0029, + "step": 94950 + }, + { + "epoch": 14.749922336129233, + "grad_norm": 0.5696905851364136, + "learning_rate": 6.931383972152982e-07, + "loss": 0.0025, + "step": 94960 + }, + { + "epoch": 14.75147561354458, + "grad_norm": 0.29018309712409973, + "learning_rate": 6.903974612309127e-07, + "loss": 0.0029, + "step": 94970 + }, + { + "epoch": 14.753028890959925, + "grad_norm": 0.3048623204231262, + "learning_rate": 6.876619176705179e-07, + "loss": 0.0027, + "step": 94980 + }, + { + "epoch": 14.754582168375272, + "grad_norm": 0.1827133744955063, + "learning_rate": 6.8493176683328e-07, + "loss": 0.0045, + "step": 94990 + }, + { + "epoch": 14.756135445790619, + "grad_norm": 1.2483352422714233, + "learning_rate": 6.822070090177657e-07, + "loss": 0.0038, + "step": 95000 + }, + { + "epoch": 14.757688723205964, + "grad_norm": 2.3296518325805664, + "learning_rate": 6.794876445219367e-07, + "loss": 0.0034, + "step": 95010 + }, + { + "epoch": 14.759242000621311, + "grad_norm": 0.07111114263534546, + "learning_rate": 6.767736736431885e-07, + "loss": 0.0022, + "step": 95020 + }, + { + "epoch": 14.760795278036657, + "grad_norm": 0.17822113633155823, + "learning_rate": 6.740650966783113e-07, + "loss": 0.0038, + "step": 95030 + }, + { + "epoch": 14.762348555452004, + "grad_norm": 0.11573588848114014, + "learning_rate": 6.713619139235239e-07, + "loss": 0.0036, + "step": 95040 + }, + { + "epoch": 14.76390183286735, + "grad_norm": 0.3163076639175415, + "learning_rate": 6.686641256744286e-07, + "loss": 0.0036, + "step": 95050 + }, + { + "epoch": 14.765455110282696, + "grad_norm": 0.13795195519924164, + "learning_rate": 6.659717322260562e-07, + "loss": 0.0036, + "step": 95060 + }, + { + "epoch": 14.767008387698043, + "grad_norm": 1.4178584814071655, + "learning_rate": 6.632847338728377e-07, + "loss": 0.0033, + "step": 95070 + }, + { + "epoch": 14.768561665113388, + "grad_norm": 0.725684404373169, + "learning_rate": 6.60603130908627e-07, + "loss": 0.0043, + "step": 95080 + }, + { + "epoch": 14.770114942528735, + "grad_norm": 0.5121039748191833, + "learning_rate": 6.579269236266783e-07, + "loss": 0.0029, + "step": 95090 + }, + { + "epoch": 14.771668219944083, + "grad_norm": 0.06809240579605103, + "learning_rate": 6.552561123196632e-07, + "loss": 0.0034, + "step": 95100 + }, + { + "epoch": 14.773221497359428, + "grad_norm": 0.20359955728054047, + "learning_rate": 6.525906972796425e-07, + "loss": 0.0029, + "step": 95110 + }, + { + "epoch": 14.774774774774775, + "grad_norm": 0.17559657990932465, + "learning_rate": 6.499306787981108e-07, + "loss": 0.0035, + "step": 95120 + }, + { + "epoch": 14.776328052190122, + "grad_norm": 0.11164633184671402, + "learning_rate": 6.472760571659631e-07, + "loss": 0.0033, + "step": 95130 + }, + { + "epoch": 14.777881329605467, + "grad_norm": 0.20498612523078918, + "learning_rate": 6.446268326735061e-07, + "loss": 0.0027, + "step": 95140 + }, + { + "epoch": 14.779434607020814, + "grad_norm": 0.1080244705080986, + "learning_rate": 6.41983005610458e-07, + "loss": 0.003, + "step": 95150 + }, + { + "epoch": 14.78098788443616, + "grad_norm": 0.1820591390132904, + "learning_rate": 6.393445762659378e-07, + "loss": 0.0026, + "step": 95160 + }, + { + "epoch": 14.782541161851507, + "grad_norm": 0.1721724420785904, + "learning_rate": 6.36711544928481e-07, + "loss": 0.0044, + "step": 95170 + }, + { + "epoch": 14.784094439266854, + "grad_norm": 0.7190316319465637, + "learning_rate": 6.340839118860298e-07, + "loss": 0.0041, + "step": 95180 + }, + { + "epoch": 14.785647716682199, + "grad_norm": 0.5420067310333252, + "learning_rate": 6.314616774259485e-07, + "loss": 0.0042, + "step": 95190 + }, + { + "epoch": 14.787200994097546, + "grad_norm": 0.10925770550966263, + "learning_rate": 6.288448418349858e-07, + "loss": 0.0017, + "step": 95200 + }, + { + "epoch": 14.788754271512893, + "grad_norm": 0.058390870690345764, + "learning_rate": 6.262334053993235e-07, + "loss": 0.0026, + "step": 95210 + }, + { + "epoch": 14.790307548928238, + "grad_norm": 0.08118493109941483, + "learning_rate": 6.236273684045501e-07, + "loss": 0.0028, + "step": 95220 + }, + { + "epoch": 14.791860826343585, + "grad_norm": 0.2346695065498352, + "learning_rate": 6.210267311356488e-07, + "loss": 0.0036, + "step": 95230 + }, + { + "epoch": 14.79341410375893, + "grad_norm": 0.3029917776584625, + "learning_rate": 6.184314938770197e-07, + "loss": 0.0042, + "step": 95240 + }, + { + "epoch": 14.794967381174278, + "grad_norm": 0.5916593074798584, + "learning_rate": 6.158416569124858e-07, + "loss": 0.0047, + "step": 95250 + }, + { + "epoch": 14.796520658589625, + "grad_norm": 0.3060728907585144, + "learning_rate": 6.132572205252541e-07, + "loss": 0.0043, + "step": 95260 + }, + { + "epoch": 14.79807393600497, + "grad_norm": 0.17308613657951355, + "learning_rate": 6.106781849979648e-07, + "loss": 0.0047, + "step": 95270 + }, + { + "epoch": 14.799627213420317, + "grad_norm": 0.09716508537530899, + "learning_rate": 6.081045506126482e-07, + "loss": 0.0041, + "step": 95280 + }, + { + "epoch": 14.801180490835662, + "grad_norm": 0.11122794449329376, + "learning_rate": 6.055363176507678e-07, + "loss": 0.0038, + "step": 95290 + }, + { + "epoch": 14.80273376825101, + "grad_norm": 0.30243873596191406, + "learning_rate": 6.029734863931602e-07, + "loss": 0.0034, + "step": 95300 + }, + { + "epoch": 14.804287045666356, + "grad_norm": 0.07881448417901993, + "learning_rate": 6.004160571201179e-07, + "loss": 0.0023, + "step": 95310 + }, + { + "epoch": 14.805840323081702, + "grad_norm": 1.5072287321090698, + "learning_rate": 5.978640301112947e-07, + "loss": 0.0045, + "step": 95320 + }, + { + "epoch": 14.807393600497049, + "grad_norm": 0.11170833557844162, + "learning_rate": 5.9531740564579e-07, + "loss": 0.003, + "step": 95330 + }, + { + "epoch": 14.808946877912396, + "grad_norm": 0.28490620851516724, + "learning_rate": 5.927761840020918e-07, + "loss": 0.0042, + "step": 95340 + }, + { + "epoch": 14.810500155327741, + "grad_norm": 0.2794519066810608, + "learning_rate": 5.902403654581113e-07, + "loss": 0.0033, + "step": 95350 + }, + { + "epoch": 14.812053432743088, + "grad_norm": 0.504582941532135, + "learning_rate": 5.877099502911542e-07, + "loss": 0.0027, + "step": 95360 + }, + { + "epoch": 14.813606710158435, + "grad_norm": 0.28163930773735046, + "learning_rate": 5.851849387779496e-07, + "loss": 0.0026, + "step": 95370 + }, + { + "epoch": 14.81515998757378, + "grad_norm": 0.04581301286816597, + "learning_rate": 5.826653311946262e-07, + "loss": 0.0014, + "step": 95380 + }, + { + "epoch": 14.816713264989128, + "grad_norm": 0.14296303689479828, + "learning_rate": 5.801511278167249e-07, + "loss": 0.0026, + "step": 95390 + }, + { + "epoch": 14.818266542404473, + "grad_norm": 0.23698538541793823, + "learning_rate": 5.776423289191868e-07, + "loss": 0.0045, + "step": 95400 + }, + { + "epoch": 14.81981981981982, + "grad_norm": 0.09622334688901901, + "learning_rate": 5.751389347763869e-07, + "loss": 0.0025, + "step": 95410 + }, + { + "epoch": 14.821373097235167, + "grad_norm": 0.06587391346693039, + "learning_rate": 5.726409456620841e-07, + "loss": 0.003, + "step": 95420 + }, + { + "epoch": 14.822926374650512, + "grad_norm": 0.1005033403635025, + "learning_rate": 5.701483618494541e-07, + "loss": 0.0031, + "step": 95430 + }, + { + "epoch": 14.82447965206586, + "grad_norm": 0.1791703701019287, + "learning_rate": 5.67661183611079e-07, + "loss": 0.0051, + "step": 95440 + }, + { + "epoch": 14.826032929481205, + "grad_norm": 0.32645347714424133, + "learning_rate": 5.651794112189579e-07, + "loss": 0.0047, + "step": 95450 + }, + { + "epoch": 14.827586206896552, + "grad_norm": 0.5349428653717041, + "learning_rate": 5.627030449444903e-07, + "loss": 0.0031, + "step": 95460 + }, + { + "epoch": 14.829139484311899, + "grad_norm": 0.1360187530517578, + "learning_rate": 5.60232085058493e-07, + "loss": 0.0026, + "step": 95470 + }, + { + "epoch": 14.830692761727244, + "grad_norm": 0.11991175264120102, + "learning_rate": 5.577665318311776e-07, + "loss": 0.0033, + "step": 95480 + }, + { + "epoch": 14.832246039142591, + "grad_norm": 0.6241487264633179, + "learning_rate": 5.553063855321838e-07, + "loss": 0.0038, + "step": 95490 + }, + { + "epoch": 14.833799316557938, + "grad_norm": 0.05672737956047058, + "learning_rate": 5.528516464305467e-07, + "loss": 0.0026, + "step": 95500 + }, + { + "epoch": 14.835352593973283, + "grad_norm": 0.22238461673259735, + "learning_rate": 5.504023147947068e-07, + "loss": 0.0045, + "step": 95510 + }, + { + "epoch": 14.83690587138863, + "grad_norm": 0.2812361717224121, + "learning_rate": 5.479583908925279e-07, + "loss": 0.0027, + "step": 95520 + }, + { + "epoch": 14.838459148803976, + "grad_norm": 0.36159417033195496, + "learning_rate": 5.455198749912626e-07, + "loss": 0.0027, + "step": 95530 + }, + { + "epoch": 14.840012426219323, + "grad_norm": 0.28500354290008545, + "learning_rate": 5.430867673575979e-07, + "loss": 0.0051, + "step": 95540 + }, + { + "epoch": 14.84156570363467, + "grad_norm": 0.8369899988174438, + "learning_rate": 5.406590682575985e-07, + "loss": 0.0035, + "step": 95550 + }, + { + "epoch": 14.843118981050015, + "grad_norm": 0.2808755338191986, + "learning_rate": 5.382367779567687e-07, + "loss": 0.0033, + "step": 95560 + }, + { + "epoch": 14.844672258465362, + "grad_norm": 0.3420087993144989, + "learning_rate": 5.358198967200023e-07, + "loss": 0.0033, + "step": 95570 + }, + { + "epoch": 14.84622553588071, + "grad_norm": 0.21522311866283417, + "learning_rate": 5.334084248115989e-07, + "loss": 0.0032, + "step": 95580 + }, + { + "epoch": 14.847778813296054, + "grad_norm": 0.12296795845031738, + "learning_rate": 5.310023624952754e-07, + "loss": 0.0032, + "step": 95590 + }, + { + "epoch": 14.849332090711401, + "grad_norm": 0.5386288166046143, + "learning_rate": 5.286017100341656e-07, + "loss": 0.0064, + "step": 95600 + }, + { + "epoch": 14.850885368126747, + "grad_norm": 0.20753666758537292, + "learning_rate": 5.262064676907874e-07, + "loss": 0.0028, + "step": 95610 + }, + { + "epoch": 14.852438645542094, + "grad_norm": 0.18066290020942688, + "learning_rate": 5.238166357270924e-07, + "loss": 0.0037, + "step": 95620 + }, + { + "epoch": 14.85399192295744, + "grad_norm": 0.07068445533514023, + "learning_rate": 5.214322144044215e-07, + "loss": 0.0022, + "step": 95630 + }, + { + "epoch": 14.855545200372786, + "grad_norm": 0.46831345558166504, + "learning_rate": 5.190532039835327e-07, + "loss": 0.0047, + "step": 95640 + }, + { + "epoch": 14.857098477788133, + "grad_norm": 0.09801820665597916, + "learning_rate": 5.166796047245903e-07, + "loss": 0.0045, + "step": 95650 + }, + { + "epoch": 14.858651755203478, + "grad_norm": 0.2636486887931824, + "learning_rate": 5.143114168871699e-07, + "loss": 0.004, + "step": 95660 + }, + { + "epoch": 14.860205032618826, + "grad_norm": 0.0952267199754715, + "learning_rate": 5.119486407302532e-07, + "loss": 0.0029, + "step": 95670 + }, + { + "epoch": 14.861758310034173, + "grad_norm": 0.16697850823402405, + "learning_rate": 5.095912765122335e-07, + "loss": 0.003, + "step": 95680 + }, + { + "epoch": 14.863311587449518, + "grad_norm": 0.09302936494350433, + "learning_rate": 5.07239324490899e-07, + "loss": 0.003, + "step": 95690 + }, + { + "epoch": 14.864864864864865, + "grad_norm": 0.2409878820180893, + "learning_rate": 5.048927849234608e-07, + "loss": 0.0031, + "step": 95700 + }, + { + "epoch": 14.866418142280212, + "grad_norm": 0.1536954641342163, + "learning_rate": 5.025516580665358e-07, + "loss": 0.0035, + "step": 95710 + }, + { + "epoch": 14.867971419695557, + "grad_norm": 0.14815691113471985, + "learning_rate": 5.002159441761412e-07, + "loss": 0.0037, + "step": 95720 + }, + { + "epoch": 14.869524697110904, + "grad_norm": 0.32916855812072754, + "learning_rate": 4.978856435077062e-07, + "loss": 0.0036, + "step": 95730 + }, + { + "epoch": 14.87107797452625, + "grad_norm": 0.2400817722082138, + "learning_rate": 4.955607563160769e-07, + "loss": 0.0037, + "step": 95740 + }, + { + "epoch": 14.872631251941597, + "grad_norm": 0.2057466357946396, + "learning_rate": 4.932412828554944e-07, + "loss": 0.0047, + "step": 95750 + }, + { + "epoch": 14.874184529356944, + "grad_norm": 0.8699962496757507, + "learning_rate": 4.909272233796114e-07, + "loss": 0.0019, + "step": 95760 + }, + { + "epoch": 14.875737806772289, + "grad_norm": 0.2780437171459198, + "learning_rate": 4.88618578141492e-07, + "loss": 0.0027, + "step": 95770 + }, + { + "epoch": 14.877291084187636, + "grad_norm": 0.05395263060927391, + "learning_rate": 4.863153473936011e-07, + "loss": 0.0043, + "step": 95780 + }, + { + "epoch": 14.878844361602983, + "grad_norm": 0.05802464112639427, + "learning_rate": 4.840175313878259e-07, + "loss": 0.0058, + "step": 95790 + }, + { + "epoch": 14.880397639018328, + "grad_norm": 0.47989970445632935, + "learning_rate": 4.817251303754433e-07, + "loss": 0.0044, + "step": 95800 + }, + { + "epoch": 14.881950916433675, + "grad_norm": 0.1551019549369812, + "learning_rate": 4.794381446071583e-07, + "loss": 0.0022, + "step": 95810 + }, + { + "epoch": 14.88350419384902, + "grad_norm": 0.8687914609909058, + "learning_rate": 4.771565743330597e-07, + "loss": 0.0043, + "step": 95820 + }, + { + "epoch": 14.885057471264368, + "grad_norm": 0.5274697542190552, + "learning_rate": 4.7488041980266463e-07, + "loss": 0.0032, + "step": 95830 + }, + { + "epoch": 14.886610748679715, + "grad_norm": 0.24297747015953064, + "learning_rate": 4.726096812648795e-07, + "loss": 0.003, + "step": 95840 + }, + { + "epoch": 14.88816402609506, + "grad_norm": 0.532416820526123, + "learning_rate": 4.703443589680445e-07, + "loss": 0.0026, + "step": 95850 + }, + { + "epoch": 14.889717303510407, + "grad_norm": 0.3937786817550659, + "learning_rate": 4.6808445315987826e-07, + "loss": 0.0034, + "step": 95860 + }, + { + "epoch": 14.891270580925752, + "grad_norm": 0.15819641947746277, + "learning_rate": 4.658299640875274e-07, + "loss": 0.0046, + "step": 95870 + }, + { + "epoch": 14.8928238583411, + "grad_norm": 0.18245626986026764, + "learning_rate": 4.635808919975393e-07, + "loss": 0.0039, + "step": 95880 + }, + { + "epoch": 14.894377135756447, + "grad_norm": 0.3579345643520355, + "learning_rate": 4.6133723713586706e-07, + "loss": 0.0025, + "step": 95890 + }, + { + "epoch": 14.895930413171792, + "grad_norm": 0.23759345710277557, + "learning_rate": 4.590989997478756e-07, + "loss": 0.004, + "step": 95900 + }, + { + "epoch": 14.897483690587139, + "grad_norm": 0.06818913668394089, + "learning_rate": 4.5686618007833027e-07, + "loss": 0.0046, + "step": 95910 + }, + { + "epoch": 14.899036968002486, + "grad_norm": 0.3308434784412384, + "learning_rate": 4.54638778371419e-07, + "loss": 0.0033, + "step": 95920 + }, + { + "epoch": 14.900590245417831, + "grad_norm": 0.29324662685394287, + "learning_rate": 4.524167948707192e-07, + "loss": 0.0034, + "step": 95930 + }, + { + "epoch": 14.902143522833178, + "grad_norm": 0.1272604763507843, + "learning_rate": 4.5020022981921986e-07, + "loss": 0.0031, + "step": 95940 + }, + { + "epoch": 14.903696800248525, + "grad_norm": 0.09165835380554199, + "learning_rate": 4.47989083459327e-07, + "loss": 0.0029, + "step": 95950 + }, + { + "epoch": 14.90525007766387, + "grad_norm": 0.4222428798675537, + "learning_rate": 4.457833560328528e-07, + "loss": 0.0052, + "step": 95960 + }, + { + "epoch": 14.906803355079218, + "grad_norm": 0.44908663630485535, + "learning_rate": 4.4358304778100434e-07, + "loss": 0.0032, + "step": 95970 + }, + { + "epoch": 14.908356632494563, + "grad_norm": 0.09220807999372482, + "learning_rate": 4.413881589444058e-07, + "loss": 0.0032, + "step": 95980 + }, + { + "epoch": 14.90990990990991, + "grad_norm": 0.032341767102479935, + "learning_rate": 4.391986897630873e-07, + "loss": 0.0023, + "step": 95990 + }, + { + "epoch": 14.911463187325257, + "grad_norm": 0.3644767999649048, + "learning_rate": 4.370146404764908e-07, + "loss": 0.0037, + "step": 96000 + }, + { + "epoch": 14.913016464740602, + "grad_norm": 0.20315690338611603, + "learning_rate": 4.3483601132345285e-07, + "loss": 0.0035, + "step": 96010 + }, + { + "epoch": 14.91456974215595, + "grad_norm": 0.28268492221832275, + "learning_rate": 4.32662802542233e-07, + "loss": 0.0023, + "step": 96020 + }, + { + "epoch": 14.916123019571295, + "grad_norm": 0.23747655749320984, + "learning_rate": 4.3049501437047444e-07, + "loss": 0.0047, + "step": 96030 + }, + { + "epoch": 14.917676296986642, + "grad_norm": 0.5859247446060181, + "learning_rate": 4.283326470452653e-07, + "loss": 0.0039, + "step": 96040 + }, + { + "epoch": 14.919229574401989, + "grad_norm": 0.10989294201135635, + "learning_rate": 4.2617570080306647e-07, + "loss": 0.0041, + "step": 96050 + }, + { + "epoch": 14.920782851817334, + "grad_norm": 0.058239225298166275, + "learning_rate": 4.2402417587975585e-07, + "loss": 0.0027, + "step": 96060 + }, + { + "epoch": 14.922336129232681, + "grad_norm": 0.22118978202342987, + "learning_rate": 4.218780725106286e-07, + "loss": 0.0024, + "step": 96070 + }, + { + "epoch": 14.923889406648028, + "grad_norm": 0.09372608363628387, + "learning_rate": 4.197373909303748e-07, + "loss": 0.0043, + "step": 96080 + }, + { + "epoch": 14.925442684063373, + "grad_norm": 0.06879892945289612, + "learning_rate": 4.17602131373096e-07, + "loss": 0.0048, + "step": 96090 + }, + { + "epoch": 14.92699596147872, + "grad_norm": 0.1283034086227417, + "learning_rate": 4.154722940723055e-07, + "loss": 0.0031, + "step": 96100 + }, + { + "epoch": 14.928549238894066, + "grad_norm": 0.17090298235416412, + "learning_rate": 4.1334787926091136e-07, + "loss": 0.0036, + "step": 96110 + }, + { + "epoch": 14.930102516309413, + "grad_norm": 0.16057994961738586, + "learning_rate": 4.1122888717124444e-07, + "loss": 0.0029, + "step": 96120 + }, + { + "epoch": 14.93165579372476, + "grad_norm": 0.25737127661705017, + "learning_rate": 4.091153180350249e-07, + "loss": 0.0024, + "step": 96130 + }, + { + "epoch": 14.933209071140105, + "grad_norm": 0.26209700107574463, + "learning_rate": 4.0700717208340125e-07, + "loss": 0.0044, + "step": 96140 + }, + { + "epoch": 14.934762348555452, + "grad_norm": 0.06690148264169693, + "learning_rate": 4.049044495469112e-07, + "loss": 0.0034, + "step": 96150 + }, + { + "epoch": 14.9363156259708, + "grad_norm": 0.3265087306499481, + "learning_rate": 4.0280715065549866e-07, + "loss": 0.006, + "step": 96160 + }, + { + "epoch": 14.937868903386144, + "grad_norm": 0.47363734245300293, + "learning_rate": 4.0071527563853575e-07, + "loss": 0.0026, + "step": 96170 + }, + { + "epoch": 14.939422180801492, + "grad_norm": 0.2770041525363922, + "learning_rate": 3.9862882472477823e-07, + "loss": 0.0034, + "step": 96180 + }, + { + "epoch": 14.940975458216837, + "grad_norm": 0.10324778407812119, + "learning_rate": 3.9654779814239375e-07, + "loss": 0.0046, + "step": 96190 + }, + { + "epoch": 14.942528735632184, + "grad_norm": 0.19238437712192535, + "learning_rate": 3.944721961189668e-07, + "loss": 0.0028, + "step": 96200 + }, + { + "epoch": 14.944082013047531, + "grad_norm": 0.0930587574839592, + "learning_rate": 3.9240201888147697e-07, + "loss": 0.0037, + "step": 96210 + }, + { + "epoch": 14.945635290462876, + "grad_norm": 1.3162051439285278, + "learning_rate": 3.9033726665632096e-07, + "loss": 0.0024, + "step": 96220 + }, + { + "epoch": 14.947188567878223, + "grad_norm": 0.18028950691223145, + "learning_rate": 3.882779396692959e-07, + "loss": 0.0047, + "step": 96230 + }, + { + "epoch": 14.948741845293569, + "grad_norm": 0.41431379318237305, + "learning_rate": 3.8622403814559947e-07, + "loss": 0.0031, + "step": 96240 + }, + { + "epoch": 14.950295122708916, + "grad_norm": 0.5085631608963013, + "learning_rate": 3.841755623098575e-07, + "loss": 0.0037, + "step": 96250 + }, + { + "epoch": 14.951848400124263, + "grad_norm": 0.10065136104822159, + "learning_rate": 3.821325123860742e-07, + "loss": 0.0033, + "step": 96260 + }, + { + "epoch": 14.953401677539608, + "grad_norm": 0.6877995729446411, + "learning_rate": 3.800948885976818e-07, + "loss": 0.0035, + "step": 96270 + }, + { + "epoch": 14.954954954954955, + "grad_norm": 0.2587689757347107, + "learning_rate": 3.780626911675078e-07, + "loss": 0.0027, + "step": 96280 + }, + { + "epoch": 14.956508232370302, + "grad_norm": 0.2861451208591461, + "learning_rate": 3.76035920317791e-07, + "loss": 0.0036, + "step": 96290 + }, + { + "epoch": 14.958061509785647, + "grad_norm": 0.1356072872877121, + "learning_rate": 3.740145762701819e-07, + "loss": 0.0023, + "step": 96300 + }, + { + "epoch": 14.959614787200994, + "grad_norm": 0.06910611689090729, + "learning_rate": 3.7199865924572585e-07, + "loss": 0.003, + "step": 96310 + }, + { + "epoch": 14.961168064616341, + "grad_norm": 0.4355801045894623, + "learning_rate": 3.699881694648799e-07, + "loss": 0.0036, + "step": 96320 + }, + { + "epoch": 14.962721342031687, + "grad_norm": 0.3192095160484314, + "learning_rate": 3.679831071475126e-07, + "loss": 0.0032, + "step": 96330 + }, + { + "epoch": 14.964274619447034, + "grad_norm": 0.21999500691890717, + "learning_rate": 3.65983472512893e-07, + "loss": 0.0036, + "step": 96340 + }, + { + "epoch": 14.965827896862379, + "grad_norm": 0.16377556324005127, + "learning_rate": 3.6398926577969616e-07, + "loss": 0.0033, + "step": 96350 + }, + { + "epoch": 14.967381174277726, + "grad_norm": 0.8456419706344604, + "learning_rate": 3.620004871660032e-07, + "loss": 0.0026, + "step": 96360 + }, + { + "epoch": 14.968934451693073, + "grad_norm": 0.04394194483757019, + "learning_rate": 3.600171368893124e-07, + "loss": 0.0019, + "step": 96370 + }, + { + "epoch": 14.970487729108418, + "grad_norm": 0.24598820507526398, + "learning_rate": 3.580392151665113e-07, + "loss": 0.0042, + "step": 96380 + }, + { + "epoch": 14.972041006523765, + "grad_norm": 0.26245859265327454, + "learning_rate": 3.5606672221391024e-07, + "loss": 0.004, + "step": 96390 + }, + { + "epoch": 14.97359428393911, + "grad_norm": 0.44241198897361755, + "learning_rate": 3.540996582472089e-07, + "loss": 0.0015, + "step": 96400 + }, + { + "epoch": 14.975147561354458, + "grad_norm": 0.44642186164855957, + "learning_rate": 3.521380234815297e-07, + "loss": 0.0034, + "step": 96410 + }, + { + "epoch": 14.976700838769805, + "grad_norm": 0.11521783471107483, + "learning_rate": 3.5018181813138985e-07, + "loss": 0.0042, + "step": 96420 + }, + { + "epoch": 14.97825411618515, + "grad_norm": 0.2376604676246643, + "learning_rate": 3.482310424107238e-07, + "loss": 0.0036, + "step": 96430 + }, + { + "epoch": 14.979807393600497, + "grad_norm": 0.32001927495002747, + "learning_rate": 3.4628569653285537e-07, + "loss": 0.0038, + "step": 96440 + }, + { + "epoch": 14.981360671015844, + "grad_norm": 0.21422912180423737, + "learning_rate": 3.44345780710531e-07, + "loss": 0.004, + "step": 96450 + }, + { + "epoch": 14.98291394843119, + "grad_norm": 0.2119804322719574, + "learning_rate": 3.424112951558978e-07, + "loss": 0.0029, + "step": 96460 + }, + { + "epoch": 14.984467225846537, + "grad_norm": 0.20752151310443878, + "learning_rate": 3.40482240080503e-07, + "loss": 0.0025, + "step": 96470 + }, + { + "epoch": 14.986020503261882, + "grad_norm": 0.2545303702354431, + "learning_rate": 3.3855861569531133e-07, + "loss": 0.0028, + "step": 96480 + }, + { + "epoch": 14.987573780677229, + "grad_norm": 0.29322826862335205, + "learning_rate": 3.3664042221068226e-07, + "loss": 0.0035, + "step": 96490 + }, + { + "epoch": 14.989127058092576, + "grad_norm": 0.11949631571769714, + "learning_rate": 3.347276598363924e-07, + "loss": 0.0027, + "step": 96500 + }, + { + "epoch": 14.990680335507921, + "grad_norm": 0.275370717048645, + "learning_rate": 3.328203287816134e-07, + "loss": 0.0044, + "step": 96510 + }, + { + "epoch": 14.992233612923268, + "grad_norm": 0.29724353551864624, + "learning_rate": 3.3091842925493386e-07, + "loss": 0.0019, + "step": 96520 + }, + { + "epoch": 14.993786890338615, + "grad_norm": 0.07179911434650421, + "learning_rate": 3.2902196146433196e-07, + "loss": 0.0033, + "step": 96530 + }, + { + "epoch": 14.99534016775396, + "grad_norm": 0.08511757850646973, + "learning_rate": 3.2713092561720836e-07, + "loss": 0.0037, + "step": 96540 + }, + { + "epoch": 14.996893445169308, + "grad_norm": 1.0898890495300293, + "learning_rate": 3.2524532192036996e-07, + "loss": 0.0042, + "step": 96550 + }, + { + "epoch": 14.998446722584653, + "grad_norm": 0.09245721995830536, + "learning_rate": 3.2336515058001836e-07, + "loss": 0.0032, + "step": 96560 + }, + { + "epoch": 15.0, + "grad_norm": 0.44544151425361633, + "learning_rate": 3.2149041180176143e-07, + "loss": 0.0028, + "step": 96570 + }, + { + "epoch": 15.001553277415347, + "grad_norm": 0.22474277019500732, + "learning_rate": 3.1962110579062954e-07, + "loss": 0.0032, + "step": 96580 + }, + { + "epoch": 15.003106554830692, + "grad_norm": 0.23460404574871063, + "learning_rate": 3.177572327510314e-07, + "loss": 0.0053, + "step": 96590 + }, + { + "epoch": 15.00465983224604, + "grad_norm": 0.0533909797668457, + "learning_rate": 3.158987928868151e-07, + "loss": 0.0047, + "step": 96600 + }, + { + "epoch": 15.006213109661385, + "grad_norm": 0.22813157737255096, + "learning_rate": 3.1404578640120144e-07, + "loss": 0.0041, + "step": 96610 + }, + { + "epoch": 15.007766387076732, + "grad_norm": 0.09009665995836258, + "learning_rate": 3.1219821349683933e-07, + "loss": 0.0021, + "step": 96620 + }, + { + "epoch": 15.009319664492079, + "grad_norm": 0.9484819769859314, + "learning_rate": 3.103560743757783e-07, + "loss": 0.004, + "step": 96630 + }, + { + "epoch": 15.010872941907424, + "grad_norm": 0.385040819644928, + "learning_rate": 3.0851936923946834e-07, + "loss": 0.0036, + "step": 96640 + }, + { + "epoch": 15.012426219322771, + "grad_norm": 0.4073924422264099, + "learning_rate": 3.06688098288771e-07, + "loss": 0.0041, + "step": 96650 + }, + { + "epoch": 15.013979496738118, + "grad_norm": 0.7850580215454102, + "learning_rate": 3.0486226172394825e-07, + "loss": 0.0037, + "step": 96660 + }, + { + "epoch": 15.015532774153463, + "grad_norm": 0.09885238856077194, + "learning_rate": 3.030418597446738e-07, + "loss": 0.003, + "step": 96670 + }, + { + "epoch": 15.01708605156881, + "grad_norm": 0.3888072967529297, + "learning_rate": 3.012268925500272e-07, + "loss": 0.0018, + "step": 96680 + }, + { + "epoch": 15.018639328984156, + "grad_norm": 0.10123437643051147, + "learning_rate": 2.994173603384831e-07, + "loss": 0.0022, + "step": 96690 + }, + { + "epoch": 15.020192606399503, + "grad_norm": 0.4217413663864136, + "learning_rate": 2.9761326330793317e-07, + "loss": 0.0024, + "step": 96700 + }, + { + "epoch": 15.02174588381485, + "grad_norm": 0.48924803733825684, + "learning_rate": 2.958146016556751e-07, + "loss": 0.0036, + "step": 96710 + }, + { + "epoch": 15.023299161230195, + "grad_norm": 0.11202267557382584, + "learning_rate": 2.94021375578396e-07, + "loss": 0.0032, + "step": 96720 + }, + { + "epoch": 15.024852438645542, + "grad_norm": 0.8313753604888916, + "learning_rate": 2.9223358527221135e-07, + "loss": 0.003, + "step": 96730 + }, + { + "epoch": 15.02640571606089, + "grad_norm": 0.05885996297001839, + "learning_rate": 2.904512309326257e-07, + "loss": 0.0034, + "step": 96740 + }, + { + "epoch": 15.027958993476235, + "grad_norm": 0.06881437450647354, + "learning_rate": 2.8867431275455545e-07, + "loss": 0.005, + "step": 96750 + }, + { + "epoch": 15.029512270891582, + "grad_norm": 0.12362848967313766, + "learning_rate": 2.86902830932323e-07, + "loss": 0.002, + "step": 96760 + }, + { + "epoch": 15.031065548306927, + "grad_norm": 0.11233799159526825, + "learning_rate": 2.851367856596565e-07, + "loss": 0.0038, + "step": 96770 + }, + { + "epoch": 15.032618825722274, + "grad_norm": 0.2174742966890335, + "learning_rate": 2.83376177129685e-07, + "loss": 0.0035, + "step": 96780 + }, + { + "epoch": 15.034172103137621, + "grad_norm": 1.0380715131759644, + "learning_rate": 2.8162100553494884e-07, + "loss": 0.0032, + "step": 96790 + }, + { + "epoch": 15.035725380552966, + "grad_norm": 0.7752842903137207, + "learning_rate": 2.798712710673834e-07, + "loss": 0.0035, + "step": 96800 + }, + { + "epoch": 15.037278657968313, + "grad_norm": 0.22375905513763428, + "learning_rate": 2.7812697391834676e-07, + "loss": 0.0027, + "step": 96810 + }, + { + "epoch": 15.03883193538366, + "grad_norm": 0.17525656521320343, + "learning_rate": 2.7638811427858627e-07, + "loss": 0.0045, + "step": 96820 + }, + { + "epoch": 15.040385212799006, + "grad_norm": 0.07267702370882034, + "learning_rate": 2.746546923382665e-07, + "loss": 0.0021, + "step": 96830 + }, + { + "epoch": 15.041938490214353, + "grad_norm": 0.11182533204555511, + "learning_rate": 2.729267082869469e-07, + "loss": 0.0028, + "step": 96840 + }, + { + "epoch": 15.043491767629698, + "grad_norm": 0.12514035403728485, + "learning_rate": 2.712041623135986e-07, + "loss": 0.0026, + "step": 96850 + }, + { + "epoch": 15.045045045045045, + "grad_norm": 0.42144709825515747, + "learning_rate": 2.694870546065931e-07, + "loss": 0.0021, + "step": 96860 + }, + { + "epoch": 15.046598322460392, + "grad_norm": 0.707868754863739, + "learning_rate": 2.677753853537246e-07, + "loss": 0.0041, + "step": 96870 + }, + { + "epoch": 15.048151599875737, + "grad_norm": 0.09395677596330643, + "learning_rate": 2.660691547421601e-07, + "loss": 0.0036, + "step": 96880 + }, + { + "epoch": 15.049704877291084, + "grad_norm": 0.2721492648124695, + "learning_rate": 2.6436836295850033e-07, + "loss": 0.0036, + "step": 96890 + }, + { + "epoch": 15.05125815470643, + "grad_norm": 0.07744193822145462, + "learning_rate": 2.6267301018874093e-07, + "loss": 0.0041, + "step": 96900 + }, + { + "epoch": 15.052811432121777, + "grad_norm": 0.12917029857635498, + "learning_rate": 2.6098309661828355e-07, + "loss": 0.0028, + "step": 96910 + }, + { + "epoch": 15.054364709537124, + "grad_norm": 0.0623326450586319, + "learning_rate": 2.592986224319305e-07, + "loss": 0.0026, + "step": 96920 + }, + { + "epoch": 15.055917986952469, + "grad_norm": 0.299805223941803, + "learning_rate": 2.576195878139009e-07, + "loss": 0.0066, + "step": 96930 + }, + { + "epoch": 15.057471264367816, + "grad_norm": 0.34658491611480713, + "learning_rate": 2.5594599294780364e-07, + "loss": 0.0021, + "step": 96940 + }, + { + "epoch": 15.059024541783163, + "grad_norm": 0.5776606202125549, + "learning_rate": 2.542778380166644e-07, + "loss": 0.0024, + "step": 96950 + }, + { + "epoch": 15.060577819198508, + "grad_norm": 0.16279922425746918, + "learning_rate": 2.526151232029095e-07, + "loss": 0.0031, + "step": 96960 + }, + { + "epoch": 15.062131096613856, + "grad_norm": 0.18089862167835236, + "learning_rate": 2.5095784868836567e-07, + "loss": 0.003, + "step": 96970 + }, + { + "epoch": 15.0636843740292, + "grad_norm": 0.644013524055481, + "learning_rate": 2.493060146542825e-07, + "loss": 0.0029, + "step": 96980 + }, + { + "epoch": 15.065237651444548, + "grad_norm": 1.0889078378677368, + "learning_rate": 2.476596212812876e-07, + "loss": 0.0039, + "step": 96990 + }, + { + "epoch": 15.066790928859895, + "grad_norm": 0.8129891157150269, + "learning_rate": 2.4601866874943703e-07, + "loss": 0.0045, + "step": 97000 + }, + { + "epoch": 15.06834420627524, + "grad_norm": 0.08489404618740082, + "learning_rate": 2.4438315723818163e-07, + "loss": 0.0025, + "step": 97010 + }, + { + "epoch": 15.069897483690587, + "grad_norm": 0.13187873363494873, + "learning_rate": 2.4275308692637833e-07, + "loss": 0.0028, + "step": 97020 + }, + { + "epoch": 15.071450761105934, + "grad_norm": 0.46506333351135254, + "learning_rate": 2.411284579922901e-07, + "loss": 0.0033, + "step": 97030 + }, + { + "epoch": 15.07300403852128, + "grad_norm": 0.25115343928337097, + "learning_rate": 2.395092706135804e-07, + "loss": 0.0037, + "step": 97040 + }, + { + "epoch": 15.074557315936627, + "grad_norm": 0.7514327764511108, + "learning_rate": 2.3789552496731872e-07, + "loss": 0.0029, + "step": 97050 + }, + { + "epoch": 15.076110593351972, + "grad_norm": 0.12432140111923218, + "learning_rate": 2.3628722122999158e-07, + "loss": 0.0036, + "step": 97060 + }, + { + "epoch": 15.077663870767319, + "grad_norm": 0.2786681056022644, + "learning_rate": 2.3468435957747503e-07, + "loss": 0.0028, + "step": 97070 + }, + { + "epoch": 15.079217148182666, + "grad_norm": 0.2597998380661011, + "learning_rate": 2.330869401850566e-07, + "loss": 0.0034, + "step": 97080 + }, + { + "epoch": 15.080770425598011, + "grad_norm": 0.09299812465906143, + "learning_rate": 2.3149496322742437e-07, + "loss": 0.0023, + "step": 97090 + }, + { + "epoch": 15.082323703013358, + "grad_norm": 0.15156519412994385, + "learning_rate": 2.2990842887868346e-07, + "loss": 0.0057, + "step": 97100 + }, + { + "epoch": 15.083876980428705, + "grad_norm": 0.5464541912078857, + "learning_rate": 2.2832733731232292e-07, + "loss": 0.004, + "step": 97110 + }, + { + "epoch": 15.08543025784405, + "grad_norm": 0.059962689876556396, + "learning_rate": 2.2675168870124886e-07, + "loss": 0.0021, + "step": 97120 + }, + { + "epoch": 15.086983535259398, + "grad_norm": 0.1444319784641266, + "learning_rate": 2.2518148321778455e-07, + "loss": 0.0035, + "step": 97130 + }, + { + "epoch": 15.088536812674743, + "grad_norm": 0.14209865033626556, + "learning_rate": 2.2361672103363706e-07, + "loss": 0.0031, + "step": 97140 + }, + { + "epoch": 15.09009009009009, + "grad_norm": 0.1931445598602295, + "learning_rate": 2.220574023199251e-07, + "loss": 0.0047, + "step": 97150 + }, + { + "epoch": 15.091643367505437, + "grad_norm": 0.3967907726764679, + "learning_rate": 2.2050352724717894e-07, + "loss": 0.0046, + "step": 97160 + }, + { + "epoch": 15.093196644920782, + "grad_norm": 0.3773222863674164, + "learning_rate": 2.1895509598532372e-07, + "loss": 0.0031, + "step": 97170 + }, + { + "epoch": 15.09474992233613, + "grad_norm": 0.14826054871082306, + "learning_rate": 2.1741210870369066e-07, + "loss": 0.0023, + "step": 97180 + }, + { + "epoch": 15.096303199751475, + "grad_norm": 0.06780056655406952, + "learning_rate": 2.1587456557102814e-07, + "loss": 0.0021, + "step": 97190 + }, + { + "epoch": 15.097856477166822, + "grad_norm": 0.20055757462978363, + "learning_rate": 2.143424667554683e-07, + "loss": 0.0048, + "step": 97200 + }, + { + "epoch": 15.099409754582169, + "grad_norm": 0.6593837738037109, + "learning_rate": 2.128158124245716e-07, + "loss": 0.0038, + "step": 97210 + }, + { + "epoch": 15.100963031997514, + "grad_norm": 0.2760092616081238, + "learning_rate": 2.1129460274527668e-07, + "loss": 0.0028, + "step": 97220 + }, + { + "epoch": 15.102516309412861, + "grad_norm": 0.32137519121170044, + "learning_rate": 2.0977883788395046e-07, + "loss": 0.0027, + "step": 97230 + }, + { + "epoch": 15.104069586828208, + "grad_norm": 0.16553521156311035, + "learning_rate": 2.0826851800634928e-07, + "loss": 0.0026, + "step": 97240 + }, + { + "epoch": 15.105622864243553, + "grad_norm": 0.8394765853881836, + "learning_rate": 2.0676364327764653e-07, + "loss": 0.0043, + "step": 97250 + }, + { + "epoch": 15.1071761416589, + "grad_norm": 0.12202318757772446, + "learning_rate": 2.0526421386240502e-07, + "loss": 0.0024, + "step": 97260 + }, + { + "epoch": 15.108729419074246, + "grad_norm": 0.5023363828659058, + "learning_rate": 2.0377022992461025e-07, + "loss": 0.0029, + "step": 97270 + }, + { + "epoch": 15.110282696489593, + "grad_norm": 0.3074291944503784, + "learning_rate": 2.0228169162763156e-07, + "loss": 0.0031, + "step": 97280 + }, + { + "epoch": 15.11183597390494, + "grad_norm": 0.07161349058151245, + "learning_rate": 2.0079859913425536e-07, + "loss": 0.0028, + "step": 97290 + }, + { + "epoch": 15.113389251320285, + "grad_norm": 0.13987624645233154, + "learning_rate": 1.9932095260667417e-07, + "loss": 0.004, + "step": 97300 + }, + { + "epoch": 15.114942528735632, + "grad_norm": 0.07378297299146652, + "learning_rate": 1.9784875220648092e-07, + "loss": 0.0032, + "step": 97310 + }, + { + "epoch": 15.11649580615098, + "grad_norm": 1.3716610670089722, + "learning_rate": 1.9638199809466907e-07, + "loss": 0.0029, + "step": 97320 + }, + { + "epoch": 15.118049083566325, + "grad_norm": 0.07220827788114548, + "learning_rate": 1.9492069043164364e-07, + "loss": 0.0046, + "step": 97330 + }, + { + "epoch": 15.119602360981672, + "grad_norm": 0.7389519810676575, + "learning_rate": 1.9346482937721567e-07, + "loss": 0.004, + "step": 97340 + }, + { + "epoch": 15.121155638397017, + "grad_norm": 0.7678717374801636, + "learning_rate": 1.920144150905856e-07, + "loss": 0.0038, + "step": 97350 + }, + { + "epoch": 15.122708915812364, + "grad_norm": 0.2065511792898178, + "learning_rate": 1.9056944773037655e-07, + "loss": 0.0036, + "step": 97360 + }, + { + "epoch": 15.124262193227711, + "grad_norm": 0.1545487344264984, + "learning_rate": 1.8912992745460102e-07, + "loss": 0.0028, + "step": 97370 + }, + { + "epoch": 15.125815470643056, + "grad_norm": 0.5639976859092712, + "learning_rate": 1.876958544206886e-07, + "loss": 0.005, + "step": 97380 + }, + { + "epoch": 15.127368748058403, + "grad_norm": 0.1415879726409912, + "learning_rate": 1.8626722878546942e-07, + "loss": 0.003, + "step": 97390 + }, + { + "epoch": 15.12892202547375, + "grad_norm": 0.3761442303657532, + "learning_rate": 1.8484405070516851e-07, + "loss": 0.0019, + "step": 97400 + }, + { + "epoch": 15.130475302889096, + "grad_norm": 0.08477209508419037, + "learning_rate": 1.8342632033542805e-07, + "loss": 0.0025, + "step": 97410 + }, + { + "epoch": 15.132028580304443, + "grad_norm": 0.26451659202575684, + "learning_rate": 1.8201403783129066e-07, + "loss": 0.0027, + "step": 97420 + }, + { + "epoch": 15.133581857719788, + "grad_norm": 0.09712015092372894, + "learning_rate": 1.806072033471884e-07, + "loss": 0.0026, + "step": 97430 + }, + { + "epoch": 15.135135135135135, + "grad_norm": 0.512046754360199, + "learning_rate": 1.79205817036987e-07, + "loss": 0.0031, + "step": 97440 + }, + { + "epoch": 15.136688412550482, + "grad_norm": 0.1472407877445221, + "learning_rate": 1.7780987905393065e-07, + "loss": 0.0031, + "step": 97450 + }, + { + "epoch": 15.138241689965827, + "grad_norm": 0.09558561444282532, + "learning_rate": 1.764193895506805e-07, + "loss": 0.0029, + "step": 97460 + }, + { + "epoch": 15.139794967381174, + "grad_norm": 0.295655220746994, + "learning_rate": 1.7503434867929824e-07, + "loss": 0.0026, + "step": 97470 + }, + { + "epoch": 15.141348244796522, + "grad_norm": 0.512282133102417, + "learning_rate": 1.7365475659125165e-07, + "loss": 0.0035, + "step": 97480 + }, + { + "epoch": 15.142901522211867, + "grad_norm": 0.11155537515878677, + "learning_rate": 1.7228061343740332e-07, + "loss": 0.0034, + "step": 97490 + }, + { + "epoch": 15.144454799627214, + "grad_norm": 0.5794085264205933, + "learning_rate": 1.7091191936803308e-07, + "loss": 0.003, + "step": 97500 + }, + { + "epoch": 15.14600807704256, + "grad_norm": 0.13835425674915314, + "learning_rate": 1.6954867453281563e-07, + "loss": 0.0029, + "step": 97510 + }, + { + "epoch": 15.147561354457906, + "grad_norm": 0.2991925776004791, + "learning_rate": 1.6819087908084285e-07, + "loss": 0.0032, + "step": 97520 + }, + { + "epoch": 15.149114631873253, + "grad_norm": 0.06663323193788528, + "learning_rate": 1.6683853316058485e-07, + "loss": 0.0045, + "step": 97530 + }, + { + "epoch": 15.150667909288599, + "grad_norm": 0.4087291657924652, + "learning_rate": 1.6549163691995106e-07, + "loss": 0.0028, + "step": 97540 + }, + { + "epoch": 15.152221186703946, + "grad_norm": 0.10247359424829483, + "learning_rate": 1.641501905062237e-07, + "loss": 0.0024, + "step": 97550 + }, + { + "epoch": 15.15377446411929, + "grad_norm": 0.24030189216136932, + "learning_rate": 1.628141940661021e-07, + "loss": 0.0054, + "step": 97560 + }, + { + "epoch": 15.155327741534638, + "grad_norm": 0.07882718741893768, + "learning_rate": 1.6148364774569158e-07, + "loss": 0.0029, + "step": 97570 + }, + { + "epoch": 15.156881018949985, + "grad_norm": 0.5444022417068481, + "learning_rate": 1.6015855169050355e-07, + "loss": 0.004, + "step": 97580 + }, + { + "epoch": 15.15843429636533, + "grad_norm": 0.7649816274642944, + "learning_rate": 1.5883890604543873e-07, + "loss": 0.0049, + "step": 97590 + }, + { + "epoch": 15.159987573780677, + "grad_norm": 0.3430691063404083, + "learning_rate": 1.5752471095482057e-07, + "loss": 0.0047, + "step": 97600 + }, + { + "epoch": 15.161540851196024, + "grad_norm": 0.22542573511600494, + "learning_rate": 1.5621596656235638e-07, + "loss": 0.0038, + "step": 97610 + }, + { + "epoch": 15.16309412861137, + "grad_norm": 0.2344464212656021, + "learning_rate": 1.5491267301117608e-07, + "loss": 0.0031, + "step": 97620 + }, + { + "epoch": 15.164647406026717, + "grad_norm": 0.060443777590990067, + "learning_rate": 1.536148304438101e-07, + "loss": 0.0033, + "step": 97630 + }, + { + "epoch": 15.166200683442062, + "grad_norm": 0.5108956098556519, + "learning_rate": 1.5232243900217823e-07, + "loss": 0.0031, + "step": 97640 + }, + { + "epoch": 15.167753960857409, + "grad_norm": 0.18865418434143066, + "learning_rate": 1.51035498827623e-07, + "loss": 0.0031, + "step": 97650 + }, + { + "epoch": 15.169307238272756, + "grad_norm": 0.13524560630321503, + "learning_rate": 1.497540100608763e-07, + "loss": 0.0033, + "step": 97660 + }, + { + "epoch": 15.170860515688101, + "grad_norm": 0.5936834812164307, + "learning_rate": 1.4847797284208153e-07, + "loss": 0.0036, + "step": 97670 + }, + { + "epoch": 15.172413793103448, + "grad_norm": 0.1387736201286316, + "learning_rate": 1.472073873107882e-07, + "loss": 0.0032, + "step": 97680 + }, + { + "epoch": 15.173967070518795, + "grad_norm": 0.276363343000412, + "learning_rate": 1.459422536059407e-07, + "loss": 0.0027, + "step": 97690 + }, + { + "epoch": 15.17552034793414, + "grad_norm": 0.062142323702573776, + "learning_rate": 1.4468257186589508e-07, + "loss": 0.0038, + "step": 97700 + }, + { + "epoch": 15.177073625349488, + "grad_norm": 0.07649683952331543, + "learning_rate": 1.4342834222840217e-07, + "loss": 0.0018, + "step": 97710 + }, + { + "epoch": 15.178626902764833, + "grad_norm": 0.958720326423645, + "learning_rate": 1.4217956483063566e-07, + "loss": 0.0026, + "step": 97720 + }, + { + "epoch": 15.18018018018018, + "grad_norm": 0.32740938663482666, + "learning_rate": 1.409362398091474e-07, + "loss": 0.0024, + "step": 97730 + }, + { + "epoch": 15.181733457595527, + "grad_norm": 0.06989184767007828, + "learning_rate": 1.3969836729990638e-07, + "loss": 0.0041, + "step": 97740 + }, + { + "epoch": 15.183286735010872, + "grad_norm": 0.23072528839111328, + "learning_rate": 1.384659474382932e-07, + "loss": 0.003, + "step": 97750 + }, + { + "epoch": 15.18484001242622, + "grad_norm": 0.13475923240184784, + "learning_rate": 1.372389803590779e-07, + "loss": 0.0016, + "step": 97760 + }, + { + "epoch": 15.186393289841567, + "grad_norm": 0.7113515734672546, + "learning_rate": 1.360174661964364e-07, + "loss": 0.0041, + "step": 97770 + }, + { + "epoch": 15.187946567256912, + "grad_norm": 0.5291273593902588, + "learning_rate": 1.3480140508396188e-07, + "loss": 0.0051, + "step": 97780 + }, + { + "epoch": 15.189499844672259, + "grad_norm": 0.38361668586730957, + "learning_rate": 1.335907971546313e-07, + "loss": 0.0032, + "step": 97790 + }, + { + "epoch": 15.191053122087604, + "grad_norm": 0.2783450484275818, + "learning_rate": 1.3238564254083875e-07, + "loss": 0.0027, + "step": 97800 + }, + { + "epoch": 15.192606399502951, + "grad_norm": 0.1737571358680725, + "learning_rate": 1.3118594137437323e-07, + "loss": 0.0025, + "step": 97810 + }, + { + "epoch": 15.194159676918298, + "grad_norm": 0.2648696303367615, + "learning_rate": 1.2999169378644094e-07, + "loss": 0.0032, + "step": 97820 + }, + { + "epoch": 15.195712954333644, + "grad_norm": 0.16331562399864197, + "learning_rate": 1.2880289990763738e-07, + "loss": 0.0036, + "step": 97830 + }, + { + "epoch": 15.19726623174899, + "grad_norm": 0.2913582921028137, + "learning_rate": 1.2761955986796968e-07, + "loss": 0.0029, + "step": 97840 + }, + { + "epoch": 15.198819509164336, + "grad_norm": 0.23572269082069397, + "learning_rate": 1.264416737968399e-07, + "loss": 0.0052, + "step": 97850 + }, + { + "epoch": 15.200372786579683, + "grad_norm": 0.5953500866889954, + "learning_rate": 1.2526924182307275e-07, + "loss": 0.0035, + "step": 97860 + }, + { + "epoch": 15.20192606399503, + "grad_norm": 0.15676727890968323, + "learning_rate": 1.2410226407487124e-07, + "loss": 0.0026, + "step": 97870 + }, + { + "epoch": 15.203479341410375, + "grad_norm": 0.1392638236284256, + "learning_rate": 1.229407406798555e-07, + "loss": 0.0025, + "step": 97880 + }, + { + "epoch": 15.205032618825722, + "grad_norm": 0.15136206150054932, + "learning_rate": 1.2178467176505726e-07, + "loss": 0.0039, + "step": 97890 + }, + { + "epoch": 15.20658589624107, + "grad_norm": 0.19304288923740387, + "learning_rate": 1.2063405745689205e-07, + "loss": 0.0044, + "step": 97900 + }, + { + "epoch": 15.208139173656415, + "grad_norm": 0.35343191027641296, + "learning_rate": 1.1948889788119812e-07, + "loss": 0.0028, + "step": 97910 + }, + { + "epoch": 15.209692451071762, + "grad_norm": 0.19521978497505188, + "learning_rate": 1.1834919316320303e-07, + "loss": 0.0043, + "step": 97920 + }, + { + "epoch": 15.211245728487107, + "grad_norm": 0.2014913409948349, + "learning_rate": 1.1721494342754048e-07, + "loss": 0.0027, + "step": 97930 + }, + { + "epoch": 15.212799005902454, + "grad_norm": 0.48483943939208984, + "learning_rate": 1.1608614879825563e-07, + "loss": 0.0032, + "step": 97940 + }, + { + "epoch": 15.214352283317801, + "grad_norm": 0.05964398384094238, + "learning_rate": 1.1496280939879422e-07, + "loss": 0.0028, + "step": 97950 + }, + { + "epoch": 15.215905560733146, + "grad_norm": 0.33862173557281494, + "learning_rate": 1.1384492535199687e-07, + "loss": 0.0038, + "step": 97960 + }, + { + "epoch": 15.217458838148493, + "grad_norm": 0.3811875581741333, + "learning_rate": 1.1273249678011578e-07, + "loss": 0.0029, + "step": 97970 + }, + { + "epoch": 15.21901211556384, + "grad_norm": 0.07544052600860596, + "learning_rate": 1.1162552380480362e-07, + "loss": 0.0028, + "step": 97980 + }, + { + "epoch": 15.220565392979186, + "grad_norm": 0.1486227810382843, + "learning_rate": 1.1052400654711359e-07, + "loss": 0.0024, + "step": 97990 + }, + { + "epoch": 15.222118670394533, + "grad_norm": 0.31837308406829834, + "learning_rate": 1.0942794512751597e-07, + "loss": 0.0024, + "step": 98000 + }, + { + "epoch": 15.223671947809878, + "grad_norm": 0.45138734579086304, + "learning_rate": 1.0833733966587045e-07, + "loss": 0.0027, + "step": 98010 + }, + { + "epoch": 15.225225225225225, + "grad_norm": 0.05923619866371155, + "learning_rate": 1.0725219028143718e-07, + "loss": 0.003, + "step": 98020 + }, + { + "epoch": 15.226778502640572, + "grad_norm": 0.26571497321128845, + "learning_rate": 1.0617249709289345e-07, + "loss": 0.0036, + "step": 98030 + }, + { + "epoch": 15.228331780055917, + "grad_norm": 0.5277324318885803, + "learning_rate": 1.0509826021831703e-07, + "loss": 0.0047, + "step": 98040 + }, + { + "epoch": 15.229885057471265, + "grad_norm": 0.30100470781326294, + "learning_rate": 1.0402947977517508e-07, + "loss": 0.005, + "step": 98050 + }, + { + "epoch": 15.231438334886612, + "grad_norm": 0.27291467785835266, + "learning_rate": 1.0296615588035185e-07, + "loss": 0.0026, + "step": 98060 + }, + { + "epoch": 15.232991612301957, + "grad_norm": 0.13546127080917358, + "learning_rate": 1.0190828865012659e-07, + "loss": 0.0031, + "step": 98070 + }, + { + "epoch": 15.234544889717304, + "grad_norm": 0.257010817527771, + "learning_rate": 1.008558782001956e-07, + "loss": 0.003, + "step": 98080 + }, + { + "epoch": 15.23609816713265, + "grad_norm": 0.21514204144477844, + "learning_rate": 9.980892464563906e-08, + "loss": 0.0025, + "step": 98090 + }, + { + "epoch": 15.237651444547996, + "grad_norm": 0.6875498294830322, + "learning_rate": 9.876742810095985e-08, + "loss": 0.0031, + "step": 98100 + }, + { + "epoch": 15.239204721963343, + "grad_norm": 0.24631240963935852, + "learning_rate": 9.77313886800446e-08, + "loss": 0.0047, + "step": 98110 + }, + { + "epoch": 15.240757999378689, + "grad_norm": 0.8222600817680359, + "learning_rate": 9.670080649619717e-08, + "loss": 0.0041, + "step": 98120 + }, + { + "epoch": 15.242311276794036, + "grad_norm": 0.07797864824533463, + "learning_rate": 9.567568166212737e-08, + "loss": 0.0031, + "step": 98130 + }, + { + "epoch": 15.243864554209381, + "grad_norm": 0.20385780930519104, + "learning_rate": 9.465601428992888e-08, + "loss": 0.0026, + "step": 98140 + }, + { + "epoch": 15.245417831624728, + "grad_norm": 1.1071218252182007, + "learning_rate": 9.364180449111803e-08, + "loss": 0.0043, + "step": 98150 + }, + { + "epoch": 15.246971109040075, + "grad_norm": 0.10761833935976028, + "learning_rate": 9.263305237661169e-08, + "loss": 0.0045, + "step": 98160 + }, + { + "epoch": 15.24852438645542, + "grad_norm": 0.290240079164505, + "learning_rate": 9.162975805671603e-08, + "loss": 0.0041, + "step": 98170 + }, + { + "epoch": 15.250077663870767, + "grad_norm": 0.164792001247406, + "learning_rate": 9.063192164115442e-08, + "loss": 0.0035, + "step": 98180 + }, + { + "epoch": 15.251630941286114, + "grad_norm": 0.13834907114505768, + "learning_rate": 8.963954323904511e-08, + "loss": 0.0027, + "step": 98190 + }, + { + "epoch": 15.25318421870146, + "grad_norm": 0.10779107362031937, + "learning_rate": 8.865262295891796e-08, + "loss": 0.0026, + "step": 98200 + }, + { + "epoch": 15.254737496116807, + "grad_norm": 0.9805048108100891, + "learning_rate": 8.767116090870331e-08, + "loss": 0.0029, + "step": 98210 + }, + { + "epoch": 15.256290773532152, + "grad_norm": 0.32387983798980713, + "learning_rate": 8.669515719572086e-08, + "loss": 0.0031, + "step": 98220 + }, + { + "epoch": 15.257844050947499, + "grad_norm": 0.4606974124908447, + "learning_rate": 8.572461192671855e-08, + "loss": 0.0046, + "step": 98230 + }, + { + "epoch": 15.259397328362846, + "grad_norm": 0.09274574369192123, + "learning_rate": 8.475952520782815e-08, + "loss": 0.0029, + "step": 98240 + }, + { + "epoch": 15.260950605778191, + "grad_norm": 0.06417399644851685, + "learning_rate": 8.379989714458747e-08, + "loss": 0.0028, + "step": 98250 + }, + { + "epoch": 15.262503883193538, + "grad_norm": 0.021588753908872604, + "learning_rate": 8.284572784194034e-08, + "loss": 0.0038, + "step": 98260 + }, + { + "epoch": 15.264057160608886, + "grad_norm": 0.05188082158565521, + "learning_rate": 8.189701740424216e-08, + "loss": 0.0029, + "step": 98270 + }, + { + "epoch": 15.26561043802423, + "grad_norm": 0.3221171796321869, + "learning_rate": 8.095376593522663e-08, + "loss": 0.0032, + "step": 98280 + }, + { + "epoch": 15.267163715439578, + "grad_norm": 0.08947868645191193, + "learning_rate": 8.001597353806123e-08, + "loss": 0.0045, + "step": 98290 + }, + { + "epoch": 15.268716992854923, + "grad_norm": 0.8091902136802673, + "learning_rate": 7.90836403152917e-08, + "loss": 0.0036, + "step": 98300 + }, + { + "epoch": 15.27027027027027, + "grad_norm": 0.19783879816532135, + "learning_rate": 7.815676636888092e-08, + "loss": 0.0033, + "step": 98310 + }, + { + "epoch": 15.271823547685617, + "grad_norm": 0.9269307255744934, + "learning_rate": 7.723535180019226e-08, + "loss": 0.0052, + "step": 98320 + }, + { + "epoch": 15.273376825100963, + "grad_norm": 1.1653780937194824, + "learning_rate": 7.6319396709984e-08, + "loss": 0.0025, + "step": 98330 + }, + { + "epoch": 15.27493010251631, + "grad_norm": 0.18258236348628998, + "learning_rate": 7.540890119843158e-08, + "loss": 0.004, + "step": 98340 + }, + { + "epoch": 15.276483379931657, + "grad_norm": 0.42159774899482727, + "learning_rate": 7.450386536509979e-08, + "loss": 0.0043, + "step": 98350 + }, + { + "epoch": 15.278036657347002, + "grad_norm": 0.4826957583427429, + "learning_rate": 7.360428930895947e-08, + "loss": 0.0024, + "step": 98360 + }, + { + "epoch": 15.279589934762349, + "grad_norm": 0.09252256155014038, + "learning_rate": 7.271017312839302e-08, + "loss": 0.0034, + "step": 98370 + }, + { + "epoch": 15.281143212177694, + "grad_norm": 0.4074574410915375, + "learning_rate": 7.182151692118333e-08, + "loss": 0.0053, + "step": 98380 + }, + { + "epoch": 15.282696489593041, + "grad_norm": 0.07087220996618271, + "learning_rate": 7.093832078449713e-08, + "loss": 0.0018, + "step": 98390 + }, + { + "epoch": 15.284249767008388, + "grad_norm": 0.3244444727897644, + "learning_rate": 7.006058481493493e-08, + "loss": 0.0045, + "step": 98400 + }, + { + "epoch": 15.285803044423734, + "grad_norm": 0.8322457671165466, + "learning_rate": 6.918830910847552e-08, + "loss": 0.0028, + "step": 98410 + }, + { + "epoch": 15.28735632183908, + "grad_norm": 1.0619118213653564, + "learning_rate": 6.832149376052033e-08, + "loss": 0.0069, + "step": 98420 + }, + { + "epoch": 15.288909599254428, + "grad_norm": 0.2852736711502075, + "learning_rate": 6.746013886584912e-08, + "loss": 0.0039, + "step": 98430 + }, + { + "epoch": 15.290462876669773, + "grad_norm": 0.9277217388153076, + "learning_rate": 6.660424451866431e-08, + "loss": 0.0025, + "step": 98440 + }, + { + "epoch": 15.29201615408512, + "grad_norm": 0.12327990680932999, + "learning_rate": 6.575381081256327e-08, + "loss": 0.0028, + "step": 98450 + }, + { + "epoch": 15.293569431500465, + "grad_norm": 0.27842628955841064, + "learning_rate": 6.49088378405549e-08, + "loss": 0.0035, + "step": 98460 + }, + { + "epoch": 15.295122708915812, + "grad_norm": 0.9838252067565918, + "learning_rate": 6.406932569503755e-08, + "loss": 0.0038, + "step": 98470 + }, + { + "epoch": 15.29667598633116, + "grad_norm": 0.2760176360607147, + "learning_rate": 6.323527446782662e-08, + "loss": 0.0037, + "step": 98480 + }, + { + "epoch": 15.298229263746505, + "grad_norm": 0.7270868420600891, + "learning_rate": 6.240668425012142e-08, + "loss": 0.0027, + "step": 98490 + }, + { + "epoch": 15.299782541161852, + "grad_norm": 0.1988147348165512, + "learning_rate": 6.158355513254388e-08, + "loss": 0.0036, + "step": 98500 + }, + { + "epoch": 15.301335818577197, + "grad_norm": 0.442686527967453, + "learning_rate": 6.076588720510534e-08, + "loss": 0.0031, + "step": 98510 + }, + { + "epoch": 15.302889095992544, + "grad_norm": 0.20536349713802338, + "learning_rate": 5.995368055722316e-08, + "loss": 0.0026, + "step": 98520 + }, + { + "epoch": 15.304442373407891, + "grad_norm": 0.10684080421924591, + "learning_rate": 5.914693527773185e-08, + "loss": 0.0022, + "step": 98530 + }, + { + "epoch": 15.305995650823236, + "grad_norm": 0.1350811868906021, + "learning_rate": 5.834565145483861e-08, + "loss": 0.006, + "step": 98540 + }, + { + "epoch": 15.307548928238583, + "grad_norm": 0.17982898652553558, + "learning_rate": 5.754982917618446e-08, + "loss": 0.0034, + "step": 98550 + }, + { + "epoch": 15.30910220565393, + "grad_norm": 0.04150005057454109, + "learning_rate": 5.675946852878866e-08, + "loss": 0.0025, + "step": 98560 + }, + { + "epoch": 15.310655483069276, + "grad_norm": 0.294128954410553, + "learning_rate": 5.59745695990932e-08, + "loss": 0.0033, + "step": 98570 + }, + { + "epoch": 15.312208760484623, + "grad_norm": 0.2480766922235489, + "learning_rate": 5.51951324729294e-08, + "loss": 0.0045, + "step": 98580 + }, + { + "epoch": 15.313762037899968, + "grad_norm": 0.054700713604688644, + "learning_rate": 5.4421157235529093e-08, + "loss": 0.0052, + "step": 98590 + }, + { + "epoch": 15.315315315315315, + "grad_norm": 0.16500301659107208, + "learning_rate": 5.365264397154679e-08, + "loss": 0.0033, + "step": 98600 + }, + { + "epoch": 15.316868592730662, + "grad_norm": 0.06540022045373917, + "learning_rate": 5.288959276501526e-08, + "loss": 0.0035, + "step": 98610 + }, + { + "epoch": 15.318421870146008, + "grad_norm": 0.2321779727935791, + "learning_rate": 5.2132003699378876e-08, + "loss": 0.003, + "step": 98620 + }, + { + "epoch": 15.319975147561355, + "grad_norm": 0.22380056977272034, + "learning_rate": 5.137987685749912e-08, + "loss": 0.0032, + "step": 98630 + }, + { + "epoch": 15.321528424976702, + "grad_norm": 0.20303721725940704, + "learning_rate": 5.0633212321610227e-08, + "loss": 0.004, + "step": 98640 + }, + { + "epoch": 15.323081702392047, + "grad_norm": 1.0115448236465454, + "learning_rate": 4.989201017338019e-08, + "loss": 0.0037, + "step": 98650 + }, + { + "epoch": 15.324634979807394, + "grad_norm": 0.19519497454166412, + "learning_rate": 4.9156270493860846e-08, + "loss": 0.0035, + "step": 98660 + }, + { + "epoch": 15.32618825722274, + "grad_norm": 0.5532851815223694, + "learning_rate": 4.842599336351561e-08, + "loss": 0.0036, + "step": 98670 + }, + { + "epoch": 15.327741534638086, + "grad_norm": 0.6086158752441406, + "learning_rate": 4.7701178862197274e-08, + "loss": 0.0037, + "step": 98680 + }, + { + "epoch": 15.329294812053433, + "grad_norm": 0.10046672075986862, + "learning_rate": 4.698182706918131e-08, + "loss": 0.0031, + "step": 98690 + }, + { + "epoch": 15.330848089468779, + "grad_norm": 0.26580554246902466, + "learning_rate": 4.6267938063121466e-08, + "loss": 0.0034, + "step": 98700 + }, + { + "epoch": 15.332401366884126, + "grad_norm": 0.8375211954116821, + "learning_rate": 4.555951192209973e-08, + "loss": 0.0035, + "step": 98710 + }, + { + "epoch": 15.333954644299471, + "grad_norm": 0.4391174614429474, + "learning_rate": 4.485654872358747e-08, + "loss": 0.0048, + "step": 98720 + }, + { + "epoch": 15.335507921714818, + "grad_norm": 0.10958421230316162, + "learning_rate": 4.415904854445097e-08, + "loss": 0.0024, + "step": 98730 + }, + { + "epoch": 15.337061199130165, + "grad_norm": 0.16304869949817657, + "learning_rate": 4.34670114609792e-08, + "loss": 0.0049, + "step": 98740 + }, + { + "epoch": 15.33861447654551, + "grad_norm": 0.15290407836437225, + "learning_rate": 4.278043754884498e-08, + "loss": 0.0029, + "step": 98750 + }, + { + "epoch": 15.340167753960857, + "grad_norm": 0.26702407002449036, + "learning_rate": 4.209932688312712e-08, + "loss": 0.0034, + "step": 98760 + }, + { + "epoch": 15.341721031376204, + "grad_norm": 0.10887522250413895, + "learning_rate": 4.14236795383216e-08, + "loss": 0.0022, + "step": 98770 + }, + { + "epoch": 15.34327430879155, + "grad_norm": 0.10826335102319717, + "learning_rate": 4.075349558830821e-08, + "loss": 0.0029, + "step": 98780 + }, + { + "epoch": 15.344827586206897, + "grad_norm": 0.30901098251342773, + "learning_rate": 4.008877510638387e-08, + "loss": 0.0031, + "step": 98790 + }, + { + "epoch": 15.346380863622242, + "grad_norm": 0.12429387122392654, + "learning_rate": 3.942951816523488e-08, + "loss": 0.0023, + "step": 98800 + }, + { + "epoch": 15.34793414103759, + "grad_norm": 0.21226190030574799, + "learning_rate": 3.8775724836959125e-08, + "loss": 0.003, + "step": 98810 + }, + { + "epoch": 15.349487418452936, + "grad_norm": 0.18401092290878296, + "learning_rate": 3.812739519305497e-08, + "loss": 0.0027, + "step": 98820 + }, + { + "epoch": 15.351040695868281, + "grad_norm": 0.0644599050283432, + "learning_rate": 3.748452930442126e-08, + "loss": 0.0038, + "step": 98830 + }, + { + "epoch": 15.352593973283629, + "grad_norm": 0.08109205216169357, + "learning_rate": 3.6847127241362855e-08, + "loss": 0.0033, + "step": 98840 + }, + { + "epoch": 15.354147250698976, + "grad_norm": 1.4977718591690063, + "learning_rate": 3.621518907358512e-08, + "loss": 0.004, + "step": 98850 + }, + { + "epoch": 15.35570052811432, + "grad_norm": 0.07448923587799072, + "learning_rate": 3.558871487019388e-08, + "loss": 0.0033, + "step": 98860 + }, + { + "epoch": 15.357253805529668, + "grad_norm": 0.35278788208961487, + "learning_rate": 3.4967704699701007e-08, + "loss": 0.0029, + "step": 98870 + }, + { + "epoch": 15.358807082945013, + "grad_norm": 0.08527693152427673, + "learning_rate": 3.435215863001884e-08, + "loss": 0.0032, + "step": 98880 + }, + { + "epoch": 15.36036036036036, + "grad_norm": 0.09209651499986649, + "learning_rate": 3.374207672846019e-08, + "loss": 0.0023, + "step": 98890 + }, + { + "epoch": 15.361913637775707, + "grad_norm": 0.2400026172399521, + "learning_rate": 3.313745906174392e-08, + "loss": 0.0029, + "step": 98900 + }, + { + "epoch": 15.363466915191053, + "grad_norm": 0.4631681740283966, + "learning_rate": 3.253830569599492e-08, + "loss": 0.0039, + "step": 98910 + }, + { + "epoch": 15.3650201926064, + "grad_norm": 0.4247596561908722, + "learning_rate": 3.1944616696727436e-08, + "loss": 0.0021, + "step": 98920 + }, + { + "epoch": 15.366573470021747, + "grad_norm": 0.3993796408176422, + "learning_rate": 3.135639212887287e-08, + "loss": 0.0026, + "step": 98930 + }, + { + "epoch": 15.368126747437092, + "grad_norm": 0.161497563123703, + "learning_rate": 3.077363205675754e-08, + "loss": 0.0025, + "step": 98940 + }, + { + "epoch": 15.369680024852439, + "grad_norm": 0.16309788823127747, + "learning_rate": 3.0196336544113804e-08, + "loss": 0.0048, + "step": 98950 + }, + { + "epoch": 15.371233302267784, + "grad_norm": 0.11830620467662811, + "learning_rate": 2.9624505654063383e-08, + "loss": 0.0019, + "step": 98960 + }, + { + "epoch": 15.372786579683131, + "grad_norm": 0.07033143937587738, + "learning_rate": 2.90581394491507e-08, + "loss": 0.003, + "step": 98970 + }, + { + "epoch": 15.374339857098478, + "grad_norm": 0.10099221765995026, + "learning_rate": 2.8497237991309545e-08, + "loss": 0.0042, + "step": 98980 + }, + { + "epoch": 15.375893134513824, + "grad_norm": 0.6602701544761658, + "learning_rate": 2.7941801341879735e-08, + "loss": 0.0041, + "step": 98990 + }, + { + "epoch": 15.37744641192917, + "grad_norm": 0.10893212258815765, + "learning_rate": 2.7391829561601578e-08, + "loss": 0.0034, + "step": 99000 + }, + { + "epoch": 15.378999689344518, + "grad_norm": 0.14711543917655945, + "learning_rate": 2.6847322710621403e-08, + "loss": 0.0027, + "step": 99010 + }, + { + "epoch": 15.380552966759863, + "grad_norm": 0.14463192224502563, + "learning_rate": 2.6308280848486022e-08, + "loss": 0.0042, + "step": 99020 + }, + { + "epoch": 15.38210624417521, + "grad_norm": 0.1871763914823532, + "learning_rate": 2.5774704034137177e-08, + "loss": 0.004, + "step": 99030 + }, + { + "epoch": 15.383659521590555, + "grad_norm": 0.14801013469696045, + "learning_rate": 2.524659232593929e-08, + "loss": 0.0027, + "step": 99040 + }, + { + "epoch": 15.385212799005902, + "grad_norm": 0.18215829133987427, + "learning_rate": 2.472394578163506e-08, + "loss": 0.003, + "step": 99050 + }, + { + "epoch": 15.38676607642125, + "grad_norm": 0.24626514315605164, + "learning_rate": 2.4206764458378772e-08, + "loss": 0.0029, + "step": 99060 + }, + { + "epoch": 15.388319353836595, + "grad_norm": 0.2775862514972687, + "learning_rate": 2.3695048412736285e-08, + "loss": 0.003, + "step": 99070 + }, + { + "epoch": 15.389872631251942, + "grad_norm": 0.16794602572917938, + "learning_rate": 2.3188797700668395e-08, + "loss": 0.0031, + "step": 99080 + }, + { + "epoch": 15.391425908667287, + "grad_norm": 0.9636713266372681, + "learning_rate": 2.268801237753082e-08, + "loss": 0.0034, + "step": 99090 + }, + { + "epoch": 15.392979186082634, + "grad_norm": 0.15870968997478485, + "learning_rate": 2.2192692498090862e-08, + "loss": 0.0031, + "step": 99100 + }, + { + "epoch": 15.394532463497981, + "grad_norm": 0.8207459449768066, + "learning_rate": 2.170283811652185e-08, + "loss": 0.0037, + "step": 99110 + }, + { + "epoch": 15.396085740913326, + "grad_norm": 0.26210030913352966, + "learning_rate": 2.12184492863865e-08, + "loss": 0.0033, + "step": 99120 + }, + { + "epoch": 15.397639018328674, + "grad_norm": 0.2942637801170349, + "learning_rate": 2.073952606066465e-08, + "loss": 0.0058, + "step": 99130 + }, + { + "epoch": 15.39919229574402, + "grad_norm": 0.07263994961977005, + "learning_rate": 2.026606849171997e-08, + "loss": 0.0023, + "step": 99140 + }, + { + "epoch": 15.400745573159366, + "grad_norm": 0.8193067908287048, + "learning_rate": 1.9798076631333262e-08, + "loss": 0.0034, + "step": 99150 + }, + { + "epoch": 15.402298850574713, + "grad_norm": 0.3319263458251953, + "learning_rate": 1.933555053069136e-08, + "loss": 0.0024, + "step": 99160 + }, + { + "epoch": 15.403852127990058, + "grad_norm": 0.16660767793655396, + "learning_rate": 1.887849024036492e-08, + "loss": 0.0036, + "step": 99170 + }, + { + "epoch": 15.405405405405405, + "grad_norm": 0.47717463970184326, + "learning_rate": 1.8426895810341736e-08, + "loss": 0.0022, + "step": 99180 + }, + { + "epoch": 15.406958682820752, + "grad_norm": 0.10517079383134842, + "learning_rate": 1.798076729000453e-08, + "loss": 0.0046, + "step": 99190 + }, + { + "epoch": 15.408511960236098, + "grad_norm": 0.15385723114013672, + "learning_rate": 1.75401047281476e-08, + "loss": 0.0025, + "step": 99200 + }, + { + "epoch": 15.410065237651445, + "grad_norm": 0.530160129070282, + "learning_rate": 1.7104908172954626e-08, + "loss": 0.0027, + "step": 99210 + }, + { + "epoch": 15.411618515066792, + "grad_norm": 0.30080580711364746, + "learning_rate": 1.6675177672020868e-08, + "loss": 0.0025, + "step": 99220 + }, + { + "epoch": 15.413171792482137, + "grad_norm": 0.3904685080051422, + "learning_rate": 1.6250913272342073e-08, + "loss": 0.0035, + "step": 99230 + }, + { + "epoch": 15.414725069897484, + "grad_norm": 0.4439985752105713, + "learning_rate": 1.5832115020314453e-08, + "loss": 0.0039, + "step": 99240 + }, + { + "epoch": 15.41627834731283, + "grad_norm": 0.06566863507032394, + "learning_rate": 1.5418782961734712e-08, + "loss": 0.0027, + "step": 99250 + }, + { + "epoch": 15.417831624728176, + "grad_norm": 0.2970873713493347, + "learning_rate": 1.501091714181113e-08, + "loss": 0.0032, + "step": 99260 + }, + { + "epoch": 15.419384902143523, + "grad_norm": 0.1386633962392807, + "learning_rate": 1.460851760513582e-08, + "loss": 0.0037, + "step": 99270 + }, + { + "epoch": 15.420938179558869, + "grad_norm": 0.5899684429168701, + "learning_rate": 1.4211584395723566e-08, + "loss": 0.0036, + "step": 99280 + }, + { + "epoch": 15.422491456974216, + "grad_norm": 0.06888259202241898, + "learning_rate": 1.382011755697854e-08, + "loss": 0.0043, + "step": 99290 + }, + { + "epoch": 15.424044734389563, + "grad_norm": 0.48863187432289124, + "learning_rate": 1.3434117131716494e-08, + "loss": 0.003, + "step": 99300 + }, + { + "epoch": 15.425598011804908, + "grad_norm": 0.5912119746208191, + "learning_rate": 1.3053583162142557e-08, + "loss": 0.0036, + "step": 99310 + }, + { + "epoch": 15.427151289220255, + "grad_norm": 0.1375468671321869, + "learning_rate": 1.2678515689873438e-08, + "loss": 0.0035, + "step": 99320 + }, + { + "epoch": 15.4287045666356, + "grad_norm": 0.08071038126945496, + "learning_rate": 1.2308914755931877e-08, + "loss": 0.0028, + "step": 99330 + }, + { + "epoch": 15.430257844050947, + "grad_norm": 0.7371296286582947, + "learning_rate": 1.1944780400729993e-08, + "loss": 0.0023, + "step": 99340 + }, + { + "epoch": 15.431811121466295, + "grad_norm": 0.08913734555244446, + "learning_rate": 1.1586112664085935e-08, + "loss": 0.0028, + "step": 99350 + }, + { + "epoch": 15.43336439888164, + "grad_norm": 0.24460722506046295, + "learning_rate": 1.1232911585234984e-08, + "loss": 0.0023, + "step": 99360 + }, + { + "epoch": 15.434917676296987, + "grad_norm": 0.08320070058107376, + "learning_rate": 1.0885177202790697e-08, + "loss": 0.0031, + "step": 99370 + }, + { + "epoch": 15.436470953712334, + "grad_norm": 0.2583865821361542, + "learning_rate": 1.0542909554789315e-08, + "loss": 0.0028, + "step": 99380 + }, + { + "epoch": 15.43802423112768, + "grad_norm": 0.22386062145233154, + "learning_rate": 1.0206108678656456e-08, + "loss": 0.0042, + "step": 99390 + }, + { + "epoch": 15.439577508543026, + "grad_norm": 0.037135086953639984, + "learning_rate": 9.874774611223769e-09, + "loss": 0.0038, + "step": 99400 + }, + { + "epoch": 15.441130785958372, + "grad_norm": 0.5797503590583801, + "learning_rate": 9.548907388728933e-09, + "loss": 0.0038, + "step": 99410 + }, + { + "epoch": 15.442684063373719, + "grad_norm": 0.09225552529096603, + "learning_rate": 9.228507046804557e-09, + "loss": 0.0022, + "step": 99420 + }, + { + "epoch": 15.444237340789066, + "grad_norm": 0.1339481770992279, + "learning_rate": 8.91357362048928e-09, + "loss": 0.0021, + "step": 99430 + }, + { + "epoch": 15.445790618204411, + "grad_norm": 0.16410352289676666, + "learning_rate": 8.604107144227769e-09, + "loss": 0.003, + "step": 99440 + }, + { + "epoch": 15.447343895619758, + "grad_norm": 0.4019947648048401, + "learning_rate": 8.300107651859623e-09, + "loss": 0.0027, + "step": 99450 + }, + { + "epoch": 15.448897173035103, + "grad_norm": 0.594746470451355, + "learning_rate": 8.001575176630472e-09, + "loss": 0.0034, + "step": 99460 + }, + { + "epoch": 15.45045045045045, + "grad_norm": 0.21980583667755127, + "learning_rate": 7.708509751186422e-09, + "loss": 0.0047, + "step": 99470 + }, + { + "epoch": 15.452003727865797, + "grad_norm": 0.16439594328403473, + "learning_rate": 7.420911407579611e-09, + "loss": 0.0029, + "step": 99480 + }, + { + "epoch": 15.453557005281143, + "grad_norm": 0.1644158959388733, + "learning_rate": 7.13878017725711e-09, + "loss": 0.0026, + "step": 99490 + }, + { + "epoch": 15.45511028269649, + "grad_norm": 0.13466979563236237, + "learning_rate": 6.8621160910720125e-09, + "loss": 0.0049, + "step": 99500 + }, + { + "epoch": 15.456663560111837, + "grad_norm": 0.4493941366672516, + "learning_rate": 6.59091917928345e-09, + "loss": 0.0031, + "step": 99510 + }, + { + "epoch": 15.458216837527182, + "grad_norm": 0.6212389469146729, + "learning_rate": 6.32518947155103e-09, + "loss": 0.0022, + "step": 99520 + }, + { + "epoch": 15.459770114942529, + "grad_norm": 0.10276412963867188, + "learning_rate": 6.064926996929288e-09, + "loss": 0.0031, + "step": 99530 + }, + { + "epoch": 15.461323392357874, + "grad_norm": 0.04135412722826004, + "learning_rate": 5.810131783884343e-09, + "loss": 0.0033, + "step": 99540 + }, + { + "epoch": 15.462876669773221, + "grad_norm": 0.087117500603199, + "learning_rate": 5.5608038602772415e-09, + "loss": 0.0031, + "step": 99550 + }, + { + "epoch": 15.464429947188568, + "grad_norm": 0.3526526689529419, + "learning_rate": 5.316943253375062e-09, + "loss": 0.0033, + "step": 99560 + }, + { + "epoch": 15.465983224603914, + "grad_norm": 0.1428777426481247, + "learning_rate": 5.07854998984536e-09, + "loss": 0.0022, + "step": 99570 + }, + { + "epoch": 15.46753650201926, + "grad_norm": 0.1764719933271408, + "learning_rate": 4.845624095756174e-09, + "loss": 0.0026, + "step": 99580 + }, + { + "epoch": 15.469089779434608, + "grad_norm": 0.10776444524526596, + "learning_rate": 4.618165596587121e-09, + "loss": 0.0028, + "step": 99590 + }, + { + "epoch": 15.470643056849953, + "grad_norm": 0.033002227544784546, + "learning_rate": 4.396174517207197e-09, + "loss": 0.0015, + "step": 99600 + }, + { + "epoch": 15.4721963342653, + "grad_norm": 0.09652025997638702, + "learning_rate": 4.179650881896979e-09, + "loss": 0.003, + "step": 99610 + }, + { + "epoch": 15.473749611680645, + "grad_norm": 0.12125375121831894, + "learning_rate": 3.968594714331975e-09, + "loss": 0.0032, + "step": 99620 + }, + { + "epoch": 15.475302889095992, + "grad_norm": 0.6547350287437439, + "learning_rate": 3.763006037593719e-09, + "loss": 0.0046, + "step": 99630 + }, + { + "epoch": 15.47685616651134, + "grad_norm": 0.46861007809638977, + "learning_rate": 3.5628848741586786e-09, + "loss": 0.0047, + "step": 99640 + }, + { + "epoch": 15.478409443926685, + "grad_norm": 0.18540506064891815, + "learning_rate": 3.368231245926001e-09, + "loss": 0.004, + "step": 99650 + }, + { + "epoch": 15.479962721342032, + "grad_norm": 0.3448718786239624, + "learning_rate": 3.179045174167561e-09, + "loss": 0.004, + "step": 99660 + }, + { + "epoch": 15.481515998757377, + "grad_norm": 0.30897465348243713, + "learning_rate": 2.9953266795834657e-09, + "loss": 0.0026, + "step": 99670 + }, + { + "epoch": 15.483069276172724, + "grad_norm": 0.6754563450813293, + "learning_rate": 2.817075782263201e-09, + "loss": 0.004, + "step": 99680 + }, + { + "epoch": 15.484622553588071, + "grad_norm": 0.1420850306749344, + "learning_rate": 2.6442925016967325e-09, + "loss": 0.0035, + "step": 99690 + }, + { + "epoch": 15.486175831003417, + "grad_norm": 0.14670760929584503, + "learning_rate": 2.4769768567800555e-09, + "loss": 0.0045, + "step": 99700 + }, + { + "epoch": 15.487729108418764, + "grad_norm": 0.15838080644607544, + "learning_rate": 2.315128865809646e-09, + "loss": 0.0041, + "step": 99710 + }, + { + "epoch": 15.48928238583411, + "grad_norm": 0.28599244356155396, + "learning_rate": 2.1587485464880097e-09, + "loss": 0.0034, + "step": 99720 + }, + { + "epoch": 15.490835663249456, + "grad_norm": 0.3823913335800171, + "learning_rate": 2.0078359159125813e-09, + "loss": 0.0019, + "step": 99730 + }, + { + "epoch": 15.492388940664803, + "grad_norm": 0.062284767627716064, + "learning_rate": 1.8623909905923776e-09, + "loss": 0.0033, + "step": 99740 + }, + { + "epoch": 15.493942218080148, + "grad_norm": 0.07854613661766052, + "learning_rate": 1.7224137864257916e-09, + "loss": 0.002, + "step": 99750 + }, + { + "epoch": 15.495495495495495, + "grad_norm": 0.18500614166259766, + "learning_rate": 1.5879043187283505e-09, + "loss": 0.003, + "step": 99760 + }, + { + "epoch": 15.497048772910842, + "grad_norm": 0.049284446984529495, + "learning_rate": 1.458862602204958e-09, + "loss": 0.0018, + "step": 99770 + }, + { + "epoch": 15.498602050326188, + "grad_norm": 0.20105713605880737, + "learning_rate": 1.3352886509720996e-09, + "loss": 0.0025, + "step": 99780 + }, + { + "epoch": 15.500155327741535, + "grad_norm": 0.14256280660629272, + "learning_rate": 1.2171824785356389e-09, + "loss": 0.0016, + "step": 99790 + }, + { + "epoch": 15.501708605156882, + "grad_norm": 0.09739773720502853, + "learning_rate": 1.1045440978185718e-09, + "loss": 0.0026, + "step": 99800 + }, + { + "epoch": 15.503261882572227, + "grad_norm": 0.5599939227104187, + "learning_rate": 9.973735211332714e-10, + "loss": 0.0034, + "step": 99810 + }, + { + "epoch": 15.504815159987574, + "grad_norm": 0.29854264855384827, + "learning_rate": 8.956707602036929e-10, + "loss": 0.0022, + "step": 99820 + }, + { + "epoch": 15.50636843740292, + "grad_norm": 0.2121099978685379, + "learning_rate": 7.994358261542712e-10, + "loss": 0.0027, + "step": 99830 + }, + { + "epoch": 15.507921714818266, + "grad_norm": 0.4298579692840576, + "learning_rate": 7.086687295043692e-10, + "loss": 0.0029, + "step": 99840 + }, + { + "epoch": 15.509474992233613, + "grad_norm": 0.11274847388267517, + "learning_rate": 6.23369480179381e-10, + "loss": 0.0024, + "step": 99850 + }, + { + "epoch": 15.511028269648959, + "grad_norm": 0.7724071741104126, + "learning_rate": 5.43538087510731e-10, + "loss": 0.0023, + "step": 99860 + }, + { + "epoch": 15.512581547064306, + "grad_norm": 0.5261003375053406, + "learning_rate": 4.691745602303233e-10, + "loss": 0.0024, + "step": 99870 + }, + { + "epoch": 15.514134824479653, + "grad_norm": 0.34441325068473816, + "learning_rate": 4.0027890646499034e-10, + "loss": 0.0024, + "step": 99880 + }, + { + "epoch": 15.515688101894998, + "grad_norm": 0.17318026721477509, + "learning_rate": 3.368511337531466e-10, + "loss": 0.0028, + "step": 99890 + }, + { + "epoch": 15.517241379310345, + "grad_norm": 0.549475371837616, + "learning_rate": 2.788912490281348e-10, + "loss": 0.0035, + "step": 99900 + }, + { + "epoch": 15.51879465672569, + "grad_norm": 0.08769059926271439, + "learning_rate": 2.2639925863487954e-10, + "loss": 0.0053, + "step": 99910 + }, + { + "epoch": 15.520347934141038, + "grad_norm": 0.03230125084519386, + "learning_rate": 1.7937516830213163e-10, + "loss": 0.0028, + "step": 99920 + }, + { + "epoch": 15.521901211556385, + "grad_norm": 0.20200347900390625, + "learning_rate": 1.3781898318687702e-10, + "loss": 0.0024, + "step": 99930 + }, + { + "epoch": 15.52345448897173, + "grad_norm": 0.42756667733192444, + "learning_rate": 1.0173070781882566e-10, + "loss": 0.0022, + "step": 99940 + }, + { + "epoch": 15.525007766387077, + "grad_norm": 0.12815611064434052, + "learning_rate": 7.111034615592261e-11, + "loss": 0.0036, + "step": 99950 + }, + { + "epoch": 15.526561043802424, + "grad_norm": 0.5566781759262085, + "learning_rate": 4.595790153993918e-11, + "loss": 0.0035, + "step": 99960 + }, + { + "epoch": 15.52811432121777, + "grad_norm": 0.19070160388946533, + "learning_rate": 2.6273376729779585e-11, + "loss": 0.0027, + "step": 99970 + }, + { + "epoch": 15.529667598633116, + "grad_norm": 0.09919342398643494, + "learning_rate": 1.2056773868174276e-11, + "loss": 0.0025, + "step": 99980 + }, + { + "epoch": 15.531220876048462, + "grad_norm": 0.5218271613121033, + "learning_rate": 3.3080945149865926e-12, + "loss": 0.0036, + "step": 99990 + }, + { + "epoch": 15.532774153463809, + "grad_norm": 0.3268229365348816, + "learning_rate": 2.7339625008337977e-14, + "loss": 0.0028, + "step": 100000 + } + ], + "logging_steps": 10, + "max_steps": 100000, + "num_input_tokens_seen": 0, + "num_train_epochs": 16, + "save_steps": 25000, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": true + }, + "attributes": {} + } + }, + "total_flos": 0.0, + "train_batch_size": 32, + "trial_name": null, + "trial_params": null +}