diff --git "a/trainer_state.json" "b/trainer_state.json" new file mode 100644--- /dev/null +++ "b/trainer_state.json" @@ -0,0 +1,15112 @@ +{ + "best_global_step": 12882, + "best_metric": 0.49088114500045776, + "best_model_checkpoint": "./results_shuffled_deep128_heads_r16_seq128_3e-5_roberta_newData/checkpoint-12882", + "epoch": 5.0, + "eval_steps": 500, + "global_step": 21470, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.002328830926874709, + "grad_norm": 7.533487796783447, + "learning_rate": 2.99874213836478e-05, + "loss": 3.2426, + "step": 10 + }, + { + "epoch": 0.004657661853749418, + "grad_norm": 5.052584171295166, + "learning_rate": 2.9973445143256466e-05, + "loss": 3.3433, + "step": 20 + }, + { + "epoch": 0.0069864927806241265, + "grad_norm": 3.4427478313446045, + "learning_rate": 2.995946890286513e-05, + "loss": 3.0442, + "step": 30 + }, + { + "epoch": 0.009315323707498836, + "grad_norm": 4.24840784072876, + "learning_rate": 2.9945492662473797e-05, + "loss": 3.1555, + "step": 40 + }, + { + "epoch": 0.011644154634373545, + "grad_norm": 4.173481464385986, + "learning_rate": 2.9931516422082464e-05, + "loss": 2.3163, + "step": 50 + }, + { + "epoch": 0.013972985561248253, + "grad_norm": 2.706721544265747, + "learning_rate": 2.9917540181691124e-05, + "loss": 2.9514, + "step": 60 + }, + { + "epoch": 0.016301816488122962, + "grad_norm": 3.585742712020874, + "learning_rate": 2.990356394129979e-05, + "loss": 2.84, + "step": 70 + }, + { + "epoch": 0.018630647414997672, + "grad_norm": 7.899261951446533, + "learning_rate": 2.9889587700908455e-05, + "loss": 3.0852, + "step": 80 + }, + { + "epoch": 0.02095947834187238, + "grad_norm": 3.6982789039611816, + "learning_rate": 2.987561146051712e-05, + "loss": 2.4326, + "step": 90 + }, + { + "epoch": 0.02328830926874709, + "grad_norm": 3.14382004737854, + "learning_rate": 2.9861635220125785e-05, + "loss": 2.6978, + "step": 100 + }, + { + "epoch": 0.025617140195621797, + "grad_norm": 6.083728313446045, + "learning_rate": 2.9847658979734452e-05, + "loss": 2.4978, + "step": 110 + }, + { + "epoch": 0.027945971122496506, + "grad_norm": 4.235475540161133, + "learning_rate": 2.9833682739343116e-05, + "loss": 2.6458, + "step": 120 + }, + { + "epoch": 0.030274802049371215, + "grad_norm": 6.495112895965576, + "learning_rate": 2.9819706498951783e-05, + "loss": 2.6736, + "step": 130 + }, + { + "epoch": 0.032603632976245925, + "grad_norm": 3.866530418395996, + "learning_rate": 2.9805730258560446e-05, + "loss": 2.355, + "step": 140 + }, + { + "epoch": 0.03493246390312063, + "grad_norm": 8.10139274597168, + "learning_rate": 2.9791754018169113e-05, + "loss": 2.7388, + "step": 150 + }, + { + "epoch": 0.037261294829995344, + "grad_norm": 9.420899391174316, + "learning_rate": 2.9777777777777777e-05, + "loss": 2.6467, + "step": 160 + }, + { + "epoch": 0.03959012575687005, + "grad_norm": 10.79451847076416, + "learning_rate": 2.9763801537386444e-05, + "loss": 2.8556, + "step": 170 + }, + { + "epoch": 0.04191895668374476, + "grad_norm": 4.488949775695801, + "learning_rate": 2.9749825296995107e-05, + "loss": 2.5636, + "step": 180 + }, + { + "epoch": 0.04424778761061947, + "grad_norm": 13.844695091247559, + "learning_rate": 2.9735849056603774e-05, + "loss": 2.3088, + "step": 190 + }, + { + "epoch": 0.04657661853749418, + "grad_norm": 5.569024562835693, + "learning_rate": 2.972187281621244e-05, + "loss": 2.7515, + "step": 200 + }, + { + "epoch": 0.04890544946436889, + "grad_norm": 3.9889473915100098, + "learning_rate": 2.9707896575821105e-05, + "loss": 2.0546, + "step": 210 + }, + { + "epoch": 0.05123428039124359, + "grad_norm": 7.793086051940918, + "learning_rate": 2.9693920335429772e-05, + "loss": 2.5026, + "step": 220 + }, + { + "epoch": 0.053563111318118306, + "grad_norm": 10.837105751037598, + "learning_rate": 2.9679944095038436e-05, + "loss": 2.7304, + "step": 230 + }, + { + "epoch": 0.05589194224499301, + "grad_norm": 5.149703025817871, + "learning_rate": 2.9665967854647103e-05, + "loss": 2.4211, + "step": 240 + }, + { + "epoch": 0.058220773171867725, + "grad_norm": 9.660852432250977, + "learning_rate": 2.9651991614255766e-05, + "loss": 2.2913, + "step": 250 + }, + { + "epoch": 0.06054960409874243, + "grad_norm": 13.585944175720215, + "learning_rate": 2.9638015373864433e-05, + "loss": 2.0449, + "step": 260 + }, + { + "epoch": 0.06287843502561714, + "grad_norm": 6.299298286437988, + "learning_rate": 2.9624039133473097e-05, + "loss": 2.2918, + "step": 270 + }, + { + "epoch": 0.06520726595249185, + "grad_norm": 3.201357364654541, + "learning_rate": 2.9610062893081764e-05, + "loss": 2.2325, + "step": 280 + }, + { + "epoch": 0.06753609687936656, + "grad_norm": 4.173748016357422, + "learning_rate": 2.9596086652690427e-05, + "loss": 2.3109, + "step": 290 + }, + { + "epoch": 0.06986492780624126, + "grad_norm": 5.973798751831055, + "learning_rate": 2.9582110412299094e-05, + "loss": 2.2817, + "step": 300 + }, + { + "epoch": 0.07219375873311598, + "grad_norm": 3.6076810359954834, + "learning_rate": 2.9568134171907758e-05, + "loss": 2.3076, + "step": 310 + }, + { + "epoch": 0.07452258965999069, + "grad_norm": 5.779540538787842, + "learning_rate": 2.955415793151642e-05, + "loss": 2.3035, + "step": 320 + }, + { + "epoch": 0.0768514205868654, + "grad_norm": 27.188243865966797, + "learning_rate": 2.9540181691125085e-05, + "loss": 2.5776, + "step": 330 + }, + { + "epoch": 0.0791802515137401, + "grad_norm": 8.513686180114746, + "learning_rate": 2.9526205450733752e-05, + "loss": 2.4484, + "step": 340 + }, + { + "epoch": 0.0815090824406148, + "grad_norm": 5.144784927368164, + "learning_rate": 2.951222921034242e-05, + "loss": 1.9469, + "step": 350 + }, + { + "epoch": 0.08383791336748952, + "grad_norm": 10.846701622009277, + "learning_rate": 2.9498252969951083e-05, + "loss": 2.301, + "step": 360 + }, + { + "epoch": 0.08616674429436423, + "grad_norm": 3.0678811073303223, + "learning_rate": 2.948427672955975e-05, + "loss": 2.4328, + "step": 370 + }, + { + "epoch": 0.08849557522123894, + "grad_norm": 4.1828413009643555, + "learning_rate": 2.9470300489168413e-05, + "loss": 1.8278, + "step": 380 + }, + { + "epoch": 0.09082440614811364, + "grad_norm": 7.489015579223633, + "learning_rate": 2.945632424877708e-05, + "loss": 1.7735, + "step": 390 + }, + { + "epoch": 0.09315323707498836, + "grad_norm": 8.380903244018555, + "learning_rate": 2.9442348008385744e-05, + "loss": 2.6727, + "step": 400 + }, + { + "epoch": 0.09548206800186307, + "grad_norm": 10.518119812011719, + "learning_rate": 2.942837176799441e-05, + "loss": 2.2046, + "step": 410 + }, + { + "epoch": 0.09781089892873777, + "grad_norm": 18.78498077392578, + "learning_rate": 2.9414395527603074e-05, + "loss": 2.3064, + "step": 420 + }, + { + "epoch": 0.10013972985561248, + "grad_norm": 4.508672714233398, + "learning_rate": 2.940041928721174e-05, + "loss": 2.1385, + "step": 430 + }, + { + "epoch": 0.10246856078248719, + "grad_norm": 12.739822387695312, + "learning_rate": 2.9386443046820405e-05, + "loss": 2.3654, + "step": 440 + }, + { + "epoch": 0.1047973917093619, + "grad_norm": 9.474522590637207, + "learning_rate": 2.9372466806429072e-05, + "loss": 1.8915, + "step": 450 + }, + { + "epoch": 0.10712622263623661, + "grad_norm": 8.7893705368042, + "learning_rate": 2.9358490566037735e-05, + "loss": 2.0838, + "step": 460 + }, + { + "epoch": 0.10945505356311132, + "grad_norm": 8.382100105285645, + "learning_rate": 2.9344514325646402e-05, + "loss": 2.2598, + "step": 470 + }, + { + "epoch": 0.11178388448998602, + "grad_norm": 10.099526405334473, + "learning_rate": 2.9330538085255066e-05, + "loss": 1.815, + "step": 480 + }, + { + "epoch": 0.11411271541686073, + "grad_norm": 7.139161109924316, + "learning_rate": 2.9316561844863733e-05, + "loss": 1.8358, + "step": 490 + }, + { + "epoch": 0.11644154634373545, + "grad_norm": 6.610865116119385, + "learning_rate": 2.93025856044724e-05, + "loss": 2.021, + "step": 500 + }, + { + "epoch": 0.11877037727061016, + "grad_norm": 8.798666000366211, + "learning_rate": 2.9288609364081064e-05, + "loss": 2.2395, + "step": 510 + }, + { + "epoch": 0.12109920819748486, + "grad_norm": 5.463034152984619, + "learning_rate": 2.927463312368973e-05, + "loss": 2.0574, + "step": 520 + }, + { + "epoch": 0.12342803912435957, + "grad_norm": 9.88456916809082, + "learning_rate": 2.9260656883298394e-05, + "loss": 2.0195, + "step": 530 + }, + { + "epoch": 0.1257568700512343, + "grad_norm": 6.4875078201293945, + "learning_rate": 2.924668064290706e-05, + "loss": 2.0221, + "step": 540 + }, + { + "epoch": 0.12808570097810898, + "grad_norm": 7.904644966125488, + "learning_rate": 2.9232704402515725e-05, + "loss": 1.9474, + "step": 550 + }, + { + "epoch": 0.1304145319049837, + "grad_norm": 6.517539024353027, + "learning_rate": 2.921872816212439e-05, + "loss": 1.9497, + "step": 560 + }, + { + "epoch": 0.13274336283185842, + "grad_norm": 13.083559036254883, + "learning_rate": 2.9204751921733055e-05, + "loss": 2.3732, + "step": 570 + }, + { + "epoch": 0.1350721937587331, + "grad_norm": 14.06116008758545, + "learning_rate": 2.9190775681341722e-05, + "loss": 2.0368, + "step": 580 + }, + { + "epoch": 0.13740102468560783, + "grad_norm": 6.1035308837890625, + "learning_rate": 2.9176799440950382e-05, + "loss": 1.7297, + "step": 590 + }, + { + "epoch": 0.13972985561248252, + "grad_norm": 22.444040298461914, + "learning_rate": 2.916282320055905e-05, + "loss": 1.9281, + "step": 600 + }, + { + "epoch": 0.14205868653935724, + "grad_norm": 8.657206535339355, + "learning_rate": 2.9148846960167713e-05, + "loss": 2.1254, + "step": 610 + }, + { + "epoch": 0.14438751746623196, + "grad_norm": 15.154228210449219, + "learning_rate": 2.913487071977638e-05, + "loss": 1.7197, + "step": 620 + }, + { + "epoch": 0.14671634839310665, + "grad_norm": 19.55249786376953, + "learning_rate": 2.9120894479385044e-05, + "loss": 2.0123, + "step": 630 + }, + { + "epoch": 0.14904517931998137, + "grad_norm": 13.391355514526367, + "learning_rate": 2.910691823899371e-05, + "loss": 1.8662, + "step": 640 + }, + { + "epoch": 0.15137401024685607, + "grad_norm": 11.850015640258789, + "learning_rate": 2.9092941998602377e-05, + "loss": 1.4906, + "step": 650 + }, + { + "epoch": 0.1537028411737308, + "grad_norm": 7.1012043952941895, + "learning_rate": 2.907896575821104e-05, + "loss": 2.0575, + "step": 660 + }, + { + "epoch": 0.1560316721006055, + "grad_norm": 6.259739875793457, + "learning_rate": 2.9064989517819708e-05, + "loss": 2.1558, + "step": 670 + }, + { + "epoch": 0.1583605030274802, + "grad_norm": 13.080730438232422, + "learning_rate": 2.905101327742837e-05, + "loss": 2.1056, + "step": 680 + }, + { + "epoch": 0.16068933395435492, + "grad_norm": 7.734439373016357, + "learning_rate": 2.903703703703704e-05, + "loss": 2.0659, + "step": 690 + }, + { + "epoch": 0.1630181648812296, + "grad_norm": 9.50960636138916, + "learning_rate": 2.9023060796645702e-05, + "loss": 2.1, + "step": 700 + }, + { + "epoch": 0.16534699580810433, + "grad_norm": 18.45827293395996, + "learning_rate": 2.900908455625437e-05, + "loss": 2.1444, + "step": 710 + }, + { + "epoch": 0.16767582673497905, + "grad_norm": 10.426290512084961, + "learning_rate": 2.8995108315863033e-05, + "loss": 2.2937, + "step": 720 + }, + { + "epoch": 0.17000465766185374, + "grad_norm": 15.96475601196289, + "learning_rate": 2.89811320754717e-05, + "loss": 2.0853, + "step": 730 + }, + { + "epoch": 0.17233348858872846, + "grad_norm": 7.1626410484313965, + "learning_rate": 2.8967155835080363e-05, + "loss": 2.1589, + "step": 740 + }, + { + "epoch": 0.17466231951560315, + "grad_norm": 16.46662139892578, + "learning_rate": 2.895317959468903e-05, + "loss": 1.5098, + "step": 750 + }, + { + "epoch": 0.17699115044247787, + "grad_norm": 12.333754539489746, + "learning_rate": 2.8939203354297694e-05, + "loss": 2.0206, + "step": 760 + }, + { + "epoch": 0.1793199813693526, + "grad_norm": 17.24032974243164, + "learning_rate": 2.892522711390636e-05, + "loss": 2.3082, + "step": 770 + }, + { + "epoch": 0.18164881229622729, + "grad_norm": 13.411188125610352, + "learning_rate": 2.8911250873515024e-05, + "loss": 1.6792, + "step": 780 + }, + { + "epoch": 0.183977643223102, + "grad_norm": 7.877953052520752, + "learning_rate": 2.889727463312369e-05, + "loss": 2.1541, + "step": 790 + }, + { + "epoch": 0.18630647414997673, + "grad_norm": 24.286130905151367, + "learning_rate": 2.888329839273236e-05, + "loss": 2.1689, + "step": 800 + }, + { + "epoch": 0.18863530507685142, + "grad_norm": 10.445568084716797, + "learning_rate": 2.8869322152341022e-05, + "loss": 1.8713, + "step": 810 + }, + { + "epoch": 0.19096413600372614, + "grad_norm": 6.439265251159668, + "learning_rate": 2.885534591194969e-05, + "loss": 1.9121, + "step": 820 + }, + { + "epoch": 0.19329296693060083, + "grad_norm": 24.311752319335938, + "learning_rate": 2.8841369671558353e-05, + "loss": 2.0249, + "step": 830 + }, + { + "epoch": 0.19562179785747555, + "grad_norm": 24.98615264892578, + "learning_rate": 2.882739343116702e-05, + "loss": 2.0608, + "step": 840 + }, + { + "epoch": 0.19795062878435027, + "grad_norm": 11.190683364868164, + "learning_rate": 2.881341719077568e-05, + "loss": 1.7668, + "step": 850 + }, + { + "epoch": 0.20027945971122496, + "grad_norm": 10.343644142150879, + "learning_rate": 2.8799440950384347e-05, + "loss": 1.8736, + "step": 860 + }, + { + "epoch": 0.20260829063809968, + "grad_norm": 4.8723530769348145, + "learning_rate": 2.878546470999301e-05, + "loss": 1.3761, + "step": 870 + }, + { + "epoch": 0.20493712156497437, + "grad_norm": 13.744318962097168, + "learning_rate": 2.8771488469601677e-05, + "loss": 1.487, + "step": 880 + }, + { + "epoch": 0.2072659524918491, + "grad_norm": 8.8036470413208, + "learning_rate": 2.875751222921034e-05, + "loss": 1.5511, + "step": 890 + }, + { + "epoch": 0.2095947834187238, + "grad_norm": 16.776226043701172, + "learning_rate": 2.8743535988819008e-05, + "loss": 1.6082, + "step": 900 + }, + { + "epoch": 0.2119236143455985, + "grad_norm": 15.255518913269043, + "learning_rate": 2.872955974842767e-05, + "loss": 1.6329, + "step": 910 + }, + { + "epoch": 0.21425244527247322, + "grad_norm": 16.834640502929688, + "learning_rate": 2.871558350803634e-05, + "loss": 1.8672, + "step": 920 + }, + { + "epoch": 0.21658127619934792, + "grad_norm": 10.923068046569824, + "learning_rate": 2.8701607267645002e-05, + "loss": 1.6011, + "step": 930 + }, + { + "epoch": 0.21891010712622264, + "grad_norm": 4.949512958526611, + "learning_rate": 2.868763102725367e-05, + "loss": 1.8381, + "step": 940 + }, + { + "epoch": 0.22123893805309736, + "grad_norm": 14.78408145904541, + "learning_rate": 2.8673654786862336e-05, + "loss": 1.7178, + "step": 950 + }, + { + "epoch": 0.22356776897997205, + "grad_norm": 10.509323120117188, + "learning_rate": 2.8659678546471e-05, + "loss": 1.5696, + "step": 960 + }, + { + "epoch": 0.22589659990684677, + "grad_norm": 13.359964370727539, + "learning_rate": 2.8645702306079667e-05, + "loss": 2.3818, + "step": 970 + }, + { + "epoch": 0.22822543083372146, + "grad_norm": 7.7604660987854, + "learning_rate": 2.863172606568833e-05, + "loss": 1.9007, + "step": 980 + }, + { + "epoch": 0.23055426176059618, + "grad_norm": 25.51626205444336, + "learning_rate": 2.8617749825296997e-05, + "loss": 1.6906, + "step": 990 + }, + { + "epoch": 0.2328830926874709, + "grad_norm": 53.45542907714844, + "learning_rate": 2.860377358490566e-05, + "loss": 2.3295, + "step": 1000 + }, + { + "epoch": 0.2352119236143456, + "grad_norm": 6.25486421585083, + "learning_rate": 2.8589797344514328e-05, + "loss": 1.3573, + "step": 1010 + }, + { + "epoch": 0.2375407545412203, + "grad_norm": 16.020435333251953, + "learning_rate": 2.857582110412299e-05, + "loss": 1.5215, + "step": 1020 + }, + { + "epoch": 0.239869585468095, + "grad_norm": 19.388914108276367, + "learning_rate": 2.8561844863731658e-05, + "loss": 2.046, + "step": 1030 + }, + { + "epoch": 0.24219841639496972, + "grad_norm": 12.028767585754395, + "learning_rate": 2.8547868623340322e-05, + "loss": 2.0784, + "step": 1040 + }, + { + "epoch": 0.24452724732184444, + "grad_norm": 20.426069259643555, + "learning_rate": 2.853389238294899e-05, + "loss": 1.8761, + "step": 1050 + }, + { + "epoch": 0.24685607824871914, + "grad_norm": 8.661053657531738, + "learning_rate": 2.8519916142557652e-05, + "loss": 1.4791, + "step": 1060 + }, + { + "epoch": 0.24918490917559386, + "grad_norm": 5.779150009155273, + "learning_rate": 2.850593990216632e-05, + "loss": 1.741, + "step": 1070 + }, + { + "epoch": 0.2515137401024686, + "grad_norm": 8.086243629455566, + "learning_rate": 2.8491963661774983e-05, + "loss": 1.5324, + "step": 1080 + }, + { + "epoch": 0.25384257102934327, + "grad_norm": 15.885953903198242, + "learning_rate": 2.847798742138365e-05, + "loss": 1.5208, + "step": 1090 + }, + { + "epoch": 0.25617140195621796, + "grad_norm": 15.936205863952637, + "learning_rate": 2.8464011180992317e-05, + "loss": 2.0363, + "step": 1100 + }, + { + "epoch": 0.2585002328830927, + "grad_norm": 17.394638061523438, + "learning_rate": 2.8450034940600977e-05, + "loss": 2.1736, + "step": 1110 + }, + { + "epoch": 0.2608290638099674, + "grad_norm": 32.21560287475586, + "learning_rate": 2.8436058700209644e-05, + "loss": 1.9567, + "step": 1120 + }, + { + "epoch": 0.2631578947368421, + "grad_norm": 21.898767471313477, + "learning_rate": 2.8422082459818308e-05, + "loss": 2.0504, + "step": 1130 + }, + { + "epoch": 0.26548672566371684, + "grad_norm": 18.948341369628906, + "learning_rate": 2.8408106219426975e-05, + "loss": 1.8841, + "step": 1140 + }, + { + "epoch": 0.26781555659059153, + "grad_norm": 14.407012939453125, + "learning_rate": 2.8394129979035638e-05, + "loss": 1.7384, + "step": 1150 + }, + { + "epoch": 0.2701443875174662, + "grad_norm": 7.401731967926025, + "learning_rate": 2.8380153738644305e-05, + "loss": 1.2154, + "step": 1160 + }, + { + "epoch": 0.2724732184443409, + "grad_norm": 10.56177806854248, + "learning_rate": 2.836617749825297e-05, + "loss": 1.4734, + "step": 1170 + }, + { + "epoch": 0.27480204937121566, + "grad_norm": 10.783820152282715, + "learning_rate": 2.8352201257861636e-05, + "loss": 1.4753, + "step": 1180 + }, + { + "epoch": 0.27713088029809035, + "grad_norm": 27.66096305847168, + "learning_rate": 2.83382250174703e-05, + "loss": 1.4075, + "step": 1190 + }, + { + "epoch": 0.27945971122496505, + "grad_norm": 39.09516143798828, + "learning_rate": 2.8324248777078966e-05, + "loss": 2.2883, + "step": 1200 + }, + { + "epoch": 0.2817885421518398, + "grad_norm": 20.759464263916016, + "learning_rate": 2.831027253668763e-05, + "loss": 2.2767, + "step": 1210 + }, + { + "epoch": 0.2841173730787145, + "grad_norm": 14.715803146362305, + "learning_rate": 2.8296296296296297e-05, + "loss": 1.7704, + "step": 1220 + }, + { + "epoch": 0.2864462040055892, + "grad_norm": 9.463580131530762, + "learning_rate": 2.828232005590496e-05, + "loss": 1.8876, + "step": 1230 + }, + { + "epoch": 0.2887750349324639, + "grad_norm": 20.55774688720703, + "learning_rate": 2.8268343815513627e-05, + "loss": 1.7778, + "step": 1240 + }, + { + "epoch": 0.2911038658593386, + "grad_norm": 4.738845348358154, + "learning_rate": 2.8254367575122294e-05, + "loss": 1.3079, + "step": 1250 + }, + { + "epoch": 0.2934326967862133, + "grad_norm": 33.04350280761719, + "learning_rate": 2.8240391334730958e-05, + "loss": 1.8545, + "step": 1260 + }, + { + "epoch": 0.29576152771308806, + "grad_norm": 10.747074127197266, + "learning_rate": 2.8226415094339625e-05, + "loss": 1.5794, + "step": 1270 + }, + { + "epoch": 0.29809035863996275, + "grad_norm": 20.86418914794922, + "learning_rate": 2.821243885394829e-05, + "loss": 1.7533, + "step": 1280 + }, + { + "epoch": 0.30041918956683744, + "grad_norm": 17.990188598632812, + "learning_rate": 2.8198462613556956e-05, + "loss": 1.7932, + "step": 1290 + }, + { + "epoch": 0.30274802049371213, + "grad_norm": 15.999037742614746, + "learning_rate": 2.818448637316562e-05, + "loss": 1.6424, + "step": 1300 + }, + { + "epoch": 0.3050768514205869, + "grad_norm": 10.242752075195312, + "learning_rate": 2.8170510132774286e-05, + "loss": 1.5599, + "step": 1310 + }, + { + "epoch": 0.3074056823474616, + "grad_norm": 56.81425857543945, + "learning_rate": 2.815653389238295e-05, + "loss": 1.7664, + "step": 1320 + }, + { + "epoch": 0.30973451327433627, + "grad_norm": 5.3695502281188965, + "learning_rate": 2.8142557651991617e-05, + "loss": 1.6969, + "step": 1330 + }, + { + "epoch": 0.312063344201211, + "grad_norm": 20.256746292114258, + "learning_rate": 2.812858141160028e-05, + "loss": 1.8104, + "step": 1340 + }, + { + "epoch": 0.3143921751280857, + "grad_norm": 14.850010871887207, + "learning_rate": 2.8114605171208947e-05, + "loss": 1.5393, + "step": 1350 + }, + { + "epoch": 0.3167210060549604, + "grad_norm": 29.18364715576172, + "learning_rate": 2.810062893081761e-05, + "loss": 2.1662, + "step": 1360 + }, + { + "epoch": 0.31904983698183514, + "grad_norm": 10.589908599853516, + "learning_rate": 2.8086652690426278e-05, + "loss": 1.3533, + "step": 1370 + }, + { + "epoch": 0.32137866790870984, + "grad_norm": 21.720638275146484, + "learning_rate": 2.8072676450034938e-05, + "loss": 1.7474, + "step": 1380 + }, + { + "epoch": 0.32370749883558453, + "grad_norm": 21.47283172607422, + "learning_rate": 2.8058700209643605e-05, + "loss": 1.7673, + "step": 1390 + }, + { + "epoch": 0.3260363297624592, + "grad_norm": 6.568681716918945, + "learning_rate": 2.8044723969252272e-05, + "loss": 1.4598, + "step": 1400 + }, + { + "epoch": 0.32836516068933397, + "grad_norm": 19.102869033813477, + "learning_rate": 2.8030747728860936e-05, + "loss": 1.6002, + "step": 1410 + }, + { + "epoch": 0.33069399161620866, + "grad_norm": 18.268447875976562, + "learning_rate": 2.8016771488469603e-05, + "loss": 1.3573, + "step": 1420 + }, + { + "epoch": 0.33302282254308335, + "grad_norm": 17.070833206176758, + "learning_rate": 2.8002795248078266e-05, + "loss": 2.402, + "step": 1430 + }, + { + "epoch": 0.3353516534699581, + "grad_norm": 9.851959228515625, + "learning_rate": 2.7988819007686933e-05, + "loss": 1.481, + "step": 1440 + }, + { + "epoch": 0.3376804843968328, + "grad_norm": 27.05587387084961, + "learning_rate": 2.7974842767295597e-05, + "loss": 1.9486, + "step": 1450 + }, + { + "epoch": 0.3400093153237075, + "grad_norm": 17.53416633605957, + "learning_rate": 2.7960866526904264e-05, + "loss": 1.6455, + "step": 1460 + }, + { + "epoch": 0.34233814625058223, + "grad_norm": 23.41783332824707, + "learning_rate": 2.7946890286512927e-05, + "loss": 1.5375, + "step": 1470 + }, + { + "epoch": 0.3446669771774569, + "grad_norm": 8.158493995666504, + "learning_rate": 2.7932914046121594e-05, + "loss": 1.3204, + "step": 1480 + }, + { + "epoch": 0.3469958081043316, + "grad_norm": 18.967121124267578, + "learning_rate": 2.7918937805730258e-05, + "loss": 1.7482, + "step": 1490 + }, + { + "epoch": 0.3493246390312063, + "grad_norm": 15.774516105651855, + "learning_rate": 2.7904961565338925e-05, + "loss": 1.6003, + "step": 1500 + }, + { + "epoch": 0.35165346995808106, + "grad_norm": 10.410409927368164, + "learning_rate": 2.789098532494759e-05, + "loss": 1.6872, + "step": 1510 + }, + { + "epoch": 0.35398230088495575, + "grad_norm": 12.958462715148926, + "learning_rate": 2.7877009084556255e-05, + "loss": 1.6192, + "step": 1520 + }, + { + "epoch": 0.35631113181183044, + "grad_norm": 8.600435256958008, + "learning_rate": 2.786303284416492e-05, + "loss": 1.2578, + "step": 1530 + }, + { + "epoch": 0.3586399627387052, + "grad_norm": 12.038568496704102, + "learning_rate": 2.7849056603773586e-05, + "loss": 1.5104, + "step": 1540 + }, + { + "epoch": 0.3609687936655799, + "grad_norm": 23.397693634033203, + "learning_rate": 2.7835080363382253e-05, + "loss": 1.5705, + "step": 1550 + }, + { + "epoch": 0.36329762459245457, + "grad_norm": 19.64638328552246, + "learning_rate": 2.7821104122990917e-05, + "loss": 1.2922, + "step": 1560 + }, + { + "epoch": 0.3656264555193293, + "grad_norm": 11.045112609863281, + "learning_rate": 2.7807127882599584e-05, + "loss": 1.354, + "step": 1570 + }, + { + "epoch": 0.367955286446204, + "grad_norm": 8.17792797088623, + "learning_rate": 2.7793151642208247e-05, + "loss": 1.4809, + "step": 1580 + }, + { + "epoch": 0.3702841173730787, + "grad_norm": 15.052375793457031, + "learning_rate": 2.7779175401816914e-05, + "loss": 1.3746, + "step": 1590 + }, + { + "epoch": 0.37261294829995345, + "grad_norm": 7.82668399810791, + "learning_rate": 2.7765199161425578e-05, + "loss": 1.3388, + "step": 1600 + }, + { + "epoch": 0.37494177922682814, + "grad_norm": 24.902265548706055, + "learning_rate": 2.7751222921034245e-05, + "loss": 1.447, + "step": 1610 + }, + { + "epoch": 0.37727061015370283, + "grad_norm": 21.929048538208008, + "learning_rate": 2.7737246680642908e-05, + "loss": 1.529, + "step": 1620 + }, + { + "epoch": 0.3795994410805775, + "grad_norm": 13.071199417114258, + "learning_rate": 2.7723270440251575e-05, + "loss": 1.7533, + "step": 1630 + }, + { + "epoch": 0.3819282720074523, + "grad_norm": 23.845232009887695, + "learning_rate": 2.7709294199860235e-05, + "loss": 1.697, + "step": 1640 + }, + { + "epoch": 0.38425710293432697, + "grad_norm": 16.655656814575195, + "learning_rate": 2.7695317959468902e-05, + "loss": 1.3866, + "step": 1650 + }, + { + "epoch": 0.38658593386120166, + "grad_norm": 5.929255962371826, + "learning_rate": 2.7681341719077566e-05, + "loss": 1.5702, + "step": 1660 + }, + { + "epoch": 0.3889147647880764, + "grad_norm": 5.026130676269531, + "learning_rate": 2.7667365478686233e-05, + "loss": 1.0261, + "step": 1670 + }, + { + "epoch": 0.3912435957149511, + "grad_norm": 48.83590316772461, + "learning_rate": 2.76533892382949e-05, + "loss": 1.9335, + "step": 1680 + }, + { + "epoch": 0.3935724266418258, + "grad_norm": 6.456787109375, + "learning_rate": 2.7639412997903564e-05, + "loss": 1.8191, + "step": 1690 + }, + { + "epoch": 0.39590125756870054, + "grad_norm": 7.918270111083984, + "learning_rate": 2.762543675751223e-05, + "loss": 1.4435, + "step": 1700 + }, + { + "epoch": 0.39823008849557523, + "grad_norm": 7.082935810089111, + "learning_rate": 2.7611460517120894e-05, + "loss": 1.1776, + "step": 1710 + }, + { + "epoch": 0.4005589194224499, + "grad_norm": 15.370290756225586, + "learning_rate": 2.759748427672956e-05, + "loss": 1.7205, + "step": 1720 + }, + { + "epoch": 0.4028877503493246, + "grad_norm": 4.2508392333984375, + "learning_rate": 2.7583508036338225e-05, + "loss": 1.4412, + "step": 1730 + }, + { + "epoch": 0.40521658127619936, + "grad_norm": 21.323366165161133, + "learning_rate": 2.756953179594689e-05, + "loss": 1.6467, + "step": 1740 + }, + { + "epoch": 0.40754541220307405, + "grad_norm": 6.784306049346924, + "learning_rate": 2.7555555555555555e-05, + "loss": 2.0458, + "step": 1750 + }, + { + "epoch": 0.40987424312994875, + "grad_norm": 5.1630988121032715, + "learning_rate": 2.7541579315164222e-05, + "loss": 1.3938, + "step": 1760 + }, + { + "epoch": 0.4122030740568235, + "grad_norm": 14.7197847366333, + "learning_rate": 2.7527603074772886e-05, + "loss": 1.5101, + "step": 1770 + }, + { + "epoch": 0.4145319049836982, + "grad_norm": 28.54730987548828, + "learning_rate": 2.7513626834381553e-05, + "loss": 1.5704, + "step": 1780 + }, + { + "epoch": 0.4168607359105729, + "grad_norm": 13.938275337219238, + "learning_rate": 2.7499650593990216e-05, + "loss": 1.446, + "step": 1790 + }, + { + "epoch": 0.4191895668374476, + "grad_norm": 45.821048736572266, + "learning_rate": 2.7485674353598883e-05, + "loss": 2.2999, + "step": 1800 + }, + { + "epoch": 0.4215183977643223, + "grad_norm": 4.642112731933594, + "learning_rate": 2.7471698113207547e-05, + "loss": 1.0337, + "step": 1810 + }, + { + "epoch": 0.423847228691197, + "grad_norm": 85.71287536621094, + "learning_rate": 2.7457721872816214e-05, + "loss": 1.5382, + "step": 1820 + }, + { + "epoch": 0.4261760596180717, + "grad_norm": 12.570047378540039, + "learning_rate": 2.744374563242488e-05, + "loss": 1.6143, + "step": 1830 + }, + { + "epoch": 0.42850489054494645, + "grad_norm": 44.61286544799805, + "learning_rate": 2.7429769392033544e-05, + "loss": 1.9973, + "step": 1840 + }, + { + "epoch": 0.43083372147182114, + "grad_norm": 10.40942096710205, + "learning_rate": 2.741579315164221e-05, + "loss": 1.6417, + "step": 1850 + }, + { + "epoch": 0.43316255239869583, + "grad_norm": 14.36320686340332, + "learning_rate": 2.7401816911250875e-05, + "loss": 1.5154, + "step": 1860 + }, + { + "epoch": 0.4354913833255706, + "grad_norm": 8.61720085144043, + "learning_rate": 2.7387840670859542e-05, + "loss": 1.4092, + "step": 1870 + }, + { + "epoch": 0.43782021425244527, + "grad_norm": 6.423306941986084, + "learning_rate": 2.7373864430468206e-05, + "loss": 1.7349, + "step": 1880 + }, + { + "epoch": 0.44014904517931996, + "grad_norm": 27.920555114746094, + "learning_rate": 2.7359888190076873e-05, + "loss": 1.2185, + "step": 1890 + }, + { + "epoch": 0.4424778761061947, + "grad_norm": 17.56697654724121, + "learning_rate": 2.7345911949685533e-05, + "loss": 2.2574, + "step": 1900 + }, + { + "epoch": 0.4448067070330694, + "grad_norm": 11.816027641296387, + "learning_rate": 2.73319357092942e-05, + "loss": 1.112, + "step": 1910 + }, + { + "epoch": 0.4471355379599441, + "grad_norm": 18.829771041870117, + "learning_rate": 2.7317959468902863e-05, + "loss": 1.5212, + "step": 1920 + }, + { + "epoch": 0.44946436888681884, + "grad_norm": 20.83513832092285, + "learning_rate": 2.730398322851153e-05, + "loss": 1.9292, + "step": 1930 + }, + { + "epoch": 0.45179319981369354, + "grad_norm": 6.374683380126953, + "learning_rate": 2.7290006988120194e-05, + "loss": 1.5452, + "step": 1940 + }, + { + "epoch": 0.4541220307405682, + "grad_norm": 10.145646095275879, + "learning_rate": 2.727603074772886e-05, + "loss": 1.779, + "step": 1950 + }, + { + "epoch": 0.4564508616674429, + "grad_norm": 15.401958465576172, + "learning_rate": 2.7262054507337525e-05, + "loss": 1.4472, + "step": 1960 + }, + { + "epoch": 0.45877969259431767, + "grad_norm": 70.80525970458984, + "learning_rate": 2.724807826694619e-05, + "loss": 1.7908, + "step": 1970 + }, + { + "epoch": 0.46110852352119236, + "grad_norm": 15.792387008666992, + "learning_rate": 2.723410202655486e-05, + "loss": 1.408, + "step": 1980 + }, + { + "epoch": 0.46343735444806705, + "grad_norm": 10.236753463745117, + "learning_rate": 2.7220125786163522e-05, + "loss": 1.2109, + "step": 1990 + }, + { + "epoch": 0.4657661853749418, + "grad_norm": 39.971900939941406, + "learning_rate": 2.720614954577219e-05, + "loss": 1.4968, + "step": 2000 + }, + { + "epoch": 0.4680950163018165, + "grad_norm": 13.20219612121582, + "learning_rate": 2.7192173305380853e-05, + "loss": 1.3401, + "step": 2010 + }, + { + "epoch": 0.4704238472286912, + "grad_norm": 20.595693588256836, + "learning_rate": 2.717819706498952e-05, + "loss": 1.3456, + "step": 2020 + }, + { + "epoch": 0.47275267815556593, + "grad_norm": 5.41182804107666, + "learning_rate": 2.7164220824598183e-05, + "loss": 1.4178, + "step": 2030 + }, + { + "epoch": 0.4750815090824406, + "grad_norm": 10.29110336303711, + "learning_rate": 2.715024458420685e-05, + "loss": 1.1662, + "step": 2040 + }, + { + "epoch": 0.4774103400093153, + "grad_norm": 8.8546781539917, + "learning_rate": 2.7136268343815514e-05, + "loss": 1.3032, + "step": 2050 + }, + { + "epoch": 0.47973917093619, + "grad_norm": 16.56825828552246, + "learning_rate": 2.712229210342418e-05, + "loss": 1.6524, + "step": 2060 + }, + { + "epoch": 0.48206800186306475, + "grad_norm": 8.039105415344238, + "learning_rate": 2.7108315863032844e-05, + "loss": 1.3149, + "step": 2070 + }, + { + "epoch": 0.48439683278993945, + "grad_norm": 7.804418087005615, + "learning_rate": 2.709433962264151e-05, + "loss": 1.2711, + "step": 2080 + }, + { + "epoch": 0.48672566371681414, + "grad_norm": 14.90645980834961, + "learning_rate": 2.7080363382250175e-05, + "loss": 1.6439, + "step": 2090 + }, + { + "epoch": 0.4890544946436889, + "grad_norm": 18.745508193969727, + "learning_rate": 2.7066387141858842e-05, + "loss": 1.3836, + "step": 2100 + }, + { + "epoch": 0.4913833255705636, + "grad_norm": 13.745448112487793, + "learning_rate": 2.7052410901467505e-05, + "loss": 1.7098, + "step": 2110 + }, + { + "epoch": 0.49371215649743827, + "grad_norm": 8.893570899963379, + "learning_rate": 2.7038434661076172e-05, + "loss": 1.5233, + "step": 2120 + }, + { + "epoch": 0.496040987424313, + "grad_norm": 5.876834869384766, + "learning_rate": 2.702445842068484e-05, + "loss": 1.7379, + "step": 2130 + }, + { + "epoch": 0.4983698183511877, + "grad_norm": 36.34437561035156, + "learning_rate": 2.7010482180293503e-05, + "loss": 1.2833, + "step": 2140 + }, + { + "epoch": 0.5006986492780624, + "grad_norm": 7.839767932891846, + "learning_rate": 2.699650593990217e-05, + "loss": 1.3774, + "step": 2150 + }, + { + "epoch": 0.5030274802049371, + "grad_norm": 20.03778648376465, + "learning_rate": 2.6982529699510834e-05, + "loss": 1.1518, + "step": 2160 + }, + { + "epoch": 0.5053563111318118, + "grad_norm": 16.76922607421875, + "learning_rate": 2.6968553459119497e-05, + "loss": 1.4198, + "step": 2170 + }, + { + "epoch": 0.5076851420586865, + "grad_norm": 15.900551795959473, + "learning_rate": 2.695457721872816e-05, + "loss": 1.8107, + "step": 2180 + }, + { + "epoch": 0.5100139729855613, + "grad_norm": 39.533077239990234, + "learning_rate": 2.6940600978336828e-05, + "loss": 1.5067, + "step": 2190 + }, + { + "epoch": 0.5123428039124359, + "grad_norm": 8.290783882141113, + "learning_rate": 2.692662473794549e-05, + "loss": 1.2505, + "step": 2200 + }, + { + "epoch": 0.5146716348393107, + "grad_norm": 11.7975492477417, + "learning_rate": 2.6912648497554158e-05, + "loss": 1.2652, + "step": 2210 + }, + { + "epoch": 0.5170004657661854, + "grad_norm": 35.9605712890625, + "learning_rate": 2.6898672257162822e-05, + "loss": 1.5008, + "step": 2220 + }, + { + "epoch": 0.51932929669306, + "grad_norm": 8.939215660095215, + "learning_rate": 2.688469601677149e-05, + "loss": 1.3159, + "step": 2230 + }, + { + "epoch": 0.5216581276199348, + "grad_norm": 29.108440399169922, + "learning_rate": 2.6870719776380152e-05, + "loss": 1.835, + "step": 2240 + }, + { + "epoch": 0.5239869585468095, + "grad_norm": 41.85350036621094, + "learning_rate": 2.685674353598882e-05, + "loss": 1.4529, + "step": 2250 + }, + { + "epoch": 0.5263157894736842, + "grad_norm": 15.356282234191895, + "learning_rate": 2.6842767295597483e-05, + "loss": 1.4199, + "step": 2260 + }, + { + "epoch": 0.5286446204005589, + "grad_norm": 10.252715110778809, + "learning_rate": 2.682879105520615e-05, + "loss": 1.171, + "step": 2270 + }, + { + "epoch": 0.5309734513274337, + "grad_norm": 23.096691131591797, + "learning_rate": 2.6814814814814817e-05, + "loss": 1.7865, + "step": 2280 + }, + { + "epoch": 0.5333022822543083, + "grad_norm": 6.9515910148620605, + "learning_rate": 2.680083857442348e-05, + "loss": 1.3458, + "step": 2290 + }, + { + "epoch": 0.5356311131811831, + "grad_norm": 39.11237335205078, + "learning_rate": 2.6786862334032148e-05, + "loss": 1.8611, + "step": 2300 + }, + { + "epoch": 0.5379599441080578, + "grad_norm": 6.406713485717773, + "learning_rate": 2.677288609364081e-05, + "loss": 1.3962, + "step": 2310 + }, + { + "epoch": 0.5402887750349324, + "grad_norm": 8.728299140930176, + "learning_rate": 2.6758909853249478e-05, + "loss": 1.7573, + "step": 2320 + }, + { + "epoch": 0.5426176059618072, + "grad_norm": 14.61226749420166, + "learning_rate": 2.674493361285814e-05, + "loss": 1.5703, + "step": 2330 + }, + { + "epoch": 0.5449464368886818, + "grad_norm": 13.667352676391602, + "learning_rate": 2.673095737246681e-05, + "loss": 1.0584, + "step": 2340 + }, + { + "epoch": 0.5472752678155566, + "grad_norm": 24.96540069580078, + "learning_rate": 2.6716981132075472e-05, + "loss": 1.4208, + "step": 2350 + }, + { + "epoch": 0.5496040987424313, + "grad_norm": 17.21762466430664, + "learning_rate": 2.670300489168414e-05, + "loss": 1.2686, + "step": 2360 + }, + { + "epoch": 0.551932929669306, + "grad_norm": 8.74630355834961, + "learning_rate": 2.6689028651292803e-05, + "loss": 1.3365, + "step": 2370 + }, + { + "epoch": 0.5542617605961807, + "grad_norm": 81.46407318115234, + "learning_rate": 2.667505241090147e-05, + "loss": 1.7805, + "step": 2380 + }, + { + "epoch": 0.5565905915230555, + "grad_norm": 7.569358825683594, + "learning_rate": 2.6661076170510133e-05, + "loss": 1.2882, + "step": 2390 + }, + { + "epoch": 0.5589194224499301, + "grad_norm": 9.290304183959961, + "learning_rate": 2.66470999301188e-05, + "loss": 1.3211, + "step": 2400 + }, + { + "epoch": 0.5612482533768048, + "grad_norm": 9.045026779174805, + "learning_rate": 2.6633123689727464e-05, + "loss": 1.4446, + "step": 2410 + }, + { + "epoch": 0.5635770843036796, + "grad_norm": 98.28316497802734, + "learning_rate": 2.661914744933613e-05, + "loss": 1.3712, + "step": 2420 + }, + { + "epoch": 0.5659059152305542, + "grad_norm": 15.944232940673828, + "learning_rate": 2.6605171208944795e-05, + "loss": 1.5916, + "step": 2430 + }, + { + "epoch": 0.568234746157429, + "grad_norm": 11.774263381958008, + "learning_rate": 2.6591194968553458e-05, + "loss": 1.3679, + "step": 2440 + }, + { + "epoch": 0.5705635770843037, + "grad_norm": 13.816254615783691, + "learning_rate": 2.6577218728162125e-05, + "loss": 1.1344, + "step": 2450 + }, + { + "epoch": 0.5728924080111784, + "grad_norm": 95.79457092285156, + "learning_rate": 2.656324248777079e-05, + "loss": 1.7998, + "step": 2460 + }, + { + "epoch": 0.5752212389380531, + "grad_norm": 9.518484115600586, + "learning_rate": 2.6549266247379456e-05, + "loss": 1.6957, + "step": 2470 + }, + { + "epoch": 0.5775500698649279, + "grad_norm": 7.290702819824219, + "learning_rate": 2.653529000698812e-05, + "loss": 1.2882, + "step": 2480 + }, + { + "epoch": 0.5798789007918025, + "grad_norm": 10.004209518432617, + "learning_rate": 2.6521313766596786e-05, + "loss": 0.9912, + "step": 2490 + }, + { + "epoch": 0.5822077317186772, + "grad_norm": 19.958898544311523, + "learning_rate": 2.650733752620545e-05, + "loss": 1.4569, + "step": 2500 + }, + { + "epoch": 0.584536562645552, + "grad_norm": 100.88062286376953, + "learning_rate": 2.6493361285814117e-05, + "loss": 1.8738, + "step": 2510 + }, + { + "epoch": 0.5868653935724266, + "grad_norm": 16.205934524536133, + "learning_rate": 2.647938504542278e-05, + "loss": 1.2735, + "step": 2520 + }, + { + "epoch": 0.5891942244993014, + "grad_norm": 26.231815338134766, + "learning_rate": 2.6465408805031447e-05, + "loss": 1.284, + "step": 2530 + }, + { + "epoch": 0.5915230554261761, + "grad_norm": 43.43105697631836, + "learning_rate": 2.645143256464011e-05, + "loss": 1.5655, + "step": 2540 + }, + { + "epoch": 0.5938518863530508, + "grad_norm": 14.103740692138672, + "learning_rate": 2.6437456324248778e-05, + "loss": 1.9522, + "step": 2550 + }, + { + "epoch": 0.5961807172799255, + "grad_norm": 24.451862335205078, + "learning_rate": 2.642348008385744e-05, + "loss": 1.9858, + "step": 2560 + }, + { + "epoch": 0.5985095482068001, + "grad_norm": 11.756208419799805, + "learning_rate": 2.640950384346611e-05, + "loss": 1.5646, + "step": 2570 + }, + { + "epoch": 0.6008383791336749, + "grad_norm": 22.268434524536133, + "learning_rate": 2.6395527603074775e-05, + "loss": 1.2753, + "step": 2580 + }, + { + "epoch": 0.6031672100605496, + "grad_norm": 10.877938270568848, + "learning_rate": 2.638155136268344e-05, + "loss": 1.7155, + "step": 2590 + }, + { + "epoch": 0.6054960409874243, + "grad_norm": 19.884037017822266, + "learning_rate": 2.6367575122292106e-05, + "loss": 1.0979, + "step": 2600 + }, + { + "epoch": 0.607824871914299, + "grad_norm": 9.810844421386719, + "learning_rate": 2.635359888190077e-05, + "loss": 1.3613, + "step": 2610 + }, + { + "epoch": 0.6101537028411738, + "grad_norm": 10.557251930236816, + "learning_rate": 2.6339622641509437e-05, + "loss": 1.3459, + "step": 2620 + }, + { + "epoch": 0.6124825337680484, + "grad_norm": 13.84223461151123, + "learning_rate": 2.63256464011181e-05, + "loss": 1.315, + "step": 2630 + }, + { + "epoch": 0.6148113646949231, + "grad_norm": 15.142064094543457, + "learning_rate": 2.6311670160726767e-05, + "loss": 1.5922, + "step": 2640 + }, + { + "epoch": 0.6171401956217979, + "grad_norm": 29.648056030273438, + "learning_rate": 2.629769392033543e-05, + "loss": 1.3471, + "step": 2650 + }, + { + "epoch": 0.6194690265486725, + "grad_norm": 5.652649879455566, + "learning_rate": 2.6283717679944098e-05, + "loss": 1.5202, + "step": 2660 + }, + { + "epoch": 0.6217978574755473, + "grad_norm": 7.965731143951416, + "learning_rate": 2.626974143955276e-05, + "loss": 1.0326, + "step": 2670 + }, + { + "epoch": 0.624126688402422, + "grad_norm": 60.36565017700195, + "learning_rate": 2.6255765199161428e-05, + "loss": 1.3599, + "step": 2680 + }, + { + "epoch": 0.6264555193292967, + "grad_norm": 12.119543075561523, + "learning_rate": 2.624178895877009e-05, + "loss": 1.4352, + "step": 2690 + }, + { + "epoch": 0.6287843502561714, + "grad_norm": 18.028051376342773, + "learning_rate": 2.6227812718378755e-05, + "loss": 1.7631, + "step": 2700 + }, + { + "epoch": 0.6311131811830462, + "grad_norm": 14.728964805603027, + "learning_rate": 2.621383647798742e-05, + "loss": 1.5176, + "step": 2710 + }, + { + "epoch": 0.6334420121099208, + "grad_norm": 46.449974060058594, + "learning_rate": 2.6199860237596086e-05, + "loss": 1.5507, + "step": 2720 + }, + { + "epoch": 0.6357708430367955, + "grad_norm": 5.887063026428223, + "learning_rate": 2.6185883997204753e-05, + "loss": 1.3399, + "step": 2730 + }, + { + "epoch": 0.6380996739636703, + "grad_norm": 9.698420524597168, + "learning_rate": 2.6171907756813417e-05, + "loss": 1.0447, + "step": 2740 + }, + { + "epoch": 0.6404285048905449, + "grad_norm": 8.20369815826416, + "learning_rate": 2.6157931516422084e-05, + "loss": 1.4727, + "step": 2750 + }, + { + "epoch": 0.6427573358174197, + "grad_norm": 17.19390869140625, + "learning_rate": 2.6143955276030747e-05, + "loss": 1.4083, + "step": 2760 + }, + { + "epoch": 0.6450861667442943, + "grad_norm": 17.035465240478516, + "learning_rate": 2.6129979035639414e-05, + "loss": 1.1995, + "step": 2770 + }, + { + "epoch": 0.6474149976711691, + "grad_norm": 7.173000812530518, + "learning_rate": 2.6116002795248078e-05, + "loss": 1.1313, + "step": 2780 + }, + { + "epoch": 0.6497438285980438, + "grad_norm": 20.684951782226562, + "learning_rate": 2.6102026554856745e-05, + "loss": 1.5742, + "step": 2790 + }, + { + "epoch": 0.6520726595249184, + "grad_norm": 9.083842277526855, + "learning_rate": 2.6088050314465408e-05, + "loss": 1.4098, + "step": 2800 + }, + { + "epoch": 0.6544014904517932, + "grad_norm": 7.859493732452393, + "learning_rate": 2.6074074074074075e-05, + "loss": 1.3814, + "step": 2810 + }, + { + "epoch": 0.6567303213786679, + "grad_norm": 5.692265033721924, + "learning_rate": 2.606009783368274e-05, + "loss": 1.7696, + "step": 2820 + }, + { + "epoch": 0.6590591523055426, + "grad_norm": 46.135520935058594, + "learning_rate": 2.6046121593291406e-05, + "loss": 2.3788, + "step": 2830 + }, + { + "epoch": 0.6613879832324173, + "grad_norm": 13.176454544067383, + "learning_rate": 2.603214535290007e-05, + "loss": 1.8681, + "step": 2840 + }, + { + "epoch": 0.6637168141592921, + "grad_norm": 13.691810607910156, + "learning_rate": 2.6018169112508736e-05, + "loss": 1.5054, + "step": 2850 + }, + { + "epoch": 0.6660456450861667, + "grad_norm": 115.35124969482422, + "learning_rate": 2.60041928721174e-05, + "loss": 1.4528, + "step": 2860 + }, + { + "epoch": 0.6683744760130415, + "grad_norm": 4.287045478820801, + "learning_rate": 2.5990216631726067e-05, + "loss": 1.1515, + "step": 2870 + }, + { + "epoch": 0.6707033069399162, + "grad_norm": 6.276763439178467, + "learning_rate": 2.5976240391334734e-05, + "loss": 1.3343, + "step": 2880 + }, + { + "epoch": 0.6730321378667908, + "grad_norm": 54.19549560546875, + "learning_rate": 2.5962264150943398e-05, + "loss": 1.9168, + "step": 2890 + }, + { + "epoch": 0.6753609687936656, + "grad_norm": 24.059030532836914, + "learning_rate": 2.5948287910552065e-05, + "loss": 1.4515, + "step": 2900 + }, + { + "epoch": 0.6776897997205403, + "grad_norm": 25.165706634521484, + "learning_rate": 2.5934311670160728e-05, + "loss": 1.6255, + "step": 2910 + }, + { + "epoch": 0.680018630647415, + "grad_norm": 16.515329360961914, + "learning_rate": 2.5920335429769395e-05, + "loss": 1.3936, + "step": 2920 + }, + { + "epoch": 0.6823474615742897, + "grad_norm": 21.545387268066406, + "learning_rate": 2.590635918937806e-05, + "loss": 1.5411, + "step": 2930 + }, + { + "epoch": 0.6846762925011645, + "grad_norm": 12.141153335571289, + "learning_rate": 2.5892382948986726e-05, + "loss": 2.0825, + "step": 2940 + }, + { + "epoch": 0.6870051234280391, + "grad_norm": 11.131927490234375, + "learning_rate": 2.587840670859539e-05, + "loss": 1.2421, + "step": 2950 + }, + { + "epoch": 0.6893339543549138, + "grad_norm": 8.020845413208008, + "learning_rate": 2.5864430468204053e-05, + "loss": 1.2091, + "step": 2960 + }, + { + "epoch": 0.6916627852817886, + "grad_norm": 26.62967300415039, + "learning_rate": 2.5850454227812716e-05, + "loss": 1.3636, + "step": 2970 + }, + { + "epoch": 0.6939916162086632, + "grad_norm": 15.849566459655762, + "learning_rate": 2.5836477987421383e-05, + "loss": 1.1627, + "step": 2980 + }, + { + "epoch": 0.696320447135538, + "grad_norm": 12.134406089782715, + "learning_rate": 2.5822501747030047e-05, + "loss": 1.4755, + "step": 2990 + }, + { + "epoch": 0.6986492780624126, + "grad_norm": 29.160348892211914, + "learning_rate": 2.5808525506638714e-05, + "loss": 1.3963, + "step": 3000 + }, + { + "epoch": 0.7009781089892874, + "grad_norm": 5.01591157913208, + "learning_rate": 2.5794549266247378e-05, + "loss": 1.144, + "step": 3010 + }, + { + "epoch": 0.7033069399161621, + "grad_norm": 14.045069694519043, + "learning_rate": 2.5780573025856045e-05, + "loss": 1.002, + "step": 3020 + }, + { + "epoch": 0.7056357708430367, + "grad_norm": 23.710153579711914, + "learning_rate": 2.576659678546471e-05, + "loss": 1.3336, + "step": 3030 + }, + { + "epoch": 0.7079646017699115, + "grad_norm": 25.294391632080078, + "learning_rate": 2.5752620545073375e-05, + "loss": 1.3291, + "step": 3040 + }, + { + "epoch": 0.7102934326967862, + "grad_norm": 15.712854385375977, + "learning_rate": 2.5738644304682042e-05, + "loss": 1.5373, + "step": 3050 + }, + { + "epoch": 0.7126222636236609, + "grad_norm": 19.031715393066406, + "learning_rate": 2.5724668064290706e-05, + "loss": 1.4451, + "step": 3060 + }, + { + "epoch": 0.7149510945505356, + "grad_norm": 15.004422187805176, + "learning_rate": 2.5710691823899373e-05, + "loss": 1.6015, + "step": 3070 + }, + { + "epoch": 0.7172799254774104, + "grad_norm": 15.895539283752441, + "learning_rate": 2.5696715583508036e-05, + "loss": 1.3032, + "step": 3080 + }, + { + "epoch": 0.719608756404285, + "grad_norm": 30.933595657348633, + "learning_rate": 2.5682739343116703e-05, + "loss": 1.2197, + "step": 3090 + }, + { + "epoch": 0.7219375873311598, + "grad_norm": 56.714134216308594, + "learning_rate": 2.5668763102725367e-05, + "loss": 1.5058, + "step": 3100 + }, + { + "epoch": 0.7242664182580345, + "grad_norm": 7.623404502868652, + "learning_rate": 2.5654786862334034e-05, + "loss": 1.1573, + "step": 3110 + }, + { + "epoch": 0.7265952491849091, + "grad_norm": 18.51123809814453, + "learning_rate": 2.5640810621942697e-05, + "loss": 1.1029, + "step": 3120 + }, + { + "epoch": 0.7289240801117839, + "grad_norm": 48.92169189453125, + "learning_rate": 2.5626834381551364e-05, + "loss": 1.245, + "step": 3130 + }, + { + "epoch": 0.7312529110386586, + "grad_norm": 11.136435508728027, + "learning_rate": 2.5612858141160028e-05, + "loss": 1.1932, + "step": 3140 + }, + { + "epoch": 0.7335817419655333, + "grad_norm": 44.02711868286133, + "learning_rate": 2.5598881900768695e-05, + "loss": 1.6498, + "step": 3150 + }, + { + "epoch": 0.735910572892408, + "grad_norm": 15.315217018127441, + "learning_rate": 2.558490566037736e-05, + "loss": 1.5206, + "step": 3160 + }, + { + "epoch": 0.7382394038192828, + "grad_norm": 13.75929069519043, + "learning_rate": 2.5570929419986025e-05, + "loss": 1.4698, + "step": 3170 + }, + { + "epoch": 0.7405682347461574, + "grad_norm": 14.253317832946777, + "learning_rate": 2.5556953179594692e-05, + "loss": 1.2942, + "step": 3180 + }, + { + "epoch": 0.7428970656730322, + "grad_norm": 54.78810501098633, + "learning_rate": 2.5542976939203356e-05, + "loss": 1.0529, + "step": 3190 + }, + { + "epoch": 0.7452258965999069, + "grad_norm": 3.501985788345337, + "learning_rate": 2.5529000698812023e-05, + "loss": 1.5216, + "step": 3200 + }, + { + "epoch": 0.7475547275267815, + "grad_norm": 10.592144966125488, + "learning_rate": 2.5515024458420687e-05, + "loss": 0.8294, + "step": 3210 + }, + { + "epoch": 0.7498835584536563, + "grad_norm": 26.526243209838867, + "learning_rate": 2.550104821802935e-05, + "loss": 1.5043, + "step": 3220 + }, + { + "epoch": 0.7522123893805309, + "grad_norm": 18.082988739013672, + "learning_rate": 2.5487071977638014e-05, + "loss": 0.922, + "step": 3230 + }, + { + "epoch": 0.7545412203074057, + "grad_norm": 9.192100524902344, + "learning_rate": 2.547309573724668e-05, + "loss": 1.5649, + "step": 3240 + }, + { + "epoch": 0.7568700512342804, + "grad_norm": 16.617460250854492, + "learning_rate": 2.5459119496855344e-05, + "loss": 1.109, + "step": 3250 + }, + { + "epoch": 0.759198882161155, + "grad_norm": 33.06241989135742, + "learning_rate": 2.544514325646401e-05, + "loss": 1.6267, + "step": 3260 + }, + { + "epoch": 0.7615277130880298, + "grad_norm": 9.304107666015625, + "learning_rate": 2.5431167016072675e-05, + "loss": 1.359, + "step": 3270 + }, + { + "epoch": 0.7638565440149045, + "grad_norm": 37.228939056396484, + "learning_rate": 2.5417190775681342e-05, + "loss": 2.4155, + "step": 3280 + }, + { + "epoch": 0.7661853749417792, + "grad_norm": 13.651851654052734, + "learning_rate": 2.5403214535290005e-05, + "loss": 1.1901, + "step": 3290 + }, + { + "epoch": 0.7685142058686539, + "grad_norm": 9.29826831817627, + "learning_rate": 2.5389238294898672e-05, + "loss": 1.3085, + "step": 3300 + }, + { + "epoch": 0.7708430367955287, + "grad_norm": 9.485541343688965, + "learning_rate": 2.5375262054507336e-05, + "loss": 1.4252, + "step": 3310 + }, + { + "epoch": 0.7731718677224033, + "grad_norm": 61.758750915527344, + "learning_rate": 2.5361285814116003e-05, + "loss": 1.7734, + "step": 3320 + }, + { + "epoch": 0.7755006986492781, + "grad_norm": 23.328594207763672, + "learning_rate": 2.534730957372467e-05, + "loss": 1.4627, + "step": 3330 + }, + { + "epoch": 0.7778295295761528, + "grad_norm": 8.944003105163574, + "learning_rate": 2.5333333333333334e-05, + "loss": 1.4042, + "step": 3340 + }, + { + "epoch": 0.7801583605030274, + "grad_norm": 10.560107231140137, + "learning_rate": 2.5319357092942e-05, + "loss": 1.7253, + "step": 3350 + }, + { + "epoch": 0.7824871914299022, + "grad_norm": 9.760632514953613, + "learning_rate": 2.5305380852550664e-05, + "loss": 1.6111, + "step": 3360 + }, + { + "epoch": 0.7848160223567769, + "grad_norm": 14.777600288391113, + "learning_rate": 2.529140461215933e-05, + "loss": 1.1902, + "step": 3370 + }, + { + "epoch": 0.7871448532836516, + "grad_norm": 7.539825916290283, + "learning_rate": 2.5277428371767995e-05, + "loss": 1.2884, + "step": 3380 + }, + { + "epoch": 0.7894736842105263, + "grad_norm": 15.00157642364502, + "learning_rate": 2.5263452131376662e-05, + "loss": 1.1476, + "step": 3390 + }, + { + "epoch": 0.7918025151374011, + "grad_norm": 17.299943923950195, + "learning_rate": 2.5249475890985325e-05, + "loss": 1.3168, + "step": 3400 + }, + { + "epoch": 0.7941313460642757, + "grad_norm": 118.13330841064453, + "learning_rate": 2.5235499650593992e-05, + "loss": 2.3259, + "step": 3410 + }, + { + "epoch": 0.7964601769911505, + "grad_norm": 14.524819374084473, + "learning_rate": 2.5221523410202656e-05, + "loss": 1.364, + "step": 3420 + }, + { + "epoch": 0.7987890079180252, + "grad_norm": 13.984453201293945, + "learning_rate": 2.5207547169811323e-05, + "loss": 1.0494, + "step": 3430 + }, + { + "epoch": 0.8011178388448998, + "grad_norm": 24.410167694091797, + "learning_rate": 2.5193570929419986e-05, + "loss": 1.1634, + "step": 3440 + }, + { + "epoch": 0.8034466697717746, + "grad_norm": 9.531725883483887, + "learning_rate": 2.5179594689028653e-05, + "loss": 1.4717, + "step": 3450 + }, + { + "epoch": 0.8057755006986492, + "grad_norm": 8.417661666870117, + "learning_rate": 2.5165618448637317e-05, + "loss": 1.006, + "step": 3460 + }, + { + "epoch": 0.808104331625524, + "grad_norm": 16.109172821044922, + "learning_rate": 2.5151642208245984e-05, + "loss": 1.1106, + "step": 3470 + }, + { + "epoch": 0.8104331625523987, + "grad_norm": 12.241450309753418, + "learning_rate": 2.5137665967854648e-05, + "loss": 1.2066, + "step": 3480 + }, + { + "epoch": 0.8127619934792734, + "grad_norm": 13.93542194366455, + "learning_rate": 2.512368972746331e-05, + "loss": 1.6023, + "step": 3490 + }, + { + "epoch": 0.8150908244061481, + "grad_norm": 33.94037628173828, + "learning_rate": 2.5109713487071978e-05, + "loss": 1.5284, + "step": 3500 + }, + { + "epoch": 0.8174196553330229, + "grad_norm": 13.690526008605957, + "learning_rate": 2.5095737246680642e-05, + "loss": 1.6503, + "step": 3510 + }, + { + "epoch": 0.8197484862598975, + "grad_norm": 11.894718170166016, + "learning_rate": 2.508176100628931e-05, + "loss": 1.118, + "step": 3520 + }, + { + "epoch": 0.8220773171867722, + "grad_norm": 18.130245208740234, + "learning_rate": 2.5067784765897972e-05, + "loss": 1.1148, + "step": 3530 + }, + { + "epoch": 0.824406148113647, + "grad_norm": 10.06002140045166, + "learning_rate": 2.505380852550664e-05, + "loss": 1.2269, + "step": 3540 + }, + { + "epoch": 0.8267349790405216, + "grad_norm": 17.191099166870117, + "learning_rate": 2.5039832285115303e-05, + "loss": 1.3745, + "step": 3550 + }, + { + "epoch": 0.8290638099673964, + "grad_norm": 23.614267349243164, + "learning_rate": 2.502585604472397e-05, + "loss": 1.2424, + "step": 3560 + }, + { + "epoch": 0.8313926408942711, + "grad_norm": 6.034205436706543, + "learning_rate": 2.5011879804332633e-05, + "loss": 1.0718, + "step": 3570 + }, + { + "epoch": 0.8337214718211458, + "grad_norm": 17.800487518310547, + "learning_rate": 2.49979035639413e-05, + "loss": 1.6048, + "step": 3580 + }, + { + "epoch": 0.8360503027480205, + "grad_norm": 5.585227012634277, + "learning_rate": 2.4983927323549964e-05, + "loss": 1.153, + "step": 3590 + }, + { + "epoch": 0.8383791336748952, + "grad_norm": 37.87883377075195, + "learning_rate": 2.496995108315863e-05, + "loss": 1.4075, + "step": 3600 + }, + { + "epoch": 0.8407079646017699, + "grad_norm": 10.529804229736328, + "learning_rate": 2.4955974842767295e-05, + "loss": 1.215, + "step": 3610 + }, + { + "epoch": 0.8430367955286446, + "grad_norm": 16.147066116333008, + "learning_rate": 2.494199860237596e-05, + "loss": 1.3325, + "step": 3620 + }, + { + "epoch": 0.8453656264555194, + "grad_norm": 8.722766876220703, + "learning_rate": 2.492802236198463e-05, + "loss": 1.0484, + "step": 3630 + }, + { + "epoch": 0.847694457382394, + "grad_norm": 2.9982659816741943, + "learning_rate": 2.4914046121593292e-05, + "loss": 1.0521, + "step": 3640 + }, + { + "epoch": 0.8500232883092688, + "grad_norm": 64.77198791503906, + "learning_rate": 2.490006988120196e-05, + "loss": 1.4048, + "step": 3650 + }, + { + "epoch": 0.8523521192361434, + "grad_norm": 10.345009803771973, + "learning_rate": 2.4886093640810623e-05, + "loss": 1.3138, + "step": 3660 + }, + { + "epoch": 0.8546809501630181, + "grad_norm": 43.743690490722656, + "learning_rate": 2.487211740041929e-05, + "loss": 1.3493, + "step": 3670 + }, + { + "epoch": 0.8570097810898929, + "grad_norm": 19.04966163635254, + "learning_rate": 2.4858141160027953e-05, + "loss": 1.2579, + "step": 3680 + }, + { + "epoch": 0.8593386120167675, + "grad_norm": 19.782512664794922, + "learning_rate": 2.484416491963662e-05, + "loss": 1.6133, + "step": 3690 + }, + { + "epoch": 0.8616674429436423, + "grad_norm": 48.4581298828125, + "learning_rate": 2.4830188679245284e-05, + "loss": 1.6982, + "step": 3700 + }, + { + "epoch": 0.863996273870517, + "grad_norm": 7.717815399169922, + "learning_rate": 2.481621243885395e-05, + "loss": 1.4523, + "step": 3710 + }, + { + "epoch": 0.8663251047973917, + "grad_norm": 22.847097396850586, + "learning_rate": 2.4802236198462614e-05, + "loss": 1.4004, + "step": 3720 + }, + { + "epoch": 0.8686539357242664, + "grad_norm": 118.66568756103516, + "learning_rate": 2.478825995807128e-05, + "loss": 1.4469, + "step": 3730 + }, + { + "epoch": 0.8709827666511412, + "grad_norm": 17.42070198059082, + "learning_rate": 2.4774283717679945e-05, + "loss": 1.8226, + "step": 3740 + }, + { + "epoch": 0.8733115975780158, + "grad_norm": 20.360017776489258, + "learning_rate": 2.476030747728861e-05, + "loss": 1.8337, + "step": 3750 + }, + { + "epoch": 0.8756404285048905, + "grad_norm": 20.167600631713867, + "learning_rate": 2.4746331236897275e-05, + "loss": 1.4437, + "step": 3760 + }, + { + "epoch": 0.8779692594317653, + "grad_norm": 9.806801795959473, + "learning_rate": 2.473235499650594e-05, + "loss": 1.0921, + "step": 3770 + }, + { + "epoch": 0.8802980903586399, + "grad_norm": 19.070938110351562, + "learning_rate": 2.4718378756114606e-05, + "loss": 1.0835, + "step": 3780 + }, + { + "epoch": 0.8826269212855147, + "grad_norm": 52.30155944824219, + "learning_rate": 2.470440251572327e-05, + "loss": 1.1373, + "step": 3790 + }, + { + "epoch": 0.8849557522123894, + "grad_norm": 31.602767944335938, + "learning_rate": 2.4690426275331937e-05, + "loss": 1.4057, + "step": 3800 + }, + { + "epoch": 0.8872845831392641, + "grad_norm": 6.680765628814697, + "learning_rate": 2.46764500349406e-05, + "loss": 1.8083, + "step": 3810 + }, + { + "epoch": 0.8896134140661388, + "grad_norm": 14.354714393615723, + "learning_rate": 2.4662473794549267e-05, + "loss": 1.1566, + "step": 3820 + }, + { + "epoch": 0.8919422449930136, + "grad_norm": 5.738077640533447, + "learning_rate": 2.464849755415793e-05, + "loss": 1.0986, + "step": 3830 + }, + { + "epoch": 0.8942710759198882, + "grad_norm": 12.508646011352539, + "learning_rate": 2.4634521313766598e-05, + "loss": 0.9977, + "step": 3840 + }, + { + "epoch": 0.8965999068467629, + "grad_norm": 12.393445014953613, + "learning_rate": 2.462054507337526e-05, + "loss": 1.6198, + "step": 3850 + }, + { + "epoch": 0.8989287377736377, + "grad_norm": 12.13697624206543, + "learning_rate": 2.460656883298393e-05, + "loss": 1.8646, + "step": 3860 + }, + { + "epoch": 0.9012575687005123, + "grad_norm": 29.515411376953125, + "learning_rate": 2.4592592592592592e-05, + "loss": 1.4014, + "step": 3870 + }, + { + "epoch": 0.9035863996273871, + "grad_norm": 9.969564437866211, + "learning_rate": 2.457861635220126e-05, + "loss": 0.9157, + "step": 3880 + }, + { + "epoch": 0.9059152305542617, + "grad_norm": 70.15406799316406, + "learning_rate": 2.4564640111809922e-05, + "loss": 1.3138, + "step": 3890 + }, + { + "epoch": 0.9082440614811365, + "grad_norm": 16.15462875366211, + "learning_rate": 2.455066387141859e-05, + "loss": 1.439, + "step": 3900 + }, + { + "epoch": 0.9105728924080112, + "grad_norm": 23.706735610961914, + "learning_rate": 2.4536687631027256e-05, + "loss": 1.3809, + "step": 3910 + }, + { + "epoch": 0.9129017233348858, + "grad_norm": 11.798118591308594, + "learning_rate": 2.452271139063592e-05, + "loss": 1.2233, + "step": 3920 + }, + { + "epoch": 0.9152305542617606, + "grad_norm": 6.833698749542236, + "learning_rate": 2.4508735150244587e-05, + "loss": 1.09, + "step": 3930 + }, + { + "epoch": 0.9175593851886353, + "grad_norm": 13.652518272399902, + "learning_rate": 2.449475890985325e-05, + "loss": 1.2642, + "step": 3940 + }, + { + "epoch": 0.91988821611551, + "grad_norm": 10.70895004272461, + "learning_rate": 2.4480782669461918e-05, + "loss": 1.7104, + "step": 3950 + }, + { + "epoch": 0.9222170470423847, + "grad_norm": 4.455838203430176, + "learning_rate": 2.446680642907058e-05, + "loss": 1.2201, + "step": 3960 + }, + { + "epoch": 0.9245458779692595, + "grad_norm": 20.450454711914062, + "learning_rate": 2.4452830188679248e-05, + "loss": 0.9957, + "step": 3970 + }, + { + "epoch": 0.9268747088961341, + "grad_norm": 40.74632263183594, + "learning_rate": 2.4438853948287912e-05, + "loss": 1.3495, + "step": 3980 + }, + { + "epoch": 0.9292035398230089, + "grad_norm": 11.148481369018555, + "learning_rate": 2.442487770789658e-05, + "loss": 1.1178, + "step": 3990 + }, + { + "epoch": 0.9315323707498836, + "grad_norm": 7.56870174407959, + "learning_rate": 2.4410901467505242e-05, + "loss": 0.9451, + "step": 4000 + }, + { + "epoch": 0.9338612016767582, + "grad_norm": 38.12516403198242, + "learning_rate": 2.4396925227113906e-05, + "loss": 1.1355, + "step": 4010 + }, + { + "epoch": 0.936190032603633, + "grad_norm": 49.5767707824707, + "learning_rate": 2.438294898672257e-05, + "loss": 1.7586, + "step": 4020 + }, + { + "epoch": 0.9385188635305077, + "grad_norm": 13.015122413635254, + "learning_rate": 2.4368972746331236e-05, + "loss": 1.873, + "step": 4030 + }, + { + "epoch": 0.9408476944573824, + "grad_norm": 26.832674026489258, + "learning_rate": 2.43549965059399e-05, + "loss": 1.553, + "step": 4040 + }, + { + "epoch": 0.9431765253842571, + "grad_norm": 13.10566520690918, + "learning_rate": 2.4341020265548567e-05, + "loss": 1.5644, + "step": 4050 + }, + { + "epoch": 0.9455053563111319, + "grad_norm": 12.45157527923584, + "learning_rate": 2.4327044025157234e-05, + "loss": 1.6215, + "step": 4060 + }, + { + "epoch": 0.9478341872380065, + "grad_norm": 28.12361717224121, + "learning_rate": 2.4313067784765898e-05, + "loss": 1.126, + "step": 4070 + }, + { + "epoch": 0.9501630181648812, + "grad_norm": 11.773497581481934, + "learning_rate": 2.4299091544374565e-05, + "loss": 1.106, + "step": 4080 + }, + { + "epoch": 0.952491849091756, + "grad_norm": 15.833434104919434, + "learning_rate": 2.4285115303983228e-05, + "loss": 1.3659, + "step": 4090 + }, + { + "epoch": 0.9548206800186306, + "grad_norm": 7.810037612915039, + "learning_rate": 2.4271139063591895e-05, + "loss": 1.4978, + "step": 4100 + }, + { + "epoch": 0.9571495109455054, + "grad_norm": 14.155339241027832, + "learning_rate": 2.425716282320056e-05, + "loss": 1.0855, + "step": 4110 + }, + { + "epoch": 0.95947834187238, + "grad_norm": 7.536425590515137, + "learning_rate": 2.4243186582809226e-05, + "loss": 1.6742, + "step": 4120 + }, + { + "epoch": 0.9618071727992548, + "grad_norm": 41.47824478149414, + "learning_rate": 2.422921034241789e-05, + "loss": 1.7182, + "step": 4130 + }, + { + "epoch": 0.9641360037261295, + "grad_norm": 14.603498458862305, + "learning_rate": 2.4215234102026556e-05, + "loss": 0.9493, + "step": 4140 + }, + { + "epoch": 0.9664648346530041, + "grad_norm": 9.240854263305664, + "learning_rate": 2.420125786163522e-05, + "loss": 0.8127, + "step": 4150 + }, + { + "epoch": 0.9687936655798789, + "grad_norm": 11.538430213928223, + "learning_rate": 2.4187281621243887e-05, + "loss": 1.1822, + "step": 4160 + }, + { + "epoch": 0.9711224965067536, + "grad_norm": 10.658085823059082, + "learning_rate": 2.417330538085255e-05, + "loss": 1.4719, + "step": 4170 + }, + { + "epoch": 0.9734513274336283, + "grad_norm": 20.83334732055664, + "learning_rate": 2.4159329140461217e-05, + "loss": 1.1781, + "step": 4180 + }, + { + "epoch": 0.975780158360503, + "grad_norm": 8.09494400024414, + "learning_rate": 2.414535290006988e-05, + "loss": 1.3265, + "step": 4190 + }, + { + "epoch": 0.9781089892873778, + "grad_norm": 8.501059532165527, + "learning_rate": 2.4131376659678548e-05, + "loss": 0.9679, + "step": 4200 + }, + { + "epoch": 0.9804378202142524, + "grad_norm": 21.922080993652344, + "learning_rate": 2.4117400419287215e-05, + "loss": 1.3786, + "step": 4210 + }, + { + "epoch": 0.9827666511411272, + "grad_norm": 8.116072654724121, + "learning_rate": 2.410342417889588e-05, + "loss": 1.0864, + "step": 4220 + }, + { + "epoch": 0.9850954820680019, + "grad_norm": 39.07789611816406, + "learning_rate": 2.4089447938504546e-05, + "loss": 1.0388, + "step": 4230 + }, + { + "epoch": 0.9874243129948765, + "grad_norm": 18.8806095123291, + "learning_rate": 2.407547169811321e-05, + "loss": 1.2614, + "step": 4240 + }, + { + "epoch": 0.9897531439217513, + "grad_norm": 30.235370635986328, + "learning_rate": 2.4061495457721876e-05, + "loss": 1.0553, + "step": 4250 + }, + { + "epoch": 0.992081974848626, + "grad_norm": 11.719708442687988, + "learning_rate": 2.404751921733054e-05, + "loss": 1.7491, + "step": 4260 + }, + { + "epoch": 0.9944108057755007, + "grad_norm": 4.49058723449707, + "learning_rate": 2.4033542976939203e-05, + "loss": 1.6797, + "step": 4270 + }, + { + "epoch": 0.9967396367023754, + "grad_norm": 9.818137168884277, + "learning_rate": 2.4019566736547867e-05, + "loss": 1.5911, + "step": 4280 + }, + { + "epoch": 0.9990684676292502, + "grad_norm": 7.682907581329346, + "learning_rate": 2.4005590496156534e-05, + "loss": 1.3785, + "step": 4290 + }, + { + "epoch": 1.0, + "eval_loss": 0.5099594593048096, + "eval_runtime": 57.748, + "eval_samples_per_second": 75.431, + "eval_steps_per_second": 1.195, + "step": 4294 + }, + { + "epoch": 1.0013972985561248, + "grad_norm": 16.090312957763672, + "learning_rate": 2.3991614255765197e-05, + "loss": 1.3029, + "step": 4300 + }, + { + "epoch": 1.0037261294829996, + "grad_norm": 10.43845272064209, + "learning_rate": 2.3977638015373864e-05, + "loss": 1.3984, + "step": 4310 + }, + { + "epoch": 1.0060549604098743, + "grad_norm": 7.594318866729736, + "learning_rate": 2.3963661774982528e-05, + "loss": 1.2071, + "step": 4320 + }, + { + "epoch": 1.008383791336749, + "grad_norm": 5.082920551300049, + "learning_rate": 2.3949685534591195e-05, + "loss": 1.1765, + "step": 4330 + }, + { + "epoch": 1.0107126222636236, + "grad_norm": 34.897769927978516, + "learning_rate": 2.393570929419986e-05, + "loss": 1.8144, + "step": 4340 + }, + { + "epoch": 1.0130414531904983, + "grad_norm": 20.48177719116211, + "learning_rate": 2.3921733053808526e-05, + "loss": 1.0724, + "step": 4350 + }, + { + "epoch": 1.015370284117373, + "grad_norm": 18.330974578857422, + "learning_rate": 2.3907756813417192e-05, + "loss": 1.4672, + "step": 4360 + }, + { + "epoch": 1.0176991150442478, + "grad_norm": 28.877042770385742, + "learning_rate": 2.3893780573025856e-05, + "loss": 1.4611, + "step": 4370 + }, + { + "epoch": 1.0200279459711226, + "grad_norm": 13.684222221374512, + "learning_rate": 2.3879804332634523e-05, + "loss": 0.9294, + "step": 4380 + }, + { + "epoch": 1.0223567768979973, + "grad_norm": 4.9023590087890625, + "learning_rate": 2.3865828092243187e-05, + "loss": 1.1675, + "step": 4390 + }, + { + "epoch": 1.0246856078248718, + "grad_norm": 9.912734031677246, + "learning_rate": 2.3851851851851854e-05, + "loss": 1.0157, + "step": 4400 + }, + { + "epoch": 1.0270144387517466, + "grad_norm": 13.99254035949707, + "learning_rate": 2.3837875611460517e-05, + "loss": 1.295, + "step": 4410 + }, + { + "epoch": 1.0293432696786213, + "grad_norm": 14.370241165161133, + "learning_rate": 2.3823899371069184e-05, + "loss": 1.6396, + "step": 4420 + }, + { + "epoch": 1.031672100605496, + "grad_norm": 20.884004592895508, + "learning_rate": 2.3809923130677848e-05, + "loss": 1.3973, + "step": 4430 + }, + { + "epoch": 1.0340009315323708, + "grad_norm": 12.322709083557129, + "learning_rate": 2.3795946890286515e-05, + "loss": 0.9916, + "step": 4440 + }, + { + "epoch": 1.0363297624592454, + "grad_norm": 54.580814361572266, + "learning_rate": 2.378197064989518e-05, + "loss": 1.2983, + "step": 4450 + }, + { + "epoch": 1.03865859338612, + "grad_norm": 45.35833740234375, + "learning_rate": 2.3767994409503845e-05, + "loss": 1.6227, + "step": 4460 + }, + { + "epoch": 1.0409874243129948, + "grad_norm": 58.20479965209961, + "learning_rate": 2.375401816911251e-05, + "loss": 1.4488, + "step": 4470 + }, + { + "epoch": 1.0433162552398696, + "grad_norm": 21.809900283813477, + "learning_rate": 2.3740041928721176e-05, + "loss": 1.5132, + "step": 4480 + }, + { + "epoch": 1.0456450861667443, + "grad_norm": 6.9720540046691895, + "learning_rate": 2.372606568832984e-05, + "loss": 0.904, + "step": 4490 + }, + { + "epoch": 1.047973917093619, + "grad_norm": 7.57560920715332, + "learning_rate": 2.3712089447938506e-05, + "loss": 1.0406, + "step": 4500 + }, + { + "epoch": 1.0503027480204936, + "grad_norm": 43.4541015625, + "learning_rate": 2.3698113207547173e-05, + "loss": 1.2533, + "step": 4510 + }, + { + "epoch": 1.0526315789473684, + "grad_norm": 17.777565002441406, + "learning_rate": 2.3684136967155837e-05, + "loss": 1.3282, + "step": 4520 + }, + { + "epoch": 1.054960409874243, + "grad_norm": 15.077924728393555, + "learning_rate": 2.3670160726764504e-05, + "loss": 1.2443, + "step": 4530 + }, + { + "epoch": 1.0572892408011179, + "grad_norm": 9.286961555480957, + "learning_rate": 2.3656184486373164e-05, + "loss": 1.0529, + "step": 4540 + }, + { + "epoch": 1.0596180717279926, + "grad_norm": 42.42253494262695, + "learning_rate": 2.364220824598183e-05, + "loss": 2.2219, + "step": 4550 + }, + { + "epoch": 1.0619469026548674, + "grad_norm": 11.062445640563965, + "learning_rate": 2.3628232005590495e-05, + "loss": 1.5294, + "step": 4560 + }, + { + "epoch": 1.0642757335817419, + "grad_norm": 15.275858879089355, + "learning_rate": 2.3614255765199162e-05, + "loss": 1.0046, + "step": 4570 + }, + { + "epoch": 1.0666045645086166, + "grad_norm": 14.448469161987305, + "learning_rate": 2.3600279524807825e-05, + "loss": 1.5815, + "step": 4580 + }, + { + "epoch": 1.0689333954354914, + "grad_norm": 4.570057392120361, + "learning_rate": 2.3586303284416492e-05, + "loss": 0.858, + "step": 4590 + }, + { + "epoch": 1.0712622263623661, + "grad_norm": 12.531164169311523, + "learning_rate": 2.3572327044025156e-05, + "loss": 0.9637, + "step": 4600 + }, + { + "epoch": 1.0735910572892409, + "grad_norm": 14.942547798156738, + "learning_rate": 2.3558350803633823e-05, + "loss": 1.1421, + "step": 4610 + }, + { + "epoch": 1.0759198882161156, + "grad_norm": 16.772188186645508, + "learning_rate": 2.3544374563242486e-05, + "loss": 1.3581, + "step": 4620 + }, + { + "epoch": 1.0782487191429901, + "grad_norm": 17.264406204223633, + "learning_rate": 2.3530398322851153e-05, + "loss": 1.8468, + "step": 4630 + }, + { + "epoch": 1.080577550069865, + "grad_norm": 50.972537994384766, + "learning_rate": 2.3516422082459817e-05, + "loss": 1.6117, + "step": 4640 + }, + { + "epoch": 1.0829063809967396, + "grad_norm": 8.686941146850586, + "learning_rate": 2.3502445842068484e-05, + "loss": 1.2004, + "step": 4650 + }, + { + "epoch": 1.0852352119236144, + "grad_norm": 5.48530387878418, + "learning_rate": 2.348846960167715e-05, + "loss": 1.0691, + "step": 4660 + }, + { + "epoch": 1.0875640428504891, + "grad_norm": 7.076744556427002, + "learning_rate": 2.3474493361285815e-05, + "loss": 1.0191, + "step": 4670 + }, + { + "epoch": 1.0898928737773637, + "grad_norm": 6.224733352661133, + "learning_rate": 2.346051712089448e-05, + "loss": 1.4263, + "step": 4680 + }, + { + "epoch": 1.0922217047042384, + "grad_norm": 7.603545665740967, + "learning_rate": 2.3446540880503145e-05, + "loss": 0.9943, + "step": 4690 + }, + { + "epoch": 1.0945505356311132, + "grad_norm": 7.480013370513916, + "learning_rate": 2.3432564640111812e-05, + "loss": 0.624, + "step": 4700 + }, + { + "epoch": 1.096879366557988, + "grad_norm": 21.913307189941406, + "learning_rate": 2.3418588399720476e-05, + "loss": 1.1937, + "step": 4710 + }, + { + "epoch": 1.0992081974848626, + "grad_norm": 8.472567558288574, + "learning_rate": 2.3404612159329143e-05, + "loss": 1.4631, + "step": 4720 + }, + { + "epoch": 1.1015370284117374, + "grad_norm": 22.3406925201416, + "learning_rate": 2.3390635918937806e-05, + "loss": 1.332, + "step": 4730 + }, + { + "epoch": 1.103865859338612, + "grad_norm": 8.574953079223633, + "learning_rate": 2.3376659678546473e-05, + "loss": 0.936, + "step": 4740 + }, + { + "epoch": 1.1061946902654867, + "grad_norm": 7.095656394958496, + "learning_rate": 2.3362683438155137e-05, + "loss": 1.1048, + "step": 4750 + }, + { + "epoch": 1.1085235211923614, + "grad_norm": 16.59130859375, + "learning_rate": 2.3348707197763804e-05, + "loss": 1.1469, + "step": 4760 + }, + { + "epoch": 1.1108523521192362, + "grad_norm": 176.99810791015625, + "learning_rate": 2.3334730957372467e-05, + "loss": 1.3763, + "step": 4770 + }, + { + "epoch": 1.113181183046111, + "grad_norm": 9.466316223144531, + "learning_rate": 2.3320754716981134e-05, + "loss": 1.5032, + "step": 4780 + }, + { + "epoch": 1.1155100139729857, + "grad_norm": 10.125219345092773, + "learning_rate": 2.3306778476589798e-05, + "loss": 1.0508, + "step": 4790 + }, + { + "epoch": 1.1178388448998602, + "grad_norm": 8.458738327026367, + "learning_rate": 2.329280223619846e-05, + "loss": 1.1744, + "step": 4800 + }, + { + "epoch": 1.120167675826735, + "grad_norm": 95.53096771240234, + "learning_rate": 2.327882599580713e-05, + "loss": 1.8102, + "step": 4810 + }, + { + "epoch": 1.1224965067536097, + "grad_norm": 7.951642036437988, + "learning_rate": 2.3264849755415792e-05, + "loss": 1.2638, + "step": 4820 + }, + { + "epoch": 1.1248253376804844, + "grad_norm": 4.9284586906433105, + "learning_rate": 2.325087351502446e-05, + "loss": 1.5309, + "step": 4830 + }, + { + "epoch": 1.1271541686073592, + "grad_norm": 23.07508659362793, + "learning_rate": 2.3236897274633123e-05, + "loss": 1.1686, + "step": 4840 + }, + { + "epoch": 1.129482999534234, + "grad_norm": 11.482251167297363, + "learning_rate": 2.322292103424179e-05, + "loss": 0.9867, + "step": 4850 + }, + { + "epoch": 1.1318118304611084, + "grad_norm": 54.271514892578125, + "learning_rate": 2.3208944793850453e-05, + "loss": 1.16, + "step": 4860 + }, + { + "epoch": 1.1341406613879832, + "grad_norm": 18.813350677490234, + "learning_rate": 2.319496855345912e-05, + "loss": 1.2557, + "step": 4870 + }, + { + "epoch": 1.136469492314858, + "grad_norm": 11.297906875610352, + "learning_rate": 2.3180992313067784e-05, + "loss": 1.4021, + "step": 4880 + }, + { + "epoch": 1.1387983232417327, + "grad_norm": 9.278247833251953, + "learning_rate": 2.316701607267645e-05, + "loss": 1.2495, + "step": 4890 + }, + { + "epoch": 1.1411271541686074, + "grad_norm": 13.483011245727539, + "learning_rate": 2.3153039832285114e-05, + "loss": 1.701, + "step": 4900 + }, + { + "epoch": 1.143455985095482, + "grad_norm": 12.479578971862793, + "learning_rate": 2.313906359189378e-05, + "loss": 1.1971, + "step": 4910 + }, + { + "epoch": 1.1457848160223567, + "grad_norm": 64.85933685302734, + "learning_rate": 2.3125087351502445e-05, + "loss": 1.8314, + "step": 4920 + }, + { + "epoch": 1.1481136469492315, + "grad_norm": 109.56636810302734, + "learning_rate": 2.3111111111111112e-05, + "loss": 1.1742, + "step": 4930 + }, + { + "epoch": 1.1504424778761062, + "grad_norm": 23.607959747314453, + "learning_rate": 2.3097134870719776e-05, + "loss": 0.9286, + "step": 4940 + }, + { + "epoch": 1.152771308802981, + "grad_norm": 7.5170207023620605, + "learning_rate": 2.3083158630328443e-05, + "loss": 1.0913, + "step": 4950 + }, + { + "epoch": 1.1551001397298557, + "grad_norm": 13.472895622253418, + "learning_rate": 2.306918238993711e-05, + "loss": 1.5781, + "step": 4960 + }, + { + "epoch": 1.1574289706567302, + "grad_norm": 73.42911529541016, + "learning_rate": 2.3055206149545773e-05, + "loss": 1.3364, + "step": 4970 + }, + { + "epoch": 1.159757801583605, + "grad_norm": 8.195212364196777, + "learning_rate": 2.304122990915444e-05, + "loss": 1.3403, + "step": 4980 + }, + { + "epoch": 1.1620866325104797, + "grad_norm": 41.242862701416016, + "learning_rate": 2.3027253668763104e-05, + "loss": 2.0296, + "step": 4990 + }, + { + "epoch": 1.1644154634373545, + "grad_norm": 15.134632110595703, + "learning_rate": 2.301327742837177e-05, + "loss": 1.4878, + "step": 5000 + }, + { + "epoch": 1.1667442943642292, + "grad_norm": 16.87616729736328, + "learning_rate": 2.2999301187980434e-05, + "loss": 1.0683, + "step": 5010 + }, + { + "epoch": 1.1690731252911037, + "grad_norm": 10.382403373718262, + "learning_rate": 2.29853249475891e-05, + "loss": 1.3711, + "step": 5020 + }, + { + "epoch": 1.1714019562179785, + "grad_norm": 10.93367862701416, + "learning_rate": 2.2971348707197765e-05, + "loss": 1.4611, + "step": 5030 + }, + { + "epoch": 1.1737307871448532, + "grad_norm": 30.08503532409668, + "learning_rate": 2.2957372466806432e-05, + "loss": 1.4648, + "step": 5040 + }, + { + "epoch": 1.176059618071728, + "grad_norm": 7.187095642089844, + "learning_rate": 2.2943396226415095e-05, + "loss": 1.407, + "step": 5050 + }, + { + "epoch": 1.1783884489986027, + "grad_norm": 22.837495803833008, + "learning_rate": 2.292941998602376e-05, + "loss": 1.284, + "step": 5060 + }, + { + "epoch": 1.1807172799254775, + "grad_norm": 34.510986328125, + "learning_rate": 2.2915443745632423e-05, + "loss": 1.0948, + "step": 5070 + }, + { + "epoch": 1.1830461108523522, + "grad_norm": 10.157007217407227, + "learning_rate": 2.290146750524109e-05, + "loss": 1.0363, + "step": 5080 + }, + { + "epoch": 1.1853749417792268, + "grad_norm": 4.127469539642334, + "learning_rate": 2.2887491264849753e-05, + "loss": 1.1315, + "step": 5090 + }, + { + "epoch": 1.1877037727061015, + "grad_norm": 8.76297378540039, + "learning_rate": 2.287351502445842e-05, + "loss": 1.2935, + "step": 5100 + }, + { + "epoch": 1.1900326036329762, + "grad_norm": 58.847904205322266, + "learning_rate": 2.2859538784067087e-05, + "loss": 2.1187, + "step": 5110 + }, + { + "epoch": 1.192361434559851, + "grad_norm": 6.2726030349731445, + "learning_rate": 2.284556254367575e-05, + "loss": 1.0023, + "step": 5120 + }, + { + "epoch": 1.1946902654867257, + "grad_norm": 13.07908821105957, + "learning_rate": 2.2831586303284418e-05, + "loss": 1.069, + "step": 5130 + }, + { + "epoch": 1.1970190964136003, + "grad_norm": 12.537034034729004, + "learning_rate": 2.281761006289308e-05, + "loss": 1.1431, + "step": 5140 + }, + { + "epoch": 1.199347927340475, + "grad_norm": 7.794961452484131, + "learning_rate": 2.2803633822501748e-05, + "loss": 1.2082, + "step": 5150 + }, + { + "epoch": 1.2016767582673498, + "grad_norm": 72.09727478027344, + "learning_rate": 2.2789657582110412e-05, + "loss": 1.2637, + "step": 5160 + }, + { + "epoch": 1.2040055891942245, + "grad_norm": 12.098929405212402, + "learning_rate": 2.277568134171908e-05, + "loss": 1.671, + "step": 5170 + }, + { + "epoch": 1.2063344201210993, + "grad_norm": 13.814184188842773, + "learning_rate": 2.2761705101327742e-05, + "loss": 1.1143, + "step": 5180 + }, + { + "epoch": 1.208663251047974, + "grad_norm": 9.710376739501953, + "learning_rate": 2.274772886093641e-05, + "loss": 1.1579, + "step": 5190 + }, + { + "epoch": 1.2109920819748485, + "grad_norm": 10.720894813537598, + "learning_rate": 2.2733752620545073e-05, + "loss": 0.9838, + "step": 5200 + }, + { + "epoch": 1.2133209129017233, + "grad_norm": 32.7349739074707, + "learning_rate": 2.271977638015374e-05, + "loss": 2.0061, + "step": 5210 + }, + { + "epoch": 1.215649743828598, + "grad_norm": 12.79881763458252, + "learning_rate": 2.2705800139762403e-05, + "loss": 1.2333, + "step": 5220 + }, + { + "epoch": 1.2179785747554728, + "grad_norm": 13.296406745910645, + "learning_rate": 2.269182389937107e-05, + "loss": 1.3174, + "step": 5230 + }, + { + "epoch": 1.2203074056823475, + "grad_norm": 9.031120300292969, + "learning_rate": 2.2677847658979734e-05, + "loss": 0.9053, + "step": 5240 + }, + { + "epoch": 1.222636236609222, + "grad_norm": 7.792122840881348, + "learning_rate": 2.26638714185884e-05, + "loss": 1.2156, + "step": 5250 + }, + { + "epoch": 1.2249650675360968, + "grad_norm": 15.26120662689209, + "learning_rate": 2.2649895178197068e-05, + "loss": 1.9811, + "step": 5260 + }, + { + "epoch": 1.2272938984629715, + "grad_norm": 18.058717727661133, + "learning_rate": 2.263591893780573e-05, + "loss": 1.1064, + "step": 5270 + }, + { + "epoch": 1.2296227293898463, + "grad_norm": 57.864341735839844, + "learning_rate": 2.26219426974144e-05, + "loss": 1.9861, + "step": 5280 + }, + { + "epoch": 1.231951560316721, + "grad_norm": 6.74455451965332, + "learning_rate": 2.2607966457023062e-05, + "loss": 1.1326, + "step": 5290 + }, + { + "epoch": 1.2342803912435958, + "grad_norm": 6.738804817199707, + "learning_rate": 2.259399021663173e-05, + "loss": 2.296, + "step": 5300 + }, + { + "epoch": 1.2366092221704705, + "grad_norm": 10.682368278503418, + "learning_rate": 2.2580013976240393e-05, + "loss": 1.7348, + "step": 5310 + }, + { + "epoch": 1.238938053097345, + "grad_norm": 9.553335189819336, + "learning_rate": 2.256603773584906e-05, + "loss": 1.1105, + "step": 5320 + }, + { + "epoch": 1.2412668840242198, + "grad_norm": 14.826432228088379, + "learning_rate": 2.255206149545772e-05, + "loss": 1.7976, + "step": 5330 + }, + { + "epoch": 1.2435957149510946, + "grad_norm": 13.372576713562012, + "learning_rate": 2.2538085255066387e-05, + "loss": 1.2633, + "step": 5340 + }, + { + "epoch": 1.2459245458779693, + "grad_norm": 14.81163215637207, + "learning_rate": 2.252410901467505e-05, + "loss": 1.2059, + "step": 5350 + }, + { + "epoch": 1.248253376804844, + "grad_norm": 9.57998275756836, + "learning_rate": 2.2510132774283717e-05, + "loss": 1.2529, + "step": 5360 + }, + { + "epoch": 1.2505822077317186, + "grad_norm": 23.91578483581543, + "learning_rate": 2.249615653389238e-05, + "loss": 1.8046, + "step": 5370 + }, + { + "epoch": 1.2529110386585933, + "grad_norm": 8.844199180603027, + "learning_rate": 2.2482180293501048e-05, + "loss": 1.0261, + "step": 5380 + }, + { + "epoch": 1.255239869585468, + "grad_norm": 7.080227851867676, + "learning_rate": 2.246820405310971e-05, + "loss": 1.0358, + "step": 5390 + }, + { + "epoch": 1.2575687005123428, + "grad_norm": 9.825217247009277, + "learning_rate": 2.245422781271838e-05, + "loss": 1.0371, + "step": 5400 + }, + { + "epoch": 1.2598975314392176, + "grad_norm": 51.84904098510742, + "learning_rate": 2.2440251572327046e-05, + "loss": 1.2432, + "step": 5410 + }, + { + "epoch": 1.2622263623660923, + "grad_norm": 6.208728313446045, + "learning_rate": 2.242627533193571e-05, + "loss": 0.8359, + "step": 5420 + }, + { + "epoch": 1.264555193292967, + "grad_norm": 8.808072090148926, + "learning_rate": 2.2412299091544376e-05, + "loss": 1.3509, + "step": 5430 + }, + { + "epoch": 1.2668840242198416, + "grad_norm": 63.0085334777832, + "learning_rate": 2.239832285115304e-05, + "loss": 1.1609, + "step": 5440 + }, + { + "epoch": 1.2692128551467163, + "grad_norm": 13.395509719848633, + "learning_rate": 2.2384346610761707e-05, + "loss": 1.8954, + "step": 5450 + }, + { + "epoch": 1.271541686073591, + "grad_norm": 42.68526840209961, + "learning_rate": 2.237037037037037e-05, + "loss": 1.2054, + "step": 5460 + }, + { + "epoch": 1.2738705170004658, + "grad_norm": 5.643659591674805, + "learning_rate": 2.2356394129979037e-05, + "loss": 1.3347, + "step": 5470 + }, + { + "epoch": 1.2761993479273404, + "grad_norm": 19.961103439331055, + "learning_rate": 2.23424178895877e-05, + "loss": 1.7996, + "step": 5480 + }, + { + "epoch": 1.278528178854215, + "grad_norm": 11.5109224319458, + "learning_rate": 2.2328441649196368e-05, + "loss": 0.8807, + "step": 5490 + }, + { + "epoch": 1.2808570097810899, + "grad_norm": 4.86500883102417, + "learning_rate": 2.231446540880503e-05, + "loss": 0.9854, + "step": 5500 + }, + { + "epoch": 1.2831858407079646, + "grad_norm": 5.852017402648926, + "learning_rate": 2.23004891684137e-05, + "loss": 1.3599, + "step": 5510 + }, + { + "epoch": 1.2855146716348393, + "grad_norm": 31.3754825592041, + "learning_rate": 2.2286512928022362e-05, + "loss": 1.3582, + "step": 5520 + }, + { + "epoch": 1.287843502561714, + "grad_norm": 16.57115364074707, + "learning_rate": 2.227253668763103e-05, + "loss": 1.4891, + "step": 5530 + }, + { + "epoch": 1.2901723334885888, + "grad_norm": 8.07555103302002, + "learning_rate": 2.2258560447239693e-05, + "loss": 1.1457, + "step": 5540 + }, + { + "epoch": 1.2925011644154634, + "grad_norm": 51.47915267944336, + "learning_rate": 2.224458420684836e-05, + "loss": 1.2228, + "step": 5550 + }, + { + "epoch": 1.2948299953423381, + "grad_norm": 61.57811737060547, + "learning_rate": 2.2230607966457026e-05, + "loss": 1.4929, + "step": 5560 + }, + { + "epoch": 1.2971588262692129, + "grad_norm": 61.08300018310547, + "learning_rate": 2.221663172606569e-05, + "loss": 1.2155, + "step": 5570 + }, + { + "epoch": 1.2994876571960876, + "grad_norm": 124.03094482421875, + "learning_rate": 2.2202655485674357e-05, + "loss": 1.7477, + "step": 5580 + }, + { + "epoch": 1.3018164881229621, + "grad_norm": 7.754411220550537, + "learning_rate": 2.2188679245283017e-05, + "loss": 1.252, + "step": 5590 + }, + { + "epoch": 1.3041453190498369, + "grad_norm": 14.382433891296387, + "learning_rate": 2.2174703004891684e-05, + "loss": 1.3252, + "step": 5600 + }, + { + "epoch": 1.3064741499767116, + "grad_norm": 15.235672950744629, + "learning_rate": 2.2160726764500348e-05, + "loss": 1.0168, + "step": 5610 + }, + { + "epoch": 1.3088029809035864, + "grad_norm": 17.89712905883789, + "learning_rate": 2.2146750524109015e-05, + "loss": 1.451, + "step": 5620 + }, + { + "epoch": 1.3111318118304611, + "grad_norm": 38.380924224853516, + "learning_rate": 2.213277428371768e-05, + "loss": 1.2927, + "step": 5630 + }, + { + "epoch": 1.3134606427573359, + "grad_norm": 13.725621223449707, + "learning_rate": 2.2118798043326345e-05, + "loss": 1.2255, + "step": 5640 + }, + { + "epoch": 1.3157894736842106, + "grad_norm": 16.346235275268555, + "learning_rate": 2.210482180293501e-05, + "loss": 1.1837, + "step": 5650 + }, + { + "epoch": 1.3181183046110854, + "grad_norm": 15.437034606933594, + "learning_rate": 2.2090845562543676e-05, + "loss": 1.4109, + "step": 5660 + }, + { + "epoch": 1.32044713553796, + "grad_norm": 13.442436218261719, + "learning_rate": 2.207686932215234e-05, + "loss": 1.2536, + "step": 5670 + }, + { + "epoch": 1.3227759664648346, + "grad_norm": 21.472286224365234, + "learning_rate": 2.2062893081761007e-05, + "loss": 1.6111, + "step": 5680 + }, + { + "epoch": 1.3251047973917094, + "grad_norm": 13.40835189819336, + "learning_rate": 2.2048916841369673e-05, + "loss": 1.5086, + "step": 5690 + }, + { + "epoch": 1.3274336283185841, + "grad_norm": 9.588897705078125, + "learning_rate": 2.2034940600978337e-05, + "loss": 1.1776, + "step": 5700 + }, + { + "epoch": 1.3297624592454587, + "grad_norm": 14.10551643371582, + "learning_rate": 2.2020964360587004e-05, + "loss": 1.3947, + "step": 5710 + }, + { + "epoch": 1.3320912901723334, + "grad_norm": 27.368440628051758, + "learning_rate": 2.2006988120195668e-05, + "loss": 1.139, + "step": 5720 + }, + { + "epoch": 1.3344201210992082, + "grad_norm": 8.45453929901123, + "learning_rate": 2.1993011879804335e-05, + "loss": 1.7411, + "step": 5730 + }, + { + "epoch": 1.336748952026083, + "grad_norm": 17.110980987548828, + "learning_rate": 2.1979035639412998e-05, + "loss": 1.9253, + "step": 5740 + }, + { + "epoch": 1.3390777829529577, + "grad_norm": 94.51256561279297, + "learning_rate": 2.1965059399021665e-05, + "loss": 1.4818, + "step": 5750 + }, + { + "epoch": 1.3414066138798324, + "grad_norm": 7.826664924621582, + "learning_rate": 2.195108315863033e-05, + "loss": 0.8781, + "step": 5760 + }, + { + "epoch": 1.3437354448067071, + "grad_norm": 7.50213623046875, + "learning_rate": 2.1937106918238996e-05, + "loss": 1.1239, + "step": 5770 + }, + { + "epoch": 1.3460642757335817, + "grad_norm": 31.289094924926758, + "learning_rate": 2.192313067784766e-05, + "loss": 1.385, + "step": 5780 + }, + { + "epoch": 1.3483931066604564, + "grad_norm": 59.284637451171875, + "learning_rate": 2.1909154437456326e-05, + "loss": 1.3212, + "step": 5790 + }, + { + "epoch": 1.3507219375873312, + "grad_norm": 6.5409040451049805, + "learning_rate": 2.189517819706499e-05, + "loss": 1.13, + "step": 5800 + }, + { + "epoch": 1.353050768514206, + "grad_norm": 5.605960369110107, + "learning_rate": 2.1881201956673657e-05, + "loss": 0.9533, + "step": 5810 + }, + { + "epoch": 1.3553795994410804, + "grad_norm": 15.074896812438965, + "learning_rate": 2.186722571628232e-05, + "loss": 1.5392, + "step": 5820 + }, + { + "epoch": 1.3577084303679552, + "grad_norm": 22.98057746887207, + "learning_rate": 2.1853249475890987e-05, + "loss": 1.467, + "step": 5830 + }, + { + "epoch": 1.36003726129483, + "grad_norm": 17.14531898498535, + "learning_rate": 2.183927323549965e-05, + "loss": 1.4293, + "step": 5840 + }, + { + "epoch": 1.3623660922217047, + "grad_norm": 11.25844955444336, + "learning_rate": 2.1825296995108315e-05, + "loss": 1.0427, + "step": 5850 + }, + { + "epoch": 1.3646949231485794, + "grad_norm": 9.811922073364258, + "learning_rate": 2.181132075471698e-05, + "loss": 1.5446, + "step": 5860 + }, + { + "epoch": 1.3670237540754542, + "grad_norm": 15.156676292419434, + "learning_rate": 2.1797344514325645e-05, + "loss": 0.7154, + "step": 5870 + }, + { + "epoch": 1.369352585002329, + "grad_norm": 10.407319068908691, + "learning_rate": 2.1783368273934312e-05, + "loss": 1.5355, + "step": 5880 + }, + { + "epoch": 1.3716814159292037, + "grad_norm": 28.042938232421875, + "learning_rate": 2.1769392033542976e-05, + "loss": 1.6673, + "step": 5890 + }, + { + "epoch": 1.3740102468560782, + "grad_norm": 6.387546062469482, + "learning_rate": 2.1755415793151643e-05, + "loss": 1.2251, + "step": 5900 + }, + { + "epoch": 1.376339077782953, + "grad_norm": 14.928718566894531, + "learning_rate": 2.1741439552760306e-05, + "loss": 1.0641, + "step": 5910 + }, + { + "epoch": 1.3786679087098277, + "grad_norm": 20.759645462036133, + "learning_rate": 2.1727463312368973e-05, + "loss": 1.6231, + "step": 5920 + }, + { + "epoch": 1.3809967396367024, + "grad_norm": 91.47080993652344, + "learning_rate": 2.1713487071977637e-05, + "loss": 1.541, + "step": 5930 + }, + { + "epoch": 1.383325570563577, + "grad_norm": 121.2904281616211, + "learning_rate": 2.1699510831586304e-05, + "loss": 1.3784, + "step": 5940 + }, + { + "epoch": 1.3856544014904517, + "grad_norm": 21.550537109375, + "learning_rate": 2.1685534591194967e-05, + "loss": 1.221, + "step": 5950 + }, + { + "epoch": 1.3879832324173265, + "grad_norm": 22.803874969482422, + "learning_rate": 2.1671558350803634e-05, + "loss": 1.239, + "step": 5960 + }, + { + "epoch": 1.3903120633442012, + "grad_norm": 3.7256100177764893, + "learning_rate": 2.1657582110412298e-05, + "loss": 0.8911, + "step": 5970 + }, + { + "epoch": 1.392640894271076, + "grad_norm": 6.628619194030762, + "learning_rate": 2.1643605870020965e-05, + "loss": 1.2682, + "step": 5980 + }, + { + "epoch": 1.3949697251979507, + "grad_norm": 13.052042961120605, + "learning_rate": 2.1629629629629632e-05, + "loss": 1.2541, + "step": 5990 + }, + { + "epoch": 1.3972985561248255, + "grad_norm": 12.446592330932617, + "learning_rate": 2.1615653389238296e-05, + "loss": 1.2242, + "step": 6000 + }, + { + "epoch": 1.3996273870517, + "grad_norm": 15.783132553100586, + "learning_rate": 2.1601677148846963e-05, + "loss": 1.069, + "step": 6010 + }, + { + "epoch": 1.4019562179785747, + "grad_norm": 19.638019561767578, + "learning_rate": 2.1587700908455626e-05, + "loss": 1.039, + "step": 6020 + }, + { + "epoch": 1.4042850489054495, + "grad_norm": 17.819883346557617, + "learning_rate": 2.1573724668064293e-05, + "loss": 1.0184, + "step": 6030 + }, + { + "epoch": 1.4066138798323242, + "grad_norm": 14.201866149902344, + "learning_rate": 2.1559748427672957e-05, + "loss": 1.3285, + "step": 6040 + }, + { + "epoch": 1.4089427107591987, + "grad_norm": 9.433976173400879, + "learning_rate": 2.1545772187281624e-05, + "loss": 0.9471, + "step": 6050 + }, + { + "epoch": 1.4112715416860735, + "grad_norm": 19.676637649536133, + "learning_rate": 2.1531795946890287e-05, + "loss": 0.9991, + "step": 6060 + }, + { + "epoch": 1.4136003726129482, + "grad_norm": 7.89212703704834, + "learning_rate": 2.1517819706498954e-05, + "loss": 1.157, + "step": 6070 + }, + { + "epoch": 1.415929203539823, + "grad_norm": 8.475332260131836, + "learning_rate": 2.1503843466107618e-05, + "loss": 1.8082, + "step": 6080 + }, + { + "epoch": 1.4182580344666977, + "grad_norm": 30.952489852905273, + "learning_rate": 2.1489867225716285e-05, + "loss": 1.3291, + "step": 6090 + }, + { + "epoch": 1.4205868653935725, + "grad_norm": 41.84647750854492, + "learning_rate": 2.147589098532495e-05, + "loss": 1.4919, + "step": 6100 + }, + { + "epoch": 1.4229156963204472, + "grad_norm": 58.389835357666016, + "learning_rate": 2.1461914744933615e-05, + "loss": 1.8779, + "step": 6110 + }, + { + "epoch": 1.425244527247322, + "grad_norm": 73.64154052734375, + "learning_rate": 2.1447938504542276e-05, + "loss": 1.2146, + "step": 6120 + }, + { + "epoch": 1.4275733581741965, + "grad_norm": 10.781891822814941, + "learning_rate": 2.1433962264150943e-05, + "loss": 1.1101, + "step": 6130 + }, + { + "epoch": 1.4299021891010713, + "grad_norm": 12.203949928283691, + "learning_rate": 2.141998602375961e-05, + "loss": 1.1829, + "step": 6140 + }, + { + "epoch": 1.432231020027946, + "grad_norm": 8.079778671264648, + "learning_rate": 2.1406009783368273e-05, + "loss": 1.6185, + "step": 6150 + }, + { + "epoch": 1.4345598509548207, + "grad_norm": 16.477624893188477, + "learning_rate": 2.139203354297694e-05, + "loss": 1.5992, + "step": 6160 + }, + { + "epoch": 1.4368886818816953, + "grad_norm": 20.183237075805664, + "learning_rate": 2.1378057302585604e-05, + "loss": 1.7741, + "step": 6170 + }, + { + "epoch": 1.43921751280857, + "grad_norm": 14.812673568725586, + "learning_rate": 2.136408106219427e-05, + "loss": 1.2379, + "step": 6180 + }, + { + "epoch": 1.4415463437354448, + "grad_norm": 12.094216346740723, + "learning_rate": 2.1350104821802934e-05, + "loss": 1.7963, + "step": 6190 + }, + { + "epoch": 1.4438751746623195, + "grad_norm": 13.727142333984375, + "learning_rate": 2.13361285814116e-05, + "loss": 1.2336, + "step": 6200 + }, + { + "epoch": 1.4462040055891943, + "grad_norm": 12.991110801696777, + "learning_rate": 2.1322152341020265e-05, + "loss": 1.0751, + "step": 6210 + }, + { + "epoch": 1.448532836516069, + "grad_norm": 13.830179214477539, + "learning_rate": 2.1308176100628932e-05, + "loss": 1.4648, + "step": 6220 + }, + { + "epoch": 1.4508616674429438, + "grad_norm": 11.587153434753418, + "learning_rate": 2.1294199860237595e-05, + "loss": 1.4138, + "step": 6230 + }, + { + "epoch": 1.4531904983698183, + "grad_norm": 7.644782543182373, + "learning_rate": 2.1280223619846262e-05, + "loss": 1.0203, + "step": 6240 + }, + { + "epoch": 1.455519329296693, + "grad_norm": 12.32254695892334, + "learning_rate": 2.1266247379454926e-05, + "loss": 1.598, + "step": 6250 + }, + { + "epoch": 1.4578481602235678, + "grad_norm": 3.885669469833374, + "learning_rate": 2.1252271139063593e-05, + "loss": 1.4226, + "step": 6260 + }, + { + "epoch": 1.4601769911504425, + "grad_norm": 27.064685821533203, + "learning_rate": 2.1238294898672257e-05, + "loss": 1.4802, + "step": 6270 + }, + { + "epoch": 1.462505822077317, + "grad_norm": 14.793761253356934, + "learning_rate": 2.1224318658280923e-05, + "loss": 1.0699, + "step": 6280 + }, + { + "epoch": 1.4648346530041918, + "grad_norm": 9.927505493164062, + "learning_rate": 2.121034241788959e-05, + "loss": 1.0639, + "step": 6290 + }, + { + "epoch": 1.4671634839310665, + "grad_norm": 10.051878929138184, + "learning_rate": 2.1196366177498254e-05, + "loss": 0.8853, + "step": 6300 + }, + { + "epoch": 1.4694923148579413, + "grad_norm": 12.47751522064209, + "learning_rate": 2.118238993710692e-05, + "loss": 1.3319, + "step": 6310 + }, + { + "epoch": 1.471821145784816, + "grad_norm": 7.797066688537598, + "learning_rate": 2.1168413696715585e-05, + "loss": 0.914, + "step": 6320 + }, + { + "epoch": 1.4741499767116908, + "grad_norm": 14.888928413391113, + "learning_rate": 2.115443745632425e-05, + "loss": 1.3441, + "step": 6330 + }, + { + "epoch": 1.4764788076385655, + "grad_norm": 7.610679626464844, + "learning_rate": 2.1140461215932915e-05, + "loss": 1.0514, + "step": 6340 + }, + { + "epoch": 1.4788076385654403, + "grad_norm": 6.923398971557617, + "learning_rate": 2.1126484975541582e-05, + "loss": 1.6134, + "step": 6350 + }, + { + "epoch": 1.4811364694923148, + "grad_norm": 24.528564453125, + "learning_rate": 2.1112508735150246e-05, + "loss": 1.5473, + "step": 6360 + }, + { + "epoch": 1.4834653004191896, + "grad_norm": 4.077647686004639, + "learning_rate": 2.1098532494758913e-05, + "loss": 1.2879, + "step": 6370 + }, + { + "epoch": 1.4857941313460643, + "grad_norm": 13.990240097045898, + "learning_rate": 2.1084556254367573e-05, + "loss": 1.6288, + "step": 6380 + }, + { + "epoch": 1.488122962272939, + "grad_norm": 19.50718116760254, + "learning_rate": 2.107058001397624e-05, + "loss": 1.522, + "step": 6390 + }, + { + "epoch": 1.4904517931998136, + "grad_norm": 11.860418319702148, + "learning_rate": 2.1056603773584904e-05, + "loss": 1.2409, + "step": 6400 + }, + { + "epoch": 1.4927806241266883, + "grad_norm": 8.834514617919922, + "learning_rate": 2.104262753319357e-05, + "loss": 1.1322, + "step": 6410 + }, + { + "epoch": 1.495109455053563, + "grad_norm": 12.05262565612793, + "learning_rate": 2.1028651292802234e-05, + "loss": 1.2565, + "step": 6420 + }, + { + "epoch": 1.4974382859804378, + "grad_norm": 47.0871696472168, + "learning_rate": 2.10146750524109e-05, + "loss": 1.6558, + "step": 6430 + }, + { + "epoch": 1.4997671169073126, + "grad_norm": 25.88595962524414, + "learning_rate": 2.1000698812019568e-05, + "loss": 1.4326, + "step": 6440 + }, + { + "epoch": 1.5020959478341873, + "grad_norm": 21.97953224182129, + "learning_rate": 2.098672257162823e-05, + "loss": 1.4276, + "step": 6450 + }, + { + "epoch": 1.504424778761062, + "grad_norm": 6.62868595123291, + "learning_rate": 2.09727463312369e-05, + "loss": 1.4069, + "step": 6460 + }, + { + "epoch": 1.5067536096879368, + "grad_norm": 16.54710578918457, + "learning_rate": 2.0958770090845562e-05, + "loss": 1.2972, + "step": 6470 + }, + { + "epoch": 1.5090824406148113, + "grad_norm": 14.595316886901855, + "learning_rate": 2.094479385045423e-05, + "loss": 1.2964, + "step": 6480 + }, + { + "epoch": 1.511411271541686, + "grad_norm": 29.083581924438477, + "learning_rate": 2.0930817610062893e-05, + "loss": 1.6111, + "step": 6490 + }, + { + "epoch": 1.5137401024685608, + "grad_norm": 17.463119506835938, + "learning_rate": 2.091684136967156e-05, + "loss": 0.8385, + "step": 6500 + }, + { + "epoch": 1.5160689333954354, + "grad_norm": 11.917354583740234, + "learning_rate": 2.0902865129280223e-05, + "loss": 0.9631, + "step": 6510 + }, + { + "epoch": 1.51839776432231, + "grad_norm": 8.570207595825195, + "learning_rate": 2.088888888888889e-05, + "loss": 1.4749, + "step": 6520 + }, + { + "epoch": 1.5207265952491849, + "grad_norm": 6.944149971008301, + "learning_rate": 2.0874912648497554e-05, + "loss": 1.1676, + "step": 6530 + }, + { + "epoch": 1.5230554261760596, + "grad_norm": 19.165542602539062, + "learning_rate": 2.086093640810622e-05, + "loss": 1.7928, + "step": 6540 + }, + { + "epoch": 1.5253842571029343, + "grad_norm": 5.844601154327393, + "learning_rate": 2.0846960167714884e-05, + "loss": 1.0746, + "step": 6550 + }, + { + "epoch": 1.527713088029809, + "grad_norm": 79.03755187988281, + "learning_rate": 2.083298392732355e-05, + "loss": 0.9438, + "step": 6560 + }, + { + "epoch": 1.5300419189566838, + "grad_norm": 70.38703155517578, + "learning_rate": 2.0819007686932215e-05, + "loss": 1.2571, + "step": 6570 + }, + { + "epoch": 1.5323707498835586, + "grad_norm": 8.431889533996582, + "learning_rate": 2.0805031446540882e-05, + "loss": 1.53, + "step": 6580 + }, + { + "epoch": 1.5346995808104331, + "grad_norm": 13.615509033203125, + "learning_rate": 2.079105520614955e-05, + "loss": 1.0836, + "step": 6590 + }, + { + "epoch": 1.5370284117373079, + "grad_norm": 10.16684341430664, + "learning_rate": 2.0777078965758213e-05, + "loss": 1.4537, + "step": 6600 + }, + { + "epoch": 1.5393572426641826, + "grad_norm": 62.860252380371094, + "learning_rate": 2.076310272536688e-05, + "loss": 1.6445, + "step": 6610 + }, + { + "epoch": 1.5416860735910571, + "grad_norm": 10.388005256652832, + "learning_rate": 2.0749126484975543e-05, + "loss": 1.8212, + "step": 6620 + }, + { + "epoch": 1.5440149045179319, + "grad_norm": 15.161338806152344, + "learning_rate": 2.073515024458421e-05, + "loss": 1.5165, + "step": 6630 + }, + { + "epoch": 1.5463437354448066, + "grad_norm": 10.338879585266113, + "learning_rate": 2.072117400419287e-05, + "loss": 0.9325, + "step": 6640 + }, + { + "epoch": 1.5486725663716814, + "grad_norm": 50.81422424316406, + "learning_rate": 2.0707197763801537e-05, + "loss": 1.3275, + "step": 6650 + }, + { + "epoch": 1.5510013972985561, + "grad_norm": 63.32241439819336, + "learning_rate": 2.06932215234102e-05, + "loss": 1.6739, + "step": 6660 + }, + { + "epoch": 1.5533302282254309, + "grad_norm": 12.45587158203125, + "learning_rate": 2.0679245283018868e-05, + "loss": 1.6098, + "step": 6670 + }, + { + "epoch": 1.5556590591523056, + "grad_norm": 17.617839813232422, + "learning_rate": 2.066526904262753e-05, + "loss": 1.3134, + "step": 6680 + }, + { + "epoch": 1.5579878900791804, + "grad_norm": 4.559420108795166, + "learning_rate": 2.06512928022362e-05, + "loss": 1.0789, + "step": 6690 + }, + { + "epoch": 1.5603167210060551, + "grad_norm": 5.019169330596924, + "learning_rate": 2.0637316561844862e-05, + "loss": 1.186, + "step": 6700 + }, + { + "epoch": 1.5626455519329296, + "grad_norm": 47.28404998779297, + "learning_rate": 2.062334032145353e-05, + "loss": 2.23, + "step": 6710 + }, + { + "epoch": 1.5649743828598044, + "grad_norm": 8.204998970031738, + "learning_rate": 2.0609364081062193e-05, + "loss": 1.5562, + "step": 6720 + }, + { + "epoch": 1.5673032137866791, + "grad_norm": 12.214827537536621, + "learning_rate": 2.059538784067086e-05, + "loss": 1.2355, + "step": 6730 + }, + { + "epoch": 1.5696320447135537, + "grad_norm": 92.05329895019531, + "learning_rate": 2.0581411600279527e-05, + "loss": 1.6341, + "step": 6740 + }, + { + "epoch": 1.5719608756404284, + "grad_norm": 9.725508689880371, + "learning_rate": 2.056743535988819e-05, + "loss": 1.6041, + "step": 6750 + }, + { + "epoch": 1.5742897065673032, + "grad_norm": 5.3793134689331055, + "learning_rate": 2.0553459119496857e-05, + "loss": 0.8478, + "step": 6760 + }, + { + "epoch": 1.576618537494178, + "grad_norm": 7.88288688659668, + "learning_rate": 2.053948287910552e-05, + "loss": 1.0917, + "step": 6770 + }, + { + "epoch": 1.5789473684210527, + "grad_norm": 59.44280242919922, + "learning_rate": 2.0525506638714188e-05, + "loss": 1.3453, + "step": 6780 + }, + { + "epoch": 1.5812761993479274, + "grad_norm": 18.248361587524414, + "learning_rate": 2.051153039832285e-05, + "loss": 1.3135, + "step": 6790 + }, + { + "epoch": 1.5836050302748022, + "grad_norm": 4.527297496795654, + "learning_rate": 2.0497554157931518e-05, + "loss": 1.6539, + "step": 6800 + }, + { + "epoch": 1.585933861201677, + "grad_norm": 6.267810821533203, + "learning_rate": 2.0483577917540182e-05, + "loss": 1.2551, + "step": 6810 + }, + { + "epoch": 1.5882626921285514, + "grad_norm": 9.992630958557129, + "learning_rate": 2.046960167714885e-05, + "loss": 0.78, + "step": 6820 + }, + { + "epoch": 1.5905915230554262, + "grad_norm": 6.590550899505615, + "learning_rate": 2.0455625436757512e-05, + "loss": 1.3155, + "step": 6830 + }, + { + "epoch": 1.592920353982301, + "grad_norm": 5.96619176864624, + "learning_rate": 2.044164919636618e-05, + "loss": 1.5254, + "step": 6840 + }, + { + "epoch": 1.5952491849091754, + "grad_norm": 15.359123229980469, + "learning_rate": 2.0427672955974843e-05, + "loss": 1.7816, + "step": 6850 + }, + { + "epoch": 1.5975780158360502, + "grad_norm": 7.312784671783447, + "learning_rate": 2.041369671558351e-05, + "loss": 1.0621, + "step": 6860 + }, + { + "epoch": 1.599906846762925, + "grad_norm": 3.967820167541504, + "learning_rate": 2.0399720475192174e-05, + "loss": 1.3336, + "step": 6870 + }, + { + "epoch": 1.6022356776897997, + "grad_norm": 10.185931205749512, + "learning_rate": 2.038574423480084e-05, + "loss": 1.0566, + "step": 6880 + }, + { + "epoch": 1.6045645086166744, + "grad_norm": 7.750170707702637, + "learning_rate": 2.0371767994409507e-05, + "loss": 1.0334, + "step": 6890 + }, + { + "epoch": 1.6068933395435492, + "grad_norm": 11.731595993041992, + "learning_rate": 2.035779175401817e-05, + "loss": 0.8421, + "step": 6900 + }, + { + "epoch": 1.609222170470424, + "grad_norm": 17.454936981201172, + "learning_rate": 2.0343815513626835e-05, + "loss": 1.2394, + "step": 6910 + }, + { + "epoch": 1.6115510013972987, + "grad_norm": 4.5072712898254395, + "learning_rate": 2.0329839273235498e-05, + "loss": 1.1167, + "step": 6920 + }, + { + "epoch": 1.6138798323241734, + "grad_norm": 5.470495700836182, + "learning_rate": 2.0315863032844165e-05, + "loss": 1.2338, + "step": 6930 + }, + { + "epoch": 1.616208663251048, + "grad_norm": 12.268614768981934, + "learning_rate": 2.030188679245283e-05, + "loss": 1.5808, + "step": 6940 + }, + { + "epoch": 1.6185374941779227, + "grad_norm": 11.641376495361328, + "learning_rate": 2.0287910552061496e-05, + "loss": 1.1396, + "step": 6950 + }, + { + "epoch": 1.6208663251047974, + "grad_norm": 12.721565246582031, + "learning_rate": 2.027393431167016e-05, + "loss": 0.9674, + "step": 6960 + }, + { + "epoch": 1.623195156031672, + "grad_norm": 18.911191940307617, + "learning_rate": 2.0259958071278826e-05, + "loss": 1.4093, + "step": 6970 + }, + { + "epoch": 1.6255239869585467, + "grad_norm": 16.1356201171875, + "learning_rate": 2.024598183088749e-05, + "loss": 1.2923, + "step": 6980 + }, + { + "epoch": 1.6278528178854215, + "grad_norm": 5.580846786499023, + "learning_rate": 2.0232005590496157e-05, + "loss": 1.4068, + "step": 6990 + }, + { + "epoch": 1.6301816488122962, + "grad_norm": 10.02719783782959, + "learning_rate": 2.021802935010482e-05, + "loss": 1.3046, + "step": 7000 + }, + { + "epoch": 1.632510479739171, + "grad_norm": 10.013179779052734, + "learning_rate": 2.0204053109713487e-05, + "loss": 1.3374, + "step": 7010 + }, + { + "epoch": 1.6348393106660457, + "grad_norm": 7.2384724617004395, + "learning_rate": 2.019007686932215e-05, + "loss": 0.9352, + "step": 7020 + }, + { + "epoch": 1.6371681415929205, + "grad_norm": 15.23076343536377, + "learning_rate": 2.0176100628930818e-05, + "loss": 1.7616, + "step": 7030 + }, + { + "epoch": 1.6394969725197952, + "grad_norm": 35.012516021728516, + "learning_rate": 2.0162124388539485e-05, + "loss": 1.6217, + "step": 7040 + }, + { + "epoch": 1.6418258034466697, + "grad_norm": 7.518862724304199, + "learning_rate": 2.014814814814815e-05, + "loss": 0.77, + "step": 7050 + }, + { + "epoch": 1.6441546343735445, + "grad_norm": 8.740734100341797, + "learning_rate": 2.0134171907756816e-05, + "loss": 0.945, + "step": 7060 + }, + { + "epoch": 1.6464834653004192, + "grad_norm": 14.992202758789062, + "learning_rate": 2.012019566736548e-05, + "loss": 1.4889, + "step": 7070 + }, + { + "epoch": 1.6488122962272937, + "grad_norm": 9.385587692260742, + "learning_rate": 2.0106219426974146e-05, + "loss": 1.4533, + "step": 7080 + }, + { + "epoch": 1.6511411271541685, + "grad_norm": 3.7752981185913086, + "learning_rate": 2.009224318658281e-05, + "loss": 1.1922, + "step": 7090 + }, + { + "epoch": 1.6534699580810432, + "grad_norm": 9.964385986328125, + "learning_rate": 2.0078266946191477e-05, + "loss": 1.009, + "step": 7100 + }, + { + "epoch": 1.655798789007918, + "grad_norm": 22.4150333404541, + "learning_rate": 2.006429070580014e-05, + "loss": 1.3856, + "step": 7110 + }, + { + "epoch": 1.6581276199347927, + "grad_norm": 6.669826984405518, + "learning_rate": 2.0050314465408807e-05, + "loss": 1.1092, + "step": 7120 + }, + { + "epoch": 1.6604564508616675, + "grad_norm": 9.804742813110352, + "learning_rate": 2.003633822501747e-05, + "loss": 1.3718, + "step": 7130 + }, + { + "epoch": 1.6627852817885422, + "grad_norm": 9.104567527770996, + "learning_rate": 2.0022361984626138e-05, + "loss": 0.7321, + "step": 7140 + }, + { + "epoch": 1.665114112715417, + "grad_norm": 64.06851196289062, + "learning_rate": 2.00083857442348e-05, + "loss": 1.7271, + "step": 7150 + }, + { + "epoch": 1.6674429436422917, + "grad_norm": 12.741191864013672, + "learning_rate": 1.999440950384347e-05, + "loss": 1.6974, + "step": 7160 + }, + { + "epoch": 1.6697717745691663, + "grad_norm": 9.56039810180664, + "learning_rate": 1.998043326345213e-05, + "loss": 1.2862, + "step": 7170 + }, + { + "epoch": 1.672100605496041, + "grad_norm": 4.528858661651611, + "learning_rate": 1.9966457023060796e-05, + "loss": 1.1691, + "step": 7180 + }, + { + "epoch": 1.6744294364229158, + "grad_norm": 17.700504302978516, + "learning_rate": 1.9952480782669463e-05, + "loss": 1.1343, + "step": 7190 + }, + { + "epoch": 1.6767582673497903, + "grad_norm": 10.610898971557617, + "learning_rate": 1.9938504542278126e-05, + "loss": 1.5065, + "step": 7200 + }, + { + "epoch": 1.679087098276665, + "grad_norm": 4.640142917633057, + "learning_rate": 1.9924528301886793e-05, + "loss": 1.9951, + "step": 7210 + }, + { + "epoch": 1.6814159292035398, + "grad_norm": 8.147830963134766, + "learning_rate": 1.9910552061495457e-05, + "loss": 1.1698, + "step": 7220 + }, + { + "epoch": 1.6837447601304145, + "grad_norm": 36.66588592529297, + "learning_rate": 1.9896575821104124e-05, + "loss": 2.1156, + "step": 7230 + }, + { + "epoch": 1.6860735910572893, + "grad_norm": 18.592336654663086, + "learning_rate": 1.9882599580712787e-05, + "loss": 1.8785, + "step": 7240 + }, + { + "epoch": 1.688402421984164, + "grad_norm": 53.298561096191406, + "learning_rate": 1.9868623340321454e-05, + "loss": 1.2684, + "step": 7250 + }, + { + "epoch": 1.6907312529110388, + "grad_norm": 16.278640747070312, + "learning_rate": 1.9854647099930118e-05, + "loss": 1.2625, + "step": 7260 + }, + { + "epoch": 1.6930600838379135, + "grad_norm": 14.667279243469238, + "learning_rate": 1.9840670859538785e-05, + "loss": 1.2811, + "step": 7270 + }, + { + "epoch": 1.695388914764788, + "grad_norm": 8.197450637817383, + "learning_rate": 1.982669461914745e-05, + "loss": 1.451, + "step": 7280 + }, + { + "epoch": 1.6977177456916628, + "grad_norm": 43.80754852294922, + "learning_rate": 1.9812718378756115e-05, + "loss": 1.1298, + "step": 7290 + }, + { + "epoch": 1.7000465766185375, + "grad_norm": 7.743311405181885, + "learning_rate": 1.979874213836478e-05, + "loss": 1.2866, + "step": 7300 + }, + { + "epoch": 1.702375407545412, + "grad_norm": 51.69382858276367, + "learning_rate": 1.9784765897973446e-05, + "loss": 1.0107, + "step": 7310 + }, + { + "epoch": 1.7047042384722868, + "grad_norm": 7.2719573974609375, + "learning_rate": 1.977078965758211e-05, + "loss": 1.162, + "step": 7320 + }, + { + "epoch": 1.7070330693991616, + "grad_norm": 7.873298645019531, + "learning_rate": 1.9756813417190777e-05, + "loss": 1.7486, + "step": 7330 + }, + { + "epoch": 1.7093619003260363, + "grad_norm": 16.609840393066406, + "learning_rate": 1.9742837176799444e-05, + "loss": 1.2125, + "step": 7340 + }, + { + "epoch": 1.711690731252911, + "grad_norm": 6.402132034301758, + "learning_rate": 1.9728860936408107e-05, + "loss": 0.9309, + "step": 7350 + }, + { + "epoch": 1.7140195621797858, + "grad_norm": 3.3988287448883057, + "learning_rate": 1.9714884696016774e-05, + "loss": 1.2475, + "step": 7360 + }, + { + "epoch": 1.7163483931066605, + "grad_norm": 13.434524536132812, + "learning_rate": 1.9700908455625438e-05, + "loss": 1.1006, + "step": 7370 + }, + { + "epoch": 1.7186772240335353, + "grad_norm": 54.3780632019043, + "learning_rate": 1.9686932215234105e-05, + "loss": 1.2499, + "step": 7380 + }, + { + "epoch": 1.72100605496041, + "grad_norm": 5.294698715209961, + "learning_rate": 1.9672955974842768e-05, + "loss": 1.5747, + "step": 7390 + }, + { + "epoch": 1.7233348858872846, + "grad_norm": 11.490097045898438, + "learning_rate": 1.9658979734451435e-05, + "loss": 1.0938, + "step": 7400 + }, + { + "epoch": 1.7256637168141593, + "grad_norm": 19.641559600830078, + "learning_rate": 1.96450034940601e-05, + "loss": 2.3794, + "step": 7410 + }, + { + "epoch": 1.7279925477410338, + "grad_norm": 14.349048614501953, + "learning_rate": 1.9631027253668766e-05, + "loss": 1.4, + "step": 7420 + }, + { + "epoch": 1.7303213786679086, + "grad_norm": 5.6987528800964355, + "learning_rate": 1.9617051013277426e-05, + "loss": 1.1306, + "step": 7430 + }, + { + "epoch": 1.7326502095947833, + "grad_norm": 5.87094259262085, + "learning_rate": 1.9603074772886093e-05, + "loss": 1.3163, + "step": 7440 + }, + { + "epoch": 1.734979040521658, + "grad_norm": 16.617454528808594, + "learning_rate": 1.9589098532494757e-05, + "loss": 1.575, + "step": 7450 + }, + { + "epoch": 1.7373078714485328, + "grad_norm": 9.811037063598633, + "learning_rate": 1.9575122292103424e-05, + "loss": 1.4052, + "step": 7460 + }, + { + "epoch": 1.7396367023754076, + "grad_norm": 12.65397834777832, + "learning_rate": 1.9561146051712087e-05, + "loss": 1.4929, + "step": 7470 + }, + { + "epoch": 1.7419655333022823, + "grad_norm": 5.476722717285156, + "learning_rate": 1.9547169811320754e-05, + "loss": 0.8258, + "step": 7480 + }, + { + "epoch": 1.744294364229157, + "grad_norm": 13.89340877532959, + "learning_rate": 1.953319357092942e-05, + "loss": 1.4055, + "step": 7490 + }, + { + "epoch": 1.7466231951560318, + "grad_norm": 8.496129035949707, + "learning_rate": 1.9519217330538085e-05, + "loss": 1.3281, + "step": 7500 + }, + { + "epoch": 1.7489520260829063, + "grad_norm": 10.355660438537598, + "learning_rate": 1.950524109014675e-05, + "loss": 0.9894, + "step": 7510 + }, + { + "epoch": 1.751280857009781, + "grad_norm": 7.4711012840271, + "learning_rate": 1.9491264849755415e-05, + "loss": 1.3505, + "step": 7520 + }, + { + "epoch": 1.7536096879366558, + "grad_norm": 23.025426864624023, + "learning_rate": 1.9477288609364082e-05, + "loss": 1.0166, + "step": 7530 + }, + { + "epoch": 1.7559385188635304, + "grad_norm": 8.058009147644043, + "learning_rate": 1.9463312368972746e-05, + "loss": 1.1184, + "step": 7540 + }, + { + "epoch": 1.758267349790405, + "grad_norm": 23.26887321472168, + "learning_rate": 1.9449336128581413e-05, + "loss": 1.0111, + "step": 7550 + }, + { + "epoch": 1.7605961807172799, + "grad_norm": 7.433112144470215, + "learning_rate": 1.9435359888190076e-05, + "loss": 0.9874, + "step": 7560 + }, + { + "epoch": 1.7629250116441546, + "grad_norm": 10.770955085754395, + "learning_rate": 1.9421383647798743e-05, + "loss": 0.6591, + "step": 7570 + }, + { + "epoch": 1.7652538425710294, + "grad_norm": 7.087027549743652, + "learning_rate": 1.9407407407407407e-05, + "loss": 1.2804, + "step": 7580 + }, + { + "epoch": 1.767582673497904, + "grad_norm": 7.6507086753845215, + "learning_rate": 1.9393431167016074e-05, + "loss": 1.3398, + "step": 7590 + }, + { + "epoch": 1.7699115044247788, + "grad_norm": 6.744988918304443, + "learning_rate": 1.9379454926624738e-05, + "loss": 1.0609, + "step": 7600 + }, + { + "epoch": 1.7722403353516536, + "grad_norm": 63.14842224121094, + "learning_rate": 1.9365478686233404e-05, + "loss": 1.5501, + "step": 7610 + }, + { + "epoch": 1.7745691662785283, + "grad_norm": 28.729961395263672, + "learning_rate": 1.9351502445842068e-05, + "loss": 2.0076, + "step": 7620 + }, + { + "epoch": 1.7768979972054029, + "grad_norm": 16.338687896728516, + "learning_rate": 1.9337526205450735e-05, + "loss": 1.2513, + "step": 7630 + }, + { + "epoch": 1.7792268281322776, + "grad_norm": 8.51009750366211, + "learning_rate": 1.9323549965059402e-05, + "loss": 1.2085, + "step": 7640 + }, + { + "epoch": 1.7815556590591521, + "grad_norm": 145.62452697753906, + "learning_rate": 1.9309573724668066e-05, + "loss": 1.2988, + "step": 7650 + }, + { + "epoch": 1.7838844899860269, + "grad_norm": 20.35512924194336, + "learning_rate": 1.9295597484276733e-05, + "loss": 1.6377, + "step": 7660 + }, + { + "epoch": 1.7862133209129016, + "grad_norm": 14.52749252319336, + "learning_rate": 1.9281621243885396e-05, + "loss": 1.4134, + "step": 7670 + }, + { + "epoch": 1.7885421518397764, + "grad_norm": 17.46294403076172, + "learning_rate": 1.9267645003494063e-05, + "loss": 1.3532, + "step": 7680 + }, + { + "epoch": 1.7908709827666511, + "grad_norm": 43.280277252197266, + "learning_rate": 1.9253668763102727e-05, + "loss": 1.7841, + "step": 7690 + }, + { + "epoch": 1.7931998136935259, + "grad_norm": 15.846488952636719, + "learning_rate": 1.923969252271139e-05, + "loss": 2.242, + "step": 7700 + }, + { + "epoch": 1.7955286446204006, + "grad_norm": 73.0659408569336, + "learning_rate": 1.9225716282320054e-05, + "loss": 1.3953, + "step": 7710 + }, + { + "epoch": 1.7978574755472754, + "grad_norm": 38.85818862915039, + "learning_rate": 1.921174004192872e-05, + "loss": 1.8649, + "step": 7720 + }, + { + "epoch": 1.8001863064741501, + "grad_norm": 5.484123706817627, + "learning_rate": 1.9197763801537384e-05, + "loss": 1.4726, + "step": 7730 + }, + { + "epoch": 1.8025151374010246, + "grad_norm": 7.886444091796875, + "learning_rate": 1.918378756114605e-05, + "loss": 1.3963, + "step": 7740 + }, + { + "epoch": 1.8048439683278994, + "grad_norm": 13.587664604187012, + "learning_rate": 1.9169811320754715e-05, + "loss": 1.5654, + "step": 7750 + }, + { + "epoch": 1.8071727992547741, + "grad_norm": 23.31873893737793, + "learning_rate": 1.9155835080363382e-05, + "loss": 1.0611, + "step": 7760 + }, + { + "epoch": 1.8095016301816487, + "grad_norm": 85.84972381591797, + "learning_rate": 1.914185883997205e-05, + "loss": 1.3399, + "step": 7770 + }, + { + "epoch": 1.8118304611085234, + "grad_norm": 12.845376014709473, + "learning_rate": 1.9127882599580713e-05, + "loss": 2.0852, + "step": 7780 + }, + { + "epoch": 1.8141592920353982, + "grad_norm": 8.165783882141113, + "learning_rate": 1.911390635918938e-05, + "loss": 1.0608, + "step": 7790 + }, + { + "epoch": 1.816488122962273, + "grad_norm": 18.634197235107422, + "learning_rate": 1.9099930118798043e-05, + "loss": 1.2293, + "step": 7800 + }, + { + "epoch": 1.8188169538891477, + "grad_norm": 4.013707160949707, + "learning_rate": 1.908595387840671e-05, + "loss": 1.1989, + "step": 7810 + }, + { + "epoch": 1.8211457848160224, + "grad_norm": 9.955910682678223, + "learning_rate": 1.9071977638015374e-05, + "loss": 1.7447, + "step": 7820 + }, + { + "epoch": 1.8234746157428972, + "grad_norm": 3.6328682899475098, + "learning_rate": 1.905800139762404e-05, + "loss": 1.3645, + "step": 7830 + }, + { + "epoch": 1.825803446669772, + "grad_norm": 5.54978609085083, + "learning_rate": 1.9044025157232704e-05, + "loss": 1.3699, + "step": 7840 + }, + { + "epoch": 1.8281322775966466, + "grad_norm": 19.170997619628906, + "learning_rate": 1.903004891684137e-05, + "loss": 1.2573, + "step": 7850 + }, + { + "epoch": 1.8304611085235212, + "grad_norm": 13.860367774963379, + "learning_rate": 1.9016072676450035e-05, + "loss": 1.1799, + "step": 7860 + }, + { + "epoch": 1.832789939450396, + "grad_norm": 10.46110725402832, + "learning_rate": 1.9002096436058702e-05, + "loss": 2.0828, + "step": 7870 + }, + { + "epoch": 1.8351187703772704, + "grad_norm": 63.465240478515625, + "learning_rate": 1.8988120195667365e-05, + "loss": 1.0285, + "step": 7880 + }, + { + "epoch": 1.8374476013041452, + "grad_norm": 10.297215461730957, + "learning_rate": 1.8974143955276032e-05, + "loss": 1.0443, + "step": 7890 + }, + { + "epoch": 1.83977643223102, + "grad_norm": 7.570961952209473, + "learning_rate": 1.8960167714884696e-05, + "loss": 1.3184, + "step": 7900 + }, + { + "epoch": 1.8421052631578947, + "grad_norm": 9.8710298538208, + "learning_rate": 1.8946191474493363e-05, + "loss": 1.0637, + "step": 7910 + }, + { + "epoch": 1.8444340940847694, + "grad_norm": 11.42401123046875, + "learning_rate": 1.8932215234102027e-05, + "loss": 1.2406, + "step": 7920 + }, + { + "epoch": 1.8467629250116442, + "grad_norm": 15.829558372497559, + "learning_rate": 1.8918238993710694e-05, + "loss": 0.9227, + "step": 7930 + }, + { + "epoch": 1.849091755938519, + "grad_norm": 20.00259017944336, + "learning_rate": 1.890426275331936e-05, + "loss": 1.1393, + "step": 7940 + }, + { + "epoch": 1.8514205868653937, + "grad_norm": 5.527038097381592, + "learning_rate": 1.8890286512928024e-05, + "loss": 1.1476, + "step": 7950 + }, + { + "epoch": 1.8537494177922684, + "grad_norm": 7.4063334465026855, + "learning_rate": 1.8876310272536688e-05, + "loss": 1.1472, + "step": 7960 + }, + { + "epoch": 1.856078248719143, + "grad_norm": 6.894715785980225, + "learning_rate": 1.886233403214535e-05, + "loss": 1.1841, + "step": 7970 + }, + { + "epoch": 1.8584070796460177, + "grad_norm": 53.53019714355469, + "learning_rate": 1.8848357791754018e-05, + "loss": 1.1964, + "step": 7980 + }, + { + "epoch": 1.8607359105728924, + "grad_norm": 50.521297454833984, + "learning_rate": 1.8834381551362682e-05, + "loss": 1.3668, + "step": 7990 + }, + { + "epoch": 1.863064741499767, + "grad_norm": 8.40929889678955, + "learning_rate": 1.882040531097135e-05, + "loss": 0.8395, + "step": 8000 + }, + { + "epoch": 1.8653935724266417, + "grad_norm": 14.613615989685059, + "learning_rate": 1.8806429070580012e-05, + "loss": 1.3538, + "step": 8010 + }, + { + "epoch": 1.8677224033535165, + "grad_norm": 5.63895320892334, + "learning_rate": 1.879245283018868e-05, + "loss": 1.0303, + "step": 8020 + }, + { + "epoch": 1.8700512342803912, + "grad_norm": 8.055392265319824, + "learning_rate": 1.8778476589797343e-05, + "loss": 1.5969, + "step": 8030 + }, + { + "epoch": 1.872380065207266, + "grad_norm": 16.90934181213379, + "learning_rate": 1.876450034940601e-05, + "loss": 1.3653, + "step": 8040 + }, + { + "epoch": 1.8747088961341407, + "grad_norm": 20.84347915649414, + "learning_rate": 1.8750524109014674e-05, + "loss": 0.9698, + "step": 8050 + }, + { + "epoch": 1.8770377270610155, + "grad_norm": 4.819990634918213, + "learning_rate": 1.873654786862334e-05, + "loss": 1.1794, + "step": 8060 + }, + { + "epoch": 1.8793665579878902, + "grad_norm": 30.6875057220459, + "learning_rate": 1.8722571628232008e-05, + "loss": 1.5583, + "step": 8070 + }, + { + "epoch": 1.8816953889147647, + "grad_norm": 5.91504430770874, + "learning_rate": 1.870859538784067e-05, + "loss": 0.9559, + "step": 8080 + }, + { + "epoch": 1.8840242198416395, + "grad_norm": 58.85013961791992, + "learning_rate": 1.8694619147449338e-05, + "loss": 1.8267, + "step": 8090 + }, + { + "epoch": 1.8863530507685142, + "grad_norm": 8.03494644165039, + "learning_rate": 1.8680642907058e-05, + "loss": 1.4692, + "step": 8100 + }, + { + "epoch": 1.8886818816953888, + "grad_norm": 7.644827842712402, + "learning_rate": 1.866666666666667e-05, + "loss": 1.3284, + "step": 8110 + }, + { + "epoch": 1.8910107126222635, + "grad_norm": 7.554988861083984, + "learning_rate": 1.8652690426275332e-05, + "loss": 1.5628, + "step": 8120 + }, + { + "epoch": 1.8933395435491382, + "grad_norm": 28.42096519470215, + "learning_rate": 1.8638714185884e-05, + "loss": 1.6219, + "step": 8130 + }, + { + "epoch": 1.895668374476013, + "grad_norm": 4.3279032707214355, + "learning_rate": 1.8624737945492663e-05, + "loss": 0.9164, + "step": 8140 + }, + { + "epoch": 1.8979972054028877, + "grad_norm": 15.174659729003906, + "learning_rate": 1.861076170510133e-05, + "loss": 1.5619, + "step": 8150 + }, + { + "epoch": 1.9003260363297625, + "grad_norm": 5.442610740661621, + "learning_rate": 1.8596785464709993e-05, + "loss": 0.9886, + "step": 8160 + }, + { + "epoch": 1.9026548672566372, + "grad_norm": 4.958669662475586, + "learning_rate": 1.858280922431866e-05, + "loss": 1.652, + "step": 8170 + }, + { + "epoch": 1.904983698183512, + "grad_norm": 17.52564811706543, + "learning_rate": 1.8568832983927324e-05, + "loss": 0.9101, + "step": 8180 + }, + { + "epoch": 1.9073125291103867, + "grad_norm": 11.019481658935547, + "learning_rate": 1.855485674353599e-05, + "loss": 1.5395, + "step": 8190 + }, + { + "epoch": 1.9096413600372613, + "grad_norm": 7.791713714599609, + "learning_rate": 1.8540880503144655e-05, + "loss": 1.3154, + "step": 8200 + }, + { + "epoch": 1.911970190964136, + "grad_norm": 19.8372802734375, + "learning_rate": 1.852690426275332e-05, + "loss": 1.1015, + "step": 8210 + }, + { + "epoch": 1.9142990218910108, + "grad_norm": 8.076462745666504, + "learning_rate": 1.8512928022361985e-05, + "loss": 1.1082, + "step": 8220 + }, + { + "epoch": 1.9166278528178853, + "grad_norm": 5.504517555236816, + "learning_rate": 1.849895178197065e-05, + "loss": 1.5135, + "step": 8230 + }, + { + "epoch": 1.91895668374476, + "grad_norm": 9.928620338439941, + "learning_rate": 1.8484975541579316e-05, + "loss": 0.9919, + "step": 8240 + }, + { + "epoch": 1.9212855146716348, + "grad_norm": 12.56198787689209, + "learning_rate": 1.847099930118798e-05, + "loss": 1.756, + "step": 8250 + }, + { + "epoch": 1.9236143455985095, + "grad_norm": 6.634976863861084, + "learning_rate": 1.8457023060796646e-05, + "loss": 1.1167, + "step": 8260 + }, + { + "epoch": 1.9259431765253843, + "grad_norm": 9.223847389221191, + "learning_rate": 1.844304682040531e-05, + "loss": 1.0864, + "step": 8270 + }, + { + "epoch": 1.928272007452259, + "grad_norm": 62.113494873046875, + "learning_rate": 1.8429070580013977e-05, + "loss": 1.324, + "step": 8280 + }, + { + "epoch": 1.9306008383791338, + "grad_norm": 8.178874969482422, + "learning_rate": 1.841509433962264e-05, + "loss": 1.0045, + "step": 8290 + }, + { + "epoch": 1.9329296693060085, + "grad_norm": 16.192407608032227, + "learning_rate": 1.8401118099231307e-05, + "loss": 1.2798, + "step": 8300 + }, + { + "epoch": 1.935258500232883, + "grad_norm": 6.274266242980957, + "learning_rate": 1.838714185883997e-05, + "loss": 1.9747, + "step": 8310 + }, + { + "epoch": 1.9375873311597578, + "grad_norm": 21.142480850219727, + "learning_rate": 1.8373165618448638e-05, + "loss": 0.9294, + "step": 8320 + }, + { + "epoch": 1.9399161620866325, + "grad_norm": 14.333216667175293, + "learning_rate": 1.83591893780573e-05, + "loss": 1.1272, + "step": 8330 + }, + { + "epoch": 1.942244993013507, + "grad_norm": 6.282688140869141, + "learning_rate": 1.834521313766597e-05, + "loss": 0.9866, + "step": 8340 + }, + { + "epoch": 1.9445738239403818, + "grad_norm": 11.034943580627441, + "learning_rate": 1.8331236897274632e-05, + "loss": 1.4079, + "step": 8350 + }, + { + "epoch": 1.9469026548672566, + "grad_norm": 6.519465923309326, + "learning_rate": 1.83172606568833e-05, + "loss": 0.7809, + "step": 8360 + }, + { + "epoch": 1.9492314857941313, + "grad_norm": 8.534490585327148, + "learning_rate": 1.8303284416491966e-05, + "loss": 1.5021, + "step": 8370 + }, + { + "epoch": 1.951560316721006, + "grad_norm": 12.68577766418457, + "learning_rate": 1.828930817610063e-05, + "loss": 1.5503, + "step": 8380 + }, + { + "epoch": 1.9538891476478808, + "grad_norm": 20.240991592407227, + "learning_rate": 1.8275331935709297e-05, + "loss": 1.5506, + "step": 8390 + }, + { + "epoch": 1.9562179785747555, + "grad_norm": 9.088332176208496, + "learning_rate": 1.826135569531796e-05, + "loss": 1.1122, + "step": 8400 + }, + { + "epoch": 1.9585468095016303, + "grad_norm": 22.946273803710938, + "learning_rate": 1.8247379454926627e-05, + "loss": 1.1421, + "step": 8410 + }, + { + "epoch": 1.960875640428505, + "grad_norm": 23.892812728881836, + "learning_rate": 1.823340321453529e-05, + "loss": 1.2118, + "step": 8420 + }, + { + "epoch": 1.9632044713553796, + "grad_norm": 9.210332870483398, + "learning_rate": 1.8219426974143958e-05, + "loss": 1.4426, + "step": 8430 + }, + { + "epoch": 1.9655333022822543, + "grad_norm": 16.855960845947266, + "learning_rate": 1.820545073375262e-05, + "loss": 1.6925, + "step": 8440 + }, + { + "epoch": 1.967862133209129, + "grad_norm": 63.36180877685547, + "learning_rate": 1.8191474493361288e-05, + "loss": 1.7084, + "step": 8450 + }, + { + "epoch": 1.9701909641360036, + "grad_norm": 7.800633430480957, + "learning_rate": 1.8177498252969952e-05, + "loss": 1.0854, + "step": 8460 + }, + { + "epoch": 1.9725197950628783, + "grad_norm": 6.054851055145264, + "learning_rate": 1.816352201257862e-05, + "loss": 1.6178, + "step": 8470 + }, + { + "epoch": 1.974848625989753, + "grad_norm": 7.2113938331604, + "learning_rate": 1.8149545772187282e-05, + "loss": 1.0403, + "step": 8480 + }, + { + "epoch": 1.9771774569166278, + "grad_norm": 13.045158386230469, + "learning_rate": 1.8135569531795946e-05, + "loss": 1.1769, + "step": 8490 + }, + { + "epoch": 1.9795062878435026, + "grad_norm": 12.509069442749023, + "learning_rate": 1.812159329140461e-05, + "loss": 1.4351, + "step": 8500 + }, + { + "epoch": 1.9818351187703773, + "grad_norm": 16.71681022644043, + "learning_rate": 1.8107617051013277e-05, + "loss": 0.8778, + "step": 8510 + }, + { + "epoch": 1.984163949697252, + "grad_norm": 10.629281997680664, + "learning_rate": 1.8093640810621944e-05, + "loss": 0.7607, + "step": 8520 + }, + { + "epoch": 1.9864927806241268, + "grad_norm": 6.857609272003174, + "learning_rate": 1.8079664570230607e-05, + "loss": 0.9464, + "step": 8530 + }, + { + "epoch": 1.9888216115510013, + "grad_norm": 6.874720573425293, + "learning_rate": 1.8065688329839274e-05, + "loss": 1.3631, + "step": 8540 + }, + { + "epoch": 1.991150442477876, + "grad_norm": 7.804827690124512, + "learning_rate": 1.8051712089447938e-05, + "loss": 0.9221, + "step": 8550 + }, + { + "epoch": 1.9934792734047508, + "grad_norm": 84.84910583496094, + "learning_rate": 1.8037735849056605e-05, + "loss": 1.2762, + "step": 8560 + }, + { + "epoch": 1.9958081043316254, + "grad_norm": 24.175487518310547, + "learning_rate": 1.8023759608665268e-05, + "loss": 1.3768, + "step": 8570 + }, + { + "epoch": 1.9981369352585001, + "grad_norm": 5.0154900550842285, + "learning_rate": 1.8009783368273935e-05, + "loss": 1.106, + "step": 8580 + }, + { + "epoch": 2.0, + "eval_loss": 0.502763032913208, + "eval_runtime": 57.2795, + "eval_samples_per_second": 76.048, + "eval_steps_per_second": 1.205, + "step": 8588 + }, + { + "epoch": 2.000465766185375, + "grad_norm": 9.15768814086914, + "learning_rate": 1.7997204751921734e-05, + "loss": 0.8738, + "step": 8590 + }, + { + "epoch": 2.0027945971122496, + "grad_norm": 19.934829711914062, + "learning_rate": 5.960857409133271e-06, + "loss": 0.9064, + "step": 8600 + }, + { + "epoch": 2.0051234280391244, + "grad_norm": 7.445651054382324, + "learning_rate": 5.932898415657036e-06, + "loss": 0.9088, + "step": 8610 + }, + { + "epoch": 2.007452258965999, + "grad_norm": 11.872694969177246, + "learning_rate": 5.904939422180802e-06, + "loss": 1.0579, + "step": 8620 + }, + { + "epoch": 2.009781089892874, + "grad_norm": 23.22488784790039, + "learning_rate": 5.8797763280521905e-06, + "loss": 1.1115, + "step": 8630 + }, + { + "epoch": 2.0121099208197486, + "grad_norm": 16.743192672729492, + "learning_rate": 5.854613233923579e-06, + "loss": 0.9759, + "step": 8640 + }, + { + "epoch": 2.0144387517466233, + "grad_norm": 6.448403358459473, + "learning_rate": 5.826654240447344e-06, + "loss": 0.9698, + "step": 8650 + }, + { + "epoch": 2.016767582673498, + "grad_norm": 22.09251594543457, + "learning_rate": 5.79869524697111e-06, + "loss": 1.0631, + "step": 8660 + }, + { + "epoch": 2.0190964136003724, + "grad_norm": 8.56945514678955, + "learning_rate": 5.770736253494874e-06, + "loss": 1.2748, + "step": 8670 + }, + { + "epoch": 2.021425244527247, + "grad_norm": 21.083589553833008, + "learning_rate": 5.742777260018639e-06, + "loss": 1.6963, + "step": 8680 + }, + { + "epoch": 2.023754075454122, + "grad_norm": 7.669532775878906, + "learning_rate": 5.714818266542404e-06, + "loss": 1.0708, + "step": 8690 + }, + { + "epoch": 2.0260829063809966, + "grad_norm": 10.819165229797363, + "learning_rate": 5.68685927306617e-06, + "loss": 0.9201, + "step": 8700 + }, + { + "epoch": 2.0284117373078714, + "grad_norm": 7.748441696166992, + "learning_rate": 5.6589002795899356e-06, + "loss": 1.4279, + "step": 8710 + }, + { + "epoch": 2.030740568234746, + "grad_norm": 5.020903587341309, + "learning_rate": 5.6309412861136996e-06, + "loss": 1.7287, + "step": 8720 + }, + { + "epoch": 2.033069399161621, + "grad_norm": 13.887042045593262, + "learning_rate": 5.602982292637465e-06, + "loss": 1.1915, + "step": 8730 + }, + { + "epoch": 2.0353982300884956, + "grad_norm": 69.59370422363281, + "learning_rate": 5.57502329916123e-06, + "loss": 1.2851, + "step": 8740 + }, + { + "epoch": 2.0377270610153704, + "grad_norm": 46.7410888671875, + "learning_rate": 5.547064305684996e-06, + "loss": 1.6063, + "step": 8750 + }, + { + "epoch": 2.040055891942245, + "grad_norm": 6.513312339782715, + "learning_rate": 5.519105312208761e-06, + "loss": 1.3997, + "step": 8760 + }, + { + "epoch": 2.04238472286912, + "grad_norm": 34.405757904052734, + "learning_rate": 5.491146318732526e-06, + "loss": 1.4069, + "step": 8770 + }, + { + "epoch": 2.0447135537959946, + "grad_norm": 18.17137336730957, + "learning_rate": 5.46318732525629e-06, + "loss": 1.1842, + "step": 8780 + }, + { + "epoch": 2.047042384722869, + "grad_norm": 50.68307876586914, + "learning_rate": 5.435228331780056e-06, + "loss": 1.3923, + "step": 8790 + }, + { + "epoch": 2.0493712156497437, + "grad_norm": 5.154820919036865, + "learning_rate": 5.407269338303822e-06, + "loss": 1.1868, + "step": 8800 + }, + { + "epoch": 2.0517000465766184, + "grad_norm": 57.07107162475586, + "learning_rate": 5.3793103448275865e-06, + "loss": 1.1506, + "step": 8810 + }, + { + "epoch": 2.054028877503493, + "grad_norm": 5.840853691101074, + "learning_rate": 5.351351351351352e-06, + "loss": 0.9782, + "step": 8820 + }, + { + "epoch": 2.056357708430368, + "grad_norm": 12.97612190246582, + "learning_rate": 5.323392357875116e-06, + "loss": 1.4827, + "step": 8830 + }, + { + "epoch": 2.0586865393572427, + "grad_norm": 16.05446434020996, + "learning_rate": 5.295433364398882e-06, + "loss": 1.2806, + "step": 8840 + }, + { + "epoch": 2.0610153702841174, + "grad_norm": 13.387228012084961, + "learning_rate": 5.267474370922647e-06, + "loss": 1.1232, + "step": 8850 + }, + { + "epoch": 2.063344201210992, + "grad_norm": 6.215972423553467, + "learning_rate": 5.239515377446412e-06, + "loss": 1.3148, + "step": 8860 + }, + { + "epoch": 2.065673032137867, + "grad_norm": 13.655588150024414, + "learning_rate": 5.211556383970177e-06, + "loss": 0.9524, + "step": 8870 + }, + { + "epoch": 2.0680018630647417, + "grad_norm": 6.583061695098877, + "learning_rate": 5.183597390493942e-06, + "loss": 1.0593, + "step": 8880 + }, + { + "epoch": 2.0703306939916164, + "grad_norm": 7.793154716491699, + "learning_rate": 5.155638397017708e-06, + "loss": 1.2788, + "step": 8890 + }, + { + "epoch": 2.0726595249184907, + "grad_norm": 17.533212661743164, + "learning_rate": 5.1276794035414726e-06, + "loss": 0.9179, + "step": 8900 + }, + { + "epoch": 2.0749883558453655, + "grad_norm": 8.611387252807617, + "learning_rate": 5.099720410065238e-06, + "loss": 1.1824, + "step": 8910 + }, + { + "epoch": 2.07731718677224, + "grad_norm": 8.162396430969238, + "learning_rate": 5.071761416589003e-06, + "loss": 0.7685, + "step": 8920 + }, + { + "epoch": 2.079646017699115, + "grad_norm": 14.802386283874512, + "learning_rate": 5.043802423112768e-06, + "loss": 0.9294, + "step": 8930 + }, + { + "epoch": 2.0819748486259897, + "grad_norm": 14.819314956665039, + "learning_rate": 5.015843429636533e-06, + "loss": 1.0837, + "step": 8940 + }, + { + "epoch": 2.0843036795528644, + "grad_norm": 45.875030517578125, + "learning_rate": 4.9878844361602984e-06, + "loss": 1.8533, + "step": 8950 + }, + { + "epoch": 2.086632510479739, + "grad_norm": 4.293564796447754, + "learning_rate": 4.959925442684063e-06, + "loss": 1.0036, + "step": 8960 + }, + { + "epoch": 2.088961341406614, + "grad_norm": 56.804134368896484, + "learning_rate": 4.931966449207829e-06, + "loss": 1.1895, + "step": 8970 + }, + { + "epoch": 2.0912901723334887, + "grad_norm": 14.47994327545166, + "learning_rate": 4.904007455731594e-06, + "loss": 1.5637, + "step": 8980 + }, + { + "epoch": 2.0936190032603634, + "grad_norm": 19.151981353759766, + "learning_rate": 4.876048462255359e-06, + "loss": 0.885, + "step": 8990 + }, + { + "epoch": 2.095947834187238, + "grad_norm": 30.888566970825195, + "learning_rate": 4.848089468779124e-06, + "loss": 1.1562, + "step": 9000 + }, + { + "epoch": 2.098276665114113, + "grad_norm": 7.732218265533447, + "learning_rate": 4.820130475302889e-06, + "loss": 0.9851, + "step": 9010 + }, + { + "epoch": 2.1006054960409872, + "grad_norm": 8.628619194030762, + "learning_rate": 4.792171481826655e-06, + "loss": 1.1773, + "step": 9020 + }, + { + "epoch": 2.102934326967862, + "grad_norm": 5.685352802276611, + "learning_rate": 4.76421248835042e-06, + "loss": 0.8583, + "step": 9030 + }, + { + "epoch": 2.1052631578947367, + "grad_norm": 59.909934997558594, + "learning_rate": 4.7362534948741845e-06, + "loss": 0.8674, + "step": 9040 + }, + { + "epoch": 2.1075919888216115, + "grad_norm": 10.20012378692627, + "learning_rate": 4.708294501397949e-06, + "loss": 1.5069, + "step": 9050 + }, + { + "epoch": 2.109920819748486, + "grad_norm": 11.948861122131348, + "learning_rate": 4.680335507921715e-06, + "loss": 0.8768, + "step": 9060 + }, + { + "epoch": 2.112249650675361, + "grad_norm": 22.48712158203125, + "learning_rate": 4.65237651444548e-06, + "loss": 1.3001, + "step": 9070 + }, + { + "epoch": 2.1145784816022357, + "grad_norm": 7.2206315994262695, + "learning_rate": 4.6244175209692456e-06, + "loss": 1.2205, + "step": 9080 + }, + { + "epoch": 2.1169073125291105, + "grad_norm": 8.07583236694336, + "learning_rate": 4.59645852749301e-06, + "loss": 1.4863, + "step": 9090 + }, + { + "epoch": 2.119236143455985, + "grad_norm": 11.95862102508545, + "learning_rate": 4.568499534016775e-06, + "loss": 1.0672, + "step": 9100 + }, + { + "epoch": 2.12156497438286, + "grad_norm": 60.69342041015625, + "learning_rate": 4.540540540540541e-06, + "loss": 1.4529, + "step": 9110 + }, + { + "epoch": 2.1238938053097347, + "grad_norm": 11.049464225769043, + "learning_rate": 4.512581547064306e-06, + "loss": 0.9201, + "step": 9120 + }, + { + "epoch": 2.1262226362366095, + "grad_norm": 15.80212116241455, + "learning_rate": 4.4846225535880715e-06, + "loss": 1.3178, + "step": 9130 + }, + { + "epoch": 2.1285514671634838, + "grad_norm": 12.804340362548828, + "learning_rate": 4.4566635601118354e-06, + "loss": 1.0052, + "step": 9140 + }, + { + "epoch": 2.1308802980903585, + "grad_norm": 52.227439880371094, + "learning_rate": 4.428704566635601e-06, + "loss": 0.6626, + "step": 9150 + }, + { + "epoch": 2.1332091290172333, + "grad_norm": 4.820531368255615, + "learning_rate": 4.400745573159366e-06, + "loss": 1.1978, + "step": 9160 + }, + { + "epoch": 2.135537959944108, + "grad_norm": 9.708754539489746, + "learning_rate": 4.372786579683132e-06, + "loss": 1.1169, + "step": 9170 + }, + { + "epoch": 2.1378667908709827, + "grad_norm": 23.4190731048584, + "learning_rate": 4.344827586206897e-06, + "loss": 0.9696, + "step": 9180 + }, + { + "epoch": 2.1401956217978575, + "grad_norm": 9.562813758850098, + "learning_rate": 4.316868592730661e-06, + "loss": 0.9916, + "step": 9190 + }, + { + "epoch": 2.1425244527247322, + "grad_norm": 51.41989517211914, + "learning_rate": 4.288909599254427e-06, + "loss": 1.6923, + "step": 9200 + }, + { + "epoch": 2.144853283651607, + "grad_norm": 18.753562927246094, + "learning_rate": 4.260950605778192e-06, + "loss": 1.283, + "step": 9210 + }, + { + "epoch": 2.1471821145784817, + "grad_norm": 6.791256904602051, + "learning_rate": 4.2329916123019575e-06, + "loss": 1.1738, + "step": 9220 + }, + { + "epoch": 2.1495109455053565, + "grad_norm": 79.94830322265625, + "learning_rate": 4.205032618825722e-06, + "loss": 0.9406, + "step": 9230 + }, + { + "epoch": 2.1518397764322312, + "grad_norm": 6.553360939025879, + "learning_rate": 4.177073625349487e-06, + "loss": 0.8378, + "step": 9240 + }, + { + "epoch": 2.1541686073591055, + "grad_norm": 5.851924419403076, + "learning_rate": 4.149114631873252e-06, + "loss": 0.9363, + "step": 9250 + }, + { + "epoch": 2.1564974382859803, + "grad_norm": 24.280319213867188, + "learning_rate": 4.121155638397018e-06, + "loss": 1.1103, + "step": 9260 + }, + { + "epoch": 2.158826269212855, + "grad_norm": 4.203216075897217, + "learning_rate": 4.093196644920783e-06, + "loss": 1.2514, + "step": 9270 + }, + { + "epoch": 2.16115510013973, + "grad_norm": 7.374353408813477, + "learning_rate": 4.065237651444548e-06, + "loss": 0.9512, + "step": 9280 + }, + { + "epoch": 2.1634839310666045, + "grad_norm": 17.918182373046875, + "learning_rate": 4.037278657968314e-06, + "loss": 0.9938, + "step": 9290 + }, + { + "epoch": 2.1658127619934793, + "grad_norm": 10.405567169189453, + "learning_rate": 4.009319664492078e-06, + "loss": 1.0328, + "step": 9300 + }, + { + "epoch": 2.168141592920354, + "grad_norm": 56.74263381958008, + "learning_rate": 3.981360671015844e-06, + "loss": 1.0014, + "step": 9310 + }, + { + "epoch": 2.1704704238472288, + "grad_norm": 67.22490692138672, + "learning_rate": 3.9534016775396084e-06, + "loss": 1.6543, + "step": 9320 + }, + { + "epoch": 2.1727992547741035, + "grad_norm": 18.301307678222656, + "learning_rate": 3.925442684063374e-06, + "loss": 1.19, + "step": 9330 + }, + { + "epoch": 2.1751280857009783, + "grad_norm": 4.8976287841796875, + "learning_rate": 3.897483690587139e-06, + "loss": 1.3139, + "step": 9340 + }, + { + "epoch": 2.177456916627853, + "grad_norm": 5.022180557250977, + "learning_rate": 3.869524697110904e-06, + "loss": 0.9576, + "step": 9350 + }, + { + "epoch": 2.1797857475547273, + "grad_norm": 17.034072875976562, + "learning_rate": 3.8415657036346695e-06, + "loss": 1.6219, + "step": 9360 + }, + { + "epoch": 2.182114578481602, + "grad_norm": 17.095535278320312, + "learning_rate": 3.8136067101584343e-06, + "loss": 1.2369, + "step": 9370 + }, + { + "epoch": 2.184443409408477, + "grad_norm": 8.286972045898438, + "learning_rate": 3.7856477166822e-06, + "loss": 1.7725, + "step": 9380 + }, + { + "epoch": 2.1867722403353516, + "grad_norm": 40.874053955078125, + "learning_rate": 3.7576887232059644e-06, + "loss": 1.0996, + "step": 9390 + }, + { + "epoch": 2.1891010712622263, + "grad_norm": 7.332772254943848, + "learning_rate": 3.72972972972973e-06, + "loss": 1.4552, + "step": 9400 + }, + { + "epoch": 2.191429902189101, + "grad_norm": 19.27867889404297, + "learning_rate": 3.701770736253495e-06, + "loss": 0.8538, + "step": 9410 + }, + { + "epoch": 2.193758733115976, + "grad_norm": 9.231902122497559, + "learning_rate": 3.67381174277726e-06, + "loss": 1.0024, + "step": 9420 + }, + { + "epoch": 2.1960875640428505, + "grad_norm": 10.635993003845215, + "learning_rate": 3.6458527493010255e-06, + "loss": 0.8558, + "step": 9430 + }, + { + "epoch": 2.1984163949697253, + "grad_norm": 8.209709167480469, + "learning_rate": 3.6178937558247903e-06, + "loss": 1.0147, + "step": 9440 + }, + { + "epoch": 2.2007452258966, + "grad_norm": 108.34497833251953, + "learning_rate": 3.5899347623485556e-06, + "loss": 1.5607, + "step": 9450 + }, + { + "epoch": 2.203074056823475, + "grad_norm": 42.276737213134766, + "learning_rate": 3.561975768872321e-06, + "loss": 1.2168, + "step": 9460 + }, + { + "epoch": 2.205402887750349, + "grad_norm": 11.989846229553223, + "learning_rate": 3.5340167753960857e-06, + "loss": 1.2395, + "step": 9470 + }, + { + "epoch": 2.207731718677224, + "grad_norm": 63.10080337524414, + "learning_rate": 3.506057781919851e-06, + "loss": 1.2162, + "step": 9480 + }, + { + "epoch": 2.2100605496040986, + "grad_norm": 5.683536529541016, + "learning_rate": 3.4780987884436158e-06, + "loss": 1.3208, + "step": 9490 + }, + { + "epoch": 2.2123893805309733, + "grad_norm": 5.439475059509277, + "learning_rate": 3.4501397949673814e-06, + "loss": 0.8779, + "step": 9500 + }, + { + "epoch": 2.214718211457848, + "grad_norm": 15.528427124023438, + "learning_rate": 3.4221808014911467e-06, + "loss": 1.3149, + "step": 9510 + }, + { + "epoch": 2.217047042384723, + "grad_norm": 16.126054763793945, + "learning_rate": 3.3942218080149115e-06, + "loss": 1.3207, + "step": 9520 + }, + { + "epoch": 2.2193758733115976, + "grad_norm": 6.91264533996582, + "learning_rate": 3.366262814538677e-06, + "loss": 1.3919, + "step": 9530 + }, + { + "epoch": 2.2217047042384723, + "grad_norm": 16.272396087646484, + "learning_rate": 3.3383038210624416e-06, + "loss": 1.2776, + "step": 9540 + }, + { + "epoch": 2.224033535165347, + "grad_norm": 6.406686782836914, + "learning_rate": 3.310344827586207e-06, + "loss": 0.7831, + "step": 9550 + }, + { + "epoch": 2.226362366092222, + "grad_norm": 17.45789909362793, + "learning_rate": 3.282385834109972e-06, + "loss": 1.6462, + "step": 9560 + }, + { + "epoch": 2.2286911970190966, + "grad_norm": 30.968273162841797, + "learning_rate": 3.254426840633737e-06, + "loss": 1.1366, + "step": 9570 + }, + { + "epoch": 2.2310200279459713, + "grad_norm": 13.682144165039062, + "learning_rate": 3.2264678471575023e-06, + "loss": 1.1012, + "step": 9580 + }, + { + "epoch": 2.233348858872846, + "grad_norm": 9.104609489440918, + "learning_rate": 3.198508853681268e-06, + "loss": 1.3075, + "step": 9590 + }, + { + "epoch": 2.2356776897997204, + "grad_norm": 79.64781188964844, + "learning_rate": 3.170549860205033e-06, + "loss": 1.2202, + "step": 9600 + }, + { + "epoch": 2.238006520726595, + "grad_norm": 6.4605631828308105, + "learning_rate": 3.142590866728798e-06, + "loss": 0.6628, + "step": 9610 + }, + { + "epoch": 2.24033535165347, + "grad_norm": 47.32363510131836, + "learning_rate": 3.114631873252563e-06, + "loss": 1.0585, + "step": 9620 + }, + { + "epoch": 2.2426641825803446, + "grad_norm": 8.3875732421875, + "learning_rate": 3.086672879776328e-06, + "loss": 1.4663, + "step": 9630 + }, + { + "epoch": 2.2449930135072194, + "grad_norm": 14.244718551635742, + "learning_rate": 3.0587138863000934e-06, + "loss": 1.2015, + "step": 9640 + }, + { + "epoch": 2.247321844434094, + "grad_norm": 22.4896297454834, + "learning_rate": 3.0307548928238582e-06, + "loss": 1.9219, + "step": 9650 + }, + { + "epoch": 2.249650675360969, + "grad_norm": 10.090749740600586, + "learning_rate": 3.0027958993476235e-06, + "loss": 0.8067, + "step": 9660 + }, + { + "epoch": 2.2519795062878436, + "grad_norm": 8.410811424255371, + "learning_rate": 2.9748369058713883e-06, + "loss": 1.0709, + "step": 9670 + }, + { + "epoch": 2.2543083372147183, + "grad_norm": 25.9169921875, + "learning_rate": 2.946877912395154e-06, + "loss": 1.3688, + "step": 9680 + }, + { + "epoch": 2.256637168141593, + "grad_norm": 24.173946380615234, + "learning_rate": 2.9189189189189193e-06, + "loss": 1.3964, + "step": 9690 + }, + { + "epoch": 2.258965999068468, + "grad_norm": 15.42428207397461, + "learning_rate": 2.890959925442684e-06, + "loss": 1.2085, + "step": 9700 + }, + { + "epoch": 2.261294829995342, + "grad_norm": 16.04413604736328, + "learning_rate": 2.8630009319664494e-06, + "loss": 1.2269, + "step": 9710 + }, + { + "epoch": 2.263623660922217, + "grad_norm": 17.178184509277344, + "learning_rate": 2.8350419384902146e-06, + "loss": 1.4462, + "step": 9720 + }, + { + "epoch": 2.2659524918490916, + "grad_norm": 6.437899589538574, + "learning_rate": 2.8070829450139795e-06, + "loss": 1.521, + "step": 9730 + }, + { + "epoch": 2.2682813227759664, + "grad_norm": 10.40018081665039, + "learning_rate": 2.7791239515377447e-06, + "loss": 1.0889, + "step": 9740 + }, + { + "epoch": 2.270610153702841, + "grad_norm": 6.229658603668213, + "learning_rate": 2.7511649580615096e-06, + "loss": 1.0799, + "step": 9750 + }, + { + "epoch": 2.272938984629716, + "grad_norm": 35.4992790222168, + "learning_rate": 2.723205964585275e-06, + "loss": 0.956, + "step": 9760 + }, + { + "epoch": 2.2752678155565906, + "grad_norm": 15.053701400756836, + "learning_rate": 2.69524697110904e-06, + "loss": 1.033, + "step": 9770 + }, + { + "epoch": 2.2775966464834654, + "grad_norm": 33.243473052978516, + "learning_rate": 2.670083876980429e-06, + "loss": 1.4346, + "step": 9780 + }, + { + "epoch": 2.27992547741034, + "grad_norm": 6.714288234710693, + "learning_rate": 2.6421248835041936e-06, + "loss": 1.1813, + "step": 9790 + }, + { + "epoch": 2.282254308337215, + "grad_norm": 11.203160285949707, + "learning_rate": 2.614165890027959e-06, + "loss": 0.7792, + "step": 9800 + }, + { + "epoch": 2.2845831392640896, + "grad_norm": 146.86062622070312, + "learning_rate": 2.5862068965517246e-06, + "loss": 1.2079, + "step": 9810 + }, + { + "epoch": 2.286911970190964, + "grad_norm": 9.571805000305176, + "learning_rate": 2.5582479030754894e-06, + "loss": 0.8902, + "step": 9820 + }, + { + "epoch": 2.2892408011178387, + "grad_norm": 15.292964935302734, + "learning_rate": 2.5302889095992547e-06, + "loss": 1.2055, + "step": 9830 + }, + { + "epoch": 2.2915696320447134, + "grad_norm": 85.07530975341797, + "learning_rate": 2.5023299161230195e-06, + "loss": 1.6654, + "step": 9840 + }, + { + "epoch": 2.293898462971588, + "grad_norm": 11.542017936706543, + "learning_rate": 2.4743709226467848e-06, + "loss": 1.3289, + "step": 9850 + }, + { + "epoch": 2.296227293898463, + "grad_norm": 7.581414222717285, + "learning_rate": 2.44641192917055e-06, + "loss": 1.1882, + "step": 9860 + }, + { + "epoch": 2.2985561248253377, + "grad_norm": 12.976031303405762, + "learning_rate": 2.418452935694315e-06, + "loss": 1.2158, + "step": 9870 + }, + { + "epoch": 2.3008849557522124, + "grad_norm": 6.565402507781982, + "learning_rate": 2.39049394221808e-06, + "loss": 1.0179, + "step": 9880 + }, + { + "epoch": 2.303213786679087, + "grad_norm": 17.8703556060791, + "learning_rate": 2.362534948741845e-06, + "loss": 1.0429, + "step": 9890 + }, + { + "epoch": 2.305542617605962, + "grad_norm": 24.56272315979004, + "learning_rate": 2.3345759552656107e-06, + "loss": 1.3841, + "step": 9900 + }, + { + "epoch": 2.3078714485328367, + "grad_norm": 13.09598159790039, + "learning_rate": 2.306616961789376e-06, + "loss": 0.9365, + "step": 9910 + }, + { + "epoch": 2.3102002794597114, + "grad_norm": 26.23815155029297, + "learning_rate": 2.2786579683131408e-06, + "loss": 1.4342, + "step": 9920 + }, + { + "epoch": 2.3125291103865857, + "grad_norm": 9.41688060760498, + "learning_rate": 2.250698974836906e-06, + "loss": 1.1212, + "step": 9930 + }, + { + "epoch": 2.3148579413134605, + "grad_norm": 7.139398097991943, + "learning_rate": 2.2227399813606713e-06, + "loss": 0.6761, + "step": 9940 + }, + { + "epoch": 2.317186772240335, + "grad_norm": 13.8255615234375, + "learning_rate": 2.194780987884436e-06, + "loss": 1.2559, + "step": 9950 + }, + { + "epoch": 2.31951560316721, + "grad_norm": 28.836498260498047, + "learning_rate": 2.1668219944082014e-06, + "loss": 1.3978, + "step": 9960 + }, + { + "epoch": 2.3218444340940847, + "grad_norm": 11.870185852050781, + "learning_rate": 2.1388630009319662e-06, + "loss": 1.5667, + "step": 9970 + }, + { + "epoch": 2.3241732650209594, + "grad_norm": 55.986881256103516, + "learning_rate": 2.1109040074557315e-06, + "loss": 1.238, + "step": 9980 + }, + { + "epoch": 2.326502095947834, + "grad_norm": 40.541297912597656, + "learning_rate": 2.082945013979497e-06, + "loss": 1.1823, + "step": 9990 + }, + { + "epoch": 2.328830926874709, + "grad_norm": 10.11244010925293, + "learning_rate": 2.054986020503262e-06, + "loss": 1.3272, + "step": 10000 + }, + { + "epoch": 2.3311597578015837, + "grad_norm": 15.212204933166504, + "learning_rate": 2.0270270270270273e-06, + "loss": 0.8434, + "step": 10010 + }, + { + "epoch": 2.3334885887284584, + "grad_norm": 7.350932598114014, + "learning_rate": 1.999068033550792e-06, + "loss": 1.1032, + "step": 10020 + }, + { + "epoch": 2.335817419655333, + "grad_norm": 32.078880310058594, + "learning_rate": 1.9711090400745574e-06, + "loss": 1.5609, + "step": 10030 + }, + { + "epoch": 2.3381462505822075, + "grad_norm": 9.110737800598145, + "learning_rate": 1.9431500465983226e-06, + "loss": 1.3581, + "step": 10040 + }, + { + "epoch": 2.3404750815090827, + "grad_norm": 43.029335021972656, + "learning_rate": 1.9151910531220875e-06, + "loss": 1.1401, + "step": 10050 + }, + { + "epoch": 2.342803912435957, + "grad_norm": 9.121164321899414, + "learning_rate": 1.8872320596458527e-06, + "loss": 1.1575, + "step": 10060 + }, + { + "epoch": 2.3451327433628317, + "grad_norm": 10.35621452331543, + "learning_rate": 1.859273066169618e-06, + "loss": 0.918, + "step": 10070 + }, + { + "epoch": 2.3474615742897065, + "grad_norm": 15.430923461914062, + "learning_rate": 1.831314072693383e-06, + "loss": 0.9122, + "step": 10080 + }, + { + "epoch": 2.3497904052165812, + "grad_norm": 58.956695556640625, + "learning_rate": 1.8033550792171483e-06, + "loss": 1.055, + "step": 10090 + }, + { + "epoch": 2.352119236143456, + "grad_norm": 4.652696132659912, + "learning_rate": 1.7753960857409133e-06, + "loss": 2.0394, + "step": 10100 + }, + { + "epoch": 2.3544480670703307, + "grad_norm": 27.606590270996094, + "learning_rate": 1.7474370922646786e-06, + "loss": 1.7843, + "step": 10110 + }, + { + "epoch": 2.3567768979972055, + "grad_norm": 13.248260498046875, + "learning_rate": 1.7194780987884436e-06, + "loss": 1.7327, + "step": 10120 + }, + { + "epoch": 2.35910572892408, + "grad_norm": 38.81512451171875, + "learning_rate": 1.6915191053122087e-06, + "loss": 1.8256, + "step": 10130 + }, + { + "epoch": 2.361434559850955, + "grad_norm": 18.427837371826172, + "learning_rate": 1.663560111835974e-06, + "loss": 0.8146, + "step": 10140 + }, + { + "epoch": 2.3637633907778297, + "grad_norm": 21.866954803466797, + "learning_rate": 1.635601118359739e-06, + "loss": 1.3319, + "step": 10150 + }, + { + "epoch": 2.3660922217047045, + "grad_norm": 12.13359546661377, + "learning_rate": 1.6076421248835043e-06, + "loss": 0.9745, + "step": 10160 + }, + { + "epoch": 2.3684210526315788, + "grad_norm": 16.050683975219727, + "learning_rate": 1.5796831314072693e-06, + "loss": 1.3594, + "step": 10170 + }, + { + "epoch": 2.3707498835584535, + "grad_norm": 9.901156425476074, + "learning_rate": 1.5517241379310346e-06, + "loss": 1.2522, + "step": 10180 + }, + { + "epoch": 2.3730787144853283, + "grad_norm": 10.21035385131836, + "learning_rate": 1.5237651444547996e-06, + "loss": 1.0524, + "step": 10190 + }, + { + "epoch": 2.375407545412203, + "grad_norm": 12.119868278503418, + "learning_rate": 1.4958061509785649e-06, + "loss": 1.3232, + "step": 10200 + }, + { + "epoch": 2.3777363763390778, + "grad_norm": 15.788361549377441, + "learning_rate": 1.46784715750233e-06, + "loss": 1.1433, + "step": 10210 + }, + { + "epoch": 2.3800652072659525, + "grad_norm": 14.479697227478027, + "learning_rate": 1.4398881640260952e-06, + "loss": 0.6549, + "step": 10220 + }, + { + "epoch": 2.3823940381928272, + "grad_norm": 5.601083278656006, + "learning_rate": 1.4119291705498603e-06, + "loss": 0.8918, + "step": 10230 + }, + { + "epoch": 2.384722869119702, + "grad_norm": 7.322697639465332, + "learning_rate": 1.3839701770736253e-06, + "loss": 1.4818, + "step": 10240 + }, + { + "epoch": 2.3870517000465767, + "grad_norm": 11.558350563049316, + "learning_rate": 1.3560111835973906e-06, + "loss": 0.9127, + "step": 10250 + }, + { + "epoch": 2.3893805309734515, + "grad_norm": 102.35279846191406, + "learning_rate": 1.3280521901211556e-06, + "loss": 1.1501, + "step": 10260 + }, + { + "epoch": 2.3917093619003262, + "grad_norm": 32.1492805480957, + "learning_rate": 1.3000931966449209e-06, + "loss": 1.2799, + "step": 10270 + }, + { + "epoch": 2.3940381928272005, + "grad_norm": 34.78493881225586, + "learning_rate": 1.272134203168686e-06, + "loss": 1.5782, + "step": 10280 + }, + { + "epoch": 2.3963670237540753, + "grad_norm": 100.30006408691406, + "learning_rate": 1.244175209692451e-06, + "loss": 1.5619, + "step": 10290 + }, + { + "epoch": 2.39869585468095, + "grad_norm": 10.30715274810791, + "learning_rate": 1.2162162162162162e-06, + "loss": 1.1595, + "step": 10300 + }, + { + "epoch": 2.401024685607825, + "grad_norm": 7.528810977935791, + "learning_rate": 1.1882572227399815e-06, + "loss": 0.8228, + "step": 10310 + }, + { + "epoch": 2.4033535165346995, + "grad_norm": 9.433389663696289, + "learning_rate": 1.1602982292637465e-06, + "loss": 1.124, + "step": 10320 + }, + { + "epoch": 2.4056823474615743, + "grad_norm": 50.19340133666992, + "learning_rate": 1.1323392357875116e-06, + "loss": 1.3402, + "step": 10330 + }, + { + "epoch": 2.408011178388449, + "grad_norm": 28.69070816040039, + "learning_rate": 1.1043802423112769e-06, + "loss": 1.1887, + "step": 10340 + }, + { + "epoch": 2.4103400093153238, + "grad_norm": 9.501694679260254, + "learning_rate": 1.0764212488350421e-06, + "loss": 1.0498, + "step": 10350 + }, + { + "epoch": 2.4126688402421985, + "grad_norm": 18.250404357910156, + "learning_rate": 1.0484622553588072e-06, + "loss": 1.7923, + "step": 10360 + }, + { + "epoch": 2.4149976711690733, + "grad_norm": 6.274858474731445, + "learning_rate": 1.0205032618825722e-06, + "loss": 1.1322, + "step": 10370 + }, + { + "epoch": 2.417326502095948, + "grad_norm": 46.66001892089844, + "learning_rate": 9.925442684063373e-07, + "loss": 1.7829, + "step": 10380 + }, + { + "epoch": 2.4196553330228223, + "grad_norm": 6.775675296783447, + "learning_rate": 9.645852749301025e-07, + "loss": 1.2486, + "step": 10390 + }, + { + "epoch": 2.421984163949697, + "grad_norm": 7.135899066925049, + "learning_rate": 9.366262814538677e-07, + "loss": 0.8247, + "step": 10400 + }, + { + "epoch": 2.424312994876572, + "grad_norm": 9.592740058898926, + "learning_rate": 9.086672879776328e-07, + "loss": 0.9868, + "step": 10410 + }, + { + "epoch": 2.4266418258034466, + "grad_norm": 8.71237564086914, + "learning_rate": 8.807082945013979e-07, + "loss": 1.6636, + "step": 10420 + }, + { + "epoch": 2.4289706567303213, + "grad_norm": 11.005659103393555, + "learning_rate": 8.527493010251631e-07, + "loss": 1.2524, + "step": 10430 + }, + { + "epoch": 2.431299487657196, + "grad_norm": 5.304808616638184, + "learning_rate": 8.247903075489282e-07, + "loss": 1.1702, + "step": 10440 + }, + { + "epoch": 2.433628318584071, + "grad_norm": 16.233470916748047, + "learning_rate": 7.968313140726935e-07, + "loss": 1.2214, + "step": 10450 + }, + { + "epoch": 2.4359571495109456, + "grad_norm": 7.547863960266113, + "learning_rate": 7.688723205964585e-07, + "loss": 0.954, + "step": 10460 + }, + { + "epoch": 2.4382859804378203, + "grad_norm": 10.951506614685059, + "learning_rate": 7.409133271202238e-07, + "loss": 1.1177, + "step": 10470 + }, + { + "epoch": 2.440614811364695, + "grad_norm": 6.370845317840576, + "learning_rate": 7.129543336439888e-07, + "loss": 0.897, + "step": 10480 + }, + { + "epoch": 2.44294364229157, + "grad_norm": 9.310226440429688, + "learning_rate": 6.84995340167754e-07, + "loss": 1.4497, + "step": 10490 + }, + { + "epoch": 2.445272473218444, + "grad_norm": 79.52318572998047, + "learning_rate": 6.570363466915191e-07, + "loss": 1.5888, + "step": 10500 + }, + { + "epoch": 2.4476013041453193, + "grad_norm": 6.686119556427002, + "learning_rate": 6.290773532152843e-07, + "loss": 1.0586, + "step": 10510 + }, + { + "epoch": 2.4499301350721936, + "grad_norm": 16.480131149291992, + "learning_rate": 6.011183597390494e-07, + "loss": 1.2573, + "step": 10520 + }, + { + "epoch": 2.4522589659990683, + "grad_norm": 18.26215362548828, + "learning_rate": 5.731593662628145e-07, + "loss": 1.4545, + "step": 10530 + }, + { + "epoch": 2.454587796925943, + "grad_norm": 6.198437213897705, + "learning_rate": 5.452003727865797e-07, + "loss": 1.0308, + "step": 10540 + }, + { + "epoch": 2.456916627852818, + "grad_norm": 8.964813232421875, + "learning_rate": 5.172413793103448e-07, + "loss": 1.3532, + "step": 10550 + }, + { + "epoch": 2.4592454587796926, + "grad_norm": 11.396198272705078, + "learning_rate": 4.892823858341099e-07, + "loss": 1.2104, + "step": 10560 + }, + { + "epoch": 2.4615742897065673, + "grad_norm": 10.318989753723145, + "learning_rate": 4.6132339235787515e-07, + "loss": 0.9936, + "step": 10570 + }, + { + "epoch": 2.463903120633442, + "grad_norm": 84.06649780273438, + "learning_rate": 4.333643988816403e-07, + "loss": 1.2898, + "step": 10580 + }, + { + "epoch": 2.466231951560317, + "grad_norm": 9.45693588256836, + "learning_rate": 4.0540540540540546e-07, + "loss": 1.3486, + "step": 10590 + }, + { + "epoch": 2.4685607824871916, + "grad_norm": 15.41531753540039, + "learning_rate": 3.774464119291705e-07, + "loss": 1.9033, + "step": 10600 + }, + { + "epoch": 2.4708896134140663, + "grad_norm": 23.50023078918457, + "learning_rate": 3.4948741845293567e-07, + "loss": 0.8877, + "step": 10610 + }, + { + "epoch": 2.473218444340941, + "grad_norm": 11.658283233642578, + "learning_rate": 3.215284249767008e-07, + "loss": 1.4179, + "step": 10620 + }, + { + "epoch": 2.4755472752678154, + "grad_norm": 9.004804611206055, + "learning_rate": 2.93569431500466e-07, + "loss": 1.254, + "step": 10630 + }, + { + "epoch": 2.47787610619469, + "grad_norm": 11.310983657836914, + "learning_rate": 2.6561043802423113e-07, + "loss": 0.9785, + "step": 10640 + }, + { + "epoch": 2.480204937121565, + "grad_norm": 24.459308624267578, + "learning_rate": 2.376514445479963e-07, + "loss": 1.8078, + "step": 10650 + }, + { + "epoch": 2.4825337680484396, + "grad_norm": 14.287947654724121, + "learning_rate": 2.0969245107176142e-07, + "loss": 1.2669, + "step": 10660 + }, + { + "epoch": 2.4848625989753144, + "grad_norm": 9.133045196533203, + "learning_rate": 1.8173345759552657e-07, + "loss": 0.8908, + "step": 10670 + }, + { + "epoch": 2.487191429902189, + "grad_norm": 45.257755279541016, + "learning_rate": 1.5377446411929173e-07, + "loss": 1.939, + "step": 10680 + }, + { + "epoch": 2.489520260829064, + "grad_norm": 14.624387741088867, + "learning_rate": 1.2581547064305686e-07, + "loss": 1.2464, + "step": 10690 + }, + { + "epoch": 2.4918490917559386, + "grad_norm": 14.175239562988281, + "learning_rate": 9.7856477166822e-08, + "loss": 1.524, + "step": 10700 + }, + { + "epoch": 2.4941779226828134, + "grad_norm": 10.873252868652344, + "learning_rate": 6.989748369058714e-08, + "loss": 1.2533, + "step": 10710 + }, + { + "epoch": 2.496506753609688, + "grad_norm": 9.596633911132812, + "learning_rate": 4.193849021435229e-08, + "loss": 1.5095, + "step": 10720 + }, + { + "epoch": 2.498835584536563, + "grad_norm": 80.26763153076172, + "learning_rate": 1.3979496738117427e-08, + "loss": 2.6379, + "step": 10730 + }, + { + "epoch": 2.501164415463437, + "grad_norm": 9.910447120666504, + "learning_rate": 0.0, + "loss": 0.9697, + "step": 10740 + }, + { + "epoch": 2.503493246390312, + "grad_norm": 11.675068855285645, + "learning_rate": 0.0, + "loss": 1.0056, + "step": 10750 + }, + { + "epoch": 2.5058220773171866, + "grad_norm": 11.493444442749023, + "learning_rate": 0.0, + "loss": 1.5176, + "step": 10760 + }, + { + "epoch": 2.5081509082440614, + "grad_norm": 12.753838539123535, + "learning_rate": 0.0, + "loss": 1.1031, + "step": 10770 + }, + { + "epoch": 2.510479739170936, + "grad_norm": 14.464014053344727, + "learning_rate": 0.0, + "loss": 0.9649, + "step": 10780 + }, + { + "epoch": 2.512808570097811, + "grad_norm": 15.377732276916504, + "learning_rate": 0.0, + "loss": 1.4322, + "step": 10790 + }, + { + "epoch": 2.5151374010246856, + "grad_norm": 9.88819694519043, + "learning_rate": 0.0, + "loss": 0.9797, + "step": 10800 + }, + { + "epoch": 2.5174662319515604, + "grad_norm": 6.533481121063232, + "learning_rate": 0.0, + "loss": 1.2964, + "step": 10810 + }, + { + "epoch": 2.519795062878435, + "grad_norm": 20.36925506591797, + "learning_rate": 0.0, + "loss": 1.6195, + "step": 10820 + }, + { + "epoch": 2.52212389380531, + "grad_norm": 6.780529022216797, + "learning_rate": 0.0, + "loss": 1.039, + "step": 10830 + }, + { + "epoch": 2.5244527247321846, + "grad_norm": 36.58348083496094, + "learning_rate": 0.0, + "loss": 1.3896, + "step": 10840 + }, + { + "epoch": 2.526781555659059, + "grad_norm": 7.8758039474487305, + "learning_rate": 0.0, + "loss": 1.4437, + "step": 10850 + }, + { + "epoch": 2.529110386585934, + "grad_norm": 8.424162864685059, + "learning_rate": 0.0, + "loss": 1.5715, + "step": 10860 + }, + { + "epoch": 2.5314392175128084, + "grad_norm": 15.89005184173584, + "learning_rate": 0.0, + "loss": 1.1113, + "step": 10870 + }, + { + "epoch": 2.533768048439683, + "grad_norm": 10.675698280334473, + "learning_rate": 0.0, + "loss": 1.5257, + "step": 10880 + }, + { + "epoch": 2.536096879366558, + "grad_norm": 36.62575912475586, + "learning_rate": 0.0, + "loss": 0.8961, + "step": 10890 + }, + { + "epoch": 2.5384257102934327, + "grad_norm": 5.308418273925781, + "learning_rate": 0.0, + "loss": 0.6685, + "step": 10900 + }, + { + "epoch": 2.5407545412203074, + "grad_norm": 29.720823287963867, + "learning_rate": 0.0, + "loss": 0.9164, + "step": 10910 + }, + { + "epoch": 2.543083372147182, + "grad_norm": 77.38935852050781, + "learning_rate": 0.0, + "loss": 2.2807, + "step": 10920 + }, + { + "epoch": 2.545412203074057, + "grad_norm": 86.90229034423828, + "learning_rate": 0.0, + "loss": 1.4119, + "step": 10930 + }, + { + "epoch": 2.5477410340009317, + "grad_norm": 7.217380046844482, + "learning_rate": 0.0, + "loss": 0.8429, + "step": 10940 + }, + { + "epoch": 2.5500698649278064, + "grad_norm": 15.611220359802246, + "learning_rate": 0.0, + "loss": 0.918, + "step": 10950 + }, + { + "epoch": 2.5523986958546807, + "grad_norm": 7.818775177001953, + "learning_rate": 0.0, + "loss": 1.1341, + "step": 10960 + }, + { + "epoch": 2.554727526781556, + "grad_norm": 10.395451545715332, + "learning_rate": 0.0, + "loss": 1.0161, + "step": 10970 + }, + { + "epoch": 2.55705635770843, + "grad_norm": 9.080485343933105, + "learning_rate": 0.0, + "loss": 0.9479, + "step": 10980 + }, + { + "epoch": 2.559385188635305, + "grad_norm": 19.17698860168457, + "learning_rate": 0.0, + "loss": 1.0788, + "step": 10990 + }, + { + "epoch": 2.5617140195621797, + "grad_norm": 9.98133659362793, + "learning_rate": 0.0, + "loss": 1.4046, + "step": 11000 + }, + { + "epoch": 2.5640428504890544, + "grad_norm": 12.91927433013916, + "learning_rate": 0.0, + "loss": 2.1021, + "step": 11010 + }, + { + "epoch": 2.566371681415929, + "grad_norm": 10.550889015197754, + "learning_rate": 0.0, + "loss": 1.4063, + "step": 11020 + }, + { + "epoch": 2.568700512342804, + "grad_norm": 20.967208862304688, + "learning_rate": 0.0, + "loss": 0.9094, + "step": 11030 + }, + { + "epoch": 2.5710293432696787, + "grad_norm": 9.758930206298828, + "learning_rate": 0.0, + "loss": 0.9526, + "step": 11040 + }, + { + "epoch": 2.5733581741965534, + "grad_norm": 10.604878425598145, + "learning_rate": 0.0, + "loss": 0.8984, + "step": 11050 + }, + { + "epoch": 2.575687005123428, + "grad_norm": 68.84051513671875, + "learning_rate": 0.0, + "loss": 1.647, + "step": 11060 + }, + { + "epoch": 2.5780158360503025, + "grad_norm": 25.205806732177734, + "learning_rate": 0.0, + "loss": 1.0534, + "step": 11070 + }, + { + "epoch": 2.5803446669771777, + "grad_norm": 31.481040954589844, + "learning_rate": 0.0, + "loss": 1.038, + "step": 11080 + }, + { + "epoch": 2.582673497904052, + "grad_norm": 8.334689140319824, + "learning_rate": 0.0, + "loss": 1.0556, + "step": 11090 + }, + { + "epoch": 2.5850023288309267, + "grad_norm": 12.273963928222656, + "learning_rate": 0.0, + "loss": 1.6612, + "step": 11100 + }, + { + "epoch": 2.5873311597578015, + "grad_norm": 29.150667190551758, + "learning_rate": 0.0, + "loss": 1.0504, + "step": 11110 + }, + { + "epoch": 2.5896599906846762, + "grad_norm": 11.735727310180664, + "learning_rate": 0.0, + "loss": 0.7439, + "step": 11120 + }, + { + "epoch": 2.591988821611551, + "grad_norm": 23.498836517333984, + "learning_rate": 0.0, + "loss": 1.6355, + "step": 11130 + }, + { + "epoch": 2.5943176525384257, + "grad_norm": 117.8069076538086, + "learning_rate": 0.0, + "loss": 1.2481, + "step": 11140 + }, + { + "epoch": 2.5966464834653005, + "grad_norm": 10.995659828186035, + "learning_rate": 0.0, + "loss": 1.3976, + "step": 11150 + }, + { + "epoch": 2.598975314392175, + "grad_norm": 5.801515102386475, + "learning_rate": 0.0, + "loss": 1.3652, + "step": 11160 + }, + { + "epoch": 2.60130414531905, + "grad_norm": 5.238818168640137, + "learning_rate": 0.0, + "loss": 1.6681, + "step": 11170 + }, + { + "epoch": 2.6036329762459243, + "grad_norm": 28.656795501708984, + "learning_rate": 0.0, + "loss": 0.8382, + "step": 11180 + }, + { + "epoch": 2.6059618071727995, + "grad_norm": 8.944382667541504, + "learning_rate": 0.0, + "loss": 1.1245, + "step": 11190 + }, + { + "epoch": 2.6082906380996738, + "grad_norm": 18.96076774597168, + "learning_rate": 0.0, + "loss": 1.0715, + "step": 11200 + }, + { + "epoch": 2.6106194690265485, + "grad_norm": 6.842470169067383, + "learning_rate": 0.0, + "loss": 1.1061, + "step": 11210 + }, + { + "epoch": 2.6129482999534233, + "grad_norm": 5.037587642669678, + "learning_rate": 0.0, + "loss": 0.8291, + "step": 11220 + }, + { + "epoch": 2.615277130880298, + "grad_norm": 33.31461715698242, + "learning_rate": 0.0, + "loss": 1.1437, + "step": 11230 + }, + { + "epoch": 2.6176059618071728, + "grad_norm": 5.52734375, + "learning_rate": 0.0, + "loss": 1.1537, + "step": 11240 + }, + { + "epoch": 2.6199347927340475, + "grad_norm": 5.281623363494873, + "learning_rate": 0.0, + "loss": 1.483, + "step": 11250 + }, + { + "epoch": 2.6222636236609222, + "grad_norm": 6.426329135894775, + "learning_rate": 0.0, + "loss": 1.2722, + "step": 11260 + }, + { + "epoch": 2.624592454587797, + "grad_norm": 8.326626777648926, + "learning_rate": 0.0, + "loss": 1.2723, + "step": 11270 + }, + { + "epoch": 2.6269212855146717, + "grad_norm": 10.69015121459961, + "learning_rate": 0.0, + "loss": 1.4028, + "step": 11280 + }, + { + "epoch": 2.6292501164415465, + "grad_norm": 73.84967803955078, + "learning_rate": 0.0, + "loss": 1.2928, + "step": 11290 + }, + { + "epoch": 2.6315789473684212, + "grad_norm": 8.578020095825195, + "learning_rate": 0.0, + "loss": 1.137, + "step": 11300 + }, + { + "epoch": 2.6339077782952955, + "grad_norm": 84.88838195800781, + "learning_rate": 0.0, + "loss": 1.819, + "step": 11310 + }, + { + "epoch": 2.6362366092221707, + "grad_norm": 10.036768913269043, + "learning_rate": 0.0, + "loss": 1.3546, + "step": 11320 + }, + { + "epoch": 2.638565440149045, + "grad_norm": 8.671600341796875, + "learning_rate": 0.0, + "loss": 1.06, + "step": 11330 + }, + { + "epoch": 2.64089427107592, + "grad_norm": 9.737736701965332, + "learning_rate": 0.0, + "loss": 1.3624, + "step": 11340 + }, + { + "epoch": 2.6432231020027945, + "grad_norm": 27.788164138793945, + "learning_rate": 0.0, + "loss": 1.4631, + "step": 11350 + }, + { + "epoch": 2.6455519329296693, + "grad_norm": 15.201400756835938, + "learning_rate": 0.0, + "loss": 0.9118, + "step": 11360 + }, + { + "epoch": 2.647880763856544, + "grad_norm": 23.824310302734375, + "learning_rate": 0.0, + "loss": 1.4152, + "step": 11370 + }, + { + "epoch": 2.6502095947834188, + "grad_norm": 14.977895736694336, + "learning_rate": 0.0, + "loss": 1.2898, + "step": 11380 + }, + { + "epoch": 2.6525384257102935, + "grad_norm": 12.554730415344238, + "learning_rate": 0.0, + "loss": 1.6473, + "step": 11390 + }, + { + "epoch": 2.6548672566371683, + "grad_norm": 11.867571830749512, + "learning_rate": 0.0, + "loss": 0.9147, + "step": 11400 + }, + { + "epoch": 2.657196087564043, + "grad_norm": 23.419448852539062, + "learning_rate": 0.0, + "loss": 1.6672, + "step": 11410 + }, + { + "epoch": 2.6595249184909173, + "grad_norm": 10.939473152160645, + "learning_rate": 0.0, + "loss": 1.1738, + "step": 11420 + }, + { + "epoch": 2.6618537494177925, + "grad_norm": 47.00886535644531, + "learning_rate": 0.0, + "loss": 1.4211, + "step": 11430 + }, + { + "epoch": 2.664182580344667, + "grad_norm": 9.351873397827148, + "learning_rate": 0.0, + "loss": 1.0582, + "step": 11440 + }, + { + "epoch": 2.6665114112715416, + "grad_norm": 19.46343994140625, + "learning_rate": 0.0, + "loss": 1.6338, + "step": 11450 + }, + { + "epoch": 2.6688402421984163, + "grad_norm": 28.46370506286621, + "learning_rate": 0.0, + "loss": 0.9037, + "step": 11460 + }, + { + "epoch": 2.671169073125291, + "grad_norm": 15.675468444824219, + "learning_rate": 0.0, + "loss": 1.2858, + "step": 11470 + }, + { + "epoch": 2.673497904052166, + "grad_norm": 38.96082305908203, + "learning_rate": 0.0, + "loss": 1.2323, + "step": 11480 + }, + { + "epoch": 2.6758267349790406, + "grad_norm": 77.10676574707031, + "learning_rate": 0.0, + "loss": 1.2334, + "step": 11490 + }, + { + "epoch": 2.6781555659059153, + "grad_norm": 10.799468994140625, + "learning_rate": 0.0, + "loss": 1.2866, + "step": 11500 + }, + { + "epoch": 2.68048439683279, + "grad_norm": 5.765465259552002, + "learning_rate": 0.0, + "loss": 0.9091, + "step": 11510 + }, + { + "epoch": 2.682813227759665, + "grad_norm": 17.54412078857422, + "learning_rate": 0.0, + "loss": 1.2162, + "step": 11520 + }, + { + "epoch": 2.685142058686539, + "grad_norm": 9.990829467773438, + "learning_rate": 0.0, + "loss": 0.9272, + "step": 11530 + }, + { + "epoch": 2.6874708896134143, + "grad_norm": 11.573451042175293, + "learning_rate": 0.0, + "loss": 1.3428, + "step": 11540 + }, + { + "epoch": 2.6897997205402886, + "grad_norm": 7.813601016998291, + "learning_rate": 0.0, + "loss": 1.3985, + "step": 11550 + }, + { + "epoch": 2.6921285514671633, + "grad_norm": 44.93555450439453, + "learning_rate": 0.0, + "loss": 1.5427, + "step": 11560 + }, + { + "epoch": 2.694457382394038, + "grad_norm": 66.35843658447266, + "learning_rate": 0.0, + "loss": 1.8103, + "step": 11570 + }, + { + "epoch": 2.696786213320913, + "grad_norm": 12.34382152557373, + "learning_rate": 0.0, + "loss": 1.2642, + "step": 11580 + }, + { + "epoch": 2.6991150442477876, + "grad_norm": 13.233556747436523, + "learning_rate": 0.0, + "loss": 0.9407, + "step": 11590 + }, + { + "epoch": 2.7014438751746623, + "grad_norm": 11.109960556030273, + "learning_rate": 0.0, + "loss": 1.0398, + "step": 11600 + }, + { + "epoch": 2.703772706101537, + "grad_norm": 18.077089309692383, + "learning_rate": 0.0, + "loss": 0.9129, + "step": 11610 + }, + { + "epoch": 2.706101537028412, + "grad_norm": 14.020467758178711, + "learning_rate": 0.0, + "loss": 1.0783, + "step": 11620 + }, + { + "epoch": 2.7084303679552866, + "grad_norm": 11.337139129638672, + "learning_rate": 0.0, + "loss": 1.5425, + "step": 11630 + }, + { + "epoch": 2.710759198882161, + "grad_norm": 20.209257125854492, + "learning_rate": 0.0, + "loss": 1.4743, + "step": 11640 + }, + { + "epoch": 2.713088029809036, + "grad_norm": 10.780107498168945, + "learning_rate": 0.0, + "loss": 1.8225, + "step": 11650 + }, + { + "epoch": 2.7154168607359104, + "grad_norm": 6.723607540130615, + "learning_rate": 0.0, + "loss": 1.4223, + "step": 11660 + }, + { + "epoch": 2.717745691662785, + "grad_norm": 28.61166000366211, + "learning_rate": 0.0, + "loss": 1.0589, + "step": 11670 + }, + { + "epoch": 2.72007452258966, + "grad_norm": 11.463347434997559, + "learning_rate": 0.0, + "loss": 1.1931, + "step": 11680 + }, + { + "epoch": 2.7224033535165346, + "grad_norm": 14.94326400756836, + "learning_rate": 0.0, + "loss": 1.0032, + "step": 11690 + }, + { + "epoch": 2.7247321844434094, + "grad_norm": 13.956268310546875, + "learning_rate": 0.0, + "loss": 0.8667, + "step": 11700 + }, + { + "epoch": 2.727061015370284, + "grad_norm": 50.78772735595703, + "learning_rate": 0.0, + "loss": 1.3266, + "step": 11710 + }, + { + "epoch": 2.729389846297159, + "grad_norm": 3.0896995067596436, + "learning_rate": 0.0, + "loss": 0.7148, + "step": 11720 + }, + { + "epoch": 2.7317186772240336, + "grad_norm": 27.836071014404297, + "learning_rate": 0.0, + "loss": 1.3137, + "step": 11730 + }, + { + "epoch": 2.7340475081509084, + "grad_norm": 7.0214715003967285, + "learning_rate": 0.0, + "loss": 0.9149, + "step": 11740 + }, + { + "epoch": 2.736376339077783, + "grad_norm": 21.97905158996582, + "learning_rate": 0.0, + "loss": 1.5909, + "step": 11750 + }, + { + "epoch": 2.738705170004658, + "grad_norm": 6.0453338623046875, + "learning_rate": 0.0, + "loss": 1.1069, + "step": 11760 + }, + { + "epoch": 2.741034000931532, + "grad_norm": 15.303627967834473, + "learning_rate": 0.0, + "loss": 1.1191, + "step": 11770 + }, + { + "epoch": 2.7433628318584073, + "grad_norm": 25.093414306640625, + "learning_rate": 0.0, + "loss": 1.4118, + "step": 11780 + }, + { + "epoch": 2.7456916627852817, + "grad_norm": 20.566513061523438, + "learning_rate": 0.0, + "loss": 1.7382, + "step": 11790 + }, + { + "epoch": 2.7480204937121564, + "grad_norm": 44.23474884033203, + "learning_rate": 0.0, + "loss": 1.36, + "step": 11800 + }, + { + "epoch": 2.750349324639031, + "grad_norm": 8.558969497680664, + "learning_rate": 0.0, + "loss": 1.4381, + "step": 11810 + }, + { + "epoch": 2.752678155565906, + "grad_norm": 15.822858810424805, + "learning_rate": 0.0, + "loss": 1.1273, + "step": 11820 + }, + { + "epoch": 2.7550069864927806, + "grad_norm": 7.357201099395752, + "learning_rate": 0.0, + "loss": 1.1263, + "step": 11830 + }, + { + "epoch": 2.7573358174196554, + "grad_norm": 74.23464965820312, + "learning_rate": 0.0, + "loss": 1.187, + "step": 11840 + }, + { + "epoch": 2.75966464834653, + "grad_norm": 23.02132797241211, + "learning_rate": 0.0, + "loss": 1.0388, + "step": 11850 + }, + { + "epoch": 2.761993479273405, + "grad_norm": 8.519207954406738, + "learning_rate": 0.0, + "loss": 0.9178, + "step": 11860 + }, + { + "epoch": 2.7643223102002796, + "grad_norm": 8.386043548583984, + "learning_rate": 0.0, + "loss": 1.3762, + "step": 11870 + }, + { + "epoch": 2.766651141127154, + "grad_norm": 12.414632797241211, + "learning_rate": 0.0, + "loss": 2.2633, + "step": 11880 + }, + { + "epoch": 2.768979972054029, + "grad_norm": 10.087637901306152, + "learning_rate": 0.0, + "loss": 1.366, + "step": 11890 + }, + { + "epoch": 2.7713088029809034, + "grad_norm": 93.13560485839844, + "learning_rate": 0.0, + "loss": 1.6345, + "step": 11900 + }, + { + "epoch": 2.773637633907778, + "grad_norm": 51.69975280761719, + "learning_rate": 0.0, + "loss": 1.0775, + "step": 11910 + }, + { + "epoch": 2.775966464834653, + "grad_norm": 47.3151741027832, + "learning_rate": 0.0, + "loss": 1.3134, + "step": 11920 + }, + { + "epoch": 2.7782952957615277, + "grad_norm": 16.13167381286621, + "learning_rate": 0.0, + "loss": 1.72, + "step": 11930 + }, + { + "epoch": 2.7806241266884024, + "grad_norm": 60.05651092529297, + "learning_rate": 0.0, + "loss": 1.5407, + "step": 11940 + }, + { + "epoch": 2.782952957615277, + "grad_norm": 7.127047538757324, + "learning_rate": 0.0, + "loss": 0.826, + "step": 11950 + }, + { + "epoch": 2.785281788542152, + "grad_norm": 8.19342041015625, + "learning_rate": 0.0, + "loss": 1.2436, + "step": 11960 + }, + { + "epoch": 2.7876106194690267, + "grad_norm": 17.038658142089844, + "learning_rate": 0.0, + "loss": 1.3329, + "step": 11970 + }, + { + "epoch": 2.7899394503959014, + "grad_norm": 45.72443771362305, + "learning_rate": 0.0, + "loss": 1.092, + "step": 11980 + }, + { + "epoch": 2.7922682813227757, + "grad_norm": 8.771135330200195, + "learning_rate": 0.0, + "loss": 1.1903, + "step": 11990 + }, + { + "epoch": 2.794597112249651, + "grad_norm": 8.184369087219238, + "learning_rate": 0.0, + "loss": 1.4032, + "step": 12000 + }, + { + "epoch": 2.796925943176525, + "grad_norm": 22.36016082763672, + "learning_rate": 0.0, + "loss": 1.0558, + "step": 12010 + }, + { + "epoch": 2.7992547741034, + "grad_norm": 6.70443058013916, + "learning_rate": 0.0, + "loss": 0.9546, + "step": 12020 + }, + { + "epoch": 2.8015836050302747, + "grad_norm": 10.176080703735352, + "learning_rate": 0.0, + "loss": 1.1962, + "step": 12030 + }, + { + "epoch": 2.8039124359571495, + "grad_norm": 17.690227508544922, + "learning_rate": 0.0, + "loss": 1.1685, + "step": 12040 + }, + { + "epoch": 2.806241266884024, + "grad_norm": 17.06946563720703, + "learning_rate": 0.0, + "loss": 1.1656, + "step": 12050 + }, + { + "epoch": 2.808570097810899, + "grad_norm": 6.972208023071289, + "learning_rate": 0.0, + "loss": 1.434, + "step": 12060 + }, + { + "epoch": 2.8108989287377737, + "grad_norm": 17.20701026916504, + "learning_rate": 0.0, + "loss": 1.1218, + "step": 12070 + }, + { + "epoch": 2.8132277596646484, + "grad_norm": 9.323033332824707, + "learning_rate": 0.0, + "loss": 1.1368, + "step": 12080 + }, + { + "epoch": 2.815556590591523, + "grad_norm": 26.10862159729004, + "learning_rate": 0.0, + "loss": 1.1541, + "step": 12090 + }, + { + "epoch": 2.8178854215183975, + "grad_norm": 12.300597190856934, + "learning_rate": 0.0, + "loss": 1.3575, + "step": 12100 + }, + { + "epoch": 2.8202142524452727, + "grad_norm": 7.925255298614502, + "learning_rate": 0.0, + "loss": 2.2617, + "step": 12110 + }, + { + "epoch": 2.822543083372147, + "grad_norm": 5.388205528259277, + "learning_rate": 0.0, + "loss": 0.9115, + "step": 12120 + }, + { + "epoch": 2.8248719142990217, + "grad_norm": 9.119213104248047, + "learning_rate": 0.0, + "loss": 1.5579, + "step": 12130 + }, + { + "epoch": 2.8272007452258965, + "grad_norm": 127.05992126464844, + "learning_rate": 0.0, + "loss": 1.0463, + "step": 12140 + }, + { + "epoch": 2.8295295761527712, + "grad_norm": 4.77928352355957, + "learning_rate": 0.0, + "loss": 0.8053, + "step": 12150 + }, + { + "epoch": 2.831858407079646, + "grad_norm": 10.260478973388672, + "learning_rate": 0.0, + "loss": 1.3177, + "step": 12160 + }, + { + "epoch": 2.8341872380065207, + "grad_norm": 6.844449996948242, + "learning_rate": 0.0, + "loss": 1.3177, + "step": 12170 + }, + { + "epoch": 2.8365160689333955, + "grad_norm": 65.62310791015625, + "learning_rate": 0.0, + "loss": 1.1177, + "step": 12180 + }, + { + "epoch": 2.83884489986027, + "grad_norm": 19.49369239807129, + "learning_rate": 0.0, + "loss": 1.2459, + "step": 12190 + }, + { + "epoch": 2.841173730787145, + "grad_norm": 17.212875366210938, + "learning_rate": 0.0, + "loss": 1.4008, + "step": 12200 + }, + { + "epoch": 2.8435025617140197, + "grad_norm": 31.808021545410156, + "learning_rate": 0.0, + "loss": 1.2806, + "step": 12210 + }, + { + "epoch": 2.8458313926408945, + "grad_norm": 9.024837493896484, + "learning_rate": 0.0, + "loss": 1.629, + "step": 12220 + }, + { + "epoch": 2.8481602235677688, + "grad_norm": 26.87761116027832, + "learning_rate": 0.0, + "loss": 1.2044, + "step": 12230 + }, + { + "epoch": 2.850489054494644, + "grad_norm": 19.435070037841797, + "learning_rate": 0.0, + "loss": 1.1629, + "step": 12240 + }, + { + "epoch": 2.8528178854215183, + "grad_norm": 15.5372953414917, + "learning_rate": 0.0, + "loss": 1.684, + "step": 12250 + }, + { + "epoch": 2.855146716348393, + "grad_norm": 4.47623872756958, + "learning_rate": 0.0, + "loss": 1.0723, + "step": 12260 + }, + { + "epoch": 2.8574755472752678, + "grad_norm": 75.65230560302734, + "learning_rate": 0.0, + "loss": 1.2155, + "step": 12270 + }, + { + "epoch": 2.8598043782021425, + "grad_norm": 8.398263931274414, + "learning_rate": 0.0, + "loss": 1.2201, + "step": 12280 + }, + { + "epoch": 2.8621332091290173, + "grad_norm": 9.25592041015625, + "learning_rate": 0.0, + "loss": 1.1723, + "step": 12290 + }, + { + "epoch": 2.864462040055892, + "grad_norm": 10.627755165100098, + "learning_rate": 0.0, + "loss": 1.167, + "step": 12300 + }, + { + "epoch": 2.8667908709827667, + "grad_norm": 13.692951202392578, + "learning_rate": 0.0, + "loss": 1.0071, + "step": 12310 + }, + { + "epoch": 2.8691197019096415, + "grad_norm": 11.026177406311035, + "learning_rate": 0.0, + "loss": 1.0744, + "step": 12320 + }, + { + "epoch": 2.8714485328365162, + "grad_norm": 5.878359317779541, + "learning_rate": 0.0, + "loss": 1.3973, + "step": 12330 + }, + { + "epoch": 2.8737773637633905, + "grad_norm": 41.004493713378906, + "learning_rate": 0.0, + "loss": 1.2759, + "step": 12340 + }, + { + "epoch": 2.8761061946902657, + "grad_norm": 11.279379844665527, + "learning_rate": 0.0, + "loss": 1.3207, + "step": 12350 + }, + { + "epoch": 2.87843502561714, + "grad_norm": 8.147149085998535, + "learning_rate": 0.0, + "loss": 1.1631, + "step": 12360 + }, + { + "epoch": 2.880763856544015, + "grad_norm": 10.414231300354004, + "learning_rate": 0.0, + "loss": 1.0849, + "step": 12370 + }, + { + "epoch": 2.8830926874708895, + "grad_norm": 16.238142013549805, + "learning_rate": 0.0, + "loss": 0.9335, + "step": 12380 + }, + { + "epoch": 2.8854215183977643, + "grad_norm": 12.68599796295166, + "learning_rate": 0.0, + "loss": 1.0928, + "step": 12390 + }, + { + "epoch": 2.887750349324639, + "grad_norm": 33.59490966796875, + "learning_rate": 0.0, + "loss": 1.1499, + "step": 12400 + }, + { + "epoch": 2.890079180251514, + "grad_norm": 4.714087963104248, + "learning_rate": 0.0, + "loss": 1.404, + "step": 12410 + }, + { + "epoch": 2.8924080111783885, + "grad_norm": 107.86946105957031, + "learning_rate": 0.0, + "loss": 1.5526, + "step": 12420 + }, + { + "epoch": 2.8947368421052633, + "grad_norm": 6.616478443145752, + "learning_rate": 0.0, + "loss": 1.0008, + "step": 12430 + }, + { + "epoch": 2.897065673032138, + "grad_norm": 10.684284210205078, + "learning_rate": 0.0, + "loss": 1.5301, + "step": 12440 + }, + { + "epoch": 2.8993945039590123, + "grad_norm": 41.24277114868164, + "learning_rate": 0.0, + "loss": 1.2296, + "step": 12450 + }, + { + "epoch": 2.9017233348858875, + "grad_norm": 72.13574981689453, + "learning_rate": 0.0, + "loss": 1.2495, + "step": 12460 + }, + { + "epoch": 2.904052165812762, + "grad_norm": 11.254340171813965, + "learning_rate": 0.0, + "loss": 1.2704, + "step": 12470 + }, + { + "epoch": 2.9063809967396366, + "grad_norm": 9.927990913391113, + "learning_rate": 0.0, + "loss": 1.4381, + "step": 12480 + }, + { + "epoch": 2.9087098276665113, + "grad_norm": 8.379936218261719, + "learning_rate": 0.0, + "loss": 1.3127, + "step": 12490 + }, + { + "epoch": 2.911038658593386, + "grad_norm": 6.6258955001831055, + "learning_rate": 0.0, + "loss": 1.0506, + "step": 12500 + }, + { + "epoch": 2.913367489520261, + "grad_norm": 13.267959594726562, + "learning_rate": 0.0, + "loss": 1.4729, + "step": 12510 + }, + { + "epoch": 2.9156963204471356, + "grad_norm": 14.169622421264648, + "learning_rate": 0.0, + "loss": 1.485, + "step": 12520 + }, + { + "epoch": 2.9180251513740103, + "grad_norm": 14.324849128723145, + "learning_rate": 0.0, + "loss": 1.3262, + "step": 12530 + }, + { + "epoch": 2.920353982300885, + "grad_norm": 139.77540588378906, + "learning_rate": 0.0, + "loss": 1.1871, + "step": 12540 + }, + { + "epoch": 2.92268281322776, + "grad_norm": 19.438474655151367, + "learning_rate": 0.0, + "loss": 1.0036, + "step": 12550 + }, + { + "epoch": 2.925011644154634, + "grad_norm": 5.162458419799805, + "learning_rate": 0.0, + "loss": 1.3947, + "step": 12560 + }, + { + "epoch": 2.9273404750815093, + "grad_norm": 9.66451644897461, + "learning_rate": 0.0, + "loss": 1.1011, + "step": 12570 + }, + { + "epoch": 2.9296693060083836, + "grad_norm": 10.375886917114258, + "learning_rate": 0.0, + "loss": 1.0098, + "step": 12580 + }, + { + "epoch": 2.9319981369352583, + "grad_norm": 13.053080558776855, + "learning_rate": 0.0, + "loss": 1.1087, + "step": 12590 + }, + { + "epoch": 2.934326967862133, + "grad_norm": 29.2081298828125, + "learning_rate": 0.0, + "loss": 1.0568, + "step": 12600 + }, + { + "epoch": 2.936655798789008, + "grad_norm": 8.402816772460938, + "learning_rate": 0.0, + "loss": 1.3496, + "step": 12610 + }, + { + "epoch": 2.9389846297158826, + "grad_norm": 12.343842506408691, + "learning_rate": 0.0, + "loss": 1.1508, + "step": 12620 + }, + { + "epoch": 2.9413134606427573, + "grad_norm": 6.9314494132995605, + "learning_rate": 0.0, + "loss": 1.6717, + "step": 12630 + }, + { + "epoch": 2.943642291569632, + "grad_norm": 18.752248764038086, + "learning_rate": 0.0, + "loss": 1.0485, + "step": 12640 + }, + { + "epoch": 2.945971122496507, + "grad_norm": 27.50632667541504, + "learning_rate": 0.0, + "loss": 1.3482, + "step": 12650 + }, + { + "epoch": 2.9482999534233816, + "grad_norm": 9.300987243652344, + "learning_rate": 0.0, + "loss": 1.2977, + "step": 12660 + }, + { + "epoch": 2.9506287843502563, + "grad_norm": 9.483819961547852, + "learning_rate": 0.0, + "loss": 1.2739, + "step": 12670 + }, + { + "epoch": 2.952957615277131, + "grad_norm": 14.791114807128906, + "learning_rate": 0.0, + "loss": 1.9703, + "step": 12680 + }, + { + "epoch": 2.9552864462040054, + "grad_norm": 14.320032119750977, + "learning_rate": 0.0, + "loss": 1.1498, + "step": 12690 + }, + { + "epoch": 2.9576152771308806, + "grad_norm": 10.36143684387207, + "learning_rate": 0.0, + "loss": 1.3839, + "step": 12700 + }, + { + "epoch": 2.959944108057755, + "grad_norm": 4.280786037445068, + "learning_rate": 0.0, + "loss": 1.1626, + "step": 12710 + }, + { + "epoch": 2.9622729389846296, + "grad_norm": 35.886634826660156, + "learning_rate": 0.0, + "loss": 1.6616, + "step": 12720 + }, + { + "epoch": 2.9646017699115044, + "grad_norm": 17.06854248046875, + "learning_rate": 0.0, + "loss": 1.0649, + "step": 12730 + }, + { + "epoch": 2.966930600838379, + "grad_norm": 11.25122356414795, + "learning_rate": 0.0, + "loss": 0.9763, + "step": 12740 + }, + { + "epoch": 2.969259431765254, + "grad_norm": 11.460931777954102, + "learning_rate": 0.0, + "loss": 1.4561, + "step": 12750 + }, + { + "epoch": 2.9715882626921286, + "grad_norm": 26.476194381713867, + "learning_rate": 0.0, + "loss": 1.814, + "step": 12760 + }, + { + "epoch": 2.9739170936190034, + "grad_norm": 19.238082885742188, + "learning_rate": 0.0, + "loss": 1.1605, + "step": 12770 + }, + { + "epoch": 2.976245924545878, + "grad_norm": 44.26625061035156, + "learning_rate": 0.0, + "loss": 1.1757, + "step": 12780 + }, + { + "epoch": 2.978574755472753, + "grad_norm": 98.12911224365234, + "learning_rate": 0.0, + "loss": 1.7095, + "step": 12790 + }, + { + "epoch": 2.980903586399627, + "grad_norm": 14.261837005615234, + "learning_rate": 0.0, + "loss": 1.3246, + "step": 12800 + }, + { + "epoch": 2.9832324173265024, + "grad_norm": 6.469521522521973, + "learning_rate": 0.0, + "loss": 1.2186, + "step": 12810 + }, + { + "epoch": 2.9855612482533767, + "grad_norm": 21.546524047851562, + "learning_rate": 0.0, + "loss": 1.3713, + "step": 12820 + }, + { + "epoch": 2.9878900791802514, + "grad_norm": 15.523602485656738, + "learning_rate": 0.0, + "loss": 1.46, + "step": 12830 + }, + { + "epoch": 2.990218910107126, + "grad_norm": 22.233688354492188, + "learning_rate": 0.0, + "loss": 1.3171, + "step": 12840 + }, + { + "epoch": 2.992547741034001, + "grad_norm": 10.238198280334473, + "learning_rate": 0.0, + "loss": 1.1856, + "step": 12850 + }, + { + "epoch": 2.9948765719608756, + "grad_norm": 32.13593292236328, + "learning_rate": 0.0, + "loss": 1.3003, + "step": 12860 + }, + { + "epoch": 2.9972054028877504, + "grad_norm": 13.810074806213379, + "learning_rate": 0.0, + "loss": 0.9724, + "step": 12870 + }, + { + "epoch": 2.999534233814625, + "grad_norm": 9.147926330566406, + "learning_rate": 0.0, + "loss": 1.052, + "step": 12880 + }, + { + "epoch": 3.0, + "eval_loss": 0.49088114500045776, + "eval_runtime": 15.2585, + "eval_samples_per_second": 285.48, + "eval_steps_per_second": 2.294, + "step": 12882 + }, + { + "epoch": 3.0018630647415, + "grad_norm": 21.883949279785156, + "learning_rate": 0.0, + "loss": 1.7736, + "step": 12890 + }, + { + "epoch": 3.0041918956683746, + "grad_norm": 7.900768280029297, + "learning_rate": 0.0, + "loss": 1.0961, + "step": 12900 + }, + { + "epoch": 3.0065207265952494, + "grad_norm": 70.86416625976562, + "learning_rate": 0.0, + "loss": 0.9499, + "step": 12910 + }, + { + "epoch": 3.0088495575221237, + "grad_norm": 9.933052062988281, + "learning_rate": 0.0, + "loss": 0.9614, + "step": 12920 + }, + { + "epoch": 3.0111783884489984, + "grad_norm": 15.00586986541748, + "learning_rate": 0.0, + "loss": 1.4427, + "step": 12930 + }, + { + "epoch": 3.013507219375873, + "grad_norm": 31.774045944213867, + "learning_rate": 0.0, + "loss": 1.567, + "step": 12940 + }, + { + "epoch": 3.015836050302748, + "grad_norm": 6.541329860687256, + "learning_rate": 0.0, + "loss": 1.0219, + "step": 12950 + }, + { + "epoch": 3.0181648812296227, + "grad_norm": 9.174057006835938, + "learning_rate": 0.0, + "loss": 1.3925, + "step": 12960 + }, + { + "epoch": 3.0204937121564974, + "grad_norm": 5.683773517608643, + "learning_rate": 0.0, + "loss": 1.072, + "step": 12970 + }, + { + "epoch": 3.022822543083372, + "grad_norm": 10.569469451904297, + "learning_rate": 0.0, + "loss": 1.3213, + "step": 12980 + }, + { + "epoch": 3.025151374010247, + "grad_norm": 11.428179740905762, + "learning_rate": 0.0, + "loss": 1.0445, + "step": 12990 + }, + { + "epoch": 3.0274802049371217, + "grad_norm": 5.285568714141846, + "learning_rate": 0.0, + "loss": 1.3264, + "step": 13000 + }, + { + "epoch": 3.0298090358639964, + "grad_norm": 24.80561065673828, + "learning_rate": 0.0, + "loss": 1.1625, + "step": 13010 + }, + { + "epoch": 3.032137866790871, + "grad_norm": 61.148956298828125, + "learning_rate": 0.0, + "loss": 1.0419, + "step": 13020 + }, + { + "epoch": 3.034466697717746, + "grad_norm": 20.153419494628906, + "learning_rate": 0.0, + "loss": 1.4652, + "step": 13030 + }, + { + "epoch": 3.03679552864462, + "grad_norm": 6.97303581237793, + "learning_rate": 0.0, + "loss": 1.2172, + "step": 13040 + }, + { + "epoch": 3.039124359571495, + "grad_norm": 8.286640167236328, + "learning_rate": 0.0, + "loss": 0.8866, + "step": 13050 + }, + { + "epoch": 3.0414531904983697, + "grad_norm": 6.162503242492676, + "learning_rate": 0.0, + "loss": 0.9947, + "step": 13060 + }, + { + "epoch": 3.0437820214252445, + "grad_norm": 32.659934997558594, + "learning_rate": 0.0, + "loss": 1.163, + "step": 13070 + }, + { + "epoch": 3.046110852352119, + "grad_norm": 24.297115325927734, + "learning_rate": 0.0, + "loss": 1.4345, + "step": 13080 + }, + { + "epoch": 3.048439683278994, + "grad_norm": 15.972943305969238, + "learning_rate": 0.0, + "loss": 0.8504, + "step": 13090 + }, + { + "epoch": 3.0507685142058687, + "grad_norm": 62.08342742919922, + "learning_rate": 0.0, + "loss": 1.1412, + "step": 13100 + }, + { + "epoch": 3.0530973451327434, + "grad_norm": 5.202879905700684, + "learning_rate": 0.0, + "loss": 1.2152, + "step": 13110 + }, + { + "epoch": 3.055426176059618, + "grad_norm": 5.640526294708252, + "learning_rate": 0.0, + "loss": 1.5043, + "step": 13120 + }, + { + "epoch": 3.057755006986493, + "grad_norm": 40.53738021850586, + "learning_rate": 0.0, + "loss": 1.3858, + "step": 13130 + }, + { + "epoch": 3.0600838379133677, + "grad_norm": 4.845460414886475, + "learning_rate": 0.0, + "loss": 1.0233, + "step": 13140 + }, + { + "epoch": 3.062412668840242, + "grad_norm": 4.9804158210754395, + "learning_rate": 0.0, + "loss": 1.4265, + "step": 13150 + }, + { + "epoch": 3.0647414997671167, + "grad_norm": 4.985668659210205, + "learning_rate": 0.0, + "loss": 1.4086, + "step": 13160 + }, + { + "epoch": 3.0670703306939915, + "grad_norm": 19.393199920654297, + "learning_rate": 0.0, + "loss": 0.9075, + "step": 13170 + }, + { + "epoch": 3.0693991616208662, + "grad_norm": 16.305225372314453, + "learning_rate": 0.0, + "loss": 1.5215, + "step": 13180 + }, + { + "epoch": 3.071727992547741, + "grad_norm": 4.610240936279297, + "learning_rate": 0.0, + "loss": 0.9681, + "step": 13190 + }, + { + "epoch": 3.0740568234746157, + "grad_norm": 3.354733467102051, + "learning_rate": 0.0, + "loss": 1.1587, + "step": 13200 + }, + { + "epoch": 3.0763856544014905, + "grad_norm": 4.949262619018555, + "learning_rate": 0.0, + "loss": 0.9106, + "step": 13210 + }, + { + "epoch": 3.0787144853283652, + "grad_norm": 12.779787063598633, + "learning_rate": 0.0, + "loss": 0.9381, + "step": 13220 + }, + { + "epoch": 3.08104331625524, + "grad_norm": 5.218024253845215, + "learning_rate": 0.0, + "loss": 0.8552, + "step": 13230 + }, + { + "epoch": 3.0833721471821147, + "grad_norm": 32.33806228637695, + "learning_rate": 0.0, + "loss": 1.4397, + "step": 13240 + }, + { + "epoch": 3.0857009781089895, + "grad_norm": 11.891190528869629, + "learning_rate": 0.0, + "loss": 1.3937, + "step": 13250 + }, + { + "epoch": 3.0880298090358638, + "grad_norm": 9.155755996704102, + "learning_rate": 0.0, + "loss": 1.0606, + "step": 13260 + }, + { + "epoch": 3.0903586399627385, + "grad_norm": 13.458362579345703, + "learning_rate": 0.0, + "loss": 1.6445, + "step": 13270 + }, + { + "epoch": 3.0926874708896133, + "grad_norm": 6.68931770324707, + "learning_rate": 0.0, + "loss": 1.1771, + "step": 13280 + }, + { + "epoch": 3.095016301816488, + "grad_norm": 19.457380294799805, + "learning_rate": 0.0, + "loss": 1.6591, + "step": 13290 + }, + { + "epoch": 3.0973451327433628, + "grad_norm": 7.39839506149292, + "learning_rate": 0.0, + "loss": 1.587, + "step": 13300 + }, + { + "epoch": 3.0996739636702375, + "grad_norm": 11.434883117675781, + "learning_rate": 0.0, + "loss": 1.0876, + "step": 13310 + }, + { + "epoch": 3.1020027945971123, + "grad_norm": 42.85976028442383, + "learning_rate": 0.0, + "loss": 1.3607, + "step": 13320 + }, + { + "epoch": 3.104331625523987, + "grad_norm": 7.510742664337158, + "learning_rate": 0.0, + "loss": 1.5981, + "step": 13330 + }, + { + "epoch": 3.1066604564508618, + "grad_norm": 9.284518241882324, + "learning_rate": 0.0, + "loss": 1.1812, + "step": 13340 + }, + { + "epoch": 3.1089892873777365, + "grad_norm": 8.393990516662598, + "learning_rate": 0.0, + "loss": 1.3099, + "step": 13350 + }, + { + "epoch": 3.1113181183046112, + "grad_norm": 8.153864860534668, + "learning_rate": 0.0, + "loss": 0.8854, + "step": 13360 + }, + { + "epoch": 3.113646949231486, + "grad_norm": 15.750825881958008, + "learning_rate": 0.0, + "loss": 1.7133, + "step": 13370 + }, + { + "epoch": 3.1159757801583603, + "grad_norm": 52.85032272338867, + "learning_rate": 0.0, + "loss": 1.0101, + "step": 13380 + }, + { + "epoch": 3.118304611085235, + "grad_norm": 6.499885559082031, + "learning_rate": 0.0, + "loss": 1.5199, + "step": 13390 + }, + { + "epoch": 3.12063344201211, + "grad_norm": 63.261356353759766, + "learning_rate": 0.0, + "loss": 1.3876, + "step": 13400 + }, + { + "epoch": 3.1229622729389845, + "grad_norm": 14.717015266418457, + "learning_rate": 0.0, + "loss": 1.735, + "step": 13410 + }, + { + "epoch": 3.1252911038658593, + "grad_norm": 5.445131778717041, + "learning_rate": 0.0, + "loss": 1.337, + "step": 13420 + }, + { + "epoch": 3.127619934792734, + "grad_norm": 35.080650329589844, + "learning_rate": 0.0, + "loss": 1.2368, + "step": 13430 + }, + { + "epoch": 3.129948765719609, + "grad_norm": 9.115435600280762, + "learning_rate": 0.0, + "loss": 0.7256, + "step": 13440 + }, + { + "epoch": 3.1322775966464835, + "grad_norm": 5.406123638153076, + "learning_rate": 0.0, + "loss": 0.8733, + "step": 13450 + }, + { + "epoch": 3.1346064275733583, + "grad_norm": 5.928583145141602, + "learning_rate": 0.0, + "loss": 0.9502, + "step": 13460 + }, + { + "epoch": 3.136935258500233, + "grad_norm": 4.724367618560791, + "learning_rate": 0.0, + "loss": 1.0116, + "step": 13470 + }, + { + "epoch": 3.1392640894271078, + "grad_norm": 22.139331817626953, + "learning_rate": 0.0, + "loss": 0.7872, + "step": 13480 + }, + { + "epoch": 3.1415929203539825, + "grad_norm": 6.987404823303223, + "learning_rate": 0.0, + "loss": 1.1415, + "step": 13490 + }, + { + "epoch": 3.143921751280857, + "grad_norm": 23.272977828979492, + "learning_rate": 0.0, + "loss": 1.1225, + "step": 13500 + }, + { + "epoch": 3.1462505822077316, + "grad_norm": 14.455467224121094, + "learning_rate": 0.0, + "loss": 1.449, + "step": 13510 + }, + { + "epoch": 3.1485794131346063, + "grad_norm": 23.383798599243164, + "learning_rate": 0.0, + "loss": 1.2972, + "step": 13520 + }, + { + "epoch": 3.150908244061481, + "grad_norm": 6.212050437927246, + "learning_rate": 0.0, + "loss": 1.0594, + "step": 13530 + }, + { + "epoch": 3.153237074988356, + "grad_norm": 5.7028422355651855, + "learning_rate": 0.0, + "loss": 0.9708, + "step": 13540 + }, + { + "epoch": 3.1555659059152306, + "grad_norm": 23.22449493408203, + "learning_rate": 0.0, + "loss": 0.9153, + "step": 13550 + }, + { + "epoch": 3.1578947368421053, + "grad_norm": 11.004061698913574, + "learning_rate": 0.0, + "loss": 0.8146, + "step": 13560 + }, + { + "epoch": 3.16022356776898, + "grad_norm": 12.053899765014648, + "learning_rate": 0.0, + "loss": 1.0754, + "step": 13570 + }, + { + "epoch": 3.162552398695855, + "grad_norm": 6.76538610458374, + "learning_rate": 0.0, + "loss": 1.383, + "step": 13580 + }, + { + "epoch": 3.1648812296227296, + "grad_norm": 17.094440460205078, + "learning_rate": 0.0, + "loss": 1.3829, + "step": 13590 + }, + { + "epoch": 3.1672100605496043, + "grad_norm": 91.07170867919922, + "learning_rate": 0.0, + "loss": 1.7515, + "step": 13600 + }, + { + "epoch": 3.1695388914764786, + "grad_norm": 5.213118076324463, + "learning_rate": 0.0, + "loss": 1.1723, + "step": 13610 + }, + { + "epoch": 3.1718677224033534, + "grad_norm": 85.04914855957031, + "learning_rate": 0.0, + "loss": 1.4981, + "step": 13620 + }, + { + "epoch": 3.174196553330228, + "grad_norm": 12.307666778564453, + "learning_rate": 0.0, + "loss": 1.4268, + "step": 13630 + }, + { + "epoch": 3.176525384257103, + "grad_norm": 9.055723190307617, + "learning_rate": 0.0, + "loss": 1.3236, + "step": 13640 + }, + { + "epoch": 3.1788542151839776, + "grad_norm": 13.058332443237305, + "learning_rate": 0.0, + "loss": 1.6913, + "step": 13650 + }, + { + "epoch": 3.1811830461108523, + "grad_norm": 12.987732887268066, + "learning_rate": 0.0, + "loss": 0.8965, + "step": 13660 + }, + { + "epoch": 3.183511877037727, + "grad_norm": 13.6783447265625, + "learning_rate": 0.0, + "loss": 1.252, + "step": 13670 + }, + { + "epoch": 3.185840707964602, + "grad_norm": 9.275946617126465, + "learning_rate": 0.0, + "loss": 1.1708, + "step": 13680 + }, + { + "epoch": 3.1881695388914766, + "grad_norm": 6.581894874572754, + "learning_rate": 0.0, + "loss": 0.994, + "step": 13690 + }, + { + "epoch": 3.1904983698183513, + "grad_norm": 9.445444107055664, + "learning_rate": 0.0, + "loss": 1.0164, + "step": 13700 + }, + { + "epoch": 3.192827200745226, + "grad_norm": 9.902006149291992, + "learning_rate": 0.0, + "loss": 1.1095, + "step": 13710 + }, + { + "epoch": 3.1951560316721004, + "grad_norm": 7.1675944328308105, + "learning_rate": 0.0, + "loss": 1.4237, + "step": 13720 + }, + { + "epoch": 3.197484862598975, + "grad_norm": 9.444397926330566, + "learning_rate": 0.0, + "loss": 1.3244, + "step": 13730 + }, + { + "epoch": 3.19981369352585, + "grad_norm": 5.586214065551758, + "learning_rate": 0.0, + "loss": 0.9574, + "step": 13740 + }, + { + "epoch": 3.2021425244527246, + "grad_norm": 5.1344218254089355, + "learning_rate": 0.0, + "loss": 1.5185, + "step": 13750 + }, + { + "epoch": 3.2044713553795994, + "grad_norm": 17.44362449645996, + "learning_rate": 0.0, + "loss": 0.8098, + "step": 13760 + }, + { + "epoch": 3.206800186306474, + "grad_norm": 13.0212984085083, + "learning_rate": 0.0, + "loss": 1.0167, + "step": 13770 + }, + { + "epoch": 3.209129017233349, + "grad_norm": 8.9563570022583, + "learning_rate": 0.0, + "loss": 1.357, + "step": 13780 + }, + { + "epoch": 3.2114578481602236, + "grad_norm": 17.59966468811035, + "learning_rate": 0.0, + "loss": 1.4481, + "step": 13790 + }, + { + "epoch": 3.2137866790870984, + "grad_norm": 46.96272659301758, + "learning_rate": 0.0, + "loss": 1.1264, + "step": 13800 + }, + { + "epoch": 3.216115510013973, + "grad_norm": 5.420903205871582, + "learning_rate": 0.0, + "loss": 0.8953, + "step": 13810 + }, + { + "epoch": 3.218444340940848, + "grad_norm": 3.29380202293396, + "learning_rate": 0.0, + "loss": 1.1815, + "step": 13820 + }, + { + "epoch": 3.2207731718677226, + "grad_norm": 5.43313455581665, + "learning_rate": 0.0, + "loss": 1.3487, + "step": 13830 + }, + { + "epoch": 3.223102002794597, + "grad_norm": 6.049877166748047, + "learning_rate": 0.0, + "loss": 0.9866, + "step": 13840 + }, + { + "epoch": 3.2254308337214717, + "grad_norm": 12.948495864868164, + "learning_rate": 0.0, + "loss": 1.2618, + "step": 13850 + }, + { + "epoch": 3.2277596646483464, + "grad_norm": 11.312189102172852, + "learning_rate": 0.0, + "loss": 1.2437, + "step": 13860 + }, + { + "epoch": 3.230088495575221, + "grad_norm": 11.164241790771484, + "learning_rate": 0.0, + "loss": 1.9565, + "step": 13870 + }, + { + "epoch": 3.232417326502096, + "grad_norm": 6.2988176345825195, + "learning_rate": 0.0, + "loss": 0.9325, + "step": 13880 + }, + { + "epoch": 3.2347461574289706, + "grad_norm": 8.579294204711914, + "learning_rate": 0.0, + "loss": 1.1787, + "step": 13890 + }, + { + "epoch": 3.2370749883558454, + "grad_norm": 7.805800437927246, + "learning_rate": 0.0, + "loss": 0.992, + "step": 13900 + }, + { + "epoch": 3.23940381928272, + "grad_norm": 8.043397903442383, + "learning_rate": 0.0, + "loss": 1.6505, + "step": 13910 + }, + { + "epoch": 3.241732650209595, + "grad_norm": 51.754547119140625, + "learning_rate": 0.0, + "loss": 1.076, + "step": 13920 + }, + { + "epoch": 3.2440614811364696, + "grad_norm": 15.905028343200684, + "learning_rate": 0.0, + "loss": 1.4288, + "step": 13930 + }, + { + "epoch": 3.2463903120633444, + "grad_norm": 8.984749794006348, + "learning_rate": 0.0, + "loss": 1.0526, + "step": 13940 + }, + { + "epoch": 3.248719142990219, + "grad_norm": 8.098136901855469, + "learning_rate": 0.0, + "loss": 1.0857, + "step": 13950 + }, + { + "epoch": 3.2510479739170934, + "grad_norm": 52.77598190307617, + "learning_rate": 0.0, + "loss": 1.1508, + "step": 13960 + }, + { + "epoch": 3.253376804843968, + "grad_norm": 9.703697204589844, + "learning_rate": 0.0, + "loss": 1.1592, + "step": 13970 + }, + { + "epoch": 3.255705635770843, + "grad_norm": 41.5420036315918, + "learning_rate": 0.0, + "loss": 1.4785, + "step": 13980 + }, + { + "epoch": 3.2580344666977177, + "grad_norm": 9.39918041229248, + "learning_rate": 0.0, + "loss": 1.4824, + "step": 13990 + }, + { + "epoch": 3.2603632976245924, + "grad_norm": 7.907922267913818, + "learning_rate": 0.0, + "loss": 0.9171, + "step": 14000 + }, + { + "epoch": 3.262692128551467, + "grad_norm": 9.973319053649902, + "learning_rate": 0.0, + "loss": 0.9349, + "step": 14010 + }, + { + "epoch": 3.265020959478342, + "grad_norm": 10.504914283752441, + "learning_rate": 0.0, + "loss": 1.0402, + "step": 14020 + }, + { + "epoch": 3.2673497904052167, + "grad_norm": 8.200088500976562, + "learning_rate": 0.0, + "loss": 1.0022, + "step": 14030 + }, + { + "epoch": 3.2696786213320914, + "grad_norm": 33.09117889404297, + "learning_rate": 0.0, + "loss": 1.3235, + "step": 14040 + }, + { + "epoch": 3.272007452258966, + "grad_norm": 9.324043273925781, + "learning_rate": 0.0, + "loss": 1.0813, + "step": 14050 + }, + { + "epoch": 3.274336283185841, + "grad_norm": 13.463679313659668, + "learning_rate": 0.0, + "loss": 0.9718, + "step": 14060 + }, + { + "epoch": 3.276665114112715, + "grad_norm": 6.217166423797607, + "learning_rate": 0.0, + "loss": 1.0963, + "step": 14070 + }, + { + "epoch": 3.27899394503959, + "grad_norm": 60.56024169921875, + "learning_rate": 0.0, + "loss": 1.0586, + "step": 14080 + }, + { + "epoch": 3.2813227759664647, + "grad_norm": 37.3182373046875, + "learning_rate": 0.0, + "loss": 1.1455, + "step": 14090 + }, + { + "epoch": 3.2836516068933395, + "grad_norm": 5.846940040588379, + "learning_rate": 0.0, + "loss": 1.5728, + "step": 14100 + }, + { + "epoch": 3.285980437820214, + "grad_norm": 33.51951217651367, + "learning_rate": 0.0, + "loss": 1.2916, + "step": 14110 + }, + { + "epoch": 3.288309268747089, + "grad_norm": 7.1149797439575195, + "learning_rate": 0.0, + "loss": 0.8535, + "step": 14120 + }, + { + "epoch": 3.2906380996739637, + "grad_norm": 9.088951110839844, + "learning_rate": 0.0, + "loss": 1.0247, + "step": 14130 + }, + { + "epoch": 3.2929669306008384, + "grad_norm": 9.655440330505371, + "learning_rate": 0.0, + "loss": 1.1175, + "step": 14140 + }, + { + "epoch": 3.295295761527713, + "grad_norm": 22.48821258544922, + "learning_rate": 0.0, + "loss": 1.1427, + "step": 14150 + }, + { + "epoch": 3.297624592454588, + "grad_norm": 51.519859313964844, + "learning_rate": 0.0, + "loss": 1.8776, + "step": 14160 + }, + { + "epoch": 3.2999534233814627, + "grad_norm": 25.946897506713867, + "learning_rate": 0.0, + "loss": 1.7691, + "step": 14170 + }, + { + "epoch": 3.302282254308337, + "grad_norm": 6.704770565032959, + "learning_rate": 0.0, + "loss": 1.6418, + "step": 14180 + }, + { + "epoch": 3.3046110852352117, + "grad_norm": 30.36494255065918, + "learning_rate": 0.0, + "loss": 1.0997, + "step": 14190 + }, + { + "epoch": 3.3069399161620865, + "grad_norm": 16.188684463500977, + "learning_rate": 0.0, + "loss": 1.4068, + "step": 14200 + }, + { + "epoch": 3.3092687470889612, + "grad_norm": 11.660002708435059, + "learning_rate": 0.0, + "loss": 0.9104, + "step": 14210 + }, + { + "epoch": 3.311597578015836, + "grad_norm": 42.77584457397461, + "learning_rate": 0.0, + "loss": 1.49, + "step": 14220 + }, + { + "epoch": 3.3139264089427107, + "grad_norm": 15.70578670501709, + "learning_rate": 0.0, + "loss": 1.204, + "step": 14230 + }, + { + "epoch": 3.3162552398695855, + "grad_norm": 28.181676864624023, + "learning_rate": 0.0, + "loss": 1.132, + "step": 14240 + }, + { + "epoch": 3.3185840707964602, + "grad_norm": 6.674837112426758, + "learning_rate": 0.0, + "loss": 1.1825, + "step": 14250 + }, + { + "epoch": 3.320912901723335, + "grad_norm": 51.54294967651367, + "learning_rate": 0.0, + "loss": 1.0609, + "step": 14260 + }, + { + "epoch": 3.3232417326502097, + "grad_norm": 11.583158493041992, + "learning_rate": 0.0, + "loss": 1.0657, + "step": 14270 + }, + { + "epoch": 3.3255705635770845, + "grad_norm": 15.019560813903809, + "learning_rate": 0.0, + "loss": 1.038, + "step": 14280 + }, + { + "epoch": 3.3278993945039588, + "grad_norm": 33.25980758666992, + "learning_rate": 0.0, + "loss": 1.1687, + "step": 14290 + }, + { + "epoch": 3.330228225430834, + "grad_norm": 28.561473846435547, + "learning_rate": 0.0, + "loss": 1.1052, + "step": 14300 + }, + { + "epoch": 3.3325570563577083, + "grad_norm": 62.97575378417969, + "learning_rate": 0.0, + "loss": 1.1649, + "step": 14310 + }, + { + "epoch": 3.334885887284583, + "grad_norm": 16.132755279541016, + "learning_rate": 0.0, + "loss": 1.1446, + "step": 14320 + }, + { + "epoch": 3.3372147182114578, + "grad_norm": 28.80936622619629, + "learning_rate": 0.0, + "loss": 1.3675, + "step": 14330 + }, + { + "epoch": 3.3395435491383325, + "grad_norm": 27.189163208007812, + "learning_rate": 0.0, + "loss": 1.2651, + "step": 14340 + }, + { + "epoch": 3.3418723800652073, + "grad_norm": 54.67479705810547, + "learning_rate": 0.0, + "loss": 1.2248, + "step": 14350 + }, + { + "epoch": 3.344201210992082, + "grad_norm": 31.085508346557617, + "learning_rate": 0.0, + "loss": 1.6074, + "step": 14360 + }, + { + "epoch": 3.3465300419189568, + "grad_norm": 4.460205078125, + "learning_rate": 0.0, + "loss": 1.4311, + "step": 14370 + }, + { + "epoch": 3.3488588728458315, + "grad_norm": 9.599722862243652, + "learning_rate": 0.0, + "loss": 1.4001, + "step": 14380 + }, + { + "epoch": 3.3511877037727063, + "grad_norm": 13.639843940734863, + "learning_rate": 0.0, + "loss": 1.2749, + "step": 14390 + }, + { + "epoch": 3.353516534699581, + "grad_norm": 34.45295715332031, + "learning_rate": 0.0, + "loss": 1.1642, + "step": 14400 + }, + { + "epoch": 3.3558453656264557, + "grad_norm": 12.180970191955566, + "learning_rate": 0.0, + "loss": 1.1084, + "step": 14410 + }, + { + "epoch": 3.35817419655333, + "grad_norm": 74.1428451538086, + "learning_rate": 0.0, + "loss": 2.0657, + "step": 14420 + }, + { + "epoch": 3.360503027480205, + "grad_norm": 9.939842224121094, + "learning_rate": 0.0, + "loss": 1.2155, + "step": 14430 + }, + { + "epoch": 3.3628318584070795, + "grad_norm": 6.518401145935059, + "learning_rate": 0.0, + "loss": 0.7601, + "step": 14440 + }, + { + "epoch": 3.3651606893339543, + "grad_norm": 64.0587387084961, + "learning_rate": 0.0, + "loss": 1.5006, + "step": 14450 + }, + { + "epoch": 3.367489520260829, + "grad_norm": 8.050665855407715, + "learning_rate": 0.0, + "loss": 1.5511, + "step": 14460 + }, + { + "epoch": 3.369818351187704, + "grad_norm": 6.347721099853516, + "learning_rate": 0.0, + "loss": 1.5936, + "step": 14470 + }, + { + "epoch": 3.3721471821145785, + "grad_norm": 16.514511108398438, + "learning_rate": 0.0, + "loss": 0.9755, + "step": 14480 + }, + { + "epoch": 3.3744760130414533, + "grad_norm": 10.741750717163086, + "learning_rate": 0.0, + "loss": 1.2707, + "step": 14490 + }, + { + "epoch": 3.376804843968328, + "grad_norm": 5.380224227905273, + "learning_rate": 0.0, + "loss": 1.6041, + "step": 14500 + }, + { + "epoch": 3.3791336748952028, + "grad_norm": 55.9947395324707, + "learning_rate": 0.0, + "loss": 1.5839, + "step": 14510 + }, + { + "epoch": 3.3814625058220775, + "grad_norm": 15.040616035461426, + "learning_rate": 0.0, + "loss": 1.0565, + "step": 14520 + }, + { + "epoch": 3.383791336748952, + "grad_norm": 7.4967522621154785, + "learning_rate": 0.0, + "loss": 1.5983, + "step": 14530 + }, + { + "epoch": 3.3861201676758266, + "grad_norm": 16.126693725585938, + "learning_rate": 0.0, + "loss": 0.9066, + "step": 14540 + }, + { + "epoch": 3.3884489986027013, + "grad_norm": 25.731449127197266, + "learning_rate": 0.0, + "loss": 0.8373, + "step": 14550 + }, + { + "epoch": 3.390777829529576, + "grad_norm": 9.732860565185547, + "learning_rate": 0.0, + "loss": 1.3208, + "step": 14560 + }, + { + "epoch": 3.393106660456451, + "grad_norm": 93.20537567138672, + "learning_rate": 0.0, + "loss": 1.3592, + "step": 14570 + }, + { + "epoch": 3.3954354913833256, + "grad_norm": 10.425060272216797, + "learning_rate": 0.0, + "loss": 1.3393, + "step": 14580 + }, + { + "epoch": 3.3977643223102003, + "grad_norm": 9.359151840209961, + "learning_rate": 0.0, + "loss": 1.4633, + "step": 14590 + }, + { + "epoch": 3.400093153237075, + "grad_norm": 19.98016357421875, + "learning_rate": 0.0, + "loss": 1.8438, + "step": 14600 + }, + { + "epoch": 3.40242198416395, + "grad_norm": 15.402310371398926, + "learning_rate": 0.0, + "loss": 1.7674, + "step": 14610 + }, + { + "epoch": 3.4047508150908246, + "grad_norm": 17.607166290283203, + "learning_rate": 0.0, + "loss": 1.1485, + "step": 14620 + }, + { + "epoch": 3.4070796460176993, + "grad_norm": 9.94525146484375, + "learning_rate": 0.0, + "loss": 0.9103, + "step": 14630 + }, + { + "epoch": 3.4094084769445736, + "grad_norm": 17.84208106994629, + "learning_rate": 0.0, + "loss": 1.1366, + "step": 14640 + }, + { + "epoch": 3.4117373078714484, + "grad_norm": 15.35504150390625, + "learning_rate": 0.0, + "loss": 1.2129, + "step": 14650 + }, + { + "epoch": 3.414066138798323, + "grad_norm": 21.50774574279785, + "learning_rate": 0.0, + "loss": 1.0228, + "step": 14660 + }, + { + "epoch": 3.416394969725198, + "grad_norm": 45.42424774169922, + "learning_rate": 0.0, + "loss": 1.6175, + "step": 14670 + }, + { + "epoch": 3.4187238006520726, + "grad_norm": 34.5196533203125, + "learning_rate": 0.0, + "loss": 1.2004, + "step": 14680 + }, + { + "epoch": 3.4210526315789473, + "grad_norm": 34.76347351074219, + "learning_rate": 0.0, + "loss": 0.9398, + "step": 14690 + }, + { + "epoch": 3.423381462505822, + "grad_norm": 12.189762115478516, + "learning_rate": 0.0, + "loss": 1.0523, + "step": 14700 + }, + { + "epoch": 3.425710293432697, + "grad_norm": 8.52271842956543, + "learning_rate": 0.0, + "loss": 0.9799, + "step": 14710 + }, + { + "epoch": 3.4280391243595716, + "grad_norm": 62.98123550415039, + "learning_rate": 0.0, + "loss": 1.4572, + "step": 14720 + }, + { + "epoch": 3.4303679552864463, + "grad_norm": 5.033658504486084, + "learning_rate": 0.0, + "loss": 1.0885, + "step": 14730 + }, + { + "epoch": 3.432696786213321, + "grad_norm": 17.944128036499023, + "learning_rate": 0.0, + "loss": 1.467, + "step": 14740 + }, + { + "epoch": 3.4350256171401954, + "grad_norm": 7.418787956237793, + "learning_rate": 0.0, + "loss": 1.0984, + "step": 14750 + }, + { + "epoch": 3.4373544480670706, + "grad_norm": 7.411730766296387, + "learning_rate": 0.0, + "loss": 0.8685, + "step": 14760 + }, + { + "epoch": 3.439683278993945, + "grad_norm": 19.966922760009766, + "learning_rate": 0.0, + "loss": 1.4455, + "step": 14770 + }, + { + "epoch": 3.4420121099208196, + "grad_norm": 10.585490226745605, + "learning_rate": 0.0, + "loss": 1.1593, + "step": 14780 + }, + { + "epoch": 3.4443409408476944, + "grad_norm": 8.054330825805664, + "learning_rate": 0.0, + "loss": 1.3982, + "step": 14790 + }, + { + "epoch": 3.446669771774569, + "grad_norm": 8.496953964233398, + "learning_rate": 0.0, + "loss": 1.1782, + "step": 14800 + }, + { + "epoch": 3.448998602701444, + "grad_norm": 9.365567207336426, + "learning_rate": 0.0, + "loss": 0.9234, + "step": 14810 + }, + { + "epoch": 3.4513274336283186, + "grad_norm": 13.222329139709473, + "learning_rate": 0.0, + "loss": 1.2217, + "step": 14820 + }, + { + "epoch": 3.4536562645551934, + "grad_norm": 28.72707748413086, + "learning_rate": 0.0, + "loss": 1.0365, + "step": 14830 + }, + { + "epoch": 3.455985095482068, + "grad_norm": 27.434940338134766, + "learning_rate": 0.0, + "loss": 1.1692, + "step": 14840 + }, + { + "epoch": 3.458313926408943, + "grad_norm": 4.920462131500244, + "learning_rate": 0.0, + "loss": 1.0483, + "step": 14850 + }, + { + "epoch": 3.4606427573358176, + "grad_norm": 12.028072357177734, + "learning_rate": 0.0, + "loss": 1.547, + "step": 14860 + }, + { + "epoch": 3.4629715882626924, + "grad_norm": 14.676789283752441, + "learning_rate": 0.0, + "loss": 1.5541, + "step": 14870 + }, + { + "epoch": 3.4653004191895667, + "grad_norm": 31.549835205078125, + "learning_rate": 0.0, + "loss": 2.1055, + "step": 14880 + }, + { + "epoch": 3.4676292501164414, + "grad_norm": 14.0533447265625, + "learning_rate": 0.0, + "loss": 1.0097, + "step": 14890 + }, + { + "epoch": 3.469958081043316, + "grad_norm": 18.530563354492188, + "learning_rate": 0.0, + "loss": 1.1465, + "step": 14900 + }, + { + "epoch": 3.472286911970191, + "grad_norm": 10.129080772399902, + "learning_rate": 0.0, + "loss": 1.3289, + "step": 14910 + }, + { + "epoch": 3.4746157428970657, + "grad_norm": 12.979293823242188, + "learning_rate": 0.0, + "loss": 1.708, + "step": 14920 + }, + { + "epoch": 3.4769445738239404, + "grad_norm": 48.22283935546875, + "learning_rate": 0.0, + "loss": 0.9649, + "step": 14930 + }, + { + "epoch": 3.479273404750815, + "grad_norm": 15.611557960510254, + "learning_rate": 0.0, + "loss": 1.407, + "step": 14940 + }, + { + "epoch": 3.48160223567769, + "grad_norm": 18.218061447143555, + "learning_rate": 0.0, + "loss": 1.6172, + "step": 14950 + }, + { + "epoch": 3.4839310666045646, + "grad_norm": 4.562121391296387, + "learning_rate": 0.0, + "loss": 0.8124, + "step": 14960 + }, + { + "epoch": 3.4862598975314394, + "grad_norm": 52.5666389465332, + "learning_rate": 0.0, + "loss": 1.391, + "step": 14970 + }, + { + "epoch": 3.488588728458314, + "grad_norm": 15.410503387451172, + "learning_rate": 0.0, + "loss": 1.4648, + "step": 14980 + }, + { + "epoch": 3.4909175593851884, + "grad_norm": 51.82364273071289, + "learning_rate": 0.0, + "loss": 1.0539, + "step": 14990 + }, + { + "epoch": 3.493246390312063, + "grad_norm": 5.764760971069336, + "learning_rate": 0.0, + "loss": 0.9975, + "step": 15000 + }, + { + "epoch": 3.495575221238938, + "grad_norm": 33.286373138427734, + "learning_rate": 0.0, + "loss": 1.4706, + "step": 15010 + }, + { + "epoch": 3.4979040521658127, + "grad_norm": 13.835527420043945, + "learning_rate": 0.0, + "loss": 1.182, + "step": 15020 + }, + { + "epoch": 3.5002328830926874, + "grad_norm": 29.695833206176758, + "learning_rate": 0.0, + "loss": 1.5128, + "step": 15030 + }, + { + "epoch": 3.502561714019562, + "grad_norm": 8.429316520690918, + "learning_rate": 0.0, + "loss": 1.1939, + "step": 15040 + }, + { + "epoch": 3.504890544946437, + "grad_norm": 34.458953857421875, + "learning_rate": 0.0, + "loss": 0.8596, + "step": 15050 + }, + { + "epoch": 3.5072193758733117, + "grad_norm": 38.27690505981445, + "learning_rate": 0.0, + "loss": 1.0778, + "step": 15060 + }, + { + "epoch": 3.5095482068001864, + "grad_norm": 11.751981735229492, + "learning_rate": 0.0, + "loss": 1.1244, + "step": 15070 + }, + { + "epoch": 3.511877037727061, + "grad_norm": 61.4968147277832, + "learning_rate": 0.0, + "loss": 1.1948, + "step": 15080 + }, + { + "epoch": 3.514205868653936, + "grad_norm": 7.1937408447265625, + "learning_rate": 0.0, + "loss": 0.9009, + "step": 15090 + }, + { + "epoch": 3.51653469958081, + "grad_norm": 116.81895446777344, + "learning_rate": 0.0, + "loss": 1.6013, + "step": 15100 + }, + { + "epoch": 3.5188635305076854, + "grad_norm": 6.943803787231445, + "learning_rate": 0.0, + "loss": 0.8098, + "step": 15110 + }, + { + "epoch": 3.5211923614345597, + "grad_norm": 12.434028625488281, + "learning_rate": 0.0, + "loss": 1.3035, + "step": 15120 + }, + { + "epoch": 3.5235211923614345, + "grad_norm": 7.836872577667236, + "learning_rate": 0.0, + "loss": 1.0687, + "step": 15130 + }, + { + "epoch": 3.525850023288309, + "grad_norm": 8.494155883789062, + "learning_rate": 0.0, + "loss": 1.2936, + "step": 15140 + }, + { + "epoch": 3.528178854215184, + "grad_norm": 27.25113868713379, + "learning_rate": 0.0, + "loss": 1.3568, + "step": 15150 + }, + { + "epoch": 3.5305076851420587, + "grad_norm": 13.736648559570312, + "learning_rate": 0.0, + "loss": 1.3675, + "step": 15160 + }, + { + "epoch": 3.5328365160689335, + "grad_norm": 8.055909156799316, + "learning_rate": 0.0, + "loss": 0.7531, + "step": 15170 + }, + { + "epoch": 3.535165346995808, + "grad_norm": 12.262843132019043, + "learning_rate": 0.0, + "loss": 1.0719, + "step": 15180 + }, + { + "epoch": 3.537494177922683, + "grad_norm": 12.301920890808105, + "learning_rate": 0.0, + "loss": 1.2898, + "step": 15190 + }, + { + "epoch": 3.5398230088495577, + "grad_norm": 22.987722396850586, + "learning_rate": 0.0, + "loss": 1.5796, + "step": 15200 + }, + { + "epoch": 3.542151839776432, + "grad_norm": 10.889913558959961, + "learning_rate": 0.0, + "loss": 0.9, + "step": 15210 + }, + { + "epoch": 3.544480670703307, + "grad_norm": 46.55030822753906, + "learning_rate": 0.0, + "loss": 1.046, + "step": 15220 + }, + { + "epoch": 3.5468095016301815, + "grad_norm": 49.14213180541992, + "learning_rate": 0.0, + "loss": 1.4466, + "step": 15230 + }, + { + "epoch": 3.5491383325570562, + "grad_norm": 9.739147186279297, + "learning_rate": 0.0, + "loss": 0.7041, + "step": 15240 + }, + { + "epoch": 3.551467163483931, + "grad_norm": 9.940231323242188, + "learning_rate": 0.0, + "loss": 1.7196, + "step": 15250 + }, + { + "epoch": 3.5537959944108057, + "grad_norm": 27.860456466674805, + "learning_rate": 0.0, + "loss": 0.9869, + "step": 15260 + }, + { + "epoch": 3.5561248253376805, + "grad_norm": 12.440192222595215, + "learning_rate": 0.0, + "loss": 0.8983, + "step": 15270 + }, + { + "epoch": 3.5584536562645552, + "grad_norm": 7.742676258087158, + "learning_rate": 0.0, + "loss": 1.1601, + "step": 15280 + }, + { + "epoch": 3.56078248719143, + "grad_norm": 10.191648483276367, + "learning_rate": 0.0, + "loss": 1.2396, + "step": 15290 + }, + { + "epoch": 3.5631113181183047, + "grad_norm": 11.127610206604004, + "learning_rate": 0.0, + "loss": 0.8596, + "step": 15300 + }, + { + "epoch": 3.5654401490451795, + "grad_norm": 22.639829635620117, + "learning_rate": 0.0, + "loss": 1.0627, + "step": 15310 + }, + { + "epoch": 3.5677689799720538, + "grad_norm": 12.301551818847656, + "learning_rate": 0.0, + "loss": 1.3009, + "step": 15320 + }, + { + "epoch": 3.570097810898929, + "grad_norm": 32.67119598388672, + "learning_rate": 0.0, + "loss": 1.1616, + "step": 15330 + }, + { + "epoch": 3.5724266418258033, + "grad_norm": 25.659578323364258, + "learning_rate": 0.0, + "loss": 1.0511, + "step": 15340 + }, + { + "epoch": 3.574755472752678, + "grad_norm": 34.240386962890625, + "learning_rate": 0.0, + "loss": 1.2238, + "step": 15350 + }, + { + "epoch": 3.5770843036795528, + "grad_norm": 8.90058422088623, + "learning_rate": 0.0, + "loss": 1.2145, + "step": 15360 + }, + { + "epoch": 3.5794131346064275, + "grad_norm": 8.673346519470215, + "learning_rate": 0.0, + "loss": 0.8024, + "step": 15370 + }, + { + "epoch": 3.5817419655333023, + "grad_norm": 4.38419246673584, + "learning_rate": 0.0, + "loss": 1.4394, + "step": 15380 + }, + { + "epoch": 3.584070796460177, + "grad_norm": 13.117408752441406, + "learning_rate": 0.0, + "loss": 1.3587, + "step": 15390 + }, + { + "epoch": 3.5863996273870518, + "grad_norm": 8.190533638000488, + "learning_rate": 0.0, + "loss": 1.2683, + "step": 15400 + }, + { + "epoch": 3.5887284583139265, + "grad_norm": 16.9916934967041, + "learning_rate": 0.0, + "loss": 1.5906, + "step": 15410 + }, + { + "epoch": 3.5910572892408013, + "grad_norm": 3.9092514514923096, + "learning_rate": 0.0, + "loss": 0.9965, + "step": 15420 + }, + { + "epoch": 3.5933861201676756, + "grad_norm": 15.621074676513672, + "learning_rate": 0.0, + "loss": 0.954, + "step": 15430 + }, + { + "epoch": 3.5957149510945507, + "grad_norm": 6.421728134155273, + "learning_rate": 0.0, + "loss": 1.135, + "step": 15440 + }, + { + "epoch": 3.598043782021425, + "grad_norm": 21.08034324645996, + "learning_rate": 0.0, + "loss": 0.9873, + "step": 15450 + }, + { + "epoch": 3.6003726129483, + "grad_norm": 27.012897491455078, + "learning_rate": 0.0, + "loss": 1.203, + "step": 15460 + }, + { + "epoch": 3.6027014438751745, + "grad_norm": 53.65129852294922, + "learning_rate": 0.0, + "loss": 1.1453, + "step": 15470 + }, + { + "epoch": 3.6050302748020493, + "grad_norm": 9.872217178344727, + "learning_rate": 0.0, + "loss": 1.0103, + "step": 15480 + }, + { + "epoch": 3.607359105728924, + "grad_norm": 6.3376569747924805, + "learning_rate": 0.0, + "loss": 0.7322, + "step": 15490 + }, + { + "epoch": 3.609687936655799, + "grad_norm": 34.383384704589844, + "learning_rate": 0.0, + "loss": 2.007, + "step": 15500 + }, + { + "epoch": 3.6120167675826735, + "grad_norm": 19.966154098510742, + "learning_rate": 0.0, + "loss": 1.0879, + "step": 15510 + }, + { + "epoch": 3.6143455985095483, + "grad_norm": 12.672451972961426, + "learning_rate": 0.0, + "loss": 1.0139, + "step": 15520 + }, + { + "epoch": 3.616674429436423, + "grad_norm": 8.703089714050293, + "learning_rate": 0.0, + "loss": 0.9467, + "step": 15530 + }, + { + "epoch": 3.619003260363298, + "grad_norm": 12.565585136413574, + "learning_rate": 0.0, + "loss": 1.2754, + "step": 15540 + }, + { + "epoch": 3.6213320912901725, + "grad_norm": 7.127060890197754, + "learning_rate": 0.0, + "loss": 1.1098, + "step": 15550 + }, + { + "epoch": 3.623660922217047, + "grad_norm": 6.133944034576416, + "learning_rate": 0.0, + "loss": 1.2271, + "step": 15560 + }, + { + "epoch": 3.625989753143922, + "grad_norm": 10.252046585083008, + "learning_rate": 0.0, + "loss": 1.0133, + "step": 15570 + }, + { + "epoch": 3.6283185840707963, + "grad_norm": 41.735809326171875, + "learning_rate": 0.0, + "loss": 1.5424, + "step": 15580 + }, + { + "epoch": 3.630647414997671, + "grad_norm": 12.260039329528809, + "learning_rate": 0.0, + "loss": 1.0404, + "step": 15590 + }, + { + "epoch": 3.632976245924546, + "grad_norm": 7.584269046783447, + "learning_rate": 0.0, + "loss": 1.5002, + "step": 15600 + }, + { + "epoch": 3.6353050768514206, + "grad_norm": 10.819036483764648, + "learning_rate": 0.0, + "loss": 1.1327, + "step": 15610 + }, + { + "epoch": 3.6376339077782953, + "grad_norm": 48.199974060058594, + "learning_rate": 0.0, + "loss": 1.1238, + "step": 15620 + }, + { + "epoch": 3.63996273870517, + "grad_norm": 35.932064056396484, + "learning_rate": 0.0, + "loss": 1.7948, + "step": 15630 + }, + { + "epoch": 3.642291569632045, + "grad_norm": 6.570189476013184, + "learning_rate": 0.0, + "loss": 1.2111, + "step": 15640 + }, + { + "epoch": 3.6446204005589196, + "grad_norm": 5.2273101806640625, + "learning_rate": 0.0, + "loss": 0.9788, + "step": 15650 + }, + { + "epoch": 3.6469492314857943, + "grad_norm": 15.28518009185791, + "learning_rate": 0.0, + "loss": 1.1353, + "step": 15660 + }, + { + "epoch": 3.6492780624126686, + "grad_norm": 11.389464378356934, + "learning_rate": 0.0, + "loss": 0.9977, + "step": 15670 + }, + { + "epoch": 3.651606893339544, + "grad_norm": 27.311847686767578, + "learning_rate": 0.0, + "loss": 1.3812, + "step": 15680 + }, + { + "epoch": 3.653935724266418, + "grad_norm": 94.494384765625, + "learning_rate": 0.0, + "loss": 1.4974, + "step": 15690 + }, + { + "epoch": 3.656264555193293, + "grad_norm": 42.005332946777344, + "learning_rate": 0.0, + "loss": 1.1392, + "step": 15700 + }, + { + "epoch": 3.6585933861201676, + "grad_norm": 40.726131439208984, + "learning_rate": 0.0, + "loss": 1.5276, + "step": 15710 + }, + { + "epoch": 3.6609222170470423, + "grad_norm": 11.148183822631836, + "learning_rate": 0.0, + "loss": 0.9784, + "step": 15720 + }, + { + "epoch": 3.663251047973917, + "grad_norm": 76.10807800292969, + "learning_rate": 0.0, + "loss": 1.2827, + "step": 15730 + }, + { + "epoch": 3.665579878900792, + "grad_norm": 198.63717651367188, + "learning_rate": 0.0, + "loss": 1.4597, + "step": 15740 + }, + { + "epoch": 3.6679087098276666, + "grad_norm": 65.95364379882812, + "learning_rate": 0.0, + "loss": 1.2488, + "step": 15750 + }, + { + "epoch": 3.6702375407545413, + "grad_norm": 31.9210262298584, + "learning_rate": 0.0, + "loss": 1.1257, + "step": 15760 + }, + { + "epoch": 3.672566371681416, + "grad_norm": 14.054140090942383, + "learning_rate": 0.0, + "loss": 0.8987, + "step": 15770 + }, + { + "epoch": 3.6748952026082904, + "grad_norm": 13.724059104919434, + "learning_rate": 0.0, + "loss": 1.4745, + "step": 15780 + }, + { + "epoch": 3.6772240335351656, + "grad_norm": 38.00988006591797, + "learning_rate": 0.0, + "loss": 1.3208, + "step": 15790 + }, + { + "epoch": 3.67955286446204, + "grad_norm": 20.750085830688477, + "learning_rate": 0.0, + "loss": 0.9753, + "step": 15800 + }, + { + "epoch": 3.6818816953889146, + "grad_norm": 6.154573440551758, + "learning_rate": 0.0, + "loss": 1.1994, + "step": 15810 + }, + { + "epoch": 3.6842105263157894, + "grad_norm": 8.296215057373047, + "learning_rate": 0.0, + "loss": 1.6545, + "step": 15820 + }, + { + "epoch": 3.686539357242664, + "grad_norm": 8.964814186096191, + "learning_rate": 0.0, + "loss": 0.8555, + "step": 15830 + }, + { + "epoch": 3.688868188169539, + "grad_norm": 72.2197494506836, + "learning_rate": 0.0, + "loss": 1.3968, + "step": 15840 + }, + { + "epoch": 3.6911970190964136, + "grad_norm": 8.642261505126953, + "learning_rate": 0.0, + "loss": 1.1964, + "step": 15850 + }, + { + "epoch": 3.6935258500232884, + "grad_norm": 9.64587688446045, + "learning_rate": 0.0, + "loss": 1.5, + "step": 15860 + }, + { + "epoch": 3.695854680950163, + "grad_norm": 9.663128852844238, + "learning_rate": 0.0, + "loss": 0.9647, + "step": 15870 + }, + { + "epoch": 3.698183511877038, + "grad_norm": 9.119319915771484, + "learning_rate": 0.0, + "loss": 0.9505, + "step": 15880 + }, + { + "epoch": 3.700512342803912, + "grad_norm": 13.707714080810547, + "learning_rate": 0.0, + "loss": 1.2409, + "step": 15890 + }, + { + "epoch": 3.7028411737307874, + "grad_norm": 9.985272407531738, + "learning_rate": 0.0, + "loss": 1.2205, + "step": 15900 + }, + { + "epoch": 3.7051700046576617, + "grad_norm": 12.141681671142578, + "learning_rate": 0.0, + "loss": 1.6891, + "step": 15910 + }, + { + "epoch": 3.7074988355845364, + "grad_norm": 12.171897888183594, + "learning_rate": 0.0, + "loss": 0.9045, + "step": 15920 + }, + { + "epoch": 3.709827666511411, + "grad_norm": 17.349260330200195, + "learning_rate": 0.0, + "loss": 1.9987, + "step": 15930 + }, + { + "epoch": 3.712156497438286, + "grad_norm": 14.914823532104492, + "learning_rate": 0.0, + "loss": 1.425, + "step": 15940 + }, + { + "epoch": 3.7144853283651607, + "grad_norm": 8.951475143432617, + "learning_rate": 0.0, + "loss": 1.1546, + "step": 15950 + }, + { + "epoch": 3.7168141592920354, + "grad_norm": 40.80722427368164, + "learning_rate": 0.0, + "loss": 1.1636, + "step": 15960 + }, + { + "epoch": 3.71914299021891, + "grad_norm": 45.113365173339844, + "learning_rate": 0.0, + "loss": 1.535, + "step": 15970 + }, + { + "epoch": 3.721471821145785, + "grad_norm": 16.32363510131836, + "learning_rate": 0.0, + "loss": 1.1573, + "step": 15980 + }, + { + "epoch": 3.7238006520726596, + "grad_norm": 14.807791709899902, + "learning_rate": 0.0, + "loss": 0.8877, + "step": 15990 + }, + { + "epoch": 3.7261294829995344, + "grad_norm": 29.387487411499023, + "learning_rate": 0.0, + "loss": 1.6075, + "step": 16000 + }, + { + "epoch": 3.728458313926409, + "grad_norm": 17.61796760559082, + "learning_rate": 0.0, + "loss": 1.144, + "step": 16010 + }, + { + "epoch": 3.7307871448532834, + "grad_norm": 9.548173904418945, + "learning_rate": 0.0, + "loss": 0.9762, + "step": 16020 + }, + { + "epoch": 3.7331159757801586, + "grad_norm": 31.526107788085938, + "learning_rate": 0.0, + "loss": 1.792, + "step": 16030 + }, + { + "epoch": 3.735444806707033, + "grad_norm": 5.914717674255371, + "learning_rate": 0.0, + "loss": 0.8789, + "step": 16040 + }, + { + "epoch": 3.7377736376339077, + "grad_norm": 16.024036407470703, + "learning_rate": 0.0, + "loss": 1.1737, + "step": 16050 + }, + { + "epoch": 3.7401024685607824, + "grad_norm": 7.287302494049072, + "learning_rate": 0.0, + "loss": 1.1577, + "step": 16060 + }, + { + "epoch": 3.742431299487657, + "grad_norm": 5.333652496337891, + "learning_rate": 0.0, + "loss": 1.2293, + "step": 16070 + }, + { + "epoch": 3.744760130414532, + "grad_norm": 18.019535064697266, + "learning_rate": 0.0, + "loss": 1.3213, + "step": 16080 + }, + { + "epoch": 3.7470889613414067, + "grad_norm": 9.020870208740234, + "learning_rate": 0.0, + "loss": 1.2686, + "step": 16090 + }, + { + "epoch": 3.7494177922682814, + "grad_norm": 5.7825822830200195, + "learning_rate": 0.0, + "loss": 1.3063, + "step": 16100 + }, + { + "epoch": 3.751746623195156, + "grad_norm": 64.28756713867188, + "learning_rate": 0.0, + "loss": 1.4442, + "step": 16110 + }, + { + "epoch": 3.754075454122031, + "grad_norm": 5.167896270751953, + "learning_rate": 0.0, + "loss": 0.839, + "step": 16120 + }, + { + "epoch": 3.7564042850489052, + "grad_norm": 10.846601486206055, + "learning_rate": 0.0, + "loss": 1.2187, + "step": 16130 + }, + { + "epoch": 3.7587331159757804, + "grad_norm": 74.83040618896484, + "learning_rate": 0.0, + "loss": 1.6439, + "step": 16140 + }, + { + "epoch": 3.7610619469026547, + "grad_norm": 8.98836612701416, + "learning_rate": 0.0, + "loss": 1.1921, + "step": 16150 + }, + { + "epoch": 3.7633907778295295, + "grad_norm": 8.449505805969238, + "learning_rate": 0.0, + "loss": 1.2248, + "step": 16160 + }, + { + "epoch": 3.765719608756404, + "grad_norm": 21.298240661621094, + "learning_rate": 0.0, + "loss": 1.1026, + "step": 16170 + }, + { + "epoch": 3.768048439683279, + "grad_norm": 7.777653217315674, + "learning_rate": 0.0, + "loss": 1.1569, + "step": 16180 + }, + { + "epoch": 3.7703772706101537, + "grad_norm": 53.60115051269531, + "learning_rate": 0.0, + "loss": 1.6678, + "step": 16190 + }, + { + "epoch": 3.7727061015370285, + "grad_norm": 14.203383445739746, + "learning_rate": 0.0, + "loss": 1.468, + "step": 16200 + }, + { + "epoch": 3.775034932463903, + "grad_norm": 24.37751007080078, + "learning_rate": 0.0, + "loss": 1.997, + "step": 16210 + }, + { + "epoch": 3.777363763390778, + "grad_norm": 17.089813232421875, + "learning_rate": 0.0, + "loss": 0.9753, + "step": 16220 + }, + { + "epoch": 3.7796925943176527, + "grad_norm": 40.838375091552734, + "learning_rate": 0.0, + "loss": 1.3751, + "step": 16230 + }, + { + "epoch": 3.782021425244527, + "grad_norm": 3.4650375843048096, + "learning_rate": 0.0, + "loss": 1.162, + "step": 16240 + }, + { + "epoch": 3.784350256171402, + "grad_norm": 20.963716506958008, + "learning_rate": 0.0, + "loss": 1.3415, + "step": 16250 + }, + { + "epoch": 3.7866790870982765, + "grad_norm": 12.60222053527832, + "learning_rate": 0.0, + "loss": 1.5009, + "step": 16260 + }, + { + "epoch": 3.7890079180251512, + "grad_norm": 16.32765007019043, + "learning_rate": 0.0, + "loss": 1.3629, + "step": 16270 + }, + { + "epoch": 3.791336748952026, + "grad_norm": 9.827703475952148, + "learning_rate": 0.0, + "loss": 0.8297, + "step": 16280 + }, + { + "epoch": 3.7936655798789007, + "grad_norm": 11.158079147338867, + "learning_rate": 0.0, + "loss": 1.051, + "step": 16290 + }, + { + "epoch": 3.7959944108057755, + "grad_norm": 7.290315628051758, + "learning_rate": 0.0, + "loss": 1.3413, + "step": 16300 + }, + { + "epoch": 3.7983232417326502, + "grad_norm": 5.4215521812438965, + "learning_rate": 0.0, + "loss": 0.7628, + "step": 16310 + }, + { + "epoch": 3.800652072659525, + "grad_norm": 8.901557922363281, + "learning_rate": 0.0, + "loss": 1.1244, + "step": 16320 + }, + { + "epoch": 3.8029809035863997, + "grad_norm": 10.666454315185547, + "learning_rate": 0.0, + "loss": 1.1982, + "step": 16330 + }, + { + "epoch": 3.8053097345132745, + "grad_norm": 32.98258590698242, + "learning_rate": 0.0, + "loss": 1.5894, + "step": 16340 + }, + { + "epoch": 3.807638565440149, + "grad_norm": 11.325212478637695, + "learning_rate": 0.0, + "loss": 1.7647, + "step": 16350 + }, + { + "epoch": 3.809967396367024, + "grad_norm": 34.25052261352539, + "learning_rate": 0.0, + "loss": 1.2847, + "step": 16360 + }, + { + "epoch": 3.8122962272938983, + "grad_norm": 12.110124588012695, + "learning_rate": 0.0, + "loss": 1.3453, + "step": 16370 + }, + { + "epoch": 3.814625058220773, + "grad_norm": 10.258769989013672, + "learning_rate": 0.0, + "loss": 0.7541, + "step": 16380 + }, + { + "epoch": 3.8169538891476478, + "grad_norm": 7.55911922454834, + "learning_rate": 0.0, + "loss": 1.1706, + "step": 16390 + }, + { + "epoch": 3.8192827200745225, + "grad_norm": 6.271566390991211, + "learning_rate": 0.0, + "loss": 0.9377, + "step": 16400 + }, + { + "epoch": 3.8216115510013973, + "grad_norm": 28.639516830444336, + "learning_rate": 0.0, + "loss": 1.2274, + "step": 16410 + }, + { + "epoch": 3.823940381928272, + "grad_norm": 5.265059471130371, + "learning_rate": 0.0, + "loss": 0.8603, + "step": 16420 + }, + { + "epoch": 3.8262692128551468, + "grad_norm": 8.896881103515625, + "learning_rate": 0.0, + "loss": 1.3604, + "step": 16430 + }, + { + "epoch": 3.8285980437820215, + "grad_norm": 8.662596702575684, + "learning_rate": 0.0, + "loss": 1.1663, + "step": 16440 + }, + { + "epoch": 3.8309268747088963, + "grad_norm": 7.28228759765625, + "learning_rate": 0.0, + "loss": 0.9696, + "step": 16450 + }, + { + "epoch": 3.833255705635771, + "grad_norm": 51.78569030761719, + "learning_rate": 0.0, + "loss": 1.6647, + "step": 16460 + }, + { + "epoch": 3.8355845365626458, + "grad_norm": 10.870829582214355, + "learning_rate": 0.0, + "loss": 1.3452, + "step": 16470 + }, + { + "epoch": 3.83791336748952, + "grad_norm": 74.10052490234375, + "learning_rate": 0.0, + "loss": 1.527, + "step": 16480 + }, + { + "epoch": 3.840242198416395, + "grad_norm": 15.562661170959473, + "learning_rate": 0.0, + "loss": 1.4266, + "step": 16490 + }, + { + "epoch": 3.8425710293432696, + "grad_norm": 8.279319763183594, + "learning_rate": 0.0, + "loss": 1.4034, + "step": 16500 + }, + { + "epoch": 3.8448998602701443, + "grad_norm": 55.923126220703125, + "learning_rate": 0.0, + "loss": 1.35, + "step": 16510 + }, + { + "epoch": 3.847228691197019, + "grad_norm": 15.255816459655762, + "learning_rate": 0.0, + "loss": 1.5191, + "step": 16520 + }, + { + "epoch": 3.849557522123894, + "grad_norm": 15.816597938537598, + "learning_rate": 0.0, + "loss": 1.0833, + "step": 16530 + }, + { + "epoch": 3.8518863530507685, + "grad_norm": 27.234275817871094, + "learning_rate": 0.0, + "loss": 0.9751, + "step": 16540 + }, + { + "epoch": 3.8542151839776433, + "grad_norm": 9.592534065246582, + "learning_rate": 0.0, + "loss": 1.2378, + "step": 16550 + }, + { + "epoch": 3.856544014904518, + "grad_norm": 5.3678388595581055, + "learning_rate": 0.0, + "loss": 1.6023, + "step": 16560 + }, + { + "epoch": 3.858872845831393, + "grad_norm": 16.641244888305664, + "learning_rate": 0.0, + "loss": 1.348, + "step": 16570 + }, + { + "epoch": 3.8612016767582675, + "grad_norm": 8.91340160369873, + "learning_rate": 0.0, + "loss": 1.0128, + "step": 16580 + }, + { + "epoch": 3.863530507685142, + "grad_norm": 27.965681076049805, + "learning_rate": 0.0, + "loss": 1.3651, + "step": 16590 + }, + { + "epoch": 3.865859338612017, + "grad_norm": 24.31036376953125, + "learning_rate": 0.0, + "loss": 1.1294, + "step": 16600 + }, + { + "epoch": 3.8681881695388913, + "grad_norm": 6.568557262420654, + "learning_rate": 0.0, + "loss": 1.6243, + "step": 16610 + }, + { + "epoch": 3.870517000465766, + "grad_norm": 8.548968315124512, + "learning_rate": 0.0, + "loss": 1.0038, + "step": 16620 + }, + { + "epoch": 3.872845831392641, + "grad_norm": 7.129677772521973, + "learning_rate": 0.0, + "loss": 1.0813, + "step": 16630 + }, + { + "epoch": 3.8751746623195156, + "grad_norm": 33.37204360961914, + "learning_rate": 0.0, + "loss": 0.9409, + "step": 16640 + }, + { + "epoch": 3.8775034932463903, + "grad_norm": 43.65248107910156, + "learning_rate": 0.0, + "loss": 1.1626, + "step": 16650 + }, + { + "epoch": 3.879832324173265, + "grad_norm": Infinity, + "learning_rate": 0.0, + "loss": 1.9948, + "step": 16660 + }, + { + "epoch": 3.88216115510014, + "grad_norm": 8.591267585754395, + "learning_rate": 0.0, + "loss": 1.1729, + "step": 16670 + }, + { + "epoch": 3.8844899860270146, + "grad_norm": 17.50629425048828, + "learning_rate": 0.0, + "loss": 1.2443, + "step": 16680 + }, + { + "epoch": 3.8868188169538893, + "grad_norm": 12.087096214294434, + "learning_rate": 0.0, + "loss": 1.3456, + "step": 16690 + }, + { + "epoch": 3.8891476478807636, + "grad_norm": 21.00797462463379, + "learning_rate": 0.0, + "loss": 1.3057, + "step": 16700 + }, + { + "epoch": 3.891476478807639, + "grad_norm": 9.580062866210938, + "learning_rate": 0.0, + "loss": 0.927, + "step": 16710 + }, + { + "epoch": 3.893805309734513, + "grad_norm": 12.416278839111328, + "learning_rate": 0.0, + "loss": 1.0927, + "step": 16720 + }, + { + "epoch": 3.896134140661388, + "grad_norm": 6.420504570007324, + "learning_rate": 0.0, + "loss": 0.8602, + "step": 16730 + }, + { + "epoch": 3.8984629715882626, + "grad_norm": 11.220664978027344, + "learning_rate": 0.0, + "loss": 1.1429, + "step": 16740 + }, + { + "epoch": 3.9007918025151374, + "grad_norm": 5.770464897155762, + "learning_rate": 0.0, + "loss": 0.6577, + "step": 16750 + }, + { + "epoch": 3.903120633442012, + "grad_norm": 17.317913055419922, + "learning_rate": 0.0, + "loss": 0.9757, + "step": 16760 + }, + { + "epoch": 3.905449464368887, + "grad_norm": 252.115234375, + "learning_rate": 0.0, + "loss": 0.9186, + "step": 16770 + }, + { + "epoch": 3.9077782952957616, + "grad_norm": 17.903982162475586, + "learning_rate": 0.0, + "loss": 1.0096, + "step": 16780 + }, + { + "epoch": 3.9101071262226363, + "grad_norm": 9.154571533203125, + "learning_rate": 0.0, + "loss": 1.1535, + "step": 16790 + }, + { + "epoch": 3.912435957149511, + "grad_norm": 5.846363544464111, + "learning_rate": 0.0, + "loss": 1.0427, + "step": 16800 + }, + { + "epoch": 3.9147647880763854, + "grad_norm": 72.26444244384766, + "learning_rate": 0.0, + "loss": 1.3003, + "step": 16810 + }, + { + "epoch": 3.9170936190032606, + "grad_norm": 11.39892578125, + "learning_rate": 0.0, + "loss": 1.1781, + "step": 16820 + }, + { + "epoch": 3.919422449930135, + "grad_norm": 4.515076160430908, + "learning_rate": 0.0, + "loss": 1.3092, + "step": 16830 + }, + { + "epoch": 3.9217512808570096, + "grad_norm": 12.91057300567627, + "learning_rate": 0.0, + "loss": 1.1982, + "step": 16840 + }, + { + "epoch": 3.9240801117838844, + "grad_norm": 11.3361234664917, + "learning_rate": 0.0, + "loss": 1.3631, + "step": 16850 + }, + { + "epoch": 3.926408942710759, + "grad_norm": 9.171483993530273, + "learning_rate": 0.0, + "loss": 0.8862, + "step": 16860 + }, + { + "epoch": 3.928737773637634, + "grad_norm": 14.986222267150879, + "learning_rate": 0.0, + "loss": 1.1989, + "step": 16870 + }, + { + "epoch": 3.9310666045645086, + "grad_norm": 35.58830642700195, + "learning_rate": 0.0, + "loss": 1.0229, + "step": 16880 + }, + { + "epoch": 3.9333954354913834, + "grad_norm": 6.8410234451293945, + "learning_rate": 0.0, + "loss": 1.095, + "step": 16890 + }, + { + "epoch": 3.935724266418258, + "grad_norm": 22.072486877441406, + "learning_rate": 0.0, + "loss": 1.9416, + "step": 16900 + }, + { + "epoch": 3.938053097345133, + "grad_norm": 14.240464210510254, + "learning_rate": 0.0, + "loss": 0.9298, + "step": 16910 + }, + { + "epoch": 3.9403819282720076, + "grad_norm": 8.36672306060791, + "learning_rate": 0.0, + "loss": 1.1451, + "step": 16920 + }, + { + "epoch": 3.9427107591988824, + "grad_norm": 7.652619361877441, + "learning_rate": 0.0, + "loss": 0.8321, + "step": 16930 + }, + { + "epoch": 3.9450395901257567, + "grad_norm": 12.397917747497559, + "learning_rate": 0.0, + "loss": 0.9605, + "step": 16940 + }, + { + "epoch": 3.9473684210526314, + "grad_norm": 5.459069728851318, + "learning_rate": 0.0, + "loss": 1.0881, + "step": 16950 + }, + { + "epoch": 3.949697251979506, + "grad_norm": 29.952774047851562, + "learning_rate": 0.0, + "loss": 1.303, + "step": 16960 + }, + { + "epoch": 3.952026082906381, + "grad_norm": 4.160209655761719, + "learning_rate": 0.0, + "loss": 1.0394, + "step": 16970 + }, + { + "epoch": 3.9543549138332557, + "grad_norm": 8.1553316116333, + "learning_rate": 0.0, + "loss": 0.9403, + "step": 16980 + }, + { + "epoch": 3.9566837447601304, + "grad_norm": 13.145912170410156, + "learning_rate": 0.0, + "loss": 1.3667, + "step": 16990 + }, + { + "epoch": 3.959012575687005, + "grad_norm": 15.345560073852539, + "learning_rate": 0.0, + "loss": 1.0222, + "step": 17000 + }, + { + "epoch": 3.96134140661388, + "grad_norm": 19.302764892578125, + "learning_rate": 0.0, + "loss": 1.044, + "step": 17010 + }, + { + "epoch": 3.9636702375407546, + "grad_norm": 7.825178146362305, + "learning_rate": 0.0, + "loss": 0.8727, + "step": 17020 + }, + { + "epoch": 3.9659990684676294, + "grad_norm": 13.018383026123047, + "learning_rate": 0.0, + "loss": 1.053, + "step": 17030 + }, + { + "epoch": 3.968327899394504, + "grad_norm": 29.119361877441406, + "learning_rate": 0.0, + "loss": 1.3638, + "step": 17040 + }, + { + "epoch": 3.9706567303213784, + "grad_norm": 8.266595840454102, + "learning_rate": 0.0, + "loss": 1.2011, + "step": 17050 + }, + { + "epoch": 3.9729855612482536, + "grad_norm": 8.626734733581543, + "learning_rate": 0.0, + "loss": 1.249, + "step": 17060 + }, + { + "epoch": 3.975314392175128, + "grad_norm": 67.26522064208984, + "learning_rate": 0.0, + "loss": 1.9316, + "step": 17070 + }, + { + "epoch": 3.9776432231020027, + "grad_norm": 8.982076644897461, + "learning_rate": 0.0, + "loss": 1.4822, + "step": 17080 + }, + { + "epoch": 3.9799720540288774, + "grad_norm": 7.744446754455566, + "learning_rate": 0.0, + "loss": 1.1842, + "step": 17090 + }, + { + "epoch": 3.982300884955752, + "grad_norm": 14.065244674682617, + "learning_rate": 0.0, + "loss": 1.3703, + "step": 17100 + }, + { + "epoch": 3.984629715882627, + "grad_norm": 56.45340347290039, + "learning_rate": 0.0, + "loss": 0.9846, + "step": 17110 + }, + { + "epoch": 3.9869585468095017, + "grad_norm": 8.13912296295166, + "learning_rate": 0.0, + "loss": 0.9176, + "step": 17120 + }, + { + "epoch": 3.9892873777363764, + "grad_norm": 12.026097297668457, + "learning_rate": 0.0, + "loss": 1.2079, + "step": 17130 + }, + { + "epoch": 3.991616208663251, + "grad_norm": 11.71669864654541, + "learning_rate": 0.0, + "loss": 1.7823, + "step": 17140 + }, + { + "epoch": 3.993945039590126, + "grad_norm": 10.856141090393066, + "learning_rate": 0.0, + "loss": 1.5076, + "step": 17150 + }, + { + "epoch": 3.9962738705170002, + "grad_norm": 111.23175048828125, + "learning_rate": 0.0, + "loss": 1.7607, + "step": 17160 + }, + { + "epoch": 3.9986027014438754, + "grad_norm": 21.408830642700195, + "learning_rate": 0.0, + "loss": 1.233, + "step": 17170 + }, + { + "epoch": 4.0, + "eval_loss": 0.49088114500045776, + "eval_runtime": 15.277, + "eval_samples_per_second": 285.134, + "eval_steps_per_second": 2.291, + "step": 17176 + }, + { + "epoch": 4.00093153237075, + "grad_norm": 42.99901580810547, + "learning_rate": 0.0, + "loss": 1.3849, + "step": 17180 + }, + { + "epoch": 4.003260363297625, + "grad_norm": 15.593567848205566, + "learning_rate": 0.0, + "loss": 0.9876, + "step": 17190 + }, + { + "epoch": 4.005589194224499, + "grad_norm": 9.418315887451172, + "learning_rate": 0.0, + "loss": 0.775, + "step": 17200 + }, + { + "epoch": 4.007918025151374, + "grad_norm": 14.26424503326416, + "learning_rate": 0.0, + "loss": 0.993, + "step": 17210 + }, + { + "epoch": 4.010246856078249, + "grad_norm": 28.059171676635742, + "learning_rate": 0.0, + "loss": 1.4448, + "step": 17220 + }, + { + "epoch": 4.012575687005123, + "grad_norm": 6.072986125946045, + "learning_rate": 0.0, + "loss": 1.2469, + "step": 17230 + }, + { + "epoch": 4.014904517931998, + "grad_norm": 10.683692932128906, + "learning_rate": 0.0, + "loss": 1.218, + "step": 17240 + }, + { + "epoch": 4.0172333488588725, + "grad_norm": 76.68214416503906, + "learning_rate": 0.0, + "loss": 1.4419, + "step": 17250 + }, + { + "epoch": 4.019562179785748, + "grad_norm": 17.647035598754883, + "learning_rate": 0.0, + "loss": 1.8136, + "step": 17260 + }, + { + "epoch": 4.021891010712622, + "grad_norm": 52.65121078491211, + "learning_rate": 0.0, + "loss": 0.8525, + "step": 17270 + }, + { + "epoch": 4.024219841639497, + "grad_norm": 10.610958099365234, + "learning_rate": 0.0, + "loss": 1.2737, + "step": 17280 + }, + { + "epoch": 4.0265486725663715, + "grad_norm": 18.813425064086914, + "learning_rate": 0.0, + "loss": 0.8997, + "step": 17290 + }, + { + "epoch": 4.028877503493247, + "grad_norm": 21.70282554626465, + "learning_rate": 0.0, + "loss": 0.9929, + "step": 17300 + }, + { + "epoch": 4.031206334420121, + "grad_norm": 5.4507036209106445, + "learning_rate": 0.0, + "loss": 1.9675, + "step": 17310 + }, + { + "epoch": 4.033535165346996, + "grad_norm": 7.815976619720459, + "learning_rate": 0.0, + "loss": 0.6941, + "step": 17320 + }, + { + "epoch": 4.0358639962738705, + "grad_norm": 21.757108688354492, + "learning_rate": 0.0, + "loss": 1.1081, + "step": 17330 + }, + { + "epoch": 4.038192827200745, + "grad_norm": 67.65559387207031, + "learning_rate": 0.0, + "loss": 1.4166, + "step": 17340 + }, + { + "epoch": 4.04052165812762, + "grad_norm": 9.288952827453613, + "learning_rate": 0.0, + "loss": 1.2824, + "step": 17350 + }, + { + "epoch": 4.042850489054494, + "grad_norm": 41.26054382324219, + "learning_rate": 0.0, + "loss": 1.3587, + "step": 17360 + }, + { + "epoch": 4.0451793199813695, + "grad_norm": 6.358667850494385, + "learning_rate": 0.0, + "loss": 0.7649, + "step": 17370 + }, + { + "epoch": 4.047508150908244, + "grad_norm": 10.879199981689453, + "learning_rate": 0.0, + "loss": 1.4232, + "step": 17380 + }, + { + "epoch": 4.049836981835119, + "grad_norm": 6.230020523071289, + "learning_rate": 0.0, + "loss": 1.0463, + "step": 17390 + }, + { + "epoch": 4.052165812761993, + "grad_norm": 15.603515625, + "learning_rate": 0.0, + "loss": 1.223, + "step": 17400 + }, + { + "epoch": 4.0544946436888685, + "grad_norm": 14.09308910369873, + "learning_rate": 0.0, + "loss": 1.0864, + "step": 17410 + }, + { + "epoch": 4.056823474615743, + "grad_norm": 65.15744018554688, + "learning_rate": 0.0, + "loss": 1.2502, + "step": 17420 + }, + { + "epoch": 4.059152305542618, + "grad_norm": 53.656063079833984, + "learning_rate": 0.0, + "loss": 0.9074, + "step": 17430 + }, + { + "epoch": 4.061481136469492, + "grad_norm": 26.082855224609375, + "learning_rate": 0.0, + "loss": 1.3669, + "step": 17440 + }, + { + "epoch": 4.063809967396367, + "grad_norm": 37.48140335083008, + "learning_rate": 0.0, + "loss": 1.2421, + "step": 17450 + }, + { + "epoch": 4.066138798323242, + "grad_norm": 8.820276260375977, + "learning_rate": 0.0, + "loss": 1.2692, + "step": 17460 + }, + { + "epoch": 4.068467629250116, + "grad_norm": 11.496400833129883, + "learning_rate": 0.0, + "loss": 1.0396, + "step": 17470 + }, + { + "epoch": 4.070796460176991, + "grad_norm": 18.882984161376953, + "learning_rate": 0.0, + "loss": 0.9758, + "step": 17480 + }, + { + "epoch": 4.073125291103866, + "grad_norm": 8.582090377807617, + "learning_rate": 0.0, + "loss": 1.1566, + "step": 17490 + }, + { + "epoch": 4.075454122030741, + "grad_norm": 13.812816619873047, + "learning_rate": 0.0, + "loss": 1.6421, + "step": 17500 + }, + { + "epoch": 4.077782952957615, + "grad_norm": 52.63260269165039, + "learning_rate": 0.0, + "loss": 1.0587, + "step": 17510 + }, + { + "epoch": 4.08011178388449, + "grad_norm": 10.908785820007324, + "learning_rate": 0.0, + "loss": 0.855, + "step": 17520 + }, + { + "epoch": 4.0824406148113646, + "grad_norm": 8.219367027282715, + "learning_rate": 0.0, + "loss": 1.009, + "step": 17530 + }, + { + "epoch": 4.08476944573824, + "grad_norm": 16.966997146606445, + "learning_rate": 0.0, + "loss": 1.2723, + "step": 17540 + }, + { + "epoch": 4.087098276665114, + "grad_norm": 14.397762298583984, + "learning_rate": 0.0, + "loss": 1.0312, + "step": 17550 + }, + { + "epoch": 4.089427107591989, + "grad_norm": 13.03594970703125, + "learning_rate": 0.0, + "loss": 1.2744, + "step": 17560 + }, + { + "epoch": 4.0917559385188635, + "grad_norm": 6.213973522186279, + "learning_rate": 0.0, + "loss": 1.0554, + "step": 17570 + }, + { + "epoch": 4.094084769445738, + "grad_norm": 42.39612579345703, + "learning_rate": 0.0, + "loss": 1.4296, + "step": 17580 + }, + { + "epoch": 4.096413600372613, + "grad_norm": 6.726439476013184, + "learning_rate": 0.0, + "loss": 1.0909, + "step": 17590 + }, + { + "epoch": 4.098742431299487, + "grad_norm": 12.621766090393066, + "learning_rate": 0.0, + "loss": 1.542, + "step": 17600 + }, + { + "epoch": 4.1010712622263625, + "grad_norm": 11.763903617858887, + "learning_rate": 0.0, + "loss": 1.6493, + "step": 17610 + }, + { + "epoch": 4.103400093153237, + "grad_norm": 6.310747146606445, + "learning_rate": 0.0, + "loss": 1.0642, + "step": 17620 + }, + { + "epoch": 4.105728924080112, + "grad_norm": 13.970367431640625, + "learning_rate": 0.0, + "loss": 1.2096, + "step": 17630 + }, + { + "epoch": 4.108057755006986, + "grad_norm": 18.80079460144043, + "learning_rate": 0.0, + "loss": 1.1508, + "step": 17640 + }, + { + "epoch": 4.1103865859338615, + "grad_norm": 5.596470832824707, + "learning_rate": 0.0, + "loss": 1.6874, + "step": 17650 + }, + { + "epoch": 4.112715416860736, + "grad_norm": 21.15366554260254, + "learning_rate": 0.0, + "loss": 1.482, + "step": 17660 + }, + { + "epoch": 4.115044247787611, + "grad_norm": 14.376870155334473, + "learning_rate": 0.0, + "loss": 1.2993, + "step": 17670 + }, + { + "epoch": 4.117373078714485, + "grad_norm": 3.591233730316162, + "learning_rate": 0.0, + "loss": 1.1712, + "step": 17680 + }, + { + "epoch": 4.11970190964136, + "grad_norm": 85.7278823852539, + "learning_rate": 0.0, + "loss": 2.0886, + "step": 17690 + }, + { + "epoch": 4.122030740568235, + "grad_norm": 24.37750816345215, + "learning_rate": 0.0, + "loss": 1.1597, + "step": 17700 + }, + { + "epoch": 4.124359571495109, + "grad_norm": 9.48483657836914, + "learning_rate": 0.0, + "loss": 0.9762, + "step": 17710 + }, + { + "epoch": 4.126688402421984, + "grad_norm": 16.925823211669922, + "learning_rate": 0.0, + "loss": 0.8279, + "step": 17720 + }, + { + "epoch": 4.129017233348859, + "grad_norm": 5.349981784820557, + "learning_rate": 0.0, + "loss": 1.1843, + "step": 17730 + }, + { + "epoch": 4.131346064275734, + "grad_norm": 6.957655906677246, + "learning_rate": 0.0, + "loss": 1.16, + "step": 17740 + }, + { + "epoch": 4.133674895202608, + "grad_norm": 8.322871208190918, + "learning_rate": 0.0, + "loss": 0.7904, + "step": 17750 + }, + { + "epoch": 4.136003726129483, + "grad_norm": 37.46146774291992, + "learning_rate": 0.0, + "loss": 1.2393, + "step": 17760 + }, + { + "epoch": 4.138332557056358, + "grad_norm": 4.923023700714111, + "learning_rate": 0.0, + "loss": 2.4185, + "step": 17770 + }, + { + "epoch": 4.140661387983233, + "grad_norm": 21.715543746948242, + "learning_rate": 0.0, + "loss": 1.0933, + "step": 17780 + }, + { + "epoch": 4.142990218910107, + "grad_norm": 26.76923179626465, + "learning_rate": 0.0, + "loss": 1.3966, + "step": 17790 + }, + { + "epoch": 4.145319049836981, + "grad_norm": 18.267608642578125, + "learning_rate": 0.0, + "loss": 1.5447, + "step": 17800 + }, + { + "epoch": 4.147647880763857, + "grad_norm": 13.733194351196289, + "learning_rate": 0.0, + "loss": 1.4679, + "step": 17810 + }, + { + "epoch": 4.149976711690731, + "grad_norm": 5.581467628479004, + "learning_rate": 0.0, + "loss": 1.0773, + "step": 17820 + }, + { + "epoch": 4.152305542617606, + "grad_norm": 8.281193733215332, + "learning_rate": 0.0, + "loss": 1.3519, + "step": 17830 + }, + { + "epoch": 4.15463437354448, + "grad_norm": 16.052032470703125, + "learning_rate": 0.0, + "loss": 1.203, + "step": 17840 + }, + { + "epoch": 4.156963204471356, + "grad_norm": 32.61860656738281, + "learning_rate": 0.0, + "loss": 1.0071, + "step": 17850 + }, + { + "epoch": 4.15929203539823, + "grad_norm": 10.298188209533691, + "learning_rate": 0.0, + "loss": 1.2432, + "step": 17860 + }, + { + "epoch": 4.161620866325105, + "grad_norm": 13.705058097839355, + "learning_rate": 0.0, + "loss": 1.2159, + "step": 17870 + }, + { + "epoch": 4.163949697251979, + "grad_norm": 7.2891669273376465, + "learning_rate": 0.0, + "loss": 1.7573, + "step": 17880 + }, + { + "epoch": 4.166278528178855, + "grad_norm": 4.742997169494629, + "learning_rate": 0.0, + "loss": 1.3245, + "step": 17890 + }, + { + "epoch": 4.168607359105729, + "grad_norm": 9.10926342010498, + "learning_rate": 0.0, + "loss": 1.3587, + "step": 17900 + }, + { + "epoch": 4.170936190032604, + "grad_norm": 34.803287506103516, + "learning_rate": 0.0, + "loss": 0.8937, + "step": 17910 + }, + { + "epoch": 4.173265020959478, + "grad_norm": 24.677139282226562, + "learning_rate": 0.0, + "loss": 1.3338, + "step": 17920 + }, + { + "epoch": 4.175593851886353, + "grad_norm": 10.664273262023926, + "learning_rate": 0.0, + "loss": 1.0591, + "step": 17930 + }, + { + "epoch": 4.177922682813228, + "grad_norm": 10.509198188781738, + "learning_rate": 0.0, + "loss": 1.0005, + "step": 17940 + }, + { + "epoch": 4.180251513740102, + "grad_norm": 7.73974609375, + "learning_rate": 0.0, + "loss": 1.1359, + "step": 17950 + }, + { + "epoch": 4.182580344666977, + "grad_norm": 6.254695415496826, + "learning_rate": 0.0, + "loss": 0.8587, + "step": 17960 + }, + { + "epoch": 4.184909175593852, + "grad_norm": 8.58466625213623, + "learning_rate": 0.0, + "loss": 0.9862, + "step": 17970 + }, + { + "epoch": 4.187238006520727, + "grad_norm": 11.516722679138184, + "learning_rate": 0.0, + "loss": 1.5512, + "step": 17980 + }, + { + "epoch": 4.189566837447601, + "grad_norm": 25.665462493896484, + "learning_rate": 0.0, + "loss": 1.3756, + "step": 17990 + }, + { + "epoch": 4.191895668374476, + "grad_norm": 205.93728637695312, + "learning_rate": 0.0, + "loss": 1.4687, + "step": 18000 + }, + { + "epoch": 4.194224499301351, + "grad_norm": 5.3130340576171875, + "learning_rate": 0.0, + "loss": 1.0205, + "step": 18010 + }, + { + "epoch": 4.196553330228226, + "grad_norm": 9.158675193786621, + "learning_rate": 0.0, + "loss": 0.903, + "step": 18020 + }, + { + "epoch": 4.1988821611551, + "grad_norm": 11.879533767700195, + "learning_rate": 0.0, + "loss": 1.3574, + "step": 18030 + }, + { + "epoch": 4.2012109920819745, + "grad_norm": 56.382537841796875, + "learning_rate": 0.0, + "loss": 1.7197, + "step": 18040 + }, + { + "epoch": 4.20353982300885, + "grad_norm": 14.329750061035156, + "learning_rate": 0.0, + "loss": 1.4055, + "step": 18050 + }, + { + "epoch": 4.205868653935724, + "grad_norm": 8.853121757507324, + "learning_rate": 0.0, + "loss": 1.519, + "step": 18060 + }, + { + "epoch": 4.208197484862599, + "grad_norm": 12.149033546447754, + "learning_rate": 0.0, + "loss": 1.1448, + "step": 18070 + }, + { + "epoch": 4.2105263157894735, + "grad_norm": 15.463021278381348, + "learning_rate": 0.0, + "loss": 0.9091, + "step": 18080 + }, + { + "epoch": 4.212855146716349, + "grad_norm": 17.922626495361328, + "learning_rate": 0.0, + "loss": 1.2816, + "step": 18090 + }, + { + "epoch": 4.215183977643223, + "grad_norm": 16.048818588256836, + "learning_rate": 0.0, + "loss": 1.2221, + "step": 18100 + }, + { + "epoch": 4.217512808570098, + "grad_norm": 7.998112201690674, + "learning_rate": 0.0, + "loss": 1.4312, + "step": 18110 + }, + { + "epoch": 4.219841639496972, + "grad_norm": 6.729568004608154, + "learning_rate": 0.0, + "loss": 0.8463, + "step": 18120 + }, + { + "epoch": 4.222170470423848, + "grad_norm": 9.382912635803223, + "learning_rate": 0.0, + "loss": 2.0268, + "step": 18130 + }, + { + "epoch": 4.224499301350722, + "grad_norm": 11.232508659362793, + "learning_rate": 0.0, + "loss": 1.6308, + "step": 18140 + }, + { + "epoch": 4.226828132277596, + "grad_norm": 7.6993536949157715, + "learning_rate": 0.0, + "loss": 0.9748, + "step": 18150 + }, + { + "epoch": 4.229156963204471, + "grad_norm": 14.037554740905762, + "learning_rate": 0.0, + "loss": 1.0376, + "step": 18160 + }, + { + "epoch": 4.231485794131346, + "grad_norm": 37.87745666503906, + "learning_rate": 0.0, + "loss": 2.3579, + "step": 18170 + }, + { + "epoch": 4.233814625058221, + "grad_norm": 51.04754638671875, + "learning_rate": 0.0, + "loss": 1.3791, + "step": 18180 + }, + { + "epoch": 4.236143455985095, + "grad_norm": 11.547856330871582, + "learning_rate": 0.0, + "loss": 0.8157, + "step": 18190 + }, + { + "epoch": 4.23847228691197, + "grad_norm": 9.100687980651855, + "learning_rate": 0.0, + "loss": 0.8207, + "step": 18200 + }, + { + "epoch": 4.240801117838845, + "grad_norm": 12.361568450927734, + "learning_rate": 0.0, + "loss": 1.0341, + "step": 18210 + }, + { + "epoch": 4.24312994876572, + "grad_norm": 15.49877643585205, + "learning_rate": 0.0, + "loss": 1.5983, + "step": 18220 + }, + { + "epoch": 4.245458779692594, + "grad_norm": 51.21619415283203, + "learning_rate": 0.0, + "loss": 1.2015, + "step": 18230 + }, + { + "epoch": 4.247787610619469, + "grad_norm": 21.072568893432617, + "learning_rate": 0.0, + "loss": 0.853, + "step": 18240 + }, + { + "epoch": 4.250116441546344, + "grad_norm": 5.333137512207031, + "learning_rate": 0.0, + "loss": 0.89, + "step": 18250 + }, + { + "epoch": 4.252445272473219, + "grad_norm": 5.594559669494629, + "learning_rate": 0.0, + "loss": 1.3852, + "step": 18260 + }, + { + "epoch": 4.254774103400093, + "grad_norm": 9.22107982635498, + "learning_rate": 0.0, + "loss": 0.9253, + "step": 18270 + }, + { + "epoch": 4.2571029343269675, + "grad_norm": 17.17836570739746, + "learning_rate": 0.0, + "loss": 0.8569, + "step": 18280 + }, + { + "epoch": 4.259431765253843, + "grad_norm": 7.7796630859375, + "learning_rate": 0.0, + "loss": 0.9931, + "step": 18290 + }, + { + "epoch": 4.261760596180717, + "grad_norm": 10.182730674743652, + "learning_rate": 0.0, + "loss": 0.9802, + "step": 18300 + }, + { + "epoch": 4.264089427107592, + "grad_norm": 6.655672073364258, + "learning_rate": 0.0, + "loss": 0.8961, + "step": 18310 + }, + { + "epoch": 4.2664182580344665, + "grad_norm": 6.27683162689209, + "learning_rate": 0.0, + "loss": 1.6628, + "step": 18320 + }, + { + "epoch": 4.268747088961342, + "grad_norm": 10.62850570678711, + "learning_rate": 0.0, + "loss": 1.5432, + "step": 18330 + }, + { + "epoch": 4.271075919888216, + "grad_norm": 25.98421287536621, + "learning_rate": 0.0, + "loss": 1.2067, + "step": 18340 + }, + { + "epoch": 4.273404750815091, + "grad_norm": 27.159278869628906, + "learning_rate": 0.0, + "loss": 1.2735, + "step": 18350 + }, + { + "epoch": 4.2757335817419655, + "grad_norm": 23.231164932250977, + "learning_rate": 0.0, + "loss": 1.3331, + "step": 18360 + }, + { + "epoch": 4.27806241266884, + "grad_norm": 96.513427734375, + "learning_rate": 0.0, + "loss": 1.5835, + "step": 18370 + }, + { + "epoch": 4.280391243595715, + "grad_norm": 6.849877834320068, + "learning_rate": 0.0, + "loss": 1.0837, + "step": 18380 + }, + { + "epoch": 4.282720074522589, + "grad_norm": 7.442630767822266, + "learning_rate": 0.0, + "loss": 1.4346, + "step": 18390 + }, + { + "epoch": 4.2850489054494645, + "grad_norm": 62.38019561767578, + "learning_rate": 0.0, + "loss": 1.2822, + "step": 18400 + }, + { + "epoch": 4.287377736376339, + "grad_norm": 7.65008020401001, + "learning_rate": 0.0, + "loss": 1.3873, + "step": 18410 + }, + { + "epoch": 4.289706567303214, + "grad_norm": 31.235200881958008, + "learning_rate": 0.0, + "loss": 1.4172, + "step": 18420 + }, + { + "epoch": 4.292035398230088, + "grad_norm": 5.858349800109863, + "learning_rate": 0.0, + "loss": 0.7914, + "step": 18430 + }, + { + "epoch": 4.2943642291569635, + "grad_norm": 8.754595756530762, + "learning_rate": 0.0, + "loss": 1.2774, + "step": 18440 + }, + { + "epoch": 4.296693060083838, + "grad_norm": 64.94791412353516, + "learning_rate": 0.0, + "loss": 1.2843, + "step": 18450 + }, + { + "epoch": 4.299021891010713, + "grad_norm": 3.240407943725586, + "learning_rate": 0.0, + "loss": 0.8924, + "step": 18460 + }, + { + "epoch": 4.301350721937587, + "grad_norm": 15.686180114746094, + "learning_rate": 0.0, + "loss": 1.1514, + "step": 18470 + }, + { + "epoch": 4.3036795528644625, + "grad_norm": 10.717156410217285, + "learning_rate": 0.0, + "loss": 1.4731, + "step": 18480 + }, + { + "epoch": 4.306008383791337, + "grad_norm": 8.399251937866211, + "learning_rate": 0.0, + "loss": 1.0628, + "step": 18490 + }, + { + "epoch": 4.308337214718211, + "grad_norm": 5.603919506072998, + "learning_rate": 0.0, + "loss": 1.1527, + "step": 18500 + }, + { + "epoch": 4.310666045645086, + "grad_norm": 30.59088706970215, + "learning_rate": 0.0, + "loss": 0.9832, + "step": 18510 + }, + { + "epoch": 4.312994876571961, + "grad_norm": 9.03952407836914, + "learning_rate": 0.0, + "loss": 1.0315, + "step": 18520 + }, + { + "epoch": 4.315323707498836, + "grad_norm": 59.89068603515625, + "learning_rate": 0.0, + "loss": 1.0991, + "step": 18530 + }, + { + "epoch": 4.31765253842571, + "grad_norm": 105.38227844238281, + "learning_rate": 0.0, + "loss": 1.2814, + "step": 18540 + }, + { + "epoch": 4.319981369352585, + "grad_norm": 15.269956588745117, + "learning_rate": 0.0, + "loss": 1.2359, + "step": 18550 + }, + { + "epoch": 4.32231020027946, + "grad_norm": 57.37482833862305, + "learning_rate": 0.0, + "loss": 1.163, + "step": 18560 + }, + { + "epoch": 4.324639031206335, + "grad_norm": 23.360641479492188, + "learning_rate": 0.0, + "loss": 0.8576, + "step": 18570 + }, + { + "epoch": 4.326967862133209, + "grad_norm": 14.31608772277832, + "learning_rate": 0.0, + "loss": 1.4788, + "step": 18580 + }, + { + "epoch": 4.329296693060084, + "grad_norm": 40.62135314941406, + "learning_rate": 0.0, + "loss": 1.2003, + "step": 18590 + }, + { + "epoch": 4.3316255239869585, + "grad_norm": 39.43674850463867, + "learning_rate": 0.0, + "loss": 1.2967, + "step": 18600 + }, + { + "epoch": 4.333954354913833, + "grad_norm": 25.50300407409668, + "learning_rate": 0.0, + "loss": 1.8007, + "step": 18610 + }, + { + "epoch": 4.336283185840708, + "grad_norm": 56.91150665283203, + "learning_rate": 0.0, + "loss": 1.2731, + "step": 18620 + }, + { + "epoch": 4.338612016767582, + "grad_norm": 13.99095344543457, + "learning_rate": 0.0, + "loss": 1.4154, + "step": 18630 + }, + { + "epoch": 4.3409408476944575, + "grad_norm": 15.199361801147461, + "learning_rate": 0.0, + "loss": 1.5806, + "step": 18640 + }, + { + "epoch": 4.343269678621332, + "grad_norm": 45.43046188354492, + "learning_rate": 0.0, + "loss": 1.4277, + "step": 18650 + }, + { + "epoch": 4.345598509548207, + "grad_norm": 4.511590480804443, + "learning_rate": 0.0, + "loss": 0.894, + "step": 18660 + }, + { + "epoch": 4.347927340475081, + "grad_norm": 51.11252212524414, + "learning_rate": 0.0, + "loss": 1.3697, + "step": 18670 + }, + { + "epoch": 4.3502561714019565, + "grad_norm": 5.312637805938721, + "learning_rate": 0.0, + "loss": 1.1775, + "step": 18680 + }, + { + "epoch": 4.352585002328831, + "grad_norm": 23.01420021057129, + "learning_rate": 0.0, + "loss": 1.209, + "step": 18690 + }, + { + "epoch": 4.354913833255706, + "grad_norm": 7.891007423400879, + "learning_rate": 0.0, + "loss": 0.8061, + "step": 18700 + }, + { + "epoch": 4.35724266418258, + "grad_norm": 11.448382377624512, + "learning_rate": 0.0, + "loss": 1.6594, + "step": 18710 + }, + { + "epoch": 4.359571495109455, + "grad_norm": 18.895126342773438, + "learning_rate": 0.0, + "loss": 1.0112, + "step": 18720 + }, + { + "epoch": 4.36190032603633, + "grad_norm": 15.76264762878418, + "learning_rate": 0.0, + "loss": 0.9219, + "step": 18730 + }, + { + "epoch": 4.364229156963204, + "grad_norm": 13.89644718170166, + "learning_rate": 0.0, + "loss": 1.4194, + "step": 18740 + }, + { + "epoch": 4.366557987890079, + "grad_norm": 7.580521106719971, + "learning_rate": 0.0, + "loss": 1.0652, + "step": 18750 + }, + { + "epoch": 4.368886818816954, + "grad_norm": 4.65724515914917, + "learning_rate": 0.0, + "loss": 1.1159, + "step": 18760 + }, + { + "epoch": 4.371215649743829, + "grad_norm": 7.993378162384033, + "learning_rate": 0.0, + "loss": 0.9724, + "step": 18770 + }, + { + "epoch": 4.373544480670703, + "grad_norm": 5.239194393157959, + "learning_rate": 0.0, + "loss": 0.9784, + "step": 18780 + }, + { + "epoch": 4.375873311597578, + "grad_norm": 7.044046878814697, + "learning_rate": 0.0, + "loss": 0.8269, + "step": 18790 + }, + { + "epoch": 4.378202142524453, + "grad_norm": 13.0368070602417, + "learning_rate": 0.0, + "loss": 2.584, + "step": 18800 + }, + { + "epoch": 4.380530973451328, + "grad_norm": 9.664212226867676, + "learning_rate": 0.0, + "loss": 1.1443, + "step": 18810 + }, + { + "epoch": 4.382859804378202, + "grad_norm": 13.87582015991211, + "learning_rate": 0.0, + "loss": 0.9475, + "step": 18820 + }, + { + "epoch": 4.385188635305077, + "grad_norm": 7.444732666015625, + "learning_rate": 0.0, + "loss": 1.1998, + "step": 18830 + }, + { + "epoch": 4.387517466231952, + "grad_norm": 17.309480667114258, + "learning_rate": 0.0, + "loss": 0.9564, + "step": 18840 + }, + { + "epoch": 4.389846297158826, + "grad_norm": 16.705272674560547, + "learning_rate": 0.0, + "loss": 1.2803, + "step": 18850 + }, + { + "epoch": 4.392175128085701, + "grad_norm": 12.118639945983887, + "learning_rate": 0.0, + "loss": 1.1282, + "step": 18860 + }, + { + "epoch": 4.394503959012575, + "grad_norm": 15.16536808013916, + "learning_rate": 0.0, + "loss": 1.1275, + "step": 18870 + }, + { + "epoch": 4.396832789939451, + "grad_norm": 4.360487461090088, + "learning_rate": 0.0, + "loss": 1.4061, + "step": 18880 + }, + { + "epoch": 4.399161620866325, + "grad_norm": 22.97736930847168, + "learning_rate": 0.0, + "loss": 1.6106, + "step": 18890 + }, + { + "epoch": 4.4014904517932, + "grad_norm": 62.73188018798828, + "learning_rate": 0.0, + "loss": 1.2969, + "step": 18900 + }, + { + "epoch": 4.403819282720074, + "grad_norm": 15.173748970031738, + "learning_rate": 0.0, + "loss": 1.6601, + "step": 18910 + }, + { + "epoch": 4.40614811364695, + "grad_norm": 12.659820556640625, + "learning_rate": 0.0, + "loss": 1.4902, + "step": 18920 + }, + { + "epoch": 4.408476944573824, + "grad_norm": 9.84178352355957, + "learning_rate": 0.0, + "loss": 1.0183, + "step": 18930 + }, + { + "epoch": 4.410805775500698, + "grad_norm": 16.406572341918945, + "learning_rate": 0.0, + "loss": 0.9501, + "step": 18940 + }, + { + "epoch": 4.413134606427573, + "grad_norm": 8.493030548095703, + "learning_rate": 0.0, + "loss": 0.9876, + "step": 18950 + }, + { + "epoch": 4.415463437354448, + "grad_norm": 8.988524436950684, + "learning_rate": 0.0, + "loss": 1.1732, + "step": 18960 + }, + { + "epoch": 4.417792268281323, + "grad_norm": 7.769420146942139, + "learning_rate": 0.0, + "loss": 1.3663, + "step": 18970 + }, + { + "epoch": 4.420121099208197, + "grad_norm": 10.500604629516602, + "learning_rate": 0.0, + "loss": 1.9919, + "step": 18980 + }, + { + "epoch": 4.422449930135072, + "grad_norm": 13.152361869812012, + "learning_rate": 0.0, + "loss": 1.5332, + "step": 18990 + }, + { + "epoch": 4.424778761061947, + "grad_norm": 29.10773468017578, + "learning_rate": 0.0, + "loss": 1.6014, + "step": 19000 + }, + { + "epoch": 4.427107591988822, + "grad_norm": 9.866641998291016, + "learning_rate": 0.0, + "loss": 1.3321, + "step": 19010 + }, + { + "epoch": 4.429436422915696, + "grad_norm": 7.406885623931885, + "learning_rate": 0.0, + "loss": 0.9355, + "step": 19020 + }, + { + "epoch": 4.431765253842571, + "grad_norm": 64.86759948730469, + "learning_rate": 0.0, + "loss": 1.4649, + "step": 19030 + }, + { + "epoch": 4.434094084769446, + "grad_norm": 24.239585876464844, + "learning_rate": 0.0, + "loss": 1.3549, + "step": 19040 + }, + { + "epoch": 4.436422915696321, + "grad_norm": 12.164912223815918, + "learning_rate": 0.0, + "loss": 0.9648, + "step": 19050 + }, + { + "epoch": 4.438751746623195, + "grad_norm": 7.17809534072876, + "learning_rate": 0.0, + "loss": 1.0822, + "step": 19060 + }, + { + "epoch": 4.4410805775500695, + "grad_norm": 15.311285972595215, + "learning_rate": 0.0, + "loss": 0.918, + "step": 19070 + }, + { + "epoch": 4.443409408476945, + "grad_norm": 6.92308235168457, + "learning_rate": 0.0, + "loss": 1.1334, + "step": 19080 + }, + { + "epoch": 4.445738239403819, + "grad_norm": 21.64579963684082, + "learning_rate": 0.0, + "loss": 1.011, + "step": 19090 + }, + { + "epoch": 4.448067070330694, + "grad_norm": 17.369667053222656, + "learning_rate": 0.0, + "loss": 1.0185, + "step": 19100 + }, + { + "epoch": 4.4503959012575685, + "grad_norm": 6.284876823425293, + "learning_rate": 0.0, + "loss": 1.3311, + "step": 19110 + }, + { + "epoch": 4.452724732184444, + "grad_norm": 4.605877876281738, + "learning_rate": 0.0, + "loss": 0.9037, + "step": 19120 + }, + { + "epoch": 4.455053563111318, + "grad_norm": 8.776147842407227, + "learning_rate": 0.0, + "loss": 1.2707, + "step": 19130 + }, + { + "epoch": 4.457382394038193, + "grad_norm": 8.23446273803711, + "learning_rate": 0.0, + "loss": 1.194, + "step": 19140 + }, + { + "epoch": 4.4597112249650674, + "grad_norm": 6.68794059753418, + "learning_rate": 0.0, + "loss": 1.1309, + "step": 19150 + }, + { + "epoch": 4.462040055891943, + "grad_norm": 5.445720672607422, + "learning_rate": 0.0, + "loss": 1.2681, + "step": 19160 + }, + { + "epoch": 4.464368886818817, + "grad_norm": 47.82493591308594, + "learning_rate": 0.0, + "loss": 1.0086, + "step": 19170 + }, + { + "epoch": 4.466697717745692, + "grad_norm": 12.19786548614502, + "learning_rate": 0.0, + "loss": 1.3866, + "step": 19180 + }, + { + "epoch": 4.469026548672566, + "grad_norm": 27.628253936767578, + "learning_rate": 0.0, + "loss": 1.8903, + "step": 19190 + }, + { + "epoch": 4.471355379599441, + "grad_norm": 6.904911041259766, + "learning_rate": 0.0, + "loss": 1.2476, + "step": 19200 + }, + { + "epoch": 4.473684210526316, + "grad_norm": 13.863846778869629, + "learning_rate": 0.0, + "loss": 0.859, + "step": 19210 + }, + { + "epoch": 4.47601304145319, + "grad_norm": 124.06563568115234, + "learning_rate": 0.0, + "loss": 1.8265, + "step": 19220 + }, + { + "epoch": 4.478341872380065, + "grad_norm": 104.48321533203125, + "learning_rate": 0.0, + "loss": 1.2933, + "step": 19230 + }, + { + "epoch": 4.48067070330694, + "grad_norm": 29.74410629272461, + "learning_rate": 0.0, + "loss": 1.3518, + "step": 19240 + }, + { + "epoch": 4.482999534233815, + "grad_norm": 9.21904468536377, + "learning_rate": 0.0, + "loss": 1.3145, + "step": 19250 + }, + { + "epoch": 4.485328365160689, + "grad_norm": 17.82377815246582, + "learning_rate": 0.0, + "loss": 1.1376, + "step": 19260 + }, + { + "epoch": 4.487657196087564, + "grad_norm": 7.17800235748291, + "learning_rate": 0.0, + "loss": 1.6021, + "step": 19270 + }, + { + "epoch": 4.489986027014439, + "grad_norm": 8.264793395996094, + "learning_rate": 0.0, + "loss": 1.0782, + "step": 19280 + }, + { + "epoch": 4.492314857941313, + "grad_norm": 10.051447868347168, + "learning_rate": 0.0, + "loss": 1.046, + "step": 19290 + }, + { + "epoch": 4.494643688868188, + "grad_norm": 9.666632652282715, + "learning_rate": 0.0, + "loss": 0.8611, + "step": 19300 + }, + { + "epoch": 4.4969725197950625, + "grad_norm": 5.8455986976623535, + "learning_rate": 0.0, + "loss": 1.1852, + "step": 19310 + }, + { + "epoch": 4.499301350721938, + "grad_norm": 40.72843933105469, + "learning_rate": 0.0, + "loss": 1.4481, + "step": 19320 + }, + { + "epoch": 4.501630181648812, + "grad_norm": 14.480032920837402, + "learning_rate": 0.0, + "loss": 1.086, + "step": 19330 + }, + { + "epoch": 4.503959012575687, + "grad_norm": 8.125821113586426, + "learning_rate": 0.0, + "loss": 1.1448, + "step": 19340 + }, + { + "epoch": 4.5062878435025615, + "grad_norm": 8.457114219665527, + "learning_rate": 0.0, + "loss": 0.9219, + "step": 19350 + }, + { + "epoch": 4.508616674429437, + "grad_norm": 46.61256790161133, + "learning_rate": 0.0, + "loss": 1.4796, + "step": 19360 + }, + { + "epoch": 4.510945505356311, + "grad_norm": 12.012411117553711, + "learning_rate": 0.0, + "loss": 0.9379, + "step": 19370 + }, + { + "epoch": 4.513274336283186, + "grad_norm": 10.32507038116455, + "learning_rate": 0.0, + "loss": 1.0856, + "step": 19380 + }, + { + "epoch": 4.5156031672100605, + "grad_norm": 18.621065139770508, + "learning_rate": 0.0, + "loss": 1.2173, + "step": 19390 + }, + { + "epoch": 4.517931998136936, + "grad_norm": 5.736082553863525, + "learning_rate": 0.0, + "loss": 0.866, + "step": 19400 + }, + { + "epoch": 4.52026082906381, + "grad_norm": 8.20267391204834, + "learning_rate": 0.0, + "loss": 1.3217, + "step": 19410 + }, + { + "epoch": 4.522589659990684, + "grad_norm": 7.2390947341918945, + "learning_rate": 0.0, + "loss": 1.6424, + "step": 19420 + }, + { + "epoch": 4.5249184909175595, + "grad_norm": 13.071420669555664, + "learning_rate": 0.0, + "loss": 0.9959, + "step": 19430 + }, + { + "epoch": 4.527247321844434, + "grad_norm": 7.594562530517578, + "learning_rate": 0.0, + "loss": 1.2662, + "step": 19440 + }, + { + "epoch": 4.529576152771309, + "grad_norm": 31.20307159423828, + "learning_rate": 0.0, + "loss": 1.7857, + "step": 19450 + }, + { + "epoch": 4.531904983698183, + "grad_norm": 11.719870567321777, + "learning_rate": 0.0, + "loss": 1.0293, + "step": 19460 + }, + { + "epoch": 4.5342338146250585, + "grad_norm": 16.678586959838867, + "learning_rate": 0.0, + "loss": 1.2137, + "step": 19470 + }, + { + "epoch": 4.536562645551933, + "grad_norm": 6.821103096008301, + "learning_rate": 0.0, + "loss": 0.7388, + "step": 19480 + }, + { + "epoch": 4.538891476478808, + "grad_norm": 22.45350456237793, + "learning_rate": 0.0, + "loss": 1.3895, + "step": 19490 + }, + { + "epoch": 4.541220307405682, + "grad_norm": 3.4364640712738037, + "learning_rate": 0.0, + "loss": 1.3045, + "step": 19500 + }, + { + "epoch": 4.543549138332557, + "grad_norm": 12.992962837219238, + "learning_rate": 0.0, + "loss": 0.9766, + "step": 19510 + }, + { + "epoch": 4.545877969259432, + "grad_norm": 8.960369110107422, + "learning_rate": 0.0, + "loss": 1.4743, + "step": 19520 + }, + { + "epoch": 4.548206800186307, + "grad_norm": 9.852701187133789, + "learning_rate": 0.0, + "loss": 1.0763, + "step": 19530 + }, + { + "epoch": 4.550535631113181, + "grad_norm": 5.533620357513428, + "learning_rate": 0.0, + "loss": 1.6823, + "step": 19540 + }, + { + "epoch": 4.552864462040056, + "grad_norm": 11.03526496887207, + "learning_rate": 0.0, + "loss": 1.3972, + "step": 19550 + }, + { + "epoch": 4.555193292966931, + "grad_norm": 11.094636917114258, + "learning_rate": 0.0, + "loss": 0.9835, + "step": 19560 + }, + { + "epoch": 4.557522123893805, + "grad_norm": 9.783246994018555, + "learning_rate": 0.0, + "loss": 1.2277, + "step": 19570 + }, + { + "epoch": 4.55985095482068, + "grad_norm": 32.54227828979492, + "learning_rate": 0.0, + "loss": 1.4958, + "step": 19580 + }, + { + "epoch": 4.562179785747555, + "grad_norm": 15.014932632446289, + "learning_rate": 0.0, + "loss": 1.0942, + "step": 19590 + }, + { + "epoch": 4.56450861667443, + "grad_norm": 10.05160140991211, + "learning_rate": 0.0, + "loss": 1.4059, + "step": 19600 + }, + { + "epoch": 4.566837447601304, + "grad_norm": 7.201435089111328, + "learning_rate": 0.0, + "loss": 1.5083, + "step": 19610 + }, + { + "epoch": 4.569166278528179, + "grad_norm": 9.957510948181152, + "learning_rate": 0.0, + "loss": 1.0267, + "step": 19620 + }, + { + "epoch": 4.5714951094550536, + "grad_norm": 38.4839973449707, + "learning_rate": 0.0, + "loss": 1.0244, + "step": 19630 + }, + { + "epoch": 4.573823940381928, + "grad_norm": 15.924169540405273, + "learning_rate": 0.0, + "loss": 1.2288, + "step": 19640 + }, + { + "epoch": 4.576152771308803, + "grad_norm": 5.762319087982178, + "learning_rate": 0.0, + "loss": 1.0638, + "step": 19650 + }, + { + "epoch": 4.578481602235677, + "grad_norm": 14.740281105041504, + "learning_rate": 0.0, + "loss": 1.151, + "step": 19660 + }, + { + "epoch": 4.5808104331625525, + "grad_norm": 11.056538581848145, + "learning_rate": 0.0, + "loss": 0.9192, + "step": 19670 + }, + { + "epoch": 4.583139264089427, + "grad_norm": 20.958890914916992, + "learning_rate": 0.0, + "loss": 1.0669, + "step": 19680 + }, + { + "epoch": 4.585468095016302, + "grad_norm": 29.393579483032227, + "learning_rate": 0.0, + "loss": 1.5113, + "step": 19690 + }, + { + "epoch": 4.587796925943176, + "grad_norm": 45.95991134643555, + "learning_rate": 0.0, + "loss": 1.4022, + "step": 19700 + }, + { + "epoch": 4.5901257568700515, + "grad_norm": 6.977542400360107, + "learning_rate": 0.0, + "loss": 0.9212, + "step": 19710 + }, + { + "epoch": 4.592454587796926, + "grad_norm": 7.12802791595459, + "learning_rate": 0.0, + "loss": 1.0174, + "step": 19720 + }, + { + "epoch": 4.594783418723801, + "grad_norm": 12.21475601196289, + "learning_rate": 0.0, + "loss": 1.7109, + "step": 19730 + }, + { + "epoch": 4.597112249650675, + "grad_norm": 21.435813903808594, + "learning_rate": 0.0, + "loss": 1.817, + "step": 19740 + }, + { + "epoch": 4.5994410805775505, + "grad_norm": 7.1112446784973145, + "learning_rate": 0.0, + "loss": 1.0705, + "step": 19750 + }, + { + "epoch": 4.601769911504425, + "grad_norm": 3.9490299224853516, + "learning_rate": 0.0, + "loss": 1.2692, + "step": 19760 + }, + { + "epoch": 4.604098742431299, + "grad_norm": 50.142601013183594, + "learning_rate": 0.0, + "loss": 1.5913, + "step": 19770 + }, + { + "epoch": 4.606427573358174, + "grad_norm": 7.299134254455566, + "learning_rate": 0.0, + "loss": 1.096, + "step": 19780 + }, + { + "epoch": 4.608756404285049, + "grad_norm": 69.69446563720703, + "learning_rate": 0.0, + "loss": 1.1911, + "step": 19790 + }, + { + "epoch": 4.611085235211924, + "grad_norm": 6.633697509765625, + "learning_rate": 0.0, + "loss": 1.1309, + "step": 19800 + }, + { + "epoch": 4.613414066138798, + "grad_norm": 10.545255661010742, + "learning_rate": 0.0, + "loss": 1.0185, + "step": 19810 + }, + { + "epoch": 4.615742897065673, + "grad_norm": 14.515172004699707, + "learning_rate": 0.0, + "loss": 1.7769, + "step": 19820 + }, + { + "epoch": 4.618071727992548, + "grad_norm": 12.217021942138672, + "learning_rate": 0.0, + "loss": 1.0632, + "step": 19830 + }, + { + "epoch": 4.620400558919423, + "grad_norm": 29.141551971435547, + "learning_rate": 0.0, + "loss": 1.2936, + "step": 19840 + }, + { + "epoch": 4.622729389846297, + "grad_norm": 18.290485382080078, + "learning_rate": 0.0, + "loss": 1.0439, + "step": 19850 + }, + { + "epoch": 4.625058220773171, + "grad_norm": 9.301356315612793, + "learning_rate": 0.0, + "loss": 1.2435, + "step": 19860 + }, + { + "epoch": 4.627387051700047, + "grad_norm": 8.590415954589844, + "learning_rate": 0.0, + "loss": 0.8793, + "step": 19870 + }, + { + "epoch": 4.629715882626921, + "grad_norm": 15.62922477722168, + "learning_rate": 0.0, + "loss": 1.1785, + "step": 19880 + }, + { + "epoch": 4.632044713553796, + "grad_norm": 6.69285249710083, + "learning_rate": 0.0, + "loss": 0.9444, + "step": 19890 + }, + { + "epoch": 4.63437354448067, + "grad_norm": 62.42060852050781, + "learning_rate": 0.0, + "loss": 1.3757, + "step": 19900 + }, + { + "epoch": 4.636702375407546, + "grad_norm": 13.940834045410156, + "learning_rate": 0.0, + "loss": 1.0621, + "step": 19910 + }, + { + "epoch": 4.63903120633442, + "grad_norm": 12.126023292541504, + "learning_rate": 0.0, + "loss": 1.0629, + "step": 19920 + }, + { + "epoch": 4.641360037261295, + "grad_norm": 19.56361198425293, + "learning_rate": 0.0, + "loss": 0.8859, + "step": 19930 + }, + { + "epoch": 4.643688868188169, + "grad_norm": 13.945938110351562, + "learning_rate": 0.0, + "loss": 1.4923, + "step": 19940 + }, + { + "epoch": 4.646017699115045, + "grad_norm": 6.583313941955566, + "learning_rate": 0.0, + "loss": 1.1695, + "step": 19950 + }, + { + "epoch": 4.648346530041919, + "grad_norm": 11.310225486755371, + "learning_rate": 0.0, + "loss": 1.7277, + "step": 19960 + }, + { + "epoch": 4.650675360968794, + "grad_norm": 104.10961151123047, + "learning_rate": 0.0, + "loss": 1.4273, + "step": 19970 + }, + { + "epoch": 4.653004191895668, + "grad_norm": 19.168676376342773, + "learning_rate": 0.0, + "loss": 1.5528, + "step": 19980 + }, + { + "epoch": 4.655333022822543, + "grad_norm": 6.544228553771973, + "learning_rate": 0.0, + "loss": 1.234, + "step": 19990 + }, + { + "epoch": 4.657661853749418, + "grad_norm": 5.276122093200684, + "learning_rate": 0.0, + "loss": 1.2581, + "step": 20000 + }, + { + "epoch": 4.659990684676292, + "grad_norm": 94.88931274414062, + "learning_rate": 0.0, + "loss": 1.3947, + "step": 20010 + }, + { + "epoch": 4.662319515603167, + "grad_norm": 8.365758895874023, + "learning_rate": 0.0, + "loss": 1.5685, + "step": 20020 + }, + { + "epoch": 4.664648346530042, + "grad_norm": 23.34221839904785, + "learning_rate": 0.0, + "loss": 1.0813, + "step": 20030 + }, + { + "epoch": 4.666977177456917, + "grad_norm": 19.669198989868164, + "learning_rate": 0.0, + "loss": 1.0667, + "step": 20040 + }, + { + "epoch": 4.669306008383791, + "grad_norm": 11.661108016967773, + "learning_rate": 0.0, + "loss": 1.1768, + "step": 20050 + }, + { + "epoch": 4.671634839310666, + "grad_norm": 10.606729507446289, + "learning_rate": 0.0, + "loss": 0.9829, + "step": 20060 + }, + { + "epoch": 4.673963670237541, + "grad_norm": 48.090267181396484, + "learning_rate": 0.0, + "loss": 1.153, + "step": 20070 + }, + { + "epoch": 4.676292501164415, + "grad_norm": 5.69580602645874, + "learning_rate": 0.0, + "loss": 1.1018, + "step": 20080 + }, + { + "epoch": 4.67862133209129, + "grad_norm": 11.6792631149292, + "learning_rate": 0.0, + "loss": 1.0087, + "step": 20090 + }, + { + "epoch": 4.680950163018165, + "grad_norm": 10.280180931091309, + "learning_rate": 0.0, + "loss": 1.0452, + "step": 20100 + }, + { + "epoch": 4.68327899394504, + "grad_norm": 25.463672637939453, + "learning_rate": 0.0, + "loss": 1.6864, + "step": 20110 + }, + { + "epoch": 4.685607824871914, + "grad_norm": 25.444168090820312, + "learning_rate": 0.0, + "loss": 1.4118, + "step": 20120 + }, + { + "epoch": 4.687936655798789, + "grad_norm": 3.998115062713623, + "learning_rate": 0.0, + "loss": 1.1572, + "step": 20130 + }, + { + "epoch": 4.6902654867256635, + "grad_norm": 54.24504852294922, + "learning_rate": 0.0, + "loss": 1.1301, + "step": 20140 + }, + { + "epoch": 4.692594317652539, + "grad_norm": 8.593291282653809, + "learning_rate": 0.0, + "loss": 1.0843, + "step": 20150 + }, + { + "epoch": 4.694923148579413, + "grad_norm": 20.388153076171875, + "learning_rate": 0.0, + "loss": 1.1582, + "step": 20160 + }, + { + "epoch": 4.697251979506288, + "grad_norm": 7.956397533416748, + "learning_rate": 0.0, + "loss": 1.1456, + "step": 20170 + }, + { + "epoch": 4.6995808104331624, + "grad_norm": 13.123111724853516, + "learning_rate": 0.0, + "loss": 1.4565, + "step": 20180 + }, + { + "epoch": 4.701909641360038, + "grad_norm": 4.361729145050049, + "learning_rate": 0.0, + "loss": 0.9043, + "step": 20190 + }, + { + "epoch": 4.704238472286912, + "grad_norm": 16.26031494140625, + "learning_rate": 0.0, + "loss": 1.3562, + "step": 20200 + }, + { + "epoch": 4.706567303213786, + "grad_norm": 39.821197509765625, + "learning_rate": 0.0, + "loss": 1.3656, + "step": 20210 + }, + { + "epoch": 4.708896134140661, + "grad_norm": 5.075674057006836, + "learning_rate": 0.0, + "loss": 1.0091, + "step": 20220 + }, + { + "epoch": 4.711224965067536, + "grad_norm": 10.941627502441406, + "learning_rate": 0.0, + "loss": 1.6002, + "step": 20230 + }, + { + "epoch": 4.713553795994411, + "grad_norm": 15.39237117767334, + "learning_rate": 0.0, + "loss": 0.986, + "step": 20240 + }, + { + "epoch": 4.715882626921285, + "grad_norm": 33.28059005737305, + "learning_rate": 0.0, + "loss": 1.1865, + "step": 20250 + }, + { + "epoch": 4.71821145784816, + "grad_norm": 20.488203048706055, + "learning_rate": 0.0, + "loss": 1.221, + "step": 20260 + }, + { + "epoch": 4.720540288775035, + "grad_norm": 31.694875717163086, + "learning_rate": 0.0, + "loss": 1.441, + "step": 20270 + }, + { + "epoch": 4.72286911970191, + "grad_norm": 11.912035942077637, + "learning_rate": 0.0, + "loss": 1.2058, + "step": 20280 + }, + { + "epoch": 4.725197950628784, + "grad_norm": 95.51344299316406, + "learning_rate": 0.0, + "loss": 1.3537, + "step": 20290 + }, + { + "epoch": 4.727526781555659, + "grad_norm": 7.430610656738281, + "learning_rate": 0.0, + "loss": 1.6614, + "step": 20300 + }, + { + "epoch": 4.729855612482534, + "grad_norm": 16.465394973754883, + "learning_rate": 0.0, + "loss": 1.3616, + "step": 20310 + }, + { + "epoch": 4.732184443409409, + "grad_norm": 8.750808715820312, + "learning_rate": 0.0, + "loss": 1.4499, + "step": 20320 + }, + { + "epoch": 4.734513274336283, + "grad_norm": 15.627165794372559, + "learning_rate": 0.0, + "loss": 1.854, + "step": 20330 + }, + { + "epoch": 4.7368421052631575, + "grad_norm": 10.590481758117676, + "learning_rate": 0.0, + "loss": 1.2535, + "step": 20340 + }, + { + "epoch": 4.739170936190033, + "grad_norm": 11.492149353027344, + "learning_rate": 0.0, + "loss": 1.2967, + "step": 20350 + }, + { + "epoch": 4.741499767116907, + "grad_norm": 21.636804580688477, + "learning_rate": 0.0, + "loss": 1.6113, + "step": 20360 + }, + { + "epoch": 4.743828598043782, + "grad_norm": 5.577965259552002, + "learning_rate": 0.0, + "loss": 0.8558, + "step": 20370 + }, + { + "epoch": 4.7461574289706565, + "grad_norm": 7.697096824645996, + "learning_rate": 0.0, + "loss": 1.211, + "step": 20380 + }, + { + "epoch": 4.748486259897532, + "grad_norm": 35.868289947509766, + "learning_rate": 0.0, + "loss": 0.9857, + "step": 20390 + }, + { + "epoch": 4.750815090824406, + "grad_norm": 5.876249313354492, + "learning_rate": 0.0, + "loss": 0.9203, + "step": 20400 + }, + { + "epoch": 4.753143921751281, + "grad_norm": 12.119768142700195, + "learning_rate": 0.0, + "loss": 1.9162, + "step": 20410 + }, + { + "epoch": 4.7554727526781555, + "grad_norm": 10.388832092285156, + "learning_rate": 0.0, + "loss": 1.0832, + "step": 20420 + }, + { + "epoch": 4.75780158360503, + "grad_norm": 10.35787296295166, + "learning_rate": 0.0, + "loss": 1.0746, + "step": 20430 + }, + { + "epoch": 4.760130414531905, + "grad_norm": 8.285577774047852, + "learning_rate": 0.0, + "loss": 0.9798, + "step": 20440 + }, + { + "epoch": 4.76245924545878, + "grad_norm": 8.039005279541016, + "learning_rate": 0.0, + "loss": 1.0026, + "step": 20450 + }, + { + "epoch": 4.7647880763856545, + "grad_norm": 14.241586685180664, + "learning_rate": 0.0, + "loss": 1.4068, + "step": 20460 + }, + { + "epoch": 4.767116907312529, + "grad_norm": 14.098332405090332, + "learning_rate": 0.0, + "loss": 2.0716, + "step": 20470 + }, + { + "epoch": 4.769445738239404, + "grad_norm": 13.193544387817383, + "learning_rate": 0.0, + "loss": 1.0225, + "step": 20480 + }, + { + "epoch": 4.771774569166278, + "grad_norm": 12.964882850646973, + "learning_rate": 0.0, + "loss": 0.9621, + "step": 20490 + }, + { + "epoch": 4.7741034000931535, + "grad_norm": 66.96244049072266, + "learning_rate": 0.0, + "loss": 1.0828, + "step": 20500 + }, + { + "epoch": 4.776432231020028, + "grad_norm": 21.851341247558594, + "learning_rate": 0.0, + "loss": 1.2226, + "step": 20510 + }, + { + "epoch": 4.778761061946903, + "grad_norm": 11.199970245361328, + "learning_rate": 0.0, + "loss": 0.9426, + "step": 20520 + }, + { + "epoch": 4.781089892873777, + "grad_norm": 40.490753173828125, + "learning_rate": 0.0, + "loss": 1.5914, + "step": 20530 + }, + { + "epoch": 4.7834187238006525, + "grad_norm": 17.287166595458984, + "learning_rate": 0.0, + "loss": 1.4681, + "step": 20540 + }, + { + "epoch": 4.785747554727527, + "grad_norm": 13.948269844055176, + "learning_rate": 0.0, + "loss": 1.2498, + "step": 20550 + }, + { + "epoch": 4.788076385654401, + "grad_norm": 57.58226776123047, + "learning_rate": 0.0, + "loss": 1.7445, + "step": 20560 + }, + { + "epoch": 4.790405216581276, + "grad_norm": 5.487739086151123, + "learning_rate": 0.0, + "loss": 0.898, + "step": 20570 + }, + { + "epoch": 4.792734047508151, + "grad_norm": 15.564270973205566, + "learning_rate": 0.0, + "loss": 1.1383, + "step": 20580 + }, + { + "epoch": 4.795062878435026, + "grad_norm": 7.663552761077881, + "learning_rate": 0.0, + "loss": 1.2161, + "step": 20590 + }, + { + "epoch": 4.7973917093619, + "grad_norm": 6.975662708282471, + "learning_rate": 0.0, + "loss": 1.0381, + "step": 20600 + }, + { + "epoch": 4.799720540288775, + "grad_norm": 68.53016662597656, + "learning_rate": 0.0, + "loss": 1.3573, + "step": 20610 + }, + { + "epoch": 4.80204937121565, + "grad_norm": 51.50092315673828, + "learning_rate": 0.0, + "loss": 1.2739, + "step": 20620 + }, + { + "epoch": 4.804378202142525, + "grad_norm": 41.12253189086914, + "learning_rate": 0.0, + "loss": 1.0314, + "step": 20630 + }, + { + "epoch": 4.806707033069399, + "grad_norm": 8.585794448852539, + "learning_rate": 0.0, + "loss": 1.1131, + "step": 20640 + }, + { + "epoch": 4.809035863996274, + "grad_norm": 15.281457901000977, + "learning_rate": 0.0, + "loss": 1.1092, + "step": 20650 + }, + { + "epoch": 4.8113646949231486, + "grad_norm": 9.958845138549805, + "learning_rate": 0.0, + "loss": 1.1242, + "step": 20660 + }, + { + "epoch": 4.813693525850024, + "grad_norm": 9.2029390335083, + "learning_rate": 0.0, + "loss": 0.8147, + "step": 20670 + }, + { + "epoch": 4.816022356776898, + "grad_norm": 8.422203063964844, + "learning_rate": 0.0, + "loss": 1.2394, + "step": 20680 + }, + { + "epoch": 4.818351187703772, + "grad_norm": 6.612034320831299, + "learning_rate": 0.0, + "loss": 0.9382, + "step": 20690 + }, + { + "epoch": 4.8206800186306475, + "grad_norm": 9.461136817932129, + "learning_rate": 0.0, + "loss": 0.8587, + "step": 20700 + }, + { + "epoch": 4.823008849557522, + "grad_norm": 10.280345916748047, + "learning_rate": 0.0, + "loss": 0.8139, + "step": 20710 + }, + { + "epoch": 4.825337680484397, + "grad_norm": 6.057394504547119, + "learning_rate": 0.0, + "loss": 1.129, + "step": 20720 + }, + { + "epoch": 4.827666511411271, + "grad_norm": 9.460865020751953, + "learning_rate": 0.0, + "loss": 0.8686, + "step": 20730 + }, + { + "epoch": 4.8299953423381465, + "grad_norm": 31.428016662597656, + "learning_rate": 0.0, + "loss": 1.4765, + "step": 20740 + }, + { + "epoch": 4.832324173265021, + "grad_norm": 6.0293426513671875, + "learning_rate": 0.0, + "loss": 1.4691, + "step": 20750 + }, + { + "epoch": 4.834653004191896, + "grad_norm": 9.366103172302246, + "learning_rate": 0.0, + "loss": 0.9476, + "step": 20760 + }, + { + "epoch": 4.83698183511877, + "grad_norm": 7.880916595458984, + "learning_rate": 0.0, + "loss": 1.6768, + "step": 20770 + }, + { + "epoch": 4.839310666045645, + "grad_norm": 5.079976558685303, + "learning_rate": 0.0, + "loss": 0.8185, + "step": 20780 + }, + { + "epoch": 4.84163949697252, + "grad_norm": 5.588425159454346, + "learning_rate": 0.0, + "loss": 0.8626, + "step": 20790 + }, + { + "epoch": 4.843968327899394, + "grad_norm": 2.570796489715576, + "learning_rate": 0.0, + "loss": 1.2061, + "step": 20800 + }, + { + "epoch": 4.846297158826269, + "grad_norm": 21.089445114135742, + "learning_rate": 0.0, + "loss": 1.4568, + "step": 20810 + }, + { + "epoch": 4.848625989753144, + "grad_norm": 8.1835355758667, + "learning_rate": 0.0, + "loss": 0.8973, + "step": 20820 + }, + { + "epoch": 4.850954820680019, + "grad_norm": 12.01059341430664, + "learning_rate": 0.0, + "loss": 0.9296, + "step": 20830 + }, + { + "epoch": 4.853283651606893, + "grad_norm": 25.412508010864258, + "learning_rate": 0.0, + "loss": 1.787, + "step": 20840 + }, + { + "epoch": 4.855612482533768, + "grad_norm": 6.831169605255127, + "learning_rate": 0.0, + "loss": 1.0895, + "step": 20850 + }, + { + "epoch": 4.857941313460643, + "grad_norm": 8.809972763061523, + "learning_rate": 0.0, + "loss": 1.2377, + "step": 20860 + }, + { + "epoch": 4.860270144387518, + "grad_norm": 6.069893836975098, + "learning_rate": 0.0, + "loss": 1.2144, + "step": 20870 + }, + { + "epoch": 4.862598975314392, + "grad_norm": 11.856980323791504, + "learning_rate": 0.0, + "loss": 1.1085, + "step": 20880 + }, + { + "epoch": 4.864927806241267, + "grad_norm": 41.11824035644531, + "learning_rate": 0.0, + "loss": 1.5651, + "step": 20890 + }, + { + "epoch": 4.867256637168142, + "grad_norm": 13.00899887084961, + "learning_rate": 0.0, + "loss": 1.0189, + "step": 20900 + }, + { + "epoch": 4.869585468095016, + "grad_norm": 11.95178508758545, + "learning_rate": 0.0, + "loss": 0.8974, + "step": 20910 + }, + { + "epoch": 4.871914299021891, + "grad_norm": 31.843204498291016, + "learning_rate": 0.0, + "loss": 1.059, + "step": 20920 + }, + { + "epoch": 4.874243129948765, + "grad_norm": 18.231882095336914, + "learning_rate": 0.0, + "loss": 1.4373, + "step": 20930 + }, + { + "epoch": 4.876571960875641, + "grad_norm": 14.467658042907715, + "learning_rate": 0.0, + "loss": 0.9063, + "step": 20940 + }, + { + "epoch": 4.878900791802515, + "grad_norm": 47.42462158203125, + "learning_rate": 0.0, + "loss": 1.6727, + "step": 20950 + }, + { + "epoch": 4.88122962272939, + "grad_norm": 12.333345413208008, + "learning_rate": 0.0, + "loss": 1.077, + "step": 20960 + }, + { + "epoch": 4.883558453656264, + "grad_norm": 25.662342071533203, + "learning_rate": 0.0, + "loss": 0.9945, + "step": 20970 + }, + { + "epoch": 4.88588728458314, + "grad_norm": 141.52272033691406, + "learning_rate": 0.0, + "loss": 0.9511, + "step": 20980 + }, + { + "epoch": 4.888216115510014, + "grad_norm": 68.67302703857422, + "learning_rate": 0.0, + "loss": 1.3785, + "step": 20990 + }, + { + "epoch": 4.890544946436888, + "grad_norm": 4.360086441040039, + "learning_rate": 0.0, + "loss": 0.931, + "step": 21000 + }, + { + "epoch": 4.892873777363763, + "grad_norm": 9.454261779785156, + "learning_rate": 0.0, + "loss": 1.0979, + "step": 21010 + }, + { + "epoch": 4.895202608290639, + "grad_norm": 15.307931900024414, + "learning_rate": 0.0, + "loss": 1.0663, + "step": 21020 + }, + { + "epoch": 4.897531439217513, + "grad_norm": 6.253856182098389, + "learning_rate": 0.0, + "loss": 1.1837, + "step": 21030 + }, + { + "epoch": 4.899860270144387, + "grad_norm": 10.178426742553711, + "learning_rate": 0.0, + "loss": 0.9836, + "step": 21040 + }, + { + "epoch": 4.902189101071262, + "grad_norm": 43.5223274230957, + "learning_rate": 0.0, + "loss": 0.9516, + "step": 21050 + }, + { + "epoch": 4.904517931998137, + "grad_norm": 63.62842559814453, + "learning_rate": 0.0, + "loss": 0.9483, + "step": 21060 + }, + { + "epoch": 4.906846762925012, + "grad_norm": 11.0003662109375, + "learning_rate": 0.0, + "loss": 0.6298, + "step": 21070 + }, + { + "epoch": 4.909175593851886, + "grad_norm": 45.8305778503418, + "learning_rate": 0.0, + "loss": 1.1087, + "step": 21080 + }, + { + "epoch": 4.911504424778761, + "grad_norm": 18.97077178955078, + "learning_rate": 0.0, + "loss": 1.3575, + "step": 21090 + }, + { + "epoch": 4.913833255705636, + "grad_norm": 12.255200386047363, + "learning_rate": 0.0, + "loss": 1.0349, + "step": 21100 + }, + { + "epoch": 4.916162086632511, + "grad_norm": 7.784099102020264, + "learning_rate": 0.0, + "loss": 1.0109, + "step": 21110 + }, + { + "epoch": 4.918490917559385, + "grad_norm": 10.180665016174316, + "learning_rate": 0.0, + "loss": 1.1474, + "step": 21120 + }, + { + "epoch": 4.9208197484862595, + "grad_norm": 8.419008255004883, + "learning_rate": 0.0, + "loss": 1.1899, + "step": 21130 + }, + { + "epoch": 4.923148579413135, + "grad_norm": 105.6771469116211, + "learning_rate": 0.0, + "loss": 1.0455, + "step": 21140 + }, + { + "epoch": 4.925477410340009, + "grad_norm": 8.955032348632812, + "learning_rate": 0.0, + "loss": 1.3199, + "step": 21150 + }, + { + "epoch": 4.927806241266884, + "grad_norm": 4.457620620727539, + "learning_rate": 0.0, + "loss": 1.0334, + "step": 21160 + }, + { + "epoch": 4.9301350721937585, + "grad_norm": 21.67259979248047, + "learning_rate": 0.0, + "loss": 1.1492, + "step": 21170 + }, + { + "epoch": 4.932463903120634, + "grad_norm": 19.54806900024414, + "learning_rate": 0.0, + "loss": 1.7704, + "step": 21180 + }, + { + "epoch": 4.934792734047508, + "grad_norm": 56.51215744018555, + "learning_rate": 0.0, + "loss": 1.379, + "step": 21190 + }, + { + "epoch": 4.937121564974383, + "grad_norm": 11.147592544555664, + "learning_rate": 0.0, + "loss": 1.6373, + "step": 21200 + }, + { + "epoch": 4.9394503959012575, + "grad_norm": 7.846542835235596, + "learning_rate": 0.0, + "loss": 1.4208, + "step": 21210 + }, + { + "epoch": 4.941779226828133, + "grad_norm": 22.17070960998535, + "learning_rate": 0.0, + "loss": 1.2205, + "step": 21220 + }, + { + "epoch": 4.944108057755007, + "grad_norm": 12.682791709899902, + "learning_rate": 0.0, + "loss": 0.9231, + "step": 21230 + }, + { + "epoch": 4.946436888681882, + "grad_norm": 10.368877410888672, + "learning_rate": 0.0, + "loss": 1.1445, + "step": 21240 + }, + { + "epoch": 4.948765719608756, + "grad_norm": 14.382050514221191, + "learning_rate": 0.0, + "loss": 2.1009, + "step": 21250 + }, + { + "epoch": 4.951094550535631, + "grad_norm": 12.668513298034668, + "learning_rate": 0.0, + "loss": 1.2432, + "step": 21260 + }, + { + "epoch": 4.953423381462506, + "grad_norm": 5.586184024810791, + "learning_rate": 0.0, + "loss": 1.1811, + "step": 21270 + }, + { + "epoch": 4.95575221238938, + "grad_norm": 7.907278060913086, + "learning_rate": 0.0, + "loss": 1.0929, + "step": 21280 + }, + { + "epoch": 4.958081043316255, + "grad_norm": 45.115882873535156, + "learning_rate": 0.0, + "loss": 1.2946, + "step": 21290 + }, + { + "epoch": 4.96040987424313, + "grad_norm": 59.10978317260742, + "learning_rate": 0.0, + "loss": 1.472, + "step": 21300 + }, + { + "epoch": 4.962738705170005, + "grad_norm": 14.696610450744629, + "learning_rate": 0.0, + "loss": 1.6996, + "step": 21310 + }, + { + "epoch": 4.965067536096879, + "grad_norm": 17.75439453125, + "learning_rate": 0.0, + "loss": 1.1951, + "step": 21320 + }, + { + "epoch": 4.967396367023754, + "grad_norm": 58.72871398925781, + "learning_rate": 0.0, + "loss": 1.2667, + "step": 21330 + }, + { + "epoch": 4.969725197950629, + "grad_norm": 19.66732406616211, + "learning_rate": 0.0, + "loss": 1.4438, + "step": 21340 + }, + { + "epoch": 4.972054028877503, + "grad_norm": 9.646601676940918, + "learning_rate": 0.0, + "loss": 1.081, + "step": 21350 + }, + { + "epoch": 4.974382859804378, + "grad_norm": 7.841832160949707, + "learning_rate": 0.0, + "loss": 0.8552, + "step": 21360 + }, + { + "epoch": 4.9767116907312525, + "grad_norm": 11.626233100891113, + "learning_rate": 0.0, + "loss": 1.0067, + "step": 21370 + }, + { + "epoch": 4.979040521658128, + "grad_norm": 39.45273971557617, + "learning_rate": 0.0, + "loss": 1.0806, + "step": 21380 + }, + { + "epoch": 4.981369352585002, + "grad_norm": 11.194560050964355, + "learning_rate": 0.0, + "loss": 0.908, + "step": 21390 + }, + { + "epoch": 4.983698183511877, + "grad_norm": 158.674072265625, + "learning_rate": 0.0, + "loss": 1.4795, + "step": 21400 + }, + { + "epoch": 4.9860270144387515, + "grad_norm": 32.56852722167969, + "learning_rate": 0.0, + "loss": 1.6169, + "step": 21410 + }, + { + "epoch": 4.988355845365627, + "grad_norm": 7.956556797027588, + "learning_rate": 0.0, + "loss": 1.3402, + "step": 21420 + }, + { + "epoch": 4.990684676292501, + "grad_norm": 29.174211502075195, + "learning_rate": 0.0, + "loss": 1.0919, + "step": 21430 + }, + { + "epoch": 4.993013507219376, + "grad_norm": 88.53599548339844, + "learning_rate": 0.0, + "loss": 1.2261, + "step": 21440 + }, + { + "epoch": 4.9953423381462505, + "grad_norm": 15.843962669372559, + "learning_rate": 0.0, + "loss": 1.2406, + "step": 21450 + }, + { + "epoch": 4.997671169073126, + "grad_norm": 38.2321662902832, + "learning_rate": 0.0, + "loss": 1.4553, + "step": 21460 + }, + { + "epoch": 5.0, + "grad_norm": 12.097282409667969, + "learning_rate": 0.0, + "loss": 1.2166, + "step": 21470 + }, + { + "epoch": 5.0, + "eval_loss": 0.49088114500045776, + "eval_runtime": 15.246, + "eval_samples_per_second": 285.715, + "eval_steps_per_second": 2.296, + "step": 21470 + } + ], + "logging_steps": 10, + "max_steps": 21470, + "num_input_tokens_seen": 0, + "num_train_epochs": 5, + "save_steps": 500, + "stateful_callbacks": { + "EarlyStoppingCallback": { + "args": { + "early_stopping_patience": 2, + "early_stopping_threshold": 0.0 + }, + "attributes": { + "early_stopping_patience_counter": 2 + } + }, + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": true + }, + "attributes": {} + } + }, + "total_flos": 0.0, + "train_batch_size": 16, + "trial_name": null, + "trial_params": null +}