diff --git "a/ProcessEval-7B/trainer_state.json" "b/ProcessEval-7B/trainer_state.json" deleted file mode 100644--- "a/ProcessEval-7B/trainer_state.json" +++ /dev/null @@ -1,6448 +0,0 @@ -{ - "best_global_step": null, - "best_metric": null, - "best_model_checkpoint": null, - "epoch": 3.0, - "eval_steps": 500, - "global_step": 9159, - "is_hyper_param_search": false, - "is_local_process_zero": true, - "is_world_process_zero": true, - "log_history": [ - { - "epoch": 0.003275466754012447, - "grad_norm": 3.4142942428588867, - "learning_rate": 9.999976175252312e-05, - "loss": 1.8511, - "step": 10 - }, - { - "epoch": 0.006550933508024894, - "grad_norm": 2.4573607444763184, - "learning_rate": 9.999893818391307e-05, - "loss": 0.6371, - "step": 20 - }, - { - "epoch": 0.00982640026203734, - "grad_norm": 1.5601880550384521, - "learning_rate": 9.99975263625302e-05, - "loss": 0.7403, - "step": 30 - }, - { - "epoch": 0.013101867016049788, - "grad_norm": 2.462369203567505, - "learning_rate": 9.999552630498499e-05, - "loss": 0.6431, - "step": 40 - }, - { - "epoch": 0.016377333770062234, - "grad_norm": 1.9137276411056519, - "learning_rate": 9.999293803480876e-05, - "loss": 0.526, - "step": 50 - }, - { - "epoch": 0.01965280052407468, - "grad_norm": 1.1909728050231934, - "learning_rate": 9.99897615824533e-05, - "loss": 0.5762, - "step": 60 - }, - { - "epoch": 0.02292826727808713, - "grad_norm": 0.7606654763221741, - "learning_rate": 9.998599698529052e-05, - "loss": 0.6087, - "step": 70 - }, - { - "epoch": 0.026203734032099576, - "grad_norm": 0.9083070158958435, - "learning_rate": 9.998164428761208e-05, - "loss": 0.5744, - "step": 80 - }, - { - "epoch": 0.02947920078611202, - "grad_norm": 0.992427408695221, - "learning_rate": 9.997670354062877e-05, - "loss": 0.6265, - "step": 90 - }, - { - "epoch": 0.03275466754012447, - "grad_norm": 1.5478897094726562, - "learning_rate": 9.997117480247e-05, - "loss": 0.5748, - "step": 100 - }, - { - "epoch": 0.036030134294136915, - "grad_norm": 1.7055308818817139, - "learning_rate": 9.996505813818308e-05, - "loss": 0.6299, - "step": 110 - }, - { - "epoch": 0.03930560104814936, - "grad_norm": 1.0064220428466797, - "learning_rate": 9.995835361973241e-05, - "loss": 0.5388, - "step": 120 - }, - { - "epoch": 0.04258106780216181, - "grad_norm": 1.6811317205429077, - "learning_rate": 9.995106132599869e-05, - "loss": 0.5092, - "step": 130 - }, - { - "epoch": 0.04585653455617426, - "grad_norm": 1.3790682554244995, - "learning_rate": 9.994318134277801e-05, - "loss": 0.6195, - "step": 140 - }, - { - "epoch": 0.049132001310186704, - "grad_norm": 0.6575820446014404, - "learning_rate": 9.993471376278074e-05, - "loss": 0.5284, - "step": 150 - }, - { - "epoch": 0.05240746806419915, - "grad_norm": 1.2911351919174194, - "learning_rate": 9.992565868563058e-05, - "loss": 0.6049, - "step": 160 - }, - { - "epoch": 0.05568293481821159, - "grad_norm": 1.5066120624542236, - "learning_rate": 9.991601621786326e-05, - "loss": 0.6721, - "step": 170 - }, - { - "epoch": 0.05895840157222404, - "grad_norm": 0.8952118754386902, - "learning_rate": 9.990578647292537e-05, - "loss": 0.5579, - "step": 180 - }, - { - "epoch": 0.06223386832623649, - "grad_norm": 0.8831691741943359, - "learning_rate": 9.989496957117297e-05, - "loss": 0.5754, - "step": 190 - }, - { - "epoch": 0.06550933508024893, - "grad_norm": 1.631256341934204, - "learning_rate": 9.988356563987021e-05, - "loss": 0.4969, - "step": 200 - }, - { - "epoch": 0.06878480183426139, - "grad_norm": 1.3874777555465698, - "learning_rate": 9.987157481318785e-05, - "loss": 0.5339, - "step": 210 - }, - { - "epoch": 0.07206026858827383, - "grad_norm": 1.3643158674240112, - "learning_rate": 9.985899723220158e-05, - "loss": 0.602, - "step": 220 - }, - { - "epoch": 0.07533573534228627, - "grad_norm": 1.5697523355484009, - "learning_rate": 9.984583304489052e-05, - "loss": 0.5897, - "step": 230 - }, - { - "epoch": 0.07861120209629872, - "grad_norm": 1.320738673210144, - "learning_rate": 9.983208240613533e-05, - "loss": 0.6451, - "step": 240 - }, - { - "epoch": 0.08188666885031116, - "grad_norm": 1.51417076587677, - "learning_rate": 9.981774547771649e-05, - "loss": 0.5811, - "step": 250 - }, - { - "epoch": 0.08516213560432362, - "grad_norm": 0.9555358290672302, - "learning_rate": 9.980282242831232e-05, - "loss": 0.522, - "step": 260 - }, - { - "epoch": 0.08843760235833606, - "grad_norm": 2.864476442337036, - "learning_rate": 9.978731343349708e-05, - "loss": 0.6185, - "step": 270 - }, - { - "epoch": 0.09171306911234851, - "grad_norm": 1.461835503578186, - "learning_rate": 9.97712186757388e-05, - "loss": 0.5321, - "step": 280 - }, - { - "epoch": 0.09498853586636095, - "grad_norm": 1.1198502779006958, - "learning_rate": 9.975453834439723e-05, - "loss": 0.6212, - "step": 290 - }, - { - "epoch": 0.09826400262037341, - "grad_norm": 1.1931099891662598, - "learning_rate": 9.973727263572157e-05, - "loss": 0.5555, - "step": 300 - }, - { - "epoch": 0.10153946937438585, - "grad_norm": 1.36172616481781, - "learning_rate": 9.971942175284815e-05, - "loss": 0.5917, - "step": 310 - }, - { - "epoch": 0.1048149361283983, - "grad_norm": 1.102199673652649, - "learning_rate": 9.970098590579805e-05, - "loss": 0.4745, - "step": 320 - }, - { - "epoch": 0.10809040288241074, - "grad_norm": 0.878726601600647, - "learning_rate": 9.968196531147464e-05, - "loss": 0.648, - "step": 330 - }, - { - "epoch": 0.11136586963642318, - "grad_norm": 1.9705549478530884, - "learning_rate": 9.966236019366104e-05, - "loss": 0.5658, - "step": 340 - }, - { - "epoch": 0.11464133639043564, - "grad_norm": 1.6200658082962036, - "learning_rate": 9.964217078301741e-05, - "loss": 0.6023, - "step": 350 - }, - { - "epoch": 0.11791680314444808, - "grad_norm": 1.1012881994247437, - "learning_rate": 9.962139731707835e-05, - "loss": 0.4407, - "step": 360 - }, - { - "epoch": 0.12119226989846053, - "grad_norm": 1.7477895021438599, - "learning_rate": 9.960004004025004e-05, - "loss": 0.4602, - "step": 370 - }, - { - "epoch": 0.12446773665247297, - "grad_norm": 1.9638822078704834, - "learning_rate": 9.957809920380733e-05, - "loss": 0.5259, - "step": 380 - }, - { - "epoch": 0.12774320340648543, - "grad_norm": 2.069183588027954, - "learning_rate": 9.955557506589091e-05, - "loss": 0.5051, - "step": 390 - }, - { - "epoch": 0.13101867016049787, - "grad_norm": 3.2607414722442627, - "learning_rate": 9.95324678915041e-05, - "loss": 0.5166, - "step": 400 - }, - { - "epoch": 0.1342941369145103, - "grad_norm": 1.1227550506591797, - "learning_rate": 9.950877795250986e-05, - "loss": 0.5482, - "step": 410 - }, - { - "epoch": 0.13756960366852278, - "grad_norm": 7.312385559082031, - "learning_rate": 9.948450552762758e-05, - "loss": 0.5393, - "step": 420 - }, - { - "epoch": 0.14084507042253522, - "grad_norm": 1.222689151763916, - "learning_rate": 9.945965090242972e-05, - "loss": 0.5394, - "step": 430 - }, - { - "epoch": 0.14412053717654766, - "grad_norm": 4.6304473876953125, - "learning_rate": 9.943421436933858e-05, - "loss": 0.5839, - "step": 440 - }, - { - "epoch": 0.1473960039305601, - "grad_norm": 2.781198740005493, - "learning_rate": 9.940819622762267e-05, - "loss": 0.5752, - "step": 450 - }, - { - "epoch": 0.15067147068457254, - "grad_norm": 1.6278905868530273, - "learning_rate": 9.938159678339343e-05, - "loss": 0.4471, - "step": 460 - }, - { - "epoch": 0.153946937438585, - "grad_norm": 0.5791640281677246, - "learning_rate": 9.935441634960141e-05, - "loss": 0.4551, - "step": 470 - }, - { - "epoch": 0.15722240419259745, - "grad_norm": 1.1955379247665405, - "learning_rate": 9.932665524603273e-05, - "loss": 0.5497, - "step": 480 - }, - { - "epoch": 0.1604978709466099, - "grad_norm": 1.944528579711914, - "learning_rate": 9.929831379930524e-05, - "loss": 0.4806, - "step": 490 - }, - { - "epoch": 0.16377333770062233, - "grad_norm": 1.2993934154510498, - "learning_rate": 9.926939234286471e-05, - "loss": 0.5159, - "step": 500 - }, - { - "epoch": 0.1670488044546348, - "grad_norm": 1.727567195892334, - "learning_rate": 9.923989121698091e-05, - "loss": 0.5308, - "step": 510 - }, - { - "epoch": 0.17032427120864724, - "grad_norm": 1.3683221340179443, - "learning_rate": 9.920981076874361e-05, - "loss": 0.543, - "step": 520 - }, - { - "epoch": 0.17359973796265968, - "grad_norm": 2.0204238891601562, - "learning_rate": 9.917915135205842e-05, - "loss": 0.468, - "step": 530 - }, - { - "epoch": 0.17687520471667212, - "grad_norm": 1.5883879661560059, - "learning_rate": 9.91479133276428e-05, - "loss": 0.4496, - "step": 540 - }, - { - "epoch": 0.18015067147068456, - "grad_norm": 1.849434494972229, - "learning_rate": 9.911609706302158e-05, - "loss": 0.5924, - "step": 550 - }, - { - "epoch": 0.18342613822469703, - "grad_norm": 1.3481714725494385, - "learning_rate": 9.90837029325229e-05, - "loss": 0.5828, - "step": 560 - }, - { - "epoch": 0.18670160497870947, - "grad_norm": 0.8173115253448486, - "learning_rate": 9.905073131727353e-05, - "loss": 0.4415, - "step": 570 - }, - { - "epoch": 0.1899770717327219, - "grad_norm": 1.248199462890625, - "learning_rate": 9.901718260519461e-05, - "loss": 0.4459, - "step": 580 - }, - { - "epoch": 0.19325253848673435, - "grad_norm": 2.4475784301757812, - "learning_rate": 9.8983057190997e-05, - "loss": 0.5182, - "step": 590 - }, - { - "epoch": 0.19652800524074682, - "grad_norm": 1.0108952522277832, - "learning_rate": 9.89483554761766e-05, - "loss": 0.4088, - "step": 600 - }, - { - "epoch": 0.19980347199475926, - "grad_norm": 2.5650863647460938, - "learning_rate": 9.891307786900967e-05, - "loss": 0.4711, - "step": 610 - }, - { - "epoch": 0.2030789387487717, - "grad_norm": 1.2226927280426025, - "learning_rate": 9.887722478454807e-05, - "loss": 0.4767, - "step": 620 - }, - { - "epoch": 0.20635440550278414, - "grad_norm": 0.4554494321346283, - "learning_rate": 9.884079664461423e-05, - "loss": 0.5237, - "step": 630 - }, - { - "epoch": 0.2096298722567966, - "grad_norm": 2.0132486820220947, - "learning_rate": 9.880379387779637e-05, - "loss": 0.5836, - "step": 640 - }, - { - "epoch": 0.21290533901080905, - "grad_norm": 2.6649820804595947, - "learning_rate": 9.876621691944333e-05, - "loss": 0.5418, - "step": 650 - }, - { - "epoch": 0.2161808057648215, - "grad_norm": 1.6513729095458984, - "learning_rate": 9.87280662116595e-05, - "loss": 0.444, - "step": 660 - }, - { - "epoch": 0.21945627251883393, - "grad_norm": 3.0045840740203857, - "learning_rate": 9.868934220329961e-05, - "loss": 0.49, - "step": 670 - }, - { - "epoch": 0.22273173927284637, - "grad_norm": 1.1606544256210327, - "learning_rate": 9.865004534996342e-05, - "loss": 0.4218, - "step": 680 - }, - { - "epoch": 0.22600720602685884, - "grad_norm": 3.1066412925720215, - "learning_rate": 9.861017611399046e-05, - "loss": 0.5331, - "step": 690 - }, - { - "epoch": 0.22928267278087128, - "grad_norm": 1.4908937215805054, - "learning_rate": 9.856973496445439e-05, - "loss": 0.541, - "step": 700 - }, - { - "epoch": 0.23255813953488372, - "grad_norm": 1.572663426399231, - "learning_rate": 9.852872237715774e-05, - "loss": 0.4687, - "step": 710 - }, - { - "epoch": 0.23583360628889616, - "grad_norm": 1.7062207460403442, - "learning_rate": 9.848713883462611e-05, - "loss": 0.5323, - "step": 720 - }, - { - "epoch": 0.23910907304290863, - "grad_norm": 3.0517234802246094, - "learning_rate": 9.844498482610254e-05, - "loss": 0.5227, - "step": 730 - }, - { - "epoch": 0.24238453979692107, - "grad_norm": 1.4975826740264893, - "learning_rate": 9.840226084754183e-05, - "loss": 0.5261, - "step": 740 - }, - { - "epoch": 0.2456600065509335, - "grad_norm": 1.4004149436950684, - "learning_rate": 9.835896740160465e-05, - "loss": 0.4678, - "step": 750 - }, - { - "epoch": 0.24893547330494595, - "grad_norm": 0.848301887512207, - "learning_rate": 9.83151049976516e-05, - "loss": 0.4945, - "step": 760 - }, - { - "epoch": 0.2522109400589584, - "grad_norm": 1.295591115951538, - "learning_rate": 9.827067415173723e-05, - "loss": 0.4544, - "step": 770 - }, - { - "epoch": 0.25548640681297086, - "grad_norm": 0.8377560377120972, - "learning_rate": 9.822567538660403e-05, - "loss": 0.4226, - "step": 780 - }, - { - "epoch": 0.25876187356698327, - "grad_norm": 1.4290851354599, - "learning_rate": 9.818010923167622e-05, - "loss": 0.4983, - "step": 790 - }, - { - "epoch": 0.26203734032099574, - "grad_norm": 0.8712912201881409, - "learning_rate": 9.813397622305349e-05, - "loss": 0.5005, - "step": 800 - }, - { - "epoch": 0.2653128070750082, - "grad_norm": 0.6377179026603699, - "learning_rate": 9.808727690350482e-05, - "loss": 0.52, - "step": 810 - }, - { - "epoch": 0.2685882738290206, - "grad_norm": 3.8083136081695557, - "learning_rate": 9.804001182246189e-05, - "loss": 0.4773, - "step": 820 - }, - { - "epoch": 0.2718637405830331, - "grad_norm": 4.870070457458496, - "learning_rate": 9.79921815360129e-05, - "loss": 0.5607, - "step": 830 - }, - { - "epoch": 0.27513920733704555, - "grad_norm": 2.2912447452545166, - "learning_rate": 9.794378660689569e-05, - "loss": 0.473, - "step": 840 - }, - { - "epoch": 0.27841467409105797, - "grad_norm": 2.352205514907837, - "learning_rate": 9.789482760449143e-05, - "loss": 0.4941, - "step": 850 - }, - { - "epoch": 0.28169014084507044, - "grad_norm": 1.8317527770996094, - "learning_rate": 9.784530510481772e-05, - "loss": 0.4432, - "step": 860 - }, - { - "epoch": 0.28496560759908285, - "grad_norm": 1.474391222000122, - "learning_rate": 9.779521969052187e-05, - "loss": 0.498, - "step": 870 - }, - { - "epoch": 0.2882410743530953, - "grad_norm": 2.2652223110198975, - "learning_rate": 9.774457195087409e-05, - "loss": 0.4967, - "step": 880 - }, - { - "epoch": 0.2915165411071078, - "grad_norm": 1.0434564352035522, - "learning_rate": 9.769336248176048e-05, - "loss": 0.3985, - "step": 890 - }, - { - "epoch": 0.2947920078611202, - "grad_norm": 3.0927841663360596, - "learning_rate": 9.76415918856761e-05, - "loss": 0.5589, - "step": 900 - }, - { - "epoch": 0.29806747461513267, - "grad_norm": 1.2934414148330688, - "learning_rate": 9.758926077171782e-05, - "loss": 0.416, - "step": 910 - }, - { - "epoch": 0.3013429413691451, - "grad_norm": 4.895192623138428, - "learning_rate": 9.753636975557715e-05, - "loss": 0.6021, - "step": 920 - }, - { - "epoch": 0.30461840812315755, - "grad_norm": 2.1728906631469727, - "learning_rate": 9.748291945953305e-05, - "loss": 0.4612, - "step": 930 - }, - { - "epoch": 0.30789387487717, - "grad_norm": 0.9599400162696838, - "learning_rate": 9.742891051244458e-05, - "loss": 0.4605, - "step": 940 - }, - { - "epoch": 0.31116934163118243, - "grad_norm": 1.9502580165863037, - "learning_rate": 9.737434354974351e-05, - "loss": 0.4943, - "step": 950 - }, - { - "epoch": 0.3144448083851949, - "grad_norm": 1.693698763847351, - "learning_rate": 9.731921921342677e-05, - "loss": 0.4324, - "step": 960 - }, - { - "epoch": 0.31772027513920736, - "grad_norm": 2.408432722091675, - "learning_rate": 9.726353815204905e-05, - "loss": 0.464, - "step": 970 - }, - { - "epoch": 0.3209957418932198, - "grad_norm": 1.7604925632476807, - "learning_rate": 9.720730102071501e-05, - "loss": 0.3936, - "step": 980 - }, - { - "epoch": 0.32427120864723225, - "grad_norm": 1.5211386680603027, - "learning_rate": 9.715050848107168e-05, - "loss": 0.4656, - "step": 990 - }, - { - "epoch": 0.32754667540124466, - "grad_norm": 0.7129203677177429, - "learning_rate": 9.709316120130063e-05, - "loss": 0.4339, - "step": 1000 - }, - { - "epoch": 0.3308221421552571, - "grad_norm": 1.4347723722457886, - "learning_rate": 9.703525985611011e-05, - "loss": 0.4857, - "step": 1010 - }, - { - "epoch": 0.3340976089092696, - "grad_norm": 0.7748245596885681, - "learning_rate": 9.697680512672712e-05, - "loss": 0.481, - "step": 1020 - }, - { - "epoch": 0.337373075663282, - "grad_norm": 0.8669755458831787, - "learning_rate": 9.691779770088941e-05, - "loss": 0.5214, - "step": 1030 - }, - { - "epoch": 0.3406485424172945, - "grad_norm": 2.17287015914917, - "learning_rate": 9.685823827283736e-05, - "loss": 0.4724, - "step": 1040 - }, - { - "epoch": 0.3439240091713069, - "grad_norm": 1.0931299924850464, - "learning_rate": 9.679812754330582e-05, - "loss": 0.585, - "step": 1050 - }, - { - "epoch": 0.34719947592531936, - "grad_norm": 4.098705768585205, - "learning_rate": 9.673746621951587e-05, - "loss": 0.3978, - "step": 1060 - }, - { - "epoch": 0.3504749426793318, - "grad_norm": 1.6805942058563232, - "learning_rate": 9.667625501516649e-05, - "loss": 0.4416, - "step": 1070 - }, - { - "epoch": 0.35375040943334424, - "grad_norm": 1.0550005435943604, - "learning_rate": 9.661449465042623e-05, - "loss": 0.4995, - "step": 1080 - }, - { - "epoch": 0.3570258761873567, - "grad_norm": 1.0647046566009521, - "learning_rate": 9.655218585192461e-05, - "loss": 0.4714, - "step": 1090 - }, - { - "epoch": 0.3603013429413691, - "grad_norm": 3.1645591259002686, - "learning_rate": 9.648932935274368e-05, - "loss": 0.4915, - "step": 1100 - }, - { - "epoch": 0.3635768096953816, - "grad_norm": 0.9103413820266724, - "learning_rate": 9.642592589240932e-05, - "loss": 0.4954, - "step": 1110 - }, - { - "epoch": 0.36685227644939405, - "grad_norm": 1.2749395370483398, - "learning_rate": 9.636197621688262e-05, - "loss": 0.4578, - "step": 1120 - }, - { - "epoch": 0.37012774320340647, - "grad_norm": 2.419579267501831, - "learning_rate": 9.629748107855104e-05, - "loss": 0.4042, - "step": 1130 - }, - { - "epoch": 0.37340320995741894, - "grad_norm": 0.9015341997146606, - "learning_rate": 9.623244123621955e-05, - "loss": 0.4193, - "step": 1140 - }, - { - "epoch": 0.3766786767114314, - "grad_norm": 1.7380903959274292, - "learning_rate": 9.616685745510177e-05, - "loss": 0.5479, - "step": 1150 - }, - { - "epoch": 0.3799541434654438, - "grad_norm": 2.135024070739746, - "learning_rate": 9.61007305068109e-05, - "loss": 0.5151, - "step": 1160 - }, - { - "epoch": 0.3832296102194563, - "grad_norm": 7.494693279266357, - "learning_rate": 9.60340611693507e-05, - "loss": 0.4688, - "step": 1170 - }, - { - "epoch": 0.3865050769734687, - "grad_norm": 1.971453070640564, - "learning_rate": 9.596685022710624e-05, - "loss": 0.4777, - "step": 1180 - }, - { - "epoch": 0.38978054372748117, - "grad_norm": 1.5618499517440796, - "learning_rate": 9.58990984708348e-05, - "loss": 0.3539, - "step": 1190 - }, - { - "epoch": 0.39305601048149363, - "grad_norm": 1.495965838432312, - "learning_rate": 9.583080669765648e-05, - "loss": 0.4653, - "step": 1200 - }, - { - "epoch": 0.39633147723550605, - "grad_norm": 2.40224027633667, - "learning_rate": 9.576197571104484e-05, - "loss": 0.4954, - "step": 1210 - }, - { - "epoch": 0.3996069439895185, - "grad_norm": 2.026282787322998, - "learning_rate": 9.569260632081744e-05, - "loss": 0.4066, - "step": 1220 - }, - { - "epoch": 0.4028824107435309, - "grad_norm": 0.7941470742225647, - "learning_rate": 9.562269934312634e-05, - "loss": 0.5392, - "step": 1230 - }, - { - "epoch": 0.4061578774975434, - "grad_norm": 3.557481050491333, - "learning_rate": 9.555225560044848e-05, - "loss": 0.4638, - "step": 1240 - }, - { - "epoch": 0.40943334425155586, - "grad_norm": 1.491097092628479, - "learning_rate": 9.548127592157598e-05, - "loss": 0.556, - "step": 1250 - }, - { - "epoch": 0.4127088110055683, - "grad_norm": 2.052738666534424, - "learning_rate": 9.540976114160647e-05, - "loss": 0.4212, - "step": 1260 - }, - { - "epoch": 0.41598427775958075, - "grad_norm": 2.0442936420440674, - "learning_rate": 9.533771210193315e-05, - "loss": 0.4298, - "step": 1270 - }, - { - "epoch": 0.4192597445135932, - "grad_norm": 3.3717715740203857, - "learning_rate": 9.526512965023497e-05, - "loss": 0.4539, - "step": 1280 - }, - { - "epoch": 0.4225352112676056, - "grad_norm": 2.036747932434082, - "learning_rate": 9.519201464046663e-05, - "loss": 0.3645, - "step": 1290 - }, - { - "epoch": 0.4258106780216181, - "grad_norm": 1.3429933786392212, - "learning_rate": 9.511836793284854e-05, - "loss": 0.4495, - "step": 1300 - }, - { - "epoch": 0.4290861447756305, - "grad_norm": 1.3036234378814697, - "learning_rate": 9.504419039385672e-05, - "loss": 0.4682, - "step": 1310 - }, - { - "epoch": 0.432361611529643, - "grad_norm": 2.7568111419677734, - "learning_rate": 9.496948289621255e-05, - "loss": 0.4913, - "step": 1320 - }, - { - "epoch": 0.43563707828365544, - "grad_norm": 1.210801362991333, - "learning_rate": 9.489424631887254e-05, - "loss": 0.5374, - "step": 1330 - }, - { - "epoch": 0.43891254503766786, - "grad_norm": 5.601550102233887, - "learning_rate": 9.4818481547018e-05, - "loss": 0.4597, - "step": 1340 - }, - { - "epoch": 0.4421880117916803, - "grad_norm": 15.908878326416016, - "learning_rate": 9.474218947204459e-05, - "loss": 0.5277, - "step": 1350 - }, - { - "epoch": 0.44546347854569274, - "grad_norm": 2.644505739212036, - "learning_rate": 9.466537099155191e-05, - "loss": 0.6062, - "step": 1360 - }, - { - "epoch": 0.4487389452997052, - "grad_norm": 1.5875250101089478, - "learning_rate": 9.458802700933278e-05, - "loss": 0.426, - "step": 1370 - }, - { - "epoch": 0.4520144120537177, - "grad_norm": 1.2542312145233154, - "learning_rate": 9.45101584353628e-05, - "loss": 0.4158, - "step": 1380 - }, - { - "epoch": 0.4552898788077301, - "grad_norm": 0.7012236714363098, - "learning_rate": 9.44317661857895e-05, - "loss": 0.4692, - "step": 1390 - }, - { - "epoch": 0.45856534556174255, - "grad_norm": 1.17298424243927, - "learning_rate": 9.435285118292163e-05, - "loss": 0.5453, - "step": 1400 - }, - { - "epoch": 0.461840812315755, - "grad_norm": 1.2099233865737915, - "learning_rate": 9.427341435521832e-05, - "loss": 0.4823, - "step": 1410 - }, - { - "epoch": 0.46511627906976744, - "grad_norm": 0.9596761465072632, - "learning_rate": 9.419345663727805e-05, - "loss": 0.4981, - "step": 1420 - }, - { - "epoch": 0.4683917458237799, - "grad_norm": 1.2681078910827637, - "learning_rate": 9.411297896982783e-05, - "loss": 0.441, - "step": 1430 - }, - { - "epoch": 0.4716672125777923, - "grad_norm": 2.0389151573181152, - "learning_rate": 9.403198229971197e-05, - "loss": 0.3656, - "step": 1440 - }, - { - "epoch": 0.4749426793318048, - "grad_norm": 3.909153699874878, - "learning_rate": 9.395046757988102e-05, - "loss": 0.4365, - "step": 1450 - }, - { - "epoch": 0.47821814608581725, - "grad_norm": 4.062755107879639, - "learning_rate": 9.386843576938056e-05, - "loss": 0.408, - "step": 1460 - }, - { - "epoch": 0.48149361283982967, - "grad_norm": 15.906514167785645, - "learning_rate": 9.378588783333986e-05, - "loss": 0.4516, - "step": 1470 - }, - { - "epoch": 0.48476907959384213, - "grad_norm": 1.8633944988250732, - "learning_rate": 9.37028247429606e-05, - "loss": 0.4167, - "step": 1480 - }, - { - "epoch": 0.48804454634785455, - "grad_norm": 0.8692916035652161, - "learning_rate": 9.36192474755054e-05, - "loss": 0.4458, - "step": 1490 - }, - { - "epoch": 0.491320013101867, - "grad_norm": 0.41129928827285767, - "learning_rate": 9.353515701428636e-05, - "loss": 0.3674, - "step": 1500 - }, - { - "epoch": 0.4945954798558795, - "grad_norm": 1.2853093147277832, - "learning_rate": 9.345055434865337e-05, - "loss": 0.4799, - "step": 1510 - }, - { - "epoch": 0.4978709466098919, - "grad_norm": 0.7302932143211365, - "learning_rate": 9.336544047398267e-05, - "loss": 0.4423, - "step": 1520 - }, - { - "epoch": 0.5011464133639043, - "grad_norm": 2.688443183898926, - "learning_rate": 9.327981639166496e-05, - "loss": 0.5434, - "step": 1530 - }, - { - "epoch": 0.5044218801179168, - "grad_norm": 1.1666405200958252, - "learning_rate": 9.31936831090937e-05, - "loss": 0.485, - "step": 1540 - }, - { - "epoch": 0.5076973468719292, - "grad_norm": 1.3045744895935059, - "learning_rate": 9.310704163965329e-05, - "loss": 0.4595, - "step": 1550 - }, - { - "epoch": 0.5109728136259417, - "grad_norm": 0.98723304271698, - "learning_rate": 9.301989300270703e-05, - "loss": 0.4388, - "step": 1560 - }, - { - "epoch": 0.5142482803799542, - "grad_norm": 6.278781890869141, - "learning_rate": 9.293223822358528e-05, - "loss": 0.393, - "step": 1570 - }, - { - "epoch": 0.5175237471339665, - "grad_norm": 4.777818202972412, - "learning_rate": 9.284407833357325e-05, - "loss": 0.4042, - "step": 1580 - }, - { - "epoch": 0.520799213887979, - "grad_norm": 1.8941549062728882, - "learning_rate": 9.275541436989899e-05, - "loss": 0.51, - "step": 1590 - }, - { - "epoch": 0.5240746806419915, - "grad_norm": 1.0570429563522339, - "learning_rate": 9.266624737572105e-05, - "loss": 0.4604, - "step": 1600 - }, - { - "epoch": 0.5273501473960039, - "grad_norm": 1.3238725662231445, - "learning_rate": 9.257657840011639e-05, - "loss": 0.5101, - "step": 1610 - }, - { - "epoch": 0.5306256141500164, - "grad_norm": 2.228785514831543, - "learning_rate": 9.248640849806787e-05, - "loss": 0.404, - "step": 1620 - }, - { - "epoch": 0.5339010809040288, - "grad_norm": 0.998704731464386, - "learning_rate": 9.239573873045192e-05, - "loss": 0.4093, - "step": 1630 - }, - { - "epoch": 0.5371765476580412, - "grad_norm": 1.920929193496704, - "learning_rate": 9.230457016402602e-05, - "loss": 0.408, - "step": 1640 - }, - { - "epoch": 0.5404520144120537, - "grad_norm": 0.7706437706947327, - "learning_rate": 9.221290387141622e-05, - "loss": 0.3845, - "step": 1650 - }, - { - "epoch": 0.5437274811660662, - "grad_norm": 9.670165061950684, - "learning_rate": 9.212074093110443e-05, - "loss": 0.4646, - "step": 1660 - }, - { - "epoch": 0.5470029479200786, - "grad_norm": 15.304981231689453, - "learning_rate": 9.202808242741574e-05, - "loss": 0.3678, - "step": 1670 - }, - { - "epoch": 0.5502784146740911, - "grad_norm": 1.0304086208343506, - "learning_rate": 9.19349294505058e-05, - "loss": 0.4433, - "step": 1680 - }, - { - "epoch": 0.5535538814281035, - "grad_norm": 2.5104544162750244, - "learning_rate": 9.184128309634776e-05, - "loss": 0.4179, - "step": 1690 - }, - { - "epoch": 0.5568293481821159, - "grad_norm": 1.2100114822387695, - "learning_rate": 9.174714446671962e-05, - "loss": 0.4384, - "step": 1700 - }, - { - "epoch": 0.5601048149361284, - "grad_norm": 1.1395717859268188, - "learning_rate": 9.165251466919106e-05, - "loss": 0.4761, - "step": 1710 - }, - { - "epoch": 0.5633802816901409, - "grad_norm": 1.6196281909942627, - "learning_rate": 9.155739481711058e-05, - "loss": 0.405, - "step": 1720 - }, - { - "epoch": 0.5666557484441533, - "grad_norm": 7.360799312591553, - "learning_rate": 9.146178602959224e-05, - "loss": 0.4315, - "step": 1730 - }, - { - "epoch": 0.5699312151981657, - "grad_norm": 1.2498565912246704, - "learning_rate": 9.136568943150265e-05, - "loss": 0.4335, - "step": 1740 - }, - { - "epoch": 0.5732066819521782, - "grad_norm": 1.4700082540512085, - "learning_rate": 9.126910615344759e-05, - "loss": 0.3774, - "step": 1750 - }, - { - "epoch": 0.5764821487061906, - "grad_norm": 2.199226140975952, - "learning_rate": 9.117203733175885e-05, - "loss": 0.3897, - "step": 1760 - }, - { - "epoch": 0.5797576154602031, - "grad_norm": 2.5813138484954834, - "learning_rate": 9.107448410848072e-05, - "loss": 0.4585, - "step": 1770 - }, - { - "epoch": 0.5830330822142156, - "grad_norm": 1.2768886089324951, - "learning_rate": 9.097644763135668e-05, - "loss": 0.4311, - "step": 1780 - }, - { - "epoch": 0.5863085489682279, - "grad_norm": 1.5592445135116577, - "learning_rate": 9.087792905381582e-05, - "loss": 0.4594, - "step": 1790 - }, - { - "epoch": 0.5895840157222404, - "grad_norm": 1.469441294670105, - "learning_rate": 9.077892953495926e-05, - "loss": 0.4883, - "step": 1800 - }, - { - "epoch": 0.5928594824762529, - "grad_norm": 1.2295513153076172, - "learning_rate": 9.067945023954656e-05, - "loss": 0.5005, - "step": 1810 - }, - { - "epoch": 0.5961349492302653, - "grad_norm": 2.6771979331970215, - "learning_rate": 9.057949233798202e-05, - "loss": 0.3879, - "step": 1820 - }, - { - "epoch": 0.5994104159842778, - "grad_norm": 2.0139994621276855, - "learning_rate": 9.047905700630088e-05, - "loss": 0.4086, - "step": 1830 - }, - { - "epoch": 0.6026858827382902, - "grad_norm": 0.9915063381195068, - "learning_rate": 9.037814542615546e-05, - "loss": 0.4029, - "step": 1840 - }, - { - "epoch": 0.6059613494923026, - "grad_norm": 1.3278034925460815, - "learning_rate": 9.027675878480131e-05, - "loss": 0.4402, - "step": 1850 - }, - { - "epoch": 0.6092368162463151, - "grad_norm": 3.338838577270508, - "learning_rate": 9.017489827508324e-05, - "loss": 0.38, - "step": 1860 - }, - { - "epoch": 0.6125122830003276, - "grad_norm": 2.513947010040283, - "learning_rate": 9.007256509542122e-05, - "loss": 0.4596, - "step": 1870 - }, - { - "epoch": 0.61578774975434, - "grad_norm": 1.1986016035079956, - "learning_rate": 8.996976044979637e-05, - "loss": 0.5269, - "step": 1880 - }, - { - "epoch": 0.6190632165083524, - "grad_norm": 1.3276673555374146, - "learning_rate": 8.986648554773674e-05, - "loss": 0.411, - "step": 1890 - }, - { - "epoch": 0.6223386832623649, - "grad_norm": 1.90061616897583, - "learning_rate": 8.976274160430304e-05, - "loss": 0.4864, - "step": 1900 - }, - { - "epoch": 0.6256141500163773, - "grad_norm": 1.219609260559082, - "learning_rate": 8.96585298400745e-05, - "loss": 0.4681, - "step": 1910 - }, - { - "epoch": 0.6288896167703898, - "grad_norm": 2.5666162967681885, - "learning_rate": 8.95538514811343e-05, - "loss": 0.405, - "step": 1920 - }, - { - "epoch": 0.6321650835244023, - "grad_norm": 29.135700225830078, - "learning_rate": 8.94487077590553e-05, - "loss": 0.4304, - "step": 1930 - }, - { - "epoch": 0.6354405502784147, - "grad_norm": 1.3458505868911743, - "learning_rate": 8.93430999108855e-05, - "loss": 0.4839, - "step": 1940 - }, - { - "epoch": 0.6387160170324271, - "grad_norm": 1.5121238231658936, - "learning_rate": 8.923702917913347e-05, - "loss": 0.4523, - "step": 1950 - }, - { - "epoch": 0.6419914837864396, - "grad_norm": 3.1551737785339355, - "learning_rate": 8.913049681175376e-05, - "loss": 0.388, - "step": 1960 - }, - { - "epoch": 0.645266950540452, - "grad_norm": 2.16231369972229, - "learning_rate": 8.902350406213221e-05, - "loss": 0.3707, - "step": 1970 - }, - { - "epoch": 0.6485424172944645, - "grad_norm": 1.1517155170440674, - "learning_rate": 8.891605218907113e-05, - "loss": 0.4232, - "step": 1980 - }, - { - "epoch": 0.651817884048477, - "grad_norm": 1.3121541738510132, - "learning_rate": 8.880814245677467e-05, - "loss": 0.4357, - "step": 1990 - }, - { - "epoch": 0.6550933508024893, - "grad_norm": 1.7689076662063599, - "learning_rate": 8.869977613483376e-05, - "loss": 0.4096, - "step": 2000 - }, - { - "epoch": 0.6583688175565018, - "grad_norm": 1.3465323448181152, - "learning_rate": 8.859095449821125e-05, - "loss": 0.493, - "step": 2010 - }, - { - "epoch": 0.6616442843105143, - "grad_norm": 1.9615005254745483, - "learning_rate": 8.848167882722693e-05, - "loss": 0.3645, - "step": 2020 - }, - { - "epoch": 0.6649197510645267, - "grad_norm": 0.8777059316635132, - "learning_rate": 8.837195040754245e-05, - "loss": 0.3735, - "step": 2030 - }, - { - "epoch": 0.6681952178185392, - "grad_norm": 4.644092559814453, - "learning_rate": 8.826177053014613e-05, - "loss": 0.4066, - "step": 2040 - }, - { - "epoch": 0.6714706845725515, - "grad_norm": 1.8295934200286865, - "learning_rate": 8.815114049133789e-05, - "loss": 0.3826, - "step": 2050 - }, - { - "epoch": 0.674746151326564, - "grad_norm": 2.737123489379883, - "learning_rate": 8.80400615927139e-05, - "loss": 0.4229, - "step": 2060 - }, - { - "epoch": 0.6780216180805765, - "grad_norm": 5.906623363494873, - "learning_rate": 8.792853514115134e-05, - "loss": 0.4671, - "step": 2070 - }, - { - "epoch": 0.681297084834589, - "grad_norm": 1.8148258924484253, - "learning_rate": 8.781656244879294e-05, - "loss": 0.4263, - "step": 2080 - }, - { - "epoch": 0.6845725515886014, - "grad_norm": 1.9900586605072021, - "learning_rate": 8.770414483303161e-05, - "loss": 0.3764, - "step": 2090 - }, - { - "epoch": 0.6878480183426138, - "grad_norm": 1.9682644605636597, - "learning_rate": 8.759128361649494e-05, - "loss": 0.4375, - "step": 2100 - }, - { - "epoch": 0.6911234850966262, - "grad_norm": 0.8583366870880127, - "learning_rate": 8.747798012702958e-05, - "loss": 0.3773, - "step": 2110 - }, - { - "epoch": 0.6943989518506387, - "grad_norm": 1.6172552108764648, - "learning_rate": 8.736423569768569e-05, - "loss": 0.4727, - "step": 2120 - }, - { - "epoch": 0.6976744186046512, - "grad_norm": 1.6621836423873901, - "learning_rate": 8.725005166670121e-05, - "loss": 0.414, - "step": 2130 - }, - { - "epoch": 0.7009498853586636, - "grad_norm": 1.3063037395477295, - "learning_rate": 8.713542937748612e-05, - "loss": 0.3476, - "step": 2140 - }, - { - "epoch": 0.704225352112676, - "grad_norm": 1.1321628093719482, - "learning_rate": 8.702037017860664e-05, - "loss": 0.4435, - "step": 2150 - }, - { - "epoch": 0.7075008188666885, - "grad_norm": 1.2167377471923828, - "learning_rate": 8.69048754237694e-05, - "loss": 0.4048, - "step": 2160 - }, - { - "epoch": 0.7107762856207009, - "grad_norm": 1.407345175743103, - "learning_rate": 8.678894647180541e-05, - "loss": 0.4064, - "step": 2170 - }, - { - "epoch": 0.7140517523747134, - "grad_norm": 2.5372154712677, - "learning_rate": 8.667258468665421e-05, - "loss": 0.4558, - "step": 2180 - }, - { - "epoch": 0.7173272191287259, - "grad_norm": 2.392592191696167, - "learning_rate": 8.655579143734773e-05, - "loss": 0.3701, - "step": 2190 - }, - { - "epoch": 0.7206026858827382, - "grad_norm": 1.34644615650177, - "learning_rate": 8.643856809799423e-05, - "loss": 0.423, - "step": 2200 - }, - { - "epoch": 0.7238781526367507, - "grad_norm": 0.5739784240722656, - "learning_rate": 8.632091604776202e-05, - "loss": 0.4103, - "step": 2210 - }, - { - "epoch": 0.7271536193907632, - "grad_norm": 3.17356538772583, - "learning_rate": 8.620283667086344e-05, - "loss": 0.5094, - "step": 2220 - }, - { - "epoch": 0.7304290861447756, - "grad_norm": 1.4579784870147705, - "learning_rate": 8.60843313565384e-05, - "loss": 0.4635, - "step": 2230 - }, - { - "epoch": 0.7337045528987881, - "grad_norm": 1.4431722164154053, - "learning_rate": 8.596540149903811e-05, - "loss": 0.4565, - "step": 2240 - }, - { - "epoch": 0.7369800196528006, - "grad_norm": 1.7317287921905518, - "learning_rate": 8.584604849760864e-05, - "loss": 0.505, - "step": 2250 - }, - { - "epoch": 0.7402554864068129, - "grad_norm": 0.7636827230453491, - "learning_rate": 8.57262737564745e-05, - "loss": 0.3418, - "step": 2260 - }, - { - "epoch": 0.7435309531608254, - "grad_norm": 4.791762828826904, - "learning_rate": 8.560607868482209e-05, - "loss": 0.4165, - "step": 2270 - }, - { - "epoch": 0.7468064199148379, - "grad_norm": 2.5725719928741455, - "learning_rate": 8.548546469678311e-05, - "loss": 0.4321, - "step": 2280 - }, - { - "epoch": 0.7500818866688503, - "grad_norm": 0.5766289830207825, - "learning_rate": 8.536443321141798e-05, - "loss": 0.4277, - "step": 2290 - }, - { - "epoch": 0.7533573534228628, - "grad_norm": 3.6899561882019043, - "learning_rate": 8.524298565269905e-05, - "loss": 0.4285, - "step": 2300 - }, - { - "epoch": 0.7566328201768752, - "grad_norm": 2.9638586044311523, - "learning_rate": 8.512112344949394e-05, - "loss": 0.4716, - "step": 2310 - }, - { - "epoch": 0.7599082869308876, - "grad_norm": 0.5847499966621399, - "learning_rate": 8.499884803554869e-05, - "loss": 0.3412, - "step": 2320 - }, - { - "epoch": 0.7631837536849001, - "grad_norm": 4.327905178070068, - "learning_rate": 8.487616084947086e-05, - "loss": 0.4231, - "step": 2330 - }, - { - "epoch": 0.7664592204389126, - "grad_norm": 2.5782723426818848, - "learning_rate": 8.475306333471263e-05, - "loss": 0.4207, - "step": 2340 - }, - { - "epoch": 0.769734687192925, - "grad_norm": 1.9452351331710815, - "learning_rate": 8.462955693955387e-05, - "loss": 0.3676, - "step": 2350 - }, - { - "epoch": 0.7730101539469374, - "grad_norm": 2.954127311706543, - "learning_rate": 8.450564311708504e-05, - "loss": 0.386, - "step": 2360 - }, - { - "epoch": 0.7762856207009499, - "grad_norm": 0.7142738699913025, - "learning_rate": 8.438132332519007e-05, - "loss": 0.4124, - "step": 2370 - }, - { - "epoch": 0.7795610874549623, - "grad_norm": 3.895556688308716, - "learning_rate": 8.425659902652928e-05, - "loss": 0.5817, - "step": 2380 - }, - { - "epoch": 0.7828365542089748, - "grad_norm": 0.7302497625350952, - "learning_rate": 8.413147168852212e-05, - "loss": 0.4036, - "step": 2390 - }, - { - "epoch": 0.7861120209629873, - "grad_norm": 4.139225959777832, - "learning_rate": 8.40059427833299e-05, - "loss": 0.3323, - "step": 2400 - }, - { - "epoch": 0.7893874877169996, - "grad_norm": 2.2886745929718018, - "learning_rate": 8.388001378783856e-05, - "loss": 0.4528, - "step": 2410 - }, - { - "epoch": 0.7926629544710121, - "grad_norm": 2.433013916015625, - "learning_rate": 8.375368618364114e-05, - "loss": 0.4841, - "step": 2420 - }, - { - "epoch": 0.7959384212250246, - "grad_norm": 2.8676037788391113, - "learning_rate": 8.362696145702051e-05, - "loss": 0.4129, - "step": 2430 - }, - { - "epoch": 0.799213887979037, - "grad_norm": 1.6046780347824097, - "learning_rate": 8.349984109893172e-05, - "loss": 0.3975, - "step": 2440 - }, - { - "epoch": 0.8024893547330495, - "grad_norm": 0.9988681674003601, - "learning_rate": 8.337232660498464e-05, - "loss": 0.3658, - "step": 2450 - }, - { - "epoch": 0.8057648214870619, - "grad_norm": 0.7486966252326965, - "learning_rate": 8.32444194754262e-05, - "loss": 0.4633, - "step": 2460 - }, - { - "epoch": 0.8090402882410743, - "grad_norm": 3.339841604232788, - "learning_rate": 8.311612121512283e-05, - "loss": 0.4518, - "step": 2470 - }, - { - "epoch": 0.8123157549950868, - "grad_norm": 2.9408583641052246, - "learning_rate": 8.298743333354272e-05, - "loss": 0.4293, - "step": 2480 - }, - { - "epoch": 0.8155912217490993, - "grad_norm": 4.075345039367676, - "learning_rate": 8.285835734473812e-05, - "loss": 0.4726, - "step": 2490 - }, - { - "epoch": 0.8188666885031117, - "grad_norm": 2.325965166091919, - "learning_rate": 8.272889476732744e-05, - "loss": 0.3978, - "step": 2500 - }, - { - "epoch": 0.8221421552571241, - "grad_norm": 0.8976953625679016, - "learning_rate": 8.25990471244774e-05, - "loss": 0.4871, - "step": 2510 - }, - { - "epoch": 0.8254176220111366, - "grad_norm": 2.086416482925415, - "learning_rate": 8.24688159438852e-05, - "loss": 0.5713, - "step": 2520 - }, - { - "epoch": 0.828693088765149, - "grad_norm": 1.157715916633606, - "learning_rate": 8.23382027577604e-05, - "loss": 0.4374, - "step": 2530 - }, - { - "epoch": 0.8319685555191615, - "grad_norm": 3.571927070617676, - "learning_rate": 8.220720910280703e-05, - "loss": 0.4597, - "step": 2540 - }, - { - "epoch": 0.835244022273174, - "grad_norm": 0.8830175995826721, - "learning_rate": 8.207583652020544e-05, - "loss": 0.4197, - "step": 2550 - }, - { - "epoch": 0.8385194890271864, - "grad_norm": 2.35368275642395, - "learning_rate": 8.194408655559418e-05, - "loss": 0.4274, - "step": 2560 - }, - { - "epoch": 0.8417949557811988, - "grad_norm": 3.0161986351013184, - "learning_rate": 8.18119607590518e-05, - "loss": 0.4448, - "step": 2570 - }, - { - "epoch": 0.8450704225352113, - "grad_norm": 0.708756148815155, - "learning_rate": 8.167946068507861e-05, - "loss": 0.4395, - "step": 2580 - }, - { - "epoch": 0.8483458892892237, - "grad_norm": 3.0698132514953613, - "learning_rate": 8.154658789257847e-05, - "loss": 0.4468, - "step": 2590 - }, - { - "epoch": 0.8516213560432362, - "grad_norm": 1.8274637460708618, - "learning_rate": 8.141334394484034e-05, - "loss": 0.368, - "step": 2600 - }, - { - "epoch": 0.8548968227972487, - "grad_norm": 1.2253950834274292, - "learning_rate": 8.127973040951995e-05, - "loss": 0.3474, - "step": 2610 - }, - { - "epoch": 0.858172289551261, - "grad_norm": 1.3629297018051147, - "learning_rate": 8.114574885862132e-05, - "loss": 0.401, - "step": 2620 - }, - { - "epoch": 0.8614477563052735, - "grad_norm": 3.098328113555908, - "learning_rate": 8.101140086847832e-05, - "loss": 0.4856, - "step": 2630 - }, - { - "epoch": 0.864723223059286, - "grad_norm": 1.9880963563919067, - "learning_rate": 8.087668801973605e-05, - "loss": 0.4137, - "step": 2640 - }, - { - "epoch": 0.8679986898132984, - "grad_norm": 0.8097736239433289, - "learning_rate": 8.074161189733231e-05, - "loss": 0.3681, - "step": 2650 - }, - { - "epoch": 0.8712741565673109, - "grad_norm": 2.174891233444214, - "learning_rate": 8.060617409047895e-05, - "loss": 0.3719, - "step": 2660 - }, - { - "epoch": 0.8745496233213232, - "grad_norm": 2.155724287033081, - "learning_rate": 8.04703761926431e-05, - "loss": 0.4674, - "step": 2670 - }, - { - "epoch": 0.8778250900753357, - "grad_norm": 1.2070379257202148, - "learning_rate": 8.033421980152847e-05, - "loss": 0.378, - "step": 2680 - }, - { - "epoch": 0.8811005568293482, - "grad_norm": 1.7812188863754272, - "learning_rate": 8.019770651905658e-05, - "loss": 0.4538, - "step": 2690 - }, - { - "epoch": 0.8843760235833606, - "grad_norm": 5.747461795806885, - "learning_rate": 8.006083795134789e-05, - "loss": 0.4049, - "step": 2700 - }, - { - "epoch": 0.8876514903373731, - "grad_norm": 1.8701963424682617, - "learning_rate": 7.992361570870288e-05, - "loss": 0.4344, - "step": 2710 - }, - { - "epoch": 0.8909269570913855, - "grad_norm": 1.7139086723327637, - "learning_rate": 7.978604140558313e-05, - "loss": 0.4481, - "step": 2720 - }, - { - "epoch": 0.8942024238453979, - "grad_norm": 2.668175458908081, - "learning_rate": 7.96481166605923e-05, - "loss": 0.4235, - "step": 2730 - }, - { - "epoch": 0.8974778905994104, - "grad_norm": 1.0785696506500244, - "learning_rate": 7.950984309645715e-05, - "loss": 0.3499, - "step": 2740 - }, - { - "epoch": 0.9007533573534229, - "grad_norm": 2.0185582637786865, - "learning_rate": 7.937122234000838e-05, - "loss": 0.4065, - "step": 2750 - }, - { - "epoch": 0.9040288241074353, - "grad_norm": 3.0428175926208496, - "learning_rate": 7.923225602216148e-05, - "loss": 0.3546, - "step": 2760 - }, - { - "epoch": 0.9073042908614477, - "grad_norm": 2.705752372741699, - "learning_rate": 7.909294577789766e-05, - "loss": 0.3767, - "step": 2770 - }, - { - "epoch": 0.9105797576154602, - "grad_norm": 4.201162815093994, - "learning_rate": 7.895329324624441e-05, - "loss": 0.3485, - "step": 2780 - }, - { - "epoch": 0.9138552243694726, - "grad_norm": 1.7414577007293701, - "learning_rate": 7.881330007025645e-05, - "loss": 0.4744, - "step": 2790 - }, - { - "epoch": 0.9171306911234851, - "grad_norm": 2.0567240715026855, - "learning_rate": 7.867296789699623e-05, - "loss": 0.3688, - "step": 2800 - }, - { - "epoch": 0.9204061578774976, - "grad_norm": 1.6167949438095093, - "learning_rate": 7.85322983775146e-05, - "loss": 0.3214, - "step": 2810 - }, - { - "epoch": 0.92368162463151, - "grad_norm": 2.3634302616119385, - "learning_rate": 7.83912931668314e-05, - "loss": 0.4194, - "step": 2820 - }, - { - "epoch": 0.9269570913855224, - "grad_norm": 1.6447968482971191, - "learning_rate": 7.8249953923916e-05, - "loss": 0.4367, - "step": 2830 - }, - { - "epoch": 0.9302325581395349, - "grad_norm": 0.7306585311889648, - "learning_rate": 7.810828231166775e-05, - "loss": 0.4073, - "step": 2840 - }, - { - "epoch": 0.9335080248935473, - "grad_norm": 2.5663485527038574, - "learning_rate": 7.79662799968964e-05, - "loss": 0.4538, - "step": 2850 - }, - { - "epoch": 0.9367834916475598, - "grad_norm": 2.1073358058929443, - "learning_rate": 7.782394865030256e-05, - "loss": 0.4704, - "step": 2860 - }, - { - "epoch": 0.9400589584015723, - "grad_norm": 2.091595411300659, - "learning_rate": 7.768128994645796e-05, - "loss": 0.4193, - "step": 2870 - }, - { - "epoch": 0.9433344251555846, - "grad_norm": 0.9281072020530701, - "learning_rate": 7.75383055637858e-05, - "loss": 0.4096, - "step": 2880 - }, - { - "epoch": 0.9466098919095971, - "grad_norm": 2.2759907245635986, - "learning_rate": 7.739499718454102e-05, - "loss": 0.416, - "step": 2890 - }, - { - "epoch": 0.9498853586636096, - "grad_norm": 2.6236648559570312, - "learning_rate": 7.725136649479043e-05, - "loss": 0.3946, - "step": 2900 - }, - { - "epoch": 0.953160825417622, - "grad_norm": 2.1200802326202393, - "learning_rate": 7.710741518439297e-05, - "loss": 0.3806, - "step": 2910 - }, - { - "epoch": 0.9564362921716345, - "grad_norm": 2.5597639083862305, - "learning_rate": 7.696314494697975e-05, - "loss": 0.4455, - "step": 2920 - }, - { - "epoch": 0.9597117589256469, - "grad_norm": 1.416804552078247, - "learning_rate": 7.681855747993416e-05, - "loss": 0.3568, - "step": 2930 - }, - { - "epoch": 0.9629872256796593, - "grad_norm": 1.9969556331634521, - "learning_rate": 7.667365448437187e-05, - "loss": 0.3652, - "step": 2940 - }, - { - "epoch": 0.9662626924336718, - "grad_norm": 2.921762704849243, - "learning_rate": 7.65284376651209e-05, - "loss": 0.3711, - "step": 2950 - }, - { - "epoch": 0.9695381591876843, - "grad_norm": 4.733334064483643, - "learning_rate": 7.638290873070145e-05, - "loss": 0.395, - "step": 2960 - }, - { - "epoch": 0.9728136259416967, - "grad_norm": 2.3838038444519043, - "learning_rate": 7.623706939330583e-05, - "loss": 0.371, - "step": 2970 - }, - { - "epoch": 0.9760890926957091, - "grad_norm": 5.851949691772461, - "learning_rate": 7.609092136877843e-05, - "loss": 0.4629, - "step": 2980 - }, - { - "epoch": 0.9793645594497216, - "grad_norm": 1.079226016998291, - "learning_rate": 7.594446637659533e-05, - "loss": 0.391, - "step": 2990 - }, - { - "epoch": 0.982640026203734, - "grad_norm": 1.7345635890960693, - "learning_rate": 7.579770613984428e-05, - "loss": 0.4165, - "step": 3000 - }, - { - "epoch": 0.9859154929577465, - "grad_norm": 3.171224594116211, - "learning_rate": 7.565064238520422e-05, - "loss": 0.457, - "step": 3010 - }, - { - "epoch": 0.989190959711759, - "grad_norm": 0.8476964235305786, - "learning_rate": 7.550327684292516e-05, - "loss": 0.332, - "step": 3020 - }, - { - "epoch": 0.9924664264657713, - "grad_norm": 0.9504492282867432, - "learning_rate": 7.535561124680766e-05, - "loss": 0.4211, - "step": 3030 - }, - { - "epoch": 0.9957418932197838, - "grad_norm": 0.6670387983322144, - "learning_rate": 7.520764733418255e-05, - "loss": 0.3673, - "step": 3040 - }, - { - "epoch": 0.9990173599737963, - "grad_norm": 0.9342664480209351, - "learning_rate": 7.505938684589047e-05, - "loss": 0.4259, - "step": 3050 - }, - { - "epoch": 1.0022928267278086, - "grad_norm": 3.07145094871521, - "learning_rate": 7.49108315262613e-05, - "loss": 0.4368, - "step": 3060 - }, - { - "epoch": 1.005568293481821, - "grad_norm": 1.4697602987289429, - "learning_rate": 7.476198312309374e-05, - "loss": 0.3947, - "step": 3070 - }, - { - "epoch": 1.0088437602358336, - "grad_norm": 2.8098597526550293, - "learning_rate": 7.46128433876347e-05, - "loss": 0.376, - "step": 3080 - }, - { - "epoch": 1.012119226989846, - "grad_norm": 0.8025636076927185, - "learning_rate": 7.446341407455871e-05, - "loss": 0.3791, - "step": 3090 - }, - { - "epoch": 1.0153946937438585, - "grad_norm": 1.9544233083724976, - "learning_rate": 7.431369694194723e-05, - "loss": 0.4074, - "step": 3100 - }, - { - "epoch": 1.018670160497871, - "grad_norm": 1.7042185068130493, - "learning_rate": 7.416369375126806e-05, - "loss": 0.4752, - "step": 3110 - }, - { - "epoch": 1.0219456272518834, - "grad_norm": 1.429700493812561, - "learning_rate": 7.401340626735454e-05, - "loss": 0.3013, - "step": 3120 - }, - { - "epoch": 1.025221094005896, - "grad_norm": 0.46176475286483765, - "learning_rate": 7.386283625838477e-05, - "loss": 0.3585, - "step": 3130 - }, - { - "epoch": 1.0284965607599084, - "grad_norm": 1.0941903591156006, - "learning_rate": 7.371198549586091e-05, - "loss": 0.3671, - "step": 3140 - }, - { - "epoch": 1.0317720275139208, - "grad_norm": 4.855766296386719, - "learning_rate": 7.356085575458824e-05, - "loss": 0.4786, - "step": 3150 - }, - { - "epoch": 1.035047494267933, - "grad_norm": 1.4299640655517578, - "learning_rate": 7.340944881265425e-05, - "loss": 0.4302, - "step": 3160 - }, - { - "epoch": 1.0383229610219455, - "grad_norm": 2.46278715133667, - "learning_rate": 7.325776645140792e-05, - "loss": 0.4645, - "step": 3170 - }, - { - "epoch": 1.041598427775958, - "grad_norm": 0.42645177245140076, - "learning_rate": 7.310581045543849e-05, - "loss": 0.3623, - "step": 3180 - }, - { - "epoch": 1.0448738945299705, - "grad_norm": 1.015130877494812, - "learning_rate": 7.295358261255464e-05, - "loss": 0.3215, - "step": 3190 - }, - { - "epoch": 1.048149361283983, - "grad_norm": 0.6677165031433105, - "learning_rate": 7.280108471376344e-05, - "loss": 0.4631, - "step": 3200 - }, - { - "epoch": 1.0514248280379954, - "grad_norm": 2.20568585395813, - "learning_rate": 7.264831855324922e-05, - "loss": 0.2762, - "step": 3210 - }, - { - "epoch": 1.0547002947920079, - "grad_norm": 10.13748836517334, - "learning_rate": 7.24952859283525e-05, - "loss": 0.4085, - "step": 3220 - }, - { - "epoch": 1.0579757615460204, - "grad_norm": 1.0852124691009521, - "learning_rate": 7.234198863954885e-05, - "loss": 0.3931, - "step": 3230 - }, - { - "epoch": 1.0612512283000328, - "grad_norm": 4.532141208648682, - "learning_rate": 7.218842849042765e-05, - "loss": 0.359, - "step": 3240 - }, - { - "epoch": 1.0645266950540453, - "grad_norm": 1.3415788412094116, - "learning_rate": 7.203460728767095e-05, - "loss": 0.329, - "step": 3250 - }, - { - "epoch": 1.0678021618080575, - "grad_norm": 20.084430694580078, - "learning_rate": 7.188052684103215e-05, - "loss": 0.2903, - "step": 3260 - }, - { - "epoch": 1.07107762856207, - "grad_norm": 2.5338168144226074, - "learning_rate": 7.172618896331476e-05, - "loss": 0.4469, - "step": 3270 - }, - { - "epoch": 1.0743530953160825, - "grad_norm": 1.9570467472076416, - "learning_rate": 7.157159547035104e-05, - "loss": 0.413, - "step": 3280 - }, - { - "epoch": 1.077628562070095, - "grad_norm": 2.034734010696411, - "learning_rate": 7.141674818098063e-05, - "loss": 0.4064, - "step": 3290 - }, - { - "epoch": 1.0809040288241074, - "grad_norm": 1.4529087543487549, - "learning_rate": 7.126164891702915e-05, - "loss": 0.4706, - "step": 3300 - }, - { - "epoch": 1.0841794955781199, - "grad_norm": 2.5306437015533447, - "learning_rate": 7.11062995032868e-05, - "loss": 0.3804, - "step": 3310 - }, - { - "epoch": 1.0874549623321323, - "grad_norm": 1.3071904182434082, - "learning_rate": 7.095070176748687e-05, - "loss": 0.4448, - "step": 3320 - }, - { - "epoch": 1.0907304290861448, - "grad_norm": 1.5244219303131104, - "learning_rate": 7.079485754028423e-05, - "loss": 0.3487, - "step": 3330 - }, - { - "epoch": 1.0940058958401573, - "grad_norm": 1.6503034830093384, - "learning_rate": 7.063876865523377e-05, - "loss": 0.4512, - "step": 3340 - }, - { - "epoch": 1.0972813625941698, - "grad_norm": 1.4571095705032349, - "learning_rate": 7.04824369487689e-05, - "loss": 0.3882, - "step": 3350 - }, - { - "epoch": 1.1005568293481822, - "grad_norm": 4.536346435546875, - "learning_rate": 7.032586426017986e-05, - "loss": 0.3647, - "step": 3360 - }, - { - "epoch": 1.1038322961021945, - "grad_norm": 3.0561070442199707, - "learning_rate": 7.016905243159215e-05, - "loss": 0.4384, - "step": 3370 - }, - { - "epoch": 1.107107762856207, - "grad_norm": 1.7853630781173706, - "learning_rate": 7.001200330794481e-05, - "loss": 0.3916, - "step": 3380 - }, - { - "epoch": 1.1103832296102194, - "grad_norm": 6.381438732147217, - "learning_rate": 6.98547187369687e-05, - "loss": 0.3069, - "step": 3390 - }, - { - "epoch": 1.1136586963642319, - "grad_norm": 2.8521006107330322, - "learning_rate": 6.969720056916487e-05, - "loss": 0.4116, - "step": 3400 - }, - { - "epoch": 1.1169341631182443, - "grad_norm": 1.0471831560134888, - "learning_rate": 6.953945065778259e-05, - "loss": 0.378, - "step": 3410 - }, - { - "epoch": 1.1202096298722568, - "grad_norm": 7.885871410369873, - "learning_rate": 6.938147085879779e-05, - "loss": 0.431, - "step": 3420 - }, - { - "epoch": 1.1234850966262693, - "grad_norm": 1.2552326917648315, - "learning_rate": 6.9223263030891e-05, - "loss": 0.4039, - "step": 3430 - }, - { - "epoch": 1.1267605633802817, - "grad_norm": 3.150535821914673, - "learning_rate": 6.906482903542562e-05, - "loss": 0.3371, - "step": 3440 - }, - { - "epoch": 1.1300360301342942, - "grad_norm": 2.9623982906341553, - "learning_rate": 6.890617073642601e-05, - "loss": 0.3809, - "step": 3450 - }, - { - "epoch": 1.1333114968883067, - "grad_norm": 0.8672639727592468, - "learning_rate": 6.874729000055545e-05, - "loss": 0.386, - "step": 3460 - }, - { - "epoch": 1.1365869636423191, - "grad_norm": 5.734320640563965, - "learning_rate": 6.858818869709434e-05, - "loss": 0.3587, - "step": 3470 - }, - { - "epoch": 1.1398624303963314, - "grad_norm": 2.558964252471924, - "learning_rate": 6.842886869791809e-05, - "loss": 0.3388, - "step": 3480 - }, - { - "epoch": 1.1431378971503439, - "grad_norm": 1.417319416999817, - "learning_rate": 6.82693318774751e-05, - "loss": 0.4509, - "step": 3490 - }, - { - "epoch": 1.1464133639043563, - "grad_norm": 4.091428756713867, - "learning_rate": 6.810958011276483e-05, - "loss": 0.3487, - "step": 3500 - }, - { - "epoch": 1.1496888306583688, - "grad_norm": 9.960776329040527, - "learning_rate": 6.794961528331549e-05, - "loss": 0.4814, - "step": 3510 - }, - { - "epoch": 1.1529642974123813, - "grad_norm": 1.4533840417861938, - "learning_rate": 6.77894392711622e-05, - "loss": 0.3077, - "step": 3520 - }, - { - "epoch": 1.1562397641663937, - "grad_norm": 1.4448729753494263, - "learning_rate": 6.76290539608246e-05, - "loss": 0.3575, - "step": 3530 - }, - { - "epoch": 1.1595152309204062, - "grad_norm": 1.107999324798584, - "learning_rate": 6.746846123928487e-05, - "loss": 0.3994, - "step": 3540 - }, - { - "epoch": 1.1627906976744187, - "grad_norm": 2.3601200580596924, - "learning_rate": 6.730766299596538e-05, - "loss": 0.4013, - "step": 3550 - }, - { - "epoch": 1.1660661644284311, - "grad_norm": 3.256369113922119, - "learning_rate": 6.714666112270657e-05, - "loss": 0.2931, - "step": 3560 - }, - { - "epoch": 1.1693416311824434, - "grad_norm": 1.749741554260254, - "learning_rate": 6.698545751374465e-05, - "loss": 0.3772, - "step": 3570 - }, - { - "epoch": 1.1726170979364559, - "grad_norm": 0.6746839284896851, - "learning_rate": 6.682405406568927e-05, - "loss": 0.2692, - "step": 3580 - }, - { - "epoch": 1.1758925646904683, - "grad_norm": 3.498533248901367, - "learning_rate": 6.666245267750126e-05, - "loss": 0.3251, - "step": 3590 - }, - { - "epoch": 1.1791680314444808, - "grad_norm": 1.594074010848999, - "learning_rate": 6.650065525047036e-05, - "loss": 0.3969, - "step": 3600 - }, - { - "epoch": 1.1824434981984933, - "grad_norm": 2.773637533187866, - "learning_rate": 6.633866368819262e-05, - "loss": 0.4042, - "step": 3610 - }, - { - "epoch": 1.1857189649525057, - "grad_norm": 0.9472955465316772, - "learning_rate": 6.617647989654825e-05, - "loss": 0.399, - "step": 3620 - }, - { - "epoch": 1.1889944317065182, - "grad_norm": 2.864743709564209, - "learning_rate": 6.601410578367911e-05, - "loss": 0.3708, - "step": 3630 - }, - { - "epoch": 1.1922698984605307, - "grad_norm": 1.3508647680282593, - "learning_rate": 6.585154325996616e-05, - "loss": 0.3877, - "step": 3640 - }, - { - "epoch": 1.1955453652145431, - "grad_norm": 1.0608330965042114, - "learning_rate": 6.568879423800716e-05, - "loss": 0.3346, - "step": 3650 - }, - { - "epoch": 1.1988208319685556, - "grad_norm": 4.288187026977539, - "learning_rate": 6.552586063259403e-05, - "loss": 0.3722, - "step": 3660 - }, - { - "epoch": 1.202096298722568, - "grad_norm": 1.2619587182998657, - "learning_rate": 6.536274436069037e-05, - "loss": 0.2926, - "step": 3670 - }, - { - "epoch": 1.2053717654765803, - "grad_norm": 0.9936348795890808, - "learning_rate": 6.519944734140896e-05, - "loss": 0.2837, - "step": 3680 - }, - { - "epoch": 1.2086472322305928, - "grad_norm": 2.6671509742736816, - "learning_rate": 6.503597149598903e-05, - "loss": 0.3847, - "step": 3690 - }, - { - "epoch": 1.2119226989846053, - "grad_norm": 1.0667388439178467, - "learning_rate": 6.487231874777387e-05, - "loss": 0.3531, - "step": 3700 - }, - { - "epoch": 1.2151981657386177, - "grad_norm": 5.38046407699585, - "learning_rate": 6.470849102218799e-05, - "loss": 0.307, - "step": 3710 - }, - { - "epoch": 1.2184736324926302, - "grad_norm": 3.895265817642212, - "learning_rate": 6.45444902467146e-05, - "loss": 0.3949, - "step": 3720 - }, - { - "epoch": 1.2217490992466427, - "grad_norm": 3.53731632232666, - "learning_rate": 6.438031835087292e-05, - "loss": 0.311, - "step": 3730 - }, - { - "epoch": 1.2250245660006551, - "grad_norm": 2.400347948074341, - "learning_rate": 6.42159772661954e-05, - "loss": 0.3836, - "step": 3740 - }, - { - "epoch": 1.2283000327546676, - "grad_norm": 1.2507903575897217, - "learning_rate": 6.405146892620508e-05, - "loss": 0.3649, - "step": 3750 - }, - { - "epoch": 1.23157549950868, - "grad_norm": 14.933448791503906, - "learning_rate": 6.388679526639282e-05, - "loss": 0.3188, - "step": 3760 - }, - { - "epoch": 1.2348509662626925, - "grad_norm": 2.4308481216430664, - "learning_rate": 6.372195822419449e-05, - "loss": 0.3949, - "step": 3770 - }, - { - "epoch": 1.238126433016705, - "grad_norm": 6.479183673858643, - "learning_rate": 6.355695973896824e-05, - "loss": 0.4244, - "step": 3780 - }, - { - "epoch": 1.2414018997707172, - "grad_norm": 3.517155170440674, - "learning_rate": 6.339180175197159e-05, - "loss": 0.4085, - "step": 3790 - }, - { - "epoch": 1.2446773665247297, - "grad_norm": 5.643728256225586, - "learning_rate": 6.322648620633872e-05, - "loss": 0.3342, - "step": 3800 - }, - { - "epoch": 1.2479528332787422, - "grad_norm": 1.671078085899353, - "learning_rate": 6.306101504705746e-05, - "loss": 0.3902, - "step": 3810 - }, - { - "epoch": 1.2512283000327546, - "grad_norm": 3.8140645027160645, - "learning_rate": 6.289539022094654e-05, - "loss": 0.3469, - "step": 3820 - }, - { - "epoch": 1.2545037667867671, - "grad_norm": 3.3500137329101562, - "learning_rate": 6.27296136766326e-05, - "loss": 0.4158, - "step": 3830 - }, - { - "epoch": 1.2577792335407796, - "grad_norm": 7.794027805328369, - "learning_rate": 6.25636873645273e-05, - "loss": 0.4174, - "step": 3840 - }, - { - "epoch": 1.261054700294792, - "grad_norm": 2.8805992603302, - "learning_rate": 6.239761323680435e-05, - "loss": 0.3501, - "step": 3850 - }, - { - "epoch": 1.2643301670488045, - "grad_norm": 12.423330307006836, - "learning_rate": 6.223139324737659e-05, - "loss": 0.4007, - "step": 3860 - }, - { - "epoch": 1.267605633802817, - "grad_norm": 1.9705214500427246, - "learning_rate": 6.206502935187291e-05, - "loss": 0.3231, - "step": 3870 - }, - { - "epoch": 1.2708811005568292, - "grad_norm": 3.554030179977417, - "learning_rate": 6.189852350761536e-05, - "loss": 0.422, - "step": 3880 - }, - { - "epoch": 1.274156567310842, - "grad_norm": 1.4976438283920288, - "learning_rate": 6.173187767359603e-05, - "loss": 0.3674, - "step": 3890 - }, - { - "epoch": 1.2774320340648542, - "grad_norm": 8.399833679199219, - "learning_rate": 6.156509381045402e-05, - "loss": 0.3402, - "step": 3900 - }, - { - "epoch": 1.2807075008188666, - "grad_norm": 2.961317539215088, - "learning_rate": 6.139817388045242e-05, - "loss": 0.3996, - "step": 3910 - }, - { - "epoch": 1.283982967572879, - "grad_norm": 3.535125970840454, - "learning_rate": 6.123111984745514e-05, - "loss": 0.3282, - "step": 3920 - }, - { - "epoch": 1.2872584343268916, - "grad_norm": 4.967417240142822, - "learning_rate": 6.106393367690393e-05, - "loss": 0.2677, - "step": 3930 - }, - { - "epoch": 1.290533901080904, - "grad_norm": 2.470003128051758, - "learning_rate": 6.089661733579507e-05, - "loss": 0.4255, - "step": 3940 - }, - { - "epoch": 1.2938093678349165, - "grad_norm": 0.9133200645446777, - "learning_rate": 6.072917279265645e-05, - "loss": 0.3475, - "step": 3950 - }, - { - "epoch": 1.297084834588929, - "grad_norm": 1.7044758796691895, - "learning_rate": 6.05616020175242e-05, - "loss": 0.4032, - "step": 3960 - }, - { - "epoch": 1.3003603013429414, - "grad_norm": 6.1042585372924805, - "learning_rate": 6.039390698191968e-05, - "loss": 0.3616, - "step": 3970 - }, - { - "epoch": 1.303635768096954, - "grad_norm": 2.1450185775756836, - "learning_rate": 6.022608965882617e-05, - "loss": 0.3967, - "step": 3980 - }, - { - "epoch": 1.3069112348509662, - "grad_norm": 0.6380177736282349, - "learning_rate": 6.0058152022665716e-05, - "loss": 0.3414, - "step": 3990 - }, - { - "epoch": 1.3101867016049786, - "grad_norm": 4.431483745574951, - "learning_rate": 5.989009604927587e-05, - "loss": 0.4099, - "step": 4000 - }, - { - "epoch": 1.313462168358991, - "grad_norm": 4.865090370178223, - "learning_rate": 5.9721923715886475e-05, - "loss": 0.3274, - "step": 4010 - }, - { - "epoch": 1.3167376351130036, - "grad_norm": 17.26369285583496, - "learning_rate": 5.9553637001096386e-05, - "loss": 0.3864, - "step": 4020 - }, - { - "epoch": 1.320013101867016, - "grad_norm": 0.9660899639129639, - "learning_rate": 5.938523788485017e-05, - "loss": 0.3634, - "step": 4030 - }, - { - "epoch": 1.3232885686210285, - "grad_norm": 5.793810844421387, - "learning_rate": 5.9216728348414855e-05, - "loss": 0.3884, - "step": 4040 - }, - { - "epoch": 1.326564035375041, - "grad_norm": 280.25372314453125, - "learning_rate": 5.904811037435656e-05, - "loss": 0.3802, - "step": 4050 - }, - { - "epoch": 1.3298395021290534, - "grad_norm": 1.9583410024642944, - "learning_rate": 5.887938594651726e-05, - "loss": 0.3261, - "step": 4060 - }, - { - "epoch": 1.333114968883066, - "grad_norm": 1.43271005153656, - "learning_rate": 5.8710557049991345e-05, - "loss": 0.3222, - "step": 4070 - }, - { - "epoch": 1.3363904356370782, - "grad_norm": 0.6183338165283203, - "learning_rate": 5.854162567110234e-05, - "loss": 0.3402, - "step": 4080 - }, - { - "epoch": 1.3396659023910908, - "grad_norm": 2.1392571926116943, - "learning_rate": 5.8372593797379474e-05, - "loss": 0.3182, - "step": 4090 - }, - { - "epoch": 1.342941369145103, - "grad_norm": 1.5215672254562378, - "learning_rate": 5.820346341753435e-05, - "loss": 0.3812, - "step": 4100 - }, - { - "epoch": 1.3462168358991156, - "grad_norm": 1.2768079042434692, - "learning_rate": 5.8034236521437554e-05, - "loss": 0.3452, - "step": 4110 - }, - { - "epoch": 1.349492302653128, - "grad_norm": 6.333654403686523, - "learning_rate": 5.7864915100095154e-05, - "loss": 0.4295, - "step": 4120 - }, - { - "epoch": 1.3527677694071405, - "grad_norm": 5.985565185546875, - "learning_rate": 5.769550114562539e-05, - "loss": 0.3587, - "step": 4130 - }, - { - "epoch": 1.356043236161153, - "grad_norm": 1.5624040365219116, - "learning_rate": 5.752599665123513e-05, - "loss": 0.3851, - "step": 4140 - }, - { - "epoch": 1.3593187029151654, - "grad_norm": 4.838425636291504, - "learning_rate": 5.735640361119652e-05, - "loss": 0.3553, - "step": 4150 - }, - { - "epoch": 1.362594169669178, - "grad_norm": 1.4126315116882324, - "learning_rate": 5.7186724020823456e-05, - "loss": 0.3165, - "step": 4160 - }, - { - "epoch": 1.3658696364231904, - "grad_norm": 2.0345327854156494, - "learning_rate": 5.7016959876448126e-05, - "loss": 0.345, - "step": 4170 - }, - { - "epoch": 1.3691451031772028, - "grad_norm": 0.852786123752594, - "learning_rate": 5.684711317539751e-05, - "loss": 0.4027, - "step": 4180 - }, - { - "epoch": 1.372420569931215, - "grad_norm": 1.806326985359192, - "learning_rate": 5.66771859159699e-05, - "loss": 0.3575, - "step": 4190 - }, - { - "epoch": 1.3756960366852278, - "grad_norm": 1.3827389478683472, - "learning_rate": 5.650718009741138e-05, - "loss": 0.359, - "step": 4200 - }, - { - "epoch": 1.37897150343924, - "grad_norm": 4.76085090637207, - "learning_rate": 5.633709771989235e-05, - "loss": 0.3918, - "step": 4210 - }, - { - "epoch": 1.3822469701932525, - "grad_norm": 4.139555931091309, - "learning_rate": 5.616694078448385e-05, - "loss": 0.3197, - "step": 4220 - }, - { - "epoch": 1.385522436947265, - "grad_norm": 2.082777976989746, - "learning_rate": 5.5996711293134205e-05, - "loss": 0.3709, - "step": 4230 - }, - { - "epoch": 1.3887979037012774, - "grad_norm": 1.9760844707489014, - "learning_rate": 5.582641124864536e-05, - "loss": 0.3646, - "step": 4240 - }, - { - "epoch": 1.39207337045529, - "grad_norm": 2.4558799266815186, - "learning_rate": 5.565604265464931e-05, - "loss": 0.3859, - "step": 4250 - }, - { - "epoch": 1.3953488372093024, - "grad_norm": 10.621898651123047, - "learning_rate": 5.548560751558461e-05, - "loss": 0.3398, - "step": 4260 - }, - { - "epoch": 1.3986243039633148, - "grad_norm": 2.439568042755127, - "learning_rate": 5.5315107836672664e-05, - "loss": 0.3953, - "step": 4270 - }, - { - "epoch": 1.4018997707173273, - "grad_norm": 7.759012699127197, - "learning_rate": 5.51445456238943e-05, - "loss": 0.3958, - "step": 4280 - }, - { - "epoch": 1.4051752374713398, - "grad_norm": 1.5926508903503418, - "learning_rate": 5.497392288396597e-05, - "loss": 0.3492, - "step": 4290 - }, - { - "epoch": 1.408450704225352, - "grad_norm": 1.8101063966751099, - "learning_rate": 5.4803241624316336e-05, - "loss": 0.4574, - "step": 4300 - }, - { - "epoch": 1.4117261709793645, - "grad_norm": 1.333472728729248, - "learning_rate": 5.463250385306253e-05, - "loss": 0.3596, - "step": 4310 - }, - { - "epoch": 1.415001637733377, - "grad_norm": 2.7605857849121094, - "learning_rate": 5.446171157898655e-05, - "loss": 0.3351, - "step": 4320 - }, - { - "epoch": 1.4182771044873894, - "grad_norm": 2.0737509727478027, - "learning_rate": 5.429086681151165e-05, - "loss": 0.3865, - "step": 4330 - }, - { - "epoch": 1.4215525712414019, - "grad_norm": 1.0666999816894531, - "learning_rate": 5.411997156067867e-05, - "loss": 0.3858, - "step": 4340 - }, - { - "epoch": 1.4248280379954144, - "grad_norm": 2.635807514190674, - "learning_rate": 5.394902783712241e-05, - "loss": 0.3308, - "step": 4350 - }, - { - "epoch": 1.4281035047494268, - "grad_norm": 5.814352512359619, - "learning_rate": 5.377803765204798e-05, - "loss": 0.3774, - "step": 4360 - }, - { - "epoch": 1.4313789715034393, - "grad_norm": 0.6657505035400391, - "learning_rate": 5.3607003017207114e-05, - "loss": 0.3702, - "step": 4370 - }, - { - "epoch": 1.4346544382574518, - "grad_norm": 12.114752769470215, - "learning_rate": 5.343592594487451e-05, - "loss": 0.4005, - "step": 4380 - }, - { - "epoch": 1.437929905011464, - "grad_norm": 1.7607343196868896, - "learning_rate": 5.3264808447824124e-05, - "loss": 0.4227, - "step": 4390 - }, - { - "epoch": 1.4412053717654767, - "grad_norm": 2.3303537368774414, - "learning_rate": 5.3093652539305584e-05, - "loss": 0.3457, - "step": 4400 - }, - { - "epoch": 1.444480838519489, - "grad_norm": 1.5805153846740723, - "learning_rate": 5.292246023302041e-05, - "loss": 0.4075, - "step": 4410 - }, - { - "epoch": 1.4477563052735014, - "grad_norm": 2.2488300800323486, - "learning_rate": 5.275123354309833e-05, - "loss": 0.3946, - "step": 4420 - }, - { - "epoch": 1.4510317720275139, - "grad_norm": 1.6756857633590698, - "learning_rate": 5.2579974484073655e-05, - "loss": 0.2896, - "step": 4430 - }, - { - "epoch": 1.4543072387815263, - "grad_norm": 1.1381936073303223, - "learning_rate": 5.2408685070861486e-05, - "loss": 0.2645, - "step": 4440 - }, - { - "epoch": 1.4575827055355388, - "grad_norm": 1.6505337953567505, - "learning_rate": 5.223736731873405e-05, - "loss": 0.389, - "step": 4450 - }, - { - "epoch": 1.4608581722895513, - "grad_norm": 5.607367038726807, - "learning_rate": 5.206602324329704e-05, - "loss": 0.292, - "step": 4460 - }, - { - "epoch": 1.4641336390435638, - "grad_norm": 1.095463514328003, - "learning_rate": 5.1894654860465796e-05, - "loss": 0.4097, - "step": 4470 - }, - { - "epoch": 1.4674091057975762, - "grad_norm": 6.031702518463135, - "learning_rate": 5.1723264186441664e-05, - "loss": 0.3869, - "step": 4480 - }, - { - "epoch": 1.4706845725515887, - "grad_norm": 6.589273929595947, - "learning_rate": 5.155185323768824e-05, - "loss": 0.3395, - "step": 4490 - }, - { - "epoch": 1.473960039305601, - "grad_norm": 3.8057470321655273, - "learning_rate": 5.138042403090769e-05, - "loss": 0.4125, - "step": 4500 - }, - { - "epoch": 1.4772355060596136, - "grad_norm": 0.8534798622131348, - "learning_rate": 5.120897858301696e-05, - "loss": 0.3345, - "step": 4510 - }, - { - "epoch": 1.4805109728136259, - "grad_norm": 2.8705549240112305, - "learning_rate": 5.103751891112409e-05, - "loss": 0.3382, - "step": 4520 - }, - { - "epoch": 1.4837864395676383, - "grad_norm": 1.631131887435913, - "learning_rate": 5.086604703250448e-05, - "loss": 0.307, - "step": 4530 - }, - { - "epoch": 1.4870619063216508, - "grad_norm": 4.411392688751221, - "learning_rate": 5.069456496457711e-05, - "loss": 0.3213, - "step": 4540 - }, - { - "epoch": 1.4903373730756633, - "grad_norm": 1.9629542827606201, - "learning_rate": 5.052307472488087e-05, - "loss": 0.3295, - "step": 4550 - }, - { - "epoch": 1.4936128398296757, - "grad_norm": 1.5351589918136597, - "learning_rate": 5.03515783310508e-05, - "loss": 0.3005, - "step": 4560 - }, - { - "epoch": 1.4968883065836882, - "grad_norm": 4.768153190612793, - "learning_rate": 5.018007780079432e-05, - "loss": 0.4036, - "step": 4570 - }, - { - "epoch": 1.5001637733377007, - "grad_norm": 6.231042385101318, - "learning_rate": 5.0008575151867545e-05, - "loss": 0.4815, - "step": 4580 - }, - { - "epoch": 1.503439240091713, - "grad_norm": 5.570981979370117, - "learning_rate": 4.9837072402051474e-05, - "loss": 0.3365, - "step": 4590 - }, - { - "epoch": 1.5067147068457256, - "grad_norm": 1.1211837530136108, - "learning_rate": 4.966557156912835e-05, - "loss": 0.4393, - "step": 4600 - }, - { - "epoch": 1.5099901735997379, - "grad_norm": 1.6216473579406738, - "learning_rate": 4.949407467085781e-05, - "loss": 0.3307, - "step": 4610 - }, - { - "epoch": 1.5132656403537506, - "grad_norm": 1.9833356142044067, - "learning_rate": 4.9322583724953234e-05, - "loss": 0.3883, - "step": 4620 - }, - { - "epoch": 1.5165411071077628, - "grad_norm": 2.1016042232513428, - "learning_rate": 4.915110074905795e-05, - "loss": 0.4067, - "step": 4630 - }, - { - "epoch": 1.5198165738617753, - "grad_norm": 2.826876640319824, - "learning_rate": 4.897962776072154e-05, - "loss": 0.3756, - "step": 4640 - }, - { - "epoch": 1.5230920406157877, - "grad_norm": 8.856269836425781, - "learning_rate": 4.880816677737605e-05, - "loss": 0.4151, - "step": 4650 - }, - { - "epoch": 1.5263675073698002, - "grad_norm": 1.4084227085113525, - "learning_rate": 4.8636719816312296e-05, - "loss": 0.3092, - "step": 4660 - }, - { - "epoch": 1.5296429741238127, - "grad_norm": 1.85322904586792, - "learning_rate": 4.846528889465612e-05, - "loss": 0.4106, - "step": 4670 - }, - { - "epoch": 1.5329184408778251, - "grad_norm": 2.41858172416687, - "learning_rate": 4.829387602934467e-05, - "loss": 0.3492, - "step": 4680 - }, - { - "epoch": 1.5361939076318376, - "grad_norm": 1.6673808097839355, - "learning_rate": 4.8122483237102613e-05, - "loss": 0.3912, - "step": 4690 - }, - { - "epoch": 1.5394693743858499, - "grad_norm": 1.8599348068237305, - "learning_rate": 4.795111253441852e-05, - "loss": 0.4306, - "step": 4700 - }, - { - "epoch": 1.5427448411398625, - "grad_norm": 2.9528512954711914, - "learning_rate": 4.777976593752098e-05, - "loss": 0.3948, - "step": 4710 - }, - { - "epoch": 1.5460203078938748, - "grad_norm": 0.6291877627372742, - "learning_rate": 4.760844546235504e-05, - "loss": 0.3693, - "step": 4720 - }, - { - "epoch": 1.5492957746478875, - "grad_norm": 0.9277985095977783, - "learning_rate": 4.743715312455841e-05, - "loss": 0.3086, - "step": 4730 - }, - { - "epoch": 1.5525712414018997, - "grad_norm": 1.523949384689331, - "learning_rate": 4.726589093943771e-05, - "loss": 0.3672, - "step": 4740 - }, - { - "epoch": 1.5558467081559122, - "grad_norm": 4.634634017944336, - "learning_rate": 4.709466092194488e-05, - "loss": 0.4547, - "step": 4750 - }, - { - "epoch": 1.5591221749099247, - "grad_norm": 2.3649632930755615, - "learning_rate": 4.69234650866533e-05, - "loss": 0.3802, - "step": 4760 - }, - { - "epoch": 1.5623976416639371, - "grad_norm": 2.4040448665618896, - "learning_rate": 4.675230544773428e-05, - "loss": 0.384, - "step": 4770 - }, - { - "epoch": 1.5656731084179496, - "grad_norm": 2.1583518981933594, - "learning_rate": 4.658118401893322e-05, - "loss": 0.319, - "step": 4780 - }, - { - "epoch": 1.5689485751719618, - "grad_norm": 2.168704032897949, - "learning_rate": 4.6410102813545964e-05, - "loss": 0.3562, - "step": 4790 - }, - { - "epoch": 1.5722240419259745, - "grad_norm": 3.7866313457489014, - "learning_rate": 4.623906384439514e-05, - "loss": 0.4019, - "step": 4800 - }, - { - "epoch": 1.5754995086799868, - "grad_norm": 11.472986221313477, - "learning_rate": 4.606806912380642e-05, - "loss": 0.3424, - "step": 4810 - }, - { - "epoch": 1.5787749754339995, - "grad_norm": 1.7527011632919312, - "learning_rate": 4.589712066358491e-05, - "loss": 0.3286, - "step": 4820 - }, - { - "epoch": 1.5820504421880117, - "grad_norm": 1.719495177268982, - "learning_rate": 4.572622047499145e-05, - "loss": 0.3368, - "step": 4830 - }, - { - "epoch": 1.5853259089420242, - "grad_norm": 3.2706644535064697, - "learning_rate": 4.5555370568718906e-05, - "loss": 0.3216, - "step": 4840 - }, - { - "epoch": 1.5886013756960367, - "grad_norm": 5.074819564819336, - "learning_rate": 4.538457295486862e-05, - "loss": 0.3557, - "step": 4850 - }, - { - "epoch": 1.5918768424500491, - "grad_norm": 0.5600479245185852, - "learning_rate": 4.521382964292663e-05, - "loss": 0.3756, - "step": 4860 - }, - { - "epoch": 1.5951523092040616, - "grad_norm": 2.0590004920959473, - "learning_rate": 4.504314264174016e-05, - "loss": 0.3697, - "step": 4870 - }, - { - "epoch": 1.598427775958074, - "grad_norm": 3.1339550018310547, - "learning_rate": 4.487251395949388e-05, - "loss": 0.3415, - "step": 4880 - }, - { - "epoch": 1.6017032427120865, - "grad_norm": 4.270013809204102, - "learning_rate": 4.470194560368633e-05, - "loss": 0.365, - "step": 4890 - }, - { - "epoch": 1.6049787094660988, - "grad_norm": 0.6680019497871399, - "learning_rate": 4.4531439581106295e-05, - "loss": 0.3257, - "step": 4900 - }, - { - "epoch": 1.6082541762201115, - "grad_norm": 1.4795156717300415, - "learning_rate": 4.4360997897809174e-05, - "loss": 0.407, - "step": 4910 - }, - { - "epoch": 1.6115296429741237, - "grad_norm": 4.313704490661621, - "learning_rate": 4.4190622559093425e-05, - "loss": 0.3297, - "step": 4920 - }, - { - "epoch": 1.6148051097281364, - "grad_norm": 2.2392537593841553, - "learning_rate": 4.402031556947693e-05, - "loss": 0.2766, - "step": 4930 - }, - { - "epoch": 1.6180805764821486, - "grad_norm": 7.899417877197266, - "learning_rate": 4.3850078932673394e-05, - "loss": 0.349, - "step": 4940 - }, - { - "epoch": 1.6213560432361611, - "grad_norm": 5.756536483764648, - "learning_rate": 4.3679914651568846e-05, - "loss": 0.2929, - "step": 4950 - }, - { - "epoch": 1.6246315099901736, - "grad_norm": 1.509802222251892, - "learning_rate": 4.350982472819799e-05, - "loss": 0.3104, - "step": 4960 - }, - { - "epoch": 1.627906976744186, - "grad_norm": 2.918820381164551, - "learning_rate": 4.333981116372071e-05, - "loss": 0.3541, - "step": 4970 - }, - { - "epoch": 1.6311824434981985, - "grad_norm": 2.5075724124908447, - "learning_rate": 4.316987595839851e-05, - "loss": 0.2626, - "step": 4980 - }, - { - "epoch": 1.634457910252211, - "grad_norm": 1.8369845151901245, - "learning_rate": 4.300002111157094e-05, - "loss": 0.3723, - "step": 4990 - }, - { - "epoch": 1.6377333770062235, - "grad_norm": 4.382946014404297, - "learning_rate": 4.283024862163216e-05, - "loss": 0.3378, - "step": 5000 - }, - { - "epoch": 1.6410088437602357, - "grad_norm": 4.092770576477051, - "learning_rate": 4.2660560486007325e-05, - "loss": 0.3686, - "step": 5010 - }, - { - "epoch": 1.6442843105142484, - "grad_norm": 1.3430540561676025, - "learning_rate": 4.249095870112916e-05, - "loss": 0.2931, - "step": 5020 - }, - { - "epoch": 1.6475597772682606, - "grad_norm": 10.310168266296387, - "learning_rate": 4.2321445262414454e-05, - "loss": 0.4021, - "step": 5030 - }, - { - "epoch": 1.6508352440222733, - "grad_norm": 3.7715585231781006, - "learning_rate": 4.2152022164240555e-05, - "loss": 0.3673, - "step": 5040 - }, - { - "epoch": 1.6541107107762856, - "grad_norm": 2.9399900436401367, - "learning_rate": 4.198269139992196e-05, - "loss": 0.3558, - "step": 5050 - }, - { - "epoch": 1.657386177530298, - "grad_norm": 0.8002705574035645, - "learning_rate": 4.181345496168678e-05, - "loss": 0.3626, - "step": 5060 - }, - { - "epoch": 1.6606616442843105, - "grad_norm": 3.334031105041504, - "learning_rate": 4.16443148406534e-05, - "loss": 0.4222, - "step": 5070 - }, - { - "epoch": 1.663937111038323, - "grad_norm": 1.9074835777282715, - "learning_rate": 4.147527302680699e-05, - "loss": 0.3411, - "step": 5080 - }, - { - "epoch": 1.6672125777923354, - "grad_norm": 7.020357608795166, - "learning_rate": 4.130633150897609e-05, - "loss": 0.382, - "step": 5090 - }, - { - "epoch": 1.6704880445463477, - "grad_norm": 1.594455599784851, - "learning_rate": 4.113749227480923e-05, - "loss": 0.367, - "step": 5100 - }, - { - "epoch": 1.6737635113003604, - "grad_norm": 3.596156597137451, - "learning_rate": 4.0968757310751556e-05, - "loss": 0.4457, - "step": 5110 - }, - { - "epoch": 1.6770389780543726, - "grad_norm": 1.4038106203079224, - "learning_rate": 4.080012860202142e-05, - "loss": 0.4062, - "step": 5120 - }, - { - "epoch": 1.6803144448083853, - "grad_norm": 2.0355899333953857, - "learning_rate": 4.0631608132587105e-05, - "loss": 0.318, - "step": 5130 - }, - { - "epoch": 1.6835899115623976, - "grad_norm": 0.6998187899589539, - "learning_rate": 4.046319788514334e-05, - "loss": 0.3185, - "step": 5140 - }, - { - "epoch": 1.68686537831641, - "grad_norm": 2.2005085945129395, - "learning_rate": 4.0294899841088135e-05, - "loss": 0.3357, - "step": 5150 - }, - { - "epoch": 1.6901408450704225, - "grad_norm": 2.754657030105591, - "learning_rate": 4.012671598049933e-05, - "loss": 0.3657, - "step": 5160 - }, - { - "epoch": 1.693416311824435, - "grad_norm": 1.1556801795959473, - "learning_rate": 3.9958648282111416e-05, - "loss": 0.3289, - "step": 5170 - }, - { - "epoch": 1.6966917785784474, - "grad_norm": 0.7944522500038147, - "learning_rate": 3.979069872329219e-05, - "loss": 0.3926, - "step": 5180 - }, - { - "epoch": 1.69996724533246, - "grad_norm": 1.9170118570327759, - "learning_rate": 3.962286928001944e-05, - "loss": 0.2898, - "step": 5190 - }, - { - "epoch": 1.7032427120864724, - "grad_norm": 3.121422290802002, - "learning_rate": 3.945516192685785e-05, - "loss": 0.3921, - "step": 5200 - }, - { - "epoch": 1.7065181788404846, - "grad_norm": 6.7555317878723145, - "learning_rate": 3.928757863693562e-05, - "loss": 0.3643, - "step": 5210 - }, - { - "epoch": 1.7097936455944973, - "grad_norm": 1.0285550355911255, - "learning_rate": 3.912012138192131e-05, - "loss": 0.3394, - "step": 5220 - }, - { - "epoch": 1.7130691123485096, - "grad_norm": 52.312110900878906, - "learning_rate": 3.895279213200067e-05, - "loss": 0.2719, - "step": 5230 - }, - { - "epoch": 1.7163445791025222, - "grad_norm": 1.516508936882019, - "learning_rate": 3.8785592855853394e-05, - "loss": 0.3542, - "step": 5240 - }, - { - "epoch": 1.7196200458565345, - "grad_norm": 2.039396047592163, - "learning_rate": 3.861852552063002e-05, - "loss": 0.4322, - "step": 5250 - }, - { - "epoch": 1.722895512610547, - "grad_norm": 4.220056056976318, - "learning_rate": 3.845159209192876e-05, - "loss": 0.3934, - "step": 5260 - }, - { - "epoch": 1.7261709793645594, - "grad_norm": 1.952877163887024, - "learning_rate": 3.8284794533772354e-05, - "loss": 0.3447, - "step": 5270 - }, - { - "epoch": 1.729446446118572, - "grad_norm": 1.4389408826828003, - "learning_rate": 3.811813480858502e-05, - "loss": 0.3976, - "step": 5280 - }, - { - "epoch": 1.7327219128725844, - "grad_norm": 1.9890987873077393, - "learning_rate": 3.7951614877169284e-05, - "loss": 0.3277, - "step": 5290 - }, - { - "epoch": 1.7359973796265968, - "grad_norm": 2.1475017070770264, - "learning_rate": 3.7785236698683e-05, - "loss": 0.3386, - "step": 5300 - }, - { - "epoch": 1.7392728463806093, - "grad_norm": 3.081803321838379, - "learning_rate": 3.7619002230616225e-05, - "loss": 0.4088, - "step": 5310 - }, - { - "epoch": 1.7425483131346216, - "grad_norm": 7.6761579513549805, - "learning_rate": 3.7452913428768224e-05, - "loss": 0.3842, - "step": 5320 - }, - { - "epoch": 1.7458237798886342, - "grad_norm": 2.025059700012207, - "learning_rate": 3.728697224722446e-05, - "loss": 0.3268, - "step": 5330 - }, - { - "epoch": 1.7490992466426465, - "grad_norm": 0.6343122124671936, - "learning_rate": 3.712118063833361e-05, - "loss": 0.3205, - "step": 5340 - }, - { - "epoch": 1.7523747133966592, - "grad_norm": 14.0781831741333, - "learning_rate": 3.695554055268455e-05, - "loss": 0.3353, - "step": 5350 - }, - { - "epoch": 1.7556501801506714, - "grad_norm": 1.599924921989441, - "learning_rate": 3.679005393908348e-05, - "loss": 0.3427, - "step": 5360 - }, - { - "epoch": 1.758925646904684, - "grad_norm": 0.9019806385040283, - "learning_rate": 3.662472274453092e-05, - "loss": 0.3699, - "step": 5370 - }, - { - "epoch": 1.7622011136586964, - "grad_norm": 2.2208123207092285, - "learning_rate": 3.6459548914198855e-05, - "loss": 0.3119, - "step": 5380 - }, - { - "epoch": 1.7654765804127088, - "grad_norm": 2.0848231315612793, - "learning_rate": 3.629453439140782e-05, - "loss": 0.3156, - "step": 5390 - }, - { - "epoch": 1.7687520471667213, - "grad_norm": 0.7760680317878723, - "learning_rate": 3.612968111760406e-05, - "loss": 0.3345, - "step": 5400 - }, - { - "epoch": 1.7720275139207335, - "grad_norm": 1.020798921585083, - "learning_rate": 3.596499103233669e-05, - "loss": 0.363, - "step": 5410 - }, - { - "epoch": 1.7753029806747462, - "grad_norm": 1.7109076976776123, - "learning_rate": 3.5800466073234825e-05, - "loss": 0.4742, - "step": 5420 - }, - { - "epoch": 1.7785784474287585, - "grad_norm": 5.3574323654174805, - "learning_rate": 3.563610817598485e-05, - "loss": 0.3143, - "step": 5430 - }, - { - "epoch": 1.7818539141827712, - "grad_norm": 0.8525514006614685, - "learning_rate": 3.547191927430759e-05, - "loss": 0.3483, - "step": 5440 - }, - { - "epoch": 1.7851293809367834, - "grad_norm": 3.113211154937744, - "learning_rate": 3.530790129993561e-05, - "loss": 0.3028, - "step": 5450 - }, - { - "epoch": 1.7884048476907959, - "grad_norm": 1.715874195098877, - "learning_rate": 3.514405618259047e-05, - "loss": 0.2859, - "step": 5460 - }, - { - "epoch": 1.7916803144448084, - "grad_norm": 2.827599287033081, - "learning_rate": 3.498038584995996e-05, - "loss": 0.301, - "step": 5470 - }, - { - "epoch": 1.7949557811988208, - "grad_norm": 2.2840957641601562, - "learning_rate": 3.4816892227675565e-05, - "loss": 0.3163, - "step": 5480 - }, - { - "epoch": 1.7982312479528333, - "grad_norm": 3.558912754058838, - "learning_rate": 3.465357723928963e-05, - "loss": 0.333, - "step": 5490 - }, - { - "epoch": 1.8015067147068458, - "grad_norm": 1.6477887630462646, - "learning_rate": 3.4490442806252864e-05, - "loss": 0.3713, - "step": 5500 - }, - { - "epoch": 1.8047821814608582, - "grad_norm": 2.674889326095581, - "learning_rate": 3.432749084789169e-05, - "loss": 0.3908, - "step": 5510 - }, - { - "epoch": 1.8080576482148705, - "grad_norm": 1.1688625812530518, - "learning_rate": 3.4164723281385624e-05, - "loss": 0.379, - "step": 5520 - }, - { - "epoch": 1.8113331149688832, - "grad_norm": 3.657430410385132, - "learning_rate": 3.4002142021744805e-05, - "loss": 0.3708, - "step": 5530 - }, - { - "epoch": 1.8146085817228954, - "grad_norm": 1.7131381034851074, - "learning_rate": 3.383974898178737e-05, - "loss": 0.354, - "step": 5540 - }, - { - "epoch": 1.817884048476908, - "grad_norm": 3.770197868347168, - "learning_rate": 3.3677546072117025e-05, - "loss": 0.3369, - "step": 5550 - }, - { - "epoch": 1.8211595152309203, - "grad_norm": 0.3596019744873047, - "learning_rate": 3.351553520110053e-05, - "loss": 0.2918, - "step": 5560 - }, - { - "epoch": 1.8244349819849328, - "grad_norm": 1.9685468673706055, - "learning_rate": 3.335371827484523e-05, - "loss": 0.3602, - "step": 5570 - }, - { - "epoch": 1.8277104487389453, - "grad_norm": 0.40883752703666687, - "learning_rate": 3.319209719717669e-05, - "loss": 0.4053, - "step": 5580 - }, - { - "epoch": 1.8309859154929577, - "grad_norm": 2.1049082279205322, - "learning_rate": 3.30306738696162e-05, - "loss": 0.3284, - "step": 5590 - }, - { - "epoch": 1.8342613822469702, - "grad_norm": 1.6273139715194702, - "learning_rate": 3.2869450191358516e-05, - "loss": 0.3108, - "step": 5600 - }, - { - "epoch": 1.8375368490009827, - "grad_norm": 1.0122137069702148, - "learning_rate": 3.2708428059249436e-05, - "loss": 0.3877, - "step": 5610 - }, - { - "epoch": 1.8408123157549952, - "grad_norm": 0.7414371967315674, - "learning_rate": 3.2547609367763496e-05, - "loss": 0.3536, - "step": 5620 - }, - { - "epoch": 1.8440877825090074, - "grad_norm": 2.060434579849243, - "learning_rate": 3.238699600898171e-05, - "loss": 0.3366, - "step": 5630 - }, - { - "epoch": 1.84736324926302, - "grad_norm": 2.741241216659546, - "learning_rate": 3.222658987256926e-05, - "loss": 0.3099, - "step": 5640 - }, - { - "epoch": 1.8506387160170323, - "grad_norm": 2.1917922496795654, - "learning_rate": 3.206639284575334e-05, - "loss": 0.3487, - "step": 5650 - }, - { - "epoch": 1.853914182771045, - "grad_norm": 2.630911111831665, - "learning_rate": 3.190640681330087e-05, - "loss": 0.325, - "step": 5660 - }, - { - "epoch": 1.8571896495250573, - "grad_norm": 5.372201919555664, - "learning_rate": 3.174663365749636e-05, - "loss": 0.3751, - "step": 5670 - }, - { - "epoch": 1.8604651162790697, - "grad_norm": 15.220906257629395, - "learning_rate": 3.158707525811979e-05, - "loss": 0.3475, - "step": 5680 - }, - { - "epoch": 1.8637405830330822, - "grad_norm": 1.3580074310302734, - "learning_rate": 3.1427733492424414e-05, - "loss": 0.3843, - "step": 5690 - }, - { - "epoch": 1.8670160497870947, - "grad_norm": 1.4969313144683838, - "learning_rate": 3.1268610235114765e-05, - "loss": 0.3371, - "step": 5700 - }, - { - "epoch": 1.8702915165411071, - "grad_norm": 5.151036262512207, - "learning_rate": 3.110970735832456e-05, - "loss": 0.3478, - "step": 5710 - }, - { - "epoch": 1.8735669832951194, - "grad_norm": 2.1546502113342285, - "learning_rate": 3.0951026731594635e-05, - "loss": 0.348, - "step": 5720 - }, - { - "epoch": 1.876842450049132, - "grad_norm": 1.2298020124435425, - "learning_rate": 3.079257022185104e-05, - "loss": 0.3526, - "step": 5730 - }, - { - "epoch": 1.8801179168031443, - "grad_norm": 6.7956624031066895, - "learning_rate": 3.0634339693382965e-05, - "loss": 0.3732, - "step": 5740 - }, - { - "epoch": 1.883393383557157, - "grad_norm": 1.4429703950881958, - "learning_rate": 3.047633700782089e-05, - "loss": 0.4111, - "step": 5750 - }, - { - "epoch": 1.8866688503111693, - "grad_norm": 1.1810033321380615, - "learning_rate": 3.0318564024114658e-05, - "loss": 0.3814, - "step": 5760 - }, - { - "epoch": 1.8899443170651817, - "grad_norm": 3.233301877975464, - "learning_rate": 3.016102259851159e-05, - "loss": 0.3621, - "step": 5770 - }, - { - "epoch": 1.8932197838191942, - "grad_norm": 1.6012914180755615, - "learning_rate": 3.000371458453466e-05, - "loss": 0.3093, - "step": 5780 - }, - { - "epoch": 1.8964952505732067, - "grad_norm": 1.3252966403961182, - "learning_rate": 2.9846641832960666e-05, - "loss": 0.3558, - "step": 5790 - }, - { - "epoch": 1.8997707173272191, - "grad_norm": 1.4198555946350098, - "learning_rate": 2.9689806191798496e-05, - "loss": 0.2476, - "step": 5800 - }, - { - "epoch": 1.9030461840812316, - "grad_norm": 0.9728915095329285, - "learning_rate": 2.9533209506267356e-05, - "loss": 0.3518, - "step": 5810 - }, - { - "epoch": 1.906321650835244, - "grad_norm": 1.678391933441162, - "learning_rate": 2.9376853618775052e-05, - "loss": 0.3168, - "step": 5820 - }, - { - "epoch": 1.9095971175892563, - "grad_norm": 15.476072311401367, - "learning_rate": 2.922074036889635e-05, - "loss": 0.3548, - "step": 5830 - }, - { - "epoch": 1.912872584343269, - "grad_norm": 1.5558778047561646, - "learning_rate": 2.9064871593351283e-05, - "loss": 0.3386, - "step": 5840 - }, - { - "epoch": 1.9161480510972813, - "grad_norm": 5.4236650466918945, - "learning_rate": 2.8909249125983572e-05, - "loss": 0.3542, - "step": 5850 - }, - { - "epoch": 1.919423517851294, - "grad_norm": 4.549374580383301, - "learning_rate": 2.875387479773912e-05, - "loss": 0.291, - "step": 5860 - }, - { - "epoch": 1.9226989846053062, - "grad_norm": 6.915854454040527, - "learning_rate": 2.859875043664425e-05, - "loss": 0.4155, - "step": 5870 - }, - { - "epoch": 1.9259744513593187, - "grad_norm": 1.5604161024093628, - "learning_rate": 2.8443877867784496e-05, - "loss": 0.4087, - "step": 5880 - }, - { - "epoch": 1.9292499181133311, - "grad_norm": 1.4447869062423706, - "learning_rate": 2.828925891328287e-05, - "loss": 0.2464, - "step": 5890 - }, - { - "epoch": 1.9325253848673436, - "grad_norm": 8.18412971496582, - "learning_rate": 2.8134895392278614e-05, - "loss": 0.3266, - "step": 5900 - }, - { - "epoch": 1.935800851621356, - "grad_norm": 1.2699623107910156, - "learning_rate": 2.7980789120905677e-05, - "loss": 0.357, - "step": 5910 - }, - { - "epoch": 1.9390763183753685, - "grad_norm": 2.9487576484680176, - "learning_rate": 2.7826941912271355e-05, - "loss": 0.3178, - "step": 5920 - }, - { - "epoch": 1.942351785129381, - "grad_norm": 3.3336455821990967, - "learning_rate": 2.767335557643509e-05, - "loss": 0.3184, - "step": 5930 - }, - { - "epoch": 1.9456272518833932, - "grad_norm": 2.4859960079193115, - "learning_rate": 2.7520031920386945e-05, - "loss": 0.431, - "step": 5940 - }, - { - "epoch": 1.948902718637406, - "grad_norm": 1.690781593322754, - "learning_rate": 2.7366972748026577e-05, - "loss": 0.3032, - "step": 5950 - }, - { - "epoch": 1.9521781853914182, - "grad_norm": 6.6911940574646, - "learning_rate": 2.721417986014192e-05, - "loss": 0.4245, - "step": 5960 - }, - { - "epoch": 1.9554536521454309, - "grad_norm": 1.7215849161148071, - "learning_rate": 2.706165505438787e-05, - "loss": 0.327, - "step": 5970 - }, - { - "epoch": 1.9587291188994431, - "grad_norm": 1.1351230144500732, - "learning_rate": 2.6909400125265392e-05, - "loss": 0.3735, - "step": 5980 - }, - { - "epoch": 1.9620045856534556, - "grad_norm": 7.951147079467773, - "learning_rate": 2.6757416864100173e-05, - "loss": 0.3993, - "step": 5990 - }, - { - "epoch": 1.965280052407468, - "grad_norm": 5.089042663574219, - "learning_rate": 2.6605707059021724e-05, - "loss": 0.407, - "step": 6000 - }, - { - "epoch": 1.9685555191614805, - "grad_norm": 3.3219149112701416, - "learning_rate": 2.645427249494219e-05, - "loss": 0.3186, - "step": 6010 - }, - { - "epoch": 1.971830985915493, - "grad_norm": 0.9335862398147583, - "learning_rate": 2.6303114953535445e-05, - "loss": 0.3418, - "step": 6020 - }, - { - "epoch": 1.9751064526695052, - "grad_norm": 0.5746390223503113, - "learning_rate": 2.6152236213216176e-05, - "loss": 0.3526, - "step": 6030 - }, - { - "epoch": 1.978381919423518, - "grad_norm": 3.2865259647369385, - "learning_rate": 2.6001638049118725e-05, - "loss": 0.3692, - "step": 6040 - }, - { - "epoch": 1.9816573861775302, - "grad_norm": 1.9871715307235718, - "learning_rate": 2.5851322233076536e-05, - "loss": 0.3449, - "step": 6050 - }, - { - "epoch": 1.9849328529315429, - "grad_norm": 4.637392520904541, - "learning_rate": 2.5701290533601107e-05, - "loss": 0.3299, - "step": 6060 - }, - { - "epoch": 1.9882083196855551, - "grad_norm": 9.569966316223145, - "learning_rate": 2.5551544715861104e-05, - "loss": 0.3677, - "step": 6070 - }, - { - "epoch": 1.9914837864395678, - "grad_norm": 3.9491708278656006, - "learning_rate": 2.540208654166184e-05, - "loss": 0.3889, - "step": 6080 - }, - { - "epoch": 1.99475925319358, - "grad_norm": 4.674814224243164, - "learning_rate": 2.5252917769424377e-05, - "loss": 0.315, - "step": 6090 - }, - { - "epoch": 1.9980347199475925, - "grad_norm": 2.508333921432495, - "learning_rate": 2.5104040154164853e-05, - "loss": 0.3957, - "step": 6100 - }, - { - "epoch": 2.001310186701605, - "grad_norm": 1.0377638339996338, - "learning_rate": 2.495545544747385e-05, - "loss": 0.3008, - "step": 6110 - }, - { - "epoch": 2.0045856534556172, - "grad_norm": 1.2737836837768555, - "learning_rate": 2.4807165397495807e-05, - "loss": 0.3732, - "step": 6120 - }, - { - "epoch": 2.00786112020963, - "grad_norm": 3.8265914916992188, - "learning_rate": 2.4659171748908456e-05, - "loss": 0.3042, - "step": 6130 - }, - { - "epoch": 2.011136586963642, - "grad_norm": 6.77216100692749, - "learning_rate": 2.4511476242902248e-05, - "loss": 0.2719, - "step": 6140 - }, - { - "epoch": 2.014412053717655, - "grad_norm": 3.4394800662994385, - "learning_rate": 2.4364080617159886e-05, - "loss": 0.2296, - "step": 6150 - }, - { - "epoch": 2.017687520471667, - "grad_norm": 1.8297233581542969, - "learning_rate": 2.4216986605835952e-05, - "loss": 0.2752, - "step": 6160 - }, - { - "epoch": 2.02096298722568, - "grad_norm": 0.9549896121025085, - "learning_rate": 2.407019593953634e-05, - "loss": 0.294, - "step": 6170 - }, - { - "epoch": 2.024238453979692, - "grad_norm": 1.4905319213867188, - "learning_rate": 2.3923710345298077e-05, - "loss": 0.3021, - "step": 6180 - }, - { - "epoch": 2.0275139207337047, - "grad_norm": 0.631741464138031, - "learning_rate": 2.3777531546568933e-05, - "loss": 0.3547, - "step": 6190 - }, - { - "epoch": 2.030789387487717, - "grad_norm": 1.5282114744186401, - "learning_rate": 2.363166126318708e-05, - "loss": 0.3025, - "step": 6200 - }, - { - "epoch": 2.0340648542417297, - "grad_norm": 3.3751862049102783, - "learning_rate": 2.348610121136094e-05, - "loss": 0.353, - "step": 6210 - }, - { - "epoch": 2.037340320995742, - "grad_norm": 1.3393089771270752, - "learning_rate": 2.3340853103648945e-05, - "loss": 0.217, - "step": 6220 - }, - { - "epoch": 2.040615787749754, - "grad_norm": 1.0742156505584717, - "learning_rate": 2.3195918648939465e-05, - "loss": 0.3253, - "step": 6230 - }, - { - "epoch": 2.043891254503767, - "grad_norm": 0.8612537980079651, - "learning_rate": 2.305129955243059e-05, - "loss": 0.309, - "step": 6240 - }, - { - "epoch": 2.047166721257779, - "grad_norm": 2.400435447692871, - "learning_rate": 2.290699751561012e-05, - "loss": 0.2939, - "step": 6250 - }, - { - "epoch": 2.050442188011792, - "grad_norm": 7.121722221374512, - "learning_rate": 2.2763014236235636e-05, - "loss": 0.2971, - "step": 6260 - }, - { - "epoch": 2.053717654765804, - "grad_norm": 1.2473032474517822, - "learning_rate": 2.2619351408314304e-05, - "loss": 0.2628, - "step": 6270 - }, - { - "epoch": 2.0569931215198167, - "grad_norm": 15.246943473815918, - "learning_rate": 2.2476010722083197e-05, - "loss": 0.3224, - "step": 6280 - }, - { - "epoch": 2.060268588273829, - "grad_norm": 1.3141411542892456, - "learning_rate": 2.2332993863989265e-05, - "loss": 0.2542, - "step": 6290 - }, - { - "epoch": 2.0635440550278417, - "grad_norm": 1.2878804206848145, - "learning_rate": 2.2190302516669493e-05, - "loss": 0.2441, - "step": 6300 - }, - { - "epoch": 2.066819521781854, - "grad_norm": 1.4133859872817993, - "learning_rate": 2.2047938358931143e-05, - "loss": 0.2866, - "step": 6310 - }, - { - "epoch": 2.070094988535866, - "grad_norm": 1.9740748405456543, - "learning_rate": 2.1905903065731974e-05, - "loss": 0.2712, - "step": 6320 - }, - { - "epoch": 2.073370455289879, - "grad_norm": 1.2161868810653687, - "learning_rate": 2.176419830816062e-05, - "loss": 0.3378, - "step": 6330 - }, - { - "epoch": 2.076645922043891, - "grad_norm": 0.7021055221557617, - "learning_rate": 2.1622825753416797e-05, - "loss": 0.3295, - "step": 6340 - }, - { - "epoch": 2.079921388797904, - "grad_norm": 0.6557903289794922, - "learning_rate": 2.148178706479176e-05, - "loss": 0.2789, - "step": 6350 - }, - { - "epoch": 2.083196855551916, - "grad_norm": 48.70767593383789, - "learning_rate": 2.134108390164878e-05, - "loss": 0.3723, - "step": 6360 - }, - { - "epoch": 2.0864723223059287, - "grad_norm": 4.277552604675293, - "learning_rate": 2.1200717919403474e-05, - "loss": 0.3226, - "step": 6370 - }, - { - "epoch": 2.089747789059941, - "grad_norm": 3.259768009185791, - "learning_rate": 2.1060690769504505e-05, - "loss": 0.3296, - "step": 6380 - }, - { - "epoch": 2.0930232558139537, - "grad_norm": 1.5467828512191772, - "learning_rate": 2.092100409941406e-05, - "loss": 0.345, - "step": 6390 - }, - { - "epoch": 2.096298722567966, - "grad_norm": 2.7095184326171875, - "learning_rate": 2.0781659552588444e-05, - "loss": 0.3672, - "step": 6400 - }, - { - "epoch": 2.0995741893219786, - "grad_norm": 1.855701208114624, - "learning_rate": 2.06426587684588e-05, - "loss": 0.3221, - "step": 6410 - }, - { - "epoch": 2.102849656075991, - "grad_norm": 5.6737961769104, - "learning_rate": 2.0504003382411768e-05, - "loss": 0.2376, - "step": 6420 - }, - { - "epoch": 2.106125122830003, - "grad_norm": 1.2713319063186646, - "learning_rate": 2.036569502577033e-05, - "loss": 0.296, - "step": 6430 - }, - { - "epoch": 2.1094005895840158, - "grad_norm": 1.7682929039001465, - "learning_rate": 2.0227735325774515e-05, - "loss": 0.2434, - "step": 6440 - }, - { - "epoch": 2.112676056338028, - "grad_norm": 3.1154940128326416, - "learning_rate": 2.0090125905562295e-05, - "loss": 0.2803, - "step": 6450 - }, - { - "epoch": 2.1159515230920407, - "grad_norm": 3.6548871994018555, - "learning_rate": 1.995286838415057e-05, - "loss": 0.2419, - "step": 6460 - }, - { - "epoch": 2.119226989846053, - "grad_norm": 1.9502896070480347, - "learning_rate": 1.98159643764159e-05, - "loss": 0.3509, - "step": 6470 - }, - { - "epoch": 2.1225024566000656, - "grad_norm": 1.707944631576538, - "learning_rate": 1.9679415493075786e-05, - "loss": 0.2906, - "step": 6480 - }, - { - "epoch": 2.125777923354078, - "grad_norm": 0.6232993602752686, - "learning_rate": 1.9543223340669542e-05, - "loss": 0.3726, - "step": 6490 - }, - { - "epoch": 2.1290533901080906, - "grad_norm": 1.6979426145553589, - "learning_rate": 1.94073895215394e-05, - "loss": 0.2825, - "step": 6500 - }, - { - "epoch": 2.132328856862103, - "grad_norm": 5.000602722167969, - "learning_rate": 1.9271915633811717e-05, - "loss": 0.3756, - "step": 6510 - }, - { - "epoch": 2.135604323616115, - "grad_norm": 1.5471789836883545, - "learning_rate": 1.913680327137812e-05, - "loss": 0.3078, - "step": 6520 - }, - { - "epoch": 2.1388797903701278, - "grad_norm": 1.6782714128494263, - "learning_rate": 1.900205402387685e-05, - "loss": 0.279, - "step": 6530 - }, - { - "epoch": 2.14215525712414, - "grad_norm": 4.149227142333984, - "learning_rate": 1.8867669476673914e-05, - "loss": 0.2807, - "step": 6540 - }, - { - "epoch": 2.1454307238781527, - "grad_norm": 1.352228045463562, - "learning_rate": 1.873365121084453e-05, - "loss": 0.318, - "step": 6550 - }, - { - "epoch": 2.148706190632165, - "grad_norm": 1.5985099077224731, - "learning_rate": 1.8600000803154565e-05, - "loss": 0.2915, - "step": 6560 - }, - { - "epoch": 2.1519816573861776, - "grad_norm": 1.272507905960083, - "learning_rate": 1.8466719826041796e-05, - "loss": 0.3084, - "step": 6570 - }, - { - "epoch": 2.15525712414019, - "grad_norm": 3.840996742248535, - "learning_rate": 1.8333809847597642e-05, - "loss": 0.3776, - "step": 6580 - }, - { - "epoch": 2.1585325908942026, - "grad_norm": 2.3710379600524902, - "learning_rate": 1.8201272431548628e-05, - "loss": 0.3641, - "step": 6590 - }, - { - "epoch": 2.161808057648215, - "grad_norm": 1.2817639112472534, - "learning_rate": 1.8069109137237827e-05, - "loss": 0.3048, - "step": 6600 - }, - { - "epoch": 2.1650835244022275, - "grad_norm": 9.000032424926758, - "learning_rate": 1.793732151960682e-05, - "loss": 0.3471, - "step": 6610 - }, - { - "epoch": 2.1683589911562398, - "grad_norm": 3.148275375366211, - "learning_rate": 1.7805911129177117e-05, - "loss": 0.2966, - "step": 6620 - }, - { - "epoch": 2.171634457910252, - "grad_norm": 2.1067912578582764, - "learning_rate": 1.7674879512032116e-05, - "loss": 0.2991, - "step": 6630 - }, - { - "epoch": 2.1749099246642647, - "grad_norm": 2.2540812492370605, - "learning_rate": 1.7544228209798793e-05, - "loss": 0.2892, - "step": 6640 - }, - { - "epoch": 2.178185391418277, - "grad_norm": 0.7971101403236389, - "learning_rate": 1.7413958759629574e-05, - "loss": 0.2391, - "step": 6650 - }, - { - "epoch": 2.1814608581722896, - "grad_norm": 3.7163023948669434, - "learning_rate": 1.728407269418437e-05, - "loss": 0.2766, - "step": 6660 - }, - { - "epoch": 2.184736324926302, - "grad_norm": 4.00792121887207, - "learning_rate": 1.7154571541612342e-05, - "loss": 0.2442, - "step": 6670 - }, - { - "epoch": 2.1880117916803146, - "grad_norm": 1.4201942682266235, - "learning_rate": 1.7025456825534097e-05, - "loss": 0.2123, - "step": 6680 - }, - { - "epoch": 2.191287258434327, - "grad_norm": 3.1648571491241455, - "learning_rate": 1.6896730065023747e-05, - "loss": 0.3516, - "step": 6690 - }, - { - "epoch": 2.1945627251883395, - "grad_norm": 2.2273266315460205, - "learning_rate": 1.6768392774590846e-05, - "loss": 0.2921, - "step": 6700 - }, - { - "epoch": 2.1978381919423517, - "grad_norm": 0.6129598617553711, - "learning_rate": 1.6640446464162857e-05, - "loss": 0.2093, - "step": 6710 - }, - { - "epoch": 2.2011136586963644, - "grad_norm": 11.434572219848633, - "learning_rate": 1.6512892639067152e-05, - "loss": 0.2708, - "step": 6720 - }, - { - "epoch": 2.2043891254503767, - "grad_norm": 0.8874580264091492, - "learning_rate": 1.6385732800013453e-05, - "loss": 0.28, - "step": 6730 - }, - { - "epoch": 2.207664592204389, - "grad_norm": 1.1818125247955322, - "learning_rate": 1.6258968443076078e-05, - "loss": 0.2914, - "step": 6740 - }, - { - "epoch": 2.2109400589584016, - "grad_norm": 4.780377388000488, - "learning_rate": 1.6132601059676376e-05, - "loss": 0.3182, - "step": 6750 - }, - { - "epoch": 2.214215525712414, - "grad_norm": 2.984995126724243, - "learning_rate": 1.6006632136565236e-05, - "loss": 0.3616, - "step": 6760 - }, - { - "epoch": 2.2174909924664266, - "grad_norm": 4.2235493659973145, - "learning_rate": 1.5881063155805437e-05, - "loss": 0.2929, - "step": 6770 - }, - { - "epoch": 2.220766459220439, - "grad_norm": 8.98217487335205, - "learning_rate": 1.5755895594754405e-05, - "loss": 0.2668, - "step": 6780 - }, - { - "epoch": 2.2240419259744515, - "grad_norm": 0.5278671979904175, - "learning_rate": 1.5631130926046744e-05, - "loss": 0.2484, - "step": 6790 - }, - { - "epoch": 2.2273173927284637, - "grad_norm": 2.6351490020751953, - "learning_rate": 1.5506770617576804e-05, - "loss": 0.3244, - "step": 6800 - }, - { - "epoch": 2.2305928594824764, - "grad_norm": 11.68897819519043, - "learning_rate": 1.5382816132481643e-05, - "loss": 0.2342, - "step": 6810 - }, - { - "epoch": 2.2338683262364887, - "grad_norm": 1.1514060497283936, - "learning_rate": 1.5259268929123573e-05, - "loss": 0.2925, - "step": 6820 - }, - { - "epoch": 2.2371437929905014, - "grad_norm": 0.7407781481742859, - "learning_rate": 1.5136130461073206e-05, - "loss": 0.3389, - "step": 6830 - }, - { - "epoch": 2.2404192597445136, - "grad_norm": 1.688888669013977, - "learning_rate": 1.5013402177092195e-05, - "loss": 0.2664, - "step": 6840 - }, - { - "epoch": 2.243694726498526, - "grad_norm": 2.2012505531311035, - "learning_rate": 1.4891085521116254e-05, - "loss": 0.295, - "step": 6850 - }, - { - "epoch": 2.2469701932525386, - "grad_norm": 2.0733485221862793, - "learning_rate": 1.476918193223823e-05, - "loss": 0.2716, - "step": 6860 - }, - { - "epoch": 2.250245660006551, - "grad_norm": 9.919839859008789, - "learning_rate": 1.4647692844691008e-05, - "loss": 0.3051, - "step": 6870 - }, - { - "epoch": 2.2535211267605635, - "grad_norm": 1.6594527959823608, - "learning_rate": 1.4526619687830822e-05, - "loss": 0.2582, - "step": 6880 - }, - { - "epoch": 2.2567965935145757, - "grad_norm": 6.103255271911621, - "learning_rate": 1.4405963886120354e-05, - "loss": 0.3067, - "step": 6890 - }, - { - "epoch": 2.2600720602685884, - "grad_norm": 2.400704860687256, - "learning_rate": 1.4285726859111897e-05, - "loss": 0.2811, - "step": 6900 - }, - { - "epoch": 2.2633475270226007, - "grad_norm": 9.178129196166992, - "learning_rate": 1.4165910021430818e-05, - "loss": 0.3131, - "step": 6910 - }, - { - "epoch": 2.2666229937766134, - "grad_norm": 0.9883064031600952, - "learning_rate": 1.4046514782758768e-05, - "loss": 0.3508, - "step": 6920 - }, - { - "epoch": 2.2698984605306256, - "grad_norm": 0.7798057794570923, - "learning_rate": 1.3927542547817219e-05, - "loss": 0.2537, - "step": 6930 - }, - { - "epoch": 2.2731739272846383, - "grad_norm": 37.888084411621094, - "learning_rate": 1.380899471635081e-05, - "loss": 0.371, - "step": 6940 - }, - { - "epoch": 2.2764493940386505, - "grad_norm": 4.047865867614746, - "learning_rate": 1.3690872683110967e-05, - "loss": 0.2141, - "step": 6950 - }, - { - "epoch": 2.279724860792663, - "grad_norm": 0.418952614068985, - "learning_rate": 1.3573177837839495e-05, - "loss": 0.2475, - "step": 6960 - }, - { - "epoch": 2.2830003275466755, - "grad_norm": 0.8674260973930359, - "learning_rate": 1.3455911565252183e-05, - "loss": 0.303, - "step": 6970 - }, - { - "epoch": 2.2862757943006877, - "grad_norm": 4.694146633148193, - "learning_rate": 1.3339075245022497e-05, - "loss": 0.3008, - "step": 6980 - }, - { - "epoch": 2.2895512610547004, - "grad_norm": 5.114990234375, - "learning_rate": 1.3222670251765474e-05, - "loss": 0.357, - "step": 6990 - }, - { - "epoch": 2.2928267278087127, - "grad_norm": 1.7379510402679443, - "learning_rate": 1.310669795502133e-05, - "loss": 0.3583, - "step": 7000 - }, - { - "epoch": 2.2961021945627254, - "grad_norm": 6.100559234619141, - "learning_rate": 1.2991159719239582e-05, - "loss": 0.3097, - "step": 7010 - }, - { - "epoch": 2.2993776613167376, - "grad_norm": 1.6555323600769043, - "learning_rate": 1.2876056903762857e-05, - "loss": 0.3146, - "step": 7020 - }, - { - "epoch": 2.30265312807075, - "grad_norm": 1.2503982782363892, - "learning_rate": 1.2761390862810907e-05, - "loss": 0.3234, - "step": 7030 - }, - { - "epoch": 2.3059285948247625, - "grad_norm": 4.169071197509766, - "learning_rate": 1.2647162945464719e-05, - "loss": 0.2697, - "step": 7040 - }, - { - "epoch": 2.3092040615787752, - "grad_norm": 1.4650436639785767, - "learning_rate": 1.253337449565059e-05, - "loss": 0.2341, - "step": 7050 - }, - { - "epoch": 2.3124795283327875, - "grad_norm": 4.525272846221924, - "learning_rate": 1.242002685212441e-05, - "loss": 0.2899, - "step": 7060 - }, - { - "epoch": 2.3157549950867997, - "grad_norm": 3.5864546298980713, - "learning_rate": 1.2307121348455792e-05, - "loss": 0.3192, - "step": 7070 - }, - { - "epoch": 2.3190304618408124, - "grad_norm": 2.348140239715576, - "learning_rate": 1.2194659313012447e-05, - "loss": 0.2387, - "step": 7080 - }, - { - "epoch": 2.3223059285948247, - "grad_norm": 3.959228277206421, - "learning_rate": 1.2082642068944594e-05, - "loss": 0.3497, - "step": 7090 - }, - { - "epoch": 2.3255813953488373, - "grad_norm": 1.4259454011917114, - "learning_rate": 1.1971070934169265e-05, - "loss": 0.3504, - "step": 7100 - }, - { - "epoch": 2.3288568621028496, - "grad_norm": 9.238195419311523, - "learning_rate": 1.1859947221354939e-05, - "loss": 0.3866, - "step": 7110 - }, - { - "epoch": 2.3321323288568623, - "grad_norm": 0.8945813775062561, - "learning_rate": 1.174927223790604e-05, - "loss": 0.2852, - "step": 7120 - }, - { - "epoch": 2.3354077956108745, - "grad_norm": 1.798998475074768, - "learning_rate": 1.1639047285947535e-05, - "loss": 0.312, - "step": 7130 - }, - { - "epoch": 2.3386832623648868, - "grad_norm": 3.453458070755005, - "learning_rate": 1.1529273662309625e-05, - "loss": 0.2563, - "step": 7140 - }, - { - "epoch": 2.3419587291188995, - "grad_norm": 6.823653221130371, - "learning_rate": 1.1419952658512495e-05, - "loss": 0.3681, - "step": 7150 - }, - { - "epoch": 2.3452341958729117, - "grad_norm": 15.647924423217773, - "learning_rate": 1.1311085560751168e-05, - "loss": 0.304, - "step": 7160 - }, - { - "epoch": 2.3485096626269244, - "grad_norm": 10.072633743286133, - "learning_rate": 1.1202673649880264e-05, - "loss": 0.322, - "step": 7170 - }, - { - "epoch": 2.3517851293809366, - "grad_norm": 0.6596316695213318, - "learning_rate": 1.1094718201399002e-05, - "loss": 0.2964, - "step": 7180 - }, - { - "epoch": 2.3550605961349493, - "grad_norm": 2.9951183795928955, - "learning_rate": 1.0987220485436234e-05, - "loss": 0.2571, - "step": 7190 - }, - { - "epoch": 2.3583360628889616, - "grad_norm": 2.3419599533081055, - "learning_rate": 1.0880181766735353e-05, - "loss": 0.2194, - "step": 7200 - }, - { - "epoch": 2.3616115296429743, - "grad_norm": 3.6008496284484863, - "learning_rate": 1.0773603304639595e-05, - "loss": 0.234, - "step": 7210 - }, - { - "epoch": 2.3648869963969865, - "grad_norm": 0.6958408951759338, - "learning_rate": 1.0667486353077133e-05, - "loss": 0.4206, - "step": 7220 - }, - { - "epoch": 2.368162463150999, - "grad_norm": 1.5056467056274414, - "learning_rate": 1.0561832160546237e-05, - "loss": 0.2707, - "step": 7230 - }, - { - "epoch": 2.3714379299050115, - "grad_norm": 1.265770673751831, - "learning_rate": 1.0456641970100795e-05, - "loss": 0.2372, - "step": 7240 - }, - { - "epoch": 2.3747133966590237, - "grad_norm": 6.484124183654785, - "learning_rate": 1.0351917019335472e-05, - "loss": 0.3406, - "step": 7250 - }, - { - "epoch": 2.3779888634130364, - "grad_norm": 4.233597278594971, - "learning_rate": 1.0247658540371318e-05, - "loss": 0.2632, - "step": 7260 - }, - { - "epoch": 2.3812643301670486, - "grad_norm": 0.9953312277793884, - "learning_rate": 1.0143867759841164e-05, - "loss": 0.3468, - "step": 7270 - }, - { - "epoch": 2.3845397969210613, - "grad_norm": 2.05120849609375, - "learning_rate": 1.0040545898875209e-05, - "loss": 0.3734, - "step": 7280 - }, - { - "epoch": 2.3878152636750736, - "grad_norm": 5.327258586883545, - "learning_rate": 9.937694173086738e-06, - "loss": 0.2544, - "step": 7290 - }, - { - "epoch": 2.3910907304290863, - "grad_norm": 2.6873862743377686, - "learning_rate": 9.835313792557644e-06, - "loss": 0.2534, - "step": 7300 - }, - { - "epoch": 2.3943661971830985, - "grad_norm": 11.175503730773926, - "learning_rate": 9.733405961824398e-06, - "loss": 0.2349, - "step": 7310 - }, - { - "epoch": 2.397641663937111, - "grad_norm": 1.960609793663025, - "learning_rate": 9.63197187986376e-06, - "loss": 0.3215, - "step": 7320 - }, - { - "epoch": 2.4009171306911234, - "grad_norm": 3.0660367012023926, - "learning_rate": 9.531012740078626e-06, - "loss": 0.2865, - "step": 7330 - }, - { - "epoch": 2.404192597445136, - "grad_norm": 5.0894999504089355, - "learning_rate": 9.430529730284138e-06, - "loss": 0.2524, - "step": 7340 - }, - { - "epoch": 2.4074680641991484, - "grad_norm": 3.0109808444976807, - "learning_rate": 9.330524032693561e-06, - "loss": 0.299, - "step": 7350 - }, - { - "epoch": 2.4107435309531606, - "grad_norm": 1.1892986297607422, - "learning_rate": 9.230996823904475e-06, - "loss": 0.3104, - "step": 7360 - }, - { - "epoch": 2.4140189977071733, - "grad_norm": 5.921483993530273, - "learning_rate": 9.131949274884877e-06, - "loss": 0.3095, - "step": 7370 - }, - { - "epoch": 2.4172944644611856, - "grad_norm": 4.141952037811279, - "learning_rate": 9.033382550959402e-06, - "loss": 0.41, - "step": 7380 - }, - { - "epoch": 2.4205699312151983, - "grad_norm": 4.587307453155518, - "learning_rate": 8.93529781179569e-06, - "loss": 0.2939, - "step": 7390 - }, - { - "epoch": 2.4238453979692105, - "grad_norm": 5.876664638519287, - "learning_rate": 8.837696211390595e-06, - "loss": 0.295, - "step": 7400 - }, - { - "epoch": 2.427120864723223, - "grad_norm": 1.5953419208526611, - "learning_rate": 8.740578898056762e-06, - "loss": 0.3099, - "step": 7410 - }, - { - "epoch": 2.4303963314772354, - "grad_norm": 2.139911413192749, - "learning_rate": 8.643947014409071e-06, - "loss": 0.2697, - "step": 7420 - }, - { - "epoch": 2.433671798231248, - "grad_norm": 1.083020567893982, - "learning_rate": 8.547801697351071e-06, - "loss": 0.2922, - "step": 7430 - }, - { - "epoch": 2.4369472649852604, - "grad_norm": 1.5458956956863403, - "learning_rate": 8.452144078061818e-06, - "loss": 0.2796, - "step": 7440 - }, - { - "epoch": 2.440222731739273, - "grad_norm": 4.247933864593506, - "learning_rate": 8.356975281982382e-06, - "loss": 0.3526, - "step": 7450 - }, - { - "epoch": 2.4434981984932853, - "grad_norm": 6.735934257507324, - "learning_rate": 8.262296428802724e-06, - "loss": 0.415, - "step": 7460 - }, - { - "epoch": 2.4467736652472976, - "grad_norm": 1.9846491813659668, - "learning_rate": 8.168108632448457e-06, - "loss": 0.3161, - "step": 7470 - }, - { - "epoch": 2.4500491320013102, - "grad_norm": 1.4496264457702637, - "learning_rate": 8.074413001067738e-06, - "loss": 0.2456, - "step": 7480 - }, - { - "epoch": 2.4533245987553225, - "grad_norm": 7.2307915687561035, - "learning_rate": 7.981210637018316e-06, - "loss": 0.3167, - "step": 7490 - }, - { - "epoch": 2.456600065509335, - "grad_norm": 0.8901455998420715, - "learning_rate": 7.888502636854411e-06, - "loss": 0.2745, - "step": 7500 - }, - { - "epoch": 2.4598755322633474, - "grad_norm": 4.472309112548828, - "learning_rate": 7.79629009131398e-06, - "loss": 0.2055, - "step": 7510 - }, - { - "epoch": 2.46315099901736, - "grad_norm": 3.1325817108154297, - "learning_rate": 7.704574085305787e-06, - "loss": 0.2287, - "step": 7520 - }, - { - "epoch": 2.4664264657713724, - "grad_norm": 1.6126749515533447, - "learning_rate": 7.613355697896602e-06, - "loss": 0.2458, - "step": 7530 - }, - { - "epoch": 2.469701932525385, - "grad_norm": 3.0331647396087646, - "learning_rate": 7.522636002298644e-06, - "loss": 0.3464, - "step": 7540 - }, - { - "epoch": 2.4729773992793973, - "grad_norm": 2.8420960903167725, - "learning_rate": 7.432416065856795e-06, - "loss": 0.2032, - "step": 7550 - }, - { - "epoch": 2.47625286603341, - "grad_norm": 3.000521183013916, - "learning_rate": 7.342696950036182e-06, - "loss": 0.2958, - "step": 7560 - }, - { - "epoch": 2.4795283327874222, - "grad_norm": 2.536396026611328, - "learning_rate": 7.253479710409583e-06, - "loss": 0.2764, - "step": 7570 - }, - { - "epoch": 2.4828037995414345, - "grad_norm": 3.343484401702881, - "learning_rate": 7.164765396645051e-06, - "loss": 0.2903, - "step": 7580 - }, - { - "epoch": 2.486079266295447, - "grad_norm": 1.3836356401443481, - "learning_rate": 7.0765550524936e-06, - "loss": 0.2862, - "step": 7590 - }, - { - "epoch": 2.4893547330494594, - "grad_norm": 1.4416383504867554, - "learning_rate": 6.988849715776818e-06, - "loss": 0.3072, - "step": 7600 - }, - { - "epoch": 2.492630199803472, - "grad_norm": 0.6436456441879272, - "learning_rate": 6.9016504183748e-06, - "loss": 0.2443, - "step": 7610 - }, - { - "epoch": 2.4959056665574844, - "grad_norm": 2.1578516960144043, - "learning_rate": 6.81495818621391e-06, - "loss": 0.3384, - "step": 7620 - }, - { - "epoch": 2.499181133311497, - "grad_norm": 1.2418651580810547, - "learning_rate": 6.728774039254693e-06, - "loss": 0.2872, - "step": 7630 - }, - { - "epoch": 2.5024566000655093, - "grad_norm": 9.140596389770508, - "learning_rate": 6.643098991479979e-06, - "loss": 0.4059, - "step": 7640 - }, - { - "epoch": 2.5057320668195215, - "grad_norm": 1.7317625284194946, - "learning_rate": 6.557934050882847e-06, - "loss": 0.2632, - "step": 7650 - }, - { - "epoch": 2.5090075335735342, - "grad_norm": 1.2443095445632935, - "learning_rate": 6.473280219454836e-06, - "loss": 0.3115, - "step": 7660 - }, - { - "epoch": 2.512283000327547, - "grad_norm": 4.098364353179932, - "learning_rate": 6.389138493174118e-06, - "loss": 0.3303, - "step": 7670 - }, - { - "epoch": 2.515558467081559, - "grad_norm": 2.1588313579559326, - "learning_rate": 6.305509861993775e-06, - "loss": 0.3233, - "step": 7680 - }, - { - "epoch": 2.5188339338355714, - "grad_norm": 0.9326940178871155, - "learning_rate": 6.22239530983022e-06, - "loss": 0.2864, - "step": 7690 - }, - { - "epoch": 2.522109400589584, - "grad_norm": 1.239442229270935, - "learning_rate": 6.139795814551486e-06, - "loss": 0.3132, - "step": 7700 - }, - { - "epoch": 2.5253848673435964, - "grad_norm": 5.176858901977539, - "learning_rate": 6.057712347965882e-06, - "loss": 0.3617, - "step": 7710 - }, - { - "epoch": 2.528660334097609, - "grad_norm": 1.6761887073516846, - "learning_rate": 5.976145875810474e-06, - "loss": 0.3024, - "step": 7720 - }, - { - "epoch": 2.5319358008516213, - "grad_norm": 6.175143718719482, - "learning_rate": 5.895097357739665e-06, - "loss": 0.3668, - "step": 7730 - }, - { - "epoch": 2.535211267605634, - "grad_norm": 2.1642215251922607, - "learning_rate": 5.814567747314048e-06, - "loss": 0.2435, - "step": 7740 - }, - { - "epoch": 2.5384867343596462, - "grad_norm": 1.4389464855194092, - "learning_rate": 5.734557991989053e-06, - "loss": 0.2986, - "step": 7750 - }, - { - "epoch": 2.5417622011136585, - "grad_norm": 1.6104439496994019, - "learning_rate": 5.655069033103899e-06, - "loss": 0.2866, - "step": 7760 - }, - { - "epoch": 2.545037667867671, - "grad_norm": 3.6179299354553223, - "learning_rate": 5.576101805870449e-06, - "loss": 0.376, - "step": 7770 - }, - { - "epoch": 2.548313134621684, - "grad_norm": 5.102657794952393, - "learning_rate": 5.497657239362225e-06, - "loss": 0.2883, - "step": 7780 - }, - { - "epoch": 2.551588601375696, - "grad_norm": 9.35090446472168, - "learning_rate": 5.4197362565035435e-06, - "loss": 0.3334, - "step": 7790 - }, - { - "epoch": 2.5548640681297083, - "grad_norm": 4.62729549407959, - "learning_rate": 5.342339774058502e-06, - "loss": 0.2948, - "step": 7800 - }, - { - "epoch": 2.558139534883721, - "grad_norm": 2.2797014713287354, - "learning_rate": 5.26546870262038e-06, - "loss": 0.3549, - "step": 7810 - }, - { - "epoch": 2.5614150016377333, - "grad_norm": 1.8159294128417969, - "learning_rate": 5.189123946600816e-06, - "loss": 0.3229, - "step": 7820 - }, - { - "epoch": 2.564690468391746, - "grad_norm": 2.3214476108551025, - "learning_rate": 5.113306404219137e-06, - "loss": 0.2853, - "step": 7830 - }, - { - "epoch": 2.567965935145758, - "grad_norm": 4.002878189086914, - "learning_rate": 5.038016967491887e-06, - "loss": 0.3169, - "step": 7840 - }, - { - "epoch": 2.571241401899771, - "grad_norm": 3.444040060043335, - "learning_rate": 4.963256522222309e-06, - "loss": 0.2777, - "step": 7850 - }, - { - "epoch": 2.574516868653783, - "grad_norm": 2.306748628616333, - "learning_rate": 4.889025947989817e-06, - "loss": 0.2988, - "step": 7860 - }, - { - "epoch": 2.5777923354077954, - "grad_norm": 4.786330699920654, - "learning_rate": 4.8153261181398125e-06, - "loss": 0.3133, - "step": 7870 - }, - { - "epoch": 2.581067802161808, - "grad_norm": 1.5634129047393799, - "learning_rate": 4.742157899773275e-06, - "loss": 0.2964, - "step": 7880 - }, - { - "epoch": 2.584343268915821, - "grad_norm": 1.6177599430084229, - "learning_rate": 4.669522153736655e-06, - "loss": 0.3588, - "step": 7890 - }, - { - "epoch": 2.587618735669833, - "grad_norm": 4.024510383605957, - "learning_rate": 4.597419734611669e-06, - "loss": 0.2645, - "step": 7900 - }, - { - "epoch": 2.5908942024238453, - "grad_norm": 3.0946905612945557, - "learning_rate": 4.52585149070528e-06, - "loss": 0.2554, - "step": 7910 - }, - { - "epoch": 2.594169669177858, - "grad_norm": 1.6684547662734985, - "learning_rate": 4.454818264039757e-06, - "loss": 0.2898, - "step": 7920 - }, - { - "epoch": 2.59744513593187, - "grad_norm": 2.170660972595215, - "learning_rate": 4.384320890342664e-06, - "loss": 0.2824, - "step": 7930 - }, - { - "epoch": 2.600720602685883, - "grad_norm": 1.775374412536621, - "learning_rate": 4.314360199037143e-06, - "loss": 0.2607, - "step": 7940 - }, - { - "epoch": 2.603996069439895, - "grad_norm": 2.0360000133514404, - "learning_rate": 4.244937013232103e-06, - "loss": 0.2951, - "step": 7950 - }, - { - "epoch": 2.607271536193908, - "grad_norm": 2.9883480072021484, - "learning_rate": 4.176052149712489e-06, - "loss": 0.2511, - "step": 7960 - }, - { - "epoch": 2.61054700294792, - "grad_norm": 1.146877408027649, - "learning_rate": 4.107706418929774e-06, - "loss": 0.3063, - "step": 7970 - }, - { - "epoch": 2.6138224697019323, - "grad_norm": 1.3895083665847778, - "learning_rate": 4.039900624992332e-06, - "loss": 0.299, - "step": 7980 - }, - { - "epoch": 2.617097936455945, - "grad_norm": 0.5333308577537537, - "learning_rate": 3.9726355656560465e-06, - "loss": 0.3232, - "step": 7990 - }, - { - "epoch": 2.6203734032099573, - "grad_norm": 2.6040537357330322, - "learning_rate": 3.905912032314868e-06, - "loss": 0.288, - "step": 8000 - }, - { - "epoch": 2.62364886996397, - "grad_norm": 2.504030227661133, - "learning_rate": 3.839730809991532e-06, - "loss": 0.2885, - "step": 8010 - }, - { - "epoch": 2.626924336717982, - "grad_norm": 7.749489784240723, - "learning_rate": 3.7740926773283502e-06, - "loss": 0.3476, - "step": 8020 - }, - { - "epoch": 2.630199803471995, - "grad_norm": 2.1124677658081055, - "learning_rate": 3.7089984065779517e-06, - "loss": 0.2971, - "step": 8030 - }, - { - "epoch": 2.633475270226007, - "grad_norm": 6.998641490936279, - "learning_rate": 3.6444487635943327e-06, - "loss": 0.2823, - "step": 8040 - }, - { - "epoch": 2.6367507369800194, - "grad_norm": 1.9063273668289185, - "learning_rate": 3.58044450782376e-06, - "loss": 0.3298, - "step": 8050 - }, - { - "epoch": 2.640026203734032, - "grad_norm": 3.506763458251953, - "learning_rate": 3.5169863922958124e-06, - "loss": 0.2504, - "step": 8060 - }, - { - "epoch": 2.6433016704880448, - "grad_norm": 1.9572142362594604, - "learning_rate": 3.454075163614623e-06, - "loss": 0.2935, - "step": 8070 - }, - { - "epoch": 2.646577137242057, - "grad_norm": 0.40890321135520935, - "learning_rate": 3.391711561949984e-06, - "loss": 0.3405, - "step": 8080 - }, - { - "epoch": 2.6498526039960693, - "grad_norm": 2.226492404937744, - "learning_rate": 3.3298963210287193e-06, - "loss": 0.3107, - "step": 8090 - }, - { - "epoch": 2.653128070750082, - "grad_norm": 1.9833711385726929, - "learning_rate": 3.2686301681260136e-06, - "loss": 0.2849, - "step": 8100 - }, - { - "epoch": 2.656403537504094, - "grad_norm": 5.058218955993652, - "learning_rate": 3.207913824056852e-06, - "loss": 0.3083, - "step": 8110 - }, - { - "epoch": 2.659679004258107, - "grad_norm": 2.2231667041778564, - "learning_rate": 3.147748003167572e-06, - "loss": 0.2815, - "step": 8120 - }, - { - "epoch": 2.662954471012119, - "grad_norm": 3.4542975425720215, - "learning_rate": 3.0881334133274097e-06, - "loss": 0.2788, - "step": 8130 - }, - { - "epoch": 2.666229937766132, - "grad_norm": 8.236302375793457, - "learning_rate": 3.029070755920216e-06, - "loss": 0.2353, - "step": 8140 - }, - { - "epoch": 2.669505404520144, - "grad_norm": 0.1895446926355362, - "learning_rate": 2.970560725836219e-06, - "loss": 0.2268, - "step": 8150 - }, - { - "epoch": 2.6727808712741563, - "grad_norm": 12.59950065612793, - "learning_rate": 2.9126040114637375e-06, - "loss": 0.3003, - "step": 8160 - }, - { - "epoch": 2.676056338028169, - "grad_norm": 3.1450276374816895, - "learning_rate": 2.855201294681248e-06, - "loss": 0.3301, - "step": 8170 - }, - { - "epoch": 2.6793318047821817, - "grad_norm": 2.6632049083709717, - "learning_rate": 2.7983532508492172e-06, - "loss": 0.2646, - "step": 8180 - }, - { - "epoch": 2.682607271536194, - "grad_norm": 2.1871113777160645, - "learning_rate": 2.7420605488022554e-06, - "loss": 0.2512, - "step": 8190 - }, - { - "epoch": 2.685882738290206, - "grad_norm": 2.605992078781128, - "learning_rate": 2.6863238508411824e-06, - "loss": 0.2433, - "step": 8200 - }, - { - "epoch": 2.689158205044219, - "grad_norm": 4.701025009155273, - "learning_rate": 2.631143812725251e-06, - "loss": 0.24, - "step": 8210 - }, - { - "epoch": 2.692433671798231, - "grad_norm": 2.5368998050689697, - "learning_rate": 2.576521083664485e-06, - "loss": 0.3311, - "step": 8220 - }, - { - "epoch": 2.695709138552244, - "grad_norm": 2.5654404163360596, - "learning_rate": 2.5224563063119433e-06, - "loss": 0.2938, - "step": 8230 - }, - { - "epoch": 2.698984605306256, - "grad_norm": 10.346571922302246, - "learning_rate": 2.468950116756241e-06, - "loss": 0.3659, - "step": 8240 - }, - { - "epoch": 2.7022600720602687, - "grad_norm": 2.1608622074127197, - "learning_rate": 2.4160031445140565e-06, - "loss": 0.228, - "step": 8250 - }, - { - "epoch": 2.705535538814281, - "grad_norm": 1.0049759149551392, - "learning_rate": 2.3636160125226568e-06, - "loss": 0.2694, - "step": 8260 - }, - { - "epoch": 2.7088110055682932, - "grad_norm": 14.081513404846191, - "learning_rate": 2.3117893371326583e-06, - "loss": 0.2249, - "step": 8270 - }, - { - "epoch": 2.712086472322306, - "grad_norm": 2.5043344497680664, - "learning_rate": 2.260523728100711e-06, - "loss": 0.3202, - "step": 8280 - }, - { - "epoch": 2.7153619390763186, - "grad_norm": 2.559095859527588, - "learning_rate": 2.209819788582368e-06, - "loss": 0.252, - "step": 8290 - }, - { - "epoch": 2.718637405830331, - "grad_norm": 0.5076512694358826, - "learning_rate": 2.1596781151249524e-06, - "loss": 0.3075, - "step": 8300 - }, - { - "epoch": 2.721912872584343, - "grad_norm": 1.9681631326675415, - "learning_rate": 2.110099297660556e-06, - "loss": 0.2022, - "step": 8310 - }, - { - "epoch": 2.725188339338356, - "grad_norm": 4.116394519805908, - "learning_rate": 2.0610839194991182e-06, - "loss": 0.2628, - "step": 8320 - }, - { - "epoch": 2.728463806092368, - "grad_norm": 2.6161558628082275, - "learning_rate": 2.012632557321509e-06, - "loss": 0.2896, - "step": 8330 - }, - { - "epoch": 2.7317392728463807, - "grad_norm": 1.7793164253234863, - "learning_rate": 1.9647457811728053e-06, - "loss": 0.3079, - "step": 8340 - }, - { - "epoch": 2.735014739600393, - "grad_norm": 1.5956727266311646, - "learning_rate": 1.9174241544555604e-06, - "loss": 0.2686, - "step": 8350 - }, - { - "epoch": 2.7382902063544057, - "grad_norm": 0.6402525305747986, - "learning_rate": 1.8706682339231397e-06, - "loss": 0.2997, - "step": 8360 - }, - { - "epoch": 2.741565673108418, - "grad_norm": 2.577993154525757, - "learning_rate": 1.8244785696732325e-06, - "loss": 0.2611, - "step": 8370 - }, - { - "epoch": 2.74484113986243, - "grad_norm": 1.288020372390747, - "learning_rate": 1.7788557051413412e-06, - "loss": 0.3083, - "step": 8380 - }, - { - "epoch": 2.748116606616443, - "grad_norm": 4.449786186218262, - "learning_rate": 1.7338001770944024e-06, - "loss": 0.3381, - "step": 8390 - }, - { - "epoch": 2.7513920733704555, - "grad_norm": 2.563095808029175, - "learning_rate": 1.6893125156244526e-06, - "loss": 0.3124, - "step": 8400 - }, - { - "epoch": 2.754667540124468, - "grad_norm": 3.3274314403533936, - "learning_rate": 1.6453932441424013e-06, - "loss": 0.2478, - "step": 8410 - }, - { - "epoch": 2.75794300687848, - "grad_norm": 1.6172996759414673, - "learning_rate": 1.6020428793719121e-06, - "loss": 0.2405, - "step": 8420 - }, - { - "epoch": 2.7612184736324927, - "grad_norm": 2.373446226119995, - "learning_rate": 1.5592619313432367e-06, - "loss": 0.3727, - "step": 8430 - }, - { - "epoch": 2.764493940386505, - "grad_norm": 1.7787894010543823, - "learning_rate": 1.517050903387296e-06, - "loss": 0.2151, - "step": 8440 - }, - { - "epoch": 2.7677694071405177, - "grad_norm": 4.724449157714844, - "learning_rate": 1.4754102921297364e-06, - "loss": 0.3432, - "step": 8450 - }, - { - "epoch": 2.77104487389453, - "grad_norm": 3.2508323192596436, - "learning_rate": 1.4343405874850447e-06, - "loss": 0.2998, - "step": 8460 - }, - { - "epoch": 2.7743203406485426, - "grad_norm": 1.8032703399658203, - "learning_rate": 1.3938422726508471e-06, - "loss": 0.3147, - "step": 8470 - }, - { - "epoch": 2.777595807402555, - "grad_norm": 1.5102208852767944, - "learning_rate": 1.3539158241021643e-06, - "loss": 0.2928, - "step": 8480 - }, - { - "epoch": 2.780871274156567, - "grad_norm": 0.5686007142066956, - "learning_rate": 1.3145617115858766e-06, - "loss": 0.226, - "step": 8490 - }, - { - "epoch": 2.78414674091058, - "grad_norm": 4.261772155761719, - "learning_rate": 1.2757803981151228e-06, - "loss": 0.2594, - "step": 8500 - }, - { - "epoch": 2.7874222076645925, - "grad_norm": 4.669197082519531, - "learning_rate": 1.2375723399638883e-06, - "loss": 0.3217, - "step": 8510 - }, - { - "epoch": 2.7906976744186047, - "grad_norm": 3.5306320190429688, - "learning_rate": 1.1999379866616588e-06, - "loss": 0.3115, - "step": 8520 - }, - { - "epoch": 2.793973141172617, - "grad_norm": 9.068894386291504, - "learning_rate": 1.1628777809880754e-06, - "loss": 0.2401, - "step": 8530 - }, - { - "epoch": 2.7972486079266297, - "grad_norm": 1.7333732843399048, - "learning_rate": 1.126392158967776e-06, - "loss": 0.2403, - "step": 8540 - }, - { - "epoch": 2.800524074680642, - "grad_norm": 1.370765209197998, - "learning_rate": 1.0904815498652576e-06, - "loss": 0.3282, - "step": 8550 - }, - { - "epoch": 2.8037995414346546, - "grad_norm": 3.9893198013305664, - "learning_rate": 1.0551463761797776e-06, - "loss": 0.3052, - "step": 8560 - }, - { - "epoch": 2.807075008188667, - "grad_norm": 0.9445925951004028, - "learning_rate": 1.0203870536404591e-06, - "loss": 0.3066, - "step": 8570 - }, - { - "epoch": 2.8103504749426795, - "grad_norm": 2.26731014251709, - "learning_rate": 9.862039912013344e-07, - "loss": 0.3479, - "step": 8580 - }, - { - "epoch": 2.813625941696692, - "grad_norm": 4.6094536781311035, - "learning_rate": 9.525975910365747e-07, - "loss": 0.3811, - "step": 8590 - }, - { - "epoch": 2.816901408450704, - "grad_norm": 0.8830626606941223, - "learning_rate": 9.195682485357349e-07, - "loss": 0.2605, - "step": 8600 - }, - { - "epoch": 2.8201768752047167, - "grad_norm": 2.182631015777588, - "learning_rate": 8.871163522991166e-07, - "loss": 0.2948, - "step": 8610 - }, - { - "epoch": 2.823452341958729, - "grad_norm": 1.5267747640609741, - "learning_rate": 8.552422841331952e-07, - "loss": 0.2531, - "step": 8620 - }, - { - "epoch": 2.8267278087127417, - "grad_norm": 4.349455833435059, - "learning_rate": 8.239464190460955e-07, - "loss": 0.2945, - "step": 8630 - }, - { - "epoch": 2.830003275466754, - "grad_norm": 4.139877796173096, - "learning_rate": 7.932291252432333e-07, - "loss": 0.3234, - "step": 8640 - }, - { - "epoch": 2.8332787422207666, - "grad_norm": 4.465481281280518, - "learning_rate": 7.63090764122959e-07, - "loss": 0.3054, - "step": 8650 - }, - { - "epoch": 2.836554208974779, - "grad_norm": 0.6089248657226562, - "learning_rate": 7.335316902722711e-07, - "loss": 0.3094, - "step": 8660 - }, - { - "epoch": 2.839829675728791, - "grad_norm": 11.503205299377441, - "learning_rate": 7.045522514627201e-07, - "loss": 0.2832, - "step": 8670 - }, - { - "epoch": 2.8431051424828038, - "grad_norm": 8.383694648742676, - "learning_rate": 6.761527886462394e-07, - "loss": 0.3258, - "step": 8680 - }, - { - "epoch": 2.8463806092368165, - "grad_norm": 8.110169410705566, - "learning_rate": 6.48333635951176e-07, - "loss": 0.3019, - "step": 8690 - }, - { - "epoch": 2.8496560759908287, - "grad_norm": 12.683677673339844, - "learning_rate": 6.210951206783666e-07, - "loss": 0.2634, - "step": 8700 - }, - { - "epoch": 2.852931542744841, - "grad_norm": 12.854812622070312, - "learning_rate": 5.944375632972509e-07, - "loss": 0.2963, - "step": 8710 - }, - { - "epoch": 2.8562070094988536, - "grad_norm": 4.972071170806885, - "learning_rate": 5.683612774421365e-07, - "loss": 0.2723, - "step": 8720 - }, - { - "epoch": 2.859482476252866, - "grad_norm": 1.6116564273834229, - "learning_rate": 5.428665699084789e-07, - "loss": 0.3976, - "step": 8730 - }, - { - "epoch": 2.8627579430068786, - "grad_norm": 2.0649640560150146, - "learning_rate": 5.179537406493018e-07, - "loss": 0.3928, - "step": 8740 - }, - { - "epoch": 2.866033409760891, - "grad_norm": 6.8376240730285645, - "learning_rate": 4.936230827716549e-07, - "loss": 0.3216, - "step": 8750 - }, - { - "epoch": 2.8693088765149035, - "grad_norm": 2.8997652530670166, - "learning_rate": 4.6987488253313893e-07, - "loss": 0.2497, - "step": 8760 - }, - { - "epoch": 2.8725843432689158, - "grad_norm": 5.959221839904785, - "learning_rate": 4.4670941933859233e-07, - "loss": 0.3114, - "step": 8770 - }, - { - "epoch": 2.875859810022928, - "grad_norm": 4.574117660522461, - "learning_rate": 4.241269657367652e-07, - "loss": 0.3167, - "step": 8780 - }, - { - "epoch": 2.8791352767769407, - "grad_norm": 1.6917626857757568, - "learning_rate": 4.021277874171059e-07, - "loss": 0.334, - "step": 8790 - }, - { - "epoch": 2.8824107435309534, - "grad_norm": 3.443678855895996, - "learning_rate": 3.8071214320667404e-07, - "loss": 0.2732, - "step": 8800 - }, - { - "epoch": 2.8856862102849656, - "grad_norm": 1.1718047857284546, - "learning_rate": 3.5988028506706017e-07, - "loss": 0.2976, - "step": 8810 - }, - { - "epoch": 2.888961677038978, - "grad_norm": 0.9828905463218689, - "learning_rate": 3.396324580914323e-07, - "loss": 0.3312, - "step": 8820 - }, - { - "epoch": 2.8922371437929906, - "grad_norm": 2.7472238540649414, - "learning_rate": 3.1996890050166596e-07, - "loss": 0.2674, - "step": 8830 - }, - { - "epoch": 2.895512610547003, - "grad_norm": 1.3765918016433716, - "learning_rate": 3.008898436455021e-07, - "loss": 0.306, - "step": 8840 - }, - { - "epoch": 2.8987880773010155, - "grad_norm": 4.040841579437256, - "learning_rate": 2.8239551199388814e-07, - "loss": 0.2718, - "step": 8850 - }, - { - "epoch": 2.9020635440550278, - "grad_norm": 1.8809579610824585, - "learning_rate": 2.6448612313826894e-07, - "loss": 0.2395, - "step": 8860 - }, - { - "epoch": 2.9053390108090404, - "grad_norm": 0.4373060166835785, - "learning_rate": 2.471618877880832e-07, - "loss": 0.3002, - "step": 8870 - }, - { - "epoch": 2.9086144775630527, - "grad_norm": 4.400331497192383, - "learning_rate": 2.3042300976824361e-07, - "loss": 0.3069, - "step": 8880 - }, - { - "epoch": 2.911889944317065, - "grad_norm": 2.0026347637176514, - "learning_rate": 2.1426968601675479e-07, - "loss": 0.2501, - "step": 8890 - }, - { - "epoch": 2.9151654110710776, - "grad_norm": 3.752847671508789, - "learning_rate": 1.9870210658241018e-07, - "loss": 0.2881, - "step": 8900 - }, - { - "epoch": 2.9184408778250903, - "grad_norm": 3.722532272338867, - "learning_rate": 1.837204546225324e-07, - "loss": 0.2459, - "step": 8910 - }, - { - "epoch": 2.9217163445791026, - "grad_norm": 1.0510586500167847, - "learning_rate": 1.693249064008362e-07, - "loss": 0.2789, - "step": 8920 - }, - { - "epoch": 2.924991811333115, - "grad_norm": 3.138674259185791, - "learning_rate": 1.5551563128533008e-07, - "loss": 0.2577, - "step": 8930 - }, - { - "epoch": 2.9282672780871275, - "grad_norm": 11.46720027923584, - "learning_rate": 1.4229279174636234e-07, - "loss": 0.2899, - "step": 8940 - }, - { - "epoch": 2.9315427448411397, - "grad_norm": 0.589900016784668, - "learning_rate": 1.296565433546726e-07, - "loss": 0.2585, - "step": 8950 - }, - { - "epoch": 2.9348182115951524, - "grad_norm": 1.5188215970993042, - "learning_rate": 1.1760703477958213e-07, - "loss": 0.2576, - "step": 8960 - }, - { - "epoch": 2.9380936783491647, - "grad_norm": 3.8869900703430176, - "learning_rate": 1.0614440778723423e-07, - "loss": 0.2754, - "step": 8970 - }, - { - "epoch": 2.9413691451031774, - "grad_norm": 5.600375652313232, - "learning_rate": 9.526879723894544e-08, - "loss": 0.3187, - "step": 8980 - }, - { - "epoch": 2.9446446118571896, - "grad_norm": 1.1915491819381714, - "learning_rate": 8.498033108957914e-08, - "loss": 0.3476, - "step": 8990 - }, - { - "epoch": 2.947920078611202, - "grad_norm": 9.471319198608398, - "learning_rate": 7.527913038609113e-08, - "loss": 0.3437, - "step": 9000 - }, - { - "epoch": 2.9511955453652146, - "grad_norm": 7.630239009857178, - "learning_rate": 6.616530926606968e-08, - "loss": 0.2756, - "step": 9010 - }, - { - "epoch": 2.9544710121192272, - "grad_norm": 3.543031692504883, - "learning_rate": 5.7638974956397695e-08, - "loss": 0.2627, - "step": 9020 - }, - { - "epoch": 2.9577464788732395, - "grad_norm": 2.3270456790924072, - "learning_rate": 4.970022777200933e-08, - "loss": 0.3101, - "step": 9030 - }, - { - "epoch": 2.9610219456272517, - "grad_norm": 2.0923027992248535, - "learning_rate": 4.234916111467979e-08, - "loss": 0.2582, - "step": 9040 - }, - { - "epoch": 2.9642974123812644, - "grad_norm": 2.730461835861206, - "learning_rate": 3.558586147195952e-08, - "loss": 0.3565, - "step": 9050 - }, - { - "epoch": 2.9675728791352767, - "grad_norm": 3.649696111679077, - "learning_rate": 2.941040841612508e-08, - "loss": 0.2722, - "step": 9060 - }, - { - "epoch": 2.9708483458892894, - "grad_norm": 3.182731866836548, - "learning_rate": 2.382287460327981e-08, - "loss": 0.3016, - "step": 9070 - }, - { - "epoch": 2.9741238126433016, - "grad_norm": 1.0557725429534912, - "learning_rate": 1.88233257724546e-08, - "loss": 0.3091, - "step": 9080 - }, - { - "epoch": 2.9773992793973143, - "grad_norm": 5.730777740478516, - "learning_rate": 1.4411820744869575e-08, - "loss": 0.348, - "step": 9090 - }, - { - "epoch": 2.9806747461513265, - "grad_norm": 1.7255222797393799, - "learning_rate": 1.0588411423234634e-08, - "loss": 0.2921, - "step": 9100 - }, - { - "epoch": 2.983950212905339, - "grad_norm": 4.160470008850098, - "learning_rate": 7.353142791122203e-09, - "loss": 0.256, - "step": 9110 - }, - { - "epoch": 2.9872256796593515, - "grad_norm": 3.834279775619507, - "learning_rate": 4.706052912456516e-09, - "loss": 0.2981, - "step": 9120 - }, - { - "epoch": 2.990501146413364, - "grad_norm": 2.5707459449768066, - "learning_rate": 2.647172931052877e-09, - "loss": 0.2123, - "step": 9130 - }, - { - "epoch": 2.9937766131673764, - "grad_norm": 2.4034626483917236, - "learning_rate": 1.176527070273492e-09, - "loss": 0.326, - "step": 9140 - }, - { - "epoch": 2.9970520799213887, - "grad_norm": 0.7634719610214233, - "learning_rate": 2.9413263271105274e-10, - "loss": 0.2781, - "step": 9150 - }, - { - "epoch": 3.0, - "step": 9159, - "total_flos": 1.1108019299619439e+19, - "train_loss": 0.37588436796957, - "train_runtime": 28593.9573, - "train_samples_per_second": 5.124, - "train_steps_per_second": 0.32 - } - ], - "logging_steps": 10, - "max_steps": 9159, - "num_input_tokens_seen": 0, - "num_train_epochs": 3, - "save_steps": 500, - "stateful_callbacks": { - "TrainerControl": { - "args": { - "should_epoch_stop": false, - "should_evaluate": false, - "should_log": false, - "should_save": true, - "should_training_stop": true - }, - "attributes": {} - } - }, - "total_flos": 1.1108019299619439e+19, - "train_batch_size": 4, - "trial_name": null, - "trial_params": null -}