{ "best_metric": null, "best_model_checkpoint": null, "epoch": 2.997759522031367, "eval_steps": 500, "global_step": 3513, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0, "eval_loss": 2.1641297340393066, "eval_runtime": 1.9384, "eval_samples_per_second": 10.318, "eval_steps_per_second": 1.032, "step": 0 }, { "epoch": 0.0008535154166222128, "grad_norm": 4.246492385864258, "learning_rate": 2.840909090909091e-08, "loss": 1.0032, "step": 1 }, { "epoch": 0.008535154166222128, "grad_norm": 4.319986820220947, "learning_rate": 2.840909090909091e-07, "loss": 0.983, "step": 10 }, { "epoch": 0.017070308332444255, "grad_norm": 3.9372925758361816, "learning_rate": 5.681818181818182e-07, "loss": 0.9573, "step": 20 }, { "epoch": 0.02560546249866638, "grad_norm": 3.070388078689575, "learning_rate": 8.522727272727273e-07, "loss": 0.9306, "step": 30 }, { "epoch": 0.03414061666488851, "grad_norm": 1.9363492727279663, "learning_rate": 1.1363636363636364e-06, "loss": 0.8481, "step": 40 }, { "epoch": 0.04267577083111064, "grad_norm": 1.282257318496704, "learning_rate": 1.4204545454545458e-06, "loss": 0.7776, "step": 50 }, { "epoch": 0.05121092499733276, "grad_norm": 1.0056549310684204, "learning_rate": 1.7045454545454546e-06, "loss": 0.7324, "step": 60 }, { "epoch": 0.05974607916355489, "grad_norm": 0.8602527379989624, "learning_rate": 1.9886363636363638e-06, "loss": 0.6952, "step": 70 }, { "epoch": 0.06828123332977702, "grad_norm": 0.822441577911377, "learning_rate": 2.2727272727272728e-06, "loss": 0.6724, "step": 80 }, { "epoch": 0.07681638749599914, "grad_norm": 0.7585397958755493, "learning_rate": 2.556818181818182e-06, "loss": 0.6635, "step": 90 }, { "epoch": 0.08535154166222128, "grad_norm": 0.7706250548362732, "learning_rate": 2.8409090909090916e-06, "loss": 0.6502, "step": 100 }, { "epoch": 0.0938866958284434, "grad_norm": 0.8293640613555908, "learning_rate": 3.125e-06, "loss": 0.6334, "step": 110 }, { "epoch": 0.10242184999466553, "grad_norm": 0.781350314617157, "learning_rate": 3.409090909090909e-06, "loss": 0.6268, "step": 120 }, { "epoch": 0.11095700416088766, "grad_norm": 0.7660292387008667, "learning_rate": 3.6931818181818186e-06, "loss": 0.6137, "step": 130 }, { "epoch": 0.11949215832710978, "grad_norm": 0.8327275514602661, "learning_rate": 3.9772727272727275e-06, "loss": 0.6159, "step": 140 }, { "epoch": 0.1280273124933319, "grad_norm": 0.8584268093109131, "learning_rate": 4.2613636363636365e-06, "loss": 0.5995, "step": 150 }, { "epoch": 0.13656246665955404, "grad_norm": 0.7809672951698303, "learning_rate": 4.5454545454545455e-06, "loss": 0.6043, "step": 160 }, { "epoch": 0.14509762082577615, "grad_norm": 0.8208410739898682, "learning_rate": 4.829545454545455e-06, "loss": 0.6015, "step": 170 }, { "epoch": 0.1536327749919983, "grad_norm": 0.828789472579956, "learning_rate": 5.113636363636364e-06, "loss": 0.5871, "step": 180 }, { "epoch": 0.16216792915822043, "grad_norm": 0.7799698114395142, "learning_rate": 5.397727272727273e-06, "loss": 0.5946, "step": 190 }, { "epoch": 0.17070308332444256, "grad_norm": 0.7569297552108765, "learning_rate": 5.681818181818183e-06, "loss": 0.5868, "step": 200 }, { "epoch": 0.17923823749066467, "grad_norm": 0.7858031392097473, "learning_rate": 5.965909090909091e-06, "loss": 0.5732, "step": 210 }, { "epoch": 0.1877733916568868, "grad_norm": 0.7812625765800476, "learning_rate": 6.25e-06, "loss": 0.5745, "step": 220 }, { "epoch": 0.19630854582310894, "grad_norm": 0.7922148704528809, "learning_rate": 6.53409090909091e-06, "loss": 0.5742, "step": 230 }, { "epoch": 0.20484369998933105, "grad_norm": 0.8420217633247375, "learning_rate": 6.818181818181818e-06, "loss": 0.5593, "step": 240 }, { "epoch": 0.2133788541555532, "grad_norm": 0.8149869441986084, "learning_rate": 7.102272727272727e-06, "loss": 0.5694, "step": 250 }, { "epoch": 0.22191400832177532, "grad_norm": 0.8200048208236694, "learning_rate": 7.386363636363637e-06, "loss": 0.5609, "step": 260 }, { "epoch": 0.23044916248799743, "grad_norm": 0.7718726396560669, "learning_rate": 7.670454545454547e-06, "loss": 0.5596, "step": 270 }, { "epoch": 0.23898431665421957, "grad_norm": 0.8079963326454163, "learning_rate": 7.954545454545455e-06, "loss": 0.5515, "step": 280 }, { "epoch": 0.2475194708204417, "grad_norm": 0.8398913741111755, "learning_rate": 8.238636363636365e-06, "loss": 0.5615, "step": 290 }, { "epoch": 0.2560546249866638, "grad_norm": 0.8522515892982483, "learning_rate": 8.522727272727273e-06, "loss": 0.5405, "step": 300 }, { "epoch": 0.26458977915288595, "grad_norm": 0.965617299079895, "learning_rate": 8.806818181818183e-06, "loss": 0.5438, "step": 310 }, { "epoch": 0.2731249333191081, "grad_norm": 0.834233283996582, "learning_rate": 9.090909090909091e-06, "loss": 0.5342, "step": 320 }, { "epoch": 0.2816600874853302, "grad_norm": 0.8867103457450867, "learning_rate": 9.375000000000001e-06, "loss": 0.5272, "step": 330 }, { "epoch": 0.2901952416515523, "grad_norm": 0.767537534236908, "learning_rate": 9.65909090909091e-06, "loss": 0.5319, "step": 340 }, { "epoch": 0.29873039581777444, "grad_norm": 0.8806461095809937, "learning_rate": 9.943181818181819e-06, "loss": 0.5322, "step": 350 }, { "epoch": 0.3072655499839966, "grad_norm": 0.7467280030250549, "learning_rate": 9.999841959480543e-06, "loss": 0.5267, "step": 360 }, { "epoch": 0.3158007041502187, "grad_norm": 0.7319239974021912, "learning_rate": 9.999199936992976e-06, "loss": 0.5221, "step": 370 }, { "epoch": 0.32433585831644085, "grad_norm": 0.7590662837028503, "learning_rate": 9.998064118371841e-06, "loss": 0.525, "step": 380 }, { "epoch": 0.332871012482663, "grad_norm": 0.7705526351928711, "learning_rate": 9.996434615807662e-06, "loss": 0.5254, "step": 390 }, { "epoch": 0.3414061666488851, "grad_norm": 0.8061032295227051, "learning_rate": 9.994311590254603e-06, "loss": 0.5221, "step": 400 }, { "epoch": 0.3499413208151072, "grad_norm": 0.7729495167732239, "learning_rate": 9.991695251414584e-06, "loss": 0.53, "step": 410 }, { "epoch": 0.35847647498132934, "grad_norm": 0.768975019454956, "learning_rate": 9.988585857716555e-06, "loss": 0.5099, "step": 420 }, { "epoch": 0.3670116291475515, "grad_norm": 0.7381827235221863, "learning_rate": 9.984983716290981e-06, "loss": 0.5182, "step": 430 }, { "epoch": 0.3755467833137736, "grad_norm": 0.8239282965660095, "learning_rate": 9.980889182939495e-06, "loss": 0.52, "step": 440 }, { "epoch": 0.38408193747999575, "grad_norm": 0.8047374486923218, "learning_rate": 9.976302662099758e-06, "loss": 0.5091, "step": 450 }, { "epoch": 0.3926170916462179, "grad_norm": 0.7568238377571106, "learning_rate": 9.971224606805514e-06, "loss": 0.5171, "step": 460 }, { "epoch": 0.40115224581243997, "grad_norm": 0.7178462147712708, "learning_rate": 9.965655518641837e-06, "loss": 0.5119, "step": 470 }, { "epoch": 0.4096873999786621, "grad_norm": 0.8041744232177734, "learning_rate": 9.959595947695588e-06, "loss": 0.506, "step": 480 }, { "epoch": 0.41822255414488424, "grad_norm": 0.7083892226219177, "learning_rate": 9.953046492501082e-06, "loss": 0.5066, "step": 490 }, { "epoch": 0.4267577083111064, "grad_norm": 0.7124646902084351, "learning_rate": 9.946007799980963e-06, "loss": 0.5098, "step": 500 }, { "epoch": 0.4352928624773285, "grad_norm": 0.7167072892189026, "learning_rate": 9.938480565382312e-06, "loss": 0.513, "step": 510 }, { "epoch": 0.44382801664355065, "grad_norm": 0.7211044430732727, "learning_rate": 9.930465532207967e-06, "loss": 0.5094, "step": 520 }, { "epoch": 0.45236317080977273, "grad_norm": 0.7868006229400635, "learning_rate": 9.921963492143083e-06, "loss": 0.5131, "step": 530 }, { "epoch": 0.46089832497599487, "grad_norm": 0.7130579948425293, "learning_rate": 9.912975284976942e-06, "loss": 0.506, "step": 540 }, { "epoch": 0.469433479142217, "grad_norm": 0.7939994931221008, "learning_rate": 9.903501798519994e-06, "loss": 0.5025, "step": 550 }, { "epoch": 0.47796863330843914, "grad_norm": 0.7138738036155701, "learning_rate": 9.893543968516166e-06, "loss": 0.5037, "step": 560 }, { "epoch": 0.4865037874746613, "grad_norm": 0.716701328754425, "learning_rate": 9.883102778550434e-06, "loss": 0.5034, "step": 570 }, { "epoch": 0.4950389416408834, "grad_norm": 0.7295582294464111, "learning_rate": 9.872179259951672e-06, "loss": 0.4949, "step": 580 }, { "epoch": 0.5035740958071055, "grad_norm": 0.7438467144966125, "learning_rate": 9.860774491690777e-06, "loss": 0.4969, "step": 590 }, { "epoch": 0.5121092499733276, "grad_norm": 0.696615993976593, "learning_rate": 9.848889600274101e-06, "loss": 0.4926, "step": 600 }, { "epoch": 0.5206444041395498, "grad_norm": 0.6948277354240417, "learning_rate": 9.836525759632171e-06, "loss": 0.5034, "step": 610 }, { "epoch": 0.5291795583057719, "grad_norm": 0.8091464042663574, "learning_rate": 9.823684191003744e-06, "loss": 0.5015, "step": 620 }, { "epoch": 0.537714712471994, "grad_norm": 0.7501440048217773, "learning_rate": 9.810366162815171e-06, "loss": 0.5024, "step": 630 }, { "epoch": 0.5462498666382162, "grad_norm": 0.7048625946044922, "learning_rate": 9.796572990555112e-06, "loss": 0.4955, "step": 640 }, { "epoch": 0.5547850208044383, "grad_norm": 0.7141663432121277, "learning_rate": 9.782306036644596e-06, "loss": 0.5045, "step": 650 }, { "epoch": 0.5633201749706604, "grad_norm": 0.7329632639884949, "learning_rate": 9.767566710302453e-06, "loss": 0.5038, "step": 660 }, { "epoch": 0.5718553291368825, "grad_norm": 0.699938952922821, "learning_rate": 9.752356467406114e-06, "loss": 0.5066, "step": 670 }, { "epoch": 0.5803904833031046, "grad_norm": 0.6912060976028442, "learning_rate": 9.736676810347807e-06, "loss": 0.4918, "step": 680 }, { "epoch": 0.5889256374693268, "grad_norm": 0.6718906164169312, "learning_rate": 9.720529287886159e-06, "loss": 0.4932, "step": 690 }, { "epoch": 0.5974607916355489, "grad_norm": 0.6830474734306335, "learning_rate": 9.703915494993215e-06, "loss": 0.489, "step": 700 }, { "epoch": 0.6059959458017711, "grad_norm": 0.6920751333236694, "learning_rate": 9.686837072696902e-06, "loss": 0.4976, "step": 710 }, { "epoch": 0.6145310999679932, "grad_norm": 0.695093035697937, "learning_rate": 9.669295707918928e-06, "loss": 0.4877, "step": 720 }, { "epoch": 0.6230662541342153, "grad_norm": 0.683686375617981, "learning_rate": 9.651293133308159e-06, "loss": 0.4939, "step": 730 }, { "epoch": 0.6316014083004374, "grad_norm": 0.7766408920288086, "learning_rate": 9.632831127069479e-06, "loss": 0.4993, "step": 740 }, { "epoch": 0.6401365624666595, "grad_norm": 0.7032186985015869, "learning_rate": 9.61391151278814e-06, "loss": 0.4896, "step": 750 }, { "epoch": 0.6486717166328817, "grad_norm": 0.7620243430137634, "learning_rate": 9.594536159249654e-06, "loss": 0.4926, "step": 760 }, { "epoch": 0.6572068707991038, "grad_norm": 0.6803762316703796, "learning_rate": 9.57470698025518e-06, "loss": 0.4902, "step": 770 }, { "epoch": 0.665742024965326, "grad_norm": 0.6878206133842468, "learning_rate": 9.55442593443251e-06, "loss": 0.4841, "step": 780 }, { "epoch": 0.674277179131548, "grad_norm": 0.7530574202537537, "learning_rate": 9.533695025042591e-06, "loss": 0.4839, "step": 790 }, { "epoch": 0.6828123332977702, "grad_norm": 0.7192838191986084, "learning_rate": 9.512516299781665e-06, "loss": 0.4841, "step": 800 }, { "epoch": 0.6913474874639923, "grad_norm": 0.7086024880409241, "learning_rate": 9.490891850578993e-06, "loss": 0.4869, "step": 810 }, { "epoch": 0.6998826416302144, "grad_norm": 0.6931254267692566, "learning_rate": 9.468823813390236e-06, "loss": 0.4834, "step": 820 }, { "epoch": 0.7084177957964366, "grad_norm": 0.6703010201454163, "learning_rate": 9.446314367986476e-06, "loss": 0.4904, "step": 830 }, { "epoch": 0.7169529499626587, "grad_norm": 0.721376359462738, "learning_rate": 9.423365737738898e-06, "loss": 0.4896, "step": 840 }, { "epoch": 0.7254881041288809, "grad_norm": 0.7349815964698792, "learning_rate": 9.399980189399191e-06, "loss": 0.4938, "step": 850 }, { "epoch": 0.734023258295103, "grad_norm": 0.7041159272193909, "learning_rate": 9.376160032875638e-06, "loss": 0.4889, "step": 860 }, { "epoch": 0.742558412461325, "grad_norm": 0.7001343965530396, "learning_rate": 9.351907621004963e-06, "loss": 0.4908, "step": 870 }, { "epoch": 0.7510935666275472, "grad_norm": 0.7233178019523621, "learning_rate": 9.327225349319922e-06, "loss": 0.4856, "step": 880 }, { "epoch": 0.7596287207937693, "grad_norm": 0.6933467984199524, "learning_rate": 9.30211565581269e-06, "loss": 0.4817, "step": 890 }, { "epoch": 0.7681638749599915, "grad_norm": 0.6627389192581177, "learning_rate": 9.276581020694048e-06, "loss": 0.4804, "step": 900 }, { "epoch": 0.7766990291262136, "grad_norm": 0.6689618229866028, "learning_rate": 9.250623966148392e-06, "loss": 0.4854, "step": 910 }, { "epoch": 0.7852341832924358, "grad_norm": 0.7105383276939392, "learning_rate": 9.224247056084617e-06, "loss": 0.4831, "step": 920 }, { "epoch": 0.7937693374586579, "grad_norm": 0.7106824517250061, "learning_rate": 9.197452895882857e-06, "loss": 0.4802, "step": 930 }, { "epoch": 0.8023044916248799, "grad_norm": 0.6922078132629395, "learning_rate": 9.170244132137143e-06, "loss": 0.4854, "step": 940 }, { "epoch": 0.8108396457911021, "grad_norm": 0.7484738230705261, "learning_rate": 9.142623452393984e-06, "loss": 0.4708, "step": 950 }, { "epoch": 0.8193747999573242, "grad_norm": 0.7449445128440857, "learning_rate": 9.114593584886902e-06, "loss": 0.4786, "step": 960 }, { "epoch": 0.8279099541235464, "grad_norm": 0.6751102209091187, "learning_rate": 9.086157298266957e-06, "loss": 0.4783, "step": 970 }, { "epoch": 0.8364451082897685, "grad_norm": 0.7621346712112427, "learning_rate": 9.057317401329271e-06, "loss": 0.4868, "step": 980 }, { "epoch": 0.8449802624559906, "grad_norm": 0.7197836637496948, "learning_rate": 9.028076742735583e-06, "loss": 0.479, "step": 990 }, { "epoch": 0.8535154166222128, "grad_norm": 0.7285887598991394, "learning_rate": 8.998438210732882e-06, "loss": 0.4851, "step": 1000 }, { "epoch": 0.8620505707884348, "grad_norm": 0.7014701962471008, "learning_rate": 8.968404732868112e-06, "loss": 0.4703, "step": 1010 }, { "epoch": 0.870585724954657, "grad_norm": 0.6849453449249268, "learning_rate": 8.937979275699012e-06, "loss": 0.4729, "step": 1020 }, { "epoch": 0.8791208791208791, "grad_norm": 0.7209568023681641, "learning_rate": 8.907164844501084e-06, "loss": 0.4777, "step": 1030 }, { "epoch": 0.8876560332871013, "grad_norm": 0.6904247403144836, "learning_rate": 8.875964482970756e-06, "loss": 0.4731, "step": 1040 }, { "epoch": 0.8961911874533234, "grad_norm": 0.6657469272613525, "learning_rate": 8.844381272924733e-06, "loss": 0.4773, "step": 1050 }, { "epoch": 0.9047263416195455, "grad_norm": 0.6649881601333618, "learning_rate": 8.812418333995597e-06, "loss": 0.4659, "step": 1060 }, { "epoch": 0.9132614957857677, "grad_norm": 0.6824942231178284, "learning_rate": 8.780078823323667e-06, "loss": 0.4736, "step": 1070 }, { "epoch": 0.9217966499519897, "grad_norm": 0.6798464059829712, "learning_rate": 8.74736593524514e-06, "loss": 0.4695, "step": 1080 }, { "epoch": 0.9303318041182119, "grad_norm": 0.6828529238700867, "learning_rate": 8.71428290097658e-06, "loss": 0.4678, "step": 1090 }, { "epoch": 0.938866958284434, "grad_norm": 0.6548129916191101, "learning_rate": 8.680832988295757e-06, "loss": 0.4748, "step": 1100 }, { "epoch": 0.9474021124506561, "grad_norm": 0.7184150815010071, "learning_rate": 8.647019501218862e-06, "loss": 0.4851, "step": 1110 }, { "epoch": 0.9559372666168783, "grad_norm": 0.6707058548927307, "learning_rate": 8.612845779674166e-06, "loss": 0.4621, "step": 1120 }, { "epoch": 0.9644724207831004, "grad_norm": 0.6430131793022156, "learning_rate": 8.578315199172105e-06, "loss": 0.4733, "step": 1130 }, { "epoch": 0.9730075749493226, "grad_norm": 0.652946949005127, "learning_rate": 8.543431170471872e-06, "loss": 0.469, "step": 1140 }, { "epoch": 0.9815427291155446, "grad_norm": 0.678126871585846, "learning_rate": 8.508197139244515e-06, "loss": 0.4645, "step": 1150 }, { "epoch": 0.9900778832817668, "grad_norm": 0.6810508966445923, "learning_rate": 8.472616585732597e-06, "loss": 0.4722, "step": 1160 }, { "epoch": 0.9986130374479889, "grad_norm": 0.6804456114768982, "learning_rate": 8.436693024406427e-06, "loss": 0.4796, "step": 1170 }, { "epoch": 1.0068281233329777, "grad_norm": 0.7665020227432251, "learning_rate": 8.400430003616924e-06, "loss": 0.4501, "step": 1180 }, { "epoch": 1.0153632774991999, "grad_norm": 0.7121028304100037, "learning_rate": 8.363831105245125e-06, "loss": 0.4436, "step": 1190 }, { "epoch": 1.023898431665422, "grad_norm": 0.7494065761566162, "learning_rate": 8.326899944348389e-06, "loss": 0.4459, "step": 1200 }, { "epoch": 1.032433585831644, "grad_norm": 0.6524202227592468, "learning_rate": 8.289640168803308e-06, "loss": 0.4445, "step": 1210 }, { "epoch": 1.0409687399978662, "grad_norm": 0.6446959972381592, "learning_rate": 8.25205545894541e-06, "loss": 0.4406, "step": 1220 }, { "epoch": 1.0495038941640884, "grad_norm": 0.6765570044517517, "learning_rate": 8.214149527205606e-06, "loss": 0.4538, "step": 1230 }, { "epoch": 1.0580390483303104, "grad_norm": 0.6525077223777771, "learning_rate": 8.17592611774352e-06, "loss": 0.4492, "step": 1240 }, { "epoch": 1.0665742024965326, "grad_norm": 0.6981220245361328, "learning_rate": 8.13738900607764e-06, "loss": 0.4428, "step": 1250 }, { "epoch": 1.0751093566627548, "grad_norm": 0.7378373742103577, "learning_rate": 8.098541998712405e-06, "loss": 0.442, "step": 1260 }, { "epoch": 1.083644510828977, "grad_norm": 0.6806818246841431, "learning_rate": 8.059388932762207e-06, "loss": 0.4428, "step": 1270 }, { "epoch": 1.092179664995199, "grad_norm": 0.704845130443573, "learning_rate": 8.019933675572389e-06, "loss": 0.4402, "step": 1280 }, { "epoch": 1.100714819161421, "grad_norm": 0.8347475528717041, "learning_rate": 7.980180124337239e-06, "loss": 0.448, "step": 1290 }, { "epoch": 1.1092499733276433, "grad_norm": 0.7030121684074402, "learning_rate": 7.940132205715053e-06, "loss": 0.4462, "step": 1300 }, { "epoch": 1.1177851274938653, "grad_norm": 0.6647533178329468, "learning_rate": 7.89979387544027e-06, "loss": 0.4481, "step": 1310 }, { "epoch": 1.1263202816600875, "grad_norm": 0.724305272102356, "learning_rate": 7.859169117932752e-06, "loss": 0.4426, "step": 1320 }, { "epoch": 1.1348554358263097, "grad_norm": 0.6487372517585754, "learning_rate": 7.818261945904226e-06, "loss": 0.4442, "step": 1330 }, { "epoch": 1.1433905899925318, "grad_norm": 0.7011922597885132, "learning_rate": 7.777076399961912e-06, "loss": 0.4486, "step": 1340 }, { "epoch": 1.1519257441587538, "grad_norm": 0.6805099844932556, "learning_rate": 7.735616548209425e-06, "loss": 0.4469, "step": 1350 }, { "epoch": 1.160460898324976, "grad_norm": 0.683447003364563, "learning_rate": 7.693886485844955e-06, "loss": 0.4553, "step": 1360 }, { "epoch": 1.1689960524911982, "grad_norm": 0.8037645816802979, "learning_rate": 7.651890334756736e-06, "loss": 0.4387, "step": 1370 }, { "epoch": 1.1775312066574202, "grad_norm": 0.668209433555603, "learning_rate": 7.609632243115941e-06, "loss": 0.4373, "step": 1380 }, { "epoch": 1.1860663608236424, "grad_norm": 0.7479778528213501, "learning_rate": 7.567116384966918e-06, "loss": 0.4525, "step": 1390 }, { "epoch": 1.1946015149898646, "grad_norm": 0.743675172328949, "learning_rate": 7.524346959814918e-06, "loss": 0.4458, "step": 1400 }, { "epoch": 1.2031366691560867, "grad_norm": 0.6677897572517395, "learning_rate": 7.481328192211278e-06, "loss": 0.4521, "step": 1410 }, { "epoch": 1.2116718233223087, "grad_norm": 0.655086100101471, "learning_rate": 7.4380643313361385e-06, "loss": 0.4439, "step": 1420 }, { "epoch": 1.220206977488531, "grad_norm": 0.7043260335922241, "learning_rate": 7.394559650578746e-06, "loss": 0.4387, "step": 1430 }, { "epoch": 1.228742131654753, "grad_norm": 0.6985997557640076, "learning_rate": 7.3508184471153274e-06, "loss": 0.4396, "step": 1440 }, { "epoch": 1.237277285820975, "grad_norm": 0.6879217028617859, "learning_rate": 7.306845041484656e-06, "loss": 0.4438, "step": 1450 }, { "epoch": 1.2458124399871973, "grad_norm": 0.6632692217826843, "learning_rate": 7.262643777161279e-06, "loss": 0.4421, "step": 1460 }, { "epoch": 1.2543475941534195, "grad_norm": 0.6956544518470764, "learning_rate": 7.218219020126491e-06, "loss": 0.45, "step": 1470 }, { "epoch": 1.2628827483196416, "grad_norm": 0.6400607824325562, "learning_rate": 7.1735751584370894e-06, "loss": 0.4372, "step": 1480 }, { "epoch": 1.2714179024858636, "grad_norm": 0.735914945602417, "learning_rate": 7.128716601791938e-06, "loss": 0.4459, "step": 1490 }, { "epoch": 1.2799530566520858, "grad_norm": 0.7256124019622803, "learning_rate": 7.083647781096404e-06, "loss": 0.4437, "step": 1500 }, { "epoch": 1.2884882108183078, "grad_norm": 0.6894016861915588, "learning_rate": 7.038373148024693e-06, "loss": 0.4468, "step": 1510 }, { "epoch": 1.29702336498453, "grad_norm": 0.676726222038269, "learning_rate": 6.992897174580133e-06, "loss": 0.4409, "step": 1520 }, { "epoch": 1.3055585191507522, "grad_norm": 0.7112148404121399, "learning_rate": 6.947224352653454e-06, "loss": 0.4387, "step": 1530 }, { "epoch": 1.3140936733169744, "grad_norm": 0.6504147052764893, "learning_rate": 6.901359193579104e-06, "loss": 0.4357, "step": 1540 }, { "epoch": 1.3226288274831965, "grad_norm": 0.6736079454421997, "learning_rate": 6.855306227689634e-06, "loss": 0.4403, "step": 1550 }, { "epoch": 1.3311639816494185, "grad_norm": 0.6631302237510681, "learning_rate": 6.809070003868221e-06, "loss": 0.4368, "step": 1560 }, { "epoch": 1.3396991358156407, "grad_norm": 0.6592603921890259, "learning_rate": 6.762655089099353e-06, "loss": 0.446, "step": 1570 }, { "epoch": 1.3482342899818627, "grad_norm": 0.6875331997871399, "learning_rate": 6.716066068017722e-06, "loss": 0.4459, "step": 1580 }, { "epoch": 1.3567694441480849, "grad_norm": 0.8026809096336365, "learning_rate": 6.669307542455369e-06, "loss": 0.4363, "step": 1590 }, { "epoch": 1.365304598314307, "grad_norm": 0.8665246367454529, "learning_rate": 6.6223841309871535e-06, "loss": 0.4415, "step": 1600 }, { "epoch": 1.3738397524805293, "grad_norm": 0.7898995280265808, "learning_rate": 6.575300468474546e-06, "loss": 0.4411, "step": 1610 }, { "epoch": 1.3823749066467514, "grad_norm": 0.6701722145080566, "learning_rate": 6.52806120560781e-06, "loss": 0.441, "step": 1620 }, { "epoch": 1.3909100608129734, "grad_norm": 0.6245713233947754, "learning_rate": 6.480671008446651e-06, "loss": 0.4357, "step": 1630 }, { "epoch": 1.3994452149791956, "grad_norm": 0.6475566029548645, "learning_rate": 6.433134557959303e-06, "loss": 0.4487, "step": 1640 }, { "epoch": 1.4079803691454176, "grad_norm": 0.7050219178199768, "learning_rate": 6.385456549560187e-06, "loss": 0.4488, "step": 1650 }, { "epoch": 1.4165155233116398, "grad_norm": 0.6392747759819031, "learning_rate": 6.337641692646106e-06, "loss": 0.4453, "step": 1660 }, { "epoch": 1.425050677477862, "grad_norm": 0.6791578531265259, "learning_rate": 6.2896947101310756e-06, "loss": 0.4406, "step": 1670 }, { "epoch": 1.4335858316440842, "grad_norm": 0.6261289119720459, "learning_rate": 6.24162033797983e-06, "loss": 0.4364, "step": 1680 }, { "epoch": 1.4421209858103061, "grad_norm": 0.6886487603187561, "learning_rate": 6.193423324740018e-06, "loss": 0.44, "step": 1690 }, { "epoch": 1.4506561399765283, "grad_norm": 0.7670247554779053, "learning_rate": 6.1451084310731545e-06, "loss": 0.4348, "step": 1700 }, { "epoch": 1.4591912941427505, "grad_norm": 0.6851808428764343, "learning_rate": 6.096680429284412e-06, "loss": 0.4358, "step": 1710 }, { "epoch": 1.4677264483089725, "grad_norm": 0.660381555557251, "learning_rate": 6.0481441028512154e-06, "loss": 0.4419, "step": 1720 }, { "epoch": 1.4762616024751947, "grad_norm": 0.6627777218818665, "learning_rate": 5.9995042459507555e-06, "loss": 0.4443, "step": 1730 }, { "epoch": 1.4847967566414169, "grad_norm": 0.7264851331710815, "learning_rate": 5.9507656629864554e-06, "loss": 0.4419, "step": 1740 }, { "epoch": 1.493331910807639, "grad_norm": 0.6353123784065247, "learning_rate": 5.901933168113405e-06, "loss": 0.4325, "step": 1750 }, { "epoch": 1.5018670649738612, "grad_norm": 0.7154319882392883, "learning_rate": 5.853011584762846e-06, "loss": 0.4413, "step": 1760 }, { "epoch": 1.5104022191400832, "grad_norm": 0.678537905216217, "learning_rate": 5.804005745165735e-06, "loss": 0.4418, "step": 1770 }, { "epoch": 1.5189373733063052, "grad_norm": 0.6648243069648743, "learning_rate": 5.754920489875447e-06, "loss": 0.438, "step": 1780 }, { "epoch": 1.5274725274725274, "grad_norm": 0.7042447924613953, "learning_rate": 5.705760667289644e-06, "loss": 0.441, "step": 1790 }, { "epoch": 1.5360076816387496, "grad_norm": 0.6617943048477173, "learning_rate": 5.65653113317137e-06, "loss": 0.4375, "step": 1800 }, { "epoch": 1.5445428358049718, "grad_norm": 0.6619049310684204, "learning_rate": 5.607236750169431e-06, "loss": 0.4395, "step": 1810 }, { "epoch": 1.553077989971194, "grad_norm": 0.6774610280990601, "learning_rate": 5.557882387338086e-06, "loss": 0.4417, "step": 1820 }, { "epoch": 1.5616131441374161, "grad_norm": 0.6360573768615723, "learning_rate": 5.508472919656101e-06, "loss": 0.4346, "step": 1830 }, { "epoch": 1.5701482983036381, "grad_norm": 0.6723465919494629, "learning_rate": 5.459013227545224e-06, "loss": 0.4387, "step": 1840 }, { "epoch": 1.57868345246986, "grad_norm": 0.6785566806793213, "learning_rate": 5.4095081963881225e-06, "loss": 0.4364, "step": 1850 }, { "epoch": 1.5872186066360823, "grad_norm": 0.6658756136894226, "learning_rate": 5.359962716045836e-06, "loss": 0.4302, "step": 1860 }, { "epoch": 1.5957537608023045, "grad_norm": 0.6586979031562805, "learning_rate": 5.310381680374767e-06, "loss": 0.4386, "step": 1870 }, { "epoch": 1.6042889149685267, "grad_norm": 0.7203266620635986, "learning_rate": 5.260769986743296e-06, "loss": 0.4377, "step": 1880 }, { "epoch": 1.6128240691347489, "grad_norm": 0.7455547451972961, "learning_rate": 5.2111325355480506e-06, "loss": 0.4428, "step": 1890 }, { "epoch": 1.6213592233009708, "grad_norm": 0.6452232599258423, "learning_rate": 5.161474229729858e-06, "loss": 0.4336, "step": 1900 }, { "epoch": 1.629894377467193, "grad_norm": 0.6487184166908264, "learning_rate": 5.111799974289462e-06, "loss": 0.4426, "step": 1910 }, { "epoch": 1.638429531633415, "grad_norm": 0.6176945567131042, "learning_rate": 5.062114675803032e-06, "loss": 0.4356, "step": 1920 }, { "epoch": 1.6469646857996372, "grad_norm": 0.6743866205215454, "learning_rate": 5.012423241937515e-06, "loss": 0.4308, "step": 1930 }, { "epoch": 1.6554998399658594, "grad_norm": 0.6723053455352783, "learning_rate": 4.962730580965879e-06, "loss": 0.4344, "step": 1940 }, { "epoch": 1.6640349941320816, "grad_norm": 0.7293416261672974, "learning_rate": 4.913041601282298e-06, "loss": 0.4229, "step": 1950 }, { "epoch": 1.6725701482983037, "grad_norm": 0.6705866456031799, "learning_rate": 4.863361210917332e-06, "loss": 0.4412, "step": 1960 }, { "epoch": 1.6811053024645257, "grad_norm": 0.676863968372345, "learning_rate": 4.813694317053122e-06, "loss": 0.4423, "step": 1970 }, { "epoch": 1.689640456630748, "grad_norm": 0.659705400466919, "learning_rate": 4.764045825538702e-06, "loss": 0.4372, "step": 1980 }, { "epoch": 1.6981756107969699, "grad_norm": 0.6498283743858337, "learning_rate": 4.714420640405406e-06, "loss": 0.4344, "step": 1990 }, { "epoch": 1.706710764963192, "grad_norm": 0.6507246494293213, "learning_rate": 4.664823663382482e-06, "loss": 0.4274, "step": 2000 }, { "epoch": 1.7152459191294143, "grad_norm": 0.705443263053894, "learning_rate": 4.61525979341292e-06, "loss": 0.4365, "step": 2010 }, { "epoch": 1.7237810732956365, "grad_norm": 0.6291791200637817, "learning_rate": 4.565733926169561e-06, "loss": 0.4354, "step": 2020 }, { "epoch": 1.7323162274618586, "grad_norm": 0.6487744450569153, "learning_rate": 4.516250953571527e-06, "loss": 0.439, "step": 2030 }, { "epoch": 1.7408513816280806, "grad_norm": 0.6847457885742188, "learning_rate": 4.466815763301014e-06, "loss": 0.4245, "step": 2040 }, { "epoch": 1.7493865357943028, "grad_norm": 0.6660692095756531, "learning_rate": 4.417433238320524e-06, "loss": 0.4345, "step": 2050 }, { "epoch": 1.7579216899605248, "grad_norm": 0.6606684327125549, "learning_rate": 4.36810825639054e-06, "loss": 0.4442, "step": 2060 }, { "epoch": 1.766456844126747, "grad_norm": 0.7084365487098694, "learning_rate": 4.318845689587727e-06, "loss": 0.4348, "step": 2070 }, { "epoch": 1.7749919982929692, "grad_norm": 0.6337569952011108, "learning_rate": 4.269650403823698e-06, "loss": 0.4384, "step": 2080 }, { "epoch": 1.7835271524591914, "grad_norm": 0.6552258133888245, "learning_rate": 4.220527258364373e-06, "loss": 0.4376, "step": 2090 }, { "epoch": 1.7920623066254135, "grad_norm": 0.6303353309631348, "learning_rate": 4.171481105350021e-06, "loss": 0.4373, "step": 2100 }, { "epoch": 1.8005974607916355, "grad_norm": 0.650391697883606, "learning_rate": 4.122516789315969e-06, "loss": 0.4437, "step": 2110 }, { "epoch": 1.8091326149578577, "grad_norm": 0.6542319059371948, "learning_rate": 4.073639146714099e-06, "loss": 0.4262, "step": 2120 }, { "epoch": 1.8176677691240797, "grad_norm": 0.651104211807251, "learning_rate": 4.024853005435121e-06, "loss": 0.4363, "step": 2130 }, { "epoch": 1.8262029232903019, "grad_norm": 0.6116424798965454, "learning_rate": 3.976163184331702e-06, "loss": 0.4291, "step": 2140 }, { "epoch": 1.834738077456524, "grad_norm": 0.8228034377098083, "learning_rate": 3.927574492742473e-06, "loss": 0.4306, "step": 2150 }, { "epoch": 1.8432732316227463, "grad_norm": 0.6177427768707275, "learning_rate": 3.879091730017003e-06, "loss": 0.4274, "step": 2160 }, { "epoch": 1.8518083857889684, "grad_norm": 0.6379538178443909, "learning_rate": 3.830719685041728e-06, "loss": 0.4344, "step": 2170 }, { "epoch": 1.8603435399551904, "grad_norm": 0.6659247279167175, "learning_rate": 3.782463135766938e-06, "loss": 0.4318, "step": 2180 }, { "epoch": 1.8688786941214126, "grad_norm": 0.6846089959144592, "learning_rate": 3.7343268487348273e-06, "loss": 0.4354, "step": 2190 }, { "epoch": 1.8774138482876346, "grad_norm": 0.6371617913246155, "learning_rate": 3.6863155786086886e-06, "loss": 0.4302, "step": 2200 }, { "epoch": 1.8859490024538568, "grad_norm": 0.6230723857879639, "learning_rate": 3.6384340677032616e-06, "loss": 0.4367, "step": 2210 }, { "epoch": 1.894484156620079, "grad_norm": 0.6282981634140015, "learning_rate": 3.5906870455163175e-06, "loss": 0.4284, "step": 2220 }, { "epoch": 1.9030193107863012, "grad_norm": 0.6560613512992859, "learning_rate": 3.5430792282614978e-06, "loss": 0.4324, "step": 2230 }, { "epoch": 1.9115544649525233, "grad_norm": 0.6765207052230835, "learning_rate": 3.4956153184024765e-06, "loss": 0.4261, "step": 2240 }, { "epoch": 1.9200896191187453, "grad_norm": 0.6762336492538452, "learning_rate": 3.448300004188463e-06, "loss": 0.4304, "step": 2250 }, { "epoch": 1.9286247732849675, "grad_norm": 0.6394000053405762, "learning_rate": 3.401137959191133e-06, "loss": 0.4294, "step": 2260 }, { "epoch": 1.9371599274511895, "grad_norm": 0.6332142949104309, "learning_rate": 3.354133841842989e-06, "loss": 0.4308, "step": 2270 }, { "epoch": 1.9456950816174117, "grad_norm": 0.6541900038719177, "learning_rate": 3.3072922949772235e-06, "loss": 0.4343, "step": 2280 }, { "epoch": 1.9542302357836339, "grad_norm": 0.652773380279541, "learning_rate": 3.2606179453691255e-06, "loss": 0.4311, "step": 2290 }, { "epoch": 1.962765389949856, "grad_norm": 0.6701607704162598, "learning_rate": 3.21411540327907e-06, "loss": 0.4244, "step": 2300 }, { "epoch": 1.9713005441160782, "grad_norm": 0.6590484976768494, "learning_rate": 3.16778926199714e-06, "loss": 0.4366, "step": 2310 }, { "epoch": 1.9798356982823002, "grad_norm": 0.6363989114761353, "learning_rate": 3.1216440973894184e-06, "loss": 0.4267, "step": 2320 }, { "epoch": 1.9883708524485222, "grad_norm": 0.6702283024787903, "learning_rate": 3.075684467446014e-06, "loss": 0.4305, "step": 2330 }, { "epoch": 1.9969060066147444, "grad_norm": 0.6762630939483643, "learning_rate": 3.029914911830846e-06, "loss": 0.4339, "step": 2340 }, { "epoch": 2.005121092499733, "grad_norm": 0.6649417877197266, "learning_rate": 2.9843399514332317e-06, "loss": 0.4193, "step": 2350 }, { "epoch": 2.0136562466659553, "grad_norm": 0.6555825471878052, "learning_rate": 2.938964087921341e-06, "loss": 0.4057, "step": 2360 }, { "epoch": 2.0221914008321775, "grad_norm": 0.6427297592163086, "learning_rate": 2.8937918032975434e-06, "loss": 0.4045, "step": 2370 }, { "epoch": 2.0307265549983997, "grad_norm": 0.6486126780509949, "learning_rate": 2.848827559455696e-06, "loss": 0.4111, "step": 2380 }, { "epoch": 2.039261709164622, "grad_norm": 0.7112558484077454, "learning_rate": 2.804075797740422e-06, "loss": 0.413, "step": 2390 }, { "epoch": 2.047796863330844, "grad_norm": 0.6829844117164612, "learning_rate": 2.759540938508415e-06, "loss": 0.4103, "step": 2400 }, { "epoch": 2.056332017497066, "grad_norm": 0.6613876819610596, "learning_rate": 2.7152273806918185e-06, "loss": 0.4158, "step": 2410 }, { "epoch": 2.064867171663288, "grad_norm": 0.6096106171607971, "learning_rate": 2.6711395013637208e-06, "loss": 0.4058, "step": 2420 }, { "epoch": 2.0734023258295102, "grad_norm": 0.6055023074150085, "learning_rate": 2.6272816553058156e-06, "loss": 0.3974, "step": 2430 }, { "epoch": 2.0819374799957324, "grad_norm": 0.6151232719421387, "learning_rate": 2.5836581745782474e-06, "loss": 0.4029, "step": 2440 }, { "epoch": 2.0904726341619546, "grad_norm": 0.6236457228660583, "learning_rate": 2.5402733680917163e-06, "loss": 0.409, "step": 2450 }, { "epoch": 2.099007788328177, "grad_norm": 0.6416062712669373, "learning_rate": 2.497131521181877e-06, "loss": 0.405, "step": 2460 }, { "epoch": 2.107542942494399, "grad_norm": 0.6345195174217224, "learning_rate": 2.4542368951860303e-06, "loss": 0.3951, "step": 2470 }, { "epoch": 2.1160780966606207, "grad_norm": 0.7058767676353455, "learning_rate": 2.4115937270222305e-06, "loss": 0.4047, "step": 2480 }, { "epoch": 2.124613250826843, "grad_norm": 0.6685160398483276, "learning_rate": 2.3692062287707795e-06, "loss": 0.4037, "step": 2490 }, { "epoch": 2.133148404993065, "grad_norm": 0.6508364677429199, "learning_rate": 2.3270785872581698e-06, "loss": 0.4044, "step": 2500 }, { "epoch": 2.1416835591592873, "grad_norm": 0.6913980841636658, "learning_rate": 2.2852149636435366e-06, "loss": 0.4071, "step": 2510 }, { "epoch": 2.1502187133255095, "grad_norm": 0.6269243955612183, "learning_rate": 2.2436194930076448e-06, "loss": 0.4131, "step": 2520 }, { "epoch": 2.1587538674917317, "grad_norm": 0.6732710599899292, "learning_rate": 2.2022962839444356e-06, "loss": 0.4043, "step": 2530 }, { "epoch": 2.167289021657954, "grad_norm": 0.658445417881012, "learning_rate": 2.161249418155199e-06, "loss": 0.4036, "step": 2540 }, { "epoch": 2.1758241758241756, "grad_norm": 0.6448598504066467, "learning_rate": 2.120482950045422e-06, "loss": 0.4139, "step": 2550 }, { "epoch": 2.184359329990398, "grad_norm": 0.6652196049690247, "learning_rate": 2.080000906324292e-06, "loss": 0.399, "step": 2560 }, { "epoch": 2.19289448415662, "grad_norm": 0.6046453714370728, "learning_rate": 2.0398072856069707e-06, "loss": 0.4063, "step": 2570 }, { "epoch": 2.201429638322842, "grad_norm": 0.6543176174163818, "learning_rate": 1.9999060580196324e-06, "loss": 0.4048, "step": 2580 }, { "epoch": 2.2099647924890644, "grad_norm": 0.6523047089576721, "learning_rate": 1.9603011648073055e-06, "loss": 0.4073, "step": 2590 }, { "epoch": 2.2184999466552866, "grad_norm": 0.6173551082611084, "learning_rate": 1.9209965179445826e-06, "loss": 0.4106, "step": 2600 }, { "epoch": 2.227035100821509, "grad_norm": 0.6489232778549194, "learning_rate": 1.881995999749215e-06, "loss": 0.4103, "step": 2610 }, { "epoch": 2.2355702549877305, "grad_norm": 0.6323755383491516, "learning_rate": 1.8433034624986335e-06, "loss": 0.4012, "step": 2620 }, { "epoch": 2.2441054091539527, "grad_norm": 0.640560507774353, "learning_rate": 1.8049227280494391e-06, "loss": 0.4048, "step": 2630 }, { "epoch": 2.252640563320175, "grad_norm": 0.6368593573570251, "learning_rate": 1.7668575874599086e-06, "loss": 0.4079, "step": 2640 }, { "epoch": 2.261175717486397, "grad_norm": 0.6442722678184509, "learning_rate": 1.7291118006155178e-06, "loss": 0.4144, "step": 2650 }, { "epoch": 2.2697108716526193, "grad_norm": 0.7017993330955505, "learning_rate": 1.6916890958575687e-06, "loss": 0.4107, "step": 2660 }, { "epoch": 2.2782460258188415, "grad_norm": 0.6456113457679749, "learning_rate": 1.6545931696149208e-06, "loss": 0.4081, "step": 2670 }, { "epoch": 2.2867811799850637, "grad_norm": 0.6799325346946716, "learning_rate": 1.6178276860388736e-06, "loss": 0.4022, "step": 2680 }, { "epoch": 2.2953163341512854, "grad_norm": 0.6276426911354065, "learning_rate": 1.5813962766412427e-06, "loss": 0.4127, "step": 2690 }, { "epoch": 2.3038514883175076, "grad_norm": 0.6316121816635132, "learning_rate": 1.5453025399356542e-06, "loss": 0.4064, "step": 2700 }, { "epoch": 2.31238664248373, "grad_norm": 0.6637789011001587, "learning_rate": 1.5095500410821085e-06, "loss": 0.4089, "step": 2710 }, { "epoch": 2.320921796649952, "grad_norm": 0.6297438740730286, "learning_rate": 1.4741423115348203e-06, "loss": 0.4012, "step": 2720 }, { "epoch": 2.329456950816174, "grad_norm": 0.6420316100120544, "learning_rate": 1.439082848693406e-06, "loss": 0.403, "step": 2730 }, { "epoch": 2.3379921049823964, "grad_norm": 0.6523280143737793, "learning_rate": 1.4043751155574331e-06, "loss": 0.4037, "step": 2740 }, { "epoch": 2.346527259148618, "grad_norm": 0.6407163739204407, "learning_rate": 1.370022540384347e-06, "loss": 0.4044, "step": 2750 }, { "epoch": 2.3550624133148403, "grad_norm": 0.6137824654579163, "learning_rate": 1.3360285163508585e-06, "loss": 0.4065, "step": 2760 }, { "epoch": 2.3635975674810625, "grad_norm": 0.6386251449584961, "learning_rate": 1.3023964012177808e-06, "loss": 0.3996, "step": 2770 }, { "epoch": 2.3721327216472847, "grad_norm": 0.6598160266876221, "learning_rate": 1.2691295169983602e-06, "loss": 0.407, "step": 2780 }, { "epoch": 2.380667875813507, "grad_norm": 0.6845166683197021, "learning_rate": 1.2362311496301465e-06, "loss": 0.4093, "step": 2790 }, { "epoch": 2.389203029979729, "grad_norm": 0.6791868209838867, "learning_rate": 1.2037045486504317e-06, "loss": 0.401, "step": 2800 }, { "epoch": 2.3977381841459513, "grad_norm": 0.6840428113937378, "learning_rate": 1.171552926875265e-06, "loss": 0.4052, "step": 2810 }, { "epoch": 2.4062733383121735, "grad_norm": 0.6226551532745361, "learning_rate": 1.1397794600821155e-06, "loss": 0.4041, "step": 2820 }, { "epoch": 2.4148084924783952, "grad_norm": 0.6312896013259888, "learning_rate": 1.1083872866961816e-06, "loss": 0.4033, "step": 2830 }, { "epoch": 2.4233436466446174, "grad_norm": 0.6522102952003479, "learning_rate": 1.0773795074803917e-06, "loss": 0.3987, "step": 2840 }, { "epoch": 2.4318788008108396, "grad_norm": 0.6765321493148804, "learning_rate": 1.0467591852291287e-06, "loss": 0.4059, "step": 2850 }, { "epoch": 2.440413954977062, "grad_norm": 0.6331756114959717, "learning_rate": 1.0165293444657037e-06, "loss": 0.401, "step": 2860 }, { "epoch": 2.448949109143284, "grad_norm": 0.6368128657341003, "learning_rate": 9.866929711436035e-07, "loss": 0.4136, "step": 2870 }, { "epoch": 2.457484263309506, "grad_norm": 0.6715298891067505, "learning_rate": 9.57253012351556e-07, "loss": 0.4157, "step": 2880 }, { "epoch": 2.466019417475728, "grad_norm": 0.6294402480125427, "learning_rate": 9.282123760224327e-07, "loss": 0.4058, "step": 2890 }, { "epoch": 2.47455457164195, "grad_norm": 0.6221645474433899, "learning_rate": 8.99573930646015e-07, "loss": 0.4014, "step": 2900 }, { "epoch": 2.4830897258081723, "grad_norm": 0.642346978187561, "learning_rate": 8.713405049856588e-07, "loss": 0.41, "step": 2910 }, { "epoch": 2.4916248799743945, "grad_norm": 0.628089189529419, "learning_rate": 8.435148877988886e-07, "loss": 0.4084, "step": 2920 }, { "epoch": 2.5001600341406167, "grad_norm": 0.637294352054596, "learning_rate": 8.160998275619309e-07, "loss": 0.4025, "step": 2930 }, { "epoch": 2.508695188306839, "grad_norm": 0.6054239273071289, "learning_rate": 7.890980321982355e-07, "loss": 0.405, "step": 2940 }, { "epoch": 2.517230342473061, "grad_norm": 0.6559164524078369, "learning_rate": 7.625121688110027e-07, "loss": 0.4043, "step": 2950 }, { "epoch": 2.5257654966392833, "grad_norm": 0.665451169013977, "learning_rate": 7.363448634197412e-07, "loss": 0.4028, "step": 2960 }, { "epoch": 2.534300650805505, "grad_norm": 0.6523730158805847, "learning_rate": 7.105987007008724e-07, "loss": 0.4068, "step": 2970 }, { "epoch": 2.5428358049717272, "grad_norm": 0.6315779089927673, "learning_rate": 6.852762237324462e-07, "loss": 0.4092, "step": 2980 }, { "epoch": 2.5513709591379494, "grad_norm": 0.6356623768806458, "learning_rate": 6.603799337429328e-07, "loss": 0.4099, "step": 2990 }, { "epoch": 2.5599061133041716, "grad_norm": 0.6033786535263062, "learning_rate": 6.359122898641717e-07, "loss": 0.4084, "step": 3000 }, { "epoch": 2.568441267470394, "grad_norm": 0.6288996338844299, "learning_rate": 6.118757088884686e-07, "loss": 0.4131, "step": 3010 }, { "epoch": 2.5769764216366156, "grad_norm": 0.6407638192176819, "learning_rate": 5.882725650298787e-07, "loss": 0.409, "step": 3020 }, { "epoch": 2.5855115758028377, "grad_norm": 0.6419194340705872, "learning_rate": 5.651051896896908e-07, "loss": 0.4038, "step": 3030 }, { "epoch": 2.59404672996906, "grad_norm": 0.6341270208358765, "learning_rate": 5.42375871226144e-07, "loss": 0.4064, "step": 3040 }, { "epoch": 2.602581884135282, "grad_norm": 0.647345244884491, "learning_rate": 5.200868547283972e-07, "loss": 0.4081, "step": 3050 }, { "epoch": 2.6111170383015043, "grad_norm": 0.6366192102432251, "learning_rate": 4.982403417947662e-07, "loss": 0.4033, "step": 3060 }, { "epoch": 2.6196521924677265, "grad_norm": 0.638552725315094, "learning_rate": 4.768384903152634e-07, "loss": 0.4096, "step": 3070 }, { "epoch": 2.6281873466339487, "grad_norm": 0.6206834316253662, "learning_rate": 4.558834142584556e-07, "loss": 0.4065, "step": 3080 }, { "epoch": 2.636722500800171, "grad_norm": 0.6011072993278503, "learning_rate": 4.3537718346264857e-07, "loss": 0.4029, "step": 3090 }, { "epoch": 2.645257654966393, "grad_norm": 0.6577970385551453, "learning_rate": 4.153218234314443e-07, "loss": 0.4119, "step": 3100 }, { "epoch": 2.653792809132615, "grad_norm": 0.6269943118095398, "learning_rate": 3.957193151336719e-07, "loss": 0.4078, "step": 3110 }, { "epoch": 2.662327963298837, "grad_norm": 0.6127352118492126, "learning_rate": 3.765715948077137e-07, "loss": 0.4014, "step": 3120 }, { "epoch": 2.6708631174650592, "grad_norm": 0.6328156590461731, "learning_rate": 3.578805537702573e-07, "loss": 0.4057, "step": 3130 }, { "epoch": 2.6793982716312814, "grad_norm": 0.6262829899787903, "learning_rate": 3.396480382294792e-07, "loss": 0.4026, "step": 3140 }, { "epoch": 2.6879334257975036, "grad_norm": 0.631956934928894, "learning_rate": 3.2187584910268464e-07, "loss": 0.4003, "step": 3150 }, { "epoch": 2.6964685799637254, "grad_norm": 0.628641664981842, "learning_rate": 3.045657418384218e-07, "loss": 0.4075, "step": 3160 }, { "epoch": 2.7050037341299475, "grad_norm": 0.6435490250587463, "learning_rate": 2.877194262430916e-07, "loss": 0.3966, "step": 3170 }, { "epoch": 2.7135388882961697, "grad_norm": 0.6484223008155823, "learning_rate": 2.7133856631205216e-07, "loss": 0.4076, "step": 3180 }, { "epoch": 2.722074042462392, "grad_norm": 0.630936324596405, "learning_rate": 2.5542478006526703e-07, "loss": 0.3984, "step": 3190 }, { "epoch": 2.730609196628614, "grad_norm": 0.6019722819328308, "learning_rate": 2.399796393874837e-07, "loss": 0.4045, "step": 3200 }, { "epoch": 2.7391443507948363, "grad_norm": 0.6216541528701782, "learning_rate": 2.2500466987296466e-07, "loss": 0.4136, "step": 3210 }, { "epoch": 2.7476795049610585, "grad_norm": 0.6513262987136841, "learning_rate": 2.105013506748027e-07, "loss": 0.405, "step": 3220 }, { "epoch": 2.7562146591272807, "grad_norm": 0.6242448091506958, "learning_rate": 1.9647111435881494e-07, "loss": 0.4, "step": 3230 }, { "epoch": 2.764749813293503, "grad_norm": 0.6355422139167786, "learning_rate": 1.829153467620426e-07, "loss": 0.402, "step": 3240 }, { "epoch": 2.7732849674597246, "grad_norm": 0.5879780054092407, "learning_rate": 1.6983538685586277e-07, "loss": 0.4015, "step": 3250 }, { "epoch": 2.781820121625947, "grad_norm": 0.6436604261398315, "learning_rate": 1.5723252661373367e-07, "loss": 0.4025, "step": 3260 }, { "epoch": 2.790355275792169, "grad_norm": 0.6310527920722961, "learning_rate": 1.451080108835784e-07, "loss": 0.406, "step": 3270 }, { "epoch": 2.798890429958391, "grad_norm": 0.6294417381286621, "learning_rate": 1.3346303726482544e-07, "loss": 0.3997, "step": 3280 }, { "epoch": 2.8074255841246134, "grad_norm": 0.6856746077537537, "learning_rate": 1.222987559901162e-07, "loss": 0.4059, "step": 3290 }, { "epoch": 2.815960738290835, "grad_norm": 0.8050450682640076, "learning_rate": 1.1161626981169072e-07, "loss": 0.4063, "step": 3300 }, { "epoch": 2.8244958924570573, "grad_norm": 0.6034400463104248, "learning_rate": 1.014166338924627e-07, "loss": 0.3985, "step": 3310 }, { "epoch": 2.8330310466232795, "grad_norm": 0.6405271291732788, "learning_rate": 9.17008557017962e-08, "loss": 0.3951, "step": 3320 }, { "epoch": 2.8415662007895017, "grad_norm": 0.6416633725166321, "learning_rate": 8.24698949159919e-08, "loss": 0.4049, "step": 3330 }, { "epoch": 2.850101354955724, "grad_norm": 0.6081260442733765, "learning_rate": 7.372466332349726e-08, "loss": 0.4023, "step": 3340 }, { "epoch": 2.858636509121946, "grad_norm": 0.6582528352737427, "learning_rate": 6.546602473484153e-08, "loss": 0.4095, "step": 3350 }, { "epoch": 2.8671716632881683, "grad_norm": 0.6263580918312073, "learning_rate": 5.769479489731666e-08, "loss": 0.4099, "step": 3360 }, { "epoch": 2.8757068174543905, "grad_norm": 0.6049970984458923, "learning_rate": 5.0411741414396756e-08, "loss": 0.3992, "step": 3370 }, { "epoch": 2.8842419716206122, "grad_norm": 0.6298121809959412, "learning_rate": 4.3617583669922636e-08, "loss": 0.4035, "step": 3380 }, { "epoch": 2.8927771257868344, "grad_norm": 0.617342472076416, "learning_rate": 3.73129927570437e-08, "loss": 0.4079, "step": 3390 }, { "epoch": 2.9013122799530566, "grad_norm": 0.6217492818832397, "learning_rate": 3.1498591411928125e-08, "loss": 0.4021, "step": 3400 }, { "epoch": 2.909847434119279, "grad_norm": 0.618754506111145, "learning_rate": 2.6174953952254334e-08, "loss": 0.3977, "step": 3410 }, { "epoch": 2.918382588285501, "grad_norm": 0.5799270868301392, "learning_rate": 2.134260622048301e-08, "loss": 0.394, "step": 3420 }, { "epoch": 2.9269177424517228, "grad_norm": 0.6517247557640076, "learning_rate": 1.700202553191588e-08, "loss": 0.4023, "step": 3430 }, { "epoch": 2.935452896617945, "grad_norm": 0.6470068693161011, "learning_rate": 1.315364062754898e-08, "loss": 0.4056, "step": 3440 }, { "epoch": 2.943988050784167, "grad_norm": 0.6544038653373718, "learning_rate": 9.797831631724319e-09, "loss": 0.4105, "step": 3450 }, { "epoch": 2.9525232049503893, "grad_norm": 0.6202620267868042, "learning_rate": 6.9349300145826745e-09, "loss": 0.4102, "step": 3460 }, { "epoch": 2.9610583591166115, "grad_norm": 0.6056284308433533, "learning_rate": 4.56521855932257e-09, "loss": 0.4006, "step": 3470 }, { "epoch": 2.9695935132828337, "grad_norm": 0.6306059956550598, "learning_rate": 2.6889313342692848e-09, "loss": 0.404, "step": 3480 }, { "epoch": 2.978128667449056, "grad_norm": 0.6009540557861328, "learning_rate": 1.3062536697533433e-09, "loss": 0.3995, "step": 3490 }, { "epoch": 2.986663821615278, "grad_norm": 0.649553656578064, "learning_rate": 4.173221398046101e-10, "loss": 0.4005, "step": 3500 }, { "epoch": 2.9951989757815003, "grad_norm": 0.6136988401412964, "learning_rate": 2.2224548664184777e-11, "loss": 0.4075, "step": 3510 } ], "logging_steps": 10, "max_steps": 3513, "num_input_tokens_seen": 0, "num_train_epochs": 3, "save_steps": 500.0, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 1.1182668447261655e+18, "train_batch_size": 4, "trial_name": null, "trial_params": null }