| { | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 2.997759522031367, | |
| "eval_steps": 500, | |
| "global_step": 3513, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0, | |
| "eval_loss": 2.1641297340393066, | |
| "eval_runtime": 1.9384, | |
| "eval_samples_per_second": 10.318, | |
| "eval_steps_per_second": 1.032, | |
| "step": 0 | |
| }, | |
| { | |
| "epoch": 0.0008535154166222128, | |
| "grad_norm": 4.246492385864258, | |
| "learning_rate": 2.840909090909091e-08, | |
| "loss": 1.0032, | |
| "step": 1 | |
| }, | |
| { | |
| "epoch": 0.008535154166222128, | |
| "grad_norm": 4.319986820220947, | |
| "learning_rate": 2.840909090909091e-07, | |
| "loss": 0.983, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.017070308332444255, | |
| "grad_norm": 3.9372925758361816, | |
| "learning_rate": 5.681818181818182e-07, | |
| "loss": 0.9573, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.02560546249866638, | |
| "grad_norm": 3.070388078689575, | |
| "learning_rate": 8.522727272727273e-07, | |
| "loss": 0.9306, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.03414061666488851, | |
| "grad_norm": 1.9363492727279663, | |
| "learning_rate": 1.1363636363636364e-06, | |
| "loss": 0.8481, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.04267577083111064, | |
| "grad_norm": 1.282257318496704, | |
| "learning_rate": 1.4204545454545458e-06, | |
| "loss": 0.7776, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.05121092499733276, | |
| "grad_norm": 1.0056549310684204, | |
| "learning_rate": 1.7045454545454546e-06, | |
| "loss": 0.7324, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.05974607916355489, | |
| "grad_norm": 0.8602527379989624, | |
| "learning_rate": 1.9886363636363638e-06, | |
| "loss": 0.6952, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 0.06828123332977702, | |
| "grad_norm": 0.822441577911377, | |
| "learning_rate": 2.2727272727272728e-06, | |
| "loss": 0.6724, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.07681638749599914, | |
| "grad_norm": 0.7585397958755493, | |
| "learning_rate": 2.556818181818182e-06, | |
| "loss": 0.6635, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 0.08535154166222128, | |
| "grad_norm": 0.7706250548362732, | |
| "learning_rate": 2.8409090909090916e-06, | |
| "loss": 0.6502, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.0938866958284434, | |
| "grad_norm": 0.8293640613555908, | |
| "learning_rate": 3.125e-06, | |
| "loss": 0.6334, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 0.10242184999466553, | |
| "grad_norm": 0.781350314617157, | |
| "learning_rate": 3.409090909090909e-06, | |
| "loss": 0.6268, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 0.11095700416088766, | |
| "grad_norm": 0.7660292387008667, | |
| "learning_rate": 3.6931818181818186e-06, | |
| "loss": 0.6137, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 0.11949215832710978, | |
| "grad_norm": 0.8327275514602661, | |
| "learning_rate": 3.9772727272727275e-06, | |
| "loss": 0.6159, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 0.1280273124933319, | |
| "grad_norm": 0.8584268093109131, | |
| "learning_rate": 4.2613636363636365e-06, | |
| "loss": 0.5995, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 0.13656246665955404, | |
| "grad_norm": 0.7809672951698303, | |
| "learning_rate": 4.5454545454545455e-06, | |
| "loss": 0.6043, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 0.14509762082577615, | |
| "grad_norm": 0.8208410739898682, | |
| "learning_rate": 4.829545454545455e-06, | |
| "loss": 0.6015, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 0.1536327749919983, | |
| "grad_norm": 0.828789472579956, | |
| "learning_rate": 5.113636363636364e-06, | |
| "loss": 0.5871, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 0.16216792915822043, | |
| "grad_norm": 0.7799698114395142, | |
| "learning_rate": 5.397727272727273e-06, | |
| "loss": 0.5946, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 0.17070308332444256, | |
| "grad_norm": 0.7569297552108765, | |
| "learning_rate": 5.681818181818183e-06, | |
| "loss": 0.5868, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.17923823749066467, | |
| "grad_norm": 0.7858031392097473, | |
| "learning_rate": 5.965909090909091e-06, | |
| "loss": 0.5732, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 0.1877733916568868, | |
| "grad_norm": 0.7812625765800476, | |
| "learning_rate": 6.25e-06, | |
| "loss": 0.5745, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 0.19630854582310894, | |
| "grad_norm": 0.7922148704528809, | |
| "learning_rate": 6.53409090909091e-06, | |
| "loss": 0.5742, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 0.20484369998933105, | |
| "grad_norm": 0.8420217633247375, | |
| "learning_rate": 6.818181818181818e-06, | |
| "loss": 0.5593, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 0.2133788541555532, | |
| "grad_norm": 0.8149869441986084, | |
| "learning_rate": 7.102272727272727e-06, | |
| "loss": 0.5694, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 0.22191400832177532, | |
| "grad_norm": 0.8200048208236694, | |
| "learning_rate": 7.386363636363637e-06, | |
| "loss": 0.5609, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 0.23044916248799743, | |
| "grad_norm": 0.7718726396560669, | |
| "learning_rate": 7.670454545454547e-06, | |
| "loss": 0.5596, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 0.23898431665421957, | |
| "grad_norm": 0.8079963326454163, | |
| "learning_rate": 7.954545454545455e-06, | |
| "loss": 0.5515, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 0.2475194708204417, | |
| "grad_norm": 0.8398913741111755, | |
| "learning_rate": 8.238636363636365e-06, | |
| "loss": 0.5615, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 0.2560546249866638, | |
| "grad_norm": 0.8522515892982483, | |
| "learning_rate": 8.522727272727273e-06, | |
| "loss": 0.5405, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.26458977915288595, | |
| "grad_norm": 0.965617299079895, | |
| "learning_rate": 8.806818181818183e-06, | |
| "loss": 0.5438, | |
| "step": 310 | |
| }, | |
| { | |
| "epoch": 0.2731249333191081, | |
| "grad_norm": 0.834233283996582, | |
| "learning_rate": 9.090909090909091e-06, | |
| "loss": 0.5342, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 0.2816600874853302, | |
| "grad_norm": 0.8867103457450867, | |
| "learning_rate": 9.375000000000001e-06, | |
| "loss": 0.5272, | |
| "step": 330 | |
| }, | |
| { | |
| "epoch": 0.2901952416515523, | |
| "grad_norm": 0.767537534236908, | |
| "learning_rate": 9.65909090909091e-06, | |
| "loss": 0.5319, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 0.29873039581777444, | |
| "grad_norm": 0.8806461095809937, | |
| "learning_rate": 9.943181818181819e-06, | |
| "loss": 0.5322, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 0.3072655499839966, | |
| "grad_norm": 0.7467280030250549, | |
| "learning_rate": 9.999841959480543e-06, | |
| "loss": 0.5267, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 0.3158007041502187, | |
| "grad_norm": 0.7319239974021912, | |
| "learning_rate": 9.999199936992976e-06, | |
| "loss": 0.5221, | |
| "step": 370 | |
| }, | |
| { | |
| "epoch": 0.32433585831644085, | |
| "grad_norm": 0.7590662837028503, | |
| "learning_rate": 9.998064118371841e-06, | |
| "loss": 0.525, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 0.332871012482663, | |
| "grad_norm": 0.7705526351928711, | |
| "learning_rate": 9.996434615807662e-06, | |
| "loss": 0.5254, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 0.3414061666488851, | |
| "grad_norm": 0.8061032295227051, | |
| "learning_rate": 9.994311590254603e-06, | |
| "loss": 0.5221, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 0.3499413208151072, | |
| "grad_norm": 0.7729495167732239, | |
| "learning_rate": 9.991695251414584e-06, | |
| "loss": 0.53, | |
| "step": 410 | |
| }, | |
| { | |
| "epoch": 0.35847647498132934, | |
| "grad_norm": 0.768975019454956, | |
| "learning_rate": 9.988585857716555e-06, | |
| "loss": 0.5099, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 0.3670116291475515, | |
| "grad_norm": 0.7381827235221863, | |
| "learning_rate": 9.984983716290981e-06, | |
| "loss": 0.5182, | |
| "step": 430 | |
| }, | |
| { | |
| "epoch": 0.3755467833137736, | |
| "grad_norm": 0.8239282965660095, | |
| "learning_rate": 9.980889182939495e-06, | |
| "loss": 0.52, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 0.38408193747999575, | |
| "grad_norm": 0.8047374486923218, | |
| "learning_rate": 9.976302662099758e-06, | |
| "loss": 0.5091, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 0.3926170916462179, | |
| "grad_norm": 0.7568238377571106, | |
| "learning_rate": 9.971224606805514e-06, | |
| "loss": 0.5171, | |
| "step": 460 | |
| }, | |
| { | |
| "epoch": 0.40115224581243997, | |
| "grad_norm": 0.7178462147712708, | |
| "learning_rate": 9.965655518641837e-06, | |
| "loss": 0.5119, | |
| "step": 470 | |
| }, | |
| { | |
| "epoch": 0.4096873999786621, | |
| "grad_norm": 0.8041744232177734, | |
| "learning_rate": 9.959595947695588e-06, | |
| "loss": 0.506, | |
| "step": 480 | |
| }, | |
| { | |
| "epoch": 0.41822255414488424, | |
| "grad_norm": 0.7083892226219177, | |
| "learning_rate": 9.953046492501082e-06, | |
| "loss": 0.5066, | |
| "step": 490 | |
| }, | |
| { | |
| "epoch": 0.4267577083111064, | |
| "grad_norm": 0.7124646902084351, | |
| "learning_rate": 9.946007799980963e-06, | |
| "loss": 0.5098, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.4352928624773285, | |
| "grad_norm": 0.7167072892189026, | |
| "learning_rate": 9.938480565382312e-06, | |
| "loss": 0.513, | |
| "step": 510 | |
| }, | |
| { | |
| "epoch": 0.44382801664355065, | |
| "grad_norm": 0.7211044430732727, | |
| "learning_rate": 9.930465532207967e-06, | |
| "loss": 0.5094, | |
| "step": 520 | |
| }, | |
| { | |
| "epoch": 0.45236317080977273, | |
| "grad_norm": 0.7868006229400635, | |
| "learning_rate": 9.921963492143083e-06, | |
| "loss": 0.5131, | |
| "step": 530 | |
| }, | |
| { | |
| "epoch": 0.46089832497599487, | |
| "grad_norm": 0.7130579948425293, | |
| "learning_rate": 9.912975284976942e-06, | |
| "loss": 0.506, | |
| "step": 540 | |
| }, | |
| { | |
| "epoch": 0.469433479142217, | |
| "grad_norm": 0.7939994931221008, | |
| "learning_rate": 9.903501798519994e-06, | |
| "loss": 0.5025, | |
| "step": 550 | |
| }, | |
| { | |
| "epoch": 0.47796863330843914, | |
| "grad_norm": 0.7138738036155701, | |
| "learning_rate": 9.893543968516166e-06, | |
| "loss": 0.5037, | |
| "step": 560 | |
| }, | |
| { | |
| "epoch": 0.4865037874746613, | |
| "grad_norm": 0.716701328754425, | |
| "learning_rate": 9.883102778550434e-06, | |
| "loss": 0.5034, | |
| "step": 570 | |
| }, | |
| { | |
| "epoch": 0.4950389416408834, | |
| "grad_norm": 0.7295582294464111, | |
| "learning_rate": 9.872179259951672e-06, | |
| "loss": 0.4949, | |
| "step": 580 | |
| }, | |
| { | |
| "epoch": 0.5035740958071055, | |
| "grad_norm": 0.7438467144966125, | |
| "learning_rate": 9.860774491690777e-06, | |
| "loss": 0.4969, | |
| "step": 590 | |
| }, | |
| { | |
| "epoch": 0.5121092499733276, | |
| "grad_norm": 0.696615993976593, | |
| "learning_rate": 9.848889600274101e-06, | |
| "loss": 0.4926, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 0.5206444041395498, | |
| "grad_norm": 0.6948277354240417, | |
| "learning_rate": 9.836525759632171e-06, | |
| "loss": 0.5034, | |
| "step": 610 | |
| }, | |
| { | |
| "epoch": 0.5291795583057719, | |
| "grad_norm": 0.8091464042663574, | |
| "learning_rate": 9.823684191003744e-06, | |
| "loss": 0.5015, | |
| "step": 620 | |
| }, | |
| { | |
| "epoch": 0.537714712471994, | |
| "grad_norm": 0.7501440048217773, | |
| "learning_rate": 9.810366162815171e-06, | |
| "loss": 0.5024, | |
| "step": 630 | |
| }, | |
| { | |
| "epoch": 0.5462498666382162, | |
| "grad_norm": 0.7048625946044922, | |
| "learning_rate": 9.796572990555112e-06, | |
| "loss": 0.4955, | |
| "step": 640 | |
| }, | |
| { | |
| "epoch": 0.5547850208044383, | |
| "grad_norm": 0.7141663432121277, | |
| "learning_rate": 9.782306036644596e-06, | |
| "loss": 0.5045, | |
| "step": 650 | |
| }, | |
| { | |
| "epoch": 0.5633201749706604, | |
| "grad_norm": 0.7329632639884949, | |
| "learning_rate": 9.767566710302453e-06, | |
| "loss": 0.5038, | |
| "step": 660 | |
| }, | |
| { | |
| "epoch": 0.5718553291368825, | |
| "grad_norm": 0.699938952922821, | |
| "learning_rate": 9.752356467406114e-06, | |
| "loss": 0.5066, | |
| "step": 670 | |
| }, | |
| { | |
| "epoch": 0.5803904833031046, | |
| "grad_norm": 0.6912060976028442, | |
| "learning_rate": 9.736676810347807e-06, | |
| "loss": 0.4918, | |
| "step": 680 | |
| }, | |
| { | |
| "epoch": 0.5889256374693268, | |
| "grad_norm": 0.6718906164169312, | |
| "learning_rate": 9.720529287886159e-06, | |
| "loss": 0.4932, | |
| "step": 690 | |
| }, | |
| { | |
| "epoch": 0.5974607916355489, | |
| "grad_norm": 0.6830474734306335, | |
| "learning_rate": 9.703915494993215e-06, | |
| "loss": 0.489, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 0.6059959458017711, | |
| "grad_norm": 0.6920751333236694, | |
| "learning_rate": 9.686837072696902e-06, | |
| "loss": 0.4976, | |
| "step": 710 | |
| }, | |
| { | |
| "epoch": 0.6145310999679932, | |
| "grad_norm": 0.695093035697937, | |
| "learning_rate": 9.669295707918928e-06, | |
| "loss": 0.4877, | |
| "step": 720 | |
| }, | |
| { | |
| "epoch": 0.6230662541342153, | |
| "grad_norm": 0.683686375617981, | |
| "learning_rate": 9.651293133308159e-06, | |
| "loss": 0.4939, | |
| "step": 730 | |
| }, | |
| { | |
| "epoch": 0.6316014083004374, | |
| "grad_norm": 0.7766408920288086, | |
| "learning_rate": 9.632831127069479e-06, | |
| "loss": 0.4993, | |
| "step": 740 | |
| }, | |
| { | |
| "epoch": 0.6401365624666595, | |
| "grad_norm": 0.7032186985015869, | |
| "learning_rate": 9.61391151278814e-06, | |
| "loss": 0.4896, | |
| "step": 750 | |
| }, | |
| { | |
| "epoch": 0.6486717166328817, | |
| "grad_norm": 0.7620243430137634, | |
| "learning_rate": 9.594536159249654e-06, | |
| "loss": 0.4926, | |
| "step": 760 | |
| }, | |
| { | |
| "epoch": 0.6572068707991038, | |
| "grad_norm": 0.6803762316703796, | |
| "learning_rate": 9.57470698025518e-06, | |
| "loss": 0.4902, | |
| "step": 770 | |
| }, | |
| { | |
| "epoch": 0.665742024965326, | |
| "grad_norm": 0.6878206133842468, | |
| "learning_rate": 9.55442593443251e-06, | |
| "loss": 0.4841, | |
| "step": 780 | |
| }, | |
| { | |
| "epoch": 0.674277179131548, | |
| "grad_norm": 0.7530574202537537, | |
| "learning_rate": 9.533695025042591e-06, | |
| "loss": 0.4839, | |
| "step": 790 | |
| }, | |
| { | |
| "epoch": 0.6828123332977702, | |
| "grad_norm": 0.7192838191986084, | |
| "learning_rate": 9.512516299781665e-06, | |
| "loss": 0.4841, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 0.6913474874639923, | |
| "grad_norm": 0.7086024880409241, | |
| "learning_rate": 9.490891850578993e-06, | |
| "loss": 0.4869, | |
| "step": 810 | |
| }, | |
| { | |
| "epoch": 0.6998826416302144, | |
| "grad_norm": 0.6931254267692566, | |
| "learning_rate": 9.468823813390236e-06, | |
| "loss": 0.4834, | |
| "step": 820 | |
| }, | |
| { | |
| "epoch": 0.7084177957964366, | |
| "grad_norm": 0.6703010201454163, | |
| "learning_rate": 9.446314367986476e-06, | |
| "loss": 0.4904, | |
| "step": 830 | |
| }, | |
| { | |
| "epoch": 0.7169529499626587, | |
| "grad_norm": 0.721376359462738, | |
| "learning_rate": 9.423365737738898e-06, | |
| "loss": 0.4896, | |
| "step": 840 | |
| }, | |
| { | |
| "epoch": 0.7254881041288809, | |
| "grad_norm": 0.7349815964698792, | |
| "learning_rate": 9.399980189399191e-06, | |
| "loss": 0.4938, | |
| "step": 850 | |
| }, | |
| { | |
| "epoch": 0.734023258295103, | |
| "grad_norm": 0.7041159272193909, | |
| "learning_rate": 9.376160032875638e-06, | |
| "loss": 0.4889, | |
| "step": 860 | |
| }, | |
| { | |
| "epoch": 0.742558412461325, | |
| "grad_norm": 0.7001343965530396, | |
| "learning_rate": 9.351907621004963e-06, | |
| "loss": 0.4908, | |
| "step": 870 | |
| }, | |
| { | |
| "epoch": 0.7510935666275472, | |
| "grad_norm": 0.7233178019523621, | |
| "learning_rate": 9.327225349319922e-06, | |
| "loss": 0.4856, | |
| "step": 880 | |
| }, | |
| { | |
| "epoch": 0.7596287207937693, | |
| "grad_norm": 0.6933467984199524, | |
| "learning_rate": 9.30211565581269e-06, | |
| "loss": 0.4817, | |
| "step": 890 | |
| }, | |
| { | |
| "epoch": 0.7681638749599915, | |
| "grad_norm": 0.6627389192581177, | |
| "learning_rate": 9.276581020694048e-06, | |
| "loss": 0.4804, | |
| "step": 900 | |
| }, | |
| { | |
| "epoch": 0.7766990291262136, | |
| "grad_norm": 0.6689618229866028, | |
| "learning_rate": 9.250623966148392e-06, | |
| "loss": 0.4854, | |
| "step": 910 | |
| }, | |
| { | |
| "epoch": 0.7852341832924358, | |
| "grad_norm": 0.7105383276939392, | |
| "learning_rate": 9.224247056084617e-06, | |
| "loss": 0.4831, | |
| "step": 920 | |
| }, | |
| { | |
| "epoch": 0.7937693374586579, | |
| "grad_norm": 0.7106824517250061, | |
| "learning_rate": 9.197452895882857e-06, | |
| "loss": 0.4802, | |
| "step": 930 | |
| }, | |
| { | |
| "epoch": 0.8023044916248799, | |
| "grad_norm": 0.6922078132629395, | |
| "learning_rate": 9.170244132137143e-06, | |
| "loss": 0.4854, | |
| "step": 940 | |
| }, | |
| { | |
| "epoch": 0.8108396457911021, | |
| "grad_norm": 0.7484738230705261, | |
| "learning_rate": 9.142623452393984e-06, | |
| "loss": 0.4708, | |
| "step": 950 | |
| }, | |
| { | |
| "epoch": 0.8193747999573242, | |
| "grad_norm": 0.7449445128440857, | |
| "learning_rate": 9.114593584886902e-06, | |
| "loss": 0.4786, | |
| "step": 960 | |
| }, | |
| { | |
| "epoch": 0.8279099541235464, | |
| "grad_norm": 0.6751102209091187, | |
| "learning_rate": 9.086157298266957e-06, | |
| "loss": 0.4783, | |
| "step": 970 | |
| }, | |
| { | |
| "epoch": 0.8364451082897685, | |
| "grad_norm": 0.7621346712112427, | |
| "learning_rate": 9.057317401329271e-06, | |
| "loss": 0.4868, | |
| "step": 980 | |
| }, | |
| { | |
| "epoch": 0.8449802624559906, | |
| "grad_norm": 0.7197836637496948, | |
| "learning_rate": 9.028076742735583e-06, | |
| "loss": 0.479, | |
| "step": 990 | |
| }, | |
| { | |
| "epoch": 0.8535154166222128, | |
| "grad_norm": 0.7285887598991394, | |
| "learning_rate": 8.998438210732882e-06, | |
| "loss": 0.4851, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 0.8620505707884348, | |
| "grad_norm": 0.7014701962471008, | |
| "learning_rate": 8.968404732868112e-06, | |
| "loss": 0.4703, | |
| "step": 1010 | |
| }, | |
| { | |
| "epoch": 0.870585724954657, | |
| "grad_norm": 0.6849453449249268, | |
| "learning_rate": 8.937979275699012e-06, | |
| "loss": 0.4729, | |
| "step": 1020 | |
| }, | |
| { | |
| "epoch": 0.8791208791208791, | |
| "grad_norm": 0.7209568023681641, | |
| "learning_rate": 8.907164844501084e-06, | |
| "loss": 0.4777, | |
| "step": 1030 | |
| }, | |
| { | |
| "epoch": 0.8876560332871013, | |
| "grad_norm": 0.6904247403144836, | |
| "learning_rate": 8.875964482970756e-06, | |
| "loss": 0.4731, | |
| "step": 1040 | |
| }, | |
| { | |
| "epoch": 0.8961911874533234, | |
| "grad_norm": 0.6657469272613525, | |
| "learning_rate": 8.844381272924733e-06, | |
| "loss": 0.4773, | |
| "step": 1050 | |
| }, | |
| { | |
| "epoch": 0.9047263416195455, | |
| "grad_norm": 0.6649881601333618, | |
| "learning_rate": 8.812418333995597e-06, | |
| "loss": 0.4659, | |
| "step": 1060 | |
| }, | |
| { | |
| "epoch": 0.9132614957857677, | |
| "grad_norm": 0.6824942231178284, | |
| "learning_rate": 8.780078823323667e-06, | |
| "loss": 0.4736, | |
| "step": 1070 | |
| }, | |
| { | |
| "epoch": 0.9217966499519897, | |
| "grad_norm": 0.6798464059829712, | |
| "learning_rate": 8.74736593524514e-06, | |
| "loss": 0.4695, | |
| "step": 1080 | |
| }, | |
| { | |
| "epoch": 0.9303318041182119, | |
| "grad_norm": 0.6828529238700867, | |
| "learning_rate": 8.71428290097658e-06, | |
| "loss": 0.4678, | |
| "step": 1090 | |
| }, | |
| { | |
| "epoch": 0.938866958284434, | |
| "grad_norm": 0.6548129916191101, | |
| "learning_rate": 8.680832988295757e-06, | |
| "loss": 0.4748, | |
| "step": 1100 | |
| }, | |
| { | |
| "epoch": 0.9474021124506561, | |
| "grad_norm": 0.7184150815010071, | |
| "learning_rate": 8.647019501218862e-06, | |
| "loss": 0.4851, | |
| "step": 1110 | |
| }, | |
| { | |
| "epoch": 0.9559372666168783, | |
| "grad_norm": 0.6707058548927307, | |
| "learning_rate": 8.612845779674166e-06, | |
| "loss": 0.4621, | |
| "step": 1120 | |
| }, | |
| { | |
| "epoch": 0.9644724207831004, | |
| "grad_norm": 0.6430131793022156, | |
| "learning_rate": 8.578315199172105e-06, | |
| "loss": 0.4733, | |
| "step": 1130 | |
| }, | |
| { | |
| "epoch": 0.9730075749493226, | |
| "grad_norm": 0.652946949005127, | |
| "learning_rate": 8.543431170471872e-06, | |
| "loss": 0.469, | |
| "step": 1140 | |
| }, | |
| { | |
| "epoch": 0.9815427291155446, | |
| "grad_norm": 0.678126871585846, | |
| "learning_rate": 8.508197139244515e-06, | |
| "loss": 0.4645, | |
| "step": 1150 | |
| }, | |
| { | |
| "epoch": 0.9900778832817668, | |
| "grad_norm": 0.6810508966445923, | |
| "learning_rate": 8.472616585732597e-06, | |
| "loss": 0.4722, | |
| "step": 1160 | |
| }, | |
| { | |
| "epoch": 0.9986130374479889, | |
| "grad_norm": 0.6804456114768982, | |
| "learning_rate": 8.436693024406427e-06, | |
| "loss": 0.4796, | |
| "step": 1170 | |
| }, | |
| { | |
| "epoch": 1.0068281233329777, | |
| "grad_norm": 0.7665020227432251, | |
| "learning_rate": 8.400430003616924e-06, | |
| "loss": 0.4501, | |
| "step": 1180 | |
| }, | |
| { | |
| "epoch": 1.0153632774991999, | |
| "grad_norm": 0.7121028304100037, | |
| "learning_rate": 8.363831105245125e-06, | |
| "loss": 0.4436, | |
| "step": 1190 | |
| }, | |
| { | |
| "epoch": 1.023898431665422, | |
| "grad_norm": 0.7494065761566162, | |
| "learning_rate": 8.326899944348389e-06, | |
| "loss": 0.4459, | |
| "step": 1200 | |
| }, | |
| { | |
| "epoch": 1.032433585831644, | |
| "grad_norm": 0.6524202227592468, | |
| "learning_rate": 8.289640168803308e-06, | |
| "loss": 0.4445, | |
| "step": 1210 | |
| }, | |
| { | |
| "epoch": 1.0409687399978662, | |
| "grad_norm": 0.6446959972381592, | |
| "learning_rate": 8.25205545894541e-06, | |
| "loss": 0.4406, | |
| "step": 1220 | |
| }, | |
| { | |
| "epoch": 1.0495038941640884, | |
| "grad_norm": 0.6765570044517517, | |
| "learning_rate": 8.214149527205606e-06, | |
| "loss": 0.4538, | |
| "step": 1230 | |
| }, | |
| { | |
| "epoch": 1.0580390483303104, | |
| "grad_norm": 0.6525077223777771, | |
| "learning_rate": 8.17592611774352e-06, | |
| "loss": 0.4492, | |
| "step": 1240 | |
| }, | |
| { | |
| "epoch": 1.0665742024965326, | |
| "grad_norm": 0.6981220245361328, | |
| "learning_rate": 8.13738900607764e-06, | |
| "loss": 0.4428, | |
| "step": 1250 | |
| }, | |
| { | |
| "epoch": 1.0751093566627548, | |
| "grad_norm": 0.7378373742103577, | |
| "learning_rate": 8.098541998712405e-06, | |
| "loss": 0.442, | |
| "step": 1260 | |
| }, | |
| { | |
| "epoch": 1.083644510828977, | |
| "grad_norm": 0.6806818246841431, | |
| "learning_rate": 8.059388932762207e-06, | |
| "loss": 0.4428, | |
| "step": 1270 | |
| }, | |
| { | |
| "epoch": 1.092179664995199, | |
| "grad_norm": 0.704845130443573, | |
| "learning_rate": 8.019933675572389e-06, | |
| "loss": 0.4402, | |
| "step": 1280 | |
| }, | |
| { | |
| "epoch": 1.100714819161421, | |
| "grad_norm": 0.8347475528717041, | |
| "learning_rate": 7.980180124337239e-06, | |
| "loss": 0.448, | |
| "step": 1290 | |
| }, | |
| { | |
| "epoch": 1.1092499733276433, | |
| "grad_norm": 0.7030121684074402, | |
| "learning_rate": 7.940132205715053e-06, | |
| "loss": 0.4462, | |
| "step": 1300 | |
| }, | |
| { | |
| "epoch": 1.1177851274938653, | |
| "grad_norm": 0.6647533178329468, | |
| "learning_rate": 7.89979387544027e-06, | |
| "loss": 0.4481, | |
| "step": 1310 | |
| }, | |
| { | |
| "epoch": 1.1263202816600875, | |
| "grad_norm": 0.724305272102356, | |
| "learning_rate": 7.859169117932752e-06, | |
| "loss": 0.4426, | |
| "step": 1320 | |
| }, | |
| { | |
| "epoch": 1.1348554358263097, | |
| "grad_norm": 0.6487372517585754, | |
| "learning_rate": 7.818261945904226e-06, | |
| "loss": 0.4442, | |
| "step": 1330 | |
| }, | |
| { | |
| "epoch": 1.1433905899925318, | |
| "grad_norm": 0.7011922597885132, | |
| "learning_rate": 7.777076399961912e-06, | |
| "loss": 0.4486, | |
| "step": 1340 | |
| }, | |
| { | |
| "epoch": 1.1519257441587538, | |
| "grad_norm": 0.6805099844932556, | |
| "learning_rate": 7.735616548209425e-06, | |
| "loss": 0.4469, | |
| "step": 1350 | |
| }, | |
| { | |
| "epoch": 1.160460898324976, | |
| "grad_norm": 0.683447003364563, | |
| "learning_rate": 7.693886485844955e-06, | |
| "loss": 0.4553, | |
| "step": 1360 | |
| }, | |
| { | |
| "epoch": 1.1689960524911982, | |
| "grad_norm": 0.8037645816802979, | |
| "learning_rate": 7.651890334756736e-06, | |
| "loss": 0.4387, | |
| "step": 1370 | |
| }, | |
| { | |
| "epoch": 1.1775312066574202, | |
| "grad_norm": 0.668209433555603, | |
| "learning_rate": 7.609632243115941e-06, | |
| "loss": 0.4373, | |
| "step": 1380 | |
| }, | |
| { | |
| "epoch": 1.1860663608236424, | |
| "grad_norm": 0.7479778528213501, | |
| "learning_rate": 7.567116384966918e-06, | |
| "loss": 0.4525, | |
| "step": 1390 | |
| }, | |
| { | |
| "epoch": 1.1946015149898646, | |
| "grad_norm": 0.743675172328949, | |
| "learning_rate": 7.524346959814918e-06, | |
| "loss": 0.4458, | |
| "step": 1400 | |
| }, | |
| { | |
| "epoch": 1.2031366691560867, | |
| "grad_norm": 0.6677897572517395, | |
| "learning_rate": 7.481328192211278e-06, | |
| "loss": 0.4521, | |
| "step": 1410 | |
| }, | |
| { | |
| "epoch": 1.2116718233223087, | |
| "grad_norm": 0.655086100101471, | |
| "learning_rate": 7.4380643313361385e-06, | |
| "loss": 0.4439, | |
| "step": 1420 | |
| }, | |
| { | |
| "epoch": 1.220206977488531, | |
| "grad_norm": 0.7043260335922241, | |
| "learning_rate": 7.394559650578746e-06, | |
| "loss": 0.4387, | |
| "step": 1430 | |
| }, | |
| { | |
| "epoch": 1.228742131654753, | |
| "grad_norm": 0.6985997557640076, | |
| "learning_rate": 7.3508184471153274e-06, | |
| "loss": 0.4396, | |
| "step": 1440 | |
| }, | |
| { | |
| "epoch": 1.237277285820975, | |
| "grad_norm": 0.6879217028617859, | |
| "learning_rate": 7.306845041484656e-06, | |
| "loss": 0.4438, | |
| "step": 1450 | |
| }, | |
| { | |
| "epoch": 1.2458124399871973, | |
| "grad_norm": 0.6632692217826843, | |
| "learning_rate": 7.262643777161279e-06, | |
| "loss": 0.4421, | |
| "step": 1460 | |
| }, | |
| { | |
| "epoch": 1.2543475941534195, | |
| "grad_norm": 0.6956544518470764, | |
| "learning_rate": 7.218219020126491e-06, | |
| "loss": 0.45, | |
| "step": 1470 | |
| }, | |
| { | |
| "epoch": 1.2628827483196416, | |
| "grad_norm": 0.6400607824325562, | |
| "learning_rate": 7.1735751584370894e-06, | |
| "loss": 0.4372, | |
| "step": 1480 | |
| }, | |
| { | |
| "epoch": 1.2714179024858636, | |
| "grad_norm": 0.735914945602417, | |
| "learning_rate": 7.128716601791938e-06, | |
| "loss": 0.4459, | |
| "step": 1490 | |
| }, | |
| { | |
| "epoch": 1.2799530566520858, | |
| "grad_norm": 0.7256124019622803, | |
| "learning_rate": 7.083647781096404e-06, | |
| "loss": 0.4437, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 1.2884882108183078, | |
| "grad_norm": 0.6894016861915588, | |
| "learning_rate": 7.038373148024693e-06, | |
| "loss": 0.4468, | |
| "step": 1510 | |
| }, | |
| { | |
| "epoch": 1.29702336498453, | |
| "grad_norm": 0.676726222038269, | |
| "learning_rate": 6.992897174580133e-06, | |
| "loss": 0.4409, | |
| "step": 1520 | |
| }, | |
| { | |
| "epoch": 1.3055585191507522, | |
| "grad_norm": 0.7112148404121399, | |
| "learning_rate": 6.947224352653454e-06, | |
| "loss": 0.4387, | |
| "step": 1530 | |
| }, | |
| { | |
| "epoch": 1.3140936733169744, | |
| "grad_norm": 0.6504147052764893, | |
| "learning_rate": 6.901359193579104e-06, | |
| "loss": 0.4357, | |
| "step": 1540 | |
| }, | |
| { | |
| "epoch": 1.3226288274831965, | |
| "grad_norm": 0.6736079454421997, | |
| "learning_rate": 6.855306227689634e-06, | |
| "loss": 0.4403, | |
| "step": 1550 | |
| }, | |
| { | |
| "epoch": 1.3311639816494185, | |
| "grad_norm": 0.6631302237510681, | |
| "learning_rate": 6.809070003868221e-06, | |
| "loss": 0.4368, | |
| "step": 1560 | |
| }, | |
| { | |
| "epoch": 1.3396991358156407, | |
| "grad_norm": 0.6592603921890259, | |
| "learning_rate": 6.762655089099353e-06, | |
| "loss": 0.446, | |
| "step": 1570 | |
| }, | |
| { | |
| "epoch": 1.3482342899818627, | |
| "grad_norm": 0.6875331997871399, | |
| "learning_rate": 6.716066068017722e-06, | |
| "loss": 0.4459, | |
| "step": 1580 | |
| }, | |
| { | |
| "epoch": 1.3567694441480849, | |
| "grad_norm": 0.8026809096336365, | |
| "learning_rate": 6.669307542455369e-06, | |
| "loss": 0.4363, | |
| "step": 1590 | |
| }, | |
| { | |
| "epoch": 1.365304598314307, | |
| "grad_norm": 0.8665246367454529, | |
| "learning_rate": 6.6223841309871535e-06, | |
| "loss": 0.4415, | |
| "step": 1600 | |
| }, | |
| { | |
| "epoch": 1.3738397524805293, | |
| "grad_norm": 0.7898995280265808, | |
| "learning_rate": 6.575300468474546e-06, | |
| "loss": 0.4411, | |
| "step": 1610 | |
| }, | |
| { | |
| "epoch": 1.3823749066467514, | |
| "grad_norm": 0.6701722145080566, | |
| "learning_rate": 6.52806120560781e-06, | |
| "loss": 0.441, | |
| "step": 1620 | |
| }, | |
| { | |
| "epoch": 1.3909100608129734, | |
| "grad_norm": 0.6245713233947754, | |
| "learning_rate": 6.480671008446651e-06, | |
| "loss": 0.4357, | |
| "step": 1630 | |
| }, | |
| { | |
| "epoch": 1.3994452149791956, | |
| "grad_norm": 0.6475566029548645, | |
| "learning_rate": 6.433134557959303e-06, | |
| "loss": 0.4487, | |
| "step": 1640 | |
| }, | |
| { | |
| "epoch": 1.4079803691454176, | |
| "grad_norm": 0.7050219178199768, | |
| "learning_rate": 6.385456549560187e-06, | |
| "loss": 0.4488, | |
| "step": 1650 | |
| }, | |
| { | |
| "epoch": 1.4165155233116398, | |
| "grad_norm": 0.6392747759819031, | |
| "learning_rate": 6.337641692646106e-06, | |
| "loss": 0.4453, | |
| "step": 1660 | |
| }, | |
| { | |
| "epoch": 1.425050677477862, | |
| "grad_norm": 0.6791578531265259, | |
| "learning_rate": 6.2896947101310756e-06, | |
| "loss": 0.4406, | |
| "step": 1670 | |
| }, | |
| { | |
| "epoch": 1.4335858316440842, | |
| "grad_norm": 0.6261289119720459, | |
| "learning_rate": 6.24162033797983e-06, | |
| "loss": 0.4364, | |
| "step": 1680 | |
| }, | |
| { | |
| "epoch": 1.4421209858103061, | |
| "grad_norm": 0.6886487603187561, | |
| "learning_rate": 6.193423324740018e-06, | |
| "loss": 0.44, | |
| "step": 1690 | |
| }, | |
| { | |
| "epoch": 1.4506561399765283, | |
| "grad_norm": 0.7670247554779053, | |
| "learning_rate": 6.1451084310731545e-06, | |
| "loss": 0.4348, | |
| "step": 1700 | |
| }, | |
| { | |
| "epoch": 1.4591912941427505, | |
| "grad_norm": 0.6851808428764343, | |
| "learning_rate": 6.096680429284412e-06, | |
| "loss": 0.4358, | |
| "step": 1710 | |
| }, | |
| { | |
| "epoch": 1.4677264483089725, | |
| "grad_norm": 0.660381555557251, | |
| "learning_rate": 6.0481441028512154e-06, | |
| "loss": 0.4419, | |
| "step": 1720 | |
| }, | |
| { | |
| "epoch": 1.4762616024751947, | |
| "grad_norm": 0.6627777218818665, | |
| "learning_rate": 5.9995042459507555e-06, | |
| "loss": 0.4443, | |
| "step": 1730 | |
| }, | |
| { | |
| "epoch": 1.4847967566414169, | |
| "grad_norm": 0.7264851331710815, | |
| "learning_rate": 5.9507656629864554e-06, | |
| "loss": 0.4419, | |
| "step": 1740 | |
| }, | |
| { | |
| "epoch": 1.493331910807639, | |
| "grad_norm": 0.6353123784065247, | |
| "learning_rate": 5.901933168113405e-06, | |
| "loss": 0.4325, | |
| "step": 1750 | |
| }, | |
| { | |
| "epoch": 1.5018670649738612, | |
| "grad_norm": 0.7154319882392883, | |
| "learning_rate": 5.853011584762846e-06, | |
| "loss": 0.4413, | |
| "step": 1760 | |
| }, | |
| { | |
| "epoch": 1.5104022191400832, | |
| "grad_norm": 0.678537905216217, | |
| "learning_rate": 5.804005745165735e-06, | |
| "loss": 0.4418, | |
| "step": 1770 | |
| }, | |
| { | |
| "epoch": 1.5189373733063052, | |
| "grad_norm": 0.6648243069648743, | |
| "learning_rate": 5.754920489875447e-06, | |
| "loss": 0.438, | |
| "step": 1780 | |
| }, | |
| { | |
| "epoch": 1.5274725274725274, | |
| "grad_norm": 0.7042447924613953, | |
| "learning_rate": 5.705760667289644e-06, | |
| "loss": 0.441, | |
| "step": 1790 | |
| }, | |
| { | |
| "epoch": 1.5360076816387496, | |
| "grad_norm": 0.6617943048477173, | |
| "learning_rate": 5.65653113317137e-06, | |
| "loss": 0.4375, | |
| "step": 1800 | |
| }, | |
| { | |
| "epoch": 1.5445428358049718, | |
| "grad_norm": 0.6619049310684204, | |
| "learning_rate": 5.607236750169431e-06, | |
| "loss": 0.4395, | |
| "step": 1810 | |
| }, | |
| { | |
| "epoch": 1.553077989971194, | |
| "grad_norm": 0.6774610280990601, | |
| "learning_rate": 5.557882387338086e-06, | |
| "loss": 0.4417, | |
| "step": 1820 | |
| }, | |
| { | |
| "epoch": 1.5616131441374161, | |
| "grad_norm": 0.6360573768615723, | |
| "learning_rate": 5.508472919656101e-06, | |
| "loss": 0.4346, | |
| "step": 1830 | |
| }, | |
| { | |
| "epoch": 1.5701482983036381, | |
| "grad_norm": 0.6723465919494629, | |
| "learning_rate": 5.459013227545224e-06, | |
| "loss": 0.4387, | |
| "step": 1840 | |
| }, | |
| { | |
| "epoch": 1.57868345246986, | |
| "grad_norm": 0.6785566806793213, | |
| "learning_rate": 5.4095081963881225e-06, | |
| "loss": 0.4364, | |
| "step": 1850 | |
| }, | |
| { | |
| "epoch": 1.5872186066360823, | |
| "grad_norm": 0.6658756136894226, | |
| "learning_rate": 5.359962716045836e-06, | |
| "loss": 0.4302, | |
| "step": 1860 | |
| }, | |
| { | |
| "epoch": 1.5957537608023045, | |
| "grad_norm": 0.6586979031562805, | |
| "learning_rate": 5.310381680374767e-06, | |
| "loss": 0.4386, | |
| "step": 1870 | |
| }, | |
| { | |
| "epoch": 1.6042889149685267, | |
| "grad_norm": 0.7203266620635986, | |
| "learning_rate": 5.260769986743296e-06, | |
| "loss": 0.4377, | |
| "step": 1880 | |
| }, | |
| { | |
| "epoch": 1.6128240691347489, | |
| "grad_norm": 0.7455547451972961, | |
| "learning_rate": 5.2111325355480506e-06, | |
| "loss": 0.4428, | |
| "step": 1890 | |
| }, | |
| { | |
| "epoch": 1.6213592233009708, | |
| "grad_norm": 0.6452232599258423, | |
| "learning_rate": 5.161474229729858e-06, | |
| "loss": 0.4336, | |
| "step": 1900 | |
| }, | |
| { | |
| "epoch": 1.629894377467193, | |
| "grad_norm": 0.6487184166908264, | |
| "learning_rate": 5.111799974289462e-06, | |
| "loss": 0.4426, | |
| "step": 1910 | |
| }, | |
| { | |
| "epoch": 1.638429531633415, | |
| "grad_norm": 0.6176945567131042, | |
| "learning_rate": 5.062114675803032e-06, | |
| "loss": 0.4356, | |
| "step": 1920 | |
| }, | |
| { | |
| "epoch": 1.6469646857996372, | |
| "grad_norm": 0.6743866205215454, | |
| "learning_rate": 5.012423241937515e-06, | |
| "loss": 0.4308, | |
| "step": 1930 | |
| }, | |
| { | |
| "epoch": 1.6554998399658594, | |
| "grad_norm": 0.6723053455352783, | |
| "learning_rate": 4.962730580965879e-06, | |
| "loss": 0.4344, | |
| "step": 1940 | |
| }, | |
| { | |
| "epoch": 1.6640349941320816, | |
| "grad_norm": 0.7293416261672974, | |
| "learning_rate": 4.913041601282298e-06, | |
| "loss": 0.4229, | |
| "step": 1950 | |
| }, | |
| { | |
| "epoch": 1.6725701482983037, | |
| "grad_norm": 0.6705866456031799, | |
| "learning_rate": 4.863361210917332e-06, | |
| "loss": 0.4412, | |
| "step": 1960 | |
| }, | |
| { | |
| "epoch": 1.6811053024645257, | |
| "grad_norm": 0.676863968372345, | |
| "learning_rate": 4.813694317053122e-06, | |
| "loss": 0.4423, | |
| "step": 1970 | |
| }, | |
| { | |
| "epoch": 1.689640456630748, | |
| "grad_norm": 0.659705400466919, | |
| "learning_rate": 4.764045825538702e-06, | |
| "loss": 0.4372, | |
| "step": 1980 | |
| }, | |
| { | |
| "epoch": 1.6981756107969699, | |
| "grad_norm": 0.6498283743858337, | |
| "learning_rate": 4.714420640405406e-06, | |
| "loss": 0.4344, | |
| "step": 1990 | |
| }, | |
| { | |
| "epoch": 1.706710764963192, | |
| "grad_norm": 0.6507246494293213, | |
| "learning_rate": 4.664823663382482e-06, | |
| "loss": 0.4274, | |
| "step": 2000 | |
| }, | |
| { | |
| "epoch": 1.7152459191294143, | |
| "grad_norm": 0.705443263053894, | |
| "learning_rate": 4.61525979341292e-06, | |
| "loss": 0.4365, | |
| "step": 2010 | |
| }, | |
| { | |
| "epoch": 1.7237810732956365, | |
| "grad_norm": 0.6291791200637817, | |
| "learning_rate": 4.565733926169561e-06, | |
| "loss": 0.4354, | |
| "step": 2020 | |
| }, | |
| { | |
| "epoch": 1.7323162274618586, | |
| "grad_norm": 0.6487744450569153, | |
| "learning_rate": 4.516250953571527e-06, | |
| "loss": 0.439, | |
| "step": 2030 | |
| }, | |
| { | |
| "epoch": 1.7408513816280806, | |
| "grad_norm": 0.6847457885742188, | |
| "learning_rate": 4.466815763301014e-06, | |
| "loss": 0.4245, | |
| "step": 2040 | |
| }, | |
| { | |
| "epoch": 1.7493865357943028, | |
| "grad_norm": 0.6660692095756531, | |
| "learning_rate": 4.417433238320524e-06, | |
| "loss": 0.4345, | |
| "step": 2050 | |
| }, | |
| { | |
| "epoch": 1.7579216899605248, | |
| "grad_norm": 0.6606684327125549, | |
| "learning_rate": 4.36810825639054e-06, | |
| "loss": 0.4442, | |
| "step": 2060 | |
| }, | |
| { | |
| "epoch": 1.766456844126747, | |
| "grad_norm": 0.7084365487098694, | |
| "learning_rate": 4.318845689587727e-06, | |
| "loss": 0.4348, | |
| "step": 2070 | |
| }, | |
| { | |
| "epoch": 1.7749919982929692, | |
| "grad_norm": 0.6337569952011108, | |
| "learning_rate": 4.269650403823698e-06, | |
| "loss": 0.4384, | |
| "step": 2080 | |
| }, | |
| { | |
| "epoch": 1.7835271524591914, | |
| "grad_norm": 0.6552258133888245, | |
| "learning_rate": 4.220527258364373e-06, | |
| "loss": 0.4376, | |
| "step": 2090 | |
| }, | |
| { | |
| "epoch": 1.7920623066254135, | |
| "grad_norm": 0.6303353309631348, | |
| "learning_rate": 4.171481105350021e-06, | |
| "loss": 0.4373, | |
| "step": 2100 | |
| }, | |
| { | |
| "epoch": 1.8005974607916355, | |
| "grad_norm": 0.650391697883606, | |
| "learning_rate": 4.122516789315969e-06, | |
| "loss": 0.4437, | |
| "step": 2110 | |
| }, | |
| { | |
| "epoch": 1.8091326149578577, | |
| "grad_norm": 0.6542319059371948, | |
| "learning_rate": 4.073639146714099e-06, | |
| "loss": 0.4262, | |
| "step": 2120 | |
| }, | |
| { | |
| "epoch": 1.8176677691240797, | |
| "grad_norm": 0.651104211807251, | |
| "learning_rate": 4.024853005435121e-06, | |
| "loss": 0.4363, | |
| "step": 2130 | |
| }, | |
| { | |
| "epoch": 1.8262029232903019, | |
| "grad_norm": 0.6116424798965454, | |
| "learning_rate": 3.976163184331702e-06, | |
| "loss": 0.4291, | |
| "step": 2140 | |
| }, | |
| { | |
| "epoch": 1.834738077456524, | |
| "grad_norm": 0.8228034377098083, | |
| "learning_rate": 3.927574492742473e-06, | |
| "loss": 0.4306, | |
| "step": 2150 | |
| }, | |
| { | |
| "epoch": 1.8432732316227463, | |
| "grad_norm": 0.6177427768707275, | |
| "learning_rate": 3.879091730017003e-06, | |
| "loss": 0.4274, | |
| "step": 2160 | |
| }, | |
| { | |
| "epoch": 1.8518083857889684, | |
| "grad_norm": 0.6379538178443909, | |
| "learning_rate": 3.830719685041728e-06, | |
| "loss": 0.4344, | |
| "step": 2170 | |
| }, | |
| { | |
| "epoch": 1.8603435399551904, | |
| "grad_norm": 0.6659247279167175, | |
| "learning_rate": 3.782463135766938e-06, | |
| "loss": 0.4318, | |
| "step": 2180 | |
| }, | |
| { | |
| "epoch": 1.8688786941214126, | |
| "grad_norm": 0.6846089959144592, | |
| "learning_rate": 3.7343268487348273e-06, | |
| "loss": 0.4354, | |
| "step": 2190 | |
| }, | |
| { | |
| "epoch": 1.8774138482876346, | |
| "grad_norm": 0.6371617913246155, | |
| "learning_rate": 3.6863155786086886e-06, | |
| "loss": 0.4302, | |
| "step": 2200 | |
| }, | |
| { | |
| "epoch": 1.8859490024538568, | |
| "grad_norm": 0.6230723857879639, | |
| "learning_rate": 3.6384340677032616e-06, | |
| "loss": 0.4367, | |
| "step": 2210 | |
| }, | |
| { | |
| "epoch": 1.894484156620079, | |
| "grad_norm": 0.6282981634140015, | |
| "learning_rate": 3.5906870455163175e-06, | |
| "loss": 0.4284, | |
| "step": 2220 | |
| }, | |
| { | |
| "epoch": 1.9030193107863012, | |
| "grad_norm": 0.6560613512992859, | |
| "learning_rate": 3.5430792282614978e-06, | |
| "loss": 0.4324, | |
| "step": 2230 | |
| }, | |
| { | |
| "epoch": 1.9115544649525233, | |
| "grad_norm": 0.6765207052230835, | |
| "learning_rate": 3.4956153184024765e-06, | |
| "loss": 0.4261, | |
| "step": 2240 | |
| }, | |
| { | |
| "epoch": 1.9200896191187453, | |
| "grad_norm": 0.6762336492538452, | |
| "learning_rate": 3.448300004188463e-06, | |
| "loss": 0.4304, | |
| "step": 2250 | |
| }, | |
| { | |
| "epoch": 1.9286247732849675, | |
| "grad_norm": 0.6394000053405762, | |
| "learning_rate": 3.401137959191133e-06, | |
| "loss": 0.4294, | |
| "step": 2260 | |
| }, | |
| { | |
| "epoch": 1.9371599274511895, | |
| "grad_norm": 0.6332142949104309, | |
| "learning_rate": 3.354133841842989e-06, | |
| "loss": 0.4308, | |
| "step": 2270 | |
| }, | |
| { | |
| "epoch": 1.9456950816174117, | |
| "grad_norm": 0.6541900038719177, | |
| "learning_rate": 3.3072922949772235e-06, | |
| "loss": 0.4343, | |
| "step": 2280 | |
| }, | |
| { | |
| "epoch": 1.9542302357836339, | |
| "grad_norm": 0.652773380279541, | |
| "learning_rate": 3.2606179453691255e-06, | |
| "loss": 0.4311, | |
| "step": 2290 | |
| }, | |
| { | |
| "epoch": 1.962765389949856, | |
| "grad_norm": 0.6701607704162598, | |
| "learning_rate": 3.21411540327907e-06, | |
| "loss": 0.4244, | |
| "step": 2300 | |
| }, | |
| { | |
| "epoch": 1.9713005441160782, | |
| "grad_norm": 0.6590484976768494, | |
| "learning_rate": 3.16778926199714e-06, | |
| "loss": 0.4366, | |
| "step": 2310 | |
| }, | |
| { | |
| "epoch": 1.9798356982823002, | |
| "grad_norm": 0.6363989114761353, | |
| "learning_rate": 3.1216440973894184e-06, | |
| "loss": 0.4267, | |
| "step": 2320 | |
| }, | |
| { | |
| "epoch": 1.9883708524485222, | |
| "grad_norm": 0.6702283024787903, | |
| "learning_rate": 3.075684467446014e-06, | |
| "loss": 0.4305, | |
| "step": 2330 | |
| }, | |
| { | |
| "epoch": 1.9969060066147444, | |
| "grad_norm": 0.6762630939483643, | |
| "learning_rate": 3.029914911830846e-06, | |
| "loss": 0.4339, | |
| "step": 2340 | |
| }, | |
| { | |
| "epoch": 2.005121092499733, | |
| "grad_norm": 0.6649417877197266, | |
| "learning_rate": 2.9843399514332317e-06, | |
| "loss": 0.4193, | |
| "step": 2350 | |
| }, | |
| { | |
| "epoch": 2.0136562466659553, | |
| "grad_norm": 0.6555825471878052, | |
| "learning_rate": 2.938964087921341e-06, | |
| "loss": 0.4057, | |
| "step": 2360 | |
| }, | |
| { | |
| "epoch": 2.0221914008321775, | |
| "grad_norm": 0.6427297592163086, | |
| "learning_rate": 2.8937918032975434e-06, | |
| "loss": 0.4045, | |
| "step": 2370 | |
| }, | |
| { | |
| "epoch": 2.0307265549983997, | |
| "grad_norm": 0.6486126780509949, | |
| "learning_rate": 2.848827559455696e-06, | |
| "loss": 0.4111, | |
| "step": 2380 | |
| }, | |
| { | |
| "epoch": 2.039261709164622, | |
| "grad_norm": 0.7112558484077454, | |
| "learning_rate": 2.804075797740422e-06, | |
| "loss": 0.413, | |
| "step": 2390 | |
| }, | |
| { | |
| "epoch": 2.047796863330844, | |
| "grad_norm": 0.6829844117164612, | |
| "learning_rate": 2.759540938508415e-06, | |
| "loss": 0.4103, | |
| "step": 2400 | |
| }, | |
| { | |
| "epoch": 2.056332017497066, | |
| "grad_norm": 0.6613876819610596, | |
| "learning_rate": 2.7152273806918185e-06, | |
| "loss": 0.4158, | |
| "step": 2410 | |
| }, | |
| { | |
| "epoch": 2.064867171663288, | |
| "grad_norm": 0.6096106171607971, | |
| "learning_rate": 2.6711395013637208e-06, | |
| "loss": 0.4058, | |
| "step": 2420 | |
| }, | |
| { | |
| "epoch": 2.0734023258295102, | |
| "grad_norm": 0.6055023074150085, | |
| "learning_rate": 2.6272816553058156e-06, | |
| "loss": 0.3974, | |
| "step": 2430 | |
| }, | |
| { | |
| "epoch": 2.0819374799957324, | |
| "grad_norm": 0.6151232719421387, | |
| "learning_rate": 2.5836581745782474e-06, | |
| "loss": 0.4029, | |
| "step": 2440 | |
| }, | |
| { | |
| "epoch": 2.0904726341619546, | |
| "grad_norm": 0.6236457228660583, | |
| "learning_rate": 2.5402733680917163e-06, | |
| "loss": 0.409, | |
| "step": 2450 | |
| }, | |
| { | |
| "epoch": 2.099007788328177, | |
| "grad_norm": 0.6416062712669373, | |
| "learning_rate": 2.497131521181877e-06, | |
| "loss": 0.405, | |
| "step": 2460 | |
| }, | |
| { | |
| "epoch": 2.107542942494399, | |
| "grad_norm": 0.6345195174217224, | |
| "learning_rate": 2.4542368951860303e-06, | |
| "loss": 0.3951, | |
| "step": 2470 | |
| }, | |
| { | |
| "epoch": 2.1160780966606207, | |
| "grad_norm": 0.7058767676353455, | |
| "learning_rate": 2.4115937270222305e-06, | |
| "loss": 0.4047, | |
| "step": 2480 | |
| }, | |
| { | |
| "epoch": 2.124613250826843, | |
| "grad_norm": 0.6685160398483276, | |
| "learning_rate": 2.3692062287707795e-06, | |
| "loss": 0.4037, | |
| "step": 2490 | |
| }, | |
| { | |
| "epoch": 2.133148404993065, | |
| "grad_norm": 0.6508364677429199, | |
| "learning_rate": 2.3270785872581698e-06, | |
| "loss": 0.4044, | |
| "step": 2500 | |
| }, | |
| { | |
| "epoch": 2.1416835591592873, | |
| "grad_norm": 0.6913980841636658, | |
| "learning_rate": 2.2852149636435366e-06, | |
| "loss": 0.4071, | |
| "step": 2510 | |
| }, | |
| { | |
| "epoch": 2.1502187133255095, | |
| "grad_norm": 0.6269243955612183, | |
| "learning_rate": 2.2436194930076448e-06, | |
| "loss": 0.4131, | |
| "step": 2520 | |
| }, | |
| { | |
| "epoch": 2.1587538674917317, | |
| "grad_norm": 0.6732710599899292, | |
| "learning_rate": 2.2022962839444356e-06, | |
| "loss": 0.4043, | |
| "step": 2530 | |
| }, | |
| { | |
| "epoch": 2.167289021657954, | |
| "grad_norm": 0.658445417881012, | |
| "learning_rate": 2.161249418155199e-06, | |
| "loss": 0.4036, | |
| "step": 2540 | |
| }, | |
| { | |
| "epoch": 2.1758241758241756, | |
| "grad_norm": 0.6448598504066467, | |
| "learning_rate": 2.120482950045422e-06, | |
| "loss": 0.4139, | |
| "step": 2550 | |
| }, | |
| { | |
| "epoch": 2.184359329990398, | |
| "grad_norm": 0.6652196049690247, | |
| "learning_rate": 2.080000906324292e-06, | |
| "loss": 0.399, | |
| "step": 2560 | |
| }, | |
| { | |
| "epoch": 2.19289448415662, | |
| "grad_norm": 0.6046453714370728, | |
| "learning_rate": 2.0398072856069707e-06, | |
| "loss": 0.4063, | |
| "step": 2570 | |
| }, | |
| { | |
| "epoch": 2.201429638322842, | |
| "grad_norm": 0.6543176174163818, | |
| "learning_rate": 1.9999060580196324e-06, | |
| "loss": 0.4048, | |
| "step": 2580 | |
| }, | |
| { | |
| "epoch": 2.2099647924890644, | |
| "grad_norm": 0.6523047089576721, | |
| "learning_rate": 1.9603011648073055e-06, | |
| "loss": 0.4073, | |
| "step": 2590 | |
| }, | |
| { | |
| "epoch": 2.2184999466552866, | |
| "grad_norm": 0.6173551082611084, | |
| "learning_rate": 1.9209965179445826e-06, | |
| "loss": 0.4106, | |
| "step": 2600 | |
| }, | |
| { | |
| "epoch": 2.227035100821509, | |
| "grad_norm": 0.6489232778549194, | |
| "learning_rate": 1.881995999749215e-06, | |
| "loss": 0.4103, | |
| "step": 2610 | |
| }, | |
| { | |
| "epoch": 2.2355702549877305, | |
| "grad_norm": 0.6323755383491516, | |
| "learning_rate": 1.8433034624986335e-06, | |
| "loss": 0.4012, | |
| "step": 2620 | |
| }, | |
| { | |
| "epoch": 2.2441054091539527, | |
| "grad_norm": 0.640560507774353, | |
| "learning_rate": 1.8049227280494391e-06, | |
| "loss": 0.4048, | |
| "step": 2630 | |
| }, | |
| { | |
| "epoch": 2.252640563320175, | |
| "grad_norm": 0.6368593573570251, | |
| "learning_rate": 1.7668575874599086e-06, | |
| "loss": 0.4079, | |
| "step": 2640 | |
| }, | |
| { | |
| "epoch": 2.261175717486397, | |
| "grad_norm": 0.6442722678184509, | |
| "learning_rate": 1.7291118006155178e-06, | |
| "loss": 0.4144, | |
| "step": 2650 | |
| }, | |
| { | |
| "epoch": 2.2697108716526193, | |
| "grad_norm": 0.7017993330955505, | |
| "learning_rate": 1.6916890958575687e-06, | |
| "loss": 0.4107, | |
| "step": 2660 | |
| }, | |
| { | |
| "epoch": 2.2782460258188415, | |
| "grad_norm": 0.6456113457679749, | |
| "learning_rate": 1.6545931696149208e-06, | |
| "loss": 0.4081, | |
| "step": 2670 | |
| }, | |
| { | |
| "epoch": 2.2867811799850637, | |
| "grad_norm": 0.6799325346946716, | |
| "learning_rate": 1.6178276860388736e-06, | |
| "loss": 0.4022, | |
| "step": 2680 | |
| }, | |
| { | |
| "epoch": 2.2953163341512854, | |
| "grad_norm": 0.6276426911354065, | |
| "learning_rate": 1.5813962766412427e-06, | |
| "loss": 0.4127, | |
| "step": 2690 | |
| }, | |
| { | |
| "epoch": 2.3038514883175076, | |
| "grad_norm": 0.6316121816635132, | |
| "learning_rate": 1.5453025399356542e-06, | |
| "loss": 0.4064, | |
| "step": 2700 | |
| }, | |
| { | |
| "epoch": 2.31238664248373, | |
| "grad_norm": 0.6637789011001587, | |
| "learning_rate": 1.5095500410821085e-06, | |
| "loss": 0.4089, | |
| "step": 2710 | |
| }, | |
| { | |
| "epoch": 2.320921796649952, | |
| "grad_norm": 0.6297438740730286, | |
| "learning_rate": 1.4741423115348203e-06, | |
| "loss": 0.4012, | |
| "step": 2720 | |
| }, | |
| { | |
| "epoch": 2.329456950816174, | |
| "grad_norm": 0.6420316100120544, | |
| "learning_rate": 1.439082848693406e-06, | |
| "loss": 0.403, | |
| "step": 2730 | |
| }, | |
| { | |
| "epoch": 2.3379921049823964, | |
| "grad_norm": 0.6523280143737793, | |
| "learning_rate": 1.4043751155574331e-06, | |
| "loss": 0.4037, | |
| "step": 2740 | |
| }, | |
| { | |
| "epoch": 2.346527259148618, | |
| "grad_norm": 0.6407163739204407, | |
| "learning_rate": 1.370022540384347e-06, | |
| "loss": 0.4044, | |
| "step": 2750 | |
| }, | |
| { | |
| "epoch": 2.3550624133148403, | |
| "grad_norm": 0.6137824654579163, | |
| "learning_rate": 1.3360285163508585e-06, | |
| "loss": 0.4065, | |
| "step": 2760 | |
| }, | |
| { | |
| "epoch": 2.3635975674810625, | |
| "grad_norm": 0.6386251449584961, | |
| "learning_rate": 1.3023964012177808e-06, | |
| "loss": 0.3996, | |
| "step": 2770 | |
| }, | |
| { | |
| "epoch": 2.3721327216472847, | |
| "grad_norm": 0.6598160266876221, | |
| "learning_rate": 1.2691295169983602e-06, | |
| "loss": 0.407, | |
| "step": 2780 | |
| }, | |
| { | |
| "epoch": 2.380667875813507, | |
| "grad_norm": 0.6845166683197021, | |
| "learning_rate": 1.2362311496301465e-06, | |
| "loss": 0.4093, | |
| "step": 2790 | |
| }, | |
| { | |
| "epoch": 2.389203029979729, | |
| "grad_norm": 0.6791868209838867, | |
| "learning_rate": 1.2037045486504317e-06, | |
| "loss": 0.401, | |
| "step": 2800 | |
| }, | |
| { | |
| "epoch": 2.3977381841459513, | |
| "grad_norm": 0.6840428113937378, | |
| "learning_rate": 1.171552926875265e-06, | |
| "loss": 0.4052, | |
| "step": 2810 | |
| }, | |
| { | |
| "epoch": 2.4062733383121735, | |
| "grad_norm": 0.6226551532745361, | |
| "learning_rate": 1.1397794600821155e-06, | |
| "loss": 0.4041, | |
| "step": 2820 | |
| }, | |
| { | |
| "epoch": 2.4148084924783952, | |
| "grad_norm": 0.6312896013259888, | |
| "learning_rate": 1.1083872866961816e-06, | |
| "loss": 0.4033, | |
| "step": 2830 | |
| }, | |
| { | |
| "epoch": 2.4233436466446174, | |
| "grad_norm": 0.6522102952003479, | |
| "learning_rate": 1.0773795074803917e-06, | |
| "loss": 0.3987, | |
| "step": 2840 | |
| }, | |
| { | |
| "epoch": 2.4318788008108396, | |
| "grad_norm": 0.6765321493148804, | |
| "learning_rate": 1.0467591852291287e-06, | |
| "loss": 0.4059, | |
| "step": 2850 | |
| }, | |
| { | |
| "epoch": 2.440413954977062, | |
| "grad_norm": 0.6331756114959717, | |
| "learning_rate": 1.0165293444657037e-06, | |
| "loss": 0.401, | |
| "step": 2860 | |
| }, | |
| { | |
| "epoch": 2.448949109143284, | |
| "grad_norm": 0.6368128657341003, | |
| "learning_rate": 9.866929711436035e-07, | |
| "loss": 0.4136, | |
| "step": 2870 | |
| }, | |
| { | |
| "epoch": 2.457484263309506, | |
| "grad_norm": 0.6715298891067505, | |
| "learning_rate": 9.57253012351556e-07, | |
| "loss": 0.4157, | |
| "step": 2880 | |
| }, | |
| { | |
| "epoch": 2.466019417475728, | |
| "grad_norm": 0.6294402480125427, | |
| "learning_rate": 9.282123760224327e-07, | |
| "loss": 0.4058, | |
| "step": 2890 | |
| }, | |
| { | |
| "epoch": 2.47455457164195, | |
| "grad_norm": 0.6221645474433899, | |
| "learning_rate": 8.99573930646015e-07, | |
| "loss": 0.4014, | |
| "step": 2900 | |
| }, | |
| { | |
| "epoch": 2.4830897258081723, | |
| "grad_norm": 0.642346978187561, | |
| "learning_rate": 8.713405049856588e-07, | |
| "loss": 0.41, | |
| "step": 2910 | |
| }, | |
| { | |
| "epoch": 2.4916248799743945, | |
| "grad_norm": 0.628089189529419, | |
| "learning_rate": 8.435148877988886e-07, | |
| "loss": 0.4084, | |
| "step": 2920 | |
| }, | |
| { | |
| "epoch": 2.5001600341406167, | |
| "grad_norm": 0.637294352054596, | |
| "learning_rate": 8.160998275619309e-07, | |
| "loss": 0.4025, | |
| "step": 2930 | |
| }, | |
| { | |
| "epoch": 2.508695188306839, | |
| "grad_norm": 0.6054239273071289, | |
| "learning_rate": 7.890980321982355e-07, | |
| "loss": 0.405, | |
| "step": 2940 | |
| }, | |
| { | |
| "epoch": 2.517230342473061, | |
| "grad_norm": 0.6559164524078369, | |
| "learning_rate": 7.625121688110027e-07, | |
| "loss": 0.4043, | |
| "step": 2950 | |
| }, | |
| { | |
| "epoch": 2.5257654966392833, | |
| "grad_norm": 0.665451169013977, | |
| "learning_rate": 7.363448634197412e-07, | |
| "loss": 0.4028, | |
| "step": 2960 | |
| }, | |
| { | |
| "epoch": 2.534300650805505, | |
| "grad_norm": 0.6523730158805847, | |
| "learning_rate": 7.105987007008724e-07, | |
| "loss": 0.4068, | |
| "step": 2970 | |
| }, | |
| { | |
| "epoch": 2.5428358049717272, | |
| "grad_norm": 0.6315779089927673, | |
| "learning_rate": 6.852762237324462e-07, | |
| "loss": 0.4092, | |
| "step": 2980 | |
| }, | |
| { | |
| "epoch": 2.5513709591379494, | |
| "grad_norm": 0.6356623768806458, | |
| "learning_rate": 6.603799337429328e-07, | |
| "loss": 0.4099, | |
| "step": 2990 | |
| }, | |
| { | |
| "epoch": 2.5599061133041716, | |
| "grad_norm": 0.6033786535263062, | |
| "learning_rate": 6.359122898641717e-07, | |
| "loss": 0.4084, | |
| "step": 3000 | |
| }, | |
| { | |
| "epoch": 2.568441267470394, | |
| "grad_norm": 0.6288996338844299, | |
| "learning_rate": 6.118757088884686e-07, | |
| "loss": 0.4131, | |
| "step": 3010 | |
| }, | |
| { | |
| "epoch": 2.5769764216366156, | |
| "grad_norm": 0.6407638192176819, | |
| "learning_rate": 5.882725650298787e-07, | |
| "loss": 0.409, | |
| "step": 3020 | |
| }, | |
| { | |
| "epoch": 2.5855115758028377, | |
| "grad_norm": 0.6419194340705872, | |
| "learning_rate": 5.651051896896908e-07, | |
| "loss": 0.4038, | |
| "step": 3030 | |
| }, | |
| { | |
| "epoch": 2.59404672996906, | |
| "grad_norm": 0.6341270208358765, | |
| "learning_rate": 5.42375871226144e-07, | |
| "loss": 0.4064, | |
| "step": 3040 | |
| }, | |
| { | |
| "epoch": 2.602581884135282, | |
| "grad_norm": 0.647345244884491, | |
| "learning_rate": 5.200868547283972e-07, | |
| "loss": 0.4081, | |
| "step": 3050 | |
| }, | |
| { | |
| "epoch": 2.6111170383015043, | |
| "grad_norm": 0.6366192102432251, | |
| "learning_rate": 4.982403417947662e-07, | |
| "loss": 0.4033, | |
| "step": 3060 | |
| }, | |
| { | |
| "epoch": 2.6196521924677265, | |
| "grad_norm": 0.638552725315094, | |
| "learning_rate": 4.768384903152634e-07, | |
| "loss": 0.4096, | |
| "step": 3070 | |
| }, | |
| { | |
| "epoch": 2.6281873466339487, | |
| "grad_norm": 0.6206834316253662, | |
| "learning_rate": 4.558834142584556e-07, | |
| "loss": 0.4065, | |
| "step": 3080 | |
| }, | |
| { | |
| "epoch": 2.636722500800171, | |
| "grad_norm": 0.6011072993278503, | |
| "learning_rate": 4.3537718346264857e-07, | |
| "loss": 0.4029, | |
| "step": 3090 | |
| }, | |
| { | |
| "epoch": 2.645257654966393, | |
| "grad_norm": 0.6577970385551453, | |
| "learning_rate": 4.153218234314443e-07, | |
| "loss": 0.4119, | |
| "step": 3100 | |
| }, | |
| { | |
| "epoch": 2.653792809132615, | |
| "grad_norm": 0.6269943118095398, | |
| "learning_rate": 3.957193151336719e-07, | |
| "loss": 0.4078, | |
| "step": 3110 | |
| }, | |
| { | |
| "epoch": 2.662327963298837, | |
| "grad_norm": 0.6127352118492126, | |
| "learning_rate": 3.765715948077137e-07, | |
| "loss": 0.4014, | |
| "step": 3120 | |
| }, | |
| { | |
| "epoch": 2.6708631174650592, | |
| "grad_norm": 0.6328156590461731, | |
| "learning_rate": 3.578805537702573e-07, | |
| "loss": 0.4057, | |
| "step": 3130 | |
| }, | |
| { | |
| "epoch": 2.6793982716312814, | |
| "grad_norm": 0.6262829899787903, | |
| "learning_rate": 3.396480382294792e-07, | |
| "loss": 0.4026, | |
| "step": 3140 | |
| }, | |
| { | |
| "epoch": 2.6879334257975036, | |
| "grad_norm": 0.631956934928894, | |
| "learning_rate": 3.2187584910268464e-07, | |
| "loss": 0.4003, | |
| "step": 3150 | |
| }, | |
| { | |
| "epoch": 2.6964685799637254, | |
| "grad_norm": 0.628641664981842, | |
| "learning_rate": 3.045657418384218e-07, | |
| "loss": 0.4075, | |
| "step": 3160 | |
| }, | |
| { | |
| "epoch": 2.7050037341299475, | |
| "grad_norm": 0.6435490250587463, | |
| "learning_rate": 2.877194262430916e-07, | |
| "loss": 0.3966, | |
| "step": 3170 | |
| }, | |
| { | |
| "epoch": 2.7135388882961697, | |
| "grad_norm": 0.6484223008155823, | |
| "learning_rate": 2.7133856631205216e-07, | |
| "loss": 0.4076, | |
| "step": 3180 | |
| }, | |
| { | |
| "epoch": 2.722074042462392, | |
| "grad_norm": 0.630936324596405, | |
| "learning_rate": 2.5542478006526703e-07, | |
| "loss": 0.3984, | |
| "step": 3190 | |
| }, | |
| { | |
| "epoch": 2.730609196628614, | |
| "grad_norm": 0.6019722819328308, | |
| "learning_rate": 2.399796393874837e-07, | |
| "loss": 0.4045, | |
| "step": 3200 | |
| }, | |
| { | |
| "epoch": 2.7391443507948363, | |
| "grad_norm": 0.6216541528701782, | |
| "learning_rate": 2.2500466987296466e-07, | |
| "loss": 0.4136, | |
| "step": 3210 | |
| }, | |
| { | |
| "epoch": 2.7476795049610585, | |
| "grad_norm": 0.6513262987136841, | |
| "learning_rate": 2.105013506748027e-07, | |
| "loss": 0.405, | |
| "step": 3220 | |
| }, | |
| { | |
| "epoch": 2.7562146591272807, | |
| "grad_norm": 0.6242448091506958, | |
| "learning_rate": 1.9647111435881494e-07, | |
| "loss": 0.4, | |
| "step": 3230 | |
| }, | |
| { | |
| "epoch": 2.764749813293503, | |
| "grad_norm": 0.6355422139167786, | |
| "learning_rate": 1.829153467620426e-07, | |
| "loss": 0.402, | |
| "step": 3240 | |
| }, | |
| { | |
| "epoch": 2.7732849674597246, | |
| "grad_norm": 0.5879780054092407, | |
| "learning_rate": 1.6983538685586277e-07, | |
| "loss": 0.4015, | |
| "step": 3250 | |
| }, | |
| { | |
| "epoch": 2.781820121625947, | |
| "grad_norm": 0.6436604261398315, | |
| "learning_rate": 1.5723252661373367e-07, | |
| "loss": 0.4025, | |
| "step": 3260 | |
| }, | |
| { | |
| "epoch": 2.790355275792169, | |
| "grad_norm": 0.6310527920722961, | |
| "learning_rate": 1.451080108835784e-07, | |
| "loss": 0.406, | |
| "step": 3270 | |
| }, | |
| { | |
| "epoch": 2.798890429958391, | |
| "grad_norm": 0.6294417381286621, | |
| "learning_rate": 1.3346303726482544e-07, | |
| "loss": 0.3997, | |
| "step": 3280 | |
| }, | |
| { | |
| "epoch": 2.8074255841246134, | |
| "grad_norm": 0.6856746077537537, | |
| "learning_rate": 1.222987559901162e-07, | |
| "loss": 0.4059, | |
| "step": 3290 | |
| }, | |
| { | |
| "epoch": 2.815960738290835, | |
| "grad_norm": 0.8050450682640076, | |
| "learning_rate": 1.1161626981169072e-07, | |
| "loss": 0.4063, | |
| "step": 3300 | |
| }, | |
| { | |
| "epoch": 2.8244958924570573, | |
| "grad_norm": 0.6034400463104248, | |
| "learning_rate": 1.014166338924627e-07, | |
| "loss": 0.3985, | |
| "step": 3310 | |
| }, | |
| { | |
| "epoch": 2.8330310466232795, | |
| "grad_norm": 0.6405271291732788, | |
| "learning_rate": 9.17008557017962e-08, | |
| "loss": 0.3951, | |
| "step": 3320 | |
| }, | |
| { | |
| "epoch": 2.8415662007895017, | |
| "grad_norm": 0.6416633725166321, | |
| "learning_rate": 8.24698949159919e-08, | |
| "loss": 0.4049, | |
| "step": 3330 | |
| }, | |
| { | |
| "epoch": 2.850101354955724, | |
| "grad_norm": 0.6081260442733765, | |
| "learning_rate": 7.372466332349726e-08, | |
| "loss": 0.4023, | |
| "step": 3340 | |
| }, | |
| { | |
| "epoch": 2.858636509121946, | |
| "grad_norm": 0.6582528352737427, | |
| "learning_rate": 6.546602473484153e-08, | |
| "loss": 0.4095, | |
| "step": 3350 | |
| }, | |
| { | |
| "epoch": 2.8671716632881683, | |
| "grad_norm": 0.6263580918312073, | |
| "learning_rate": 5.769479489731666e-08, | |
| "loss": 0.4099, | |
| "step": 3360 | |
| }, | |
| { | |
| "epoch": 2.8757068174543905, | |
| "grad_norm": 0.6049970984458923, | |
| "learning_rate": 5.0411741414396756e-08, | |
| "loss": 0.3992, | |
| "step": 3370 | |
| }, | |
| { | |
| "epoch": 2.8842419716206122, | |
| "grad_norm": 0.6298121809959412, | |
| "learning_rate": 4.3617583669922636e-08, | |
| "loss": 0.4035, | |
| "step": 3380 | |
| }, | |
| { | |
| "epoch": 2.8927771257868344, | |
| "grad_norm": 0.617342472076416, | |
| "learning_rate": 3.73129927570437e-08, | |
| "loss": 0.4079, | |
| "step": 3390 | |
| }, | |
| { | |
| "epoch": 2.9013122799530566, | |
| "grad_norm": 0.6217492818832397, | |
| "learning_rate": 3.1498591411928125e-08, | |
| "loss": 0.4021, | |
| "step": 3400 | |
| }, | |
| { | |
| "epoch": 2.909847434119279, | |
| "grad_norm": 0.618754506111145, | |
| "learning_rate": 2.6174953952254334e-08, | |
| "loss": 0.3977, | |
| "step": 3410 | |
| }, | |
| { | |
| "epoch": 2.918382588285501, | |
| "grad_norm": 0.5799270868301392, | |
| "learning_rate": 2.134260622048301e-08, | |
| "loss": 0.394, | |
| "step": 3420 | |
| }, | |
| { | |
| "epoch": 2.9269177424517228, | |
| "grad_norm": 0.6517247557640076, | |
| "learning_rate": 1.700202553191588e-08, | |
| "loss": 0.4023, | |
| "step": 3430 | |
| }, | |
| { | |
| "epoch": 2.935452896617945, | |
| "grad_norm": 0.6470068693161011, | |
| "learning_rate": 1.315364062754898e-08, | |
| "loss": 0.4056, | |
| "step": 3440 | |
| }, | |
| { | |
| "epoch": 2.943988050784167, | |
| "grad_norm": 0.6544038653373718, | |
| "learning_rate": 9.797831631724319e-09, | |
| "loss": 0.4105, | |
| "step": 3450 | |
| }, | |
| { | |
| "epoch": 2.9525232049503893, | |
| "grad_norm": 0.6202620267868042, | |
| "learning_rate": 6.9349300145826745e-09, | |
| "loss": 0.4102, | |
| "step": 3460 | |
| }, | |
| { | |
| "epoch": 2.9610583591166115, | |
| "grad_norm": 0.6056284308433533, | |
| "learning_rate": 4.56521855932257e-09, | |
| "loss": 0.4006, | |
| "step": 3470 | |
| }, | |
| { | |
| "epoch": 2.9695935132828337, | |
| "grad_norm": 0.6306059956550598, | |
| "learning_rate": 2.6889313342692848e-09, | |
| "loss": 0.404, | |
| "step": 3480 | |
| }, | |
| { | |
| "epoch": 2.978128667449056, | |
| "grad_norm": 0.6009540557861328, | |
| "learning_rate": 1.3062536697533433e-09, | |
| "loss": 0.3995, | |
| "step": 3490 | |
| }, | |
| { | |
| "epoch": 2.986663821615278, | |
| "grad_norm": 0.649553656578064, | |
| "learning_rate": 4.173221398046101e-10, | |
| "loss": 0.4005, | |
| "step": 3500 | |
| }, | |
| { | |
| "epoch": 2.9951989757815003, | |
| "grad_norm": 0.6136988401412964, | |
| "learning_rate": 2.2224548664184777e-11, | |
| "loss": 0.4075, | |
| "step": 3510 | |
| } | |
| ], | |
| "logging_steps": 10, | |
| "max_steps": 3513, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 3, | |
| "save_steps": 500.0, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": true | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 1.1182668447261655e+18, | |
| "train_batch_size": 4, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |