{ "best_metric": 0.02887474000453949, "best_model_checkpoint": "saves/psy-course/MentaLLaMA-chat-7B/train/fold2/checkpoint-1900", "epoch": 5.0, "eval_steps": 50, "global_step": 3225, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.015503875968992248, "grad_norm": 1.8097485303878784, "learning_rate": 3.0959752321981426e-06, "loss": 1.678, "step": 10 }, { "epoch": 0.031007751937984496, "grad_norm": 1.9774603843688965, "learning_rate": 6.191950464396285e-06, "loss": 1.7145, "step": 20 }, { "epoch": 0.046511627906976744, "grad_norm": 2.036001443862915, "learning_rate": 9.287925696594429e-06, "loss": 1.5924, "step": 30 }, { "epoch": 0.06201550387596899, "grad_norm": 3.265268087387085, "learning_rate": 1.238390092879257e-05, "loss": 1.2457, "step": 40 }, { "epoch": 0.07751937984496124, "grad_norm": 1.1102505922317505, "learning_rate": 1.5479876160990712e-05, "loss": 0.7519, "step": 50 }, { "epoch": 0.07751937984496124, "eval_loss": 0.638915479183197, "eval_runtime": 177.4395, "eval_samples_per_second": 6.464, "eval_steps_per_second": 6.464, "step": 50 }, { "epoch": 0.09302325581395349, "grad_norm": 0.6308065056800842, "learning_rate": 1.8575851393188857e-05, "loss": 0.6351, "step": 60 }, { "epoch": 0.10852713178294573, "grad_norm": 0.8012827634811401, "learning_rate": 2.1671826625387e-05, "loss": 0.4782, "step": 70 }, { "epoch": 0.12403100775193798, "grad_norm": 0.6199682354927063, "learning_rate": 2.476780185758514e-05, "loss": 0.2779, "step": 80 }, { "epoch": 0.13953488372093023, "grad_norm": 0.5094736218452454, "learning_rate": 2.7863777089783283e-05, "loss": 0.1896, "step": 90 }, { "epoch": 0.15503875968992248, "grad_norm": 0.5959056615829468, "learning_rate": 3.0959752321981425e-05, "loss": 0.1485, "step": 100 }, { "epoch": 0.15503875968992248, "eval_loss": 0.11244650185108185, "eval_runtime": 177.3837, "eval_samples_per_second": 6.466, "eval_steps_per_second": 6.466, "step": 100 }, { "epoch": 0.17054263565891473, "grad_norm": 0.919588029384613, "learning_rate": 3.4055727554179566e-05, "loss": 0.1271, "step": 110 }, { "epoch": 0.18604651162790697, "grad_norm": 0.9113105535507202, "learning_rate": 3.7151702786377715e-05, "loss": 0.101, "step": 120 }, { "epoch": 0.20155038759689922, "grad_norm": 0.74349045753479, "learning_rate": 4.024767801857585e-05, "loss": 0.1009, "step": 130 }, { "epoch": 0.21705426356589147, "grad_norm": 0.6830365061759949, "learning_rate": 4.3343653250774e-05, "loss": 0.096, "step": 140 }, { "epoch": 0.23255813953488372, "grad_norm": 0.5473061203956604, "learning_rate": 4.6439628482972134e-05, "loss": 0.074, "step": 150 }, { "epoch": 0.23255813953488372, "eval_loss": 0.06496970355510712, "eval_runtime": 177.3867, "eval_samples_per_second": 6.466, "eval_steps_per_second": 6.466, "step": 150 }, { "epoch": 0.24806201550387597, "grad_norm": 0.45949482917785645, "learning_rate": 4.953560371517028e-05, "loss": 0.076, "step": 160 }, { "epoch": 0.26356589147286824, "grad_norm": 0.6822295188903809, "learning_rate": 5.2631578947368424e-05, "loss": 0.0689, "step": 170 }, { "epoch": 0.27906976744186046, "grad_norm": 0.5395657420158386, "learning_rate": 5.5727554179566566e-05, "loss": 0.0634, "step": 180 }, { "epoch": 0.29457364341085274, "grad_norm": 1.2032052278518677, "learning_rate": 5.882352941176471e-05, "loss": 0.0687, "step": 190 }, { "epoch": 0.31007751937984496, "grad_norm": 0.8278073072433472, "learning_rate": 6.191950464396285e-05, "loss": 0.0655, "step": 200 }, { "epoch": 0.31007751937984496, "eval_loss": 0.06194576248526573, "eval_runtime": 177.5104, "eval_samples_per_second": 6.462, "eval_steps_per_second": 6.462, "step": 200 }, { "epoch": 0.32558139534883723, "grad_norm": 0.48666349053382874, "learning_rate": 6.501547987616098e-05, "loss": 0.0591, "step": 210 }, { "epoch": 0.34108527131782945, "grad_norm": 0.6362942457199097, "learning_rate": 6.811145510835913e-05, "loss": 0.0557, "step": 220 }, { "epoch": 0.35658914728682173, "grad_norm": 0.44466376304626465, "learning_rate": 7.120743034055728e-05, "loss": 0.0531, "step": 230 }, { "epoch": 0.37209302325581395, "grad_norm": 0.473111093044281, "learning_rate": 7.430340557275543e-05, "loss": 0.0704, "step": 240 }, { "epoch": 0.3875968992248062, "grad_norm": 0.6141083836555481, "learning_rate": 7.739938080495357e-05, "loss": 0.0598, "step": 250 }, { "epoch": 0.3875968992248062, "eval_loss": 0.05116499215364456, "eval_runtime": 177.5255, "eval_samples_per_second": 6.461, "eval_steps_per_second": 6.461, "step": 250 }, { "epoch": 0.40310077519379844, "grad_norm": 0.34960028529167175, "learning_rate": 8.04953560371517e-05, "loss": 0.0502, "step": 260 }, { "epoch": 0.4186046511627907, "grad_norm": 0.6346738338470459, "learning_rate": 8.359133126934985e-05, "loss": 0.0469, "step": 270 }, { "epoch": 0.43410852713178294, "grad_norm": 0.34611088037490845, "learning_rate": 8.6687306501548e-05, "loss": 0.0769, "step": 280 }, { "epoch": 0.4496124031007752, "grad_norm": 0.746246337890625, "learning_rate": 8.978328173374613e-05, "loss": 0.0547, "step": 290 }, { "epoch": 0.46511627906976744, "grad_norm": 0.3229033350944519, "learning_rate": 9.287925696594427e-05, "loss": 0.0414, "step": 300 }, { "epoch": 0.46511627906976744, "eval_loss": 0.04494742304086685, "eval_runtime": 177.4386, "eval_samples_per_second": 6.464, "eval_steps_per_second": 6.464, "step": 300 }, { "epoch": 0.4806201550387597, "grad_norm": 0.7018037438392639, "learning_rate": 9.597523219814242e-05, "loss": 0.0469, "step": 310 }, { "epoch": 0.49612403100775193, "grad_norm": 0.3599799871444702, "learning_rate": 9.907120743034056e-05, "loss": 0.0377, "step": 320 }, { "epoch": 0.5116279069767442, "grad_norm": 0.3153678774833679, "learning_rate": 9.999856438185238e-05, "loss": 0.0441, "step": 330 }, { "epoch": 0.5271317829457365, "grad_norm": 0.2962203621864319, "learning_rate": 9.999153298122152e-05, "loss": 0.0522, "step": 340 }, { "epoch": 0.5426356589147286, "grad_norm": 0.1656709760427475, "learning_rate": 9.997864293614054e-05, "loss": 0.0427, "step": 350 }, { "epoch": 0.5426356589147286, "eval_loss": 0.04140669107437134, "eval_runtime": 177.5437, "eval_samples_per_second": 6.46, "eval_steps_per_second": 6.46, "step": 350 }, { "epoch": 0.5581395348837209, "grad_norm": 0.2917194068431854, "learning_rate": 9.995989575722902e-05, "loss": 0.047, "step": 360 }, { "epoch": 0.5736434108527132, "grad_norm": 0.3477674424648285, "learning_rate": 9.993529364152018e-05, "loss": 0.0501, "step": 370 }, { "epoch": 0.5891472868217055, "grad_norm": 0.48425808548927307, "learning_rate": 9.990483947220319e-05, "loss": 0.0593, "step": 380 }, { "epoch": 0.6046511627906976, "grad_norm": 0.21694518625736237, "learning_rate": 9.986853681828546e-05, "loss": 0.0513, "step": 390 }, { "epoch": 0.6201550387596899, "grad_norm": 0.5227199792861938, "learning_rate": 9.982638993417425e-05, "loss": 0.0471, "step": 400 }, { "epoch": 0.6201550387596899, "eval_loss": 0.038663338869810104, "eval_runtime": 177.5346, "eval_samples_per_second": 6.461, "eval_steps_per_second": 6.461, "step": 400 }, { "epoch": 0.6356589147286822, "grad_norm": 0.25696733593940735, "learning_rate": 9.977840375917817e-05, "loss": 0.0475, "step": 410 }, { "epoch": 0.6511627906976745, "grad_norm": 0.22936882078647614, "learning_rate": 9.972458391692827e-05, "loss": 0.0405, "step": 420 }, { "epoch": 0.6666666666666666, "grad_norm": 0.31321045756340027, "learning_rate": 9.966493671471904e-05, "loss": 0.0449, "step": 430 }, { "epoch": 0.6821705426356589, "grad_norm": 0.3151184916496277, "learning_rate": 9.959946914276922e-05, "loss": 0.056, "step": 440 }, { "epoch": 0.6976744186046512, "grad_norm": 0.50681471824646, "learning_rate": 9.952818887340257e-05, "loss": 0.0433, "step": 450 }, { "epoch": 0.6976744186046512, "eval_loss": 0.03619127348065376, "eval_runtime": 177.5548, "eval_samples_per_second": 6.46, "eval_steps_per_second": 6.46, "step": 450 }, { "epoch": 0.7131782945736435, "grad_norm": 0.31854957342147827, "learning_rate": 9.945110426014878e-05, "loss": 0.0494, "step": 460 }, { "epoch": 0.7286821705426356, "grad_norm": 0.28269049525260925, "learning_rate": 9.936822433676444e-05, "loss": 0.0387, "step": 470 }, { "epoch": 0.7441860465116279, "grad_norm": 0.24601627886295319, "learning_rate": 9.927955881617444e-05, "loss": 0.0479, "step": 480 }, { "epoch": 0.7596899224806202, "grad_norm": 0.19982942938804626, "learning_rate": 9.918511808933358e-05, "loss": 0.0378, "step": 490 }, { "epoch": 0.7751937984496124, "grad_norm": 0.20261038839817047, "learning_rate": 9.908491322400885e-05, "loss": 0.0432, "step": 500 }, { "epoch": 0.7751937984496124, "eval_loss": 0.03534514829516411, "eval_runtime": 177.4652, "eval_samples_per_second": 6.463, "eval_steps_per_second": 6.463, "step": 500 }, { "epoch": 0.7906976744186046, "grad_norm": 0.1283186674118042, "learning_rate": 9.897895596348247e-05, "loss": 0.0395, "step": 510 }, { "epoch": 0.8062015503875969, "grad_norm": 0.2949117124080658, "learning_rate": 9.886725872517552e-05, "loss": 0.0326, "step": 520 }, { "epoch": 0.8217054263565892, "grad_norm": 0.42478370666503906, "learning_rate": 9.874983459919277e-05, "loss": 0.0442, "step": 530 }, { "epoch": 0.8372093023255814, "grad_norm": 0.270944207906723, "learning_rate": 9.862669734678867e-05, "loss": 0.038, "step": 540 }, { "epoch": 0.8527131782945736, "grad_norm": 0.19820506870746613, "learning_rate": 9.849786139875452e-05, "loss": 0.0445, "step": 550 }, { "epoch": 0.8527131782945736, "eval_loss": 0.03533301129937172, "eval_runtime": 177.5388, "eval_samples_per_second": 6.461, "eval_steps_per_second": 6.461, "step": 550 }, { "epoch": 0.8682170542635659, "grad_norm": 0.33879902958869934, "learning_rate": 9.836334185372738e-05, "loss": 0.0464, "step": 560 }, { "epoch": 0.8837209302325582, "grad_norm": 0.330963134765625, "learning_rate": 9.822315447642056e-05, "loss": 0.0368, "step": 570 }, { "epoch": 0.8992248062015504, "grad_norm": 0.22305184602737427, "learning_rate": 9.807731569577615e-05, "loss": 0.0296, "step": 580 }, { "epoch": 0.9147286821705426, "grad_norm": 0.28611594438552856, "learning_rate": 9.792584260303964e-05, "loss": 0.0364, "step": 590 }, { "epoch": 0.9302325581395349, "grad_norm": 0.24084076285362244, "learning_rate": 9.776875294975698e-05, "loss": 0.0529, "step": 600 }, { "epoch": 0.9302325581395349, "eval_loss": 0.03533976525068283, "eval_runtime": 177.6164, "eval_samples_per_second": 6.458, "eval_steps_per_second": 6.458, "step": 600 }, { "epoch": 0.9457364341085271, "grad_norm": 0.28942936658859253, "learning_rate": 9.760606514569416e-05, "loss": 0.0381, "step": 610 }, { "epoch": 0.9612403100775194, "grad_norm": 0.2056342214345932, "learning_rate": 9.743779825667984e-05, "loss": 0.0337, "step": 620 }, { "epoch": 0.9767441860465116, "grad_norm": 0.14191323518753052, "learning_rate": 9.726397200237085e-05, "loss": 0.0368, "step": 630 }, { "epoch": 0.9922480620155039, "grad_norm": 0.33825528621673584, "learning_rate": 9.708460675394126e-05, "loss": 0.0465, "step": 640 }, { "epoch": 1.0077519379844961, "grad_norm": 0.12557874619960785, "learning_rate": 9.689972353169499e-05, "loss": 0.0313, "step": 650 }, { "epoch": 1.0077519379844961, "eval_loss": 0.03178540617227554, "eval_runtime": 177.627, "eval_samples_per_second": 6.457, "eval_steps_per_second": 6.457, "step": 650 }, { "epoch": 1.0232558139534884, "grad_norm": 0.35606396198272705, "learning_rate": 9.67093440026024e-05, "loss": 0.0303, "step": 660 }, { "epoch": 1.0387596899224807, "grad_norm": 0.183203786611557, "learning_rate": 9.651349047776112e-05, "loss": 0.0386, "step": 670 }, { "epoch": 1.054263565891473, "grad_norm": 0.23599869012832642, "learning_rate": 9.631218590978126e-05, "loss": 0.0222, "step": 680 }, { "epoch": 1.069767441860465, "grad_norm": 0.13795703649520874, "learning_rate": 9.610545389009562e-05, "loss": 0.0307, "step": 690 }, { "epoch": 1.0852713178294573, "grad_norm": 0.2756401002407074, "learning_rate": 9.589331864619486e-05, "loss": 0.0301, "step": 700 }, { "epoch": 1.0852713178294573, "eval_loss": 0.03220707178115845, "eval_runtime": 177.6556, "eval_samples_per_second": 6.456, "eval_steps_per_second": 6.456, "step": 700 }, { "epoch": 1.1007751937984496, "grad_norm": 0.22982142865657806, "learning_rate": 9.567580503878833e-05, "loss": 0.025, "step": 710 }, { "epoch": 1.1162790697674418, "grad_norm": 0.21597911417484283, "learning_rate": 9.545293855889043e-05, "loss": 0.0342, "step": 720 }, { "epoch": 1.1317829457364341, "grad_norm": 0.16282838582992554, "learning_rate": 9.522474532483337e-05, "loss": 0.0275, "step": 730 }, { "epoch": 1.1472868217054264, "grad_norm": 0.32020196318626404, "learning_rate": 9.499125207920622e-05, "loss": 0.0402, "step": 740 }, { "epoch": 1.1627906976744187, "grad_norm": 0.12987101078033447, "learning_rate": 9.475248618572096e-05, "loss": 0.0289, "step": 750 }, { "epoch": 1.1627906976744187, "eval_loss": 0.03383904695510864, "eval_runtime": 177.667, "eval_samples_per_second": 6.456, "eval_steps_per_second": 6.456, "step": 750 }, { "epoch": 1.178294573643411, "grad_norm": 0.37133800983428955, "learning_rate": 9.45084756260055e-05, "loss": 0.035, "step": 760 }, { "epoch": 1.193798449612403, "grad_norm": 0.07524913549423218, "learning_rate": 9.425924899632458e-05, "loss": 0.0291, "step": 770 }, { "epoch": 1.2093023255813953, "grad_norm": 0.14250116050243378, "learning_rate": 9.400483550422845e-05, "loss": 0.0221, "step": 780 }, { "epoch": 1.2248062015503876, "grad_norm": 0.33380234241485596, "learning_rate": 9.374526496512992e-05, "loss": 0.0303, "step": 790 }, { "epoch": 1.2403100775193798, "grad_norm": 0.6035688519477844, "learning_rate": 9.348056779881025e-05, "loss": 0.0267, "step": 800 }, { "epoch": 1.2403100775193798, "eval_loss": 0.03135337680578232, "eval_runtime": 177.6172, "eval_samples_per_second": 6.458, "eval_steps_per_second": 6.458, "step": 800 }, { "epoch": 1.255813953488372, "grad_norm": 0.25495055317878723, "learning_rate": 9.321077502585417e-05, "loss": 0.0267, "step": 810 }, { "epoch": 1.2713178294573644, "grad_norm": 0.352885365486145, "learning_rate": 9.293591826401451e-05, "loss": 0.0351, "step": 820 }, { "epoch": 1.2868217054263567, "grad_norm": 0.25123533606529236, "learning_rate": 9.265602972450673e-05, "loss": 0.0328, "step": 830 }, { "epoch": 1.302325581395349, "grad_norm": 0.47959160804748535, "learning_rate": 9.237114220823413e-05, "loss": 0.0326, "step": 840 }, { "epoch": 1.3178294573643412, "grad_norm": 0.13225722312927246, "learning_rate": 9.208128910194377e-05, "loss": 0.0314, "step": 850 }, { "epoch": 1.3178294573643412, "eval_loss": 0.03171215206384659, "eval_runtime": 177.6409, "eval_samples_per_second": 6.457, "eval_steps_per_second": 6.457, "step": 850 }, { "epoch": 1.3333333333333333, "grad_norm": 0.27190810441970825, "learning_rate": 9.17865043743138e-05, "loss": 0.031, "step": 860 }, { "epoch": 1.3488372093023255, "grad_norm": 0.10395022481679916, "learning_rate": 9.148682257197245e-05, "loss": 0.0238, "step": 870 }, { "epoch": 1.3643410852713178, "grad_norm": 0.33697041869163513, "learning_rate": 9.118227881544967e-05, "loss": 0.0397, "step": 880 }, { "epoch": 1.37984496124031, "grad_norm": 0.2214643359184265, "learning_rate": 9.087290879506104e-05, "loss": 0.0281, "step": 890 }, { "epoch": 1.3953488372093024, "grad_norm": 0.11761048436164856, "learning_rate": 9.055874876672519e-05, "loss": 0.0382, "step": 900 }, { "epoch": 1.3953488372093024, "eval_loss": 0.03273608908057213, "eval_runtime": 177.7052, "eval_samples_per_second": 6.455, "eval_steps_per_second": 6.455, "step": 900 }, { "epoch": 1.4108527131782946, "grad_norm": 0.17550455033779144, "learning_rate": 9.023983554771492e-05, "loss": 0.0263, "step": 910 }, { "epoch": 1.4263565891472867, "grad_norm": 0.21578268706798553, "learning_rate": 8.991620651234242e-05, "loss": 0.0261, "step": 920 }, { "epoch": 1.441860465116279, "grad_norm": 0.21228443086147308, "learning_rate": 8.958789958757928e-05, "loss": 0.029, "step": 930 }, { "epoch": 1.4573643410852712, "grad_norm": 0.2734227180480957, "learning_rate": 8.925495324861177e-05, "loss": 0.0286, "step": 940 }, { "epoch": 1.4728682170542635, "grad_norm": 0.4174731373786926, "learning_rate": 8.89174065143318e-05, "loss": 0.0354, "step": 950 }, { "epoch": 1.4728682170542635, "eval_loss": 0.031981565058231354, "eval_runtime": 177.7226, "eval_samples_per_second": 6.454, "eval_steps_per_second": 6.454, "step": 950 }, { "epoch": 1.4883720930232558, "grad_norm": 0.3433147370815277, "learning_rate": 8.857529894276412e-05, "loss": 0.0352, "step": 960 }, { "epoch": 1.503875968992248, "grad_norm": 0.21089136600494385, "learning_rate": 8.822867062643059e-05, "loss": 0.0298, "step": 970 }, { "epoch": 1.5193798449612403, "grad_norm": 0.1780281960964203, "learning_rate": 8.78775621876514e-05, "loss": 0.0319, "step": 980 }, { "epoch": 1.5348837209302326, "grad_norm": 0.3001708686351776, "learning_rate": 8.752201477378459e-05, "loss": 0.0217, "step": 990 }, { "epoch": 1.550387596899225, "grad_norm": 0.15130403637886047, "learning_rate": 8.716207005240382e-05, "loss": 0.0265, "step": 1000 }, { "epoch": 1.550387596899225, "eval_loss": 0.03209313005208969, "eval_runtime": 177.6849, "eval_samples_per_second": 6.455, "eval_steps_per_second": 6.455, "step": 1000 }, { "epoch": 1.5658914728682172, "grad_norm": 0.28896772861480713, "learning_rate": 8.679777020641525e-05, "loss": 0.0345, "step": 1010 }, { "epoch": 1.5813953488372094, "grad_norm": 0.4542350471019745, "learning_rate": 8.6429157929114e-05, "loss": 0.033, "step": 1020 }, { "epoch": 1.5968992248062015, "grad_norm": 0.26365530490875244, "learning_rate": 8.60562764191808e-05, "loss": 0.0277, "step": 1030 }, { "epoch": 1.6124031007751938, "grad_norm": 0.17537464201450348, "learning_rate": 8.567916937561944e-05, "loss": 0.0307, "step": 1040 }, { "epoch": 1.627906976744186, "grad_norm": 0.39150458574295044, "learning_rate": 8.529788099263554e-05, "loss": 0.0301, "step": 1050 }, { "epoch": 1.627906976744186, "eval_loss": 0.03326583281159401, "eval_runtime": 177.6473, "eval_samples_per_second": 6.457, "eval_steps_per_second": 6.457, "step": 1050 }, { "epoch": 1.6434108527131783, "grad_norm": 0.2252993881702423, "learning_rate": 8.49124559544573e-05, "loss": 0.0279, "step": 1060 }, { "epoch": 1.6589147286821704, "grad_norm": 0.3254898488521576, "learning_rate": 8.452293943009889e-05, "loss": 0.0269, "step": 1070 }, { "epoch": 1.6744186046511627, "grad_norm": 0.2006569355726242, "learning_rate": 8.412937706806692e-05, "loss": 0.0361, "step": 1080 }, { "epoch": 1.689922480620155, "grad_norm": 0.2423427700996399, "learning_rate": 8.373181499101077e-05, "loss": 0.0301, "step": 1090 }, { "epoch": 1.7054263565891472, "grad_norm": 0.36214545369148254, "learning_rate": 8.333029979031737e-05, "loss": 0.0262, "step": 1100 }, { "epoch": 1.7054263565891472, "eval_loss": 0.031156621873378754, "eval_runtime": 177.8049, "eval_samples_per_second": 6.451, "eval_steps_per_second": 6.451, "step": 1100 }, { "epoch": 1.7209302325581395, "grad_norm": 0.17312179505825043, "learning_rate": 8.292487852065104e-05, "loss": 0.0392, "step": 1110 }, { "epoch": 1.7364341085271318, "grad_norm": 0.275713711977005, "learning_rate": 8.251559869443897e-05, "loss": 0.0322, "step": 1120 }, { "epoch": 1.751937984496124, "grad_norm": 0.33000582456588745, "learning_rate": 8.210250827630313e-05, "loss": 0.029, "step": 1130 }, { "epoch": 1.7674418604651163, "grad_norm": 0.29563483595848083, "learning_rate": 8.168565567743924e-05, "loss": 0.0263, "step": 1140 }, { "epoch": 1.7829457364341086, "grad_norm": 0.31916943192481995, "learning_rate": 8.126508974994321e-05, "loss": 0.0273, "step": 1150 }, { "epoch": 1.7829457364341086, "eval_loss": 0.030605774372816086, "eval_runtime": 177.725, "eval_samples_per_second": 6.454, "eval_steps_per_second": 6.454, "step": 1150 }, { "epoch": 1.7984496124031009, "grad_norm": 0.1885177344083786, "learning_rate": 8.084085978108611e-05, "loss": 0.0169, "step": 1160 }, { "epoch": 1.8139534883720931, "grad_norm": 0.15678612887859344, "learning_rate": 8.041301548753804e-05, "loss": 0.0284, "step": 1170 }, { "epoch": 1.8294573643410854, "grad_norm": 0.24891909956932068, "learning_rate": 7.998160700954165e-05, "loss": 0.0244, "step": 1180 }, { "epoch": 1.8449612403100775, "grad_norm": 0.27802011370658875, "learning_rate": 7.954668490503621e-05, "loss": 0.027, "step": 1190 }, { "epoch": 1.8604651162790697, "grad_norm": 0.2340599000453949, "learning_rate": 7.910830014373237e-05, "loss": 0.0283, "step": 1200 }, { "epoch": 1.8604651162790697, "eval_loss": 0.029726749286055565, "eval_runtime": 177.8483, "eval_samples_per_second": 6.449, "eval_steps_per_second": 6.449, "step": 1200 }, { "epoch": 1.875968992248062, "grad_norm": 0.13624481856822968, "learning_rate": 7.866650410113907e-05, "loss": 0.0243, "step": 1210 }, { "epoch": 1.8914728682170543, "grad_norm": 0.30038079619407654, "learning_rate": 7.822134855254263e-05, "loss": 0.0296, "step": 1220 }, { "epoch": 1.9069767441860463, "grad_norm": 0.23166057467460632, "learning_rate": 7.777288566693905e-05, "loss": 0.0273, "step": 1230 }, { "epoch": 1.9224806201550386, "grad_norm": 0.26034536957740784, "learning_rate": 7.732116800092018e-05, "loss": 0.029, "step": 1240 }, { "epoch": 1.937984496124031, "grad_norm": 0.2269134223461151, "learning_rate": 7.686624849251453e-05, "loss": 0.0381, "step": 1250 }, { "epoch": 1.937984496124031, "eval_loss": 0.029867732897400856, "eval_runtime": 177.8147, "eval_samples_per_second": 6.451, "eval_steps_per_second": 6.451, "step": 1250 }, { "epoch": 1.9534883720930232, "grad_norm": 0.30028459429740906, "learning_rate": 7.640818045498324e-05, "loss": 0.0378, "step": 1260 }, { "epoch": 1.9689922480620154, "grad_norm": 0.10794702172279358, "learning_rate": 7.59470175705722e-05, "loss": 0.0303, "step": 1270 }, { "epoch": 1.9844961240310077, "grad_norm": 0.27119511365890503, "learning_rate": 7.548281388422088e-05, "loss": 0.0361, "step": 1280 }, { "epoch": 2.0, "grad_norm": 0.22147968411445618, "learning_rate": 7.50156237972286e-05, "loss": 0.0263, "step": 1290 }, { "epoch": 2.0155038759689923, "grad_norm": 0.12922464311122894, "learning_rate": 7.454550206087921e-05, "loss": 0.0207, "step": 1300 }, { "epoch": 2.0155038759689923, "eval_loss": 0.029372563585639, "eval_runtime": 177.7077, "eval_samples_per_second": 6.454, "eval_steps_per_second": 6.454, "step": 1300 }, { "epoch": 2.0310077519379846, "grad_norm": 0.20252908766269684, "learning_rate": 7.407250377002451e-05, "loss": 0.0184, "step": 1310 }, { "epoch": 2.046511627906977, "grad_norm": 0.20628713071346283, "learning_rate": 7.359668435662757e-05, "loss": 0.0216, "step": 1320 }, { "epoch": 2.062015503875969, "grad_norm": 0.16820959746837616, "learning_rate": 7.311809958326659e-05, "loss": 0.0153, "step": 1330 }, { "epoch": 2.0775193798449614, "grad_norm": 0.28085780143737793, "learning_rate": 7.263680553659983e-05, "loss": 0.0172, "step": 1340 }, { "epoch": 2.0930232558139537, "grad_norm": 0.4147201180458069, "learning_rate": 7.21528586207927e-05, "loss": 0.0163, "step": 1350 }, { "epoch": 2.0930232558139537, "eval_loss": 0.03293036296963692, "eval_runtime": 177.2709, "eval_samples_per_second": 6.47, "eval_steps_per_second": 6.47, "step": 1350 }, { "epoch": 2.108527131782946, "grad_norm": 0.2516096830368042, "learning_rate": 7.166631555090759e-05, "loss": 0.025, "step": 1360 }, { "epoch": 2.124031007751938, "grad_norm": 0.26919105648994446, "learning_rate": 7.117723334625731e-05, "loss": 0.024, "step": 1370 }, { "epoch": 2.13953488372093, "grad_norm": 0.228290855884552, "learning_rate": 7.068566932372278e-05, "loss": 0.0183, "step": 1380 }, { "epoch": 2.1550387596899223, "grad_norm": 0.28446343541145325, "learning_rate": 7.019168109103599e-05, "loss": 0.0162, "step": 1390 }, { "epoch": 2.1705426356589146, "grad_norm": 0.19133096933364868, "learning_rate": 6.96953265400287e-05, "loss": 0.0236, "step": 1400 }, { "epoch": 2.1705426356589146, "eval_loss": 0.031053410843014717, "eval_runtime": 176.3297, "eval_samples_per_second": 6.505, "eval_steps_per_second": 6.505, "step": 1400 }, { "epoch": 2.186046511627907, "grad_norm": 0.14185263216495514, "learning_rate": 6.9196663839848e-05, "loss": 0.0142, "step": 1410 }, { "epoch": 2.201550387596899, "grad_norm": 0.2277035117149353, "learning_rate": 6.869575143013934e-05, "loss": 0.0166, "step": 1420 }, { "epoch": 2.2170542635658914, "grad_norm": 0.24037382006645203, "learning_rate": 6.819264801419771e-05, "loss": 0.0274, "step": 1430 }, { "epoch": 2.2325581395348837, "grad_norm": 0.10074973106384277, "learning_rate": 6.768741255208813e-05, "loss": 0.0154, "step": 1440 }, { "epoch": 2.248062015503876, "grad_norm": 0.24921472370624542, "learning_rate": 6.718010425373599e-05, "loss": 0.0191, "step": 1450 }, { "epoch": 2.248062015503876, "eval_loss": 0.030957119539380074, "eval_runtime": 175.6681, "eval_samples_per_second": 6.529, "eval_steps_per_second": 6.529, "step": 1450 }, { "epoch": 2.2635658914728682, "grad_norm": 0.22070886194705963, "learning_rate": 6.667078257198796e-05, "loss": 0.0184, "step": 1460 }, { "epoch": 2.2790697674418605, "grad_norm": 0.39727696776390076, "learning_rate": 6.615950719564466e-05, "loss": 0.0142, "step": 1470 }, { "epoch": 2.294573643410853, "grad_norm": 0.31478437781333923, "learning_rate": 6.56463380424655e-05, "loss": 0.0252, "step": 1480 }, { "epoch": 2.310077519379845, "grad_norm": 0.1986107975244522, "learning_rate": 6.51313352521468e-05, "loss": 0.0151, "step": 1490 }, { "epoch": 2.3255813953488373, "grad_norm": 0.3045531213283539, "learning_rate": 6.46145591792738e-05, "loss": 0.0243, "step": 1500 }, { "epoch": 2.3255813953488373, "eval_loss": 0.03084620088338852, "eval_runtime": 175.2668, "eval_samples_per_second": 6.544, "eval_steps_per_second": 6.544, "step": 1500 }, { "epoch": 2.3410852713178296, "grad_norm": 0.08667182922363281, "learning_rate": 6.409607038624759e-05, "loss": 0.0142, "step": 1510 }, { "epoch": 2.356589147286822, "grad_norm": 0.17333078384399414, "learning_rate": 6.357592963618761e-05, "loss": 0.0241, "step": 1520 }, { "epoch": 2.3720930232558137, "grad_norm": 0.20276910066604614, "learning_rate": 6.305419788581067e-05, "loss": 0.0185, "step": 1530 }, { "epoch": 2.387596899224806, "grad_norm": 0.287349134683609, "learning_rate": 6.253093627828724e-05, "loss": 0.0176, "step": 1540 }, { "epoch": 2.4031007751937983, "grad_norm": 0.32019197940826416, "learning_rate": 6.200620613607597e-05, "loss": 0.0165, "step": 1550 }, { "epoch": 2.4031007751937983, "eval_loss": 0.03274994716048241, "eval_runtime": 174.5607, "eval_samples_per_second": 6.571, "eval_steps_per_second": 6.571, "step": 1550 }, { "epoch": 2.4186046511627906, "grad_norm": 0.34726911783218384, "learning_rate": 6.148006895373706e-05, "loss": 0.0196, "step": 1560 }, { "epoch": 2.434108527131783, "grad_norm": 0.15877018868923187, "learning_rate": 6.0952586390725644e-05, "loss": 0.0171, "step": 1570 }, { "epoch": 2.449612403100775, "grad_norm": 0.21150444447994232, "learning_rate": 6.042382026416563e-05, "loss": 0.0164, "step": 1580 }, { "epoch": 2.4651162790697674, "grad_norm": 0.19103744626045227, "learning_rate": 5.989383254160529e-05, "loss": 0.0171, "step": 1590 }, { "epoch": 2.4806201550387597, "grad_norm": 0.44117987155914307, "learning_rate": 5.936268533375506e-05, "loss": 0.0224, "step": 1600 }, { "epoch": 2.4806201550387597, "eval_loss": 0.03294115141034126, "eval_runtime": 174.4995, "eval_samples_per_second": 6.573, "eval_steps_per_second": 6.573, "step": 1600 }, { "epoch": 2.496124031007752, "grad_norm": 0.3606320321559906, "learning_rate": 5.883044088720865e-05, "loss": 0.0214, "step": 1610 }, { "epoch": 2.511627906976744, "grad_norm": 0.13814851641654968, "learning_rate": 5.829716157714814e-05, "loss": 0.0177, "step": 1620 }, { "epoch": 2.5271317829457365, "grad_norm": 0.23484356701374054, "learning_rate": 5.77629099000341e-05, "loss": 0.0263, "step": 1630 }, { "epoch": 2.5426356589147288, "grad_norm": 0.21404676139354706, "learning_rate": 5.7227748466281486e-05, "loss": 0.0187, "step": 1640 }, { "epoch": 2.558139534883721, "grad_norm": 0.3276318609714508, "learning_rate": 5.669173999292208e-05, "loss": 0.0289, "step": 1650 }, { "epoch": 2.558139534883721, "eval_loss": 0.03190125152468681, "eval_runtime": 174.4148, "eval_samples_per_second": 6.576, "eval_steps_per_second": 6.576, "step": 1650 }, { "epoch": 2.5736434108527133, "grad_norm": 0.41611889004707336, "learning_rate": 5.615494729625458e-05, "loss": 0.0237, "step": 1660 }, { "epoch": 2.5891472868217056, "grad_norm": 0.372302383184433, "learning_rate": 5.561743328448296e-05, "loss": 0.0246, "step": 1670 }, { "epoch": 2.604651162790698, "grad_norm": 0.1654350608587265, "learning_rate": 5.5079260950344035e-05, "loss": 0.019, "step": 1680 }, { "epoch": 2.62015503875969, "grad_norm": 0.09135215729475021, "learning_rate": 5.454049336372531e-05, "loss": 0.0165, "step": 1690 }, { "epoch": 2.6356589147286824, "grad_norm": 0.24403300881385803, "learning_rate": 5.4001193664273454e-05, "loss": 0.014, "step": 1700 }, { "epoch": 2.6356589147286824, "eval_loss": 0.03161098435521126, "eval_runtime": 174.2648, "eval_samples_per_second": 6.582, "eval_steps_per_second": 6.582, "step": 1700 }, { "epoch": 2.6511627906976747, "grad_norm": 0.19514554738998413, "learning_rate": 5.346142505399495e-05, "loss": 0.0198, "step": 1710 }, { "epoch": 2.6666666666666665, "grad_norm": 0.13203318417072296, "learning_rate": 5.292125078984925e-05, "loss": 0.0181, "step": 1720 }, { "epoch": 2.682170542635659, "grad_norm": 0.09618645161390305, "learning_rate": 5.2380734176335425e-05, "loss": 0.0171, "step": 1730 }, { "epoch": 2.697674418604651, "grad_norm": 0.17840640246868134, "learning_rate": 5.183993855807343e-05, "loss": 0.0186, "step": 1740 }, { "epoch": 2.7131782945736433, "grad_norm": 0.12565214931964874, "learning_rate": 5.1298927312380586e-05, "loss": 0.0182, "step": 1750 }, { "epoch": 2.7131782945736433, "eval_loss": 0.03337998315691948, "eval_runtime": 174.2061, "eval_samples_per_second": 6.584, "eval_steps_per_second": 6.584, "step": 1750 }, { "epoch": 2.7286821705426356, "grad_norm": 0.1932082623243332, "learning_rate": 5.075776384184411e-05, "loss": 0.0275, "step": 1760 }, { "epoch": 2.744186046511628, "grad_norm": 0.14624637365341187, "learning_rate": 5.021651156689094e-05, "loss": 0.0284, "step": 1770 }, { "epoch": 2.75968992248062, "grad_norm": 0.20881560444831848, "learning_rate": 4.967523391835521e-05, "loss": 0.0217, "step": 1780 }, { "epoch": 2.7751937984496124, "grad_norm": 0.2833938002586365, "learning_rate": 4.9133994330044644e-05, "loss": 0.0234, "step": 1790 }, { "epoch": 2.7906976744186047, "grad_norm": 0.23201899230480194, "learning_rate": 4.85928562313066e-05, "loss": 0.0175, "step": 1800 }, { "epoch": 2.7906976744186047, "eval_loss": 0.029777053743600845, "eval_runtime": 174.1005, "eval_samples_per_second": 6.588, "eval_steps_per_second": 6.588, "step": 1800 }, { "epoch": 2.806201550387597, "grad_norm": 0.2119215726852417, "learning_rate": 4.8051883039594616e-05, "loss": 0.0151, "step": 1810 }, { "epoch": 2.8217054263565893, "grad_norm": 0.30587702989578247, "learning_rate": 4.751113815303624e-05, "loss": 0.0187, "step": 1820 }, { "epoch": 2.8372093023255816, "grad_norm": 0.18018727004528046, "learning_rate": 4.697068494300343e-05, "loss": 0.0151, "step": 1830 }, { "epoch": 2.8527131782945734, "grad_norm": 0.21248644590377808, "learning_rate": 4.6430586746685724e-05, "loss": 0.0222, "step": 1840 }, { "epoch": 2.8682170542635657, "grad_norm": 0.18737761676311493, "learning_rate": 4.589090685966758e-05, "loss": 0.0218, "step": 1850 }, { "epoch": 2.8682170542635657, "eval_loss": 0.029682571068406105, "eval_runtime": 174.0789, "eval_samples_per_second": 6.589, "eval_steps_per_second": 6.589, "step": 1850 }, { "epoch": 2.883720930232558, "grad_norm": 0.3078942894935608, "learning_rate": 4.535170852851073e-05, "loss": 0.0265, "step": 1860 }, { "epoch": 2.89922480620155, "grad_norm": 0.2036602944135666, "learning_rate": 4.481305494334201e-05, "loss": 0.0167, "step": 1870 }, { "epoch": 2.9147286821705425, "grad_norm": 0.08146923035383224, "learning_rate": 4.427500923044801e-05, "loss": 0.0199, "step": 1880 }, { "epoch": 2.9302325581395348, "grad_norm": 0.12505419552326202, "learning_rate": 4.373763444487705e-05, "loss": 0.0179, "step": 1890 }, { "epoch": 2.945736434108527, "grad_norm": 0.18068896234035492, "learning_rate": 4.3200993563049725e-05, "loss": 0.018, "step": 1900 }, { "epoch": 2.945736434108527, "eval_loss": 0.02887474000453949, "eval_runtime": 173.8738, "eval_samples_per_second": 6.597, "eval_steps_per_second": 6.597, "step": 1900 }, { "epoch": 2.9612403100775193, "grad_norm": 0.19300121068954468, "learning_rate": 4.266514947537839e-05, "loss": 0.0157, "step": 1910 }, { "epoch": 2.9767441860465116, "grad_norm": 0.1891685128211975, "learning_rate": 4.2130164978896916e-05, "loss": 0.0188, "step": 1920 }, { "epoch": 2.992248062015504, "grad_norm": 0.027009131386876106, "learning_rate": 4.159610276990137e-05, "loss": 0.0118, "step": 1930 }, { "epoch": 3.007751937984496, "grad_norm": 0.11215299367904663, "learning_rate": 4.106302543660235e-05, "loss": 0.0164, "step": 1940 }, { "epoch": 3.0232558139534884, "grad_norm": 0.07163859158754349, "learning_rate": 4.053099545179028e-05, "loss": 0.01, "step": 1950 }, { "epoch": 3.0232558139534884, "eval_loss": 0.030872860923409462, "eval_runtime": 173.945, "eval_samples_per_second": 6.594, "eval_steps_per_second": 6.594, "step": 1950 }, { "epoch": 3.0387596899224807, "grad_norm": 0.12158221006393433, "learning_rate": 4.0000075165513845e-05, "loss": 0.014, "step": 1960 }, { "epoch": 3.054263565891473, "grad_norm": 0.17903171479701996, "learning_rate": 3.9470326797773216e-05, "loss": 0.0085, "step": 1970 }, { "epoch": 3.0697674418604652, "grad_norm": 0.23486894369125366, "learning_rate": 3.8941812431228166e-05, "loss": 0.012, "step": 1980 }, { "epoch": 3.0852713178294575, "grad_norm": 0.19141437113285065, "learning_rate": 3.8414594003922515e-05, "loss": 0.0141, "step": 1990 }, { "epoch": 3.10077519379845, "grad_norm": 0.047276828438043594, "learning_rate": 3.788873330202544e-05, "loss": 0.0109, "step": 2000 }, { "epoch": 3.10077519379845, "eval_loss": 0.033768460154533386, "eval_runtime": 173.8687, "eval_samples_per_second": 6.597, "eval_steps_per_second": 6.597, "step": 2000 }, { "epoch": 3.116279069767442, "grad_norm": 0.3079964518547058, "learning_rate": 3.736429195259051e-05, "loss": 0.0069, "step": 2010 }, { "epoch": 3.1317829457364343, "grad_norm": 0.207097128033638, "learning_rate": 3.684133141633358e-05, "loss": 0.009, "step": 2020 }, { "epoch": 3.147286821705426, "grad_norm": 0.16703131794929504, "learning_rate": 3.6319912980429846e-05, "loss": 0.0098, "step": 2030 }, { "epoch": 3.1627906976744184, "grad_norm": 0.1511267125606537, "learning_rate": 3.580009775133168e-05, "loss": 0.0076, "step": 2040 }, { "epoch": 3.1782945736434107, "grad_norm": 0.184535413980484, "learning_rate": 3.528194664760714e-05, "loss": 0.0076, "step": 2050 }, { "epoch": 3.1782945736434107, "eval_loss": 0.03470822051167488, "eval_runtime": 173.7828, "eval_samples_per_second": 6.6, "eval_steps_per_second": 6.6, "step": 2050 }, { "epoch": 3.193798449612403, "grad_norm": 0.19592364132404327, "learning_rate": 3.476552039280096e-05, "loss": 0.0089, "step": 2060 }, { "epoch": 3.2093023255813953, "grad_norm": 0.1496165543794632, "learning_rate": 3.42508795083181e-05, "loss": 0.0092, "step": 2070 }, { "epoch": 3.2248062015503876, "grad_norm": 0.19381049275398254, "learning_rate": 3.373808430633106e-05, "loss": 0.0086, "step": 2080 }, { "epoch": 3.24031007751938, "grad_norm": 0.30487295985221863, "learning_rate": 3.32271948827118e-05, "loss": 0.0099, "step": 2090 }, { "epoch": 3.255813953488372, "grad_norm": 0.21724191308021545, "learning_rate": 3.2718271109988863e-05, "loss": 0.0087, "step": 2100 }, { "epoch": 3.255813953488372, "eval_loss": 0.03577126935124397, "eval_runtime": 173.7017, "eval_samples_per_second": 6.603, "eval_steps_per_second": 6.603, "step": 2100 }, { "epoch": 3.2713178294573644, "grad_norm": 0.8536518812179565, "learning_rate": 3.2211372630330835e-05, "loss": 0.0094, "step": 2110 }, { "epoch": 3.2868217054263567, "grad_norm": 0.23373013734817505, "learning_rate": 3.170655884855661e-05, "loss": 0.0121, "step": 2120 }, { "epoch": 3.302325581395349, "grad_norm": 0.09981165081262589, "learning_rate": 3.120388892517368e-05, "loss": 0.0193, "step": 2130 }, { "epoch": 3.317829457364341, "grad_norm": 0.2072608470916748, "learning_rate": 3.070342176944494e-05, "loss": 0.0135, "step": 2140 }, { "epoch": 3.3333333333333335, "grad_norm": 0.07687430083751678, "learning_rate": 3.0205216032484805e-05, "loss": 0.0092, "step": 2150 }, { "epoch": 3.3333333333333335, "eval_loss": 0.032335054129362106, "eval_runtime": 173.7494, "eval_samples_per_second": 6.601, "eval_steps_per_second": 6.601, "step": 2150 }, { "epoch": 3.3488372093023258, "grad_norm": 0.2054612785577774, "learning_rate": 2.970933010038599e-05, "loss": 0.013, "step": 2160 }, { "epoch": 3.3643410852713176, "grad_norm": 0.17977559566497803, "learning_rate": 2.921582208737681e-05, "loss": 0.0088, "step": 2170 }, { "epoch": 3.37984496124031, "grad_norm": 0.23377564549446106, "learning_rate": 2.872474982901081e-05, "loss": 0.0136, "step": 2180 }, { "epoch": 3.395348837209302, "grad_norm": 0.16202200949192047, "learning_rate": 2.8236170875388744e-05, "loss": 0.01, "step": 2190 }, { "epoch": 3.4108527131782944, "grad_norm": 0.2739998698234558, "learning_rate": 2.775014248441422e-05, "loss": 0.0078, "step": 2200 }, { "epoch": 3.4108527131782944, "eval_loss": 0.03312241658568382, "eval_runtime": 173.7097, "eval_samples_per_second": 6.603, "eval_steps_per_second": 6.603, "step": 2200 }, { "epoch": 3.4263565891472867, "grad_norm": 0.10180757939815521, "learning_rate": 2.726672161508341e-05, "loss": 0.0098, "step": 2210 }, { "epoch": 3.441860465116279, "grad_norm": 0.27830150723457336, "learning_rate": 2.678596492080984e-05, "loss": 0.0077, "step": 2220 }, { "epoch": 3.4573643410852712, "grad_norm": 0.25859320163726807, "learning_rate": 2.630792874278516e-05, "loss": 0.0099, "step": 2230 }, { "epoch": 3.4728682170542635, "grad_norm": 0.4165274500846863, "learning_rate": 2.583266910337624e-05, "loss": 0.0095, "step": 2240 }, { "epoch": 3.488372093023256, "grad_norm": 0.5020641684532166, "learning_rate": 2.5360241699559816e-05, "loss": 0.0109, "step": 2250 }, { "epoch": 3.488372093023256, "eval_loss": 0.035617418587207794, "eval_runtime": 173.6138, "eval_samples_per_second": 6.607, "eval_steps_per_second": 6.607, "step": 2250 }, { "epoch": 3.503875968992248, "grad_norm": 0.19337324798107147, "learning_rate": 2.4890701896395146e-05, "loss": 0.0104, "step": 2260 }, { "epoch": 3.5193798449612403, "grad_norm": 0.1879683881998062, "learning_rate": 2.4424104720535735e-05, "loss": 0.0101, "step": 2270 }, { "epoch": 3.5348837209302326, "grad_norm": 0.063876673579216, "learning_rate": 2.3960504853780462e-05, "loss": 0.0115, "step": 2280 }, { "epoch": 3.550387596899225, "grad_norm": 0.11843185126781464, "learning_rate": 2.349995662666547e-05, "loss": 0.0085, "step": 2290 }, { "epoch": 3.565891472868217, "grad_norm": 0.21422308683395386, "learning_rate": 2.3042514012096843e-05, "loss": 0.0137, "step": 2300 }, { "epoch": 3.565891472868217, "eval_loss": 0.035960037261247635, "eval_runtime": 173.6653, "eval_samples_per_second": 6.605, "eval_steps_per_second": 6.605, "step": 2300 }, { "epoch": 3.5813953488372094, "grad_norm": 0.354839563369751, "learning_rate": 2.2588230619025407e-05, "loss": 0.0159, "step": 2310 }, { "epoch": 3.5968992248062017, "grad_norm": 0.2568114995956421, "learning_rate": 2.213715968616425e-05, "loss": 0.0109, "step": 2320 }, { "epoch": 3.612403100775194, "grad_norm": 0.06852281093597412, "learning_rate": 2.16893540757494e-05, "loss": 0.0091, "step": 2330 }, { "epoch": 3.6279069767441863, "grad_norm": 0.194693461060524, "learning_rate": 2.1244866267344866e-05, "loss": 0.0084, "step": 2340 }, { "epoch": 3.6434108527131785, "grad_norm": 0.3488699495792389, "learning_rate": 2.080374835169235e-05, "loss": 0.013, "step": 2350 }, { "epoch": 3.6434108527131785, "eval_loss": 0.03496725112199783, "eval_runtime": 173.6559, "eval_samples_per_second": 6.605, "eval_steps_per_second": 6.605, "step": 2350 }, { "epoch": 3.6589147286821704, "grad_norm": 0.16774657368659973, "learning_rate": 2.0366052024606612e-05, "loss": 0.0082, "step": 2360 }, { "epoch": 3.6744186046511627, "grad_norm": 0.2563646733760834, "learning_rate": 1.9931828580917107e-05, "loss": 0.0114, "step": 2370 }, { "epoch": 3.689922480620155, "grad_norm": 0.254089891910553, "learning_rate": 1.9501128908456523e-05, "loss": 0.0087, "step": 2380 }, { "epoch": 3.705426356589147, "grad_norm": 0.2673187255859375, "learning_rate": 1.9074003482097258e-05, "loss": 0.0093, "step": 2390 }, { "epoch": 3.7209302325581395, "grad_norm": 0.47327151894569397, "learning_rate": 1.8650502357835925e-05, "loss": 0.0133, "step": 2400 }, { "epoch": 3.7209302325581395, "eval_loss": 0.03533458709716797, "eval_runtime": 173.6028, "eval_samples_per_second": 6.607, "eval_steps_per_second": 6.607, "step": 2400 }, { "epoch": 3.7364341085271318, "grad_norm": 0.4366307556629181, "learning_rate": 1.82306751669274e-05, "loss": 0.0085, "step": 2410 }, { "epoch": 3.751937984496124, "grad_norm": 0.39723125100135803, "learning_rate": 1.7814571110068135e-05, "loss": 0.0123, "step": 2420 }, { "epoch": 3.7674418604651163, "grad_norm": 0.16450275480747223, "learning_rate": 1.740223895163039e-05, "loss": 0.0091, "step": 2430 }, { "epoch": 3.7829457364341086, "grad_norm": 0.06844169646501541, "learning_rate": 1.6993727013947336e-05, "loss": 0.0109, "step": 2440 }, { "epoch": 3.798449612403101, "grad_norm": 0.29116564989089966, "learning_rate": 1.6589083171649977e-05, "loss": 0.0068, "step": 2450 }, { "epoch": 3.798449612403101, "eval_loss": 0.035682931542396545, "eval_runtime": 173.5613, "eval_samples_per_second": 6.609, "eval_steps_per_second": 6.609, "step": 2450 }, { "epoch": 3.813953488372093, "grad_norm": 0.1012692078948021, "learning_rate": 1.6188354846056698e-05, "loss": 0.008, "step": 2460 }, { "epoch": 3.8294573643410854, "grad_norm": 0.1340951770544052, "learning_rate": 1.579158899961575e-05, "loss": 0.0054, "step": 2470 }, { "epoch": 3.8449612403100772, "grad_norm": 0.22337476909160614, "learning_rate": 1.5398832130401637e-05, "loss": 0.0147, "step": 2480 }, { "epoch": 3.8604651162790695, "grad_norm": 0.18346606194972992, "learning_rate": 1.5010130266665807e-05, "loss": 0.0108, "step": 2490 }, { "epoch": 3.875968992248062, "grad_norm": 0.2582853436470032, "learning_rate": 1.4625528961442591e-05, "loss": 0.012, "step": 2500 }, { "epoch": 3.875968992248062, "eval_loss": 0.034770578145980835, "eval_runtime": 173.5234, "eval_samples_per_second": 6.61, "eval_steps_per_second": 6.61, "step": 2500 }, { "epoch": 3.891472868217054, "grad_norm": 0.22837190330028534, "learning_rate": 1.4245073287210608e-05, "loss": 0.0098, "step": 2510 }, { "epoch": 3.9069767441860463, "grad_norm": 0.49539613723754883, "learning_rate": 1.3868807830610653e-05, "loss": 0.016, "step": 2520 }, { "epoch": 3.9224806201550386, "grad_norm": 0.17993128299713135, "learning_rate": 1.3496776687220514e-05, "loss": 0.0129, "step": 2530 }, { "epoch": 3.937984496124031, "grad_norm": 0.25489673018455505, "learning_rate": 1.3129023456387151e-05, "loss": 0.0068, "step": 2540 }, { "epoch": 3.953488372093023, "grad_norm": 0.06681143492460251, "learning_rate": 1.2765591236117324e-05, "loss": 0.0088, "step": 2550 }, { "epoch": 3.953488372093023, "eval_loss": 0.034369926899671555, "eval_runtime": 173.574, "eval_samples_per_second": 6.608, "eval_steps_per_second": 6.608, "step": 2550 }, { "epoch": 3.9689922480620154, "grad_norm": 0.14447127282619476, "learning_rate": 1.240652261802669e-05, "loss": 0.0105, "step": 2560 }, { "epoch": 3.9844961240310077, "grad_norm": 0.2608015239238739, "learning_rate": 1.205185968234847e-05, "loss": 0.0136, "step": 2570 }, { "epoch": 4.0, "grad_norm": 0.17687876522541046, "learning_rate": 1.1701643993001916e-05, "loss": 0.0093, "step": 2580 }, { "epoch": 4.015503875968992, "grad_norm": 0.1070009246468544, "learning_rate": 1.1355916592721316e-05, "loss": 0.0045, "step": 2590 }, { "epoch": 4.0310077519379846, "grad_norm": 0.11302992701530457, "learning_rate": 1.1014717998246099e-05, "loss": 0.0066, "step": 2600 }, { "epoch": 4.0310077519379846, "eval_loss": 0.03459889069199562, "eval_runtime": 173.5445, "eval_samples_per_second": 6.609, "eval_steps_per_second": 6.609, "step": 2600 }, { "epoch": 4.046511627906977, "grad_norm": 0.107364721596241, "learning_rate": 1.0678088195572517e-05, "loss": 0.0053, "step": 2610 }, { "epoch": 4.062015503875969, "grad_norm": 0.2821330428123474, "learning_rate": 1.0346066635267676e-05, "loss": 0.0069, "step": 2620 }, { "epoch": 4.077519379844961, "grad_norm": 0.17317578196525574, "learning_rate": 1.001869222784611e-05, "loss": 0.0064, "step": 2630 }, { "epoch": 4.093023255813954, "grad_norm": 0.08701343834400177, "learning_rate": 9.696003339209819e-06, "loss": 0.007, "step": 2640 }, { "epoch": 4.108527131782946, "grad_norm": 0.23731686174869537, "learning_rate": 9.378037786152055e-06, "loss": 0.0052, "step": 2650 }, { "epoch": 4.108527131782946, "eval_loss": 0.03613542765378952, "eval_runtime": 173.8232, "eval_samples_per_second": 6.599, "eval_steps_per_second": 6.599, "step": 2650 }, { "epoch": 4.124031007751938, "grad_norm": 0.13787880539894104, "learning_rate": 9.064832831925441e-06, "loss": 0.0054, "step": 2660 }, { "epoch": 4.1395348837209305, "grad_norm": 0.1381884068250656, "learning_rate": 8.756425181875028e-06, "loss": 0.0035, "step": 2670 }, { "epoch": 4.155038759689923, "grad_norm": 0.16911053657531738, "learning_rate": 8.452850979136617e-06, "loss": 0.0045, "step": 2680 }, { "epoch": 4.170542635658915, "grad_norm": 0.04475264623761177, "learning_rate": 8.15414580040117e-06, "loss": 0.0032, "step": 2690 }, { "epoch": 4.186046511627907, "grad_norm": 0.09773126989603043, "learning_rate": 7.860344651745366e-06, "loss": 0.008, "step": 2700 }, { "epoch": 4.186046511627907, "eval_loss": 0.037369657307863235, "eval_runtime": 173.9253, "eval_samples_per_second": 6.595, "eval_steps_per_second": 6.595, "step": 2700 }, { "epoch": 4.2015503875969, "grad_norm": 0.1751956194639206, "learning_rate": 7.571481964529226e-06, "loss": 0.0057, "step": 2710 }, { "epoch": 4.217054263565892, "grad_norm": 0.06625787168741226, "learning_rate": 7.287591591360926e-06, "loss": 0.0028, "step": 2720 }, { "epoch": 4.232558139534884, "grad_norm": 0.1029408872127533, "learning_rate": 7.008706802129606e-06, "loss": 0.0069, "step": 2730 }, { "epoch": 4.248062015503876, "grad_norm": 0.01715020090341568, "learning_rate": 6.734860280106292e-06, "loss": 0.0021, "step": 2740 }, { "epoch": 4.263565891472869, "grad_norm": 0.5773080587387085, "learning_rate": 6.466084118113735e-06, "loss": 0.0062, "step": 2750 }, { "epoch": 4.263565891472869, "eval_loss": 0.0383269228041172, "eval_runtime": 173.7893, "eval_samples_per_second": 6.6, "eval_steps_per_second": 6.6, "step": 2750 }, { "epoch": 4.27906976744186, "grad_norm": 0.09289313852787018, "learning_rate": 6.202409814765325e-06, "loss": 0.0043, "step": 2760 }, { "epoch": 4.294573643410852, "grad_norm": 0.16086533665657043, "learning_rate": 5.9438682707736725e-06, "loss": 0.0064, "step": 2770 }, { "epoch": 4.310077519379845, "grad_norm": 0.175432026386261, "learning_rate": 5.690489785329301e-06, "loss": 0.0052, "step": 2780 }, { "epoch": 4.325581395348837, "grad_norm": 0.03287180885672569, "learning_rate": 5.442304052549752e-06, "loss": 0.0037, "step": 2790 }, { "epoch": 4.341085271317829, "grad_norm": 0.09991113096475601, "learning_rate": 5.199340157999733e-06, "loss": 0.005, "step": 2800 }, { "epoch": 4.341085271317829, "eval_loss": 0.038646504282951355, "eval_runtime": 173.5908, "eval_samples_per_second": 6.607, "eval_steps_per_second": 6.607, "step": 2800 }, { "epoch": 4.3565891472868215, "grad_norm": 0.07446504384279251, "learning_rate": 4.961626575282396e-06, "loss": 0.0028, "step": 2810 }, { "epoch": 4.372093023255814, "grad_norm": 0.1563124656677246, "learning_rate": 4.7291911627025235e-06, "loss": 0.0045, "step": 2820 }, { "epoch": 4.387596899224806, "grad_norm": 0.09954715520143509, "learning_rate": 4.502061160001725e-06, "loss": 0.0038, "step": 2830 }, { "epoch": 4.403100775193798, "grad_norm": 0.03728143870830536, "learning_rate": 4.280263185166078e-06, "loss": 0.007, "step": 2840 }, { "epoch": 4.4186046511627906, "grad_norm": 0.14879867434501648, "learning_rate": 4.063823231306757e-06, "loss": 0.004, "step": 2850 }, { "epoch": 4.4186046511627906, "eval_loss": 0.03950056806206703, "eval_runtime": 173.7546, "eval_samples_per_second": 6.601, "eval_steps_per_second": 6.601, "step": 2850 }, { "epoch": 4.434108527131783, "grad_norm": 0.04011745750904083, "learning_rate": 3.8527666636137885e-06, "loss": 0.0039, "step": 2860 }, { "epoch": 4.449612403100775, "grad_norm": 0.16904973983764648, "learning_rate": 3.6471182163834825e-06, "loss": 0.0075, "step": 2870 }, { "epoch": 4.465116279069767, "grad_norm": 0.05882781371474266, "learning_rate": 3.4469019901197054e-06, "loss": 0.0027, "step": 2880 }, { "epoch": 4.48062015503876, "grad_norm": 0.36288192868232727, "learning_rate": 3.252141448709495e-06, "loss": 0.0052, "step": 2890 }, { "epoch": 4.496124031007752, "grad_norm": 0.25636863708496094, "learning_rate": 3.0628594166732693e-06, "loss": 0.0075, "step": 2900 }, { "epoch": 4.496124031007752, "eval_loss": 0.040012069046497345, "eval_runtime": 173.5684, "eval_samples_per_second": 6.608, "eval_steps_per_second": 6.608, "step": 2900 }, { "epoch": 4.511627906976744, "grad_norm": 0.07002587616443634, "learning_rate": 2.8790780764899384e-06, "loss": 0.0053, "step": 2910 }, { "epoch": 4.5271317829457365, "grad_norm": 0.18481026589870453, "learning_rate": 2.700818965997315e-06, "loss": 0.0064, "step": 2920 }, { "epoch": 4.542635658914729, "grad_norm": 0.1123792976140976, "learning_rate": 2.528102975867991e-06, "loss": 0.0033, "step": 2930 }, { "epoch": 4.558139534883721, "grad_norm": 0.039597250521183014, "learning_rate": 2.3609503471611284e-06, "loss": 0.0041, "step": 2940 }, { "epoch": 4.573643410852713, "grad_norm": 0.05195772275328636, "learning_rate": 2.1993806689503738e-06, "loss": 0.003, "step": 2950 }, { "epoch": 4.573643410852713, "eval_loss": 0.040207840502262115, "eval_runtime": 173.5139, "eval_samples_per_second": 6.61, "eval_steps_per_second": 6.61, "step": 2950 }, { "epoch": 4.589147286821706, "grad_norm": 0.06371759623289108, "learning_rate": 2.0434128760281167e-06, "loss": 0.0021, "step": 2960 }, { "epoch": 4.604651162790698, "grad_norm": 0.0519726499915123, "learning_rate": 1.893065246686504e-06, "loss": 0.0029, "step": 2970 }, { "epoch": 4.62015503875969, "grad_norm": 0.05901438370347023, "learning_rate": 1.74835540057533e-06, "loss": 0.0067, "step": 2980 }, { "epoch": 4.635658914728682, "grad_norm": 0.16322703659534454, "learning_rate": 1.6093002966371617e-06, "loss": 0.0044, "step": 2990 }, { "epoch": 4.651162790697675, "grad_norm": 0.25651177763938904, "learning_rate": 1.4759162311198783e-06, "loss": 0.0066, "step": 3000 }, { "epoch": 4.651162790697675, "eval_loss": 0.0405086986720562, "eval_runtime": 173.4642, "eval_samples_per_second": 6.612, "eval_steps_per_second": 6.612, "step": 3000 }, { "epoch": 4.666666666666667, "grad_norm": 0.009503856301307678, "learning_rate": 1.348218835666859e-06, "loss": 0.0032, "step": 3010 }, { "epoch": 4.682170542635659, "grad_norm": 0.12524276971817017, "learning_rate": 1.2262230754850445e-06, "loss": 0.0032, "step": 3020 }, { "epoch": 4.6976744186046515, "grad_norm": 0.06170937791466713, "learning_rate": 1.109943247591172e-06, "loss": 0.0045, "step": 3030 }, { "epoch": 4.713178294573644, "grad_norm": 0.12521466612815857, "learning_rate": 9.993929791362323e-07, "loss": 0.0052, "step": 3040 }, { "epoch": 4.728682170542635, "grad_norm": 0.17083865404129028, "learning_rate": 8.945852258084863e-07, "loss": 0.005, "step": 3050 }, { "epoch": 4.728682170542635, "eval_loss": 0.04061457887291908, "eval_runtime": 173.5187, "eval_samples_per_second": 6.61, "eval_steps_per_second": 6.61, "step": 3050 }, { "epoch": 4.7441860465116275, "grad_norm": 0.35159188508987427, "learning_rate": 7.955322703151358e-07, "loss": 0.0046, "step": 3060 }, { "epoch": 4.75968992248062, "grad_norm": 0.2449195384979248, "learning_rate": 7.022457209428901e-07, "loss": 0.0044, "step": 3070 }, { "epoch": 4.775193798449612, "grad_norm": 0.2619343101978302, "learning_rate": 6.147365101975666e-07, "loss": 0.0035, "step": 3080 }, { "epoch": 4.790697674418604, "grad_norm": 0.09077528119087219, "learning_rate": 5.33014893522854e-07, "loss": 0.0032, "step": 3090 }, { "epoch": 4.8062015503875966, "grad_norm": 0.14284414052963257, "learning_rate": 4.57090448098485e-07, "loss": 0.0067, "step": 3100 }, { "epoch": 4.8062015503875966, "eval_loss": 0.04071947932243347, "eval_runtime": 173.7232, "eval_samples_per_second": 6.602, "eval_steps_per_second": 6.602, "step": 3100 }, { "epoch": 4.821705426356589, "grad_norm": 0.10482044517993927, "learning_rate": 3.8697207171781714e-07, "loss": 0.0037, "step": 3110 }, { "epoch": 4.837209302325581, "grad_norm": 0.03253592178225517, "learning_rate": 3.2266798174512837e-07, "loss": 0.0045, "step": 3120 }, { "epoch": 4.852713178294573, "grad_norm": 0.03622818365693092, "learning_rate": 2.6418571415255387e-07, "loss": 0.0093, "step": 3130 }, { "epoch": 4.868217054263566, "grad_norm": 0.09220151603221893, "learning_rate": 2.1153212263695378e-07, "loss": 0.0046, "step": 3140 }, { "epoch": 4.883720930232558, "grad_norm": 0.1942395716905594, "learning_rate": 1.6471337781669982e-07, "loss": 0.0067, "step": 3150 }, { "epoch": 4.883720930232558, "eval_loss": 0.040707945823669434, "eval_runtime": 173.7396, "eval_samples_per_second": 6.602, "eval_steps_per_second": 6.602, "step": 3150 }, { "epoch": 4.89922480620155, "grad_norm": 0.09173876792192459, "learning_rate": 1.237349665085097e-07, "loss": 0.0049, "step": 3160 }, { "epoch": 4.9147286821705425, "grad_norm": 0.045584529638290405, "learning_rate": 8.86016910844667e-08, "loss": 0.003, "step": 3170 }, { "epoch": 4.930232558139535, "grad_norm": 0.46576347947120667, "learning_rate": 5.9317668909192323e-08, "loss": 0.0061, "step": 3180 }, { "epoch": 4.945736434108527, "grad_norm": 0.059987977147102356, "learning_rate": 3.588633185730994e-08, "loss": 0.0062, "step": 3190 }, { "epoch": 4.961240310077519, "grad_norm": 0.2531404197216034, "learning_rate": 1.8310425911294283e-08, "loss": 0.006, "step": 3200 }, { "epoch": 4.961240310077519, "eval_loss": 0.04073048010468483, "eval_runtime": 173.54, "eval_samples_per_second": 6.609, "eval_steps_per_second": 6.609, "step": 3200 }, { "epoch": 4.976744186046512, "grad_norm": 0.24996767938137054, "learning_rate": 6.592010839612251e-09, "loss": 0.0076, "step": 3210 }, { "epoch": 4.992248062015504, "grad_norm": 0.020084669813513756, "learning_rate": 7.324599553770739e-10, "loss": 0.0025, "step": 3220 }, { "epoch": 5.0, "step": 3225, "total_flos": 8.718478050646426e+17, "train_loss": 0.049027663960821866, "train_runtime": 36714.2196, "train_samples_per_second": 1.405, "train_steps_per_second": 0.088 } ], "logging_steps": 10, "max_steps": 3225, "num_input_tokens_seen": 0, "num_train_epochs": 5, "save_steps": 50, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 8.718478050646426e+17, "train_batch_size": 1, "trial_name": null, "trial_params": null }