{ "best_global_step": 4662, "best_metric": 0.9187405866743169, "best_model_checkpoint": "./automod-model/model_versions/v4-masked/checkpoint-4662", "epoch": 7.0, "eval_steps": 500, "global_step": 4662, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.015015015015015015, "grad_norm": 17.097240447998047, "learning_rate": 3.6e-07, "loss": 2.9671, "step": 10 }, { "epoch": 0.03003003003003003, "grad_norm": 26.079153060913086, "learning_rate": 7.6e-07, "loss": 3.0981, "step": 20 }, { "epoch": 0.04504504504504504, "grad_norm": 80.04959106445312, "learning_rate": 1.1600000000000001e-06, "loss": 2.5492, "step": 30 }, { "epoch": 0.06006006006006006, "grad_norm": 20.597259521484375, "learning_rate": 1.56e-06, "loss": 2.1283, "step": 40 }, { "epoch": 0.07507507507507508, "grad_norm": 33.89236831665039, "learning_rate": 1.9600000000000003e-06, "loss": 1.821, "step": 50 }, { "epoch": 0.09009009009009009, "grad_norm": 21.228269577026367, "learning_rate": 2.3600000000000003e-06, "loss": 1.0551, "step": 60 }, { "epoch": 0.10510510510510511, "grad_norm": 29.945491790771484, "learning_rate": 2.7600000000000003e-06, "loss": 0.906, "step": 70 }, { "epoch": 0.12012012012012012, "grad_norm": 21.597726821899414, "learning_rate": 3.1600000000000002e-06, "loss": 0.7201, "step": 80 }, { "epoch": 0.13513513513513514, "grad_norm": 21.56252670288086, "learning_rate": 3.5600000000000002e-06, "loss": 0.3313, "step": 90 }, { "epoch": 0.15015015015015015, "grad_norm": 9.781754493713379, "learning_rate": 3.96e-06, "loss": 0.3786, "step": 100 }, { "epoch": 0.16516516516516516, "grad_norm": 13.030562400817871, "learning_rate": 4.360000000000001e-06, "loss": 0.3193, "step": 110 }, { "epoch": 0.18018018018018017, "grad_norm": 18.68300437927246, "learning_rate": 4.76e-06, "loss": 0.2353, "step": 120 }, { "epoch": 0.19519519519519518, "grad_norm": 12.98038101196289, "learning_rate": 5.1600000000000006e-06, "loss": 0.2529, "step": 130 }, { "epoch": 0.21021021021021022, "grad_norm": 14.075587272644043, "learning_rate": 5.560000000000001e-06, "loss": 0.2724, "step": 140 }, { "epoch": 0.22522522522522523, "grad_norm": 13.491477966308594, "learning_rate": 5.9600000000000005e-06, "loss": 0.2253, "step": 150 }, { "epoch": 0.24024024024024024, "grad_norm": 21.4825439453125, "learning_rate": 6.360000000000001e-06, "loss": 0.1718, "step": 160 }, { "epoch": 0.2552552552552553, "grad_norm": 12.049821853637695, "learning_rate": 6.760000000000001e-06, "loss": 0.2332, "step": 170 }, { "epoch": 0.2702702702702703, "grad_norm": 13.302751541137695, "learning_rate": 7.16e-06, "loss": 0.1648, "step": 180 }, { "epoch": 0.2852852852852853, "grad_norm": 9.489476203918457, "learning_rate": 7.5600000000000005e-06, "loss": 0.1742, "step": 190 }, { "epoch": 0.3003003003003003, "grad_norm": 7.511724948883057, "learning_rate": 7.960000000000002e-06, "loss": 0.1392, "step": 200 }, { "epoch": 0.3153153153153153, "grad_norm": 6.69696044921875, "learning_rate": 8.36e-06, "loss": 0.1368, "step": 210 }, { "epoch": 0.3303303303303303, "grad_norm": 19.48908042907715, "learning_rate": 8.76e-06, "loss": 0.1694, "step": 220 }, { "epoch": 0.34534534534534533, "grad_norm": 9.891258239746094, "learning_rate": 9.16e-06, "loss": 0.106, "step": 230 }, { "epoch": 0.36036036036036034, "grad_norm": 6.785653114318848, "learning_rate": 9.56e-06, "loss": 0.2347, "step": 240 }, { "epoch": 0.37537537537537535, "grad_norm": 11.124282836914062, "learning_rate": 9.960000000000001e-06, "loss": 0.1421, "step": 250 }, { "epoch": 0.39039039039039036, "grad_norm": 13.130078315734863, "learning_rate": 1.036e-05, "loss": 0.1036, "step": 260 }, { "epoch": 0.40540540540540543, "grad_norm": 6.414909362792969, "learning_rate": 1.0760000000000002e-05, "loss": 0.1492, "step": 270 }, { "epoch": 0.42042042042042044, "grad_norm": 9.707856178283691, "learning_rate": 1.1160000000000002e-05, "loss": 0.1308, "step": 280 }, { "epoch": 0.43543543543543545, "grad_norm": 7.036477565765381, "learning_rate": 1.156e-05, "loss": 0.1186, "step": 290 }, { "epoch": 0.45045045045045046, "grad_norm": 5.08420991897583, "learning_rate": 1.196e-05, "loss": 0.1248, "step": 300 }, { "epoch": 0.46546546546546547, "grad_norm": 6.691461563110352, "learning_rate": 1.236e-05, "loss": 0.1096, "step": 310 }, { "epoch": 0.4804804804804805, "grad_norm": 8.506667137145996, "learning_rate": 1.2760000000000001e-05, "loss": 0.1377, "step": 320 }, { "epoch": 0.4954954954954955, "grad_norm": 7.334710597991943, "learning_rate": 1.3160000000000001e-05, "loss": 0.1051, "step": 330 }, { "epoch": 0.5105105105105106, "grad_norm": 12.647950172424316, "learning_rate": 1.3560000000000002e-05, "loss": 0.1136, "step": 340 }, { "epoch": 0.5255255255255256, "grad_norm": 4.561657905578613, "learning_rate": 1.396e-05, "loss": 0.1091, "step": 350 }, { "epoch": 0.5405405405405406, "grad_norm": 6.011996269226074, "learning_rate": 1.4360000000000001e-05, "loss": 0.0891, "step": 360 }, { "epoch": 0.5555555555555556, "grad_norm": 9.521617889404297, "learning_rate": 1.4760000000000001e-05, "loss": 0.0702, "step": 370 }, { "epoch": 0.5705705705705706, "grad_norm": 4.111408710479736, "learning_rate": 1.516e-05, "loss": 0.105, "step": 380 }, { "epoch": 0.5855855855855856, "grad_norm": 3.560914993286133, "learning_rate": 1.556e-05, "loss": 0.0766, "step": 390 }, { "epoch": 0.6006006006006006, "grad_norm": 2.4863393306732178, "learning_rate": 1.5960000000000003e-05, "loss": 0.0804, "step": 400 }, { "epoch": 0.6156156156156156, "grad_norm": 8.447433471679688, "learning_rate": 1.636e-05, "loss": 0.0755, "step": 410 }, { "epoch": 0.6306306306306306, "grad_norm": 6.882669448852539, "learning_rate": 1.6760000000000002e-05, "loss": 0.0712, "step": 420 }, { "epoch": 0.6456456456456456, "grad_norm": 3.4703919887542725, "learning_rate": 1.7160000000000002e-05, "loss": 0.0546, "step": 430 }, { "epoch": 0.6606606606606606, "grad_norm": 1.0200519561767578, "learning_rate": 1.756e-05, "loss": 0.0514, "step": 440 }, { "epoch": 0.6756756756756757, "grad_norm": 5.531317710876465, "learning_rate": 1.796e-05, "loss": 0.0762, "step": 450 }, { "epoch": 0.6906906906906907, "grad_norm": 0.6142644882202148, "learning_rate": 1.8360000000000004e-05, "loss": 0.0354, "step": 460 }, { "epoch": 0.7057057057057057, "grad_norm": 6.253670692443848, "learning_rate": 1.876e-05, "loss": 0.0394, "step": 470 }, { "epoch": 0.7207207207207207, "grad_norm": 6.93025016784668, "learning_rate": 1.916e-05, "loss": 0.0477, "step": 480 }, { "epoch": 0.7357357357357357, "grad_norm": 6.35200309753418, "learning_rate": 1.9560000000000002e-05, "loss": 0.0327, "step": 490 }, { "epoch": 0.7507507507507507, "grad_norm": 5.108675003051758, "learning_rate": 1.9960000000000002e-05, "loss": 0.0575, "step": 500 }, { "epoch": 0.7657657657657657, "grad_norm": 4.811997413635254, "learning_rate": 1.9962717481358742e-05, "loss": 0.0545, "step": 510 }, { "epoch": 0.7807807807807807, "grad_norm": 3.6220436096191406, "learning_rate": 1.992129246064623e-05, "loss": 0.0202, "step": 520 }, { "epoch": 0.7957957957957958, "grad_norm": 6.347226142883301, "learning_rate": 1.9879867439933723e-05, "loss": 0.0431, "step": 530 }, { "epoch": 0.8108108108108109, "grad_norm": 0.9845024347305298, "learning_rate": 1.983844241922121e-05, "loss": 0.0353, "step": 540 }, { "epoch": 0.8258258258258259, "grad_norm": 6.9723663330078125, "learning_rate": 1.97970173985087e-05, "loss": 0.0495, "step": 550 }, { "epoch": 0.8408408408408409, "grad_norm": 0.35112687945365906, "learning_rate": 1.9755592377796192e-05, "loss": 0.0249, "step": 560 }, { "epoch": 0.8558558558558559, "grad_norm": 6.217226505279541, "learning_rate": 1.971416735708368e-05, "loss": 0.0367, "step": 570 }, { "epoch": 0.8708708708708709, "grad_norm": 1.6721233129501343, "learning_rate": 1.967274233637117e-05, "loss": 0.0279, "step": 580 }, { "epoch": 0.8858858858858859, "grad_norm": 6.501187324523926, "learning_rate": 1.963131731565866e-05, "loss": 0.037, "step": 590 }, { "epoch": 0.9009009009009009, "grad_norm": 3.4928908348083496, "learning_rate": 1.9589892294946147e-05, "loss": 0.0341, "step": 600 }, { "epoch": 0.9159159159159159, "grad_norm": 4.611274719238281, "learning_rate": 1.954846727423364e-05, "loss": 0.0475, "step": 610 }, { "epoch": 0.9309309309309309, "grad_norm": 7.894728660583496, "learning_rate": 1.9507042253521128e-05, "loss": 0.0478, "step": 620 }, { "epoch": 0.9459459459459459, "grad_norm": 4.463768005371094, "learning_rate": 1.9465617232808617e-05, "loss": 0.0303, "step": 630 }, { "epoch": 0.960960960960961, "grad_norm": 3.8476738929748535, "learning_rate": 1.9424192212096106e-05, "loss": 0.0216, "step": 640 }, { "epoch": 0.975975975975976, "grad_norm": 1.4353941679000854, "learning_rate": 1.9382767191383598e-05, "loss": 0.0115, "step": 650 }, { "epoch": 0.990990990990991, "grad_norm": 0.5383431315422058, "learning_rate": 1.9341342170671087e-05, "loss": 0.0316, "step": 660 }, { "epoch": 1.0, "eval_accuracy": 0.844590780809031, "eval_f1": 0.9061312298314809, "eval_loss": 0.21899376809597015, "eval_roc_auc": 0.9366743383999476, "eval_runtime": 7.3751, "eval_samples_per_second": 720.673, "eval_steps_per_second": 22.644, "step": 666 }, { "epoch": 1.006006006006006, "grad_norm": 2.482923984527588, "learning_rate": 1.9299917149958575e-05, "loss": 0.0243, "step": 670 }, { "epoch": 1.021021021021021, "grad_norm": 0.28770723938941956, "learning_rate": 1.9258492129246067e-05, "loss": 0.0174, "step": 680 }, { "epoch": 1.0360360360360361, "grad_norm": 3.5773494243621826, "learning_rate": 1.9217067108533556e-05, "loss": 0.0317, "step": 690 }, { "epoch": 1.0510510510510511, "grad_norm": 7.554081439971924, "learning_rate": 1.9175642087821045e-05, "loss": 0.0322, "step": 700 }, { "epoch": 1.0660660660660661, "grad_norm": 0.5560117959976196, "learning_rate": 1.9134217067108537e-05, "loss": 0.0192, "step": 710 }, { "epoch": 1.0810810810810811, "grad_norm": 2.8596460819244385, "learning_rate": 1.9092792046396026e-05, "loss": 0.0201, "step": 720 }, { "epoch": 1.0960960960960962, "grad_norm": 2.245648145675659, "learning_rate": 1.9051367025683514e-05, "loss": 0.0206, "step": 730 }, { "epoch": 1.1111111111111112, "grad_norm": 1.595831274986267, "learning_rate": 1.9009942004971003e-05, "loss": 0.0145, "step": 740 }, { "epoch": 1.1261261261261262, "grad_norm": 3.3559751510620117, "learning_rate": 1.8968516984258495e-05, "loss": 0.0259, "step": 750 }, { "epoch": 1.1411411411411412, "grad_norm": 8.872766494750977, "learning_rate": 1.8927091963545984e-05, "loss": 0.0316, "step": 760 }, { "epoch": 1.1561561561561562, "grad_norm": 1.1749708652496338, "learning_rate": 1.8885666942833473e-05, "loss": 0.0335, "step": 770 }, { "epoch": 1.1711711711711712, "grad_norm": 2.3856115341186523, "learning_rate": 1.884424192212096e-05, "loss": 0.0258, "step": 780 }, { "epoch": 1.1861861861861862, "grad_norm": 0.8851059675216675, "learning_rate": 1.880281690140845e-05, "loss": 0.0135, "step": 790 }, { "epoch": 1.2012012012012012, "grad_norm": 3.2970917224884033, "learning_rate": 1.8761391880695942e-05, "loss": 0.0234, "step": 800 }, { "epoch": 1.2162162162162162, "grad_norm": 5.862252712249756, "learning_rate": 1.871996685998343e-05, "loss": 0.0265, "step": 810 }, { "epoch": 1.2312312312312312, "grad_norm": 6.934352874755859, "learning_rate": 1.867854183927092e-05, "loss": 0.0209, "step": 820 }, { "epoch": 1.2462462462462462, "grad_norm": 5.499032974243164, "learning_rate": 1.8637116818558412e-05, "loss": 0.0381, "step": 830 }, { "epoch": 1.2612612612612613, "grad_norm": 1.9378541707992554, "learning_rate": 1.85956917978459e-05, "loss": 0.0355, "step": 840 }, { "epoch": 1.2762762762762763, "grad_norm": 3.2807366847991943, "learning_rate": 1.855426677713339e-05, "loss": 0.0252, "step": 850 }, { "epoch": 1.2912912912912913, "grad_norm": 0.08878616243600845, "learning_rate": 1.851284175642088e-05, "loss": 0.0184, "step": 860 }, { "epoch": 1.3063063063063063, "grad_norm": 0.21245735883712769, "learning_rate": 1.847141673570837e-05, "loss": 0.0107, "step": 870 }, { "epoch": 1.3213213213213213, "grad_norm": 0.23962010443210602, "learning_rate": 1.842999171499586e-05, "loss": 0.0236, "step": 880 }, { "epoch": 1.3363363363363363, "grad_norm": 5.516748905181885, "learning_rate": 1.838856669428335e-05, "loss": 0.0175, "step": 890 }, { "epoch": 1.3513513513513513, "grad_norm": 0.09256813675165176, "learning_rate": 1.834714167357084e-05, "loss": 0.0153, "step": 900 }, { "epoch": 1.3663663663663663, "grad_norm": 4.04671049118042, "learning_rate": 1.830571665285833e-05, "loss": 0.0304, "step": 910 }, { "epoch": 1.3813813813813813, "grad_norm": 0.05709666758775711, "learning_rate": 1.8264291632145817e-05, "loss": 0.0226, "step": 920 }, { "epoch": 1.3963963963963963, "grad_norm": 2.605480194091797, "learning_rate": 1.8222866611433306e-05, "loss": 0.0191, "step": 930 }, { "epoch": 1.4114114114114114, "grad_norm": 12.805084228515625, "learning_rate": 1.8181441590720795e-05, "loss": 0.023, "step": 940 }, { "epoch": 1.4264264264264264, "grad_norm": 0.10644625872373581, "learning_rate": 1.8140016570008287e-05, "loss": 0.0138, "step": 950 }, { "epoch": 1.4414414414414414, "grad_norm": 0.4645148515701294, "learning_rate": 1.8098591549295776e-05, "loss": 0.0207, "step": 960 }, { "epoch": 1.4564564564564564, "grad_norm": 6.761512279510498, "learning_rate": 1.8057166528583265e-05, "loss": 0.0228, "step": 970 }, { "epoch": 1.4714714714714714, "grad_norm": 0.7149978876113892, "learning_rate": 1.8015741507870757e-05, "loss": 0.0113, "step": 980 }, { "epoch": 1.4864864864864864, "grad_norm": 4.93594217300415, "learning_rate": 1.7974316487158245e-05, "loss": 0.0156, "step": 990 }, { "epoch": 1.5015015015015014, "grad_norm": 0.030841587111353874, "learning_rate": 1.7932891466445734e-05, "loss": 0.0224, "step": 1000 }, { "epoch": 1.5165165165165164, "grad_norm": 3.3840746879577637, "learning_rate": 1.7891466445733223e-05, "loss": 0.027, "step": 1010 }, { "epoch": 1.5315315315315314, "grad_norm": 0.03548554331064224, "learning_rate": 1.7850041425020715e-05, "loss": 0.0232, "step": 1020 }, { "epoch": 1.5465465465465464, "grad_norm": 3.062281847000122, "learning_rate": 1.7808616404308204e-05, "loss": 0.0199, "step": 1030 }, { "epoch": 1.5615615615615615, "grad_norm": 0.028117358684539795, "learning_rate": 1.7767191383595693e-05, "loss": 0.0196, "step": 1040 }, { "epoch": 1.5765765765765765, "grad_norm": 1.8692775964736938, "learning_rate": 1.7725766362883185e-05, "loss": 0.0091, "step": 1050 }, { "epoch": 1.5915915915915915, "grad_norm": 3.6717379093170166, "learning_rate": 1.7684341342170673e-05, "loss": 0.0081, "step": 1060 }, { "epoch": 1.6066066066066065, "grad_norm": 0.24632543325424194, "learning_rate": 1.7642916321458162e-05, "loss": 0.0211, "step": 1070 }, { "epoch": 1.6216216216216215, "grad_norm": 0.46948179602622986, "learning_rate": 1.760149130074565e-05, "loss": 0.0237, "step": 1080 }, { "epoch": 1.6366366366366365, "grad_norm": 5.49431037902832, "learning_rate": 1.756006628003314e-05, "loss": 0.0292, "step": 1090 }, { "epoch": 1.6516516516516515, "grad_norm": 1.7562233209609985, "learning_rate": 1.7518641259320632e-05, "loss": 0.0107, "step": 1100 }, { "epoch": 1.6666666666666665, "grad_norm": 4.193264961242676, "learning_rate": 1.747721623860812e-05, "loss": 0.0088, "step": 1110 }, { "epoch": 1.6816816816816815, "grad_norm": 2.0336873531341553, "learning_rate": 1.743579121789561e-05, "loss": 0.0159, "step": 1120 }, { "epoch": 1.6966966966966965, "grad_norm": 0.03742986172437668, "learning_rate": 1.7394366197183098e-05, "loss": 0.0227, "step": 1130 }, { "epoch": 1.7117117117117115, "grad_norm": 2.9706671237945557, "learning_rate": 1.735294117647059e-05, "loss": 0.0133, "step": 1140 }, { "epoch": 1.7267267267267268, "grad_norm": 0.054504282772541046, "learning_rate": 1.731151615575808e-05, "loss": 0.018, "step": 1150 }, { "epoch": 1.7417417417417418, "grad_norm": 0.3289143145084381, "learning_rate": 1.7270091135045568e-05, "loss": 0.0164, "step": 1160 }, { "epoch": 1.7567567567567568, "grad_norm": 5.590298652648926, "learning_rate": 1.722866611433306e-05, "loss": 0.0266, "step": 1170 }, { "epoch": 1.7717717717717718, "grad_norm": 3.789137601852417, "learning_rate": 1.718724109362055e-05, "loss": 0.0325, "step": 1180 }, { "epoch": 1.7867867867867868, "grad_norm": 1.8751722574234009, "learning_rate": 1.7145816072908037e-05, "loss": 0.0303, "step": 1190 }, { "epoch": 1.8018018018018018, "grad_norm": 0.3353935778141022, "learning_rate": 1.710439105219553e-05, "loss": 0.0328, "step": 1200 }, { "epoch": 1.8168168168168168, "grad_norm": 2.441638231277466, "learning_rate": 1.7062966031483018e-05, "loss": 0.0105, "step": 1210 }, { "epoch": 1.8318318318318318, "grad_norm": 1.8479589223861694, "learning_rate": 1.7021541010770507e-05, "loss": 0.0429, "step": 1220 }, { "epoch": 1.8468468468468469, "grad_norm": 0.5461903214454651, "learning_rate": 1.6980115990057996e-05, "loss": 0.0208, "step": 1230 }, { "epoch": 1.8618618618618619, "grad_norm": 0.17233674228191376, "learning_rate": 1.6938690969345484e-05, "loss": 0.0279, "step": 1240 }, { "epoch": 1.8768768768768769, "grad_norm": 1.572726845741272, "learning_rate": 1.6897265948632976e-05, "loss": 0.0085, "step": 1250 }, { "epoch": 1.8918918918918919, "grad_norm": 13.602294921875, "learning_rate": 1.6855840927920465e-05, "loss": 0.0304, "step": 1260 }, { "epoch": 1.906906906906907, "grad_norm": 0.2937522530555725, "learning_rate": 1.6814415907207954e-05, "loss": 0.0182, "step": 1270 }, { "epoch": 1.921921921921922, "grad_norm": 8.252511024475098, "learning_rate": 1.6772990886495443e-05, "loss": 0.0191, "step": 1280 }, { "epoch": 1.936936936936937, "grad_norm": 3.633695602416992, "learning_rate": 1.6731565865782935e-05, "loss": 0.0209, "step": 1290 }, { "epoch": 1.951951951951952, "grad_norm": 2.4778058528900146, "learning_rate": 1.6690140845070424e-05, "loss": 0.0283, "step": 1300 }, { "epoch": 1.966966966966967, "grad_norm": 4.573657989501953, "learning_rate": 1.6648715824357912e-05, "loss": 0.0242, "step": 1310 }, { "epoch": 1.981981981981982, "grad_norm": 10.803725242614746, "learning_rate": 1.6607290803645404e-05, "loss": 0.0154, "step": 1320 }, { "epoch": 1.996996996996997, "grad_norm": 1.1839510202407837, "learning_rate": 1.6565865782932893e-05, "loss": 0.0164, "step": 1330 }, { "epoch": 2.0, "eval_accuracy": 0.844590780809031, "eval_f1": 0.9083219840871622, "eval_loss": 0.21870091557502747, "eval_roc_auc": 0.9382724799051716, "eval_runtime": 7.4074, "eval_samples_per_second": 717.526, "eval_steps_per_second": 22.545, "step": 1332 }, { "epoch": 2.012012012012012, "grad_norm": 2.07466197013855, "learning_rate": 1.6524440762220382e-05, "loss": 0.0341, "step": 1340 }, { "epoch": 2.027027027027027, "grad_norm": 0.3642808794975281, "learning_rate": 1.6483015741507874e-05, "loss": 0.0068, "step": 1350 }, { "epoch": 2.042042042042042, "grad_norm": 3.976308822631836, "learning_rate": 1.6441590720795363e-05, "loss": 0.0232, "step": 1360 }, { "epoch": 2.057057057057057, "grad_norm": 0.7623342275619507, "learning_rate": 1.640016570008285e-05, "loss": 0.0055, "step": 1370 }, { "epoch": 2.0720720720720722, "grad_norm": 2.239234685897827, "learning_rate": 1.635874067937034e-05, "loss": 0.0124, "step": 1380 }, { "epoch": 2.0870870870870872, "grad_norm": 5.9343061447143555, "learning_rate": 1.6317315658657832e-05, "loss": 0.0116, "step": 1390 }, { "epoch": 2.1021021021021022, "grad_norm": 2.6489689350128174, "learning_rate": 1.627589063794532e-05, "loss": 0.0126, "step": 1400 }, { "epoch": 2.1171171171171173, "grad_norm": 1.19527006149292, "learning_rate": 1.623446561723281e-05, "loss": 0.0209, "step": 1410 }, { "epoch": 2.1321321321321323, "grad_norm": 2.1353812217712402, "learning_rate": 1.61930405965203e-05, "loss": 0.0146, "step": 1420 }, { "epoch": 2.1471471471471473, "grad_norm": 0.1424659639596939, "learning_rate": 1.6151615575807787e-05, "loss": 0.0266, "step": 1430 }, { "epoch": 2.1621621621621623, "grad_norm": 4.604438304901123, "learning_rate": 1.611019055509528e-05, "loss": 0.0131, "step": 1440 }, { "epoch": 2.1771771771771773, "grad_norm": 0.18747694790363312, "learning_rate": 1.6068765534382768e-05, "loss": 0.0166, "step": 1450 }, { "epoch": 2.1921921921921923, "grad_norm": 0.10579628497362137, "learning_rate": 1.6027340513670257e-05, "loss": 0.0141, "step": 1460 }, { "epoch": 2.2072072072072073, "grad_norm": 0.06953544914722443, "learning_rate": 1.598591549295775e-05, "loss": 0.0066, "step": 1470 }, { "epoch": 2.2222222222222223, "grad_norm": 1.7048633098602295, "learning_rate": 1.5944490472245238e-05, "loss": 0.026, "step": 1480 }, { "epoch": 2.2372372372372373, "grad_norm": 1.1399219036102295, "learning_rate": 1.5903065451532727e-05, "loss": 0.0178, "step": 1490 }, { "epoch": 2.2522522522522523, "grad_norm": 0.9059074521064758, "learning_rate": 1.586164043082022e-05, "loss": 0.0122, "step": 1500 }, { "epoch": 2.2672672672672673, "grad_norm": 0.6121827960014343, "learning_rate": 1.5820215410107707e-05, "loss": 0.007, "step": 1510 }, { "epoch": 2.2822822822822824, "grad_norm": 4.591891288757324, "learning_rate": 1.5778790389395196e-05, "loss": 0.0189, "step": 1520 }, { "epoch": 2.2972972972972974, "grad_norm": 6.620510101318359, "learning_rate": 1.5737365368682688e-05, "loss": 0.0214, "step": 1530 }, { "epoch": 2.3123123123123124, "grad_norm": 3.0007002353668213, "learning_rate": 1.5695940347970177e-05, "loss": 0.0368, "step": 1540 }, { "epoch": 2.3273273273273274, "grad_norm": 4.294153213500977, "learning_rate": 1.5654515327257666e-05, "loss": 0.0134, "step": 1550 }, { "epoch": 2.3423423423423424, "grad_norm": 12.665278434753418, "learning_rate": 1.5613090306545154e-05, "loss": 0.0094, "step": 1560 }, { "epoch": 2.3573573573573574, "grad_norm": 0.5548607707023621, "learning_rate": 1.5571665285832643e-05, "loss": 0.0145, "step": 1570 }, { "epoch": 2.3723723723723724, "grad_norm": 5.412700176239014, "learning_rate": 1.5530240265120132e-05, "loss": 0.0199, "step": 1580 }, { "epoch": 2.3873873873873874, "grad_norm": 2.9422519207000732, "learning_rate": 1.5488815244407624e-05, "loss": 0.0122, "step": 1590 }, { "epoch": 2.4024024024024024, "grad_norm": 4.299681186676025, "learning_rate": 1.5447390223695113e-05, "loss": 0.0287, "step": 1600 }, { "epoch": 2.4174174174174174, "grad_norm": 3.635817766189575, "learning_rate": 1.54059652029826e-05, "loss": 0.0145, "step": 1610 }, { "epoch": 2.4324324324324325, "grad_norm": 0.06890556961297989, "learning_rate": 1.5364540182270094e-05, "loss": 0.0155, "step": 1620 }, { "epoch": 2.4474474474474475, "grad_norm": 1.9247612953186035, "learning_rate": 1.5323115161557582e-05, "loss": 0.0208, "step": 1630 }, { "epoch": 2.4624624624624625, "grad_norm": 0.0777619257569313, "learning_rate": 1.528169014084507e-05, "loss": 0.0194, "step": 1640 }, { "epoch": 2.4774774774774775, "grad_norm": 0.09070724248886108, "learning_rate": 1.524026512013256e-05, "loss": 0.0106, "step": 1650 }, { "epoch": 2.4924924924924925, "grad_norm": 3.718015670776367, "learning_rate": 1.5198840099420052e-05, "loss": 0.0186, "step": 1660 }, { "epoch": 2.5075075075075075, "grad_norm": 0.07736914604902267, "learning_rate": 1.515741507870754e-05, "loss": 0.0061, "step": 1670 }, { "epoch": 2.5225225225225225, "grad_norm": 7.326141834259033, "learning_rate": 1.511599005799503e-05, "loss": 0.0205, "step": 1680 }, { "epoch": 2.5375375375375375, "grad_norm": 0.2035406529903412, "learning_rate": 1.507456503728252e-05, "loss": 0.0055, "step": 1690 }, { "epoch": 2.5525525525525525, "grad_norm": 0.1642528474330902, "learning_rate": 1.5033140016570009e-05, "loss": 0.0247, "step": 1700 }, { "epoch": 2.5675675675675675, "grad_norm": 0.04008019343018532, "learning_rate": 1.4991714995857497e-05, "loss": 0.0134, "step": 1710 }, { "epoch": 2.5825825825825826, "grad_norm": 2.158572196960449, "learning_rate": 1.495028997514499e-05, "loss": 0.0139, "step": 1720 }, { "epoch": 2.5975975975975976, "grad_norm": 6.9125494956970215, "learning_rate": 1.4908864954432478e-05, "loss": 0.0141, "step": 1730 }, { "epoch": 2.6126126126126126, "grad_norm": 0.05439571291208267, "learning_rate": 1.4867439933719967e-05, "loss": 0.0172, "step": 1740 }, { "epoch": 2.6276276276276276, "grad_norm": 1.7265421152114868, "learning_rate": 1.4826014913007457e-05, "loss": 0.029, "step": 1750 }, { "epoch": 2.6426426426426426, "grad_norm": 0.15565641224384308, "learning_rate": 1.4784589892294948e-05, "loss": 0.0164, "step": 1760 }, { "epoch": 2.6576576576576576, "grad_norm": 1.2956058979034424, "learning_rate": 1.4743164871582437e-05, "loss": 0.0225, "step": 1770 }, { "epoch": 2.6726726726726726, "grad_norm": 0.5900393724441528, "learning_rate": 1.4701739850869927e-05, "loss": 0.0128, "step": 1780 }, { "epoch": 2.6876876876876876, "grad_norm": 2.6349215507507324, "learning_rate": 1.4660314830157416e-05, "loss": 0.0048, "step": 1790 }, { "epoch": 2.7027027027027026, "grad_norm": 4.905094146728516, "learning_rate": 1.4618889809444905e-05, "loss": 0.0148, "step": 1800 }, { "epoch": 2.7177177177177176, "grad_norm": 0.15017619729042053, "learning_rate": 1.4577464788732397e-05, "loss": 0.022, "step": 1810 }, { "epoch": 2.7327327327327327, "grad_norm": 0.6810951828956604, "learning_rate": 1.4536039768019885e-05, "loss": 0.0348, "step": 1820 }, { "epoch": 2.7477477477477477, "grad_norm": 0.09275978058576584, "learning_rate": 1.4494614747307374e-05, "loss": 0.0103, "step": 1830 }, { "epoch": 2.7627627627627627, "grad_norm": 0.9358335733413696, "learning_rate": 1.4453189726594865e-05, "loss": 0.007, "step": 1840 }, { "epoch": 2.7777777777777777, "grad_norm": 0.21646398305892944, "learning_rate": 1.4411764705882353e-05, "loss": 0.0111, "step": 1850 }, { "epoch": 2.7927927927927927, "grad_norm": 3.0269370079040527, "learning_rate": 1.4370339685169842e-05, "loss": 0.0064, "step": 1860 }, { "epoch": 2.8078078078078077, "grad_norm": 0.03600875288248062, "learning_rate": 1.4328914664457334e-05, "loss": 0.0173, "step": 1870 }, { "epoch": 2.8228228228228227, "grad_norm": 0.03154575824737549, "learning_rate": 1.4287489643744823e-05, "loss": 0.0057, "step": 1880 }, { "epoch": 2.8378378378378377, "grad_norm": 2.526102304458618, "learning_rate": 1.4246064623032312e-05, "loss": 0.004, "step": 1890 }, { "epoch": 2.8528528528528527, "grad_norm": 1.8955109119415283, "learning_rate": 1.4204639602319802e-05, "loss": 0.0161, "step": 1900 }, { "epoch": 2.8678678678678677, "grad_norm": 0.41396257281303406, "learning_rate": 1.4163214581607293e-05, "loss": 0.0143, "step": 1910 }, { "epoch": 2.8828828828828827, "grad_norm": 0.08320903033018112, "learning_rate": 1.4121789560894781e-05, "loss": 0.024, "step": 1920 }, { "epoch": 2.8978978978978978, "grad_norm": 0.22456814348697662, "learning_rate": 1.4080364540182272e-05, "loss": 0.0111, "step": 1930 }, { "epoch": 2.9129129129129128, "grad_norm": 0.3686913847923279, "learning_rate": 1.403893951946976e-05, "loss": 0.0244, "step": 1940 }, { "epoch": 2.9279279279279278, "grad_norm": 0.21892797946929932, "learning_rate": 1.399751449875725e-05, "loss": 0.0176, "step": 1950 }, { "epoch": 2.942942942942943, "grad_norm": 3.7241322994232178, "learning_rate": 1.3956089478044741e-05, "loss": 0.0103, "step": 1960 }, { "epoch": 2.957957957957958, "grad_norm": 2.0452113151550293, "learning_rate": 1.391466445733223e-05, "loss": 0.039, "step": 1970 }, { "epoch": 2.972972972972973, "grad_norm": 2.3977973461151123, "learning_rate": 1.3873239436619719e-05, "loss": 0.0106, "step": 1980 }, { "epoch": 2.987987987987988, "grad_norm": 0.21678972244262695, "learning_rate": 1.383181441590721e-05, "loss": 0.0173, "step": 1990 }, { "epoch": 3.0, "eval_accuracy": 0.8553151458137347, "eval_f1": 0.9143430578453158, "eval_loss": 0.21451762318611145, "eval_roc_auc": 0.9423584374850139, "eval_runtime": 7.3964, "eval_samples_per_second": 718.591, "eval_steps_per_second": 22.579, "step": 1998 }, { "epoch": 3.003003003003003, "grad_norm": 1.4772316217422485, "learning_rate": 1.3790389395194698e-05, "loss": 0.0103, "step": 2000 }, { "epoch": 3.018018018018018, "grad_norm": 0.7159767746925354, "learning_rate": 1.3748964374482188e-05, "loss": 0.0107, "step": 2010 }, { "epoch": 3.033033033033033, "grad_norm": 0.16911160945892334, "learning_rate": 1.3707539353769679e-05, "loss": 0.0063, "step": 2020 }, { "epoch": 3.048048048048048, "grad_norm": 0.11155136674642563, "learning_rate": 1.3666114333057168e-05, "loss": 0.0337, "step": 2030 }, { "epoch": 3.063063063063063, "grad_norm": 0.32412320375442505, "learning_rate": 1.3624689312344656e-05, "loss": 0.0238, "step": 2040 }, { "epoch": 3.078078078078078, "grad_norm": 0.03618387132883072, "learning_rate": 1.3583264291632149e-05, "loss": 0.0162, "step": 2050 }, { "epoch": 3.093093093093093, "grad_norm": 0.7066205739974976, "learning_rate": 1.3541839270919637e-05, "loss": 0.0103, "step": 2060 }, { "epoch": 3.108108108108108, "grad_norm": 0.11609063297510147, "learning_rate": 1.3500414250207126e-05, "loss": 0.0105, "step": 2070 }, { "epoch": 3.123123123123123, "grad_norm": 0.32982802391052246, "learning_rate": 1.3458989229494616e-05, "loss": 0.0102, "step": 2080 }, { "epoch": 3.138138138138138, "grad_norm": 3.4004058837890625, "learning_rate": 1.3417564208782105e-05, "loss": 0.0159, "step": 2090 }, { "epoch": 3.153153153153153, "grad_norm": 6.178320407867432, "learning_rate": 1.3376139188069594e-05, "loss": 0.0232, "step": 2100 }, { "epoch": 3.1681681681681684, "grad_norm": 1.1276676654815674, "learning_rate": 1.3334714167357086e-05, "loss": 0.0069, "step": 2110 }, { "epoch": 3.1831831831831834, "grad_norm": 0.02346782386302948, "learning_rate": 1.3293289146644575e-05, "loss": 0.0059, "step": 2120 }, { "epoch": 3.1981981981981984, "grad_norm": 4.738574028015137, "learning_rate": 1.3251864125932064e-05, "loss": 0.0202, "step": 2130 }, { "epoch": 3.2132132132132134, "grad_norm": 6.053224086761475, "learning_rate": 1.3210439105219554e-05, "loss": 0.0198, "step": 2140 }, { "epoch": 3.2282282282282284, "grad_norm": 3.8819286823272705, "learning_rate": 1.3169014084507044e-05, "loss": 0.0073, "step": 2150 }, { "epoch": 3.2432432432432434, "grad_norm": 2.7025258541107178, "learning_rate": 1.3127589063794533e-05, "loss": 0.0213, "step": 2160 }, { "epoch": 3.2582582582582584, "grad_norm": 3.3367843627929688, "learning_rate": 1.3086164043082024e-05, "loss": 0.01, "step": 2170 }, { "epoch": 3.2732732732732734, "grad_norm": 0.6221895217895508, "learning_rate": 1.3044739022369512e-05, "loss": 0.0046, "step": 2180 }, { "epoch": 3.2882882882882885, "grad_norm": 4.412055969238281, "learning_rate": 1.3003314001657001e-05, "loss": 0.0207, "step": 2190 }, { "epoch": 3.3033033033033035, "grad_norm": 5.719205379486084, "learning_rate": 1.2961888980944493e-05, "loss": 0.01, "step": 2200 }, { "epoch": 3.3183183183183185, "grad_norm": 1.341784119606018, "learning_rate": 1.2920463960231982e-05, "loss": 0.0079, "step": 2210 }, { "epoch": 3.3333333333333335, "grad_norm": 0.6478399038314819, "learning_rate": 1.287903893951947e-05, "loss": 0.0104, "step": 2220 }, { "epoch": 3.3483483483483485, "grad_norm": 1.90777587890625, "learning_rate": 1.2837613918806961e-05, "loss": 0.0063, "step": 2230 }, { "epoch": 3.3633633633633635, "grad_norm": 0.18736779689788818, "learning_rate": 1.279618889809445e-05, "loss": 0.0142, "step": 2240 }, { "epoch": 3.3783783783783785, "grad_norm": 1.921953797340393, "learning_rate": 1.2754763877381939e-05, "loss": 0.0056, "step": 2250 }, { "epoch": 3.3933933933933935, "grad_norm": 0.2626243531703949, "learning_rate": 1.271333885666943e-05, "loss": 0.0135, "step": 2260 }, { "epoch": 3.4084084084084085, "grad_norm": 2.2788760662078857, "learning_rate": 1.267191383595692e-05, "loss": 0.0085, "step": 2270 }, { "epoch": 3.4234234234234235, "grad_norm": 0.39181146025657654, "learning_rate": 1.2630488815244408e-05, "loss": 0.0148, "step": 2280 }, { "epoch": 3.4384384384384385, "grad_norm": 8.13824462890625, "learning_rate": 1.2589063794531897e-05, "loss": 0.0091, "step": 2290 }, { "epoch": 3.4534534534534536, "grad_norm": 5.898563385009766, "learning_rate": 1.2547638773819389e-05, "loss": 0.0172, "step": 2300 }, { "epoch": 3.4684684684684686, "grad_norm": 0.0695483535528183, "learning_rate": 1.2506213753106878e-05, "loss": 0.008, "step": 2310 }, { "epoch": 3.4834834834834836, "grad_norm": 0.5626207590103149, "learning_rate": 1.2464788732394367e-05, "loss": 0.0191, "step": 2320 }, { "epoch": 3.4984984984984986, "grad_norm": 0.9743795990943909, "learning_rate": 1.2423363711681857e-05, "loss": 0.0064, "step": 2330 }, { "epoch": 3.5135135135135136, "grad_norm": 0.03379818797111511, "learning_rate": 1.2381938690969346e-05, "loss": 0.0063, "step": 2340 }, { "epoch": 3.5285285285285286, "grad_norm": 2.3289623260498047, "learning_rate": 1.2340513670256834e-05, "loss": 0.0074, "step": 2350 }, { "epoch": 3.5435435435435436, "grad_norm": 0.7672750949859619, "learning_rate": 1.2299088649544327e-05, "loss": 0.0137, "step": 2360 }, { "epoch": 3.5585585585585586, "grad_norm": 0.019895190373063087, "learning_rate": 1.2257663628831815e-05, "loss": 0.0117, "step": 2370 }, { "epoch": 3.5735735735735736, "grad_norm": 2.079782485961914, "learning_rate": 1.2216238608119304e-05, "loss": 0.0097, "step": 2380 }, { "epoch": 3.5885885885885886, "grad_norm": 7.858700752258301, "learning_rate": 1.2174813587406794e-05, "loss": 0.0148, "step": 2390 }, { "epoch": 3.6036036036036037, "grad_norm": 1.9548249244689941, "learning_rate": 1.2133388566694285e-05, "loss": 0.0132, "step": 2400 }, { "epoch": 3.6186186186186187, "grad_norm": 3.6328041553497314, "learning_rate": 1.2091963545981774e-05, "loss": 0.0102, "step": 2410 }, { "epoch": 3.6336336336336337, "grad_norm": 0.6240374445915222, "learning_rate": 1.2050538525269264e-05, "loss": 0.0082, "step": 2420 }, { "epoch": 3.6486486486486487, "grad_norm": 0.9374043941497803, "learning_rate": 1.2009113504556753e-05, "loss": 0.0129, "step": 2430 }, { "epoch": 3.6636636636636637, "grad_norm": 1.3998479843139648, "learning_rate": 1.1967688483844242e-05, "loss": 0.0112, "step": 2440 }, { "epoch": 3.6786786786786787, "grad_norm": 6.0802507400512695, "learning_rate": 1.1926263463131734e-05, "loss": 0.0147, "step": 2450 }, { "epoch": 3.6936936936936937, "grad_norm": 1.1334019899368286, "learning_rate": 1.1884838442419222e-05, "loss": 0.011, "step": 2460 }, { "epoch": 3.7087087087087087, "grad_norm": 2.0716891288757324, "learning_rate": 1.1843413421706711e-05, "loss": 0.019, "step": 2470 }, { "epoch": 3.7237237237237237, "grad_norm": 0.058690622448921204, "learning_rate": 1.1801988400994202e-05, "loss": 0.0142, "step": 2480 }, { "epoch": 3.7387387387387387, "grad_norm": 0.4929686486721039, "learning_rate": 1.176056338028169e-05, "loss": 0.0098, "step": 2490 }, { "epoch": 3.7537537537537538, "grad_norm": 0.36563634872436523, "learning_rate": 1.1719138359569179e-05, "loss": 0.0074, "step": 2500 }, { "epoch": 3.7687687687687688, "grad_norm": 0.20172318816184998, "learning_rate": 1.1677713338856671e-05, "loss": 0.0076, "step": 2510 }, { "epoch": 3.7837837837837838, "grad_norm": 1.6785756349563599, "learning_rate": 1.163628831814416e-05, "loss": 0.0087, "step": 2520 }, { "epoch": 3.798798798798799, "grad_norm": 0.06056467816233635, "learning_rate": 1.1594863297431649e-05, "loss": 0.0071, "step": 2530 }, { "epoch": 3.813813813813814, "grad_norm": 0.4795999526977539, "learning_rate": 1.155343827671914e-05, "loss": 0.0068, "step": 2540 }, { "epoch": 3.828828828828829, "grad_norm": 1.2181782722473145, "learning_rate": 1.151201325600663e-05, "loss": 0.0088, "step": 2550 }, { "epoch": 3.843843843843844, "grad_norm": 0.036575447767972946, "learning_rate": 1.1470588235294118e-05, "loss": 0.009, "step": 2560 }, { "epoch": 3.858858858858859, "grad_norm": 0.7162961363792419, "learning_rate": 1.1429163214581609e-05, "loss": 0.011, "step": 2570 }, { "epoch": 3.873873873873874, "grad_norm": 0.8698685765266418, "learning_rate": 1.1387738193869098e-05, "loss": 0.0063, "step": 2580 }, { "epoch": 3.888888888888889, "grad_norm": 0.37558797001838684, "learning_rate": 1.1346313173156586e-05, "loss": 0.0064, "step": 2590 }, { "epoch": 3.903903903903904, "grad_norm": 5.980177402496338, "learning_rate": 1.1304888152444078e-05, "loss": 0.0253, "step": 2600 }, { "epoch": 3.918918918918919, "grad_norm": 0.07595256716012955, "learning_rate": 1.1263463131731567e-05, "loss": 0.0122, "step": 2610 }, { "epoch": 3.933933933933934, "grad_norm": 3.3737852573394775, "learning_rate": 1.1222038111019056e-05, "loss": 0.0122, "step": 2620 }, { "epoch": 3.948948948948949, "grad_norm": 0.67271488904953, "learning_rate": 1.1180613090306546e-05, "loss": 0.0112, "step": 2630 }, { "epoch": 3.963963963963964, "grad_norm": 5.097310543060303, "learning_rate": 1.1139188069594035e-05, "loss": 0.0117, "step": 2640 }, { "epoch": 3.978978978978979, "grad_norm": 0.0951702669262886, "learning_rate": 1.1097763048881525e-05, "loss": 0.0038, "step": 2650 }, { "epoch": 3.993993993993994, "grad_norm": 2.21738862991333, "learning_rate": 1.1056338028169016e-05, "loss": 0.0095, "step": 2660 }, { "epoch": 4.0, "eval_accuracy": 0.8568203198494826, "eval_f1": 0.9156453809216656, "eval_loss": 0.21418116986751556, "eval_roc_auc": 0.9432802901372163, "eval_runtime": 7.387, "eval_samples_per_second": 719.51, "eval_steps_per_second": 22.607, "step": 2664 }, { "epoch": 4.009009009009009, "grad_norm": 1.2217696905136108, "learning_rate": 1.1014913007456505e-05, "loss": 0.0082, "step": 2670 }, { "epoch": 4.024024024024024, "grad_norm": 0.018607912585139275, "learning_rate": 1.0973487986743993e-05, "loss": 0.0143, "step": 2680 }, { "epoch": 4.039039039039039, "grad_norm": 0.2839476764202118, "learning_rate": 1.0932062966031486e-05, "loss": 0.0081, "step": 2690 }, { "epoch": 4.054054054054054, "grad_norm": 0.2138938009738922, "learning_rate": 1.0890637945318974e-05, "loss": 0.0108, "step": 2700 }, { "epoch": 4.069069069069069, "grad_norm": 1.584167718887329, "learning_rate": 1.0849212924606463e-05, "loss": 0.0131, "step": 2710 }, { "epoch": 4.084084084084084, "grad_norm": 0.10289608687162399, "learning_rate": 1.0807787903893953e-05, "loss": 0.0078, "step": 2720 }, { "epoch": 4.099099099099099, "grad_norm": 0.9854161739349365, "learning_rate": 1.0766362883181442e-05, "loss": 0.0073, "step": 2730 }, { "epoch": 4.114114114114114, "grad_norm": 0.760006844997406, "learning_rate": 1.0724937862468931e-05, "loss": 0.0085, "step": 2740 }, { "epoch": 4.129129129129129, "grad_norm": 2.729097604751587, "learning_rate": 1.0683512841756423e-05, "loss": 0.0086, "step": 2750 }, { "epoch": 4.1441441441441444, "grad_norm": 0.011590493842959404, "learning_rate": 1.0642087821043912e-05, "loss": 0.0182, "step": 2760 }, { "epoch": 4.1591591591591595, "grad_norm": 1.401490569114685, "learning_rate": 1.06006628003314e-05, "loss": 0.0052, "step": 2770 }, { "epoch": 4.1741741741741745, "grad_norm": 9.527831077575684, "learning_rate": 1.0559237779618891e-05, "loss": 0.0058, "step": 2780 }, { "epoch": 4.1891891891891895, "grad_norm": 0.16799849271774292, "learning_rate": 1.051781275890638e-05, "loss": 0.0019, "step": 2790 }, { "epoch": 4.2042042042042045, "grad_norm": 0.15852217376232147, "learning_rate": 1.047638773819387e-05, "loss": 0.0215, "step": 2800 }, { "epoch": 4.2192192192192195, "grad_norm": 0.03776758536696434, "learning_rate": 1.043496271748136e-05, "loss": 0.0087, "step": 2810 }, { "epoch": 4.2342342342342345, "grad_norm": 0.038916442543268204, "learning_rate": 1.039353769676885e-05, "loss": 0.0048, "step": 2820 }, { "epoch": 4.2492492492492495, "grad_norm": 0.1437731832265854, "learning_rate": 1.0352112676056338e-05, "loss": 0.007, "step": 2830 }, { "epoch": 4.2642642642642645, "grad_norm": 0.5450798273086548, "learning_rate": 1.031068765534383e-05, "loss": 0.0061, "step": 2840 }, { "epoch": 4.2792792792792795, "grad_norm": 0.19095036387443542, "learning_rate": 1.0269262634631319e-05, "loss": 0.003, "step": 2850 }, { "epoch": 4.2942942942942945, "grad_norm": 0.08140315115451813, "learning_rate": 1.0227837613918808e-05, "loss": 0.0088, "step": 2860 }, { "epoch": 4.3093093093093096, "grad_norm": 6.2226667404174805, "learning_rate": 1.0186412593206298e-05, "loss": 0.0038, "step": 2870 }, { "epoch": 4.324324324324325, "grad_norm": 0.4789693355560303, "learning_rate": 1.0144987572493787e-05, "loss": 0.0087, "step": 2880 }, { "epoch": 4.33933933933934, "grad_norm": 0.22297614812850952, "learning_rate": 1.0103562551781276e-05, "loss": 0.0104, "step": 2890 }, { "epoch": 4.354354354354355, "grad_norm": 3.0525765419006348, "learning_rate": 1.0062137531068768e-05, "loss": 0.0095, "step": 2900 }, { "epoch": 4.36936936936937, "grad_norm": 0.9432957768440247, "learning_rate": 1.0020712510356256e-05, "loss": 0.0063, "step": 2910 }, { "epoch": 4.384384384384385, "grad_norm": 0.20507968962192535, "learning_rate": 9.979287489643745e-06, "loss": 0.0081, "step": 2920 }, { "epoch": 4.3993993993994, "grad_norm": 0.13641659915447235, "learning_rate": 9.937862468931236e-06, "loss": 0.0162, "step": 2930 }, { "epoch": 4.414414414414415, "grad_norm": 0.16560953855514526, "learning_rate": 9.896437448218726e-06, "loss": 0.0084, "step": 2940 }, { "epoch": 4.42942942942943, "grad_norm": 0.11404027044773102, "learning_rate": 9.855012427506215e-06, "loss": 0.0104, "step": 2950 }, { "epoch": 4.444444444444445, "grad_norm": 0.016587508842349052, "learning_rate": 9.813587406793704e-06, "loss": 0.0116, "step": 2960 }, { "epoch": 4.45945945945946, "grad_norm": 0.623416006565094, "learning_rate": 9.772162386081194e-06, "loss": 0.0105, "step": 2970 }, { "epoch": 4.474474474474475, "grad_norm": 0.0888320803642273, "learning_rate": 9.730737365368683e-06, "loss": 0.0063, "step": 2980 }, { "epoch": 4.48948948948949, "grad_norm": 1.374616026878357, "learning_rate": 9.689312344656173e-06, "loss": 0.0077, "step": 2990 }, { "epoch": 4.504504504504505, "grad_norm": 0.029000842943787575, "learning_rate": 9.647887323943664e-06, "loss": 0.0098, "step": 3000 }, { "epoch": 4.51951951951952, "grad_norm": 2.8579955101013184, "learning_rate": 9.606462303231152e-06, "loss": 0.0043, "step": 3010 }, { "epoch": 4.534534534534535, "grad_norm": 3.9038970470428467, "learning_rate": 9.565037282518643e-06, "loss": 0.0194, "step": 3020 }, { "epoch": 4.54954954954955, "grad_norm": 0.4741935431957245, "learning_rate": 9.523612261806132e-06, "loss": 0.0065, "step": 3030 }, { "epoch": 4.564564564564565, "grad_norm": 1.3021470308303833, "learning_rate": 9.482187241093622e-06, "loss": 0.0048, "step": 3040 }, { "epoch": 4.57957957957958, "grad_norm": 0.17900153994560242, "learning_rate": 9.44076222038111e-06, "loss": 0.0103, "step": 3050 }, { "epoch": 4.594594594594595, "grad_norm": 0.8498497605323792, "learning_rate": 9.399337199668601e-06, "loss": 0.0084, "step": 3060 }, { "epoch": 4.60960960960961, "grad_norm": 0.6373823285102844, "learning_rate": 9.35791217895609e-06, "loss": 0.0061, "step": 3070 }, { "epoch": 4.624624624624625, "grad_norm": 0.9934857487678528, "learning_rate": 9.31648715824358e-06, "loss": 0.0069, "step": 3080 }, { "epoch": 4.63963963963964, "grad_norm": 0.12153011560440063, "learning_rate": 9.275062137531069e-06, "loss": 0.0166, "step": 3090 }, { "epoch": 4.654654654654655, "grad_norm": 0.3886626660823822, "learning_rate": 9.23363711681856e-06, "loss": 0.0124, "step": 3100 }, { "epoch": 4.66966966966967, "grad_norm": 3.08005690574646, "learning_rate": 9.192212096106048e-06, "loss": 0.0126, "step": 3110 }, { "epoch": 4.684684684684685, "grad_norm": 2.155261278152466, "learning_rate": 9.150787075393539e-06, "loss": 0.0189, "step": 3120 }, { "epoch": 4.6996996996997, "grad_norm": 3.455352544784546, "learning_rate": 9.109362054681027e-06, "loss": 0.0043, "step": 3130 }, { "epoch": 4.714714714714715, "grad_norm": 0.19982695579528809, "learning_rate": 9.067937033968518e-06, "loss": 0.0144, "step": 3140 }, { "epoch": 4.72972972972973, "grad_norm": 0.28114157915115356, "learning_rate": 9.026512013256007e-06, "loss": 0.0081, "step": 3150 }, { "epoch": 4.744744744744745, "grad_norm": 3.1953513622283936, "learning_rate": 8.985086992543497e-06, "loss": 0.0125, "step": 3160 }, { "epoch": 4.75975975975976, "grad_norm": 9.547475814819336, "learning_rate": 8.943661971830987e-06, "loss": 0.0107, "step": 3170 }, { "epoch": 4.774774774774775, "grad_norm": 0.3070663809776306, "learning_rate": 8.902236951118476e-06, "loss": 0.0094, "step": 3180 }, { "epoch": 4.78978978978979, "grad_norm": 1.0773664712905884, "learning_rate": 8.860811930405967e-06, "loss": 0.0151, "step": 3190 }, { "epoch": 4.804804804804805, "grad_norm": 0.05860641971230507, "learning_rate": 8.819386909693455e-06, "loss": 0.0091, "step": 3200 }, { "epoch": 4.81981981981982, "grad_norm": 0.1941542774438858, "learning_rate": 8.777961888980944e-06, "loss": 0.0081, "step": 3210 }, { "epoch": 4.834834834834835, "grad_norm": 0.0734286680817604, "learning_rate": 8.736536868268435e-06, "loss": 0.0055, "step": 3220 }, { "epoch": 4.84984984984985, "grad_norm": 0.18101628124713898, "learning_rate": 8.695111847555925e-06, "loss": 0.0103, "step": 3230 }, { "epoch": 4.864864864864865, "grad_norm": 0.027951765805482864, "learning_rate": 8.653686826843414e-06, "loss": 0.0034, "step": 3240 }, { "epoch": 4.87987987987988, "grad_norm": 0.8830456137657166, "learning_rate": 8.612261806130904e-06, "loss": 0.0152, "step": 3250 }, { "epoch": 4.894894894894895, "grad_norm": 0.04218738153576851, "learning_rate": 8.570836785418395e-06, "loss": 0.0141, "step": 3260 }, { "epoch": 4.90990990990991, "grad_norm": 0.07865418493747711, "learning_rate": 8.529411764705883e-06, "loss": 0.0039, "step": 3270 }, { "epoch": 4.924924924924925, "grad_norm": 2.8771729469299316, "learning_rate": 8.487986743993372e-06, "loss": 0.0092, "step": 3280 }, { "epoch": 4.93993993993994, "grad_norm": 0.7154805064201355, "learning_rate": 8.446561723280862e-06, "loss": 0.0191, "step": 3290 }, { "epoch": 4.954954954954955, "grad_norm": 1.2003430128097534, "learning_rate": 8.405136702568351e-06, "loss": 0.0165, "step": 3300 }, { "epoch": 4.96996996996997, "grad_norm": 0.19323813915252686, "learning_rate": 8.363711681855842e-06, "loss": 0.0157, "step": 3310 }, { "epoch": 4.984984984984985, "grad_norm": 0.06954075396060944, "learning_rate": 8.322286661143332e-06, "loss": 0.0116, "step": 3320 }, { "epoch": 5.0, "grad_norm": 0.9249535202980042, "learning_rate": 8.28086164043082e-06, "loss": 0.0151, "step": 3330 }, { "epoch": 5.0, "eval_accuracy": 0.8613358419567263, "eval_f1": 0.9182426116344287, "eval_loss": 0.21194002032279968, "eval_roc_auc": 0.9441613070085622, "eval_runtime": 7.442, "eval_samples_per_second": 714.187, "eval_steps_per_second": 22.44, "step": 3330 }, { "epoch": 5.015015015015015, "grad_norm": 1.1550884246826172, "learning_rate": 8.239436619718311e-06, "loss": 0.0097, "step": 3340 }, { "epoch": 5.03003003003003, "grad_norm": 3.4667506217956543, "learning_rate": 8.1980115990058e-06, "loss": 0.0076, "step": 3350 }, { "epoch": 5.045045045045045, "grad_norm": 0.022501222789287567, "learning_rate": 8.15658657829329e-06, "loss": 0.0174, "step": 3360 }, { "epoch": 5.06006006006006, "grad_norm": 0.2297457605600357, "learning_rate": 8.11516155758078e-06, "loss": 0.0121, "step": 3370 }, { "epoch": 5.075075075075075, "grad_norm": 1.9849693775177002, "learning_rate": 8.07373653686827e-06, "loss": 0.0069, "step": 3380 }, { "epoch": 5.09009009009009, "grad_norm": 3.121706962585449, "learning_rate": 8.032311516155758e-06, "loss": 0.0069, "step": 3390 }, { "epoch": 5.105105105105105, "grad_norm": 0.28827306628227234, "learning_rate": 7.990886495443249e-06, "loss": 0.0037, "step": 3400 }, { "epoch": 5.12012012012012, "grad_norm": 0.26308292150497437, "learning_rate": 7.949461474730738e-06, "loss": 0.003, "step": 3410 }, { "epoch": 5.135135135135135, "grad_norm": 1.9140377044677734, "learning_rate": 7.908036454018228e-06, "loss": 0.0051, "step": 3420 }, { "epoch": 5.15015015015015, "grad_norm": 0.019069673493504524, "learning_rate": 7.866611433305717e-06, "loss": 0.0087, "step": 3430 }, { "epoch": 5.165165165165165, "grad_norm": 0.09611427038908005, "learning_rate": 7.825186412593207e-06, "loss": 0.0042, "step": 3440 }, { "epoch": 5.18018018018018, "grad_norm": 0.04649114981293678, "learning_rate": 7.783761391880696e-06, "loss": 0.0075, "step": 3450 }, { "epoch": 5.195195195195195, "grad_norm": 0.7499408721923828, "learning_rate": 7.742336371168186e-06, "loss": 0.0064, "step": 3460 }, { "epoch": 5.21021021021021, "grad_norm": 0.08816402405500412, "learning_rate": 7.700911350455675e-06, "loss": 0.0052, "step": 3470 }, { "epoch": 5.225225225225225, "grad_norm": 0.6211869120597839, "learning_rate": 7.659486329743165e-06, "loss": 0.0123, "step": 3480 }, { "epoch": 5.24024024024024, "grad_norm": 2.1001930236816406, "learning_rate": 7.618061309030655e-06, "loss": 0.0053, "step": 3490 }, { "epoch": 5.255255255255255, "grad_norm": 5.740816593170166, "learning_rate": 7.576636288318145e-06, "loss": 0.0064, "step": 3500 }, { "epoch": 5.27027027027027, "grad_norm": 3.0299530029296875, "learning_rate": 7.535211267605634e-06, "loss": 0.007, "step": 3510 }, { "epoch": 5.285285285285285, "grad_norm": 0.03936946764588356, "learning_rate": 7.493786246893125e-06, "loss": 0.015, "step": 3520 }, { "epoch": 5.3003003003003, "grad_norm": 0.11476922035217285, "learning_rate": 7.4523612261806134e-06, "loss": 0.0218, "step": 3530 }, { "epoch": 5.315315315315315, "grad_norm": 0.18106217682361603, "learning_rate": 7.410936205468103e-06, "loss": 0.0034, "step": 3540 }, { "epoch": 5.33033033033033, "grad_norm": 3.0002665519714355, "learning_rate": 7.3695111847555935e-06, "loss": 0.0104, "step": 3550 }, { "epoch": 5.345345345345345, "grad_norm": 0.30738332867622375, "learning_rate": 7.328086164043082e-06, "loss": 0.0148, "step": 3560 }, { "epoch": 5.36036036036036, "grad_norm": 0.021894821897149086, "learning_rate": 7.286661143330573e-06, "loss": 0.0178, "step": 3570 }, { "epoch": 5.375375375375375, "grad_norm": 8.09498405456543, "learning_rate": 7.245236122618062e-06, "loss": 0.0114, "step": 3580 }, { "epoch": 5.39039039039039, "grad_norm": 0.6404626369476318, "learning_rate": 7.203811101905551e-06, "loss": 0.0145, "step": 3590 }, { "epoch": 5.405405405405405, "grad_norm": 0.25979307293891907, "learning_rate": 7.162386081193041e-06, "loss": 0.0134, "step": 3600 }, { "epoch": 5.42042042042042, "grad_norm": 4.571927547454834, "learning_rate": 7.120961060480531e-06, "loss": 0.0146, "step": 3610 }, { "epoch": 5.435435435435435, "grad_norm": 0.42414915561676025, "learning_rate": 7.0795360397680206e-06, "loss": 0.002, "step": 3620 }, { "epoch": 5.45045045045045, "grad_norm": 0.13032342493534088, "learning_rate": 7.03811101905551e-06, "loss": 0.0042, "step": 3630 }, { "epoch": 5.465465465465465, "grad_norm": 4.202887535095215, "learning_rate": 6.996685998343001e-06, "loss": 0.0068, "step": 3640 }, { "epoch": 5.48048048048048, "grad_norm": 1.762563943862915, "learning_rate": 6.955260977630489e-06, "loss": 0.0082, "step": 3650 }, { "epoch": 5.495495495495495, "grad_norm": 0.03661961853504181, "learning_rate": 6.913835956917979e-06, "loss": 0.0176, "step": 3660 }, { "epoch": 5.51051051051051, "grad_norm": 0.043978456407785416, "learning_rate": 6.872410936205469e-06, "loss": 0.0027, "step": 3670 }, { "epoch": 5.525525525525525, "grad_norm": 0.4701463282108307, "learning_rate": 6.830985915492958e-06, "loss": 0.0135, "step": 3680 }, { "epoch": 5.54054054054054, "grad_norm": 2.628462076187134, "learning_rate": 6.7895608947804485e-06, "loss": 0.0068, "step": 3690 }, { "epoch": 5.555555555555555, "grad_norm": 0.09407415241003036, "learning_rate": 6.748135874067938e-06, "loss": 0.0167, "step": 3700 }, { "epoch": 5.57057057057057, "grad_norm": 0.7712174654006958, "learning_rate": 6.706710853355427e-06, "loss": 0.0116, "step": 3710 }, { "epoch": 5.585585585585585, "grad_norm": 0.07994990795850754, "learning_rate": 6.665285832642917e-06, "loss": 0.0077, "step": 3720 }, { "epoch": 5.6006006006006, "grad_norm": 0.057626720517873764, "learning_rate": 6.623860811930406e-06, "loss": 0.0168, "step": 3730 }, { "epoch": 5.615615615615615, "grad_norm": 3.927039861679077, "learning_rate": 6.582435791217896e-06, "loss": 0.0073, "step": 3740 }, { "epoch": 5.63063063063063, "grad_norm": 0.04220356419682503, "learning_rate": 6.541010770505386e-06, "loss": 0.0126, "step": 3750 }, { "epoch": 5.645645645645645, "grad_norm": 0.30007079243659973, "learning_rate": 6.499585749792875e-06, "loss": 0.0092, "step": 3760 }, { "epoch": 5.66066066066066, "grad_norm": 0.011833683587610722, "learning_rate": 6.458160729080365e-06, "loss": 0.0077, "step": 3770 }, { "epoch": 5.675675675675675, "grad_norm": 0.1129905954003334, "learning_rate": 6.416735708367855e-06, "loss": 0.0019, "step": 3780 }, { "epoch": 5.6906906906906904, "grad_norm": 0.40713539719581604, "learning_rate": 6.3753106876553436e-06, "loss": 0.0112, "step": 3790 }, { "epoch": 5.7057057057057055, "grad_norm": 0.6869781017303467, "learning_rate": 6.333885666942834e-06, "loss": 0.0044, "step": 3800 }, { "epoch": 5.7207207207207205, "grad_norm": 2.5867929458618164, "learning_rate": 6.292460646230324e-06, "loss": 0.0098, "step": 3810 }, { "epoch": 5.7357357357357355, "grad_norm": 0.9201477766036987, "learning_rate": 6.251035625517813e-06, "loss": 0.0093, "step": 3820 }, { "epoch": 5.7507507507507505, "grad_norm": 1.0136293172836304, "learning_rate": 6.209610604805303e-06, "loss": 0.0166, "step": 3830 }, { "epoch": 5.7657657657657655, "grad_norm": 6.813093662261963, "learning_rate": 6.168185584092793e-06, "loss": 0.0066, "step": 3840 }, { "epoch": 5.7807807807807805, "grad_norm": 0.02762005664408207, "learning_rate": 6.126760563380282e-06, "loss": 0.0039, "step": 3850 }, { "epoch": 5.7957957957957955, "grad_norm": 2.779971122741699, "learning_rate": 6.0853355426677715e-06, "loss": 0.023, "step": 3860 }, { "epoch": 5.8108108108108105, "grad_norm": 5.017881393432617, "learning_rate": 6.043910521955262e-06, "loss": 0.0131, "step": 3870 }, { "epoch": 5.8258258258258255, "grad_norm": 0.02220802940428257, "learning_rate": 6.002485501242751e-06, "loss": 0.0069, "step": 3880 }, { "epoch": 5.8408408408408405, "grad_norm": 0.02759486250579357, "learning_rate": 5.961060480530241e-06, "loss": 0.0026, "step": 3890 }, { "epoch": 5.8558558558558556, "grad_norm": 0.3462941646575928, "learning_rate": 5.919635459817731e-06, "loss": 0.0092, "step": 3900 }, { "epoch": 5.870870870870871, "grad_norm": 0.22936759889125824, "learning_rate": 5.8782104391052195e-06, "loss": 0.0043, "step": 3910 }, { "epoch": 5.885885885885886, "grad_norm": 0.4390968978404999, "learning_rate": 5.83678541839271e-06, "loss": 0.0052, "step": 3920 }, { "epoch": 5.900900900900901, "grad_norm": 0.6948964595794678, "learning_rate": 5.7953603976801995e-06, "loss": 0.0047, "step": 3930 }, { "epoch": 5.915915915915916, "grad_norm": 0.47237828373908997, "learning_rate": 5.753935376967689e-06, "loss": 0.0031, "step": 3940 }, { "epoch": 5.930930930930931, "grad_norm": 0.18768872320652008, "learning_rate": 5.712510356255179e-06, "loss": 0.007, "step": 3950 }, { "epoch": 5.945945945945946, "grad_norm": 0.328825443983078, "learning_rate": 5.671085335542669e-06, "loss": 0.0109, "step": 3960 }, { "epoch": 5.960960960960961, "grad_norm": 0.2502267360687256, "learning_rate": 5.629660314830158e-06, "loss": 0.0078, "step": 3970 }, { "epoch": 5.975975975975976, "grad_norm": 1.2609971761703491, "learning_rate": 5.588235294117647e-06, "loss": 0.0162, "step": 3980 }, { "epoch": 5.990990990990991, "grad_norm": 0.03614773973822594, "learning_rate": 5.546810273405138e-06, "loss": 0.0038, "step": 3990 }, { "epoch": 6.0, "eval_accuracy": 0.863593603010348, "eval_f1": 0.9184759166067578, "eval_loss": 0.21304447948932648, "eval_roc_auc": 0.9443766695354825, "eval_runtime": 7.4272, "eval_samples_per_second": 715.612, "eval_steps_per_second": 22.485, "step": 3996 }, { "epoch": 6.006006006006006, "grad_norm": 0.12015757709741592, "learning_rate": 5.505385252692627e-06, "loss": 0.0053, "step": 4000 }, { "epoch": 6.021021021021021, "grad_norm": 4.061729907989502, "learning_rate": 5.463960231980116e-06, "loss": 0.0082, "step": 4010 }, { "epoch": 6.036036036036036, "grad_norm": 0.02439393475651741, "learning_rate": 5.422535211267607e-06, "loss": 0.0034, "step": 4020 }, { "epoch": 6.051051051051051, "grad_norm": 0.24121862649917603, "learning_rate": 5.381110190555095e-06, "loss": 0.0049, "step": 4030 }, { "epoch": 6.066066066066066, "grad_norm": 0.1391245424747467, "learning_rate": 5.339685169842586e-06, "loss": 0.0032, "step": 4040 }, { "epoch": 6.081081081081081, "grad_norm": 1.9418208599090576, "learning_rate": 5.2982601491300745e-06, "loss": 0.0099, "step": 4050 }, { "epoch": 6.096096096096096, "grad_norm": 0.706449568271637, "learning_rate": 5.256835128417564e-06, "loss": 0.0088, "step": 4060 }, { "epoch": 6.111111111111111, "grad_norm": 0.6291218996047974, "learning_rate": 5.2154101077050546e-06, "loss": 0.0065, "step": 4070 }, { "epoch": 6.126126126126126, "grad_norm": 0.024717414751648903, "learning_rate": 5.173985086992543e-06, "loss": 0.0044, "step": 4080 }, { "epoch": 6.141141141141141, "grad_norm": 0.2399992197751999, "learning_rate": 5.132560066280034e-06, "loss": 0.0024, "step": 4090 }, { "epoch": 6.156156156156156, "grad_norm": 0.02905816026031971, "learning_rate": 5.091135045567523e-06, "loss": 0.0051, "step": 4100 }, { "epoch": 6.171171171171171, "grad_norm": 1.4288544654846191, "learning_rate": 5.049710024855012e-06, "loss": 0.008, "step": 4110 }, { "epoch": 6.186186186186186, "grad_norm": 0.4185940623283386, "learning_rate": 5.0082850041425025e-06, "loss": 0.0063, "step": 4120 }, { "epoch": 6.201201201201201, "grad_norm": 1.1119290590286255, "learning_rate": 4.966859983429992e-06, "loss": 0.0112, "step": 4130 }, { "epoch": 6.216216216216216, "grad_norm": 0.015655577182769775, "learning_rate": 4.925434962717482e-06, "loss": 0.0067, "step": 4140 }, { "epoch": 6.231231231231231, "grad_norm": 3.4215376377105713, "learning_rate": 4.884009942004971e-06, "loss": 0.0076, "step": 4150 }, { "epoch": 6.246246246246246, "grad_norm": 0.013630112633109093, "learning_rate": 4.842584921292461e-06, "loss": 0.0025, "step": 4160 }, { "epoch": 6.261261261261261, "grad_norm": 0.8447656035423279, "learning_rate": 4.8011599005799504e-06, "loss": 0.0065, "step": 4170 }, { "epoch": 6.276276276276276, "grad_norm": 0.1344449371099472, "learning_rate": 4.75973487986744e-06, "loss": 0.0084, "step": 4180 }, { "epoch": 6.291291291291291, "grad_norm": 2.4140686988830566, "learning_rate": 4.71830985915493e-06, "loss": 0.0034, "step": 4190 }, { "epoch": 6.306306306306306, "grad_norm": 0.021810833364725113, "learning_rate": 4.67688483844242e-06, "loss": 0.0029, "step": 4200 }, { "epoch": 6.321321321321321, "grad_norm": 0.031651388853788376, "learning_rate": 4.63545981772991e-06, "loss": 0.0045, "step": 4210 }, { "epoch": 6.336336336336337, "grad_norm": 0.2835276424884796, "learning_rate": 4.594034797017398e-06, "loss": 0.017, "step": 4220 }, { "epoch": 6.351351351351352, "grad_norm": 0.08151044696569443, "learning_rate": 4.552609776304889e-06, "loss": 0.0017, "step": 4230 }, { "epoch": 6.366366366366367, "grad_norm": 0.6036203503608704, "learning_rate": 4.511184755592378e-06, "loss": 0.0029, "step": 4240 }, { "epoch": 6.381381381381382, "grad_norm": 0.13752371072769165, "learning_rate": 4.469759734879868e-06, "loss": 0.0033, "step": 4250 }, { "epoch": 6.396396396396397, "grad_norm": 0.14697255194187164, "learning_rate": 4.4283347141673576e-06, "loss": 0.0083, "step": 4260 }, { "epoch": 6.411411411411412, "grad_norm": 4.969063758850098, "learning_rate": 4.386909693454847e-06, "loss": 0.0092, "step": 4270 }, { "epoch": 6.426426426426427, "grad_norm": 0.26685309410095215, "learning_rate": 4.345484672742337e-06, "loss": 0.0101, "step": 4280 }, { "epoch": 6.441441441441442, "grad_norm": 0.030430737882852554, "learning_rate": 4.304059652029826e-06, "loss": 0.003, "step": 4290 }, { "epoch": 6.456456456456457, "grad_norm": 0.5854690074920654, "learning_rate": 4.262634631317316e-06, "loss": 0.0043, "step": 4300 }, { "epoch": 6.471471471471472, "grad_norm": 0.02023538574576378, "learning_rate": 4.2212096106048055e-06, "loss": 0.0093, "step": 4310 }, { "epoch": 6.486486486486487, "grad_norm": 0.02463540807366371, "learning_rate": 4.179784589892295e-06, "loss": 0.0023, "step": 4320 }, { "epoch": 6.501501501501502, "grad_norm": 0.25623008608818054, "learning_rate": 4.138359569179785e-06, "loss": 0.0083, "step": 4330 }, { "epoch": 6.516516516516517, "grad_norm": 4.038478851318359, "learning_rate": 4.096934548467274e-06, "loss": 0.0076, "step": 4340 }, { "epoch": 6.531531531531532, "grad_norm": 3.8557655811309814, "learning_rate": 4.055509527754764e-06, "loss": 0.0117, "step": 4350 }, { "epoch": 6.546546546546547, "grad_norm": 0.7435543537139893, "learning_rate": 4.014084507042254e-06, "loss": 0.0076, "step": 4360 }, { "epoch": 6.561561561561562, "grad_norm": 0.22783824801445007, "learning_rate": 3.972659486329744e-06, "loss": 0.0047, "step": 4370 }, { "epoch": 6.576576576576577, "grad_norm": 4.226299285888672, "learning_rate": 3.931234465617233e-06, "loss": 0.0121, "step": 4380 }, { "epoch": 6.591591591591592, "grad_norm": 0.0962708368897438, "learning_rate": 3.889809444904723e-06, "loss": 0.0088, "step": 4390 }, { "epoch": 6.606606606606607, "grad_norm": 2.5976169109344482, "learning_rate": 3.848384424192213e-06, "loss": 0.0067, "step": 4400 }, { "epoch": 6.621621621621622, "grad_norm": 0.026430007070302963, "learning_rate": 3.806959403479702e-06, "loss": 0.0029, "step": 4410 }, { "epoch": 6.636636636636637, "grad_norm": 0.3470136523246765, "learning_rate": 3.765534382767192e-06, "loss": 0.0047, "step": 4420 }, { "epoch": 6.651651651651652, "grad_norm": 0.1191195622086525, "learning_rate": 3.7241093620546814e-06, "loss": 0.0025, "step": 4430 }, { "epoch": 6.666666666666667, "grad_norm": 0.15261198580265045, "learning_rate": 3.682684341342171e-06, "loss": 0.015, "step": 4440 }, { "epoch": 6.681681681681682, "grad_norm": 1.6035737991333008, "learning_rate": 3.64125932062966e-06, "loss": 0.0128, "step": 4450 }, { "epoch": 6.696696696696697, "grad_norm": 0.059660863131284714, "learning_rate": 3.5998342999171506e-06, "loss": 0.0032, "step": 4460 }, { "epoch": 6.711711711711712, "grad_norm": 0.0641784518957138, "learning_rate": 3.5584092792046398e-06, "loss": 0.0105, "step": 4470 }, { "epoch": 6.726726726726727, "grad_norm": 0.028454996645450592, "learning_rate": 3.5169842584921293e-06, "loss": 0.0033, "step": 4480 }, { "epoch": 6.741741741741742, "grad_norm": 0.8149629235267639, "learning_rate": 3.4755592377796194e-06, "loss": 0.0031, "step": 4490 }, { "epoch": 6.756756756756757, "grad_norm": 0.023963823914527893, "learning_rate": 3.434134217067109e-06, "loss": 0.0098, "step": 4500 }, { "epoch": 6.771771771771772, "grad_norm": 0.24265415966510773, "learning_rate": 3.392709196354598e-06, "loss": 0.0071, "step": 4510 }, { "epoch": 6.786786786786787, "grad_norm": 0.018639802932739258, "learning_rate": 3.351284175642088e-06, "loss": 0.0072, "step": 4520 }, { "epoch": 6.801801801801802, "grad_norm": 2.4583423137664795, "learning_rate": 3.3098591549295777e-06, "loss": 0.0175, "step": 4530 }, { "epoch": 6.816816816816817, "grad_norm": 0.10339108109474182, "learning_rate": 3.2684341342170673e-06, "loss": 0.0162, "step": 4540 }, { "epoch": 6.831831831831832, "grad_norm": 0.41008320450782776, "learning_rate": 3.2270091135045573e-06, "loss": 0.0034, "step": 4550 }, { "epoch": 6.846846846846847, "grad_norm": 0.056475620716810226, "learning_rate": 3.185584092792047e-06, "loss": 0.0092, "step": 4560 }, { "epoch": 6.861861861861862, "grad_norm": 0.3767000138759613, "learning_rate": 3.144159072079536e-06, "loss": 0.0097, "step": 4570 }, { "epoch": 6.876876876876877, "grad_norm": 0.08370574563741684, "learning_rate": 3.102734051367026e-06, "loss": 0.0063, "step": 4580 }, { "epoch": 6.891891891891892, "grad_norm": 0.25333690643310547, "learning_rate": 3.0613090306545157e-06, "loss": 0.0031, "step": 4590 }, { "epoch": 6.906906906906907, "grad_norm": 0.06466072797775269, "learning_rate": 3.0198840099420052e-06, "loss": 0.0091, "step": 4600 }, { "epoch": 6.921921921921922, "grad_norm": 0.45550736784935, "learning_rate": 2.9784589892294944e-06, "loss": 0.0063, "step": 4610 }, { "epoch": 6.936936936936937, "grad_norm": 0.2586912512779236, "learning_rate": 2.937033968516985e-06, "loss": 0.0038, "step": 4620 }, { "epoch": 6.951951951951952, "grad_norm": 0.020179102197289467, "learning_rate": 2.895608947804474e-06, "loss": 0.0084, "step": 4630 }, { "epoch": 6.966966966966967, "grad_norm": 0.016084246337413788, "learning_rate": 2.8541839270919636e-06, "loss": 0.0112, "step": 4640 }, { "epoch": 6.981981981981982, "grad_norm": 2.5298447608947754, "learning_rate": 2.8127589063794536e-06, "loss": 0.0085, "step": 4650 }, { "epoch": 6.996996996996997, "grad_norm": 1.6728397607803345, "learning_rate": 2.771333885666943e-06, "loss": 0.0092, "step": 4660 }, { "epoch": 7.0, "eval_accuracy": 0.8630291627469426, "eval_f1": 0.9187405866743169, "eval_loss": 0.21907110512256622, "eval_roc_auc": 0.9451891020630947, "eval_runtime": 7.3889, "eval_samples_per_second": 719.318, "eval_steps_per_second": 22.601, "step": 4662 } ], "logging_steps": 10, "max_steps": 5328, "num_input_tokens_seen": 0, "num_train_epochs": 8, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 4368118206154752.0, "train_batch_size": 32, "trial_name": null, "trial_params": null }