{ "best_global_step": 2000, "best_metric": 1.216275863572211, "best_model_checkpoint": "./SALAMA_NEWMED8/checkpoint-2000", "epoch": 0.7990012484394506, "eval_steps": 2000, "global_step": 2000, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.003995006242197253, "grad_norm": 1.0434356927871704, "learning_rate": 1.8e-07, "loss": 0.0117, "step": 10 }, { "epoch": 0.007990012484394507, "grad_norm": 1.660996675491333, "learning_rate": 3.8e-07, "loss": 0.0129, "step": 20 }, { "epoch": 0.01198501872659176, "grad_norm": 1.3255594968795776, "learning_rate": 5.800000000000001e-07, "loss": 0.0149, "step": 30 }, { "epoch": 0.015980024968789013, "grad_norm": 1.7757971286773682, "learning_rate": 7.8e-07, "loss": 0.0169, "step": 40 }, { "epoch": 0.019975031210986267, "grad_norm": 0.890116810798645, "learning_rate": 9.800000000000001e-07, "loss": 0.0098, "step": 50 }, { "epoch": 0.02397003745318352, "grad_norm": 0.7157179117202759, "learning_rate": 1.1800000000000001e-06, "loss": 0.0135, "step": 60 }, { "epoch": 0.027965043695380776, "grad_norm": 0.7145370244979858, "learning_rate": 1.3800000000000001e-06, "loss": 0.0129, "step": 70 }, { "epoch": 0.031960049937578026, "grad_norm": 1.092121958732605, "learning_rate": 1.5800000000000001e-06, "loss": 0.0135, "step": 80 }, { "epoch": 0.035955056179775284, "grad_norm": 1.9209226369857788, "learning_rate": 1.7800000000000001e-06, "loss": 0.0087, "step": 90 }, { "epoch": 0.039950062421972535, "grad_norm": 0.5864148139953613, "learning_rate": 1.98e-06, "loss": 0.0193, "step": 100 }, { "epoch": 0.043945068664169785, "grad_norm": 1.5507863759994507, "learning_rate": 2.1800000000000003e-06, "loss": 0.0083, "step": 110 }, { "epoch": 0.04794007490636704, "grad_norm": 0.3510238826274872, "learning_rate": 2.38e-06, "loss": 0.0096, "step": 120 }, { "epoch": 0.051935081148564294, "grad_norm": 1.3097003698349, "learning_rate": 2.5800000000000003e-06, "loss": 0.0187, "step": 130 }, { "epoch": 0.05593008739076155, "grad_norm": 1.5574332475662231, "learning_rate": 2.7800000000000005e-06, "loss": 0.0134, "step": 140 }, { "epoch": 0.0599250936329588, "grad_norm": 2.5471532344818115, "learning_rate": 2.9800000000000003e-06, "loss": 0.0132, "step": 150 }, { "epoch": 0.06392009987515605, "grad_norm": 1.05274498462677, "learning_rate": 3.1800000000000005e-06, "loss": 0.0114, "step": 160 }, { "epoch": 0.06791510611735331, "grad_norm": 0.276777446269989, "learning_rate": 3.3800000000000007e-06, "loss": 0.0152, "step": 170 }, { "epoch": 0.07191011235955057, "grad_norm": 1.5912141799926758, "learning_rate": 3.58e-06, "loss": 0.0174, "step": 180 }, { "epoch": 0.07590511860174781, "grad_norm": 1.230015754699707, "learning_rate": 3.7800000000000002e-06, "loss": 0.0087, "step": 190 }, { "epoch": 0.07990012484394507, "grad_norm": 1.1366710662841797, "learning_rate": 3.980000000000001e-06, "loss": 0.0124, "step": 200 }, { "epoch": 0.08389513108614233, "grad_norm": 2.4593794345855713, "learning_rate": 4.18e-06, "loss": 0.0166, "step": 210 }, { "epoch": 0.08789013732833957, "grad_norm": 1.208004355430603, "learning_rate": 4.38e-06, "loss": 0.0137, "step": 220 }, { "epoch": 0.09188514357053683, "grad_norm": 1.0044916868209839, "learning_rate": 4.58e-06, "loss": 0.0189, "step": 230 }, { "epoch": 0.09588014981273409, "grad_norm": 1.5824278593063354, "learning_rate": 4.78e-06, "loss": 0.0098, "step": 240 }, { "epoch": 0.09987515605493133, "grad_norm": 0.8815672397613525, "learning_rate": 4.980000000000001e-06, "loss": 0.0231, "step": 250 }, { "epoch": 0.10387016229712859, "grad_norm": 0.8806136846542358, "learning_rate": 5.18e-06, "loss": 0.0094, "step": 260 }, { "epoch": 0.10786516853932585, "grad_norm": 0.7804479002952576, "learning_rate": 5.380000000000001e-06, "loss": 0.0146, "step": 270 }, { "epoch": 0.1118601747815231, "grad_norm": 1.9678781032562256, "learning_rate": 5.580000000000001e-06, "loss": 0.0243, "step": 280 }, { "epoch": 0.11585518102372035, "grad_norm": 1.9052259922027588, "learning_rate": 5.78e-06, "loss": 0.0168, "step": 290 }, { "epoch": 0.1198501872659176, "grad_norm": 0.9115912914276123, "learning_rate": 5.98e-06, "loss": 0.0125, "step": 300 }, { "epoch": 0.12384519350811486, "grad_norm": 2.4003071784973145, "learning_rate": 6.18e-06, "loss": 0.0276, "step": 310 }, { "epoch": 0.1278401997503121, "grad_norm": 1.5195643901824951, "learning_rate": 6.380000000000001e-06, "loss": 0.0209, "step": 320 }, { "epoch": 0.13183520599250936, "grad_norm": 1.8138670921325684, "learning_rate": 6.5800000000000005e-06, "loss": 0.0148, "step": 330 }, { "epoch": 0.13583021223470662, "grad_norm": 1.7891756296157837, "learning_rate": 6.780000000000001e-06, "loss": 0.0127, "step": 340 }, { "epoch": 0.13982521847690388, "grad_norm": 2.7678897380828857, "learning_rate": 6.98e-06, "loss": 0.0232, "step": 350 }, { "epoch": 0.14382022471910114, "grad_norm": 1.1360479593276978, "learning_rate": 7.180000000000001e-06, "loss": 0.0149, "step": 360 }, { "epoch": 0.14781523096129837, "grad_norm": 1.1238574981689453, "learning_rate": 7.3800000000000005e-06, "loss": 0.0155, "step": 370 }, { "epoch": 0.15181023720349562, "grad_norm": 1.681251049041748, "learning_rate": 7.58e-06, "loss": 0.0229, "step": 380 }, { "epoch": 0.15580524344569288, "grad_norm": 1.4169082641601562, "learning_rate": 7.78e-06, "loss": 0.0202, "step": 390 }, { "epoch": 0.15980024968789014, "grad_norm": 0.7510893940925598, "learning_rate": 7.980000000000002e-06, "loss": 0.0144, "step": 400 }, { "epoch": 0.1637952559300874, "grad_norm": 1.6169250011444092, "learning_rate": 8.18e-06, "loss": 0.0139, "step": 410 }, { "epoch": 0.16779026217228465, "grad_norm": 2.5482001304626465, "learning_rate": 8.380000000000001e-06, "loss": 0.0297, "step": 420 }, { "epoch": 0.17178526841448188, "grad_norm": 1.1548925638198853, "learning_rate": 8.580000000000001e-06, "loss": 0.0214, "step": 430 }, { "epoch": 0.17578027465667914, "grad_norm": 1.8763704299926758, "learning_rate": 8.78e-06, "loss": 0.0118, "step": 440 }, { "epoch": 0.1797752808988764, "grad_norm": 1.5481230020523071, "learning_rate": 8.98e-06, "loss": 0.0223, "step": 450 }, { "epoch": 0.18377028714107366, "grad_norm": 0.8432000279426575, "learning_rate": 9.180000000000002e-06, "loss": 0.0249, "step": 460 }, { "epoch": 0.18776529338327091, "grad_norm": 1.0781527757644653, "learning_rate": 9.38e-06, "loss": 0.0213, "step": 470 }, { "epoch": 0.19176029962546817, "grad_norm": 2.022068977355957, "learning_rate": 9.58e-06, "loss": 0.0244, "step": 480 }, { "epoch": 0.19575530586766543, "grad_norm": 1.7390263080596924, "learning_rate": 9.780000000000001e-06, "loss": 0.0206, "step": 490 }, { "epoch": 0.19975031210986266, "grad_norm": 0.7592054605484009, "learning_rate": 9.980000000000001e-06, "loss": 0.0288, "step": 500 }, { "epoch": 0.20374531835205992, "grad_norm": 1.2262804508209229, "learning_rate": 9.980035492457854e-06, "loss": 0.0184, "step": 510 }, { "epoch": 0.20774032459425718, "grad_norm": 1.5795695781707764, "learning_rate": 9.957852706299913e-06, "loss": 0.0137, "step": 520 }, { "epoch": 0.21173533083645443, "grad_norm": 2.214512586593628, "learning_rate": 9.935669920141971e-06, "loss": 0.0229, "step": 530 }, { "epoch": 0.2157303370786517, "grad_norm": 1.125135064125061, "learning_rate": 9.91348713398403e-06, "loss": 0.0197, "step": 540 }, { "epoch": 0.21972534332084895, "grad_norm": 1.3488562107086182, "learning_rate": 9.891304347826088e-06, "loss": 0.0265, "step": 550 }, { "epoch": 0.2237203495630462, "grad_norm": 1.836803913116455, "learning_rate": 9.869121561668146e-06, "loss": 0.0161, "step": 560 }, { "epoch": 0.22771535580524344, "grad_norm": 1.634339451789856, "learning_rate": 9.846938775510205e-06, "loss": 0.0262, "step": 570 }, { "epoch": 0.2317103620474407, "grad_norm": 1.905315637588501, "learning_rate": 9.824755989352263e-06, "loss": 0.0185, "step": 580 }, { "epoch": 0.23570536828963795, "grad_norm": 2.3678486347198486, "learning_rate": 9.802573203194321e-06, "loss": 0.019, "step": 590 }, { "epoch": 0.2397003745318352, "grad_norm": 2.1958577632904053, "learning_rate": 9.78039041703638e-06, "loss": 0.0271, "step": 600 }, { "epoch": 0.24369538077403247, "grad_norm": 2.0940821170806885, "learning_rate": 9.758207630878438e-06, "loss": 0.0248, "step": 610 }, { "epoch": 0.24769038701622972, "grad_norm": 1.8483749628067017, "learning_rate": 9.736024844720497e-06, "loss": 0.0267, "step": 620 }, { "epoch": 0.251685393258427, "grad_norm": 1.6134873628616333, "learning_rate": 9.713842058562555e-06, "loss": 0.0305, "step": 630 }, { "epoch": 0.2556803995006242, "grad_norm": 1.7308474779129028, "learning_rate": 9.691659272404614e-06, "loss": 0.0365, "step": 640 }, { "epoch": 0.2596754057428215, "grad_norm": 1.9092700481414795, "learning_rate": 9.669476486246674e-06, "loss": 0.0224, "step": 650 }, { "epoch": 0.2636704119850187, "grad_norm": 1.6482040882110596, "learning_rate": 9.647293700088732e-06, "loss": 0.0231, "step": 660 }, { "epoch": 0.26766541822721596, "grad_norm": 1.7020848989486694, "learning_rate": 9.62511091393079e-06, "loss": 0.0279, "step": 670 }, { "epoch": 0.27166042446941324, "grad_norm": 1.9563522338867188, "learning_rate": 9.602928127772849e-06, "loss": 0.0238, "step": 680 }, { "epoch": 0.27565543071161047, "grad_norm": 1.9009273052215576, "learning_rate": 9.580745341614907e-06, "loss": 0.0206, "step": 690 }, { "epoch": 0.27965043695380776, "grad_norm": 1.7510108947753906, "learning_rate": 9.558562555456966e-06, "loss": 0.0164, "step": 700 }, { "epoch": 0.283645443196005, "grad_norm": 1.257036566734314, "learning_rate": 9.536379769299024e-06, "loss": 0.0262, "step": 710 }, { "epoch": 0.2876404494382023, "grad_norm": 1.376685380935669, "learning_rate": 9.514196983141083e-06, "loss": 0.0208, "step": 720 }, { "epoch": 0.2916354556803995, "grad_norm": 2.9449803829193115, "learning_rate": 9.492014196983141e-06, "loss": 0.0259, "step": 730 }, { "epoch": 0.29563046192259673, "grad_norm": 1.7352170944213867, "learning_rate": 9.469831410825201e-06, "loss": 0.027, "step": 740 }, { "epoch": 0.299625468164794, "grad_norm": 2.675891876220703, "learning_rate": 9.44764862466726e-06, "loss": 0.0308, "step": 750 }, { "epoch": 0.30362047440699125, "grad_norm": 2.65049147605896, "learning_rate": 9.425465838509318e-06, "loss": 0.0288, "step": 760 }, { "epoch": 0.30761548064918853, "grad_norm": 1.4568003416061401, "learning_rate": 9.403283052351377e-06, "loss": 0.0207, "step": 770 }, { "epoch": 0.31161048689138576, "grad_norm": 1.7025985717773438, "learning_rate": 9.381100266193435e-06, "loss": 0.0217, "step": 780 }, { "epoch": 0.31560549313358305, "grad_norm": 1.9456532001495361, "learning_rate": 9.358917480035493e-06, "loss": 0.0232, "step": 790 }, { "epoch": 0.3196004993757803, "grad_norm": 1.8600986003875732, "learning_rate": 9.336734693877552e-06, "loss": 0.0323, "step": 800 }, { "epoch": 0.3235955056179775, "grad_norm": 1.3675034046173096, "learning_rate": 9.31455190771961e-06, "loss": 0.0222, "step": 810 }, { "epoch": 0.3275905118601748, "grad_norm": 2.452406883239746, "learning_rate": 9.292369121561669e-06, "loss": 0.0252, "step": 820 }, { "epoch": 0.331585518102372, "grad_norm": 1.9519559144973755, "learning_rate": 9.270186335403727e-06, "loss": 0.0211, "step": 830 }, { "epoch": 0.3355805243445693, "grad_norm": 1.6819555759429932, "learning_rate": 9.248003549245787e-06, "loss": 0.0318, "step": 840 }, { "epoch": 0.33957553058676654, "grad_norm": 9.365504264831543, "learning_rate": 9.225820763087846e-06, "loss": 0.0287, "step": 850 }, { "epoch": 0.34357053682896377, "grad_norm": 2.786489248275757, "learning_rate": 9.203637976929904e-06, "loss": 0.0212, "step": 860 }, { "epoch": 0.34756554307116105, "grad_norm": 2.4852006435394287, "learning_rate": 9.181455190771963e-06, "loss": 0.0313, "step": 870 }, { "epoch": 0.3515605493133583, "grad_norm": 1.6487276554107666, "learning_rate": 9.159272404614021e-06, "loss": 0.016, "step": 880 }, { "epoch": 0.35555555555555557, "grad_norm": 2.3104565143585205, "learning_rate": 9.13708961845608e-06, "loss": 0.0232, "step": 890 }, { "epoch": 0.3595505617977528, "grad_norm": 1.7769932746887207, "learning_rate": 9.114906832298138e-06, "loss": 0.0198, "step": 900 }, { "epoch": 0.3635455680399501, "grad_norm": 2.558542490005493, "learning_rate": 9.092724046140196e-06, "loss": 0.0297, "step": 910 }, { "epoch": 0.3675405742821473, "grad_norm": 1.9595332145690918, "learning_rate": 9.070541259982255e-06, "loss": 0.021, "step": 920 }, { "epoch": 0.37153558052434454, "grad_norm": 2.161356210708618, "learning_rate": 9.048358473824313e-06, "loss": 0.0333, "step": 930 }, { "epoch": 0.37553058676654183, "grad_norm": 1.427983283996582, "learning_rate": 9.026175687666371e-06, "loss": 0.0159, "step": 940 }, { "epoch": 0.37952559300873906, "grad_norm": 2.5176424980163574, "learning_rate": 9.00399290150843e-06, "loss": 0.0337, "step": 950 }, { "epoch": 0.38352059925093634, "grad_norm": 2.1165552139282227, "learning_rate": 8.981810115350488e-06, "loss": 0.0382, "step": 960 }, { "epoch": 0.3875156054931336, "grad_norm": 1.4064266681671143, "learning_rate": 8.959627329192547e-06, "loss": 0.0228, "step": 970 }, { "epoch": 0.39151061173533086, "grad_norm": 2.283217668533325, "learning_rate": 8.937444543034605e-06, "loss": 0.0233, "step": 980 }, { "epoch": 0.3955056179775281, "grad_norm": 1.8586297035217285, "learning_rate": 8.915261756876664e-06, "loss": 0.0285, "step": 990 }, { "epoch": 0.3995006242197253, "grad_norm": 2.8438491821289062, "learning_rate": 8.893078970718722e-06, "loss": 0.0241, "step": 1000 }, { "epoch": 0.4034956304619226, "grad_norm": 2.190476655960083, "learning_rate": 8.87089618456078e-06, "loss": 0.0322, "step": 1010 }, { "epoch": 0.40749063670411984, "grad_norm": 1.7664802074432373, "learning_rate": 8.84871339840284e-06, "loss": 0.032, "step": 1020 }, { "epoch": 0.4114856429463171, "grad_norm": 2.394594430923462, "learning_rate": 8.826530612244899e-06, "loss": 0.0298, "step": 1030 }, { "epoch": 0.41548064918851435, "grad_norm": 3.4191994667053223, "learning_rate": 8.804347826086957e-06, "loss": 0.0266, "step": 1040 }, { "epoch": 0.41947565543071164, "grad_norm": 1.0744972229003906, "learning_rate": 8.782165039929016e-06, "loss": 0.0294, "step": 1050 }, { "epoch": 0.42347066167290887, "grad_norm": 1.4667426347732544, "learning_rate": 8.759982253771074e-06, "loss": 0.0244, "step": 1060 }, { "epoch": 0.4274656679151061, "grad_norm": 1.7679630517959595, "learning_rate": 8.737799467613133e-06, "loss": 0.0182, "step": 1070 }, { "epoch": 0.4314606741573034, "grad_norm": 1.9316728115081787, "learning_rate": 8.715616681455191e-06, "loss": 0.0295, "step": 1080 }, { "epoch": 0.4354556803995006, "grad_norm": 1.9184544086456299, "learning_rate": 8.69343389529725e-06, "loss": 0.0193, "step": 1090 }, { "epoch": 0.4394506866416979, "grad_norm": 2.1781647205352783, "learning_rate": 8.671251109139308e-06, "loss": 0.0196, "step": 1100 }, { "epoch": 0.4434456928838951, "grad_norm": 1.777761459350586, "learning_rate": 8.649068322981368e-06, "loss": 0.0352, "step": 1110 }, { "epoch": 0.4474406991260924, "grad_norm": 1.3648267984390259, "learning_rate": 8.626885536823427e-06, "loss": 0.0228, "step": 1120 }, { "epoch": 0.45143570536828964, "grad_norm": 1.1893045902252197, "learning_rate": 8.604702750665485e-06, "loss": 0.0204, "step": 1130 }, { "epoch": 0.45543071161048687, "grad_norm": 2.0382745265960693, "learning_rate": 8.582519964507543e-06, "loss": 0.0268, "step": 1140 }, { "epoch": 0.45942571785268416, "grad_norm": 1.841737985610962, "learning_rate": 8.560337178349602e-06, "loss": 0.0253, "step": 1150 }, { "epoch": 0.4634207240948814, "grad_norm": 2.250272750854492, "learning_rate": 8.53815439219166e-06, "loss": 0.0194, "step": 1160 }, { "epoch": 0.46741573033707867, "grad_norm": 3.142618417739868, "learning_rate": 8.515971606033719e-06, "loss": 0.0239, "step": 1170 }, { "epoch": 0.4714107365792759, "grad_norm": 1.7657616138458252, "learning_rate": 8.493788819875777e-06, "loss": 0.0241, "step": 1180 }, { "epoch": 0.47540574282147313, "grad_norm": 1.2238365411758423, "learning_rate": 8.471606033717836e-06, "loss": 0.0188, "step": 1190 }, { "epoch": 0.4794007490636704, "grad_norm": 2.4322609901428223, "learning_rate": 8.449423247559894e-06, "loss": 0.0245, "step": 1200 }, { "epoch": 0.48339575530586765, "grad_norm": 2.3233511447906494, "learning_rate": 8.427240461401952e-06, "loss": 0.025, "step": 1210 }, { "epoch": 0.48739076154806493, "grad_norm": 1.8041311502456665, "learning_rate": 8.40505767524401e-06, "loss": 0.0249, "step": 1220 }, { "epoch": 0.49138576779026216, "grad_norm": 2.726259469985962, "learning_rate": 8.38287488908607e-06, "loss": 0.0247, "step": 1230 }, { "epoch": 0.49538077403245945, "grad_norm": 1.5258828401565552, "learning_rate": 8.360692102928128e-06, "loss": 0.0212, "step": 1240 }, { "epoch": 0.4993757802746567, "grad_norm": 1.2424405813217163, "learning_rate": 8.338509316770186e-06, "loss": 0.0284, "step": 1250 }, { "epoch": 0.503370786516854, "grad_norm": 1.3515020608901978, "learning_rate": 8.316326530612246e-06, "loss": 0.0197, "step": 1260 }, { "epoch": 0.5073657927590511, "grad_norm": 1.9467544555664062, "learning_rate": 8.294143744454305e-06, "loss": 0.0193, "step": 1270 }, { "epoch": 0.5113607990012484, "grad_norm": 1.0850805044174194, "learning_rate": 8.271960958296363e-06, "loss": 0.0182, "step": 1280 }, { "epoch": 0.5153558052434457, "grad_norm": 1.8913824558258057, "learning_rate": 8.249778172138421e-06, "loss": 0.0198, "step": 1290 }, { "epoch": 0.519350811485643, "grad_norm": 1.2548930644989014, "learning_rate": 8.22759538598048e-06, "loss": 0.0221, "step": 1300 }, { "epoch": 0.5233458177278402, "grad_norm": 1.1515015363693237, "learning_rate": 8.205412599822538e-06, "loss": 0.0216, "step": 1310 }, { "epoch": 0.5273408239700375, "grad_norm": 1.1634601354599, "learning_rate": 8.183229813664597e-06, "loss": 0.0179, "step": 1320 }, { "epoch": 0.5313358302122347, "grad_norm": 1.7621314525604248, "learning_rate": 8.161047027506655e-06, "loss": 0.028, "step": 1330 }, { "epoch": 0.5353308364544319, "grad_norm": 2.0318105220794678, "learning_rate": 8.138864241348714e-06, "loss": 0.0253, "step": 1340 }, { "epoch": 0.5393258426966292, "grad_norm": 1.7418289184570312, "learning_rate": 8.116681455190772e-06, "loss": 0.0289, "step": 1350 }, { "epoch": 0.5433208489388265, "grad_norm": 2.0739269256591797, "learning_rate": 8.09449866903283e-06, "loss": 0.0219, "step": 1360 }, { "epoch": 0.5473158551810238, "grad_norm": 3.1342859268188477, "learning_rate": 8.072315882874889e-06, "loss": 0.0315, "step": 1370 }, { "epoch": 0.5513108614232209, "grad_norm": 1.7212834358215332, "learning_rate": 8.050133096716947e-06, "loss": 0.016, "step": 1380 }, { "epoch": 0.5553058676654182, "grad_norm": 1.869975209236145, "learning_rate": 8.027950310559007e-06, "loss": 0.0207, "step": 1390 }, { "epoch": 0.5593008739076155, "grad_norm": 2.3647046089172363, "learning_rate": 8.005767524401066e-06, "loss": 0.02, "step": 1400 }, { "epoch": 0.5632958801498127, "grad_norm": 1.8435038328170776, "learning_rate": 7.983584738243124e-06, "loss": 0.024, "step": 1410 }, { "epoch": 0.56729088639201, "grad_norm": 1.6630502939224243, "learning_rate": 7.961401952085183e-06, "loss": 0.0329, "step": 1420 }, { "epoch": 0.5712858926342073, "grad_norm": 1.3610472679138184, "learning_rate": 7.939219165927241e-06, "loss": 0.0184, "step": 1430 }, { "epoch": 0.5752808988764045, "grad_norm": 1.3569384813308716, "learning_rate": 7.9170363797693e-06, "loss": 0.0173, "step": 1440 }, { "epoch": 0.5792759051186017, "grad_norm": 1.6595990657806396, "learning_rate": 7.894853593611358e-06, "loss": 0.0197, "step": 1450 }, { "epoch": 0.583270911360799, "grad_norm": 1.4524695873260498, "learning_rate": 7.872670807453416e-06, "loss": 0.0215, "step": 1460 }, { "epoch": 0.5872659176029963, "grad_norm": 2.361285924911499, "learning_rate": 7.850488021295475e-06, "loss": 0.0285, "step": 1470 }, { "epoch": 0.5912609238451935, "grad_norm": 1.3114995956420898, "learning_rate": 7.828305235137535e-06, "loss": 0.0201, "step": 1480 }, { "epoch": 0.5952559300873907, "grad_norm": 1.7692322731018066, "learning_rate": 7.806122448979593e-06, "loss": 0.0208, "step": 1490 }, { "epoch": 0.599250936329588, "grad_norm": 4.138287544250488, "learning_rate": 7.783939662821652e-06, "loss": 0.0278, "step": 1500 }, { "epoch": 0.6032459425717853, "grad_norm": 0.6977252960205078, "learning_rate": 7.76175687666371e-06, "loss": 0.024, "step": 1510 }, { "epoch": 0.6072409488139825, "grad_norm": 2.3383960723876953, "learning_rate": 7.739574090505769e-06, "loss": 0.022, "step": 1520 }, { "epoch": 0.6112359550561798, "grad_norm": 2.005115032196045, "learning_rate": 7.717391304347827e-06, "loss": 0.0263, "step": 1530 }, { "epoch": 0.6152309612983771, "grad_norm": 1.0198010206222534, "learning_rate": 7.695208518189886e-06, "loss": 0.0283, "step": 1540 }, { "epoch": 0.6192259675405742, "grad_norm": 1.7714745998382568, "learning_rate": 7.673025732031944e-06, "loss": 0.0189, "step": 1550 }, { "epoch": 0.6232209737827715, "grad_norm": 0.724946141242981, "learning_rate": 7.650842945874002e-06, "loss": 0.0212, "step": 1560 }, { "epoch": 0.6272159800249688, "grad_norm": 2.007723569869995, "learning_rate": 7.628660159716062e-06, "loss": 0.0296, "step": 1570 }, { "epoch": 0.6312109862671661, "grad_norm": 1.758449912071228, "learning_rate": 7.60647737355812e-06, "loss": 0.0265, "step": 1580 }, { "epoch": 0.6352059925093633, "grad_norm": 1.2592417001724243, "learning_rate": 7.5842945874001785e-06, "loss": 0.0189, "step": 1590 }, { "epoch": 0.6392009987515606, "grad_norm": 2.694614887237549, "learning_rate": 7.562111801242237e-06, "loss": 0.0229, "step": 1600 }, { "epoch": 0.6431960049937578, "grad_norm": 0.9839036464691162, "learning_rate": 7.539929015084295e-06, "loss": 0.0261, "step": 1610 }, { "epoch": 0.647191011235955, "grad_norm": 0.8098208904266357, "learning_rate": 7.517746228926354e-06, "loss": 0.0223, "step": 1620 }, { "epoch": 0.6511860174781523, "grad_norm": 1.9389737844467163, "learning_rate": 7.495563442768412e-06, "loss": 0.0161, "step": 1630 }, { "epoch": 0.6551810237203496, "grad_norm": 1.6782140731811523, "learning_rate": 7.473380656610471e-06, "loss": 0.0245, "step": 1640 }, { "epoch": 0.6591760299625468, "grad_norm": 1.6728957891464233, "learning_rate": 7.451197870452529e-06, "loss": 0.0147, "step": 1650 }, { "epoch": 0.663171036204744, "grad_norm": 2.187469005584717, "learning_rate": 7.429015084294588e-06, "loss": 0.0277, "step": 1660 }, { "epoch": 0.6671660424469413, "grad_norm": 1.527753472328186, "learning_rate": 7.406832298136647e-06, "loss": 0.0212, "step": 1670 }, { "epoch": 0.6711610486891386, "grad_norm": 1.698236346244812, "learning_rate": 7.384649511978705e-06, "loss": 0.0232, "step": 1680 }, { "epoch": 0.6751560549313358, "grad_norm": 3.0273208618164062, "learning_rate": 7.362466725820764e-06, "loss": 0.0199, "step": 1690 }, { "epoch": 0.6791510611735331, "grad_norm": 1.5876965522766113, "learning_rate": 7.340283939662822e-06, "loss": 0.0186, "step": 1700 }, { "epoch": 0.6831460674157304, "grad_norm": 3.0659890174865723, "learning_rate": 7.3181011535048805e-06, "loss": 0.0224, "step": 1710 }, { "epoch": 0.6871410736579275, "grad_norm": 1.5666126012802124, "learning_rate": 7.295918367346939e-06, "loss": 0.0256, "step": 1720 }, { "epoch": 0.6911360799001248, "grad_norm": 2.3663785457611084, "learning_rate": 7.273735581188997e-06, "loss": 0.0258, "step": 1730 }, { "epoch": 0.6951310861423221, "grad_norm": 2.1256401538848877, "learning_rate": 7.251552795031056e-06, "loss": 0.0231, "step": 1740 }, { "epoch": 0.6991260923845194, "grad_norm": 1.7226366996765137, "learning_rate": 7.229370008873116e-06, "loss": 0.0242, "step": 1750 }, { "epoch": 0.7031210986267166, "grad_norm": 1.5627435445785522, "learning_rate": 7.207187222715174e-06, "loss": 0.019, "step": 1760 }, { "epoch": 0.7071161048689139, "grad_norm": 1.726256251335144, "learning_rate": 7.185004436557233e-06, "loss": 0.0214, "step": 1770 }, { "epoch": 0.7111111111111111, "grad_norm": 2.170620918273926, "learning_rate": 7.162821650399291e-06, "loss": 0.0268, "step": 1780 }, { "epoch": 0.7151061173533083, "grad_norm": 1.2729148864746094, "learning_rate": 7.1406388642413496e-06, "loss": 0.0233, "step": 1790 }, { "epoch": 0.7191011235955056, "grad_norm": 1.9662100076675415, "learning_rate": 7.118456078083408e-06, "loss": 0.0183, "step": 1800 }, { "epoch": 0.7230961298377029, "grad_norm": 2.146286725997925, "learning_rate": 7.0962732919254664e-06, "loss": 0.0168, "step": 1810 }, { "epoch": 0.7270911360799002, "grad_norm": 2.4282994270324707, "learning_rate": 7.074090505767525e-06, "loss": 0.0262, "step": 1820 }, { "epoch": 0.7310861423220973, "grad_norm": 0.9622244834899902, "learning_rate": 7.051907719609583e-06, "loss": 0.0161, "step": 1830 }, { "epoch": 0.7350811485642946, "grad_norm": 1.1307194232940674, "learning_rate": 7.0297249334516426e-06, "loss": 0.0112, "step": 1840 }, { "epoch": 0.7390761548064919, "grad_norm": 2.2122983932495117, "learning_rate": 7.007542147293701e-06, "loss": 0.0249, "step": 1850 }, { "epoch": 0.7430711610486891, "grad_norm": 1.137358546257019, "learning_rate": 6.985359361135759e-06, "loss": 0.0187, "step": 1860 }, { "epoch": 0.7470661672908864, "grad_norm": 1.170784831047058, "learning_rate": 6.963176574977818e-06, "loss": 0.0258, "step": 1870 }, { "epoch": 0.7510611735330837, "grad_norm": 2.369542121887207, "learning_rate": 6.940993788819876e-06, "loss": 0.0186, "step": 1880 }, { "epoch": 0.755056179775281, "grad_norm": 1.8876056671142578, "learning_rate": 6.918811002661935e-06, "loss": 0.0173, "step": 1890 }, { "epoch": 0.7590511860174781, "grad_norm": 0.8910442590713501, "learning_rate": 6.896628216503993e-06, "loss": 0.0338, "step": 1900 }, { "epoch": 0.7630461922596754, "grad_norm": 1.264801025390625, "learning_rate": 6.8744454303460515e-06, "loss": 0.0173, "step": 1910 }, { "epoch": 0.7670411985018727, "grad_norm": 1.1293343305587769, "learning_rate": 6.85226264418811e-06, "loss": 0.0214, "step": 1920 }, { "epoch": 0.7710362047440699, "grad_norm": 1.2280560731887817, "learning_rate": 6.830079858030168e-06, "loss": 0.0238, "step": 1930 }, { "epoch": 0.7750312109862671, "grad_norm": 2.1744937896728516, "learning_rate": 6.8078970718722285e-06, "loss": 0.0233, "step": 1940 }, { "epoch": 0.7790262172284644, "grad_norm": 1.7281086444854736, "learning_rate": 6.785714285714287e-06, "loss": 0.0253, "step": 1950 }, { "epoch": 0.7830212234706617, "grad_norm": 1.299452543258667, "learning_rate": 6.763531499556345e-06, "loss": 0.0187, "step": 1960 }, { "epoch": 0.7870162297128589, "grad_norm": 2.433673620223999, "learning_rate": 6.741348713398404e-06, "loss": 0.0144, "step": 1970 }, { "epoch": 0.7910112359550562, "grad_norm": 1.6157279014587402, "learning_rate": 6.719165927240462e-06, "loss": 0.0173, "step": 1980 }, { "epoch": 0.7950062421972535, "grad_norm": 2.246328830718994, "learning_rate": 6.696983141082521e-06, "loss": 0.0213, "step": 1990 }, { "epoch": 0.7990012484394506, "grad_norm": 1.133818507194519, "learning_rate": 6.674800354924579e-06, "loss": 0.0153, "step": 2000 }, { "epoch": 0.7990012484394506, "eval_loss": 0.01269734650850296, "eval_runtime": 14611.2635, "eval_samples_per_second": 1.371, "eval_steps_per_second": 0.171, "eval_wer": 1.216275863572211, "step": 2000 } ], "logging_steps": 10, "max_steps": 5008, "num_input_tokens_seen": 0, "num_train_epochs": 2, "save_steps": 2000, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 6.531871408128e+19, "train_batch_size": 4, "trial_name": null, "trial_params": null }