| { | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 1.0, | |
| "eval_steps": 500, | |
| "global_step": 1036, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.009652509652509652, | |
| "grad_norm": 1.4631391763687134, | |
| "learning_rate": 0.0001, | |
| "loss": 0.664, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.019305019305019305, | |
| "grad_norm": 1.8917441368103027, | |
| "learning_rate": 0.0002, | |
| "loss": 0.1575, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.02895752895752896, | |
| "grad_norm": 1.5227961540222168, | |
| "learning_rate": 0.00019803149606299213, | |
| "loss": 0.1233, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.03861003861003861, | |
| "grad_norm": 0.9390913248062134, | |
| "learning_rate": 0.00019606299212598428, | |
| "loss": 0.1106, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.04826254826254826, | |
| "grad_norm": 0.8157206177711487, | |
| "learning_rate": 0.0001940944881889764, | |
| "loss": 0.0869, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.05791505791505792, | |
| "grad_norm": 0.5291563868522644, | |
| "learning_rate": 0.0001921259842519685, | |
| "loss": 0.0903, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.06756756756756757, | |
| "grad_norm": 1.155928373336792, | |
| "learning_rate": 0.00019015748031496065, | |
| "loss": 0.1186, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 0.07722007722007722, | |
| "grad_norm": 0.5892627239227295, | |
| "learning_rate": 0.00018818897637795277, | |
| "loss": 0.0914, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.08687258687258688, | |
| "grad_norm": 0.5989981889724731, | |
| "learning_rate": 0.0001862204724409449, | |
| "loss": 0.0917, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 0.09652509652509653, | |
| "grad_norm": 0.9171364903450012, | |
| "learning_rate": 0.000184251968503937, | |
| "loss": 0.0888, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.10617760617760617, | |
| "grad_norm": 0.37568220496177673, | |
| "learning_rate": 0.00018228346456692916, | |
| "loss": 0.0865, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 0.11583011583011583, | |
| "grad_norm": 0.5606808662414551, | |
| "learning_rate": 0.00018031496062992125, | |
| "loss": 0.0898, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 0.12548262548262548, | |
| "grad_norm": 0.6872594356536865, | |
| "learning_rate": 0.00017834645669291338, | |
| "loss": 0.0851, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 0.13513513513513514, | |
| "grad_norm": 0.6313247680664062, | |
| "learning_rate": 0.00017637795275590552, | |
| "loss": 0.088, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 0.14478764478764478, | |
| "grad_norm": 0.7841348648071289, | |
| "learning_rate": 0.00017440944881889765, | |
| "loss": 0.0966, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 0.15444015444015444, | |
| "grad_norm": 0.8097236752510071, | |
| "learning_rate": 0.00017244094488188977, | |
| "loss": 0.0626, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 0.1640926640926641, | |
| "grad_norm": 0.7307904362678528, | |
| "learning_rate": 0.00017047244094488192, | |
| "loss": 0.084, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 0.17374517374517376, | |
| "grad_norm": 0.5725339651107788, | |
| "learning_rate": 0.000168503937007874, | |
| "loss": 0.071, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 0.1833976833976834, | |
| "grad_norm": 0.7641273736953735, | |
| "learning_rate": 0.00016653543307086613, | |
| "loss": 0.0958, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 0.19305019305019305, | |
| "grad_norm": 0.6539800763130188, | |
| "learning_rate": 0.00016456692913385828, | |
| "loss": 0.0803, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.20270270270270271, | |
| "grad_norm": 0.6908950805664062, | |
| "learning_rate": 0.0001625984251968504, | |
| "loss": 0.0885, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 0.21235521235521235, | |
| "grad_norm": 0.31019431352615356, | |
| "learning_rate": 0.00016062992125984252, | |
| "loss": 0.0588, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 0.222007722007722, | |
| "grad_norm": 0.7413542866706848, | |
| "learning_rate": 0.00015866141732283467, | |
| "loss": 0.0827, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 0.23166023166023167, | |
| "grad_norm": 0.7324005365371704, | |
| "learning_rate": 0.0001566929133858268, | |
| "loss": 0.0926, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 0.2413127413127413, | |
| "grad_norm": 0.5140088796615601, | |
| "learning_rate": 0.0001547244094488189, | |
| "loss": 0.0768, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 0.25096525096525096, | |
| "grad_norm": 0.3749833405017853, | |
| "learning_rate": 0.00015275590551181104, | |
| "loss": 0.0741, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 0.2606177606177606, | |
| "grad_norm": 0.5479316711425781, | |
| "learning_rate": 0.00015078740157480316, | |
| "loss": 0.068, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 0.2702702702702703, | |
| "grad_norm": 0.5078131556510925, | |
| "learning_rate": 0.00014881889763779528, | |
| "loss": 0.0763, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 0.2799227799227799, | |
| "grad_norm": 0.5565307140350342, | |
| "learning_rate": 0.0001468503937007874, | |
| "loss": 0.0657, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 0.28957528957528955, | |
| "grad_norm": 0.5333523750305176, | |
| "learning_rate": 0.00014488188976377955, | |
| "loss": 0.0467, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.29922779922779924, | |
| "grad_norm": 0.9170491695404053, | |
| "learning_rate": 0.00014291338582677165, | |
| "loss": 0.0677, | |
| "step": 310 | |
| }, | |
| { | |
| "epoch": 0.3088803088803089, | |
| "grad_norm": 0.6521899104118347, | |
| "learning_rate": 0.00014094488188976377, | |
| "loss": 0.0692, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 0.3185328185328185, | |
| "grad_norm": 0.5061705708503723, | |
| "learning_rate": 0.00013897637795275592, | |
| "loss": 0.086, | |
| "step": 330 | |
| }, | |
| { | |
| "epoch": 0.3281853281853282, | |
| "grad_norm": 0.643752932548523, | |
| "learning_rate": 0.00013700787401574804, | |
| "loss": 0.0621, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 0.33783783783783783, | |
| "grad_norm": 0.5780313014984131, | |
| "learning_rate": 0.00013503937007874016, | |
| "loss": 0.0691, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 0.3474903474903475, | |
| "grad_norm": 0.36350390315055847, | |
| "learning_rate": 0.0001330708661417323, | |
| "loss": 0.0679, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 0.35714285714285715, | |
| "grad_norm": 0.6011677384376526, | |
| "learning_rate": 0.0001311023622047244, | |
| "loss": 0.0707, | |
| "step": 370 | |
| }, | |
| { | |
| "epoch": 0.3667953667953668, | |
| "grad_norm": 0.5352160930633545, | |
| "learning_rate": 0.00012913385826771653, | |
| "loss": 0.0543, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 0.3764478764478765, | |
| "grad_norm": 0.4793628752231598, | |
| "learning_rate": 0.00012716535433070867, | |
| "loss": 0.058, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 0.3861003861003861, | |
| "grad_norm": 0.6923650503158569, | |
| "learning_rate": 0.0001251968503937008, | |
| "loss": 0.0691, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 0.39575289575289574, | |
| "grad_norm": 0.49190232157707214, | |
| "learning_rate": 0.00012322834645669292, | |
| "loss": 0.0579, | |
| "step": 410 | |
| }, | |
| { | |
| "epoch": 0.40540540540540543, | |
| "grad_norm": 0.23852688074111938, | |
| "learning_rate": 0.00012125984251968505, | |
| "loss": 0.0524, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 0.41505791505791506, | |
| "grad_norm": 0.6561082005500793, | |
| "learning_rate": 0.00011929133858267719, | |
| "loss": 0.057, | |
| "step": 430 | |
| }, | |
| { | |
| "epoch": 0.4247104247104247, | |
| "grad_norm": 0.613944411277771, | |
| "learning_rate": 0.00011732283464566928, | |
| "loss": 0.0575, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 0.4343629343629344, | |
| "grad_norm": 0.3899982273578644, | |
| "learning_rate": 0.00011535433070866142, | |
| "loss": 0.0527, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 0.444015444015444, | |
| "grad_norm": 0.5278599262237549, | |
| "learning_rate": 0.00011338582677165355, | |
| "loss": 0.0589, | |
| "step": 460 | |
| }, | |
| { | |
| "epoch": 0.45366795366795365, | |
| "grad_norm": 0.29473400115966797, | |
| "learning_rate": 0.00011141732283464567, | |
| "loss": 0.0271, | |
| "step": 470 | |
| }, | |
| { | |
| "epoch": 0.46332046332046334, | |
| "grad_norm": 0.6821677088737488, | |
| "learning_rate": 0.00010944881889763781, | |
| "loss": 0.0772, | |
| "step": 480 | |
| }, | |
| { | |
| "epoch": 0.47297297297297297, | |
| "grad_norm": 0.3420783281326294, | |
| "learning_rate": 0.00010748031496062993, | |
| "loss": 0.0543, | |
| "step": 490 | |
| }, | |
| { | |
| "epoch": 0.4826254826254826, | |
| "grad_norm": 0.5966827869415283, | |
| "learning_rate": 0.00010551181102362204, | |
| "loss": 0.0552, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.4922779922779923, | |
| "grad_norm": 0.656173825263977, | |
| "learning_rate": 0.00010354330708661417, | |
| "loss": 0.0692, | |
| "step": 510 | |
| }, | |
| { | |
| "epoch": 0.5019305019305019, | |
| "grad_norm": 0.46832337975502014, | |
| "learning_rate": 0.0001015748031496063, | |
| "loss": 0.0456, | |
| "step": 520 | |
| }, | |
| { | |
| "epoch": 0.5115830115830116, | |
| "grad_norm": 0.5552840828895569, | |
| "learning_rate": 9.960629921259843e-05, | |
| "loss": 0.0581, | |
| "step": 530 | |
| }, | |
| { | |
| "epoch": 0.5212355212355212, | |
| "grad_norm": 0.7662914395332336, | |
| "learning_rate": 9.763779527559055e-05, | |
| "loss": 0.0542, | |
| "step": 540 | |
| }, | |
| { | |
| "epoch": 0.5308880308880309, | |
| "grad_norm": 0.5977205634117126, | |
| "learning_rate": 9.566929133858268e-05, | |
| "loss": 0.0432, | |
| "step": 550 | |
| }, | |
| { | |
| "epoch": 0.5405405405405406, | |
| "grad_norm": 0.41738152503967285, | |
| "learning_rate": 9.370078740157481e-05, | |
| "loss": 0.0606, | |
| "step": 560 | |
| }, | |
| { | |
| "epoch": 0.5501930501930502, | |
| "grad_norm": 0.3176082670688629, | |
| "learning_rate": 9.173228346456693e-05, | |
| "loss": 0.053, | |
| "step": 570 | |
| }, | |
| { | |
| "epoch": 0.5598455598455598, | |
| "grad_norm": 0.5319856405258179, | |
| "learning_rate": 8.976377952755905e-05, | |
| "loss": 0.0548, | |
| "step": 580 | |
| }, | |
| { | |
| "epoch": 0.5694980694980695, | |
| "grad_norm": 0.6625571250915527, | |
| "learning_rate": 8.779527559055119e-05, | |
| "loss": 0.0735, | |
| "step": 590 | |
| }, | |
| { | |
| "epoch": 0.5791505791505791, | |
| "grad_norm": 0.6264726519584656, | |
| "learning_rate": 8.582677165354331e-05, | |
| "loss": 0.0574, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 0.5888030888030888, | |
| "grad_norm": 0.49182403087615967, | |
| "learning_rate": 8.385826771653543e-05, | |
| "loss": 0.046, | |
| "step": 610 | |
| }, | |
| { | |
| "epoch": 0.5984555984555985, | |
| "grad_norm": 0.712940514087677, | |
| "learning_rate": 8.188976377952757e-05, | |
| "loss": 0.0451, | |
| "step": 620 | |
| }, | |
| { | |
| "epoch": 0.6081081081081081, | |
| "grad_norm": 0.5095399022102356, | |
| "learning_rate": 7.992125984251969e-05, | |
| "loss": 0.0527, | |
| "step": 630 | |
| }, | |
| { | |
| "epoch": 0.6177606177606177, | |
| "grad_norm": 0.4520854949951172, | |
| "learning_rate": 7.795275590551181e-05, | |
| "loss": 0.0498, | |
| "step": 640 | |
| }, | |
| { | |
| "epoch": 0.6274131274131274, | |
| "grad_norm": 0.7156594395637512, | |
| "learning_rate": 7.598425196850393e-05, | |
| "loss": 0.0482, | |
| "step": 650 | |
| }, | |
| { | |
| "epoch": 0.637065637065637, | |
| "grad_norm": 0.3221661150455475, | |
| "learning_rate": 7.401574803149607e-05, | |
| "loss": 0.0399, | |
| "step": 660 | |
| }, | |
| { | |
| "epoch": 0.6467181467181468, | |
| "grad_norm": 0.22723270952701569, | |
| "learning_rate": 7.20472440944882e-05, | |
| "loss": 0.0388, | |
| "step": 670 | |
| }, | |
| { | |
| "epoch": 0.6563706563706564, | |
| "grad_norm": 0.5212529301643372, | |
| "learning_rate": 7.007874015748031e-05, | |
| "loss": 0.0389, | |
| "step": 680 | |
| }, | |
| { | |
| "epoch": 0.666023166023166, | |
| "grad_norm": 0.5126072764396667, | |
| "learning_rate": 6.811023622047245e-05, | |
| "loss": 0.0525, | |
| "step": 690 | |
| }, | |
| { | |
| "epoch": 0.6756756756756757, | |
| "grad_norm": 0.7248953580856323, | |
| "learning_rate": 6.614173228346457e-05, | |
| "loss": 0.0619, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 0.6853281853281853, | |
| "grad_norm": 0.3085402548313141, | |
| "learning_rate": 6.417322834645669e-05, | |
| "loss": 0.053, | |
| "step": 710 | |
| }, | |
| { | |
| "epoch": 0.694980694980695, | |
| "grad_norm": 0.3684390187263489, | |
| "learning_rate": 6.220472440944882e-05, | |
| "loss": 0.0352, | |
| "step": 720 | |
| }, | |
| { | |
| "epoch": 0.7046332046332047, | |
| "grad_norm": 0.5736101269721985, | |
| "learning_rate": 6.0236220472440953e-05, | |
| "loss": 0.0524, | |
| "step": 730 | |
| }, | |
| { | |
| "epoch": 0.7142857142857143, | |
| "grad_norm": 0.24105365574359894, | |
| "learning_rate": 5.826771653543307e-05, | |
| "loss": 0.0361, | |
| "step": 740 | |
| }, | |
| { | |
| "epoch": 0.7239382239382239, | |
| "grad_norm": 0.4142036736011505, | |
| "learning_rate": 5.62992125984252e-05, | |
| "loss": 0.0423, | |
| "step": 750 | |
| }, | |
| { | |
| "epoch": 0.7335907335907336, | |
| "grad_norm": 0.43454453349113464, | |
| "learning_rate": 5.433070866141733e-05, | |
| "loss": 0.0573, | |
| "step": 760 | |
| }, | |
| { | |
| "epoch": 0.7432432432432432, | |
| "grad_norm": 0.6893749833106995, | |
| "learning_rate": 5.236220472440945e-05, | |
| "loss": 0.0578, | |
| "step": 770 | |
| }, | |
| { | |
| "epoch": 0.752895752895753, | |
| "grad_norm": 0.38158169388771057, | |
| "learning_rate": 5.0393700787401575e-05, | |
| "loss": 0.042, | |
| "step": 780 | |
| }, | |
| { | |
| "epoch": 0.7625482625482626, | |
| "grad_norm": 0.6105143427848816, | |
| "learning_rate": 4.84251968503937e-05, | |
| "loss": 0.0632, | |
| "step": 790 | |
| }, | |
| { | |
| "epoch": 0.7722007722007722, | |
| "grad_norm": 0.28586989641189575, | |
| "learning_rate": 4.645669291338583e-05, | |
| "loss": 0.0453, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 0.7818532818532818, | |
| "grad_norm": 0.4869031012058258, | |
| "learning_rate": 4.4488188976377954e-05, | |
| "loss": 0.0426, | |
| "step": 810 | |
| }, | |
| { | |
| "epoch": 0.7915057915057915, | |
| "grad_norm": 0.09411308914422989, | |
| "learning_rate": 4.251968503937008e-05, | |
| "loss": 0.034, | |
| "step": 820 | |
| }, | |
| { | |
| "epoch": 0.8011583011583011, | |
| "grad_norm": 0.43962639570236206, | |
| "learning_rate": 4.0551181102362204e-05, | |
| "loss": 0.0528, | |
| "step": 830 | |
| }, | |
| { | |
| "epoch": 0.8108108108108109, | |
| "grad_norm": 0.33696189522743225, | |
| "learning_rate": 3.858267716535433e-05, | |
| "loss": 0.0331, | |
| "step": 840 | |
| }, | |
| { | |
| "epoch": 0.8204633204633205, | |
| "grad_norm": 0.4021511673927307, | |
| "learning_rate": 3.661417322834646e-05, | |
| "loss": 0.0406, | |
| "step": 850 | |
| }, | |
| { | |
| "epoch": 0.8301158301158301, | |
| "grad_norm": 0.6140969395637512, | |
| "learning_rate": 3.464566929133858e-05, | |
| "loss": 0.0466, | |
| "step": 860 | |
| }, | |
| { | |
| "epoch": 0.8397683397683398, | |
| "grad_norm": 0.4614850878715515, | |
| "learning_rate": 3.2677165354330704e-05, | |
| "loss": 0.0285, | |
| "step": 870 | |
| }, | |
| { | |
| "epoch": 0.8494208494208494, | |
| "grad_norm": 0.6398894786834717, | |
| "learning_rate": 3.070866141732284e-05, | |
| "loss": 0.0515, | |
| "step": 880 | |
| }, | |
| { | |
| "epoch": 0.859073359073359, | |
| "grad_norm": 0.3512854278087616, | |
| "learning_rate": 2.874015748031496e-05, | |
| "loss": 0.0336, | |
| "step": 890 | |
| }, | |
| { | |
| "epoch": 0.8687258687258688, | |
| "grad_norm": 0.2678048610687256, | |
| "learning_rate": 2.677165354330709e-05, | |
| "loss": 0.0286, | |
| "step": 900 | |
| }, | |
| { | |
| "epoch": 0.8783783783783784, | |
| "grad_norm": 0.5848315358161926, | |
| "learning_rate": 2.4803149606299215e-05, | |
| "loss": 0.0384, | |
| "step": 910 | |
| }, | |
| { | |
| "epoch": 0.888030888030888, | |
| "grad_norm": 0.2881620228290558, | |
| "learning_rate": 2.283464566929134e-05, | |
| "loss": 0.0424, | |
| "step": 920 | |
| }, | |
| { | |
| "epoch": 0.8976833976833977, | |
| "grad_norm": 0.3219210207462311, | |
| "learning_rate": 2.0866141732283465e-05, | |
| "loss": 0.0488, | |
| "step": 930 | |
| }, | |
| { | |
| "epoch": 0.9073359073359073, | |
| "grad_norm": 0.408877432346344, | |
| "learning_rate": 1.889763779527559e-05, | |
| "loss": 0.0345, | |
| "step": 940 | |
| }, | |
| { | |
| "epoch": 0.916988416988417, | |
| "grad_norm": 0.20342448353767395, | |
| "learning_rate": 1.692913385826772e-05, | |
| "loss": 0.049, | |
| "step": 950 | |
| }, | |
| { | |
| "epoch": 0.9266409266409267, | |
| "grad_norm": 0.4167528748512268, | |
| "learning_rate": 1.4960629921259845e-05, | |
| "loss": 0.043, | |
| "step": 960 | |
| }, | |
| { | |
| "epoch": 0.9362934362934363, | |
| "grad_norm": 0.49826258420944214, | |
| "learning_rate": 1.2992125984251968e-05, | |
| "loss": 0.0412, | |
| "step": 970 | |
| }, | |
| { | |
| "epoch": 0.9459459459459459, | |
| "grad_norm": 0.5426783561706543, | |
| "learning_rate": 1.1023622047244095e-05, | |
| "loss": 0.0538, | |
| "step": 980 | |
| }, | |
| { | |
| "epoch": 0.9555984555984556, | |
| "grad_norm": 0.170461505651474, | |
| "learning_rate": 9.055118110236222e-06, | |
| "loss": 0.0483, | |
| "step": 990 | |
| }, | |
| { | |
| "epoch": 0.9652509652509652, | |
| "grad_norm": 0.3618116080760956, | |
| "learning_rate": 7.086614173228347e-06, | |
| "loss": 0.038, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 0.974903474903475, | |
| "grad_norm": 0.417107492685318, | |
| "learning_rate": 5.118110236220473e-06, | |
| "loss": 0.0373, | |
| "step": 1010 | |
| }, | |
| { | |
| "epoch": 0.9845559845559846, | |
| "grad_norm": 0.4384624660015106, | |
| "learning_rate": 3.1496062992125985e-06, | |
| "loss": 0.0534, | |
| "step": 1020 | |
| }, | |
| { | |
| "epoch": 0.9942084942084942, | |
| "grad_norm": 0.7793737649917603, | |
| "learning_rate": 1.1811023622047244e-06, | |
| "loss": 0.0406, | |
| "step": 1030 | |
| } | |
| ], | |
| "logging_steps": 10, | |
| "max_steps": 1036, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 1, | |
| "save_steps": 50, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": true | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 1.0789424666431488e+16, | |
| "train_batch_size": 2, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |