{ "best_metric": null, "best_model_checkpoint": null, "epoch": 9.654545454545454, "eval_steps": 30, "global_step": 270, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.07272727272727272, "grad_norm": 0.42279353737831116, "learning_rate": 3.3333333333333333e-06, "loss": 0.4041, "step": 2 }, { "epoch": 0.14545454545454545, "grad_norm": 0.39341622591018677, "learning_rate": 4.999826945767665e-06, "loss": 0.4252, "step": 4 }, { "epoch": 0.21818181818181817, "grad_norm": 0.28936225175857544, "learning_rate": 4.998442655654946e-06, "loss": 0.4065, "step": 6 }, { "epoch": 0.2909090909090909, "grad_norm": 0.1611073613166809, "learning_rate": 4.995674841986217e-06, "loss": 0.3861, "step": 8 }, { "epoch": 0.36363636363636365, "grad_norm": 0.1466752290725708, "learning_rate": 4.991525037450412e-06, "loss": 0.3941, "step": 10 }, { "epoch": 0.43636363636363634, "grad_norm": 0.23374205827713013, "learning_rate": 4.985995540019956e-06, "loss": 0.385, "step": 12 }, { "epoch": 0.509090909090909, "grad_norm": 0.2576417326927185, "learning_rate": 4.979089411678252e-06, "loss": 0.3865, "step": 14 }, { "epoch": 0.5818181818181818, "grad_norm": 0.2165990173816681, "learning_rate": 4.970810476724097e-06, "loss": 0.3914, "step": 16 }, { "epoch": 0.6545454545454545, "grad_norm": 0.13620640337467194, "learning_rate": 4.961163319653959e-06, "loss": 0.3682, "step": 18 }, { "epoch": 0.7272727272727273, "grad_norm": 0.09898896515369415, "learning_rate": 4.950153282623289e-06, "loss": 0.3763, "step": 20 }, { "epoch": 0.8, "grad_norm": 0.08760473132133484, "learning_rate": 4.937786462488284e-06, "loss": 0.3793, "step": 22 }, { "epoch": 0.8727272727272727, "grad_norm": 0.10171981155872345, "learning_rate": 4.9240697074297205e-06, "loss": 0.3823, "step": 24 }, { "epoch": 0.9454545454545454, "grad_norm": 0.10795366764068604, "learning_rate": 4.909010613160751e-06, "loss": 0.3805, "step": 26 }, { "epoch": 1.0, "grad_norm": 0.08594109863042831, "learning_rate": 4.892617518720737e-06, "loss": 0.3662, "step": 28 }, { "epoch": 1.0727272727272728, "grad_norm": 0.08396578580141068, "learning_rate": 4.874899501857477e-06, "loss": 0.368, "step": 30 }, { "epoch": 1.0727272727272728, "eval_loss": 0.40457433462142944, "eval_runtime": 16.4893, "eval_samples_per_second": 6.065, "eval_steps_per_second": 0.425, "step": 30 }, { "epoch": 1.1454545454545455, "grad_norm": 0.06160604581236839, "learning_rate": 4.85586637400036e-06, "loss": 0.3608, "step": 32 }, { "epoch": 1.2181818181818183, "grad_norm": 0.053206440061330795, "learning_rate": 4.8355286748272405e-06, "loss": 0.3423, "step": 34 }, { "epoch": 1.290909090909091, "grad_norm": 0.05600437521934509, "learning_rate": 4.813897666428054e-06, "loss": 0.349, "step": 36 }, { "epoch": 1.3636363636363638, "grad_norm": 0.06277399510145187, "learning_rate": 4.790985327068376e-06, "loss": 0.3569, "step": 38 }, { "epoch": 1.4363636363636363, "grad_norm": 0.0717867836356163, "learning_rate": 4.766804344556414e-06, "loss": 0.359, "step": 40 }, { "epoch": 1.509090909090909, "grad_norm": 0.051569655537605286, "learning_rate": 4.741368109217072e-06, "loss": 0.3489, "step": 42 }, { "epoch": 1.5818181818181818, "grad_norm": 0.04518551751971245, "learning_rate": 4.714690706477e-06, "loss": 0.3382, "step": 44 }, { "epoch": 1.6545454545454545, "grad_norm": 0.0480523444712162, "learning_rate": 4.68678690906473e-06, "loss": 0.3412, "step": 46 }, { "epoch": 1.7272727272727273, "grad_norm": 0.04666323587298393, "learning_rate": 4.657672168830211e-06, "loss": 0.3536, "step": 48 }, { "epoch": 1.8, "grad_norm": 0.05208882689476013, "learning_rate": 4.627362608188281e-06, "loss": 0.3483, "step": 50 }, { "epoch": 1.8727272727272726, "grad_norm": 0.05304299667477608, "learning_rate": 4.5958750111908065e-06, "loss": 0.3408, "step": 52 }, { "epoch": 1.9454545454545453, "grad_norm": 0.04273353889584541, "learning_rate": 4.563226814232444e-06, "loss": 0.3728, "step": 54 }, { "epoch": 2.0, "grad_norm": 0.08124157786369324, "learning_rate": 4.529436096395157e-06, "loss": 0.3783, "step": 56 }, { "epoch": 2.0727272727272728, "grad_norm": 0.044167324900627136, "learning_rate": 4.494521569436845e-06, "loss": 0.3326, "step": 58 }, { "epoch": 2.1454545454545455, "grad_norm": 0.041194308549165726, "learning_rate": 4.4585025674296315e-06, "loss": 0.3302, "step": 60 }, { "epoch": 2.1454545454545455, "eval_loss": 0.4037678837776184, "eval_runtime": 15.9657, "eval_samples_per_second": 6.263, "eval_steps_per_second": 0.438, "step": 60 }, { "epoch": 2.2181818181818183, "grad_norm": 0.050775595009326935, "learning_rate": 4.4213990360535274e-06, "loss": 0.3303, "step": 62 }, { "epoch": 2.290909090909091, "grad_norm": 0.03840049356222153, "learning_rate": 4.383231521551432e-06, "loss": 0.3142, "step": 64 }, { "epoch": 2.3636363636363638, "grad_norm": 0.04364444687962532, "learning_rate": 4.3440211593515556e-06, "loss": 0.3329, "step": 66 }, { "epoch": 2.4363636363636365, "grad_norm": 0.040008459240198135, "learning_rate": 4.303789662363587e-06, "loss": 0.3426, "step": 68 }, { "epoch": 2.509090909090909, "grad_norm": 0.04201684519648552, "learning_rate": 4.262559308955072e-06, "loss": 0.3439, "step": 70 }, { "epoch": 2.581818181818182, "grad_norm": 0.04128064960241318, "learning_rate": 4.220352930614672e-06, "loss": 0.3391, "step": 72 }, { "epoch": 2.6545454545454543, "grad_norm": 0.03866910934448242, "learning_rate": 4.177193899309127e-06, "loss": 0.3299, "step": 74 }, { "epoch": 2.7272727272727275, "grad_norm": 0.0435042642056942, "learning_rate": 4.133106114540923e-06, "loss": 0.323, "step": 76 }, { "epoch": 2.8, "grad_norm": 0.03538796305656433, "learning_rate": 4.088113990113846e-06, "loss": 0.322, "step": 78 }, { "epoch": 2.8727272727272726, "grad_norm": 0.04087506979703903, "learning_rate": 4.042242440613724e-06, "loss": 0.3247, "step": 80 }, { "epoch": 2.9454545454545453, "grad_norm": 0.046281322836875916, "learning_rate": 3.995516867611865e-06, "loss": 0.3438, "step": 82 }, { "epoch": 3.0, "grad_norm": 0.04055408388376236, "learning_rate": 3.947963145598833e-06, "loss": 0.3276, "step": 84 }, { "epoch": 3.0727272727272728, "grad_norm": 0.041513592004776, "learning_rate": 3.899607607656334e-06, "loss": 0.3252, "step": 86 }, { "epoch": 3.1454545454545455, "grad_norm": 0.03839336708188057, "learning_rate": 3.850477030875147e-06, "loss": 0.3115, "step": 88 }, { "epoch": 3.2181818181818183, "grad_norm": 0.04176229238510132, "learning_rate": 3.8005986215272056e-06, "loss": 0.309, "step": 90 }, { "epoch": 3.2181818181818183, "eval_loss": 0.40795081853866577, "eval_runtime": 15.995, "eval_samples_per_second": 6.252, "eval_steps_per_second": 0.438, "step": 90 }, { "epoch": 3.290909090909091, "grad_norm": 0.04411574825644493, "learning_rate": 3.7500000000000005e-06, "loss": 0.3077, "step": 92 }, { "epoch": 3.3636363636363638, "grad_norm": 0.038224343210458755, "learning_rate": 3.6987091855016667e-06, "loss": 0.2897, "step": 94 }, { "epoch": 3.4363636363636365, "grad_norm": 0.048621904104948044, "learning_rate": 3.6467545805452266e-06, "loss": 0.3019, "step": 96 }, { "epoch": 3.509090909090909, "grad_norm": 0.041004959493875504, "learning_rate": 3.594164955220577e-06, "loss": 0.3105, "step": 98 }, { "epoch": 3.581818181818182, "grad_norm": 0.040131937712430954, "learning_rate": 3.5409694312629193e-06, "loss": 0.3201, "step": 100 }, { "epoch": 3.6545454545454543, "grad_norm": 0.04099489003419876, "learning_rate": 3.4871974659264786e-06, "loss": 0.3149, "step": 102 }, { "epoch": 3.7272727272727275, "grad_norm": 0.04340096190571785, "learning_rate": 3.4328788356724135e-06, "loss": 0.3044, "step": 104 }, { "epoch": 3.8, "grad_norm": 0.04188128933310509, "learning_rate": 3.378043619679974e-06, "loss": 0.3103, "step": 106 }, { "epoch": 3.8727272727272726, "grad_norm": 0.037973225116729736, "learning_rate": 3.322722183190025e-06, "loss": 0.2999, "step": 108 }, { "epoch": 3.9454545454545453, "grad_norm": 0.038421131670475006, "learning_rate": 3.26694516069016e-06, "loss": 0.312, "step": 110 }, { "epoch": 4.0, "grad_norm": 0.06082421541213989, "learning_rate": 3.210743438950718e-06, "loss": 0.276, "step": 112 }, { "epoch": 4.072727272727272, "grad_norm": 0.037277910858392715, "learning_rate": 3.154148139921102e-06, "loss": 0.2977, "step": 114 }, { "epoch": 4.1454545454545455, "grad_norm": 0.037613365799188614, "learning_rate": 3.0971906034958616e-06, "loss": 0.3004, "step": 116 }, { "epoch": 4.218181818181818, "grad_norm": 0.03691912069916725, "learning_rate": 3.0399023701600903e-06, "loss": 0.2673, "step": 118 }, { "epoch": 4.290909090909091, "grad_norm": 0.03884879872202873, "learning_rate": 2.9823151635237424e-06, "loss": 0.2901, "step": 120 }, { "epoch": 4.290909090909091, "eval_loss": 0.41295385360717773, "eval_runtime": 15.9766, "eval_samples_per_second": 6.259, "eval_steps_per_second": 0.438, "step": 120 }, { "epoch": 4.363636363636363, "grad_norm": 0.040829263627529144, "learning_rate": 2.924460872754547e-06, "loss": 0.2902, "step": 122 }, { "epoch": 4.4363636363636365, "grad_norm": 0.04225178807973862, "learning_rate": 2.8663715349192388e-06, "loss": 0.2951, "step": 124 }, { "epoch": 4.509090909090909, "grad_norm": 0.036058977246284485, "learning_rate": 2.8080793172428965e-06, "loss": 0.2919, "step": 126 }, { "epoch": 4.581818181818182, "grad_norm": 0.03855273872613907, "learning_rate": 2.7496164992961995e-06, "loss": 0.2897, "step": 128 }, { "epoch": 4.654545454545454, "grad_norm": 0.039659108966588974, "learning_rate": 2.691015455120468e-06, "loss": 0.2932, "step": 130 }, { "epoch": 4.7272727272727275, "grad_norm": 0.042663708329200745, "learning_rate": 2.6323086353004077e-06, "loss": 0.2725, "step": 132 }, { "epoch": 4.8, "grad_norm": 0.03936437889933586, "learning_rate": 2.573528548994449e-06, "loss": 0.2939, "step": 134 }, { "epoch": 4.872727272727273, "grad_norm": 0.0572555810213089, "learning_rate": 2.5147077459326556e-06, "loss": 0.2776, "step": 136 }, { "epoch": 4.945454545454545, "grad_norm": 0.0380295105278492, "learning_rate": 2.455878798392179e-06, "loss": 0.2979, "step": 138 }, { "epoch": 5.0, "grad_norm": 0.035944413393735886, "learning_rate": 2.397074283160206e-06, "loss": 0.2715, "step": 140 }, { "epoch": 5.072727272727272, "grad_norm": 0.0400959774851799, "learning_rate": 2.338326763494429e-06, "loss": 0.2746, "step": 142 }, { "epoch": 5.1454545454545455, "grad_norm": 0.047982338815927505, "learning_rate": 2.2796687710909966e-06, "loss": 0.2705, "step": 144 }, { "epoch": 5.218181818181818, "grad_norm": 0.04013433679938316, "learning_rate": 2.2211327880699392e-06, "loss": 0.2549, "step": 146 }, { "epoch": 5.290909090909091, "grad_norm": 0.0582863911986351, "learning_rate": 2.162751228988063e-06, "loss": 0.2731, "step": 148 }, { "epoch": 5.363636363636363, "grad_norm": 0.03681691363453865, "learning_rate": 2.1045564228892404e-06, "loss": 0.2705, "step": 150 }, { "epoch": 5.363636363636363, "eval_loss": 0.41707590222358704, "eval_runtime": 15.9844, "eval_samples_per_second": 6.256, "eval_steps_per_second": 0.438, "step": 150 }, { "epoch": 5.4363636363636365, "grad_norm": 0.040616922080516815, "learning_rate": 2.04658059540206e-06, "loss": 0.2603, "step": 152 }, { "epoch": 5.509090909090909, "grad_norm": 0.04648638516664505, "learning_rate": 1.9888558508947496e-06, "loss": 0.2665, "step": 154 }, { "epoch": 5.581818181818182, "grad_norm": 0.03560300171375275, "learning_rate": 1.9314141546972345e-06, "loss": 0.2682, "step": 156 }, { "epoch": 5.654545454545454, "grad_norm": 0.03742830082774162, "learning_rate": 1.8742873154002007e-06, "loss": 0.2726, "step": 158 }, { "epoch": 5.7272727272727275, "grad_norm": 0.03804292157292366, "learning_rate": 1.8175069672409476e-06, "loss": 0.2581, "step": 160 }, { "epoch": 5.8, "grad_norm": 0.03460890054702759, "learning_rate": 1.7611045525857902e-06, "loss": 0.2834, "step": 162 }, { "epoch": 5.872727272727273, "grad_norm": 0.038920674473047256, "learning_rate": 1.7051113045187123e-06, "loss": 0.2746, "step": 164 }, { "epoch": 5.945454545454545, "grad_norm": 0.04531797394156456, "learning_rate": 1.6495582295459081e-06, "loss": 0.2666, "step": 166 }, { "epoch": 6.0, "grad_norm": 0.06146521493792534, "learning_rate": 1.5944760904257944e-06, "loss": 0.2675, "step": 168 }, { "epoch": 6.072727272727272, "grad_norm": 0.0393206886947155, "learning_rate": 1.5398953891339972e-06, "loss": 0.2619, "step": 170 }, { "epoch": 6.1454545454545455, "grad_norm": 0.03849950432777405, "learning_rate": 1.485846349972751e-06, "loss": 0.2582, "step": 172 }, { "epoch": 6.218181818181818, "grad_norm": 0.03983594849705696, "learning_rate": 1.4323589028340598e-06, "loss": 0.2736, "step": 174 }, { "epoch": 6.290909090909091, "grad_norm": 0.03905485197901726, "learning_rate": 1.3794626666258868e-06, "loss": 0.2428, "step": 176 }, { "epoch": 6.363636363636363, "grad_norm": 0.04092865809798241, "learning_rate": 1.3271869328705517e-06, "loss": 0.2555, "step": 178 }, { "epoch": 6.4363636363636365, "grad_norm": 0.036916088312864304, "learning_rate": 1.2755606494844294e-06, "loss": 0.2615, "step": 180 }, { "epoch": 6.4363636363636365, "eval_loss": 0.4211080074310303, "eval_runtime": 15.9838, "eval_samples_per_second": 6.256, "eval_steps_per_second": 0.438, "step": 180 }, { "epoch": 6.509090909090909, "grad_norm": 0.03942929953336716, "learning_rate": 1.2246124047479074e-06, "loss": 0.2553, "step": 182 }, { "epoch": 6.581818181818182, "grad_norm": 0.038776326924562454, "learning_rate": 1.174370411474503e-06, "loss": 0.2523, "step": 184 }, { "epoch": 6.654545454545454, "grad_norm": 0.039575546979904175, "learning_rate": 1.1248624913878966e-06, "loss": 0.2616, "step": 186 }, { "epoch": 6.7272727272727275, "grad_norm": 0.03467780724167824, "learning_rate": 1.0761160597155288e-06, "loss": 0.2675, "step": 188 }, { "epoch": 6.8, "grad_norm": 0.036917053163051605, "learning_rate": 1.028158110007294e-06, "loss": 0.268, "step": 190 }, { "epoch": 6.872727272727273, "grad_norm": 0.03764381632208824, "learning_rate": 9.81015199187753e-07, "loss": 0.2618, "step": 192 }, { "epoch": 6.945454545454545, "grad_norm": 0.03786522150039673, "learning_rate": 9.347134328501098e-07, "loss": 0.2468, "step": 194 }, { "epoch": 7.0, "grad_norm": 0.03737010434269905, "learning_rate": 8.892784508001343e-07, "loss": 0.2299, "step": 196 }, { "epoch": 7.072727272727272, "grad_norm": 0.04024997726082802, "learning_rate": 8.44735412857999e-07, "loss": 0.2625, "step": 198 }, { "epoch": 7.1454545454545455, "grad_norm": 0.042676106095314026, "learning_rate": 8.011089849259263e-07, "loss": 0.2615, "step": 200 }, { "epoch": 7.218181818181818, "grad_norm": 0.03763395920395851, "learning_rate": 7.584233253293327e-07, "loss": 0.2476, "step": 202 }, { "epoch": 7.290909090909091, "grad_norm": 0.03564968332648277, "learning_rate": 7.167020714390502e-07, "loss": 0.2453, "step": 204 }, { "epoch": 7.363636363636363, "grad_norm": 0.03487633168697357, "learning_rate": 6.759683265820294e-07, "loss": 0.2557, "step": 206 }, { "epoch": 7.4363636363636365, "grad_norm": 0.038702890276908875, "learning_rate": 6.36244647247774e-07, "loss": 0.2537, "step": 208 }, { "epoch": 7.509090909090909, "grad_norm": 0.03509435057640076, "learning_rate": 5.975530305975808e-07, "loss": 0.2417, "step": 210 }, { "epoch": 7.509090909090909, "eval_loss": 0.4235740005970001, "eval_runtime": 15.9573, "eval_samples_per_second": 6.267, "eval_steps_per_second": 0.439, "step": 210 }, { "epoch": 7.581818181818182, "grad_norm": 0.03575809672474861, "learning_rate": 5.599149022835201e-07, "loss": 0.2521, "step": 212 }, { "epoch": 7.654545454545454, "grad_norm": 0.038031384348869324, "learning_rate": 5.233511045838846e-07, "loss": 0.2645, "step": 214 }, { "epoch": 7.7272727272727275, "grad_norm": 0.03550686687231064, "learning_rate": 4.878818848616861e-07, "loss": 0.2416, "step": 216 }, { "epoch": 7.8, "grad_norm": 0.035883497446775436, "learning_rate": 4.5352688435259084e-07, "loss": 0.2523, "step": 218 }, { "epoch": 7.872727272727273, "grad_norm": 0.039348237216472626, "learning_rate": 4.2030512728849946e-07, "loss": 0.2476, "step": 220 }, { "epoch": 7.945454545454545, "grad_norm": 0.03574049472808838, "learning_rate": 3.882350103627952e-07, "loss": 0.2514, "step": 222 }, { "epoch": 8.0, "grad_norm": 0.05936720594763756, "learning_rate": 3.5733429254309253e-07, "loss": 0.241, "step": 224 }, { "epoch": 8.072727272727272, "grad_norm": 0.03717917203903198, "learning_rate": 3.276200852371339e-07, "loss": 0.2488, "step": 226 }, { "epoch": 8.145454545454545, "grad_norm": 0.03638555109500885, "learning_rate": 2.9910884281727225e-07, "loss": 0.2479, "step": 228 }, { "epoch": 8.218181818181819, "grad_norm": 0.03493395075201988, "learning_rate": 2.7181635350878645e-07, "loss": 0.2398, "step": 230 }, { "epoch": 8.290909090909091, "grad_norm": 0.0358763113617897, "learning_rate": 2.4575773064708904e-07, "loss": 0.2537, "step": 232 }, { "epoch": 8.363636363636363, "grad_norm": 0.03697650507092476, "learning_rate": 2.2094740430864569e-07, "loss": 0.237, "step": 234 }, { "epoch": 8.436363636363636, "grad_norm": 0.03936561197042465, "learning_rate": 1.9739911332025796e-07, "loss": 0.2358, "step": 236 }, { "epoch": 8.50909090909091, "grad_norm": 0.03290196508169174, "learning_rate": 1.7512589765112998e-07, "loss": 0.2338, "step": 238 }, { "epoch": 8.581818181818182, "grad_norm": 0.0346793606877327, "learning_rate": 1.5414009119192635e-07, "loss": 0.2569, "step": 240 }, { "epoch": 8.581818181818182, "eval_loss": 0.42472872138023376, "eval_runtime": 15.9951, "eval_samples_per_second": 6.252, "eval_steps_per_second": 0.438, "step": 240 }, { "epoch": 8.654545454545454, "grad_norm": 0.040124256163835526, "learning_rate": 1.3445331492482617e-07, "loss": 0.2667, "step": 242 }, { "epoch": 8.727272727272727, "grad_norm": 0.03517503663897514, "learning_rate": 1.1607647048835463e-07, "loss": 0.2355, "step": 244 }, { "epoch": 8.8, "grad_norm": 0.038074642419815063, "learning_rate": 9.901973414055188e-08, "loss": 0.2439, "step": 246 }, { "epoch": 8.872727272727273, "grad_norm": 0.03690966218709946, "learning_rate": 8.329255112382666e-08, "loss": 0.2487, "step": 248 }, { "epoch": 8.945454545454545, "grad_norm": 0.03566785156726837, "learning_rate": 6.890363043461051e-08, "loss": 0.252, "step": 250 }, { "epoch": 9.0, "grad_norm": 0.0351359024643898, "learning_rate": 5.5860940000714016e-08, "loss": 0.2637, "step": 252 }, { "epoch": 9.072727272727272, "grad_norm": 0.03571401536464691, "learning_rate": 4.4171702269051874e-08, "loss": 0.2386, "step": 254 }, { "epoch": 9.145454545454545, "grad_norm": 0.04017382860183716, "learning_rate": 3.3842390206180186e-08, "loss": 0.2527, "step": 256 }, { "epoch": 9.218181818181819, "grad_norm": 0.035146769136190414, "learning_rate": 2.487872371386424e-08, "loss": 0.2461, "step": 258 }, { "epoch": 9.290909090909091, "grad_norm": 0.03707621991634369, "learning_rate": 1.728566646165747e-08, "loss": 0.2537, "step": 260 }, { "epoch": 9.363636363636363, "grad_norm": 0.03554265946149826, "learning_rate": 1.1067423138247103e-08, "loss": 0.2353, "step": 262 }, { "epoch": 9.436363636363636, "grad_norm": 0.03387082368135452, "learning_rate": 6.2274371230905405e-09, "loss": 0.2422, "step": 264 }, { "epoch": 9.50909090909091, "grad_norm": 0.03694632276892662, "learning_rate": 2.7683885796273014e-09, "loss": 0.2448, "step": 266 }, { "epoch": 9.581818181818182, "grad_norm": 0.03442605957388878, "learning_rate": 6.921929711287134e-10, "loss": 0.2482, "step": 268 }, { "epoch": 9.654545454545454, "grad_norm": 0.03988894447684288, "learning_rate": 0.0, "loss": 0.2565, "step": 270 }, { "epoch": 9.654545454545454, "eval_loss": 0.42475754022598267, "eval_runtime": 15.9878, "eval_samples_per_second": 6.255, "eval_steps_per_second": 0.438, "step": 270 } ], "logging_steps": 2, "max_steps": 270, "num_input_tokens_seen": 0, "num_train_epochs": 10, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 5.17181948959916e+17, "train_batch_size": 1, "trial_name": null, "trial_params": null }