| { | |
| "best_global_step": null, | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 6.0, | |
| "eval_steps": 500, | |
| "global_step": 3426, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.035056967572304996, | |
| "grad_norm": 12.451324462890625, | |
| "learning_rate": 1.8446601941747574e-06, | |
| "loss": 3.237, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.07011393514460999, | |
| "grad_norm": 4.131791114807129, | |
| "learning_rate": 3.7864077669902915e-06, | |
| "loss": 2.5266, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.10517090271691498, | |
| "grad_norm": 2.2451958656311035, | |
| "learning_rate": 5.728155339805825e-06, | |
| "loss": 1.8755, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.14022787028921999, | |
| "grad_norm": 1.4421552419662476, | |
| "learning_rate": 7.66990291262136e-06, | |
| "loss": 1.3649, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.175284837861525, | |
| "grad_norm": 0.9530179500579834, | |
| "learning_rate": 9.611650485436894e-06, | |
| "loss": 1.0674, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.21034180543382996, | |
| "grad_norm": 0.7194477319717407, | |
| "learning_rate": 9.99942798060303e-06, | |
| "loss": 0.9241, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 0.24539877300613497, | |
| "grad_norm": 0.6556061506271362, | |
| "learning_rate": 9.997104376116195e-06, | |
| "loss": 0.8575, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 0.28045574057843997, | |
| "grad_norm": 0.5718048810958862, | |
| "learning_rate": 9.992994265395959e-06, | |
| "loss": 0.829, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 0.31551270815074495, | |
| "grad_norm": 0.4922148287296295, | |
| "learning_rate": 9.987099117840969e-06, | |
| "loss": 0.8034, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 0.35056967572305, | |
| "grad_norm": 0.47302234172821045, | |
| "learning_rate": 9.979421041015336e-06, | |
| "loss": 0.7839, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.38562664329535495, | |
| "grad_norm": 0.49009189009666443, | |
| "learning_rate": 9.969962779895172e-06, | |
| "loss": 0.768, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 0.42068361086765993, | |
| "grad_norm": 0.4963654577732086, | |
| "learning_rate": 9.958727715887218e-06, | |
| "loss": 0.7628, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 0.45574057843996496, | |
| "grad_norm": 0.5206854343414307, | |
| "learning_rate": 9.94571986561998e-06, | |
| "loss": 0.7488, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 0.49079754601226994, | |
| "grad_norm": 0.48924869298934937, | |
| "learning_rate": 9.930943879507748e-06, | |
| "loss": 0.7436, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 0.5258545135845749, | |
| "grad_norm": 0.43540337681770325, | |
| "learning_rate": 9.914405040088026e-06, | |
| "loss": 0.7375, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.5609114811568799, | |
| "grad_norm": 0.44258421659469604, | |
| "learning_rate": 9.896109260132993e-06, | |
| "loss": 0.7277, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 0.595968448729185, | |
| "grad_norm": 0.4955386519432068, | |
| "learning_rate": 9.876063080535627e-06, | |
| "loss": 0.7284, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 0.6310254163014899, | |
| "grad_norm": 0.5027541518211365, | |
| "learning_rate": 9.85427366797129e-06, | |
| "loss": 0.7231, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 0.6660823838737949, | |
| "grad_norm": 0.4675957262516022, | |
| "learning_rate": 9.830748812335576e-06, | |
| "loss": 0.7212, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 0.7011393514461, | |
| "grad_norm": 0.4283595383167267, | |
| "learning_rate": 9.805496923959363e-06, | |
| "loss": 0.7164, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 0.7361963190184049, | |
| "grad_norm": 0.452084481716156, | |
| "learning_rate": 9.778527030602049e-06, | |
| "loss": 0.711, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 0.7712532865907099, | |
| "grad_norm": 0.4737929105758667, | |
| "learning_rate": 9.74984877422405e-06, | |
| "loss": 0.7084, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 0.8063102541630149, | |
| "grad_norm": 0.4964485466480255, | |
| "learning_rate": 9.719472407539725e-06, | |
| "loss": 0.7028, | |
| "step": 460 | |
| }, | |
| { | |
| "epoch": 0.8413672217353199, | |
| "grad_norm": 0.44363030791282654, | |
| "learning_rate": 9.68740879035194e-06, | |
| "loss": 0.7045, | |
| "step": 480 | |
| }, | |
| { | |
| "epoch": 0.8764241893076249, | |
| "grad_norm": 0.5004998445510864, | |
| "learning_rate": 9.6536693856696e-06, | |
| "loss": 0.6937, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.9114811568799299, | |
| "grad_norm": 0.4564264118671417, | |
| "learning_rate": 9.618266255609533e-06, | |
| "loss": 0.699, | |
| "step": 520 | |
| }, | |
| { | |
| "epoch": 0.9465381244522348, | |
| "grad_norm": 0.4558616280555725, | |
| "learning_rate": 9.58121205708418e-06, | |
| "loss": 0.691, | |
| "step": 540 | |
| }, | |
| { | |
| "epoch": 0.9815950920245399, | |
| "grad_norm": 0.413114458322525, | |
| "learning_rate": 9.542520037276636e-06, | |
| "loss": 0.6891, | |
| "step": 560 | |
| }, | |
| { | |
| "epoch": 1.0157756354075373, | |
| "grad_norm": 0.403679758310318, | |
| "learning_rate": 9.502204028904687e-06, | |
| "loss": 0.6812, | |
| "step": 580 | |
| }, | |
| { | |
| "epoch": 1.0508326029798423, | |
| "grad_norm": 0.40308722853660583, | |
| "learning_rate": 9.46027844527549e-06, | |
| "loss": 0.6791, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 1.0858895705521472, | |
| "grad_norm": 0.4085083603858948, | |
| "learning_rate": 9.416758275132693e-06, | |
| "loss": 0.6803, | |
| "step": 620 | |
| }, | |
| { | |
| "epoch": 1.1209465381244523, | |
| "grad_norm": 0.4475920796394348, | |
| "learning_rate": 9.371659077297843e-06, | |
| "loss": 0.6789, | |
| "step": 640 | |
| }, | |
| { | |
| "epoch": 1.1560035056967572, | |
| "grad_norm": 0.4604188799858093, | |
| "learning_rate": 9.324996975107978e-06, | |
| "loss": 0.674, | |
| "step": 660 | |
| }, | |
| { | |
| "epoch": 1.1910604732690622, | |
| "grad_norm": 0.4190482795238495, | |
| "learning_rate": 9.276788650651392e-06, | |
| "loss": 0.6746, | |
| "step": 680 | |
| }, | |
| { | |
| "epoch": 1.2261174408413673, | |
| "grad_norm": 0.420953631401062, | |
| "learning_rate": 9.227051338803656e-06, | |
| "loss": 0.6692, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 1.2611744084136722, | |
| "grad_norm": 0.4463854432106018, | |
| "learning_rate": 9.175802821066009e-06, | |
| "loss": 0.6737, | |
| "step": 720 | |
| }, | |
| { | |
| "epoch": 1.2962313759859772, | |
| "grad_norm": 0.44004735350608826, | |
| "learning_rate": 9.12306141920832e-06, | |
| "loss": 0.6673, | |
| "step": 740 | |
| }, | |
| { | |
| "epoch": 1.331288343558282, | |
| "grad_norm": 0.42015475034713745, | |
| "learning_rate": 9.068845988718906e-06, | |
| "loss": 0.6676, | |
| "step": 760 | |
| }, | |
| { | |
| "epoch": 1.3663453111305872, | |
| "grad_norm": 0.43683475255966187, | |
| "learning_rate": 9.013175912063534e-06, | |
| "loss": 0.6649, | |
| "step": 780 | |
| }, | |
| { | |
| "epoch": 1.4014022787028921, | |
| "grad_norm": 0.4281805753707886, | |
| "learning_rate": 8.956071091756036e-06, | |
| "loss": 0.6658, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 1.4364592462751973, | |
| "grad_norm": 0.4270734190940857, | |
| "learning_rate": 8.89755194324299e-06, | |
| "loss": 0.6646, | |
| "step": 820 | |
| }, | |
| { | |
| "epoch": 1.4715162138475022, | |
| "grad_norm": 0.4163481593132019, | |
| "learning_rate": 8.837639387605031e-06, | |
| "loss": 0.6658, | |
| "step": 840 | |
| }, | |
| { | |
| "epoch": 1.5065731814198071, | |
| "grad_norm": 0.45280900597572327, | |
| "learning_rate": 8.776354844077389e-06, | |
| "loss": 0.6592, | |
| "step": 860 | |
| }, | |
| { | |
| "epoch": 1.541630148992112, | |
| "grad_norm": 0.40485361218452454, | |
| "learning_rate": 8.713720222392338e-06, | |
| "loss": 0.6579, | |
| "step": 880 | |
| }, | |
| { | |
| "epoch": 1.5766871165644172, | |
| "grad_norm": 0.42039763927459717, | |
| "learning_rate": 8.649757914946284e-06, | |
| "loss": 0.6616, | |
| "step": 900 | |
| }, | |
| { | |
| "epoch": 1.6117440841367223, | |
| "grad_norm": 0.4760454595088959, | |
| "learning_rate": 8.584490788794296e-06, | |
| "loss": 0.6572, | |
| "step": 920 | |
| }, | |
| { | |
| "epoch": 1.6468010517090272, | |
| "grad_norm": 0.43802690505981445, | |
| "learning_rate": 8.517942177474943e-06, | |
| "loss": 0.6548, | |
| "step": 940 | |
| }, | |
| { | |
| "epoch": 1.6818580192813322, | |
| "grad_norm": 0.5002708435058594, | |
| "learning_rate": 8.450135872668369e-06, | |
| "loss": 0.6557, | |
| "step": 960 | |
| }, | |
| { | |
| "epoch": 1.716914986853637, | |
| "grad_norm": 0.4160609543323517, | |
| "learning_rate": 8.38109611569056e-06, | |
| "loss": 0.6529, | |
| "step": 980 | |
| }, | |
| { | |
| "epoch": 1.751971954425942, | |
| "grad_norm": 0.43179649114608765, | |
| "learning_rate": 8.310847588826876e-06, | |
| "loss": 0.6529, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 1.7870289219982471, | |
| "grad_norm": 0.4322780668735504, | |
| "learning_rate": 8.239415406507934e-06, | |
| "loss": 0.6535, | |
| "step": 1020 | |
| }, | |
| { | |
| "epoch": 1.8220858895705523, | |
| "grad_norm": 0.4642186462879181, | |
| "learning_rate": 8.166825106330985e-06, | |
| "loss": 0.649, | |
| "step": 1040 | |
| }, | |
| { | |
| "epoch": 1.8571428571428572, | |
| "grad_norm": 0.42697349190711975, | |
| "learning_rate": 8.093102639930013e-06, | |
| "loss": 0.65, | |
| "step": 1060 | |
| }, | |
| { | |
| "epoch": 1.8921998247151621, | |
| "grad_norm": 0.4486387372016907, | |
| "learning_rate": 8.01827436369781e-06, | |
| "loss": 0.6492, | |
| "step": 1080 | |
| }, | |
| { | |
| "epoch": 1.927256792287467, | |
| "grad_norm": 0.42962825298309326, | |
| "learning_rate": 7.942367029363351e-06, | |
| "loss": 0.6518, | |
| "step": 1100 | |
| }, | |
| { | |
| "epoch": 1.962313759859772, | |
| "grad_norm": 0.43645408749580383, | |
| "learning_rate": 7.865407774427828e-06, | |
| "loss": 0.6475, | |
| "step": 1120 | |
| }, | |
| { | |
| "epoch": 1.997370727432077, | |
| "grad_norm": 0.4662039875984192, | |
| "learning_rate": 7.787424112462758e-06, | |
| "loss": 0.649, | |
| "step": 1140 | |
| }, | |
| { | |
| "epoch": 2.0315512708150747, | |
| "grad_norm": 0.44152551889419556, | |
| "learning_rate": 7.708443923273671e-06, | |
| "loss": 0.6401, | |
| "step": 1160 | |
| }, | |
| { | |
| "epoch": 2.0666082383873796, | |
| "grad_norm": 0.39342495799064636, | |
| "learning_rate": 7.628495442932838e-06, | |
| "loss": 0.6386, | |
| "step": 1180 | |
| }, | |
| { | |
| "epoch": 2.1016652059596845, | |
| "grad_norm": 0.4356766641139984, | |
| "learning_rate": 7.54760725368464e-06, | |
| "loss": 0.6391, | |
| "step": 1200 | |
| }, | |
| { | |
| "epoch": 2.1367221735319895, | |
| "grad_norm": 0.3897708058357239, | |
| "learning_rate": 7.465808273727182e-06, | |
| "loss": 0.6383, | |
| "step": 1220 | |
| }, | |
| { | |
| "epoch": 2.1717791411042944, | |
| "grad_norm": 0.4168529510498047, | |
| "learning_rate": 7.383127746873796e-06, | |
| "loss": 0.6361, | |
| "step": 1240 | |
| }, | |
| { | |
| "epoch": 2.2068361086765993, | |
| "grad_norm": 0.39462465047836304, | |
| "learning_rate": 7.2995952320981356e-06, | |
| "loss": 0.6371, | |
| "step": 1260 | |
| }, | |
| { | |
| "epoch": 2.2418930762489047, | |
| "grad_norm": 0.42870041728019714, | |
| "learning_rate": 7.215240592966603e-06, | |
| "loss": 0.6318, | |
| "step": 1280 | |
| }, | |
| { | |
| "epoch": 2.2769500438212096, | |
| "grad_norm": 0.46848800778388977, | |
| "learning_rate": 7.130093986961868e-06, | |
| "loss": 0.633, | |
| "step": 1300 | |
| }, | |
| { | |
| "epoch": 2.3120070113935145, | |
| "grad_norm": 0.415912002325058, | |
| "learning_rate": 7.044185854701321e-06, | |
| "loss": 0.6367, | |
| "step": 1320 | |
| }, | |
| { | |
| "epoch": 2.3470639789658194, | |
| "grad_norm": 0.4347931444644928, | |
| "learning_rate": 6.957546909054304e-06, | |
| "loss": 0.6374, | |
| "step": 1340 | |
| }, | |
| { | |
| "epoch": 2.3821209465381243, | |
| "grad_norm": 0.4282444417476654, | |
| "learning_rate": 6.870208124161998e-06, | |
| "loss": 0.6353, | |
| "step": 1360 | |
| }, | |
| { | |
| "epoch": 2.4171779141104293, | |
| "grad_norm": 0.43224233388900757, | |
| "learning_rate": 6.78220072436392e-06, | |
| "loss": 0.6348, | |
| "step": 1380 | |
| }, | |
| { | |
| "epoch": 2.4522348816827346, | |
| "grad_norm": 0.4176190495491028, | |
| "learning_rate": 6.693556173034953e-06, | |
| "loss": 0.633, | |
| "step": 1400 | |
| }, | |
| { | |
| "epoch": 2.4872918492550395, | |
| "grad_norm": 0.4181615710258484, | |
| "learning_rate": 6.6043061613369356e-06, | |
| "loss": 0.6324, | |
| "step": 1420 | |
| }, | |
| { | |
| "epoch": 2.5223488168273445, | |
| "grad_norm": 0.38148173689842224, | |
| "learning_rate": 6.514482596888807e-06, | |
| "loss": 0.6301, | |
| "step": 1440 | |
| }, | |
| { | |
| "epoch": 2.5574057843996494, | |
| "grad_norm": 0.41031816601753235, | |
| "learning_rate": 6.424117592359367e-06, | |
| "loss": 0.6332, | |
| "step": 1460 | |
| }, | |
| { | |
| "epoch": 2.5924627519719543, | |
| "grad_norm": 0.4005562961101532, | |
| "learning_rate": 6.333243453986734e-06, | |
| "loss": 0.6265, | |
| "step": 1480 | |
| }, | |
| { | |
| "epoch": 2.6275197195442592, | |
| "grad_norm": 0.3940238058567047, | |
| "learning_rate": 6.241892670028595e-06, | |
| "loss": 0.6315, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 2.662576687116564, | |
| "grad_norm": 0.4001730680465698, | |
| "learning_rate": 6.150097899147384e-06, | |
| "loss": 0.6299, | |
| "step": 1520 | |
| }, | |
| { | |
| "epoch": 2.6976336546888695, | |
| "grad_norm": 0.3857872188091278, | |
| "learning_rate": 6.057891958734538e-06, | |
| "loss": 0.6304, | |
| "step": 1540 | |
| }, | |
| { | |
| "epoch": 2.7326906222611744, | |
| "grad_norm": 0.4169263243675232, | |
| "learning_rate": 5.965307813178015e-06, | |
| "loss": 0.6315, | |
| "step": 1560 | |
| }, | |
| { | |
| "epoch": 2.7677475898334793, | |
| "grad_norm": 0.4010975658893585, | |
| "learning_rate": 5.872378562077241e-06, | |
| "loss": 0.6297, | |
| "step": 1580 | |
| }, | |
| { | |
| "epoch": 2.8028045574057843, | |
| "grad_norm": 0.4302142262458801, | |
| "learning_rate": 5.779137428409738e-06, | |
| "loss": 0.6302, | |
| "step": 1600 | |
| }, | |
| { | |
| "epoch": 2.8378615249780896, | |
| "grad_norm": 0.39000585675239563, | |
| "learning_rate": 5.685617746653629e-06, | |
| "loss": 0.6312, | |
| "step": 1620 | |
| }, | |
| { | |
| "epoch": 2.8729184925503946, | |
| "grad_norm": 0.4292212724685669, | |
| "learning_rate": 5.591852950870287e-06, | |
| "loss": 0.6312, | |
| "step": 1640 | |
| }, | |
| { | |
| "epoch": 2.9079754601226995, | |
| "grad_norm": 0.3838886320590973, | |
| "learning_rate": 5.497876562751384e-06, | |
| "loss": 0.6302, | |
| "step": 1660 | |
| }, | |
| { | |
| "epoch": 2.9430324276950044, | |
| "grad_norm": 0.36835259199142456, | |
| "learning_rate": 5.403722179634602e-06, | |
| "loss": 0.6292, | |
| "step": 1680 | |
| }, | |
| { | |
| "epoch": 2.9780893952673093, | |
| "grad_norm": 0.3884848654270172, | |
| "learning_rate": 5.309423462492314e-06, | |
| "loss": 0.6261, | |
| "step": 1700 | |
| }, | |
| { | |
| "epoch": 3.0122699386503067, | |
| "grad_norm": 0.3762246072292328, | |
| "learning_rate": 5.215014123897504e-06, | |
| "loss": 0.6202, | |
| "step": 1720 | |
| }, | |
| { | |
| "epoch": 3.0473269062226116, | |
| "grad_norm": 0.38138872385025024, | |
| "learning_rate": 5.120527915971235e-06, | |
| "loss": 0.6205, | |
| "step": 1740 | |
| }, | |
| { | |
| "epoch": 3.0823838737949165, | |
| "grad_norm": 0.38698920607566833, | |
| "learning_rate": 5.0259986183160006e-06, | |
| "loss": 0.6186, | |
| "step": 1760 | |
| }, | |
| { | |
| "epoch": 3.117440841367222, | |
| "grad_norm": 0.378830224275589, | |
| "learning_rate": 4.931460025939226e-06, | |
| "loss": 0.6214, | |
| "step": 1780 | |
| }, | |
| { | |
| "epoch": 3.152497808939527, | |
| "grad_norm": 0.3751004934310913, | |
| "learning_rate": 4.836945937171279e-06, | |
| "loss": 0.6209, | |
| "step": 1800 | |
| }, | |
| { | |
| "epoch": 3.1875547765118317, | |
| "grad_norm": 0.3829745054244995, | |
| "learning_rate": 4.742490141582279e-06, | |
| "loss": 0.6213, | |
| "step": 1820 | |
| }, | |
| { | |
| "epoch": 3.2226117440841366, | |
| "grad_norm": 0.4228389859199524, | |
| "learning_rate": 4.648126407902058e-06, | |
| "loss": 0.6193, | |
| "step": 1840 | |
| }, | |
| { | |
| "epoch": 3.2576687116564416, | |
| "grad_norm": 0.367960125207901, | |
| "learning_rate": 4.553888471947546e-06, | |
| "loss": 0.6198, | |
| "step": 1860 | |
| }, | |
| { | |
| "epoch": 3.292725679228747, | |
| "grad_norm": 0.39815646409988403, | |
| "learning_rate": 4.4598100245619505e-06, | |
| "loss": 0.6187, | |
| "step": 1880 | |
| }, | |
| { | |
| "epoch": 3.327782646801052, | |
| "grad_norm": 0.3625248074531555, | |
| "learning_rate": 4.3659246995699845e-06, | |
| "loss": 0.6176, | |
| "step": 1900 | |
| }, | |
| { | |
| "epoch": 3.3628396143733568, | |
| "grad_norm": 0.37671083211898804, | |
| "learning_rate": 4.2722660617535105e-06, | |
| "loss": 0.6182, | |
| "step": 1920 | |
| }, | |
| { | |
| "epoch": 3.3978965819456617, | |
| "grad_norm": 0.3727245032787323, | |
| "learning_rate": 4.178867594851849e-06, | |
| "loss": 0.616, | |
| "step": 1940 | |
| }, | |
| { | |
| "epoch": 3.4329535495179666, | |
| "grad_norm": 0.361914724111557, | |
| "learning_rate": 4.085762689591054e-06, | |
| "loss": 0.6157, | |
| "step": 1960 | |
| }, | |
| { | |
| "epoch": 3.4680105170902715, | |
| "grad_norm": 0.3587988018989563, | |
| "learning_rate": 3.992984631746469e-06, | |
| "loss": 0.6188, | |
| "step": 1980 | |
| }, | |
| { | |
| "epoch": 3.5030674846625764, | |
| "grad_norm": 0.3729381561279297, | |
| "learning_rate": 3.9005665902427695e-06, | |
| "loss": 0.6208, | |
| "step": 2000 | |
| }, | |
| { | |
| "epoch": 3.538124452234882, | |
| "grad_norm": 0.35941800475120544, | |
| "learning_rate": 3.8085416052958107e-06, | |
| "loss": 0.6183, | |
| "step": 2020 | |
| }, | |
| { | |
| "epoch": 3.5731814198071867, | |
| "grad_norm": 0.35061511397361755, | |
| "learning_rate": 3.7169425766004653e-06, | |
| "loss": 0.6148, | |
| "step": 2040 | |
| }, | |
| { | |
| "epoch": 3.6082383873794917, | |
| "grad_norm": 0.36660343408584595, | |
| "learning_rate": 3.6258022515687215e-06, | |
| "loss": 0.6186, | |
| "step": 2060 | |
| }, | |
| { | |
| "epoch": 3.6432953549517966, | |
| "grad_norm": 0.4035237729549408, | |
| "learning_rate": 3.5351532136222012e-06, | |
| "loss": 0.616, | |
| "step": 2080 | |
| }, | |
| { | |
| "epoch": 3.678352322524102, | |
| "grad_norm": 0.3646794855594635, | |
| "learning_rate": 3.445027870543323e-06, | |
| "loss": 0.6145, | |
| "step": 2100 | |
| }, | |
| { | |
| "epoch": 3.713409290096407, | |
| "grad_norm": 0.38012927770614624, | |
| "learning_rate": 3.3554584428892488e-06, | |
| "loss": 0.6181, | |
| "step": 2120 | |
| }, | |
| { | |
| "epoch": 3.7484662576687118, | |
| "grad_norm": 0.36509743332862854, | |
| "learning_rate": 3.2664769524727712e-06, | |
| "loss": 0.616, | |
| "step": 2140 | |
| }, | |
| { | |
| "epoch": 3.7835232252410167, | |
| "grad_norm": 0.38520926237106323, | |
| "learning_rate": 3.178115210914242e-06, | |
| "loss": 0.6169, | |
| "step": 2160 | |
| }, | |
| { | |
| "epoch": 3.8185801928133216, | |
| "grad_norm": 0.3496517837047577, | |
| "learning_rate": 3.0904048082686655e-06, | |
| "loss": 0.6158, | |
| "step": 2180 | |
| }, | |
| { | |
| "epoch": 3.8536371603856265, | |
| "grad_norm": 0.37065935134887695, | |
| "learning_rate": 3.00337710173198e-06, | |
| "loss": 0.6162, | |
| "step": 2200 | |
| }, | |
| { | |
| "epoch": 3.8886941279579315, | |
| "grad_norm": 0.339855819940567, | |
| "learning_rate": 2.9170632044306137e-06, | |
| "loss": 0.6156, | |
| "step": 2220 | |
| }, | |
| { | |
| "epoch": 3.9237510955302364, | |
| "grad_norm": 0.3591175675392151, | |
| "learning_rate": 2.8314939742982673e-06, | |
| "loss": 0.6144, | |
| "step": 2240 | |
| }, | |
| { | |
| "epoch": 3.9588080631025417, | |
| "grad_norm": 0.3600142300128937, | |
| "learning_rate": 2.746700003043964e-06, | |
| "loss": 0.6195, | |
| "step": 2260 | |
| }, | |
| { | |
| "epoch": 3.9938650306748467, | |
| "grad_norm": 0.3738831579685211, | |
| "learning_rate": 2.662711605215248e-06, | |
| "loss": 0.6135, | |
| "step": 2280 | |
| }, | |
| { | |
| "epoch": 4.028045574057844, | |
| "grad_norm": 0.3281383514404297, | |
| "learning_rate": 2.579558807360489e-06, | |
| "loss": 0.6104, | |
| "step": 2300 | |
| }, | |
| { | |
| "epoch": 4.063102541630149, | |
| "grad_norm": 0.3486866354942322, | |
| "learning_rate": 2.4972713372941406e-06, | |
| "loss": 0.6095, | |
| "step": 2320 | |
| }, | |
| { | |
| "epoch": 4.098159509202454, | |
| "grad_norm": 0.3392680287361145, | |
| "learning_rate": 2.4158786134687966e-06, | |
| "loss": 0.6108, | |
| "step": 2340 | |
| }, | |
| { | |
| "epoch": 4.133216476774759, | |
| "grad_norm": 0.3901905417442322, | |
| "learning_rate": 2.3354097344578565e-06, | |
| "loss": 0.6123, | |
| "step": 2360 | |
| }, | |
| { | |
| "epoch": 4.168273444347064, | |
| "grad_norm": 0.33477798104286194, | |
| "learning_rate": 2.25589346855254e-06, | |
| "loss": 0.6079, | |
| "step": 2380 | |
| }, | |
| { | |
| "epoch": 4.203330411919369, | |
| "grad_norm": 0.35764065384864807, | |
| "learning_rate": 2.1773582434769854e-06, | |
| "loss": 0.6066, | |
| "step": 2400 | |
| }, | |
| { | |
| "epoch": 4.238387379491674, | |
| "grad_norm": 0.34379830956459045, | |
| "learning_rate": 2.0998321362251036e-06, | |
| "loss": 0.608, | |
| "step": 2420 | |
| }, | |
| { | |
| "epoch": 4.273444347063979, | |
| "grad_norm": 0.3394622206687927, | |
| "learning_rate": 2.023342863022819e-06, | |
| "loss": 0.6142, | |
| "step": 2440 | |
| }, | |
| { | |
| "epoch": 4.308501314636284, | |
| "grad_norm": 0.35136643052101135, | |
| "learning_rate": 1.94791776941929e-06, | |
| "loss": 0.6115, | |
| "step": 2460 | |
| }, | |
| { | |
| "epoch": 4.343558282208589, | |
| "grad_norm": 0.3298383355140686, | |
| "learning_rate": 1.873583820510647e-06, | |
| "loss": 0.6089, | |
| "step": 2480 | |
| }, | |
| { | |
| "epoch": 4.378615249780894, | |
| "grad_norm": 0.3390386402606964, | |
| "learning_rate": 1.8003675912997487e-06, | |
| "loss": 0.612, | |
| "step": 2500 | |
| }, | |
| { | |
| "epoch": 4.413672217353199, | |
| "grad_norm": 0.3399540185928345, | |
| "learning_rate": 1.7282952571953987e-06, | |
| "loss": 0.6123, | |
| "step": 2520 | |
| }, | |
| { | |
| "epoch": 4.448729184925504, | |
| "grad_norm": 0.3504091203212738, | |
| "learning_rate": 1.657392584654412e-06, | |
| "loss": 0.6078, | |
| "step": 2540 | |
| }, | |
| { | |
| "epoch": 4.483786152497809, | |
| "grad_norm": 0.3221462070941925, | |
| "learning_rate": 1.587684921969912e-06, | |
| "loss": 0.6142, | |
| "step": 2560 | |
| }, | |
| { | |
| "epoch": 4.518843120070114, | |
| "grad_norm": 0.34956008195877075, | |
| "learning_rate": 1.5191971902090797e-06, | |
| "loss": 0.6085, | |
| "step": 2580 | |
| }, | |
| { | |
| "epoch": 4.553900087642419, | |
| "grad_norm": 0.34091153740882874, | |
| "learning_rate": 1.4519538743036927e-06, | |
| "loss": 0.6117, | |
| "step": 2600 | |
| }, | |
| { | |
| "epoch": 4.588957055214724, | |
| "grad_norm": 0.34645262360572815, | |
| "learning_rate": 1.385979014296533e-06, | |
| "loss": 0.6076, | |
| "step": 2620 | |
| }, | |
| { | |
| "epoch": 4.624014022787029, | |
| "grad_norm": 0.3348851799964905, | |
| "learning_rate": 1.3212961967468985e-06, | |
| "loss": 0.6116, | |
| "step": 2640 | |
| }, | |
| { | |
| "epoch": 4.659070990359334, | |
| "grad_norm": 0.32889384031295776, | |
| "learning_rate": 1.2579285462981855e-06, | |
| "loss": 0.6129, | |
| "step": 2660 | |
| }, | |
| { | |
| "epoch": 4.694127957931639, | |
| "grad_norm": 0.3461220860481262, | |
| "learning_rate": 1.195898717410664e-06, | |
| "loss": 0.6106, | |
| "step": 2680 | |
| }, | |
| { | |
| "epoch": 4.729184925503944, | |
| "grad_norm": 0.319545716047287, | |
| "learning_rate": 1.1352288862622968e-06, | |
| "loss": 0.6068, | |
| "step": 2700 | |
| }, | |
| { | |
| "epoch": 4.764241893076249, | |
| "grad_norm": 0.3272876739501953, | |
| "learning_rate": 1.075940742820588e-06, | |
| "loss": 0.6082, | |
| "step": 2720 | |
| }, | |
| { | |
| "epoch": 4.799298860648554, | |
| "grad_norm": 0.3398887515068054, | |
| "learning_rate": 1.0180554830882333e-06, | |
| "loss": 0.6069, | |
| "step": 2740 | |
| }, | |
| { | |
| "epoch": 4.8343558282208585, | |
| "grad_norm": 0.32064250111579895, | |
| "learning_rate": 9.615938015253723e-07, | |
| "loss": 0.6086, | |
| "step": 2760 | |
| }, | |
| { | |
| "epoch": 4.869412795793163, | |
| "grad_norm": 0.32411012053489685, | |
| "learning_rate": 9.065758836511556e-07, | |
| "loss": 0.6083, | |
| "step": 2780 | |
| }, | |
| { | |
| "epoch": 4.904469763365469, | |
| "grad_norm": 0.31640660762786865, | |
| "learning_rate": 8.53021398827239e-07, | |
| "loss": 0.6094, | |
| "step": 2800 | |
| }, | |
| { | |
| "epoch": 4.939526730937774, | |
| "grad_norm": 0.3243345320224762, | |
| "learning_rate": 8.009494932258427e-07, | |
| "loss": 0.6104, | |
| "step": 2820 | |
| }, | |
| { | |
| "epoch": 4.974583698510079, | |
| "grad_norm": 0.31760430335998535, | |
| "learning_rate": 7.503787829848191e-07, | |
| "loss": 0.6078, | |
| "step": 2840 | |
| }, | |
| { | |
| "epoch": 5.0087642418930765, | |
| "grad_norm": 0.32336053252220154, | |
| "learning_rate": 7.013273475522392e-07, | |
| "loss": 0.609, | |
| "step": 2860 | |
| }, | |
| { | |
| "epoch": 5.043821209465381, | |
| "grad_norm": 0.31958338618278503, | |
| "learning_rate": 6.53812723222838e-07, | |
| "loss": 0.6073, | |
| "step": 2880 | |
| }, | |
| { | |
| "epoch": 5.078878177037686, | |
| "grad_norm": 0.319396436214447, | |
| "learning_rate": 6.078518968686426e-07, | |
| "loss": 0.6096, | |
| "step": 2900 | |
| }, | |
| { | |
| "epoch": 5.113935144609991, | |
| "grad_norm": 0.33287620544433594, | |
| "learning_rate": 5.634612998660249e-07, | |
| "loss": 0.6011, | |
| "step": 2920 | |
| }, | |
| { | |
| "epoch": 5.148992112182296, | |
| "grad_norm": 0.32154789566993713, | |
| "learning_rate": 5.206568022213482e-07, | |
| "loss": 0.6081, | |
| "step": 2940 | |
| }, | |
| { | |
| "epoch": 5.184049079754601, | |
| "grad_norm": 0.3217925727367401, | |
| "learning_rate": 4.794537068973093e-07, | |
| "loss": 0.607, | |
| "step": 2960 | |
| }, | |
| { | |
| "epoch": 5.219106047326906, | |
| "grad_norm": 0.3143027722835541, | |
| "learning_rate": 4.398667443420029e-07, | |
| "loss": 0.6065, | |
| "step": 2980 | |
| }, | |
| { | |
| "epoch": 5.254163014899211, | |
| "grad_norm": 0.3175284266471863, | |
| "learning_rate": 4.019100672226617e-07, | |
| "loss": 0.6056, | |
| "step": 3000 | |
| }, | |
| { | |
| "epoch": 5.289219982471516, | |
| "grad_norm": 0.32859277725219727, | |
| "learning_rate": 3.65597245365964e-07, | |
| "loss": 0.6063, | |
| "step": 3020 | |
| }, | |
| { | |
| "epoch": 5.324276950043822, | |
| "grad_norm": 0.3200486898422241, | |
| "learning_rate": 3.3094126090670477e-07, | |
| "loss": 0.6074, | |
| "step": 3040 | |
| }, | |
| { | |
| "epoch": 5.3593339176161265, | |
| "grad_norm": 0.3193705976009369, | |
| "learning_rate": 2.9795450364657865e-07, | |
| "loss": 0.607, | |
| "step": 3060 | |
| }, | |
| { | |
| "epoch": 5.3943908851884315, | |
| "grad_norm": 0.32681772112846375, | |
| "learning_rate": 2.6664876662471697e-07, | |
| "loss": 0.6053, | |
| "step": 3080 | |
| }, | |
| { | |
| "epoch": 5.429447852760736, | |
| "grad_norm": 0.31692323088645935, | |
| "learning_rate": 2.370352419015892e-07, | |
| "loss": 0.608, | |
| "step": 3100 | |
| }, | |
| { | |
| "epoch": 5.464504820333041, | |
| "grad_norm": 0.32039353251457214, | |
| "learning_rate": 2.091245165577349e-07, | |
| "loss": 0.6079, | |
| "step": 3120 | |
| }, | |
| { | |
| "epoch": 5.499561787905346, | |
| "grad_norm": 0.31619492173194885, | |
| "learning_rate": 1.8292656890880722e-07, | |
| "loss": 0.6045, | |
| "step": 3140 | |
| }, | |
| { | |
| "epoch": 5.534618755477651, | |
| "grad_norm": 0.31279516220092773, | |
| "learning_rate": 1.5845076493823331e-07, | |
| "loss": 0.6066, | |
| "step": 3160 | |
| }, | |
| { | |
| "epoch": 5.569675723049956, | |
| "grad_norm": 0.3182794451713562, | |
| "learning_rate": 1.3570585494880328e-07, | |
| "loss": 0.6085, | |
| "step": 3180 | |
| }, | |
| { | |
| "epoch": 5.604732690622261, | |
| "grad_norm": 0.31176891922950745, | |
| "learning_rate": 1.1469997043436154e-07, | |
| "loss": 0.6048, | |
| "step": 3200 | |
| }, | |
| { | |
| "epoch": 5.639789658194566, | |
| "grad_norm": 0.31051337718963623, | |
| "learning_rate": 9.544062117273045e-08, | |
| "loss": 0.607, | |
| "step": 3220 | |
| }, | |
| { | |
| "epoch": 5.674846625766871, | |
| "grad_norm": 0.31119874119758606, | |
| "learning_rate": 7.793469254090524e-08, | |
| "loss": 0.6054, | |
| "step": 3240 | |
| }, | |
| { | |
| "epoch": 5.709903593339176, | |
| "grad_norm": 0.3019949793815613, | |
| "learning_rate": 6.218844305346916e-08, | |
| "loss": 0.6074, | |
| "step": 3260 | |
| }, | |
| { | |
| "epoch": 5.7449605609114816, | |
| "grad_norm": 0.30909818410873413, | |
| "learning_rate": 4.820750212513048e-08, | |
| "loss": 0.6085, | |
| "step": 3280 | |
| }, | |
| { | |
| "epoch": 5.7800175284837865, | |
| "grad_norm": 0.30439695715904236, | |
| "learning_rate": 3.599686805815128e-08, | |
| "loss": 0.6101, | |
| "step": 3300 | |
| }, | |
| { | |
| "epoch": 5.815074496056091, | |
| "grad_norm": 0.3216073215007782, | |
| "learning_rate": 2.5560906255420737e-08, | |
| "loss": 0.6092, | |
| "step": 3320 | |
| }, | |
| { | |
| "epoch": 5.850131463628396, | |
| "grad_norm": 0.3213990330696106, | |
| "learning_rate": 1.6903347659781856e-08, | |
| "loss": 0.6071, | |
| "step": 3340 | |
| }, | |
| { | |
| "epoch": 5.885188431200701, | |
| "grad_norm": 0.3153989613056183, | |
| "learning_rate": 1.0027287420192322e-08, | |
| "loss": 0.6074, | |
| "step": 3360 | |
| }, | |
| { | |
| "epoch": 5.920245398773006, | |
| "grad_norm": 0.308703750371933, | |
| "learning_rate": 4.935183785180209e-09, | |
| "loss": 0.6058, | |
| "step": 3380 | |
| }, | |
| { | |
| "epoch": 5.955302366345311, | |
| "grad_norm": 0.3108364939689636, | |
| "learning_rate": 1.6288572240014123e-09, | |
| "loss": 0.6038, | |
| "step": 3400 | |
| }, | |
| { | |
| "epoch": 5.990359333917616, | |
| "grad_norm": 0.313579261302948, | |
| "learning_rate": 1.0948977580638176e-10, | |
| "loss": 0.6079, | |
| "step": 3420 | |
| }, | |
| { | |
| "epoch": 6.0, | |
| "step": 3426, | |
| "total_flos": 9.485754681343869e+18, | |
| "train_loss": 0.6827397172519218, | |
| "train_runtime": 47545.1809, | |
| "train_samples_per_second": 4.607, | |
| "train_steps_per_second": 0.072 | |
| } | |
| ], | |
| "logging_steps": 20, | |
| "max_steps": 3426, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 6, | |
| "save_steps": 500, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": true | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 9.485754681343869e+18, | |
| "train_batch_size": 2, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |