| { | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 3.0, | |
| "eval_steps": 500, | |
| "global_step": 939, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.03194888178913738, | |
| "grad_norm": 4.9730574359401585, | |
| "learning_rate": 5e-06, | |
| "loss": 0.9178, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.06389776357827476, | |
| "grad_norm": 2.414364422517365, | |
| "learning_rate": 5e-06, | |
| "loss": 0.8185, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.09584664536741214, | |
| "grad_norm": 2.355627570386528, | |
| "learning_rate": 5e-06, | |
| "loss": 0.7955, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.12779552715654952, | |
| "grad_norm": 0.9368274495065052, | |
| "learning_rate": 5e-06, | |
| "loss": 0.777, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.1597444089456869, | |
| "grad_norm": 1.0303864547714687, | |
| "learning_rate": 5e-06, | |
| "loss": 0.7544, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.19169329073482427, | |
| "grad_norm": 0.8799746481470822, | |
| "learning_rate": 5e-06, | |
| "loss": 0.7414, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.22364217252396165, | |
| "grad_norm": 2.571145704302205, | |
| "learning_rate": 5e-06, | |
| "loss": 0.7335, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 0.25559105431309903, | |
| "grad_norm": 1.4093624036341388, | |
| "learning_rate": 5e-06, | |
| "loss": 0.7174, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.28753993610223644, | |
| "grad_norm": 0.8615478628829397, | |
| "learning_rate": 5e-06, | |
| "loss": 0.7318, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 0.3194888178913738, | |
| "grad_norm": 0.5831162653054314, | |
| "learning_rate": 5e-06, | |
| "loss": 0.72, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.3514376996805112, | |
| "grad_norm": 0.6173922126615862, | |
| "learning_rate": 5e-06, | |
| "loss": 0.716, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 0.38338658146964855, | |
| "grad_norm": 0.8888214325930446, | |
| "learning_rate": 5e-06, | |
| "loss": 0.7134, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 0.41533546325878595, | |
| "grad_norm": 0.6979078907766634, | |
| "learning_rate": 5e-06, | |
| "loss": 0.7021, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 0.4472843450479233, | |
| "grad_norm": 0.6491577783946951, | |
| "learning_rate": 5e-06, | |
| "loss": 0.6991, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 0.4792332268370607, | |
| "grad_norm": 0.5395727376776203, | |
| "learning_rate": 5e-06, | |
| "loss": 0.7068, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 0.5111821086261981, | |
| "grad_norm": 0.689753166346745, | |
| "learning_rate": 5e-06, | |
| "loss": 0.6935, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 0.5431309904153354, | |
| "grad_norm": 0.5521859868964065, | |
| "learning_rate": 5e-06, | |
| "loss": 0.705, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 0.5750798722044729, | |
| "grad_norm": 0.5708494730680989, | |
| "learning_rate": 5e-06, | |
| "loss": 0.6961, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 0.6070287539936102, | |
| "grad_norm": 0.7742403437422439, | |
| "learning_rate": 5e-06, | |
| "loss": 0.7045, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 0.6389776357827476, | |
| "grad_norm": 0.6022796802365044, | |
| "learning_rate": 5e-06, | |
| "loss": 0.6911, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.670926517571885, | |
| "grad_norm": 0.5364796054217524, | |
| "learning_rate": 5e-06, | |
| "loss": 0.695, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 0.7028753993610224, | |
| "grad_norm": 0.47930189735330775, | |
| "learning_rate": 5e-06, | |
| "loss": 0.6961, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 0.7348242811501597, | |
| "grad_norm": 0.5376571771457028, | |
| "learning_rate": 5e-06, | |
| "loss": 0.7035, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 0.7667731629392971, | |
| "grad_norm": 0.7570921090383397, | |
| "learning_rate": 5e-06, | |
| "loss": 0.6801, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 0.7987220447284346, | |
| "grad_norm": 0.5375423682771385, | |
| "learning_rate": 5e-06, | |
| "loss": 0.6809, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 0.8306709265175719, | |
| "grad_norm": 0.5836884654889642, | |
| "learning_rate": 5e-06, | |
| "loss": 0.6864, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 0.8626198083067093, | |
| "grad_norm": 0.598608895789362, | |
| "learning_rate": 5e-06, | |
| "loss": 0.6902, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 0.8945686900958466, | |
| "grad_norm": 0.5331459973786111, | |
| "learning_rate": 5e-06, | |
| "loss": 0.6879, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 0.9265175718849841, | |
| "grad_norm": 0.5067341091409512, | |
| "learning_rate": 5e-06, | |
| "loss": 0.6881, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 0.9584664536741214, | |
| "grad_norm": 0.5948122722251207, | |
| "learning_rate": 5e-06, | |
| "loss": 0.6868, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.9904153354632588, | |
| "grad_norm": 0.6883144495229085, | |
| "learning_rate": 5e-06, | |
| "loss": 0.6818, | |
| "step": 310 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "eval_loss": 0.680176854133606, | |
| "eval_runtime": 30.2696, | |
| "eval_samples_per_second": 277.968, | |
| "eval_steps_per_second": 1.09, | |
| "step": 313 | |
| }, | |
| { | |
| "epoch": 1.0223642172523961, | |
| "grad_norm": 0.7468894293006159, | |
| "learning_rate": 5e-06, | |
| "loss": 0.656, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 1.0543130990415335, | |
| "grad_norm": 0.6942190077761362, | |
| "learning_rate": 5e-06, | |
| "loss": 0.6387, | |
| "step": 330 | |
| }, | |
| { | |
| "epoch": 1.0862619808306708, | |
| "grad_norm": 0.6242193449068775, | |
| "learning_rate": 5e-06, | |
| "loss": 0.6407, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 1.1182108626198084, | |
| "grad_norm": 0.6329861406612104, | |
| "learning_rate": 5e-06, | |
| "loss": 0.6428, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 1.1501597444089458, | |
| "grad_norm": 0.6142460964851904, | |
| "learning_rate": 5e-06, | |
| "loss": 0.6356, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 1.182108626198083, | |
| "grad_norm": 0.6850789950403501, | |
| "learning_rate": 5e-06, | |
| "loss": 0.636, | |
| "step": 370 | |
| }, | |
| { | |
| "epoch": 1.2140575079872205, | |
| "grad_norm": 0.5355828508870474, | |
| "learning_rate": 5e-06, | |
| "loss": 0.6382, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 1.2460063897763578, | |
| "grad_norm": 0.5629126693045496, | |
| "learning_rate": 5e-06, | |
| "loss": 0.6344, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 1.2779552715654952, | |
| "grad_norm": 0.5204488937445614, | |
| "learning_rate": 5e-06, | |
| "loss": 0.6401, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 1.3099041533546325, | |
| "grad_norm": 0.5396176523426488, | |
| "learning_rate": 5e-06, | |
| "loss": 0.6382, | |
| "step": 410 | |
| }, | |
| { | |
| "epoch": 1.34185303514377, | |
| "grad_norm": 0.6023672235090743, | |
| "learning_rate": 5e-06, | |
| "loss": 0.6417, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 1.3738019169329074, | |
| "grad_norm": 0.5459880912136352, | |
| "learning_rate": 5e-06, | |
| "loss": 0.6434, | |
| "step": 430 | |
| }, | |
| { | |
| "epoch": 1.4057507987220448, | |
| "grad_norm": 0.5092923506095731, | |
| "learning_rate": 5e-06, | |
| "loss": 0.6366, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 1.4376996805111821, | |
| "grad_norm": 0.6382479698360073, | |
| "learning_rate": 5e-06, | |
| "loss": 0.6353, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 1.4696485623003195, | |
| "grad_norm": 0.7533985947305522, | |
| "learning_rate": 5e-06, | |
| "loss": 0.6392, | |
| "step": 460 | |
| }, | |
| { | |
| "epoch": 1.5015974440894568, | |
| "grad_norm": 0.7229864015275371, | |
| "learning_rate": 5e-06, | |
| "loss": 0.6381, | |
| "step": 470 | |
| }, | |
| { | |
| "epoch": 1.5335463258785942, | |
| "grad_norm": 0.5758239404648058, | |
| "learning_rate": 5e-06, | |
| "loss": 0.636, | |
| "step": 480 | |
| }, | |
| { | |
| "epoch": 1.5654952076677318, | |
| "grad_norm": 0.5422584017870065, | |
| "learning_rate": 5e-06, | |
| "loss": 0.6393, | |
| "step": 490 | |
| }, | |
| { | |
| "epoch": 1.5974440894568689, | |
| "grad_norm": 0.500259651199132, | |
| "learning_rate": 5e-06, | |
| "loss": 0.6441, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 1.6293929712460065, | |
| "grad_norm": 0.8813431746686585, | |
| "learning_rate": 5e-06, | |
| "loss": 0.6308, | |
| "step": 510 | |
| }, | |
| { | |
| "epoch": 1.6613418530351438, | |
| "grad_norm": 0.49923586083798, | |
| "learning_rate": 5e-06, | |
| "loss": 0.6365, | |
| "step": 520 | |
| }, | |
| { | |
| "epoch": 1.6932907348242812, | |
| "grad_norm": 0.46329712814558677, | |
| "learning_rate": 5e-06, | |
| "loss": 0.6279, | |
| "step": 530 | |
| }, | |
| { | |
| "epoch": 1.7252396166134185, | |
| "grad_norm": 0.7606638693711257, | |
| "learning_rate": 5e-06, | |
| "loss": 0.6341, | |
| "step": 540 | |
| }, | |
| { | |
| "epoch": 1.7571884984025559, | |
| "grad_norm": 0.5490152416615772, | |
| "learning_rate": 5e-06, | |
| "loss": 0.6378, | |
| "step": 550 | |
| }, | |
| { | |
| "epoch": 1.7891373801916934, | |
| "grad_norm": 0.5845892096604395, | |
| "learning_rate": 5e-06, | |
| "loss": 0.6395, | |
| "step": 560 | |
| }, | |
| { | |
| "epoch": 1.8210862619808306, | |
| "grad_norm": 0.5147171055711276, | |
| "learning_rate": 5e-06, | |
| "loss": 0.6297, | |
| "step": 570 | |
| }, | |
| { | |
| "epoch": 1.8530351437699681, | |
| "grad_norm": 0.5052763678219258, | |
| "learning_rate": 5e-06, | |
| "loss": 0.6342, | |
| "step": 580 | |
| }, | |
| { | |
| "epoch": 1.8849840255591053, | |
| "grad_norm": 0.5390734384968956, | |
| "learning_rate": 5e-06, | |
| "loss": 0.6369, | |
| "step": 590 | |
| }, | |
| { | |
| "epoch": 1.9169329073482428, | |
| "grad_norm": 0.6413516519865096, | |
| "learning_rate": 5e-06, | |
| "loss": 0.6418, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 1.9488817891373802, | |
| "grad_norm": 0.5239174298332788, | |
| "learning_rate": 5e-06, | |
| "loss": 0.6274, | |
| "step": 610 | |
| }, | |
| { | |
| "epoch": 1.9808306709265175, | |
| "grad_norm": 0.7039020492664947, | |
| "learning_rate": 5e-06, | |
| "loss": 0.6385, | |
| "step": 620 | |
| }, | |
| { | |
| "epoch": 2.0, | |
| "eval_loss": 0.6707362532615662, | |
| "eval_runtime": 29.7735, | |
| "eval_samples_per_second": 282.6, | |
| "eval_steps_per_second": 1.108, | |
| "step": 626 | |
| }, | |
| { | |
| "epoch": 2.012779552715655, | |
| "grad_norm": 0.9555964669975626, | |
| "learning_rate": 5e-06, | |
| "loss": 0.6152, | |
| "step": 630 | |
| }, | |
| { | |
| "epoch": 2.0447284345047922, | |
| "grad_norm": 0.6929923210576585, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5881, | |
| "step": 640 | |
| }, | |
| { | |
| "epoch": 2.07667731629393, | |
| "grad_norm": 0.5186537877634542, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5871, | |
| "step": 650 | |
| }, | |
| { | |
| "epoch": 2.108626198083067, | |
| "grad_norm": 0.5532963474572669, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5896, | |
| "step": 660 | |
| }, | |
| { | |
| "epoch": 2.1405750798722045, | |
| "grad_norm": 0.5667781850304423, | |
| "learning_rate": 5e-06, | |
| "loss": 0.589, | |
| "step": 670 | |
| }, | |
| { | |
| "epoch": 2.1725239616613417, | |
| "grad_norm": 0.5285098402420302, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5901, | |
| "step": 680 | |
| }, | |
| { | |
| "epoch": 2.2044728434504792, | |
| "grad_norm": 0.536646702155888, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5863, | |
| "step": 690 | |
| }, | |
| { | |
| "epoch": 2.236421725239617, | |
| "grad_norm": 0.5015237529663761, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5846, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 2.268370607028754, | |
| "grad_norm": 0.6152245702781334, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5885, | |
| "step": 710 | |
| }, | |
| { | |
| "epoch": 2.3003194888178915, | |
| "grad_norm": 0.6897591535184826, | |
| "learning_rate": 5e-06, | |
| "loss": 0.584, | |
| "step": 720 | |
| }, | |
| { | |
| "epoch": 2.3322683706070286, | |
| "grad_norm": 0.48574756268499075, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5888, | |
| "step": 730 | |
| }, | |
| { | |
| "epoch": 2.364217252396166, | |
| "grad_norm": 0.7776088479205505, | |
| "learning_rate": 5e-06, | |
| "loss": 0.591, | |
| "step": 740 | |
| }, | |
| { | |
| "epoch": 2.3961661341853033, | |
| "grad_norm": 0.5643419759040877, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5975, | |
| "step": 750 | |
| }, | |
| { | |
| "epoch": 2.428115015974441, | |
| "grad_norm": 0.7618894474745995, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5907, | |
| "step": 760 | |
| }, | |
| { | |
| "epoch": 2.460063897763578, | |
| "grad_norm": 0.566521654030305, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5902, | |
| "step": 770 | |
| }, | |
| { | |
| "epoch": 2.4920127795527156, | |
| "grad_norm": 0.539754953993636, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5884, | |
| "step": 780 | |
| }, | |
| { | |
| "epoch": 2.523961661341853, | |
| "grad_norm": 0.5625428853361434, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5818, | |
| "step": 790 | |
| }, | |
| { | |
| "epoch": 2.5559105431309903, | |
| "grad_norm": 0.6279591943173818, | |
| "learning_rate": 5e-06, | |
| "loss": 0.588, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 2.587859424920128, | |
| "grad_norm": 0.6224175939518917, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5934, | |
| "step": 810 | |
| }, | |
| { | |
| "epoch": 2.619808306709265, | |
| "grad_norm": 0.5642613523825399, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5865, | |
| "step": 820 | |
| }, | |
| { | |
| "epoch": 2.6517571884984026, | |
| "grad_norm": 0.6865432310936253, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5926, | |
| "step": 830 | |
| }, | |
| { | |
| "epoch": 2.68370607028754, | |
| "grad_norm": 0.5373522424367686, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5921, | |
| "step": 840 | |
| }, | |
| { | |
| "epoch": 2.7156549520766773, | |
| "grad_norm": 0.567457215981538, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5954, | |
| "step": 850 | |
| }, | |
| { | |
| "epoch": 2.747603833865815, | |
| "grad_norm": 0.6426061724785515, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5988, | |
| "step": 860 | |
| }, | |
| { | |
| "epoch": 2.779552715654952, | |
| "grad_norm": 0.59442339830324, | |
| "learning_rate": 5e-06, | |
| "loss": 0.6008, | |
| "step": 870 | |
| }, | |
| { | |
| "epoch": 2.8115015974440896, | |
| "grad_norm": 0.5592579452365836, | |
| "learning_rate": 5e-06, | |
| "loss": 0.6022, | |
| "step": 880 | |
| }, | |
| { | |
| "epoch": 2.8434504792332267, | |
| "grad_norm": 0.5094979594863669, | |
| "learning_rate": 5e-06, | |
| "loss": 0.592, | |
| "step": 890 | |
| }, | |
| { | |
| "epoch": 2.8753993610223643, | |
| "grad_norm": 0.5091387458726695, | |
| "learning_rate": 5e-06, | |
| "loss": 0.6, | |
| "step": 900 | |
| }, | |
| { | |
| "epoch": 2.9073482428115014, | |
| "grad_norm": 0.5094298110498509, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5938, | |
| "step": 910 | |
| }, | |
| { | |
| "epoch": 2.939297124600639, | |
| "grad_norm": 0.5130578982623665, | |
| "learning_rate": 5e-06, | |
| "loss": 0.586, | |
| "step": 920 | |
| }, | |
| { | |
| "epoch": 2.9712460063897765, | |
| "grad_norm": 0.4928009047818368, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5911, | |
| "step": 930 | |
| }, | |
| { | |
| "epoch": 3.0, | |
| "eval_loss": 0.6763173341751099, | |
| "eval_runtime": 29.5519, | |
| "eval_samples_per_second": 284.719, | |
| "eval_steps_per_second": 1.117, | |
| "step": 939 | |
| }, | |
| { | |
| "epoch": 3.0, | |
| "step": 939, | |
| "total_flos": 1572859973468160.0, | |
| "train_loss": 0.648996304907103, | |
| "train_runtime": 5978.5297, | |
| "train_samples_per_second": 80.212, | |
| "train_steps_per_second": 0.157 | |
| } | |
| ], | |
| "logging_steps": 10, | |
| "max_steps": 939, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 3, | |
| "save_steps": 500, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": true | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 1572859973468160.0, | |
| "train_batch_size": 16, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |