| { | |
| "best_global_step": null, | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 14.993576017130621, | |
| "eval_steps": 500, | |
| "global_step": 1740, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.08565310492505353, | |
| "grad_norm": 5.411635875701904, | |
| "learning_rate": 9.999339889379647e-06, | |
| "loss": 1.5609, | |
| "num_input_tokens_seen": 255328, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.17130620985010706, | |
| "grad_norm": 0.7214002013206482, | |
| "learning_rate": 9.997058249278764e-06, | |
| "loss": 0.2055, | |
| "num_input_tokens_seen": 510640, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.2569593147751606, | |
| "grad_norm": 0.47501006722450256, | |
| "learning_rate": 9.993147673772869e-06, | |
| "loss": 0.184, | |
| "num_input_tokens_seen": 763808, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.3426124197002141, | |
| "grad_norm": 0.20641829073429108, | |
| "learning_rate": 9.987609437626955e-06, | |
| "loss": 0.1767, | |
| "num_input_tokens_seen": 1017472, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.4282655246252677, | |
| "grad_norm": 0.49966660141944885, | |
| "learning_rate": 9.98044534618898e-06, | |
| "loss": 0.1821, | |
| "num_input_tokens_seen": 1273488, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.5139186295503212, | |
| "grad_norm": 0.8994255065917969, | |
| "learning_rate": 9.971657734801385e-06, | |
| "loss": 0.1819, | |
| "num_input_tokens_seen": 1527744, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.5995717344753747, | |
| "grad_norm": 1.7688418626785278, | |
| "learning_rate": 9.961249468039806e-06, | |
| "loss": 0.1797, | |
| "num_input_tokens_seen": 1785520, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 0.6852248394004282, | |
| "grad_norm": 0.7211973071098328, | |
| "learning_rate": 9.949223938779286e-06, | |
| "loss": 0.1765, | |
| "num_input_tokens_seen": 2037648, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.7708779443254818, | |
| "grad_norm": 0.94338458776474, | |
| "learning_rate": 9.935585067088276e-06, | |
| "loss": 0.1766, | |
| "num_input_tokens_seen": 2292464, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 0.8565310492505354, | |
| "grad_norm": 0.6227909326553345, | |
| "learning_rate": 9.920337298950767e-06, | |
| "loss": 0.1714, | |
| "num_input_tokens_seen": 2547872, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.9421841541755889, | |
| "grad_norm": 0.5941164493560791, | |
| "learning_rate": 9.903485604816993e-06, | |
| "loss": 0.1728, | |
| "num_input_tokens_seen": 2801536, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 1.0342612419700214, | |
| "grad_norm": 0.4057783782482147, | |
| "learning_rate": 9.885035477983184e-06, | |
| "loss": 0.1868, | |
| "num_input_tokens_seen": 3074416, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 1.119914346895075, | |
| "grad_norm": 0.5549430251121521, | |
| "learning_rate": 9.864992932800845e-06, | |
| "loss": 0.1679, | |
| "num_input_tokens_seen": 3327568, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 1.2055674518201285, | |
| "grad_norm": 0.41472136974334717, | |
| "learning_rate": 9.843364502716225e-06, | |
| "loss": 0.1671, | |
| "num_input_tokens_seen": 3582240, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 1.291220556745182, | |
| "grad_norm": 0.5156757235527039, | |
| "learning_rate": 9.820157238140535e-06, | |
| "loss": 0.1682, | |
| "num_input_tokens_seen": 3838160, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 1.3768736616702355, | |
| "grad_norm": 0.5046593546867371, | |
| "learning_rate": 9.795378704151675e-06, | |
| "loss": 0.1651, | |
| "num_input_tokens_seen": 4092304, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 1.462526766595289, | |
| "grad_norm": 0.5588434338569641, | |
| "learning_rate": 9.76903697802817e-06, | |
| "loss": 0.1649, | |
| "num_input_tokens_seen": 4346640, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 1.5481798715203428, | |
| "grad_norm": 0.46262454986572266, | |
| "learning_rate": 9.741140646616161e-06, | |
| "loss": 0.1669, | |
| "num_input_tokens_seen": 4602192, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 1.633832976445396, | |
| "grad_norm": 0.45427972078323364, | |
| "learning_rate": 9.711698803530253e-06, | |
| "loss": 0.1674, | |
| "num_input_tokens_seen": 4858240, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 1.7194860813704498, | |
| "grad_norm": 0.4514879882335663, | |
| "learning_rate": 9.68072104618921e-06, | |
| "loss": 0.1641, | |
| "num_input_tokens_seen": 5113408, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 1.805139186295503, | |
| "grad_norm": 0.7933849692344666, | |
| "learning_rate": 9.648217472687385e-06, | |
| "loss": 0.1614, | |
| "num_input_tokens_seen": 5368352, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 1.8907922912205568, | |
| "grad_norm": 0.6207934021949768, | |
| "learning_rate": 9.614198678502965e-06, | |
| "loss": 0.163, | |
| "num_input_tokens_seen": 5622128, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 1.9764453961456103, | |
| "grad_norm": 0.8193040490150452, | |
| "learning_rate": 9.57867575304406e-06, | |
| "loss": 0.1589, | |
| "num_input_tokens_seen": 5876816, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 2.068522483940043, | |
| "grad_norm": 1.0469202995300293, | |
| "learning_rate": 9.541660276033795e-06, | |
| "loss": 0.1755, | |
| "num_input_tokens_seen": 6145392, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 2.154175588865096, | |
| "grad_norm": 0.9274189472198486, | |
| "learning_rate": 9.503164313735566e-06, | |
| "loss": 0.1595, | |
| "num_input_tokens_seen": 6399504, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 2.23982869379015, | |
| "grad_norm": 0.6875982880592346, | |
| "learning_rate": 9.46320041501969e-06, | |
| "loss": 0.1563, | |
| "num_input_tokens_seen": 6654160, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 2.325481798715203, | |
| "grad_norm": 0.5835751295089722, | |
| "learning_rate": 9.421781607272741e-06, | |
| "loss": 0.1554, | |
| "num_input_tokens_seen": 6910752, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 2.411134903640257, | |
| "grad_norm": 0.6475698351860046, | |
| "learning_rate": 9.378921392150893e-06, | |
| "loss": 0.1579, | |
| "num_input_tokens_seen": 7166960, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 2.4967880085653107, | |
| "grad_norm": 0.6029316782951355, | |
| "learning_rate": 9.33463374117867e-06, | |
| "loss": 0.1577, | |
| "num_input_tokens_seen": 7420288, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 2.582441113490364, | |
| "grad_norm": 0.6444355845451355, | |
| "learning_rate": 9.288933091194524e-06, | |
| "loss": 0.1564, | |
| "num_input_tokens_seen": 7675184, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 2.6680942184154177, | |
| "grad_norm": 0.5541071891784668, | |
| "learning_rate": 9.241834339644726e-06, | |
| "loss": 0.1528, | |
| "num_input_tokens_seen": 7926976, | |
| "step": 310 | |
| }, | |
| { | |
| "epoch": 2.753747323340471, | |
| "grad_norm": 0.6703725457191467, | |
| "learning_rate": 9.193352839727122e-06, | |
| "loss": 0.1549, | |
| "num_input_tokens_seen": 8184992, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 2.8394004282655247, | |
| "grad_norm": 0.6515584588050842, | |
| "learning_rate": 9.143504395386302e-06, | |
| "loss": 0.157, | |
| "num_input_tokens_seen": 8439712, | |
| "step": 330 | |
| }, | |
| { | |
| "epoch": 2.925053533190578, | |
| "grad_norm": 0.5527693629264832, | |
| "learning_rate": 9.09230525616186e-06, | |
| "loss": 0.157, | |
| "num_input_tokens_seen": 8694080, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 3.017130620985011, | |
| "grad_norm": 0.6593677401542664, | |
| "learning_rate": 9.039772111891383e-06, | |
| "loss": 0.1672, | |
| "num_input_tokens_seen": 8965488, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 3.102783725910064, | |
| "grad_norm": 0.5042828917503357, | |
| "learning_rate": 8.985922087269916e-06, | |
| "loss": 0.1483, | |
| "num_input_tokens_seen": 9220480, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 3.188436830835118, | |
| "grad_norm": 0.4123888611793518, | |
| "learning_rate": 8.930772736267675e-06, | |
| "loss": 0.1532, | |
| "num_input_tokens_seen": 9477024, | |
| "step": 370 | |
| }, | |
| { | |
| "epoch": 3.274089935760171, | |
| "grad_norm": 0.7851901054382324, | |
| "learning_rate": 8.874342036407815e-06, | |
| "loss": 0.1508, | |
| "num_input_tokens_seen": 9731840, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 3.359743040685225, | |
| "grad_norm": 0.7545840740203857, | |
| "learning_rate": 8.816648382906154e-06, | |
| "loss": 0.1516, | |
| "num_input_tokens_seen": 9986704, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 3.445396145610278, | |
| "grad_norm": 0.7439327239990234, | |
| "learning_rate": 8.757710582674708e-06, | |
| "loss": 0.1506, | |
| "num_input_tokens_seen": 10238720, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 3.531049250535332, | |
| "grad_norm": 0.8343164920806885, | |
| "learning_rate": 8.697547848191037e-06, | |
| "loss": 0.1516, | |
| "num_input_tokens_seen": 10491856, | |
| "step": 410 | |
| }, | |
| { | |
| "epoch": 3.6167023554603857, | |
| "grad_norm": 0.817565381526947, | |
| "learning_rate": 8.63617979123539e-06, | |
| "loss": 0.1542, | |
| "num_input_tokens_seen": 10744240, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 3.702355460385439, | |
| "grad_norm": 0.5334470272064209, | |
| "learning_rate": 8.573626416497669e-06, | |
| "loss": 0.1446, | |
| "num_input_tokens_seen": 10996768, | |
| "step": 430 | |
| }, | |
| { | |
| "epoch": 3.7880085653104922, | |
| "grad_norm": 0.9441611766815186, | |
| "learning_rate": 8.509908115056334e-06, | |
| "loss": 0.1515, | |
| "num_input_tokens_seen": 11254560, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 3.873661670235546, | |
| "grad_norm": 0.6177489757537842, | |
| "learning_rate": 8.445045657731329e-06, | |
| "loss": 0.1513, | |
| "num_input_tokens_seen": 11512992, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 3.9593147751605997, | |
| "grad_norm": 0.5743350982666016, | |
| "learning_rate": 8.379060188313244e-06, | |
| "loss": 0.1458, | |
| "num_input_tokens_seen": 11765808, | |
| "step": 460 | |
| }, | |
| { | |
| "epoch": 4.0513918629550325, | |
| "grad_norm": 0.8525713086128235, | |
| "learning_rate": 8.311973216670888e-06, | |
| "loss": 0.1598, | |
| "num_input_tokens_seen": 12036784, | |
| "step": 470 | |
| }, | |
| { | |
| "epoch": 4.137044967880086, | |
| "grad_norm": 0.6399952173233032, | |
| "learning_rate": 8.243806611739516e-06, | |
| "loss": 0.1448, | |
| "num_input_tokens_seen": 12290592, | |
| "step": 480 | |
| }, | |
| { | |
| "epoch": 4.222698072805139, | |
| "grad_norm": 0.657546877861023, | |
| "learning_rate": 8.17458259439202e-06, | |
| "loss": 0.144, | |
| "num_input_tokens_seen": 12542464, | |
| "step": 490 | |
| }, | |
| { | |
| "epoch": 4.308351177730192, | |
| "grad_norm": 0.6414650678634644, | |
| "learning_rate": 8.104323730195407e-06, | |
| "loss": 0.1406, | |
| "num_input_tokens_seen": 12796848, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 4.394004282655247, | |
| "grad_norm": 0.7480872869491577, | |
| "learning_rate": 8.033052922054882e-06, | |
| "loss": 0.1436, | |
| "num_input_tokens_seen": 13051760, | |
| "step": 510 | |
| }, | |
| { | |
| "epoch": 4.4796573875803, | |
| "grad_norm": 0.7025752067565918, | |
| "learning_rate": 7.960793402748001e-06, | |
| "loss": 0.147, | |
| "num_input_tokens_seen": 13305808, | |
| "step": 520 | |
| }, | |
| { | |
| "epoch": 4.565310492505353, | |
| "grad_norm": 0.5708986520767212, | |
| "learning_rate": 7.887568727351262e-06, | |
| "loss": 0.1456, | |
| "num_input_tokens_seen": 13563056, | |
| "step": 530 | |
| }, | |
| { | |
| "epoch": 4.650963597430406, | |
| "grad_norm": 0.6903087496757507, | |
| "learning_rate": 7.813402765561664e-06, | |
| "loss": 0.143, | |
| "num_input_tokens_seen": 13816992, | |
| "step": 540 | |
| }, | |
| { | |
| "epoch": 4.736616702355461, | |
| "grad_norm": 0.6083903908729553, | |
| "learning_rate": 7.738319693915673e-06, | |
| "loss": 0.1439, | |
| "num_input_tokens_seen": 14071936, | |
| "step": 550 | |
| }, | |
| { | |
| "epoch": 4.822269807280514, | |
| "grad_norm": 0.6583831906318665, | |
| "learning_rate": 7.662343987908195e-06, | |
| "loss": 0.147, | |
| "num_input_tokens_seen": 14327440, | |
| "step": 560 | |
| }, | |
| { | |
| "epoch": 4.907922912205567, | |
| "grad_norm": 0.8827478885650635, | |
| "learning_rate": 7.585500414014077e-06, | |
| "loss": 0.1467, | |
| "num_input_tokens_seen": 14582832, | |
| "step": 570 | |
| }, | |
| { | |
| "epoch": 4.993576017130621, | |
| "grad_norm": 0.8274891972541809, | |
| "learning_rate": 7.507814021614761e-06, | |
| "loss": 0.1478, | |
| "num_input_tokens_seen": 14839136, | |
| "step": 580 | |
| }, | |
| { | |
| "epoch": 5.085653104925053, | |
| "grad_norm": 1.3195112943649292, | |
| "learning_rate": 7.429310134832709e-06, | |
| "loss": 0.1517, | |
| "num_input_tokens_seen": 15109264, | |
| "step": 590 | |
| }, | |
| { | |
| "epoch": 5.1713062098501075, | |
| "grad_norm": 0.7981224656105042, | |
| "learning_rate": 7.35001434427628e-06, | |
| "loss": 0.1396, | |
| "num_input_tokens_seen": 15363824, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 5.256959314775161, | |
| "grad_norm": 0.6522560715675354, | |
| "learning_rate": 7.269952498697734e-06, | |
| "loss": 0.142, | |
| "num_input_tokens_seen": 15618576, | |
| "step": 610 | |
| }, | |
| { | |
| "epoch": 5.342612419700214, | |
| "grad_norm": 0.7629905343055725, | |
| "learning_rate": 7.189150696567081e-06, | |
| "loss": 0.1384, | |
| "num_input_tokens_seen": 15871056, | |
| "step": 620 | |
| }, | |
| { | |
| "epoch": 5.428265524625267, | |
| "grad_norm": 0.9554848670959473, | |
| "learning_rate": 7.10763527756453e-06, | |
| "loss": 0.1405, | |
| "num_input_tokens_seen": 16124976, | |
| "step": 630 | |
| }, | |
| { | |
| "epoch": 5.5139186295503215, | |
| "grad_norm": 0.8175866603851318, | |
| "learning_rate": 7.025432813994315e-06, | |
| "loss": 0.1357, | |
| "num_input_tokens_seen": 16381680, | |
| "step": 640 | |
| }, | |
| { | |
| "epoch": 5.599571734475375, | |
| "grad_norm": 0.7990790009498596, | |
| "learning_rate": 6.942570102122679e-06, | |
| "loss": 0.1387, | |
| "num_input_tokens_seen": 16638048, | |
| "step": 650 | |
| }, | |
| { | |
| "epoch": 5.685224839400428, | |
| "grad_norm": 0.9116854667663574, | |
| "learning_rate": 6.859074153442864e-06, | |
| "loss": 0.1414, | |
| "num_input_tokens_seen": 16894688, | |
| "step": 660 | |
| }, | |
| { | |
| "epoch": 5.770877944325482, | |
| "grad_norm": 0.7633938789367676, | |
| "learning_rate": 6.774972185869928e-06, | |
| "loss": 0.1389, | |
| "num_input_tokens_seen": 17147808, | |
| "step": 670 | |
| }, | |
| { | |
| "epoch": 5.856531049250536, | |
| "grad_norm": 0.8924551606178284, | |
| "learning_rate": 6.690291614868287e-06, | |
| "loss": 0.1361, | |
| "num_input_tokens_seen": 17403280, | |
| "step": 680 | |
| }, | |
| { | |
| "epoch": 5.942184154175589, | |
| "grad_norm": 0.8566009998321533, | |
| "learning_rate": 6.60506004451485e-06, | |
| "loss": 0.1356, | |
| "num_input_tokens_seen": 17657888, | |
| "step": 690 | |
| }, | |
| { | |
| "epoch": 6.034261241970022, | |
| "grad_norm": 0.9057173132896423, | |
| "learning_rate": 6.5193052585006666e-06, | |
| "loss": 0.1483, | |
| "num_input_tokens_seen": 17927520, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 6.119914346895075, | |
| "grad_norm": 0.9895085692405701, | |
| "learning_rate": 6.433055211074042e-06, | |
| "loss": 0.1308, | |
| "num_input_tokens_seen": 18184352, | |
| "step": 710 | |
| }, | |
| { | |
| "epoch": 6.205567451820128, | |
| "grad_norm": 1.0845868587493896, | |
| "learning_rate": 6.346338017928036e-06, | |
| "loss": 0.1269, | |
| "num_input_tokens_seen": 18437792, | |
| "step": 720 | |
| }, | |
| { | |
| "epoch": 6.291220556745182, | |
| "grad_norm": 1.021283745765686, | |
| "learning_rate": 6.2591819470353424e-06, | |
| "loss": 0.1301, | |
| "num_input_tokens_seen": 18690144, | |
| "step": 730 | |
| }, | |
| { | |
| "epoch": 6.376873661670236, | |
| "grad_norm": 1.1350120306015015, | |
| "learning_rate": 6.171615409433525e-06, | |
| "loss": 0.1275, | |
| "num_input_tokens_seen": 18944688, | |
| "step": 740 | |
| }, | |
| { | |
| "epoch": 6.462526766595289, | |
| "grad_norm": 1.0572874546051025, | |
| "learning_rate": 6.0836669499636255e-06, | |
| "loss": 0.1264, | |
| "num_input_tokens_seen": 19199984, | |
| "step": 750 | |
| }, | |
| { | |
| "epoch": 6.548179871520342, | |
| "grad_norm": 1.1884225606918335, | |
| "learning_rate": 5.995365237965144e-06, | |
| "loss": 0.1294, | |
| "num_input_tokens_seen": 19452032, | |
| "step": 760 | |
| }, | |
| { | |
| "epoch": 6.6338329764453965, | |
| "grad_norm": 0.9745492339134216, | |
| "learning_rate": 5.906739057930439e-06, | |
| "loss": 0.1262, | |
| "num_input_tokens_seen": 19707040, | |
| "step": 770 | |
| }, | |
| { | |
| "epoch": 6.71948608137045, | |
| "grad_norm": 1.090391755104065, | |
| "learning_rate": 5.817817300121592e-06, | |
| "loss": 0.1266, | |
| "num_input_tokens_seen": 19962960, | |
| "step": 780 | |
| }, | |
| { | |
| "epoch": 6.805139186295503, | |
| "grad_norm": 1.1640676259994507, | |
| "learning_rate": 5.728628951152799e-06, | |
| "loss": 0.1324, | |
| "num_input_tokens_seen": 20219008, | |
| "step": 790 | |
| }, | |
| { | |
| "epoch": 6.890792291220556, | |
| "grad_norm": 0.9813507199287415, | |
| "learning_rate": 5.639203084541338e-06, | |
| "loss": 0.1338, | |
| "num_input_tokens_seen": 20473664, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 6.9764453961456105, | |
| "grad_norm": 1.11289644241333, | |
| "learning_rate": 5.549568851230219e-06, | |
| "loss": 0.1273, | |
| "num_input_tokens_seen": 20727296, | |
| "step": 810 | |
| }, | |
| { | |
| "epoch": 7.0685224839400425, | |
| "grad_norm": 1.5624918937683105, | |
| "learning_rate": 5.459755470085595e-06, | |
| "loss": 0.1332, | |
| "num_input_tokens_seen": 20996432, | |
| "step": 820 | |
| }, | |
| { | |
| "epoch": 7.154175588865097, | |
| "grad_norm": 1.3339862823486328, | |
| "learning_rate": 5.369792218372026e-06, | |
| "loss": 0.1104, | |
| "num_input_tokens_seen": 21252272, | |
| "step": 830 | |
| }, | |
| { | |
| "epoch": 7.23982869379015, | |
| "grad_norm": 1.5236716270446777, | |
| "learning_rate": 5.2797084222087105e-06, | |
| "loss": 0.1114, | |
| "num_input_tokens_seen": 21508208, | |
| "step": 840 | |
| }, | |
| { | |
| "epoch": 7.325481798715203, | |
| "grad_norm": 1.4154669046401978, | |
| "learning_rate": 5.189533447009795e-06, | |
| "loss": 0.1134, | |
| "num_input_tokens_seen": 21765536, | |
| "step": 850 | |
| }, | |
| { | |
| "epoch": 7.4111349036402565, | |
| "grad_norm": 1.5260732173919678, | |
| "learning_rate": 5.099296687911858e-06, | |
| "loss": 0.1102, | |
| "num_input_tokens_seen": 22020160, | |
| "step": 860 | |
| }, | |
| { | |
| "epoch": 7.496788008565311, | |
| "grad_norm": 1.2989623546600342, | |
| "learning_rate": 5.009027560191732e-06, | |
| "loss": 0.1122, | |
| "num_input_tokens_seen": 22274400, | |
| "step": 870 | |
| }, | |
| { | |
| "epoch": 7.582441113490364, | |
| "grad_norm": 1.4925442934036255, | |
| "learning_rate": 4.918755489677729e-06, | |
| "loss": 0.1094, | |
| "num_input_tokens_seen": 22526464, | |
| "step": 880 | |
| }, | |
| { | |
| "epoch": 7.668094218415417, | |
| "grad_norm": 1.3059921264648438, | |
| "learning_rate": 4.828509903157451e-06, | |
| "loss": 0.1128, | |
| "num_input_tokens_seen": 22779664, | |
| "step": 890 | |
| }, | |
| { | |
| "epoch": 7.7537473233404715, | |
| "grad_norm": 1.6819276809692383, | |
| "learning_rate": 4.738320218785281e-06, | |
| "loss": 0.1146, | |
| "num_input_tokens_seen": 23036160, | |
| "step": 900 | |
| }, | |
| { | |
| "epoch": 7.839400428265525, | |
| "grad_norm": 1.3909580707550049, | |
| "learning_rate": 4.648215836492682e-06, | |
| "loss": 0.1145, | |
| "num_input_tokens_seen": 23292016, | |
| "step": 910 | |
| }, | |
| { | |
| "epoch": 7.925053533190578, | |
| "grad_norm": 1.7210851907730103, | |
| "learning_rate": 4.5582261284044385e-06, | |
| "loss": 0.1156, | |
| "num_input_tokens_seen": 23544800, | |
| "step": 920 | |
| }, | |
| { | |
| "epoch": 8.01713062098501, | |
| "grad_norm": 1.2723944187164307, | |
| "learning_rate": 4.468380429263973e-06, | |
| "loss": 0.1197, | |
| "num_input_tokens_seen": 23816288, | |
| "step": 930 | |
| }, | |
| { | |
| "epoch": 8.102783725910065, | |
| "grad_norm": 1.9091925621032715, | |
| "learning_rate": 4.378708026870825e-06, | |
| "loss": 0.0916, | |
| "num_input_tokens_seen": 24071488, | |
| "step": 940 | |
| }, | |
| { | |
| "epoch": 8.188436830835117, | |
| "grad_norm": 1.7839370965957642, | |
| "learning_rate": 4.289238152533465e-06, | |
| "loss": 0.0893, | |
| "num_input_tokens_seen": 24324720, | |
| "step": 950 | |
| }, | |
| { | |
| "epoch": 8.274089935760172, | |
| "grad_norm": 2.00311541557312, | |
| "learning_rate": 4.199999971540489e-06, | |
| "loss": 0.0889, | |
| "num_input_tokens_seen": 24579648, | |
| "step": 960 | |
| }, | |
| { | |
| "epoch": 8.359743040685224, | |
| "grad_norm": 2.047337293624878, | |
| "learning_rate": 4.111022573653366e-06, | |
| "loss": 0.0873, | |
| "num_input_tokens_seen": 24833840, | |
| "step": 970 | |
| }, | |
| { | |
| "epoch": 8.445396145610278, | |
| "grad_norm": 1.9115785360336304, | |
| "learning_rate": 4.0223349636237766e-06, | |
| "loss": 0.0904, | |
| "num_input_tokens_seen": 25089776, | |
| "step": 980 | |
| }, | |
| { | |
| "epoch": 8.531049250535332, | |
| "grad_norm": 1.8445810079574585, | |
| "learning_rate": 3.933966051738684e-06, | |
| "loss": 0.088, | |
| "num_input_tokens_seen": 25345264, | |
| "step": 990 | |
| }, | |
| { | |
| "epoch": 8.616702355460385, | |
| "grad_norm": 1.6529115438461304, | |
| "learning_rate": 3.845944644396194e-06, | |
| "loss": 0.0919, | |
| "num_input_tokens_seen": 25598112, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 8.702355460385439, | |
| "grad_norm": 2.129995346069336, | |
| "learning_rate": 3.758299434715268e-06, | |
| "loss": 0.0906, | |
| "num_input_tokens_seen": 25851728, | |
| "step": 1010 | |
| }, | |
| { | |
| "epoch": 8.788008565310493, | |
| "grad_norm": 2.1039373874664307, | |
| "learning_rate": 3.6710589931823837e-06, | |
| "loss": 0.0895, | |
| "num_input_tokens_seen": 26104704, | |
| "step": 1020 | |
| }, | |
| { | |
| "epoch": 8.873661670235546, | |
| "grad_norm": 2.058598518371582, | |
| "learning_rate": 3.584251758338151e-06, | |
| "loss": 0.0923, | |
| "num_input_tokens_seen": 26361680, | |
| "step": 1030 | |
| }, | |
| { | |
| "epoch": 8.9593147751606, | |
| "grad_norm": 1.8930065631866455, | |
| "learning_rate": 3.4979060275069576e-06, | |
| "loss": 0.0908, | |
| "num_input_tokens_seen": 26617536, | |
| "step": 1040 | |
| }, | |
| { | |
| "epoch": 9.051391862955033, | |
| "grad_norm": 1.8233646154403687, | |
| "learning_rate": 3.4120499475726266e-06, | |
| "loss": 0.0847, | |
| "num_input_tokens_seen": 26888160, | |
| "step": 1050 | |
| }, | |
| { | |
| "epoch": 9.137044967880085, | |
| "grad_norm": 2.1758053302764893, | |
| "learning_rate": 3.3267115058031418e-06, | |
| "loss": 0.0657, | |
| "num_input_tokens_seen": 27142528, | |
| "step": 1060 | |
| }, | |
| { | |
| "epoch": 9.222698072805139, | |
| "grad_norm": 2.0327367782592773, | |
| "learning_rate": 3.2419185207273816e-06, | |
| "loss": 0.0662, | |
| "num_input_tokens_seen": 27394144, | |
| "step": 1070 | |
| }, | |
| { | |
| "epoch": 9.308351177730193, | |
| "grad_norm": 2.2035434246063232, | |
| "learning_rate": 3.157698633066863e-06, | |
| "loss": 0.0665, | |
| "num_input_tokens_seen": 27649488, | |
| "step": 1080 | |
| }, | |
| { | |
| "epoch": 9.394004282655246, | |
| "grad_norm": 1.9066494703292847, | |
| "learning_rate": 3.0740792967254606e-06, | |
| "loss": 0.0642, | |
| "num_input_tokens_seen": 27904992, | |
| "step": 1090 | |
| }, | |
| { | |
| "epoch": 9.4796573875803, | |
| "grad_norm": 2.2175674438476562, | |
| "learning_rate": 2.991087769840001e-06, | |
| "loss": 0.0625, | |
| "num_input_tokens_seen": 28160336, | |
| "step": 1100 | |
| }, | |
| { | |
| "epoch": 9.565310492505354, | |
| "grad_norm": 2.435115337371826, | |
| "learning_rate": 2.9087511058947014e-06, | |
| "loss": 0.0643, | |
| "num_input_tokens_seen": 28417360, | |
| "step": 1110 | |
| }, | |
| { | |
| "epoch": 9.650963597430406, | |
| "grad_norm": 2.237015724182129, | |
| "learning_rate": 2.827096144902289e-06, | |
| "loss": 0.0645, | |
| "num_input_tokens_seen": 28670512, | |
| "step": 1120 | |
| }, | |
| { | |
| "epoch": 9.73661670235546, | |
| "grad_norm": 2.473604202270508, | |
| "learning_rate": 2.7461495046547436e-06, | |
| "loss": 0.068, | |
| "num_input_tokens_seen": 28927232, | |
| "step": 1130 | |
| }, | |
| { | |
| "epoch": 9.822269807280513, | |
| "grad_norm": 2.220705270767212, | |
| "learning_rate": 2.665937572046432e-06, | |
| "loss": 0.0647, | |
| "num_input_tokens_seen": 29182768, | |
| "step": 1140 | |
| }, | |
| { | |
| "epoch": 9.907922912205567, | |
| "grad_norm": 2.652024269104004, | |
| "learning_rate": 2.586486494472572e-06, | |
| "loss": 0.0644, | |
| "num_input_tokens_seen": 29437936, | |
| "step": 1150 | |
| }, | |
| { | |
| "epoch": 9.993576017130621, | |
| "grad_norm": 2.180983304977417, | |
| "learning_rate": 2.5078221713057048e-06, | |
| "loss": 0.0658, | |
| "num_input_tokens_seen": 29690944, | |
| "step": 1160 | |
| }, | |
| { | |
| "epoch": 10.085653104925054, | |
| "grad_norm": 1.6538355350494385, | |
| "learning_rate": 2.4299702454530605e-06, | |
| "loss": 0.053, | |
| "num_input_tokens_seen": 29964448, | |
| "step": 1170 | |
| }, | |
| { | |
| "epoch": 10.171306209850107, | |
| "grad_norm": 2.229673147201538, | |
| "learning_rate": 2.3529560949975184e-06, | |
| "loss": 0.0446, | |
| "num_input_tokens_seen": 30215952, | |
| "step": 1180 | |
| }, | |
| { | |
| "epoch": 10.25695931477516, | |
| "grad_norm": 1.8106822967529297, | |
| "learning_rate": 2.2768048249248648e-06, | |
| "loss": 0.0449, | |
| "num_input_tokens_seen": 30471952, | |
| "step": 1190 | |
| }, | |
| { | |
| "epoch": 10.342612419700215, | |
| "grad_norm": 2.150508403778076, | |
| "learning_rate": 2.201541258940129e-06, | |
| "loss": 0.0422, | |
| "num_input_tokens_seen": 30727376, | |
| "step": 1200 | |
| }, | |
| { | |
| "epoch": 10.428265524625267, | |
| "grad_norm": 2.0471906661987305, | |
| "learning_rate": 2.12718993137555e-06, | |
| "loss": 0.0461, | |
| "num_input_tokens_seen": 30983760, | |
| "step": 1210 | |
| }, | |
| { | |
| "epoch": 10.513918629550322, | |
| "grad_norm": 2.299278497695923, | |
| "learning_rate": 2.0537750791929296e-06, | |
| "loss": 0.0458, | |
| "num_input_tokens_seen": 31238720, | |
| "step": 1220 | |
| }, | |
| { | |
| "epoch": 10.599571734475374, | |
| "grad_norm": 2.1924257278442383, | |
| "learning_rate": 1.981320634082873e-06, | |
| "loss": 0.0434, | |
| "num_input_tokens_seen": 31494560, | |
| "step": 1230 | |
| }, | |
| { | |
| "epoch": 10.685224839400428, | |
| "grad_norm": 2.3524584770202637, | |
| "learning_rate": 1.909850214663575e-06, | |
| "loss": 0.0452, | |
| "num_input_tokens_seen": 31750784, | |
| "step": 1240 | |
| }, | |
| { | |
| "epoch": 10.770877944325482, | |
| "grad_norm": 2.2468934059143066, | |
| "learning_rate": 1.8393871187816526e-06, | |
| "loss": 0.0447, | |
| "num_input_tokens_seen": 32005120, | |
| "step": 1250 | |
| }, | |
| { | |
| "epoch": 10.856531049250535, | |
| "grad_norm": 2.448117971420288, | |
| "learning_rate": 1.7699543159175215e-06, | |
| "loss": 0.0449, | |
| "num_input_tokens_seen": 32258480, | |
| "step": 1260 | |
| }, | |
| { | |
| "epoch": 10.942184154175589, | |
| "grad_norm": 2.0848143100738525, | |
| "learning_rate": 1.7015744396978557e-06, | |
| "loss": 0.0442, | |
| "num_input_tokens_seen": 32510944, | |
| "step": 1270 | |
| }, | |
| { | |
| "epoch": 11.034261241970022, | |
| "grad_norm": 1.6036432981491089, | |
| "learning_rate": 1.634269780517483e-06, | |
| "loss": 0.0435, | |
| "num_input_tokens_seen": 32780608, | |
| "step": 1280 | |
| }, | |
| { | |
| "epoch": 11.119914346895074, | |
| "grad_norm": 3.015963315963745, | |
| "learning_rate": 1.568062278273197e-06, | |
| "loss": 0.0286, | |
| "num_input_tokens_seen": 33034112, | |
| "step": 1290 | |
| }, | |
| { | |
| "epoch": 11.205567451820128, | |
| "grad_norm": 1.6929532289505005, | |
| "learning_rate": 1.5029735152118125e-06, | |
| "loss": 0.0308, | |
| "num_input_tokens_seen": 33290224, | |
| "step": 1300 | |
| }, | |
| { | |
| "epoch": 11.291220556745182, | |
| "grad_norm": 1.9741885662078857, | |
| "learning_rate": 1.4390247088948073e-06, | |
| "loss": 0.0309, | |
| "num_input_tokens_seen": 33544448, | |
| "step": 1310 | |
| }, | |
| { | |
| "epoch": 11.376873661670235, | |
| "grad_norm": 1.5955508947372437, | |
| "learning_rate": 1.3762367052818527e-06, | |
| "loss": 0.0275, | |
| "num_input_tokens_seen": 33799536, | |
| "step": 1320 | |
| }, | |
| { | |
| "epoch": 11.462526766595289, | |
| "grad_norm": 2.293123245239258, | |
| "learning_rate": 1.3146299719354544e-06, | |
| "loss": 0.0304, | |
| "num_input_tokens_seen": 34055952, | |
| "step": 1330 | |
| }, | |
| { | |
| "epoch": 11.548179871520343, | |
| "grad_norm": 1.8011912107467651, | |
| "learning_rate": 1.254224591348983e-06, | |
| "loss": 0.0299, | |
| "num_input_tokens_seen": 34310000, | |
| "step": 1340 | |
| }, | |
| { | |
| "epoch": 11.633832976445396, | |
| "grad_norm": 1.8339879512786865, | |
| "learning_rate": 1.1950402544001849e-06, | |
| "loss": 0.0311, | |
| "num_input_tokens_seen": 34565680, | |
| "step": 1350 | |
| }, | |
| { | |
| "epoch": 11.71948608137045, | |
| "grad_norm": 1.6808807849884033, | |
| "learning_rate": 1.1370962539323837e-06, | |
| "loss": 0.0314, | |
| "num_input_tokens_seen": 34820768, | |
| "step": 1360 | |
| }, | |
| { | |
| "epoch": 11.805139186295504, | |
| "grad_norm": 1.7647879123687744, | |
| "learning_rate": 1.0804114784654158e-06, | |
| "loss": 0.0311, | |
| "num_input_tokens_seen": 35074016, | |
| "step": 1370 | |
| }, | |
| { | |
| "epoch": 11.890792291220556, | |
| "grad_norm": 1.753990650177002, | |
| "learning_rate": 1.0250044060383734e-06, | |
| "loss": 0.0299, | |
| "num_input_tokens_seen": 35328272, | |
| "step": 1380 | |
| }, | |
| { | |
| "epoch": 11.97644539614561, | |
| "grad_norm": 2.10841965675354, | |
| "learning_rate": 9.708930981861603e-07, | |
| "loss": 0.03, | |
| "num_input_tokens_seen": 35582880, | |
| "step": 1390 | |
| }, | |
| { | |
| "epoch": 12.068522483940043, | |
| "grad_norm": 1.4194451570510864, | |
| "learning_rate": 9.180951940518002e-07, | |
| "loss": 0.026, | |
| "num_input_tokens_seen": 35853280, | |
| "step": 1400 | |
| }, | |
| { | |
| "epoch": 12.154175588865096, | |
| "grad_norm": 1.612318515777588, | |
| "learning_rate": 8.666279046364595e-07, | |
| "loss": 0.0208, | |
| "num_input_tokens_seen": 36106816, | |
| "step": 1410 | |
| }, | |
| { | |
| "epoch": 12.23982869379015, | |
| "grad_norm": 1.6022765636444092, | |
| "learning_rate": 8.165080071890208e-07, | |
| "loss": 0.0205, | |
| "num_input_tokens_seen": 36359232, | |
| "step": 1420 | |
| }, | |
| { | |
| "epoch": 12.325481798715204, | |
| "grad_norm": 1.608430027961731, | |
| "learning_rate": 7.677518397370548e-07, | |
| "loss": 0.0228, | |
| "num_input_tokens_seen": 36614176, | |
| "step": 1430 | |
| }, | |
| { | |
| "epoch": 12.411134903640257, | |
| "grad_norm": 1.4423803091049194, | |
| "learning_rate": 7.203752957609672e-07, | |
| "loss": 0.0207, | |
| "num_input_tokens_seen": 36868400, | |
| "step": 1440 | |
| }, | |
| { | |
| "epoch": 12.49678800856531, | |
| "grad_norm": 1.6684809923171997, | |
| "learning_rate": 6.743938190130616e-07, | |
| "loss": 0.0215, | |
| "num_input_tokens_seen": 37121536, | |
| "step": 1450 | |
| }, | |
| { | |
| "epoch": 12.582441113490365, | |
| "grad_norm": 1.7179003953933716, | |
| "learning_rate": 6.298223984832047e-07, | |
| "loss": 0.0216, | |
| "num_input_tokens_seen": 37377168, | |
| "step": 1460 | |
| }, | |
| { | |
| "epoch": 12.668094218415417, | |
| "grad_norm": 1.6454778909683228, | |
| "learning_rate": 5.866755635127247e-07, | |
| "loss": 0.0207, | |
| "num_input_tokens_seen": 37632992, | |
| "step": 1470 | |
| }, | |
| { | |
| "epoch": 12.753747323340471, | |
| "grad_norm": 1.8044767379760742, | |
| "learning_rate": 5.449673790581611e-07, | |
| "loss": 0.0217, | |
| "num_input_tokens_seen": 37888640, | |
| "step": 1480 | |
| }, | |
| { | |
| "epoch": 12.839400428265524, | |
| "grad_norm": 1.874295711517334, | |
| "learning_rate": 5.04711441106382e-07, | |
| "loss": 0.0197, | |
| "num_input_tokens_seen": 38143760, | |
| "step": 1490 | |
| }, | |
| { | |
| "epoch": 12.925053533190578, | |
| "grad_norm": 1.3250926733016968, | |
| "learning_rate": 4.659208722425806e-07, | |
| "loss": 0.0207, | |
| "num_input_tokens_seen": 38398560, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 13.01713062098501, | |
| "grad_norm": 1.2411588430404663, | |
| "learning_rate": 4.2860831737258857e-07, | |
| "loss": 0.0216, | |
| "num_input_tokens_seen": 38670912, | |
| "step": 1510 | |
| }, | |
| { | |
| "epoch": 13.102783725910065, | |
| "grad_norm": 1.3138427734375, | |
| "learning_rate": 3.9278593960090873e-07, | |
| "loss": 0.0167, | |
| "num_input_tokens_seen": 38925872, | |
| "step": 1520 | |
| }, | |
| { | |
| "epoch": 13.188436830835117, | |
| "grad_norm": 1.362457036972046, | |
| "learning_rate": 3.5846541626579026e-07, | |
| "loss": 0.0159, | |
| "num_input_tokens_seen": 39183632, | |
| "step": 1530 | |
| }, | |
| { | |
| "epoch": 13.274089935760172, | |
| "grad_norm": 1.515376091003418, | |
| "learning_rate": 3.256579351326744e-07, | |
| "loss": 0.0156, | |
| "num_input_tokens_seen": 39440864, | |
| "step": 1540 | |
| }, | |
| { | |
| "epoch": 13.359743040685224, | |
| "grad_norm": 1.4070255756378174, | |
| "learning_rate": 2.94374190747212e-07, | |
| "loss": 0.0166, | |
| "num_input_tokens_seen": 39695712, | |
| "step": 1550 | |
| }, | |
| { | |
| "epoch": 13.445396145610278, | |
| "grad_norm": 1.4853448867797852, | |
| "learning_rate": 2.64624380949069e-07, | |
| "loss": 0.0173, | |
| "num_input_tokens_seen": 39950304, | |
| "step": 1560 | |
| }, | |
| { | |
| "epoch": 13.531049250535332, | |
| "grad_norm": 1.542286992073059, | |
| "learning_rate": 2.3641820354764755e-07, | |
| "loss": 0.0165, | |
| "num_input_tokens_seen": 40203616, | |
| "step": 1570 | |
| }, | |
| { | |
| "epoch": 13.616702355460385, | |
| "grad_norm": 1.565663456916809, | |
| "learning_rate": 2.0976485316080375e-07, | |
| "loss": 0.0167, | |
| "num_input_tokens_seen": 40458464, | |
| "step": 1580 | |
| }, | |
| { | |
| "epoch": 13.702355460385439, | |
| "grad_norm": 1.3701163530349731, | |
| "learning_rate": 1.846730182175993e-07, | |
| "loss": 0.017, | |
| "num_input_tokens_seen": 40711216, | |
| "step": 1590 | |
| }, | |
| { | |
| "epoch": 13.788008565310493, | |
| "grad_norm": 1.4886751174926758, | |
| "learning_rate": 1.6115087812605123e-07, | |
| "loss": 0.015, | |
| "num_input_tokens_seen": 40965856, | |
| "step": 1600 | |
| }, | |
| { | |
| "epoch": 13.873661670235546, | |
| "grad_norm": 1.2140471935272217, | |
| "learning_rate": 1.392061006068246e-07, | |
| "loss": 0.0169, | |
| "num_input_tokens_seen": 41220736, | |
| "step": 1610 | |
| }, | |
| { | |
| "epoch": 13.9593147751606, | |
| "grad_norm": 1.314063549041748, | |
| "learning_rate": 1.1884583919371251e-07, | |
| "loss": 0.0164, | |
| "num_input_tokens_seen": 41473952, | |
| "step": 1620 | |
| }, | |
| { | |
| "epoch": 14.051391862955033, | |
| "grad_norm": 1.2103674411773682, | |
| "learning_rate": 1.0007673090173808e-07, | |
| "loss": 0.0168, | |
| "num_input_tokens_seen": 41742832, | |
| "step": 1630 | |
| }, | |
| { | |
| "epoch": 14.137044967880085, | |
| "grad_norm": 1.250216007232666, | |
| "learning_rate": 8.29048940636279e-08, | |
| "loss": 0.0153, | |
| "num_input_tokens_seen": 41998320, | |
| "step": 1640 | |
| }, | |
| { | |
| "epoch": 14.222698072805139, | |
| "grad_norm": 1.114964485168457, | |
| "learning_rate": 6.733592633536124e-08, | |
| "loss": 0.0148, | |
| "num_input_tokens_seen": 42253104, | |
| "step": 1650 | |
| }, | |
| { | |
| "epoch": 14.308351177730193, | |
| "grad_norm": 1.3133609294891357, | |
| "learning_rate": 5.3374902871456965e-08, | |
| "loss": 0.0151, | |
| "num_input_tokens_seen": 42509584, | |
| "step": 1660 | |
| }, | |
| { | |
| "epoch": 14.394004282655246, | |
| "grad_norm": 1.3046901226043701, | |
| "learning_rate": 4.102637467057746e-08, | |
| "loss": 0.0144, | |
| "num_input_tokens_seen": 42764768, | |
| "step": 1670 | |
| }, | |
| { | |
| "epoch": 14.4796573875803, | |
| "grad_norm": 1.3270611763000488, | |
| "learning_rate": 3.029436709200084e-08, | |
| "loss": 0.0142, | |
| "num_input_tokens_seen": 43019376, | |
| "step": 1680 | |
| }, | |
| { | |
| "epoch": 14.565310492505354, | |
| "grad_norm": 1.1487038135528564, | |
| "learning_rate": 2.1182378543438408e-08, | |
| "loss": 0.0159, | |
| "num_input_tokens_seen": 43273248, | |
| "step": 1690 | |
| }, | |
| { | |
| "epoch": 14.650963597430406, | |
| "grad_norm": 1.1392930746078491, | |
| "learning_rate": 1.3693379340626867e-08, | |
| "loss": 0.0148, | |
| "num_input_tokens_seen": 43529200, | |
| "step": 1700 | |
| }, | |
| { | |
| "epoch": 14.73661670235546, | |
| "grad_norm": 1.24246084690094, | |
| "learning_rate": 7.829810739069521e-09, | |
| "loss": 0.0144, | |
| "num_input_tokens_seen": 43781760, | |
| "step": 1710 | |
| }, | |
| { | |
| "epoch": 14.822269807280513, | |
| "grad_norm": 1.2764571905136108, | |
| "learning_rate": 3.593584138237294e-09, | |
| "loss": 0.0142, | |
| "num_input_tokens_seen": 44036144, | |
| "step": 1720 | |
| }, | |
| { | |
| "epoch": 14.907922912205567, | |
| "grad_norm": 1.4254299402236938, | |
| "learning_rate": 9.860804584937988e-10, | |
| "loss": 0.0144, | |
| "num_input_tokens_seen": 44292256, | |
| "step": 1730 | |
| }, | |
| { | |
| "epoch": 14.993576017130621, | |
| "grad_norm": 1.1011109352111816, | |
| "learning_rate": 8.149690943204391e-12, | |
| "loss": 0.014, | |
| "num_input_tokens_seen": 44548112, | |
| "step": 1740 | |
| } | |
| ], | |
| "logging_steps": 10, | |
| "max_steps": 1740, | |
| "num_input_tokens_seen": 44548112, | |
| "num_train_epochs": 15, | |
| "save_steps": 500, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": true | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 1.9454424851110953e+18, | |
| "train_batch_size": 2, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |