| { |
| "best_global_step": null, |
| "best_metric": null, |
| "best_model_checkpoint": null, |
| "epoch": 2.7181719260065287, |
| "eval_steps": 500, |
| "global_step": 2000, |
| "is_hyper_param_search": false, |
| "is_local_process_zero": true, |
| "is_world_process_zero": true, |
| "log_history": [ |
| { |
| "epoch": 0.013601741022850925, |
| "grad_norm": 2.812274217605591, |
| "learning_rate": 2.0000000000000003e-06, |
| "loss": 5.2086, |
| "step": 10 |
| }, |
| { |
| "epoch": 0.02720348204570185, |
| "grad_norm": 2.162468433380127, |
| "learning_rate": 4.000000000000001e-06, |
| "loss": 5.2051, |
| "step": 20 |
| }, |
| { |
| "epoch": 0.040805223068552776, |
| "grad_norm": 1.7208396196365356, |
| "learning_rate": 6e-06, |
| "loss": 5.2105, |
| "step": 30 |
| }, |
| { |
| "epoch": 0.0544069640914037, |
| "grad_norm": 2.556692361831665, |
| "learning_rate": 8.000000000000001e-06, |
| "loss": 5.1625, |
| "step": 40 |
| }, |
| { |
| "epoch": 0.06800870511425462, |
| "grad_norm": 8.235758781433105, |
| "learning_rate": 1e-05, |
| "loss": 4.6379, |
| "step": 50 |
| }, |
| { |
| "epoch": 0.08161044613710555, |
| "grad_norm": 10.163893699645996, |
| "learning_rate": 1.2e-05, |
| "loss": 3.8391, |
| "step": 60 |
| }, |
| { |
| "epoch": 0.09521218715995647, |
| "grad_norm": 11.116437911987305, |
| "learning_rate": 1.4000000000000001e-05, |
| "loss": 3.0016, |
| "step": 70 |
| }, |
| { |
| "epoch": 0.1088139281828074, |
| "grad_norm": 10.696512222290039, |
| "learning_rate": 1.6000000000000003e-05, |
| "loss": 2.4004, |
| "step": 80 |
| }, |
| { |
| "epoch": 0.12241566920565833, |
| "grad_norm": 12.367400169372559, |
| "learning_rate": 1.8e-05, |
| "loss": 1.8969, |
| "step": 90 |
| }, |
| { |
| "epoch": 0.13601741022850924, |
| "grad_norm": 13.723505973815918, |
| "learning_rate": 2e-05, |
| "loss": 1.4616, |
| "step": 100 |
| }, |
| { |
| "epoch": 0.14961915125136016, |
| "grad_norm": 22.034290313720703, |
| "learning_rate": 2.2000000000000003e-05, |
| "loss": 1.2379, |
| "step": 110 |
| }, |
| { |
| "epoch": 0.1632208922742111, |
| "grad_norm": 8.55447769165039, |
| "learning_rate": 2.4e-05, |
| "loss": 1.0951, |
| "step": 120 |
| }, |
| { |
| "epoch": 0.17682263329706202, |
| "grad_norm": 11.120705604553223, |
| "learning_rate": 2.6000000000000002e-05, |
| "loss": 0.9654, |
| "step": 130 |
| }, |
| { |
| "epoch": 0.19042437431991294, |
| "grad_norm": 13.087006568908691, |
| "learning_rate": 2.8000000000000003e-05, |
| "loss": 0.8391, |
| "step": 140 |
| }, |
| { |
| "epoch": 0.20402611534276388, |
| "grad_norm": 12.556931495666504, |
| "learning_rate": 3e-05, |
| "loss": 0.7075, |
| "step": 150 |
| }, |
| { |
| "epoch": 0.2176278563656148, |
| "grad_norm": 14.245359420776367, |
| "learning_rate": 3.2000000000000005e-05, |
| "loss": 0.6853, |
| "step": 160 |
| }, |
| { |
| "epoch": 0.2312295973884657, |
| "grad_norm": 12.10800838470459, |
| "learning_rate": 3.4000000000000007e-05, |
| "loss": 0.606, |
| "step": 170 |
| }, |
| { |
| "epoch": 0.24483133841131666, |
| "grad_norm": 11.180713653564453, |
| "learning_rate": 3.6e-05, |
| "loss": 0.5734, |
| "step": 180 |
| }, |
| { |
| "epoch": 0.2584330794341676, |
| "grad_norm": 12.638975143432617, |
| "learning_rate": 3.8e-05, |
| "loss": 0.545, |
| "step": 190 |
| }, |
| { |
| "epoch": 0.2720348204570185, |
| "grad_norm": 11.650677680969238, |
| "learning_rate": 4e-05, |
| "loss": 0.5162, |
| "step": 200 |
| }, |
| { |
| "epoch": 0.28563656147986943, |
| "grad_norm": 11.61837387084961, |
| "learning_rate": 4.2e-05, |
| "loss": 0.491, |
| "step": 210 |
| }, |
| { |
| "epoch": 0.2992383025027203, |
| "grad_norm": 5.453939914703369, |
| "learning_rate": 4.4000000000000006e-05, |
| "loss": 0.4378, |
| "step": 220 |
| }, |
| { |
| "epoch": 0.31284004352557127, |
| "grad_norm": 7.908788204193115, |
| "learning_rate": 4.600000000000001e-05, |
| "loss": 0.4299, |
| "step": 230 |
| }, |
| { |
| "epoch": 0.3264417845484222, |
| "grad_norm": 13.816142082214355, |
| "learning_rate": 4.8e-05, |
| "loss": 0.4113, |
| "step": 240 |
| }, |
| { |
| "epoch": 0.3400435255712731, |
| "grad_norm": 11.430989265441895, |
| "learning_rate": 5e-05, |
| "loss": 0.4371, |
| "step": 250 |
| }, |
| { |
| "epoch": 0.35364526659412404, |
| "grad_norm": 6.766596794128418, |
| "learning_rate": 5.2000000000000004e-05, |
| "loss": 0.387, |
| "step": 260 |
| }, |
| { |
| "epoch": 0.367247007616975, |
| "grad_norm": 10.817462921142578, |
| "learning_rate": 5.4000000000000005e-05, |
| "loss": 0.3643, |
| "step": 270 |
| }, |
| { |
| "epoch": 0.3808487486398259, |
| "grad_norm": 6.278716564178467, |
| "learning_rate": 5.6000000000000006e-05, |
| "loss": 0.3477, |
| "step": 280 |
| }, |
| { |
| "epoch": 0.3944504896626768, |
| "grad_norm": 13.244380950927734, |
| "learning_rate": 5.8e-05, |
| "loss": 0.3763, |
| "step": 290 |
| }, |
| { |
| "epoch": 0.40805223068552776, |
| "grad_norm": 4.513805866241455, |
| "learning_rate": 6e-05, |
| "loss": 0.3093, |
| "step": 300 |
| }, |
| { |
| "epoch": 0.42165397170837865, |
| "grad_norm": 4.978238582611084, |
| "learning_rate": 6.2e-05, |
| "loss": 0.306, |
| "step": 310 |
| }, |
| { |
| "epoch": 0.4352557127312296, |
| "grad_norm": 9.654162406921387, |
| "learning_rate": 6.400000000000001e-05, |
| "loss": 0.3157, |
| "step": 320 |
| }, |
| { |
| "epoch": 0.44885745375408054, |
| "grad_norm": 5.806159496307373, |
| "learning_rate": 6.6e-05, |
| "loss": 0.302, |
| "step": 330 |
| }, |
| { |
| "epoch": 0.4624591947769314, |
| "grad_norm": 6.283710479736328, |
| "learning_rate": 6.800000000000001e-05, |
| "loss": 0.2885, |
| "step": 340 |
| }, |
| { |
| "epoch": 0.47606093579978237, |
| "grad_norm": 6.1870951652526855, |
| "learning_rate": 7e-05, |
| "loss": 0.289, |
| "step": 350 |
| }, |
| { |
| "epoch": 0.4896626768226333, |
| "grad_norm": 3.082080364227295, |
| "learning_rate": 7.2e-05, |
| "loss": 0.2584, |
| "step": 360 |
| }, |
| { |
| "epoch": 0.5032644178454843, |
| "grad_norm": 5.254792213439941, |
| "learning_rate": 7.4e-05, |
| "loss": 0.2752, |
| "step": 370 |
| }, |
| { |
| "epoch": 0.5168661588683352, |
| "grad_norm": 6.462503433227539, |
| "learning_rate": 7.6e-05, |
| "loss": 0.2639, |
| "step": 380 |
| }, |
| { |
| "epoch": 0.530467899891186, |
| "grad_norm": 10.89343547821045, |
| "learning_rate": 7.800000000000001e-05, |
| "loss": 0.2893, |
| "step": 390 |
| }, |
| { |
| "epoch": 0.544069640914037, |
| "grad_norm": 5.2178192138671875, |
| "learning_rate": 8e-05, |
| "loss": 0.2415, |
| "step": 400 |
| }, |
| { |
| "epoch": 0.5576713819368879, |
| "grad_norm": 8.687518119812012, |
| "learning_rate": 8.2e-05, |
| "loss": 0.2385, |
| "step": 410 |
| }, |
| { |
| "epoch": 0.5712731229597389, |
| "grad_norm": 4.670180320739746, |
| "learning_rate": 8.4e-05, |
| "loss": 0.2297, |
| "step": 420 |
| }, |
| { |
| "epoch": 0.5848748639825898, |
| "grad_norm": 3.699598550796509, |
| "learning_rate": 8.6e-05, |
| "loss": 0.2241, |
| "step": 430 |
| }, |
| { |
| "epoch": 0.5984766050054406, |
| "grad_norm": 3.260232925415039, |
| "learning_rate": 8.800000000000001e-05, |
| "loss": 0.216, |
| "step": 440 |
| }, |
| { |
| "epoch": 0.6120783460282916, |
| "grad_norm": 3.4559569358825684, |
| "learning_rate": 9e-05, |
| "loss": 0.2275, |
| "step": 450 |
| }, |
| { |
| "epoch": 0.6256800870511425, |
| "grad_norm": 3.7167000770568848, |
| "learning_rate": 9.200000000000001e-05, |
| "loss": 0.2133, |
| "step": 460 |
| }, |
| { |
| "epoch": 0.6392818280739935, |
| "grad_norm": 4.1776628494262695, |
| "learning_rate": 9.4e-05, |
| "loss": 0.2194, |
| "step": 470 |
| }, |
| { |
| "epoch": 0.6528835690968444, |
| "grad_norm": 4.609129428863525, |
| "learning_rate": 9.6e-05, |
| "loss": 0.2088, |
| "step": 480 |
| }, |
| { |
| "epoch": 0.6664853101196954, |
| "grad_norm": 5.062413215637207, |
| "learning_rate": 9.8e-05, |
| "loss": 0.2185, |
| "step": 490 |
| }, |
| { |
| "epoch": 0.6800870511425462, |
| "grad_norm": 3.3144760131835938, |
| "learning_rate": 0.0001, |
| "loss": 0.2176, |
| "step": 500 |
| }, |
| { |
| "epoch": 0.6936887921653971, |
| "grad_norm": 2.5662150382995605, |
| "learning_rate": 0.00010200000000000001, |
| "loss": 0.2108, |
| "step": 510 |
| }, |
| { |
| "epoch": 0.7072905331882481, |
| "grad_norm": 2.718777656555176, |
| "learning_rate": 0.00010400000000000001, |
| "loss": 0.1734, |
| "step": 520 |
| }, |
| { |
| "epoch": 0.720892274211099, |
| "grad_norm": 2.934107780456543, |
| "learning_rate": 0.00010600000000000002, |
| "loss": 0.209, |
| "step": 530 |
| }, |
| { |
| "epoch": 0.73449401523395, |
| "grad_norm": 2.3642656803131104, |
| "learning_rate": 0.00010800000000000001, |
| "loss": 0.1958, |
| "step": 540 |
| }, |
| { |
| "epoch": 0.7480957562568009, |
| "grad_norm": 2.765012502670288, |
| "learning_rate": 0.00011000000000000002, |
| "loss": 0.1857, |
| "step": 550 |
| }, |
| { |
| "epoch": 0.7616974972796517, |
| "grad_norm": 2.482921600341797, |
| "learning_rate": 0.00011200000000000001, |
| "loss": 0.186, |
| "step": 560 |
| }, |
| { |
| "epoch": 0.7752992383025027, |
| "grad_norm": 2.26837420463562, |
| "learning_rate": 0.00011399999999999999, |
| "loss": 0.1783, |
| "step": 570 |
| }, |
| { |
| "epoch": 0.7889009793253536, |
| "grad_norm": 2.1319069862365723, |
| "learning_rate": 0.000116, |
| "loss": 0.1731, |
| "step": 580 |
| }, |
| { |
| "epoch": 0.8025027203482046, |
| "grad_norm": 1.4668622016906738, |
| "learning_rate": 0.000118, |
| "loss": 0.1854, |
| "step": 590 |
| }, |
| { |
| "epoch": 0.8161044613710555, |
| "grad_norm": 2.2079055309295654, |
| "learning_rate": 0.00012, |
| "loss": 0.1806, |
| "step": 600 |
| }, |
| { |
| "epoch": 0.8297062023939065, |
| "grad_norm": 1.9991375207901, |
| "learning_rate": 0.000122, |
| "loss": 0.1682, |
| "step": 610 |
| }, |
| { |
| "epoch": 0.8433079434167573, |
| "grad_norm": 1.889404296875, |
| "learning_rate": 0.000124, |
| "loss": 0.162, |
| "step": 620 |
| }, |
| { |
| "epoch": 0.8569096844396082, |
| "grad_norm": 1.7021093368530273, |
| "learning_rate": 0.000126, |
| "loss": 0.1521, |
| "step": 630 |
| }, |
| { |
| "epoch": 0.8705114254624592, |
| "grad_norm": 2.153775215148926, |
| "learning_rate": 0.00012800000000000002, |
| "loss": 0.1656, |
| "step": 640 |
| }, |
| { |
| "epoch": 0.8841131664853101, |
| "grad_norm": 2.285336494445801, |
| "learning_rate": 0.00013000000000000002, |
| "loss": 0.1609, |
| "step": 650 |
| }, |
| { |
| "epoch": 0.8977149075081611, |
| "grad_norm": 1.8330312967300415, |
| "learning_rate": 0.000132, |
| "loss": 0.151, |
| "step": 660 |
| }, |
| { |
| "epoch": 0.911316648531012, |
| "grad_norm": 1.712404489517212, |
| "learning_rate": 0.000134, |
| "loss": 0.1577, |
| "step": 670 |
| }, |
| { |
| "epoch": 0.9249183895538629, |
| "grad_norm": 1.6210945844650269, |
| "learning_rate": 0.00013600000000000003, |
| "loss": 0.1641, |
| "step": 680 |
| }, |
| { |
| "epoch": 0.9385201305767138, |
| "grad_norm": 1.661000370979309, |
| "learning_rate": 0.000138, |
| "loss": 0.1543, |
| "step": 690 |
| }, |
| { |
| "epoch": 0.9521218715995647, |
| "grad_norm": 1.754515290260315, |
| "learning_rate": 0.00014, |
| "loss": 0.1621, |
| "step": 700 |
| }, |
| { |
| "epoch": 0.9657236126224157, |
| "grad_norm": 1.6415514945983887, |
| "learning_rate": 0.000142, |
| "loss": 0.144, |
| "step": 710 |
| }, |
| { |
| "epoch": 0.9793253536452666, |
| "grad_norm": 2.2498250007629395, |
| "learning_rate": 0.000144, |
| "loss": 0.1502, |
| "step": 720 |
| }, |
| { |
| "epoch": 0.9929270946681176, |
| "grad_norm": 1.6925517320632935, |
| "learning_rate": 0.000146, |
| "loss": 0.1546, |
| "step": 730 |
| }, |
| { |
| "epoch": 1.0054406964091405, |
| "grad_norm": 1.8807954788208008, |
| "learning_rate": 0.000148, |
| "loss": 0.1409, |
| "step": 740 |
| }, |
| { |
| "epoch": 1.0190424374319913, |
| "grad_norm": 1.611089825630188, |
| "learning_rate": 0.00015000000000000001, |
| "loss": 0.1578, |
| "step": 750 |
| }, |
| { |
| "epoch": 1.0326441784548421, |
| "grad_norm": 1.1505298614501953, |
| "learning_rate": 0.000152, |
| "loss": 0.1371, |
| "step": 760 |
| }, |
| { |
| "epoch": 1.0462459194776932, |
| "grad_norm": 1.5354456901550293, |
| "learning_rate": 0.000154, |
| "loss": 0.1407, |
| "step": 770 |
| }, |
| { |
| "epoch": 1.059847660500544, |
| "grad_norm": 1.178806185722351, |
| "learning_rate": 0.00015600000000000002, |
| "loss": 0.141, |
| "step": 780 |
| }, |
| { |
| "epoch": 1.073449401523395, |
| "grad_norm": 1.3755444288253784, |
| "learning_rate": 0.00015800000000000002, |
| "loss": 0.1287, |
| "step": 790 |
| }, |
| { |
| "epoch": 1.087051142546246, |
| "grad_norm": 1.5800870656967163, |
| "learning_rate": 0.00016, |
| "loss": 0.1382, |
| "step": 800 |
| }, |
| { |
| "epoch": 1.1006528835690967, |
| "grad_norm": 1.1685993671417236, |
| "learning_rate": 0.000162, |
| "loss": 0.1363, |
| "step": 810 |
| }, |
| { |
| "epoch": 1.1142546245919478, |
| "grad_norm": 1.1938755512237549, |
| "learning_rate": 0.000164, |
| "loss": 0.1316, |
| "step": 820 |
| }, |
| { |
| "epoch": 1.1278563656147986, |
| "grad_norm": 1.2022426128387451, |
| "learning_rate": 0.000166, |
| "loss": 0.1313, |
| "step": 830 |
| }, |
| { |
| "epoch": 1.1414581066376497, |
| "grad_norm": 1.1900382041931152, |
| "learning_rate": 0.000168, |
| "loss": 0.1256, |
| "step": 840 |
| }, |
| { |
| "epoch": 1.1550598476605005, |
| "grad_norm": 1.067172884941101, |
| "learning_rate": 0.00017, |
| "loss": 0.1385, |
| "step": 850 |
| }, |
| { |
| "epoch": 1.1686615886833516, |
| "grad_norm": 1.4434224367141724, |
| "learning_rate": 0.000172, |
| "loss": 0.1382, |
| "step": 860 |
| }, |
| { |
| "epoch": 1.1822633297062024, |
| "grad_norm": 1.0884168148040771, |
| "learning_rate": 0.000174, |
| "loss": 0.1266, |
| "step": 870 |
| }, |
| { |
| "epoch": 1.1958650707290532, |
| "grad_norm": 1.3909893035888672, |
| "learning_rate": 0.00017600000000000002, |
| "loss": 0.1215, |
| "step": 880 |
| }, |
| { |
| "epoch": 1.2094668117519043, |
| "grad_norm": 0.9344027042388916, |
| "learning_rate": 0.00017800000000000002, |
| "loss": 0.119, |
| "step": 890 |
| }, |
| { |
| "epoch": 1.2230685527747551, |
| "grad_norm": 1.1694083213806152, |
| "learning_rate": 0.00018, |
| "loss": 0.1192, |
| "step": 900 |
| }, |
| { |
| "epoch": 1.2366702937976062, |
| "grad_norm": 0.9874443411827087, |
| "learning_rate": 0.000182, |
| "loss": 0.1278, |
| "step": 910 |
| }, |
| { |
| "epoch": 1.250272034820457, |
| "grad_norm": 1.2893680334091187, |
| "learning_rate": 0.00018400000000000003, |
| "loss": 0.1318, |
| "step": 920 |
| }, |
| { |
| "epoch": 1.263873775843308, |
| "grad_norm": 1.3346811532974243, |
| "learning_rate": 0.00018600000000000002, |
| "loss": 0.128, |
| "step": 930 |
| }, |
| { |
| "epoch": 1.277475516866159, |
| "grad_norm": 0.9889335632324219, |
| "learning_rate": 0.000188, |
| "loss": 0.1372, |
| "step": 940 |
| }, |
| { |
| "epoch": 1.2910772578890097, |
| "grad_norm": 1.2218222618103027, |
| "learning_rate": 0.00019, |
| "loss": 0.1307, |
| "step": 950 |
| }, |
| { |
| "epoch": 1.3046789989118608, |
| "grad_norm": 0.897546112537384, |
| "learning_rate": 0.000192, |
| "loss": 0.1217, |
| "step": 960 |
| }, |
| { |
| "epoch": 1.3182807399347116, |
| "grad_norm": 0.9673519730567932, |
| "learning_rate": 0.000194, |
| "loss": 0.1094, |
| "step": 970 |
| }, |
| { |
| "epoch": 1.3318824809575625, |
| "grad_norm": 0.9263612031936646, |
| "learning_rate": 0.000196, |
| "loss": 0.1111, |
| "step": 980 |
| }, |
| { |
| "epoch": 1.3454842219804135, |
| "grad_norm": 1.0785088539123535, |
| "learning_rate": 0.00019800000000000002, |
| "loss": 0.12, |
| "step": 990 |
| }, |
| { |
| "epoch": 1.3590859630032643, |
| "grad_norm": 0.8844039440155029, |
| "learning_rate": 0.0002, |
| "loss": 0.1211, |
| "step": 1000 |
| }, |
| { |
| "epoch": 1.3726877040261154, |
| "grad_norm": 0.8651229739189148, |
| "learning_rate": 0.0001999999906373993, |
| "loss": 0.1291, |
| "step": 1010 |
| }, |
| { |
| "epoch": 1.3862894450489662, |
| "grad_norm": 1.0854979753494263, |
| "learning_rate": 0.000199999962549599, |
| "loss": 0.125, |
| "step": 1020 |
| }, |
| { |
| "epoch": 1.3998911860718173, |
| "grad_norm": 1.303252935409546, |
| "learning_rate": 0.00019999991573660427, |
| "loss": 0.1295, |
| "step": 1030 |
| }, |
| { |
| "epoch": 1.4134929270946681, |
| "grad_norm": 1.0601307153701782, |
| "learning_rate": 0.00019999985019842397, |
| "loss": 0.1363, |
| "step": 1040 |
| }, |
| { |
| "epoch": 1.427094668117519, |
| "grad_norm": 0.8127331733703613, |
| "learning_rate": 0.0001999997659350703, |
| "loss": 0.1124, |
| "step": 1050 |
| }, |
| { |
| "epoch": 1.44069640914037, |
| "grad_norm": 0.867559015750885, |
| "learning_rate": 0.0001999996629465591, |
| "loss": 0.1157, |
| "step": 1060 |
| }, |
| { |
| "epoch": 1.4542981501632208, |
| "grad_norm": 0.8963221907615662, |
| "learning_rate": 0.0001999995412329096, |
| "loss": 0.1197, |
| "step": 1070 |
| }, |
| { |
| "epoch": 1.467899891186072, |
| "grad_norm": 0.7211653590202332, |
| "learning_rate": 0.00019999940079414464, |
| "loss": 0.11, |
| "step": 1080 |
| }, |
| { |
| "epoch": 1.4815016322089227, |
| "grad_norm": 0.8746846914291382, |
| "learning_rate": 0.00019999924163029048, |
| "loss": 0.104, |
| "step": 1090 |
| }, |
| { |
| "epoch": 1.4951033732317738, |
| "grad_norm": 0.6718381643295288, |
| "learning_rate": 0.00019999906374137694, |
| "loss": 0.1041, |
| "step": 1100 |
| }, |
| { |
| "epoch": 1.5087051142546246, |
| "grad_norm": 0.7146100997924805, |
| "learning_rate": 0.00019999886712743732, |
| "loss": 0.0894, |
| "step": 1110 |
| }, |
| { |
| "epoch": 1.5223068552774754, |
| "grad_norm": 0.784561276435852, |
| "learning_rate": 0.00019999865178850845, |
| "loss": 0.0885, |
| "step": 1120 |
| }, |
| { |
| "epoch": 1.5359085963003265, |
| "grad_norm": 0.610625147819519, |
| "learning_rate": 0.00019999841772463066, |
| "loss": 0.0853, |
| "step": 1130 |
| }, |
| { |
| "epoch": 1.5495103373231773, |
| "grad_norm": 0.5619096159934998, |
| "learning_rate": 0.00019999816493584775, |
| "loss": 0.0788, |
| "step": 1140 |
| }, |
| { |
| "epoch": 1.5631120783460282, |
| "grad_norm": 0.5502200126647949, |
| "learning_rate": 0.00019999789342220708, |
| "loss": 0.0802, |
| "step": 1150 |
| }, |
| { |
| "epoch": 1.5767138193688792, |
| "grad_norm": 0.6030136942863464, |
| "learning_rate": 0.0001999976031837595, |
| "loss": 0.083, |
| "step": 1160 |
| }, |
| { |
| "epoch": 1.5903155603917303, |
| "grad_norm": 0.647160530090332, |
| "learning_rate": 0.00019999729422055928, |
| "loss": 0.0897, |
| "step": 1170 |
| }, |
| { |
| "epoch": 1.603917301414581, |
| "grad_norm": 0.7512218952178955, |
| "learning_rate": 0.00019999696653266437, |
| "loss": 0.08, |
| "step": 1180 |
| }, |
| { |
| "epoch": 1.617519042437432, |
| "grad_norm": 0.5823985934257507, |
| "learning_rate": 0.00019999662012013612, |
| "loss": 0.0772, |
| "step": 1190 |
| }, |
| { |
| "epoch": 1.631120783460283, |
| "grad_norm": 0.6494550108909607, |
| "learning_rate": 0.00019999625498303932, |
| "loss": 0.0772, |
| "step": 1200 |
| }, |
| { |
| "epoch": 1.6447225244831338, |
| "grad_norm": 0.5662053823471069, |
| "learning_rate": 0.00019999587112144244, |
| "loss": 0.0695, |
| "step": 1210 |
| }, |
| { |
| "epoch": 1.6583242655059847, |
| "grad_norm": 0.4981078803539276, |
| "learning_rate": 0.0001999954685354173, |
| "loss": 0.0742, |
| "step": 1220 |
| }, |
| { |
| "epoch": 1.6719260065288357, |
| "grad_norm": 0.5595643520355225, |
| "learning_rate": 0.00019999504722503927, |
| "loss": 0.0693, |
| "step": 1230 |
| }, |
| { |
| "epoch": 1.6855277475516868, |
| "grad_norm": 0.49906110763549805, |
| "learning_rate": 0.00019999460719038732, |
| "loss": 0.0692, |
| "step": 1240 |
| }, |
| { |
| "epoch": 1.6991294885745374, |
| "grad_norm": 0.5288122296333313, |
| "learning_rate": 0.00019999414843154375, |
| "loss": 0.0689, |
| "step": 1250 |
| }, |
| { |
| "epoch": 1.7127312295973884, |
| "grad_norm": 0.48555830121040344, |
| "learning_rate": 0.00019999367094859452, |
| "loss": 0.0655, |
| "step": 1260 |
| }, |
| { |
| "epoch": 1.7263329706202395, |
| "grad_norm": 0.5002060532569885, |
| "learning_rate": 0.00019999317474162905, |
| "loss": 0.0707, |
| "step": 1270 |
| }, |
| { |
| "epoch": 1.7399347116430903, |
| "grad_norm": 0.4510345458984375, |
| "learning_rate": 0.0001999926598107402, |
| "loss": 0.06, |
| "step": 1280 |
| }, |
| { |
| "epoch": 1.7535364526659412, |
| "grad_norm": 0.5075559020042419, |
| "learning_rate": 0.00019999212615602445, |
| "loss": 0.0675, |
| "step": 1290 |
| }, |
| { |
| "epoch": 1.7671381936887922, |
| "grad_norm": 0.5471305251121521, |
| "learning_rate": 0.0001999915737775817, |
| "loss": 0.0661, |
| "step": 1300 |
| }, |
| { |
| "epoch": 1.780739934711643, |
| "grad_norm": 0.6649473905563354, |
| "learning_rate": 0.00019999100267551538, |
| "loss": 0.0746, |
| "step": 1310 |
| }, |
| { |
| "epoch": 1.7943416757344939, |
| "grad_norm": 0.6705607175827026, |
| "learning_rate": 0.00019999041284993245, |
| "loss": 0.075, |
| "step": 1320 |
| }, |
| { |
| "epoch": 1.807943416757345, |
| "grad_norm": 0.645964503288269, |
| "learning_rate": 0.00019998980430094334, |
| "loss": 0.0825, |
| "step": 1330 |
| }, |
| { |
| "epoch": 1.821545157780196, |
| "grad_norm": 0.48304426670074463, |
| "learning_rate": 0.00019998917702866202, |
| "loss": 0.0726, |
| "step": 1340 |
| }, |
| { |
| "epoch": 1.8351468988030468, |
| "grad_norm": 0.5829260349273682, |
| "learning_rate": 0.00019998853103320592, |
| "loss": 0.0674, |
| "step": 1350 |
| }, |
| { |
| "epoch": 1.8487486398258977, |
| "grad_norm": 0.6563496589660645, |
| "learning_rate": 0.00019998786631469603, |
| "loss": 0.0666, |
| "step": 1360 |
| }, |
| { |
| "epoch": 1.8623503808487487, |
| "grad_norm": 0.5735076069831848, |
| "learning_rate": 0.00019998718287325676, |
| "loss": 0.0686, |
| "step": 1370 |
| }, |
| { |
| "epoch": 1.8759521218715995, |
| "grad_norm": 0.5888078212738037, |
| "learning_rate": 0.0001999864807090162, |
| "loss": 0.0728, |
| "step": 1380 |
| }, |
| { |
| "epoch": 1.8895538628944504, |
| "grad_norm": 0.5383855700492859, |
| "learning_rate": 0.00019998575982210572, |
| "loss": 0.0691, |
| "step": 1390 |
| }, |
| { |
| "epoch": 1.9031556039173014, |
| "grad_norm": 0.4812714159488678, |
| "learning_rate": 0.0001999850202126604, |
| "loss": 0.0591, |
| "step": 1400 |
| }, |
| { |
| "epoch": 1.9167573449401525, |
| "grad_norm": 0.5054184794425964, |
| "learning_rate": 0.00019998426188081865, |
| "loss": 0.0657, |
| "step": 1410 |
| }, |
| { |
| "epoch": 1.9303590859630033, |
| "grad_norm": 0.4280984103679657, |
| "learning_rate": 0.0001999834848267225, |
| "loss": 0.0619, |
| "step": 1420 |
| }, |
| { |
| "epoch": 1.9439608269858542, |
| "grad_norm": 0.7245299816131592, |
| "learning_rate": 0.0001999826890505175, |
| "loss": 0.064, |
| "step": 1430 |
| }, |
| { |
| "epoch": 1.9575625680087052, |
| "grad_norm": 0.5748353004455566, |
| "learning_rate": 0.0001999818745523526, |
| "loss": 0.0646, |
| "step": 1440 |
| }, |
| { |
| "epoch": 1.971164309031556, |
| "grad_norm": 0.5662197470664978, |
| "learning_rate": 0.00019998104133238034, |
| "loss": 0.0646, |
| "step": 1450 |
| }, |
| { |
| "epoch": 1.9847660500544069, |
| "grad_norm": 0.47684717178344727, |
| "learning_rate": 0.00019998018939075673, |
| "loss": 0.0626, |
| "step": 1460 |
| }, |
| { |
| "epoch": 1.998367791077258, |
| "grad_norm": 0.4988132417201996, |
| "learning_rate": 0.00019997931872764132, |
| "loss": 0.0596, |
| "step": 1470 |
| }, |
| { |
| "epoch": 2.010881392818281, |
| "grad_norm": 0.48226118087768555, |
| "learning_rate": 0.0001999784293431971, |
| "loss": 0.0579, |
| "step": 1480 |
| }, |
| { |
| "epoch": 2.0244831338411315, |
| "grad_norm": 0.44725948572158813, |
| "learning_rate": 0.0001999775212375907, |
| "loss": 0.0575, |
| "step": 1490 |
| }, |
| { |
| "epoch": 2.0380848748639826, |
| "grad_norm": 0.5599634051322937, |
| "learning_rate": 0.00019997659441099206, |
| "loss": 0.0594, |
| "step": 1500 |
| }, |
| { |
| "epoch": 2.0516866158868337, |
| "grad_norm": 0.47687241435050964, |
| "learning_rate": 0.00019997564886357476, |
| "loss": 0.0601, |
| "step": 1510 |
| }, |
| { |
| "epoch": 2.0652883569096843, |
| "grad_norm": 0.38273605704307556, |
| "learning_rate": 0.0001999746845955159, |
| "loss": 0.0551, |
| "step": 1520 |
| }, |
| { |
| "epoch": 2.0788900979325353, |
| "grad_norm": 0.5022798180580139, |
| "learning_rate": 0.00019997370160699602, |
| "loss": 0.0628, |
| "step": 1530 |
| }, |
| { |
| "epoch": 2.0924918389553864, |
| "grad_norm": 0.4850836992263794, |
| "learning_rate": 0.00019997269989819916, |
| "loss": 0.0562, |
| "step": 1540 |
| }, |
| { |
| "epoch": 2.1060935799782374, |
| "grad_norm": 0.45668402314186096, |
| "learning_rate": 0.0001999716794693129, |
| "loss": 0.0542, |
| "step": 1550 |
| }, |
| { |
| "epoch": 2.119695321001088, |
| "grad_norm": 0.4036906361579895, |
| "learning_rate": 0.00019997064032052837, |
| "loss": 0.0556, |
| "step": 1560 |
| }, |
| { |
| "epoch": 2.133297062023939, |
| "grad_norm": 0.46593400835990906, |
| "learning_rate": 0.00019996958245204009, |
| "loss": 0.0563, |
| "step": 1570 |
| }, |
| { |
| "epoch": 2.14689880304679, |
| "grad_norm": 0.4298928380012512, |
| "learning_rate": 0.00019996850586404615, |
| "loss": 0.0531, |
| "step": 1580 |
| }, |
| { |
| "epoch": 2.1605005440696408, |
| "grad_norm": 0.40037456154823303, |
| "learning_rate": 0.00019996741055674816, |
| "loss": 0.0519, |
| "step": 1590 |
| }, |
| { |
| "epoch": 2.174102285092492, |
| "grad_norm": 0.4456872344017029, |
| "learning_rate": 0.00019996629653035126, |
| "loss": 0.0477, |
| "step": 1600 |
| }, |
| { |
| "epoch": 2.187704026115343, |
| "grad_norm": 0.5066975355148315, |
| "learning_rate": 0.000199965163785064, |
| "loss": 0.049, |
| "step": 1610 |
| }, |
| { |
| "epoch": 2.2013057671381935, |
| "grad_norm": 0.39652055501937866, |
| "learning_rate": 0.0001999640123210985, |
| "loss": 0.0485, |
| "step": 1620 |
| }, |
| { |
| "epoch": 2.2149075081610445, |
| "grad_norm": 0.41883769631385803, |
| "learning_rate": 0.00019996284213867033, |
| "loss": 0.051, |
| "step": 1630 |
| }, |
| { |
| "epoch": 2.2285092491838956, |
| "grad_norm": 0.4462110996246338, |
| "learning_rate": 0.0001999616532379987, |
| "loss": 0.0515, |
| "step": 1640 |
| }, |
| { |
| "epoch": 2.2421109902067466, |
| "grad_norm": 0.5779175758361816, |
| "learning_rate": 0.00019996044561930622, |
| "loss": 0.0569, |
| "step": 1650 |
| }, |
| { |
| "epoch": 2.2557127312295973, |
| "grad_norm": 0.42402154207229614, |
| "learning_rate": 0.00019995921928281894, |
| "loss": 0.0519, |
| "step": 1660 |
| }, |
| { |
| "epoch": 2.2693144722524483, |
| "grad_norm": 0.642371654510498, |
| "learning_rate": 0.00019995797422876654, |
| "loss": 0.0612, |
| "step": 1670 |
| }, |
| { |
| "epoch": 2.2829162132752994, |
| "grad_norm": 0.44712305068969727, |
| "learning_rate": 0.0001999567104573822, |
| "loss": 0.0555, |
| "step": 1680 |
| }, |
| { |
| "epoch": 2.29651795429815, |
| "grad_norm": 0.5107985138893127, |
| "learning_rate": 0.0001999554279689025, |
| "loss": 0.0629, |
| "step": 1690 |
| }, |
| { |
| "epoch": 2.310119695321001, |
| "grad_norm": 0.4252520203590393, |
| "learning_rate": 0.00019995412676356762, |
| "loss": 0.0511, |
| "step": 1700 |
| }, |
| { |
| "epoch": 2.323721436343852, |
| "grad_norm": 0.413331001996994, |
| "learning_rate": 0.0001999528068416212, |
| "loss": 0.0504, |
| "step": 1710 |
| }, |
| { |
| "epoch": 2.337323177366703, |
| "grad_norm": 0.4216318726539612, |
| "learning_rate": 0.0001999514682033104, |
| "loss": 0.0539, |
| "step": 1720 |
| }, |
| { |
| "epoch": 2.3509249183895538, |
| "grad_norm": 0.5241215825080872, |
| "learning_rate": 0.0001999501108488859, |
| "loss": 0.0551, |
| "step": 1730 |
| }, |
| { |
| "epoch": 2.364526659412405, |
| "grad_norm": 0.48192480206489563, |
| "learning_rate": 0.00019994873477860185, |
| "loss": 0.0591, |
| "step": 1740 |
| }, |
| { |
| "epoch": 2.378128400435256, |
| "grad_norm": 0.5567557215690613, |
| "learning_rate": 0.00019994733999271596, |
| "loss": 0.0522, |
| "step": 1750 |
| }, |
| { |
| "epoch": 2.3917301414581065, |
| "grad_norm": 0.4027807414531708, |
| "learning_rate": 0.00019994592649148933, |
| "loss": 0.0524, |
| "step": 1760 |
| }, |
| { |
| "epoch": 2.4053318824809575, |
| "grad_norm": 0.40923604369163513, |
| "learning_rate": 0.0001999444942751867, |
| "loss": 0.0533, |
| "step": 1770 |
| }, |
| { |
| "epoch": 2.4189336235038086, |
| "grad_norm": 0.37654557824134827, |
| "learning_rate": 0.00019994304334407622, |
| "loss": 0.0482, |
| "step": 1780 |
| }, |
| { |
| "epoch": 2.432535364526659, |
| "grad_norm": 0.37279561161994934, |
| "learning_rate": 0.00019994157369842964, |
| "loss": 0.0448, |
| "step": 1790 |
| }, |
| { |
| "epoch": 2.4461371055495102, |
| "grad_norm": 0.34426528215408325, |
| "learning_rate": 0.0001999400853385221, |
| "loss": 0.0467, |
| "step": 1800 |
| }, |
| { |
| "epoch": 2.4597388465723613, |
| "grad_norm": 0.4583146870136261, |
| "learning_rate": 0.00019993857826463231, |
| "loss": 0.0501, |
| "step": 1810 |
| }, |
| { |
| "epoch": 2.4733405875952124, |
| "grad_norm": 0.4296802878379822, |
| "learning_rate": 0.00019993705247704245, |
| "loss": 0.0475, |
| "step": 1820 |
| }, |
| { |
| "epoch": 2.486942328618063, |
| "grad_norm": 0.3417333960533142, |
| "learning_rate": 0.00019993550797603828, |
| "loss": 0.0475, |
| "step": 1830 |
| }, |
| { |
| "epoch": 2.500544069640914, |
| "grad_norm": 0.3391024172306061, |
| "learning_rate": 0.000199933944761909, |
| "loss": 0.0428, |
| "step": 1840 |
| }, |
| { |
| "epoch": 2.514145810663765, |
| "grad_norm": 0.4173300862312317, |
| "learning_rate": 0.00019993236283494728, |
| "loss": 0.0487, |
| "step": 1850 |
| }, |
| { |
| "epoch": 2.527747551686616, |
| "grad_norm": 0.41140830516815186, |
| "learning_rate": 0.00019993076219544938, |
| "loss": 0.0499, |
| "step": 1860 |
| }, |
| { |
| "epoch": 2.5413492927094667, |
| "grad_norm": 0.4615647494792938, |
| "learning_rate": 0.00019992914284371497, |
| "loss": 0.0485, |
| "step": 1870 |
| }, |
| { |
| "epoch": 2.554951033732318, |
| "grad_norm": 0.4390008747577667, |
| "learning_rate": 0.00019992750478004738, |
| "loss": 0.0476, |
| "step": 1880 |
| }, |
| { |
| "epoch": 2.568552774755169, |
| "grad_norm": 0.36641502380371094, |
| "learning_rate": 0.00019992584800475322, |
| "loss": 0.0421, |
| "step": 1890 |
| }, |
| { |
| "epoch": 2.5821545157780195, |
| "grad_norm": 0.43274542689323425, |
| "learning_rate": 0.00019992417251814282, |
| "loss": 0.048, |
| "step": 1900 |
| }, |
| { |
| "epoch": 2.5957562568008705, |
| "grad_norm": 0.39262470602989197, |
| "learning_rate": 0.0001999224783205299, |
| "loss": 0.0482, |
| "step": 1910 |
| }, |
| { |
| "epoch": 2.6093579978237216, |
| "grad_norm": 0.3618634343147278, |
| "learning_rate": 0.0001999207654122316, |
| "loss": 0.0464, |
| "step": 1920 |
| }, |
| { |
| "epoch": 2.622959738846572, |
| "grad_norm": 0.3481171727180481, |
| "learning_rate": 0.0001999190337935688, |
| "loss": 0.0478, |
| "step": 1930 |
| }, |
| { |
| "epoch": 2.6365614798694232, |
| "grad_norm": 0.33801841735839844, |
| "learning_rate": 0.0001999172834648657, |
| "loss": 0.0444, |
| "step": 1940 |
| }, |
| { |
| "epoch": 2.6501632208922743, |
| "grad_norm": 0.337028443813324, |
| "learning_rate": 0.00019991551442645006, |
| "loss": 0.0403, |
| "step": 1950 |
| }, |
| { |
| "epoch": 2.663764961915125, |
| "grad_norm": 0.3380604684352875, |
| "learning_rate": 0.0001999137266786531, |
| "loss": 0.0483, |
| "step": 1960 |
| }, |
| { |
| "epoch": 2.677366702937976, |
| "grad_norm": 0.44155481457710266, |
| "learning_rate": 0.0001999119202218096, |
| "loss": 0.045, |
| "step": 1970 |
| }, |
| { |
| "epoch": 2.690968443960827, |
| "grad_norm": 0.43442797660827637, |
| "learning_rate": 0.00019991009505625784, |
| "loss": 0.0442, |
| "step": 1980 |
| }, |
| { |
| "epoch": 2.704570184983678, |
| "grad_norm": 0.4352855980396271, |
| "learning_rate": 0.00019990825118233957, |
| "loss": 0.0463, |
| "step": 1990 |
| }, |
| { |
| "epoch": 2.7181719260065287, |
| "grad_norm": 0.5073570609092712, |
| "learning_rate": 0.00019990638860040006, |
| "loss": 0.0492, |
| "step": 2000 |
| } |
| ], |
| "logging_steps": 10, |
| "max_steps": 73600, |
| "num_input_tokens_seen": 0, |
| "num_train_epochs": 100, |
| "save_steps": 2000, |
| "stateful_callbacks": { |
| "TrainerControl": { |
| "args": { |
| "should_epoch_stop": false, |
| "should_evaluate": false, |
| "should_log": false, |
| "should_save": true, |
| "should_training_stop": false |
| }, |
| "attributes": {} |
| } |
| }, |
| "total_flos": 5.54846097526789e+21, |
| "train_batch_size": 32, |
| "trial_name": null, |
| "trial_params": null |
| } |
|
|