| { | |
| "best_global_step": null, | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 0.5699629524080935, | |
| "eval_steps": 500, | |
| "global_step": 1000, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.005699629524080934, | |
| "grad_norm": 19.25, | |
| "learning_rate": 9.982896237172178e-06, | |
| "loss": 2.0698, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.011399259048161869, | |
| "grad_norm": 6.96875, | |
| "learning_rate": 9.963892056252377e-06, | |
| "loss": 1.5026, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.017098888572242805, | |
| "grad_norm": 6.5, | |
| "learning_rate": 9.944887875332574e-06, | |
| "loss": 1.4607, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.022798518096323737, | |
| "grad_norm": 6.5625, | |
| "learning_rate": 9.925883694412771e-06, | |
| "loss": 1.3855, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.028498147620404674, | |
| "grad_norm": 6.0625, | |
| "learning_rate": 9.90687951349297e-06, | |
| "loss": 1.4154, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.03419777714448561, | |
| "grad_norm": 5.90625, | |
| "learning_rate": 9.887875332573167e-06, | |
| "loss": 1.3327, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.039897406668566546, | |
| "grad_norm": 12.9375, | |
| "learning_rate": 9.868871151653364e-06, | |
| "loss": 1.3797, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 0.045597036192647475, | |
| "grad_norm": 6.3125, | |
| "learning_rate": 9.849866970733563e-06, | |
| "loss": 1.361, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.05129666571672841, | |
| "grad_norm": 7.0, | |
| "learning_rate": 9.83086278981376e-06, | |
| "loss": 1.4097, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 0.05699629524080935, | |
| "grad_norm": 6.46875, | |
| "learning_rate": 9.811858608893958e-06, | |
| "loss": 1.3784, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.06269592476489028, | |
| "grad_norm": 6.09375, | |
| "learning_rate": 9.792854427974155e-06, | |
| "loss": 1.3886, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 0.06839555428897122, | |
| "grad_norm": 7.0625, | |
| "learning_rate": 9.773850247054353e-06, | |
| "loss": 1.2982, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 0.07409518381305215, | |
| "grad_norm": 6.0625, | |
| "learning_rate": 9.75484606613455e-06, | |
| "loss": 1.3887, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 0.07979481333713309, | |
| "grad_norm": 5.46875, | |
| "learning_rate": 9.735841885214748e-06, | |
| "loss": 1.3336, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 0.08549444286121402, | |
| "grad_norm": 6.78125, | |
| "learning_rate": 9.716837704294946e-06, | |
| "loss": 1.3523, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 0.09119407238529495, | |
| "grad_norm": 7.96875, | |
| "learning_rate": 9.697833523375144e-06, | |
| "loss": 1.3483, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 0.09689370190937589, | |
| "grad_norm": 6.96875, | |
| "learning_rate": 9.678829342455342e-06, | |
| "loss": 1.3502, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 0.10259333143345682, | |
| "grad_norm": 7.0, | |
| "learning_rate": 9.65982516153554e-06, | |
| "loss": 1.3553, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 0.10829296095753776, | |
| "grad_norm": 6.3125, | |
| "learning_rate": 9.640820980615736e-06, | |
| "loss": 1.338, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 0.1139925904816187, | |
| "grad_norm": 7.3125, | |
| "learning_rate": 9.621816799695934e-06, | |
| "loss": 1.3868, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.11969222000569962, | |
| "grad_norm": 6.53125, | |
| "learning_rate": 9.602812618776132e-06, | |
| "loss": 1.3996, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 0.12539184952978055, | |
| "grad_norm": 7.8125, | |
| "learning_rate": 9.58380843785633e-06, | |
| "loss": 1.3638, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 0.1310914790538615, | |
| "grad_norm": 6.90625, | |
| "learning_rate": 9.564804256936528e-06, | |
| "loss": 1.2893, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 0.13679110857794244, | |
| "grad_norm": 6.9375, | |
| "learning_rate": 9.545800076016724e-06, | |
| "loss": 1.3408, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 0.14249073810202337, | |
| "grad_norm": 6.4375, | |
| "learning_rate": 9.526795895096922e-06, | |
| "loss": 1.3522, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 0.1481903676261043, | |
| "grad_norm": 6.59375, | |
| "learning_rate": 9.50779171417712e-06, | |
| "loss": 1.3595, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 0.15388999715018523, | |
| "grad_norm": 7.125, | |
| "learning_rate": 9.488787533257318e-06, | |
| "loss": 1.381, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 0.15958962667426618, | |
| "grad_norm": 7.1875, | |
| "learning_rate": 9.469783352337516e-06, | |
| "loss": 1.3109, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 0.1652892561983471, | |
| "grad_norm": 7.09375, | |
| "learning_rate": 9.450779171417712e-06, | |
| "loss": 1.3793, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 0.17098888572242804, | |
| "grad_norm": 7.125, | |
| "learning_rate": 9.43177499049791e-06, | |
| "loss": 1.3373, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.17668851524650897, | |
| "grad_norm": 6.5625, | |
| "learning_rate": 9.412770809578108e-06, | |
| "loss": 1.3474, | |
| "step": 310 | |
| }, | |
| { | |
| "epoch": 0.1823881447705899, | |
| "grad_norm": 7.03125, | |
| "learning_rate": 9.393766628658306e-06, | |
| "loss": 1.3076, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 0.18808777429467086, | |
| "grad_norm": 7.71875, | |
| "learning_rate": 9.374762447738504e-06, | |
| "loss": 1.2929, | |
| "step": 330 | |
| }, | |
| { | |
| "epoch": 0.19378740381875179, | |
| "grad_norm": 6.8125, | |
| "learning_rate": 9.3557582668187e-06, | |
| "loss": 1.3652, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 0.19948703334283271, | |
| "grad_norm": 7.15625, | |
| "learning_rate": 9.336754085898898e-06, | |
| "loss": 1.3535, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 0.20518666286691364, | |
| "grad_norm": 5.9375, | |
| "learning_rate": 9.317749904979096e-06, | |
| "loss": 1.3388, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 0.21088629239099457, | |
| "grad_norm": 7.6875, | |
| "learning_rate": 9.298745724059294e-06, | |
| "loss": 1.3083, | |
| "step": 370 | |
| }, | |
| { | |
| "epoch": 0.21658592191507553, | |
| "grad_norm": 7.125, | |
| "learning_rate": 9.279741543139492e-06, | |
| "loss": 1.3409, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 0.22228555143915646, | |
| "grad_norm": 6.78125, | |
| "learning_rate": 9.26073736221969e-06, | |
| "loss": 1.3149, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 0.2279851809632374, | |
| "grad_norm": 6.3125, | |
| "learning_rate": 9.241733181299886e-06, | |
| "loss": 1.3525, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 0.23368481048731832, | |
| "grad_norm": 7.34375, | |
| "learning_rate": 9.222729000380084e-06, | |
| "loss": 1.3343, | |
| "step": 410 | |
| }, | |
| { | |
| "epoch": 0.23938444001139925, | |
| "grad_norm": 6.875, | |
| "learning_rate": 9.203724819460282e-06, | |
| "loss": 1.3098, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 0.2450840695354802, | |
| "grad_norm": 5.8125, | |
| "learning_rate": 9.18472063854048e-06, | |
| "loss": 1.2878, | |
| "step": 430 | |
| }, | |
| { | |
| "epoch": 0.2507836990595611, | |
| "grad_norm": 6.0625, | |
| "learning_rate": 9.165716457620676e-06, | |
| "loss": 1.2681, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 0.2564833285836421, | |
| "grad_norm": 7.09375, | |
| "learning_rate": 9.146712276700876e-06, | |
| "loss": 1.2974, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 0.262182958107723, | |
| "grad_norm": 6.96875, | |
| "learning_rate": 9.127708095781072e-06, | |
| "loss": 1.3047, | |
| "step": 460 | |
| }, | |
| { | |
| "epoch": 0.26788258763180395, | |
| "grad_norm": 7.40625, | |
| "learning_rate": 9.10870391486127e-06, | |
| "loss": 1.2063, | |
| "step": 470 | |
| }, | |
| { | |
| "epoch": 0.2735822171558849, | |
| "grad_norm": 7.59375, | |
| "learning_rate": 9.089699733941468e-06, | |
| "loss": 1.3046, | |
| "step": 480 | |
| }, | |
| { | |
| "epoch": 0.2792818466799658, | |
| "grad_norm": 5.9375, | |
| "learning_rate": 9.070695553021666e-06, | |
| "loss": 1.2633, | |
| "step": 490 | |
| }, | |
| { | |
| "epoch": 0.28498147620404674, | |
| "grad_norm": 7.3125, | |
| "learning_rate": 9.051691372101862e-06, | |
| "loss": 1.3394, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.29068110572812766, | |
| "grad_norm": 5.75, | |
| "learning_rate": 9.032687191182062e-06, | |
| "loss": 1.2653, | |
| "step": 510 | |
| }, | |
| { | |
| "epoch": 0.2963807352522086, | |
| "grad_norm": 6.6875, | |
| "learning_rate": 9.013683010262258e-06, | |
| "loss": 1.2359, | |
| "step": 520 | |
| }, | |
| { | |
| "epoch": 0.3020803647762895, | |
| "grad_norm": 7.21875, | |
| "learning_rate": 8.994678829342456e-06, | |
| "loss": 1.2021, | |
| "step": 530 | |
| }, | |
| { | |
| "epoch": 0.30777999430037045, | |
| "grad_norm": 6.8125, | |
| "learning_rate": 8.975674648422654e-06, | |
| "loss": 1.233, | |
| "step": 540 | |
| }, | |
| { | |
| "epoch": 0.31347962382445144, | |
| "grad_norm": 7.5, | |
| "learning_rate": 8.956670467502852e-06, | |
| "loss": 1.3498, | |
| "step": 550 | |
| }, | |
| { | |
| "epoch": 0.31917925334853237, | |
| "grad_norm": 6.5625, | |
| "learning_rate": 8.937666286583048e-06, | |
| "loss": 1.2261, | |
| "step": 560 | |
| }, | |
| { | |
| "epoch": 0.3248788828726133, | |
| "grad_norm": 6.6875, | |
| "learning_rate": 8.918662105663248e-06, | |
| "loss": 1.3059, | |
| "step": 570 | |
| }, | |
| { | |
| "epoch": 0.3305785123966942, | |
| "grad_norm": 6.5, | |
| "learning_rate": 8.899657924743444e-06, | |
| "loss": 1.241, | |
| "step": 580 | |
| }, | |
| { | |
| "epoch": 0.33627814192077515, | |
| "grad_norm": 7.28125, | |
| "learning_rate": 8.880653743823642e-06, | |
| "loss": 1.3466, | |
| "step": 590 | |
| }, | |
| { | |
| "epoch": 0.3419777714448561, | |
| "grad_norm": 6.75, | |
| "learning_rate": 8.861649562903838e-06, | |
| "loss": 1.314, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 0.347677400968937, | |
| "grad_norm": 6.6875, | |
| "learning_rate": 8.842645381984038e-06, | |
| "loss": 1.2967, | |
| "step": 610 | |
| }, | |
| { | |
| "epoch": 0.35337703049301794, | |
| "grad_norm": 7.40625, | |
| "learning_rate": 8.823641201064234e-06, | |
| "loss": 1.3309, | |
| "step": 620 | |
| }, | |
| { | |
| "epoch": 0.35907666001709887, | |
| "grad_norm": 7.375, | |
| "learning_rate": 8.804637020144432e-06, | |
| "loss": 1.248, | |
| "step": 630 | |
| }, | |
| { | |
| "epoch": 0.3647762895411798, | |
| "grad_norm": 6.8125, | |
| "learning_rate": 8.78563283922463e-06, | |
| "loss": 1.2862, | |
| "step": 640 | |
| }, | |
| { | |
| "epoch": 0.3704759190652608, | |
| "grad_norm": 7.3125, | |
| "learning_rate": 8.766628658304828e-06, | |
| "loss": 1.2528, | |
| "step": 650 | |
| }, | |
| { | |
| "epoch": 0.3761755485893417, | |
| "grad_norm": 6.75, | |
| "learning_rate": 8.747624477385024e-06, | |
| "loss": 1.2739, | |
| "step": 660 | |
| }, | |
| { | |
| "epoch": 0.38187517811342264, | |
| "grad_norm": 7.40625, | |
| "learning_rate": 8.728620296465224e-06, | |
| "loss": 1.2521, | |
| "step": 670 | |
| }, | |
| { | |
| "epoch": 0.38757480763750357, | |
| "grad_norm": 7.71875, | |
| "learning_rate": 8.70961611554542e-06, | |
| "loss": 1.2068, | |
| "step": 680 | |
| }, | |
| { | |
| "epoch": 0.3932744371615845, | |
| "grad_norm": 5.96875, | |
| "learning_rate": 8.690611934625618e-06, | |
| "loss": 1.3239, | |
| "step": 690 | |
| }, | |
| { | |
| "epoch": 0.39897406668566543, | |
| "grad_norm": 7.28125, | |
| "learning_rate": 8.671607753705816e-06, | |
| "loss": 1.2795, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 0.40467369620974636, | |
| "grad_norm": 6.90625, | |
| "learning_rate": 8.652603572786014e-06, | |
| "loss": 1.2848, | |
| "step": 710 | |
| }, | |
| { | |
| "epoch": 0.4103733257338273, | |
| "grad_norm": 6.59375, | |
| "learning_rate": 8.63359939186621e-06, | |
| "loss": 1.2491, | |
| "step": 720 | |
| }, | |
| { | |
| "epoch": 0.4160729552579082, | |
| "grad_norm": 6.5625, | |
| "learning_rate": 8.61459521094641e-06, | |
| "loss": 1.229, | |
| "step": 730 | |
| }, | |
| { | |
| "epoch": 0.42177258478198915, | |
| "grad_norm": 6.34375, | |
| "learning_rate": 8.595591030026606e-06, | |
| "loss": 1.2394, | |
| "step": 740 | |
| }, | |
| { | |
| "epoch": 0.42747221430607013, | |
| "grad_norm": 7.0625, | |
| "learning_rate": 8.576586849106804e-06, | |
| "loss": 1.3423, | |
| "step": 750 | |
| }, | |
| { | |
| "epoch": 0.43317184383015106, | |
| "grad_norm": 7.03125, | |
| "learning_rate": 8.557582668187002e-06, | |
| "loss": 1.3051, | |
| "step": 760 | |
| }, | |
| { | |
| "epoch": 0.438871473354232, | |
| "grad_norm": 6.21875, | |
| "learning_rate": 8.5385784872672e-06, | |
| "loss": 1.2504, | |
| "step": 770 | |
| }, | |
| { | |
| "epoch": 0.4445711028783129, | |
| "grad_norm": 7.5, | |
| "learning_rate": 8.519574306347396e-06, | |
| "loss": 1.2255, | |
| "step": 780 | |
| }, | |
| { | |
| "epoch": 0.45027073240239385, | |
| "grad_norm": 7.34375, | |
| "learning_rate": 8.500570125427594e-06, | |
| "loss": 1.2367, | |
| "step": 790 | |
| }, | |
| { | |
| "epoch": 0.4559703619264748, | |
| "grad_norm": 6.46875, | |
| "learning_rate": 8.481565944507792e-06, | |
| "loss": 1.2855, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 0.4616699914505557, | |
| "grad_norm": 6.6875, | |
| "learning_rate": 8.46256176358799e-06, | |
| "loss": 1.2875, | |
| "step": 810 | |
| }, | |
| { | |
| "epoch": 0.46736962097463663, | |
| "grad_norm": 6.75, | |
| "learning_rate": 8.443557582668188e-06, | |
| "loss": 1.2946, | |
| "step": 820 | |
| }, | |
| { | |
| "epoch": 0.47306925049871756, | |
| "grad_norm": 7.03125, | |
| "learning_rate": 8.424553401748386e-06, | |
| "loss": 1.2672, | |
| "step": 830 | |
| }, | |
| { | |
| "epoch": 0.4787688800227985, | |
| "grad_norm": 6.78125, | |
| "learning_rate": 8.405549220828583e-06, | |
| "loss": 1.2279, | |
| "step": 840 | |
| }, | |
| { | |
| "epoch": 0.4844685095468795, | |
| "grad_norm": 7.59375, | |
| "learning_rate": 8.38654503990878e-06, | |
| "loss": 1.2593, | |
| "step": 850 | |
| }, | |
| { | |
| "epoch": 0.4901681390709604, | |
| "grad_norm": 7.0, | |
| "learning_rate": 8.367540858988978e-06, | |
| "loss": 1.2662, | |
| "step": 860 | |
| }, | |
| { | |
| "epoch": 0.49586776859504134, | |
| "grad_norm": 7.8125, | |
| "learning_rate": 8.348536678069176e-06, | |
| "loss": 1.2404, | |
| "step": 870 | |
| }, | |
| { | |
| "epoch": 0.5015673981191222, | |
| "grad_norm": 6.21875, | |
| "learning_rate": 8.329532497149374e-06, | |
| "loss": 1.3002, | |
| "step": 880 | |
| }, | |
| { | |
| "epoch": 0.5072670276432032, | |
| "grad_norm": 6.5, | |
| "learning_rate": 8.31052831622957e-06, | |
| "loss": 1.2786, | |
| "step": 890 | |
| }, | |
| { | |
| "epoch": 0.5129666571672842, | |
| "grad_norm": 7.125, | |
| "learning_rate": 8.291524135309769e-06, | |
| "loss": 1.2384, | |
| "step": 900 | |
| }, | |
| { | |
| "epoch": 0.518666286691365, | |
| "grad_norm": 7.375, | |
| "learning_rate": 8.272519954389967e-06, | |
| "loss": 1.2383, | |
| "step": 910 | |
| }, | |
| { | |
| "epoch": 0.524365916215446, | |
| "grad_norm": 7.375, | |
| "learning_rate": 8.253515773470164e-06, | |
| "loss": 1.2136, | |
| "step": 920 | |
| }, | |
| { | |
| "epoch": 0.5300655457395269, | |
| "grad_norm": 5.84375, | |
| "learning_rate": 8.234511592550362e-06, | |
| "loss": 1.2139, | |
| "step": 930 | |
| }, | |
| { | |
| "epoch": 0.5357651752636079, | |
| "grad_norm": 6.8125, | |
| "learning_rate": 8.215507411630559e-06, | |
| "loss": 1.2511, | |
| "step": 940 | |
| }, | |
| { | |
| "epoch": 0.5414648047876888, | |
| "grad_norm": 7.65625, | |
| "learning_rate": 8.196503230710757e-06, | |
| "loss": 1.2315, | |
| "step": 950 | |
| }, | |
| { | |
| "epoch": 0.5471644343117698, | |
| "grad_norm": 6.875, | |
| "learning_rate": 8.177499049790955e-06, | |
| "loss": 1.2691, | |
| "step": 960 | |
| }, | |
| { | |
| "epoch": 0.5528640638358506, | |
| "grad_norm": 6.78125, | |
| "learning_rate": 8.158494868871153e-06, | |
| "loss": 1.2067, | |
| "step": 970 | |
| }, | |
| { | |
| "epoch": 0.5585636933599316, | |
| "grad_norm": 6.125, | |
| "learning_rate": 8.13949068795135e-06, | |
| "loss": 1.1774, | |
| "step": 980 | |
| }, | |
| { | |
| "epoch": 0.5642633228840125, | |
| "grad_norm": 6.90625, | |
| "learning_rate": 8.120486507031547e-06, | |
| "loss": 1.2299, | |
| "step": 990 | |
| }, | |
| { | |
| "epoch": 0.5699629524080935, | |
| "grad_norm": 7.6875, | |
| "learning_rate": 8.101482326111745e-06, | |
| "loss": 1.2525, | |
| "step": 1000 | |
| } | |
| ], | |
| "logging_steps": 10, | |
| "max_steps": 5262, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 3, | |
| "save_steps": 500, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": false | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 4.3299712794624e+16, | |
| "train_batch_size": 4, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |