{ "best_global_step": null, "best_metric": null, "best_model_checkpoint": null, "epoch": 1.0, "eval_steps": 500, "global_step": 457, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.01095290251916758, "grad_norm": 1.3321008682250977, "learning_rate": 1.0434782608695653e-06, "loss": 1.3073, "step": 5 }, { "epoch": 0.02190580503833516, "grad_norm": 1.1412581205368042, "learning_rate": 2.347826086956522e-06, "loss": 1.3358, "step": 10 }, { "epoch": 0.03285870755750274, "grad_norm": 0.9217167496681213, "learning_rate": 3.6521739130434787e-06, "loss": 1.2979, "step": 15 }, { "epoch": 0.04381161007667032, "grad_norm": 0.6123777627944946, "learning_rate": 4.956521739130435e-06, "loss": 1.2163, "step": 20 }, { "epoch": 0.054764512595837894, "grad_norm": 0.9250121116638184, "learning_rate": 6.260869565217391e-06, "loss": 1.264, "step": 25 }, { "epoch": 0.06571741511500548, "grad_norm": 0.5497902035713196, "learning_rate": 7.5652173913043475e-06, "loss": 1.225, "step": 30 }, { "epoch": 0.07667031763417305, "grad_norm": 0.602593183517456, "learning_rate": 8.869565217391306e-06, "loss": 1.1959, "step": 35 }, { "epoch": 0.08762322015334063, "grad_norm": 0.5248430371284485, "learning_rate": 1.017391304347826e-05, "loss": 1.2186, "step": 40 }, { "epoch": 0.09857612267250822, "grad_norm": 0.5515073537826538, "learning_rate": 1.1478260869565218e-05, "loss": 1.1472, "step": 45 }, { "epoch": 0.10952902519167579, "grad_norm": 1.2510318756103516, "learning_rate": 1.2782608695652173e-05, "loss": 1.1666, "step": 50 }, { "epoch": 0.12048192771084337, "grad_norm": 0.49298274517059326, "learning_rate": 1.4086956521739131e-05, "loss": 1.2019, "step": 55 }, { "epoch": 0.13143483023001096, "grad_norm": 0.4494762122631073, "learning_rate": 1.5391304347826088e-05, "loss": 1.173, "step": 60 }, { "epoch": 0.14238773274917854, "grad_norm": 0.6092376112937927, "learning_rate": 1.6695652173913044e-05, "loss": 1.1323, "step": 65 }, { "epoch": 0.1533406352683461, "grad_norm": 0.4575347602367401, "learning_rate": 1.8e-05, "loss": 1.1164, "step": 70 }, { "epoch": 0.16429353778751368, "grad_norm": 1.1347095966339111, "learning_rate": 1.9304347826086957e-05, "loss": 1.193, "step": 75 }, { "epoch": 0.17524644030668127, "grad_norm": 0.45952585339546204, "learning_rate": 2.0608695652173913e-05, "loss": 1.1403, "step": 80 }, { "epoch": 0.18619934282584885, "grad_norm": 0.49114152789115906, "learning_rate": 2.191304347826087e-05, "loss": 1.1535, "step": 85 }, { "epoch": 0.19715224534501644, "grad_norm": 0.5684127807617188, "learning_rate": 2.3217391304347826e-05, "loss": 1.1398, "step": 90 }, { "epoch": 0.20810514786418402, "grad_norm": 0.5454962849617004, "learning_rate": 2.4521739130434786e-05, "loss": 1.0928, "step": 95 }, { "epoch": 0.21905805038335158, "grad_norm": 0.491813600063324, "learning_rate": 2.582608695652174e-05, "loss": 1.0858, "step": 100 }, { "epoch": 0.23001095290251916, "grad_norm": 0.5732094645500183, "learning_rate": 2.7130434782608695e-05, "loss": 1.0765, "step": 105 }, { "epoch": 0.24096385542168675, "grad_norm": 0.5950366258621216, "learning_rate": 2.8434782608695652e-05, "loss": 1.149, "step": 110 }, { "epoch": 0.25191675794085433, "grad_norm": 0.536044180393219, "learning_rate": 2.973913043478261e-05, "loss": 1.0859, "step": 115 }, { "epoch": 0.2628696604600219, "grad_norm": 0.5769652724266052, "learning_rate": 2.999974848707377e-05, "loss": 1.0433, "step": 120 }, { "epoch": 0.2738225629791895, "grad_norm": 0.5500301718711853, "learning_rate": 2.9998726730266534e-05, "loss": 1.051, "step": 125 }, { "epoch": 0.2847754654983571, "grad_norm": 0.5901659727096558, "learning_rate": 2.9996919063517692e-05, "loss": 1.0166, "step": 130 }, { "epoch": 0.29572836801752467, "grad_norm": 0.6402804851531982, "learning_rate": 2.999432558154617e-05, "loss": 1.078, "step": 135 }, { "epoch": 0.3066812705366922, "grad_norm": 0.5208508372306824, "learning_rate": 2.9990946420246395e-05, "loss": 1.0167, "step": 140 }, { "epoch": 0.3176341730558598, "grad_norm": 0.6010607481002808, "learning_rate": 2.9986781756681146e-05, "loss": 1.0342, "step": 145 }, { "epoch": 0.32858707557502737, "grad_norm": 0.5925593376159668, "learning_rate": 2.9981831809072318e-05, "loss": 0.9813, "step": 150 }, { "epoch": 0.33953997809419495, "grad_norm": 0.7121065855026245, "learning_rate": 2.997609683678947e-05, "loss": 0.9768, "step": 155 }, { "epoch": 0.35049288061336253, "grad_norm": 0.5926089882850647, "learning_rate": 2.996957714033622e-05, "loss": 1.0256, "step": 160 }, { "epoch": 0.3614457831325301, "grad_norm": 0.6746991276741028, "learning_rate": 2.9962273061334515e-05, "loss": 0.9901, "step": 165 }, { "epoch": 0.3723986856516977, "grad_norm": 0.6542229056358337, "learning_rate": 2.9954184982506734e-05, "loss": 0.9707, "step": 170 }, { "epoch": 0.3833515881708653, "grad_norm": 0.7646006345748901, "learning_rate": 2.9945313327655622e-05, "loss": 0.9826, "step": 175 }, { "epoch": 0.39430449069003287, "grad_norm": 0.6529242396354675, "learning_rate": 2.9935658561642083e-05, "loss": 0.9659, "step": 180 }, { "epoch": 0.40525739320920046, "grad_norm": 0.6744837760925293, "learning_rate": 2.9925221190360836e-05, "loss": 0.9517, "step": 185 }, { "epoch": 0.41621029572836804, "grad_norm": 0.6999198198318481, "learning_rate": 2.9914001760713898e-05, "loss": 1.0202, "step": 190 }, { "epoch": 0.42716319824753557, "grad_norm": 0.7155759334564209, "learning_rate": 2.9902000860581916e-05, "loss": 0.9947, "step": 195 }, { "epoch": 0.43811610076670315, "grad_norm": 0.6805759072303772, "learning_rate": 2.9889219118793393e-05, "loss": 0.9751, "step": 200 }, { "epoch": 0.44906900328587074, "grad_norm": 0.7391874194145203, "learning_rate": 2.9875657205091703e-05, "loss": 1.0113, "step": 205 }, { "epoch": 0.4600219058050383, "grad_norm": 0.7946105003356934, "learning_rate": 2.9861315830100023e-05, "loss": 0.952, "step": 210 }, { "epoch": 0.4709748083242059, "grad_norm": 0.6928731203079224, "learning_rate": 2.984619574528408e-05, "loss": 0.9207, "step": 215 }, { "epoch": 0.4819277108433735, "grad_norm": 0.7348776459693909, "learning_rate": 2.9830297742912798e-05, "loss": 0.933, "step": 220 }, { "epoch": 0.4928806133625411, "grad_norm": 0.7400885820388794, "learning_rate": 2.981362265601676e-05, "loss": 0.8837, "step": 225 }, { "epoch": 0.5038335158817087, "grad_norm": 0.79625004529953, "learning_rate": 2.9796171358344578e-05, "loss": 0.9132, "step": 230 }, { "epoch": 0.5147864184008762, "grad_norm": 0.8105800747871399, "learning_rate": 2.9777944764317082e-05, "loss": 0.9753, "step": 235 }, { "epoch": 0.5257393209200438, "grad_norm": 0.8556815981864929, "learning_rate": 2.9758943828979444e-05, "loss": 0.9439, "step": 240 }, { "epoch": 0.5366922234392114, "grad_norm": 0.8349775671958923, "learning_rate": 2.9739169547951116e-05, "loss": 0.897, "step": 245 }, { "epoch": 0.547645125958379, "grad_norm": 0.6672224402427673, "learning_rate": 2.971862295737364e-05, "loss": 0.8832, "step": 250 }, { "epoch": 0.5585980284775466, "grad_norm": 0.7600496411323547, "learning_rate": 2.9697305133856394e-05, "loss": 0.8834, "step": 255 }, { "epoch": 0.5695509309967142, "grad_norm": 0.8782379031181335, "learning_rate": 2.9675217194420174e-05, "loss": 0.906, "step": 260 }, { "epoch": 0.5805038335158818, "grad_norm": 0.7369331121444702, "learning_rate": 2.9652360296438612e-05, "loss": 0.8909, "step": 265 }, { "epoch": 0.5914567360350493, "grad_norm": 0.7942291498184204, "learning_rate": 2.9628735637577612e-05, "loss": 0.8707, "step": 270 }, { "epoch": 0.6024096385542169, "grad_norm": 0.782035231590271, "learning_rate": 2.960434445573253e-05, "loss": 0.8819, "step": 275 }, { "epoch": 0.6133625410733844, "grad_norm": 0.9603850841522217, "learning_rate": 2.957918802896334e-05, "loss": 0.8689, "step": 280 }, { "epoch": 0.624315443592552, "grad_norm": 0.7741917967796326, "learning_rate": 2.9553267675427667e-05, "loss": 0.8543, "step": 285 }, { "epoch": 0.6352683461117196, "grad_norm": 0.7536712884902954, "learning_rate": 2.952658475331169e-05, "loss": 0.8515, "step": 290 }, { "epoch": 0.6462212486308871, "grad_norm": 0.8260005712509155, "learning_rate": 2.9499140660759018e-05, "loss": 0.8181, "step": 295 }, { "epoch": 0.6571741511500547, "grad_norm": 0.8287461400032043, "learning_rate": 2.9470936835797392e-05, "loss": 0.8527, "step": 300 }, { "epoch": 0.6681270536692223, "grad_norm": 0.8062161207199097, "learning_rate": 2.944197475626335e-05, "loss": 0.8389, "step": 305 }, { "epoch": 0.6790799561883899, "grad_norm": 0.8538839817047119, "learning_rate": 2.94122559397248e-05, "loss": 0.8177, "step": 310 }, { "epoch": 0.6900328587075575, "grad_norm": 0.7861185073852539, "learning_rate": 2.9381781943401477e-05, "loss": 0.839, "step": 315 }, { "epoch": 0.7009857612267251, "grad_norm": 0.9167773723602295, "learning_rate": 2.935055436408337e-05, "loss": 0.7825, "step": 320 }, { "epoch": 0.7119386637458927, "grad_norm": 0.807101309299469, "learning_rate": 2.9318574838047045e-05, "loss": 0.7862, "step": 325 }, { "epoch": 0.7228915662650602, "grad_norm": 0.8477482199668884, "learning_rate": 2.9285845040969906e-05, "loss": 0.8001, "step": 330 }, { "epoch": 0.7338444687842278, "grad_norm": 1.5396567583084106, "learning_rate": 2.9252366687842383e-05, "loss": 0.7666, "step": 335 }, { "epoch": 0.7447973713033954, "grad_norm": 0.8536133170127869, "learning_rate": 2.921814153287809e-05, "loss": 0.767, "step": 340 }, { "epoch": 0.755750273822563, "grad_norm": 0.8579494953155518, "learning_rate": 2.918317136942189e-05, "loss": 0.8055, "step": 345 }, { "epoch": 0.7667031763417306, "grad_norm": 0.8801035284996033, "learning_rate": 2.9147458029855933e-05, "loss": 0.7404, "step": 350 }, { "epoch": 0.7776560788608982, "grad_norm": 0.8523620963096619, "learning_rate": 2.911100338550364e-05, "loss": 0.7868, "step": 355 }, { "epoch": 0.7886089813800657, "grad_norm": 0.8762844800949097, "learning_rate": 2.9073809346531645e-05, "loss": 0.786, "step": 360 }, { "epoch": 0.7995618838992333, "grad_norm": 0.8992785215377808, "learning_rate": 2.903587786184972e-05, "loss": 0.7726, "step": 365 }, { "epoch": 0.8105147864184009, "grad_norm": 0.8587476015090942, "learning_rate": 2.899721091900863e-05, "loss": 0.7611, "step": 370 }, { "epoch": 0.8214676889375685, "grad_norm": 0.9412994980812073, "learning_rate": 2.895781054409602e-05, "loss": 0.7366, "step": 375 }, { "epoch": 0.8324205914567361, "grad_norm": 0.8805333375930786, "learning_rate": 2.891767880163022e-05, "loss": 0.7364, "step": 380 }, { "epoch": 0.8433734939759037, "grad_norm": 0.944273054599762, "learning_rate": 2.8876817794452087e-05, "loss": 0.7449, "step": 385 }, { "epoch": 0.8543263964950711, "grad_norm": 0.8578236103057861, "learning_rate": 2.883522966361482e-05, "loss": 0.7288, "step": 390 }, { "epoch": 0.8652792990142387, "grad_norm": 0.8459666967391968, "learning_rate": 2.8792916588271762e-05, "loss": 0.7062, "step": 395 }, { "epoch": 0.8762322015334063, "grad_norm": 0.9263506531715393, "learning_rate": 2.874988078556221e-05, "loss": 0.7203, "step": 400 }, { "epoch": 0.8871851040525739, "grad_norm": 0.9023187756538391, "learning_rate": 2.8706124510495263e-05, "loss": 0.6927, "step": 405 }, { "epoch": 0.8981380065717415, "grad_norm": 0.9095348119735718, "learning_rate": 2.8661650055831645e-05, "loss": 0.7178, "step": 410 }, { "epoch": 0.9090909090909091, "grad_norm": 0.9606519937515259, "learning_rate": 2.8616459751963583e-05, "loss": 0.7349, "step": 415 }, { "epoch": 0.9200438116100766, "grad_norm": 0.9363222122192383, "learning_rate": 2.8570555966792677e-05, "loss": 0.6886, "step": 420 }, { "epoch": 0.9309967141292442, "grad_norm": 0.8668004870414734, "learning_rate": 2.852394110560583e-05, "loss": 0.689, "step": 425 }, { "epoch": 0.9419496166484118, "grad_norm": 0.8470218181610107, "learning_rate": 2.8476617610949246e-05, "loss": 0.7109, "step": 430 }, { "epoch": 0.9529025191675794, "grad_norm": 1.0149718523025513, "learning_rate": 2.84285879625004e-05, "loss": 0.6739, "step": 435 }, { "epoch": 0.963855421686747, "grad_norm": 0.8750612139701843, "learning_rate": 2.8379854676938137e-05, "loss": 0.6981, "step": 440 }, { "epoch": 0.9748083242059146, "grad_norm": 0.9368709325790405, "learning_rate": 2.8330420307810784e-05, "loss": 0.6762, "step": 445 }, { "epoch": 0.9857612267250822, "grad_norm": 1.0392800569534302, "learning_rate": 2.828028744540236e-05, "loss": 0.6816, "step": 450 }, { "epoch": 0.9967141292442497, "grad_norm": 1.062735915184021, "learning_rate": 2.822945871659685e-05, "loss": 0.6951, "step": 455 } ], "logging_steps": 5, "max_steps": 2285, "num_input_tokens_seen": 0, "num_train_epochs": 5, "save_steps": 2000, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 6.804863634510971e+17, "train_batch_size": 2, "trial_name": null, "trial_params": null }