| { |
| "best_metric": null, |
| "best_model_checkpoint": null, |
| "epoch": 5.136986301369863, |
| "eval_steps": 500, |
| "global_step": 750, |
| "is_hyper_param_search": false, |
| "is_local_process_zero": true, |
| "is_world_process_zero": true, |
| "log_history": [ |
| { |
| "epoch": 0.0684931506849315, |
| "grad_norm": 6.335273265838623, |
| "learning_rate": 8.000000000000001e-06, |
| "loss": 0.954, |
| "step": 10 |
| }, |
| { |
| "epoch": 0.136986301369863, |
| "grad_norm": 2.381432056427002, |
| "learning_rate": 1.6000000000000003e-05, |
| "loss": 0.7291, |
| "step": 20 |
| }, |
| { |
| "epoch": 0.2054794520547945, |
| "grad_norm": 2.1213607788085938, |
| "learning_rate": 2.4e-05, |
| "loss": 0.3867, |
| "step": 30 |
| }, |
| { |
| "epoch": 0.273972602739726, |
| "grad_norm": 1.0599077939987183, |
| "learning_rate": 3.2000000000000005e-05, |
| "loss": 0.2747, |
| "step": 40 |
| }, |
| { |
| "epoch": 0.3424657534246575, |
| "grad_norm": 1.1717661619186401, |
| "learning_rate": 4e-05, |
| "loss": 0.2013, |
| "step": 50 |
| }, |
| { |
| "epoch": 0.410958904109589, |
| "grad_norm": 1.1004457473754883, |
| "learning_rate": 4.8e-05, |
| "loss": 0.1569, |
| "step": 60 |
| }, |
| { |
| "epoch": 0.4794520547945205, |
| "grad_norm": 1.0647892951965332, |
| "learning_rate": 5.6000000000000006e-05, |
| "loss": 0.1391, |
| "step": 70 |
| }, |
| { |
| "epoch": 0.547945205479452, |
| "grad_norm": 0.9715467095375061, |
| "learning_rate": 6.400000000000001e-05, |
| "loss": 0.1181, |
| "step": 80 |
| }, |
| { |
| "epoch": 0.6164383561643836, |
| "grad_norm": 0.8772512674331665, |
| "learning_rate": 7.2e-05, |
| "loss": 0.1051, |
| "step": 90 |
| }, |
| { |
| "epoch": 0.684931506849315, |
| "grad_norm": 1.2750014066696167, |
| "learning_rate": 8e-05, |
| "loss": 0.0947, |
| "step": 100 |
| }, |
| { |
| "epoch": 0.7534246575342466, |
| "grad_norm": 0.6222656965255737, |
| "learning_rate": 8.800000000000001e-05, |
| "loss": 0.0978, |
| "step": 110 |
| }, |
| { |
| "epoch": 0.821917808219178, |
| "grad_norm": 0.6259370446205139, |
| "learning_rate": 9.6e-05, |
| "loss": 0.0781, |
| "step": 120 |
| }, |
| { |
| "epoch": 0.8904109589041096, |
| "grad_norm": 0.6297749876976013, |
| "learning_rate": 9.999890641901125e-05, |
| "loss": 0.0738, |
| "step": 130 |
| }, |
| { |
| "epoch": 0.958904109589041, |
| "grad_norm": 0.4628015160560608, |
| "learning_rate": 9.999015805811965e-05, |
| "loss": 0.0748, |
| "step": 140 |
| }, |
| { |
| "epoch": 1.0273972602739727, |
| "grad_norm": 0.6147599816322327, |
| "learning_rate": 9.997266286704631e-05, |
| "loss": 0.0641, |
| "step": 150 |
| }, |
| { |
| "epoch": 1.095890410958904, |
| "grad_norm": 0.5399206280708313, |
| "learning_rate": 9.994642390694308e-05, |
| "loss": 0.0662, |
| "step": 160 |
| }, |
| { |
| "epoch": 1.1643835616438356, |
| "grad_norm": 0.44403907656669617, |
| "learning_rate": 9.991144576886823e-05, |
| "loss": 0.0592, |
| "step": 170 |
| }, |
| { |
| "epoch": 1.2328767123287672, |
| "grad_norm": 0.571832537651062, |
| "learning_rate": 9.986773457298311e-05, |
| "loss": 0.0556, |
| "step": 180 |
| }, |
| { |
| "epoch": 1.3013698630136985, |
| "grad_norm": 0.561576247215271, |
| "learning_rate": 9.981529796748134e-05, |
| "loss": 0.0596, |
| "step": 190 |
| }, |
| { |
| "epoch": 1.36986301369863, |
| "grad_norm": 0.4030166268348694, |
| "learning_rate": 9.975414512725057e-05, |
| "loss": 0.0539, |
| "step": 200 |
| }, |
| { |
| "epoch": 1.4383561643835616, |
| "grad_norm": 0.34468358755111694, |
| "learning_rate": 9.968428675226714e-05, |
| "loss": 0.0577, |
| "step": 210 |
| }, |
| { |
| "epoch": 1.5068493150684932, |
| "grad_norm": 0.483900785446167, |
| "learning_rate": 9.96057350657239e-05, |
| "loss": 0.052, |
| "step": 220 |
| }, |
| { |
| "epoch": 1.5753424657534247, |
| "grad_norm": 0.4898487627506256, |
| "learning_rate": 9.95185038118915e-05, |
| "loss": 0.0477, |
| "step": 230 |
| }, |
| { |
| "epoch": 1.643835616438356, |
| "grad_norm": 0.5045291781425476, |
| "learning_rate": 9.942260825371358e-05, |
| "loss": 0.0528, |
| "step": 240 |
| }, |
| { |
| "epoch": 1.7123287671232876, |
| "grad_norm": 0.5695160031318665, |
| "learning_rate": 9.931806517013612e-05, |
| "loss": 0.0439, |
| "step": 250 |
| }, |
| { |
| "epoch": 1.7808219178082192, |
| "grad_norm": 0.33678483963012695, |
| "learning_rate": 9.92048928531717e-05, |
| "loss": 0.0488, |
| "step": 260 |
| }, |
| { |
| "epoch": 1.8493150684931505, |
| "grad_norm": 0.3467772305011749, |
| "learning_rate": 9.90831111046988e-05, |
| "loss": 0.0448, |
| "step": 270 |
| }, |
| { |
| "epoch": 1.9178082191780823, |
| "grad_norm": 0.4071213901042938, |
| "learning_rate": 9.895274123299723e-05, |
| "loss": 0.0431, |
| "step": 280 |
| }, |
| { |
| "epoch": 1.9863013698630136, |
| "grad_norm": 0.30543234944343567, |
| "learning_rate": 9.881380604901964e-05, |
| "loss": 0.0429, |
| "step": 290 |
| }, |
| { |
| "epoch": 2.0547945205479454, |
| "grad_norm": 0.6378459334373474, |
| "learning_rate": 9.86663298624003e-05, |
| "loss": 0.0464, |
| "step": 300 |
| }, |
| { |
| "epoch": 2.1232876712328768, |
| "grad_norm": 0.4035351276397705, |
| "learning_rate": 9.851033847720166e-05, |
| "loss": 0.0488, |
| "step": 310 |
| }, |
| { |
| "epoch": 2.191780821917808, |
| "grad_norm": 0.3094054162502289, |
| "learning_rate": 9.834585918739936e-05, |
| "loss": 0.0364, |
| "step": 320 |
| }, |
| { |
| "epoch": 2.26027397260274, |
| "grad_norm": 0.45768025517463684, |
| "learning_rate": 9.817292077210659e-05, |
| "loss": 0.0373, |
| "step": 330 |
| }, |
| { |
| "epoch": 2.328767123287671, |
| "grad_norm": 0.504475474357605, |
| "learning_rate": 9.799155349053851e-05, |
| "loss": 0.047, |
| "step": 340 |
| }, |
| { |
| "epoch": 2.3972602739726026, |
| "grad_norm": 0.36838439106941223, |
| "learning_rate": 9.780178907671789e-05, |
| "loss": 0.0396, |
| "step": 350 |
| }, |
| { |
| "epoch": 2.4657534246575343, |
| "grad_norm": 0.3892686665058136, |
| "learning_rate": 9.760366073392246e-05, |
| "loss": 0.04, |
| "step": 360 |
| }, |
| { |
| "epoch": 2.5342465753424657, |
| "grad_norm": 0.376288503408432, |
| "learning_rate": 9.739720312887535e-05, |
| "loss": 0.0352, |
| "step": 370 |
| }, |
| { |
| "epoch": 2.602739726027397, |
| "grad_norm": 0.33039844036102295, |
| "learning_rate": 9.718245238567939e-05, |
| "loss": 0.0364, |
| "step": 380 |
| }, |
| { |
| "epoch": 2.671232876712329, |
| "grad_norm": 0.38675811886787415, |
| "learning_rate": 9.695944607949649e-05, |
| "loss": 0.0367, |
| "step": 390 |
| }, |
| { |
| "epoch": 2.73972602739726, |
| "grad_norm": 0.3453963100910187, |
| "learning_rate": 9.672822322997305e-05, |
| "loss": 0.0337, |
| "step": 400 |
| }, |
| { |
| "epoch": 2.808219178082192, |
| "grad_norm": 0.5285444259643555, |
| "learning_rate": 9.648882429441257e-05, |
| "loss": 0.0414, |
| "step": 410 |
| }, |
| { |
| "epoch": 2.8767123287671232, |
| "grad_norm": 0.38015297055244446, |
| "learning_rate": 9.624129116069694e-05, |
| "loss": 0.0348, |
| "step": 420 |
| }, |
| { |
| "epoch": 2.9452054794520546, |
| "grad_norm": 0.44161203503608704, |
| "learning_rate": 9.598566713995718e-05, |
| "loss": 0.0377, |
| "step": 430 |
| }, |
| { |
| "epoch": 3.0136986301369864, |
| "grad_norm": 0.4227602779865265, |
| "learning_rate": 9.572199695899522e-05, |
| "loss": 0.0323, |
| "step": 440 |
| }, |
| { |
| "epoch": 3.0821917808219177, |
| "grad_norm": 0.2722010314464569, |
| "learning_rate": 9.545032675245813e-05, |
| "loss": 0.0345, |
| "step": 450 |
| }, |
| { |
| "epoch": 3.1506849315068495, |
| "grad_norm": 0.4375353753566742, |
| "learning_rate": 9.517070405476575e-05, |
| "loss": 0.0315, |
| "step": 460 |
| }, |
| { |
| "epoch": 3.219178082191781, |
| "grad_norm": 0.4301266372203827, |
| "learning_rate": 9.488317779179361e-05, |
| "loss": 0.0348, |
| "step": 470 |
| }, |
| { |
| "epoch": 3.287671232876712, |
| "grad_norm": 0.458469957113266, |
| "learning_rate": 9.458779827231237e-05, |
| "loss": 0.0348, |
| "step": 480 |
| }, |
| { |
| "epoch": 3.356164383561644, |
| "grad_norm": 0.3810347616672516, |
| "learning_rate": 9.428461717918511e-05, |
| "loss": 0.0304, |
| "step": 490 |
| }, |
| { |
| "epoch": 3.4246575342465753, |
| "grad_norm": 0.3728407919406891, |
| "learning_rate": 9.397368756032445e-05, |
| "loss": 0.0332, |
| "step": 500 |
| }, |
| { |
| "epoch": 3.493150684931507, |
| "grad_norm": 0.3293081223964691, |
| "learning_rate": 9.365506381941066e-05, |
| "loss": 0.0305, |
| "step": 510 |
| }, |
| { |
| "epoch": 3.5616438356164384, |
| "grad_norm": 0.35907241702079773, |
| "learning_rate": 9.332880170637252e-05, |
| "loss": 0.0319, |
| "step": 520 |
| }, |
| { |
| "epoch": 3.6301369863013697, |
| "grad_norm": 0.3396986126899719, |
| "learning_rate": 9.299495830763286e-05, |
| "loss": 0.0319, |
| "step": 530 |
| }, |
| { |
| "epoch": 3.6986301369863015, |
| "grad_norm": 0.22632896900177002, |
| "learning_rate": 9.265359203611987e-05, |
| "loss": 0.0318, |
| "step": 540 |
| }, |
| { |
| "epoch": 3.767123287671233, |
| "grad_norm": 0.2579493522644043, |
| "learning_rate": 9.230476262104677e-05, |
| "loss": 0.0295, |
| "step": 550 |
| }, |
| { |
| "epoch": 3.8356164383561646, |
| "grad_norm": 0.2979259788990021, |
| "learning_rate": 9.194853109746074e-05, |
| "loss": 0.0317, |
| "step": 560 |
| }, |
| { |
| "epoch": 3.904109589041096, |
| "grad_norm": 0.28193944692611694, |
| "learning_rate": 9.158495979556358e-05, |
| "loss": 0.0324, |
| "step": 570 |
| }, |
| { |
| "epoch": 3.9726027397260273, |
| "grad_norm": 0.30196675658226013, |
| "learning_rate": 9.121411232980588e-05, |
| "loss": 0.0282, |
| "step": 580 |
| }, |
| { |
| "epoch": 4.041095890410959, |
| "grad_norm": 0.2790427505970001, |
| "learning_rate": 9.083605358775612e-05, |
| "loss": 0.0322, |
| "step": 590 |
| }, |
| { |
| "epoch": 4.109589041095891, |
| "grad_norm": 0.3180605173110962, |
| "learning_rate": 9.045084971874738e-05, |
| "loss": 0.0311, |
| "step": 600 |
| }, |
| { |
| "epoch": 4.178082191780822, |
| "grad_norm": 0.23422259092330933, |
| "learning_rate": 9.005856812230304e-05, |
| "loss": 0.0313, |
| "step": 610 |
| }, |
| { |
| "epoch": 4.2465753424657535, |
| "grad_norm": 0.1887388378381729, |
| "learning_rate": 8.965927743634391e-05, |
| "loss": 0.0297, |
| "step": 620 |
| }, |
| { |
| "epoch": 4.315068493150685, |
| "grad_norm": 0.21844805777072906, |
| "learning_rate": 8.92530475251784e-05, |
| "loss": 0.0306, |
| "step": 630 |
| }, |
| { |
| "epoch": 4.383561643835616, |
| "grad_norm": 0.45273661613464355, |
| "learning_rate": 8.883994946727849e-05, |
| "loss": 0.0292, |
| "step": 640 |
| }, |
| { |
| "epoch": 4.4520547945205475, |
| "grad_norm": 0.39670127630233765, |
| "learning_rate": 8.842005554284296e-05, |
| "loss": 0.0331, |
| "step": 650 |
| }, |
| { |
| "epoch": 4.52054794520548, |
| "grad_norm": 0.25612273812294006, |
| "learning_rate": 8.799343922115044e-05, |
| "loss": 0.0305, |
| "step": 660 |
| }, |
| { |
| "epoch": 4.589041095890411, |
| "grad_norm": 0.22720636427402496, |
| "learning_rate": 8.756017514770443e-05, |
| "loss": 0.0284, |
| "step": 670 |
| }, |
| { |
| "epoch": 4.657534246575342, |
| "grad_norm": 0.39644864201545715, |
| "learning_rate": 8.71203391311725e-05, |
| "loss": 0.0286, |
| "step": 680 |
| }, |
| { |
| "epoch": 4.726027397260274, |
| "grad_norm": 0.4030825197696686, |
| "learning_rate": 8.6674008130122e-05, |
| "loss": 0.0268, |
| "step": 690 |
| }, |
| { |
| "epoch": 4.794520547945205, |
| "grad_norm": 0.33811622858047485, |
| "learning_rate": 8.622126023955446e-05, |
| "loss": 0.0274, |
| "step": 700 |
| }, |
| { |
| "epoch": 4.863013698630137, |
| "grad_norm": 0.27626654505729675, |
| "learning_rate": 8.576217467724128e-05, |
| "loss": 0.0267, |
| "step": 710 |
| }, |
| { |
| "epoch": 4.931506849315069, |
| "grad_norm": 0.3249574899673462, |
| "learning_rate": 8.529683176986295e-05, |
| "loss": 0.0297, |
| "step": 720 |
| }, |
| { |
| "epoch": 5.0, |
| "grad_norm": 0.6223666071891785, |
| "learning_rate": 8.482531293895412e-05, |
| "loss": 0.0289, |
| "step": 730 |
| }, |
| { |
| "epoch": 5.068493150684931, |
| "grad_norm": 0.4200395941734314, |
| "learning_rate": 8.434770068665723e-05, |
| "loss": 0.0293, |
| "step": 740 |
| }, |
| { |
| "epoch": 5.136986301369863, |
| "grad_norm": 0.33773040771484375, |
| "learning_rate": 8.386407858128706e-05, |
| "loss": 0.0243, |
| "step": 750 |
| } |
| ], |
| "logging_steps": 10, |
| "max_steps": 2500, |
| "num_input_tokens_seen": 0, |
| "num_train_epochs": 18, |
| "save_steps": 125, |
| "stateful_callbacks": { |
| "TrainerControl": { |
| "args": { |
| "should_epoch_stop": false, |
| "should_evaluate": false, |
| "should_log": false, |
| "should_save": true, |
| "should_training_stop": false |
| }, |
| "attributes": {} |
| } |
| }, |
| "total_flos": 1.0243918948403232e+17, |
| "train_batch_size": 64, |
| "trial_name": null, |
| "trial_params": null |
| } |
|
|