{ "best_global_step": null, "best_metric": null, "best_model_checkpoint": null, "epoch": 8.0, "eval_steps": 50, "global_step": 1104, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.007246376811594203, "grad_norm": 1.9687033891677856, "learning_rate": 0.0, "loss": 19.2682, "step": 1 }, { "epoch": 0.07246376811594203, "grad_norm": 1.703112006187439, "learning_rate": 1.8e-05, "loss": 12.4529, "step": 10 }, { "epoch": 0.14492753623188406, "grad_norm": 3.6209306716918945, "learning_rate": 3.8e-05, "loss": 17.5169, "step": 20 }, { "epoch": 0.21739130434782608, "grad_norm": 2.09627366065979, "learning_rate": 5.8e-05, "loss": 16.0619, "step": 30 }, { "epoch": 0.2898550724637681, "grad_norm": 1.9405275583267212, "learning_rate": 7.800000000000001e-05, "loss": 20.228, "step": 40 }, { "epoch": 0.36231884057971014, "grad_norm": 8.32076644897461, "learning_rate": 9.8e-05, "loss": 16.0457, "step": 50 }, { "epoch": 0.43478260869565216, "grad_norm": 1.1994603872299194, "learning_rate": 9.9146110056926e-05, "loss": 15.9109, "step": 60 }, { "epoch": 0.5072463768115942, "grad_norm": 1.3328566551208496, "learning_rate": 9.819734345351043e-05, "loss": 15.6408, "step": 70 }, { "epoch": 0.5797101449275363, "grad_norm": 6.663642406463623, "learning_rate": 9.724857685009489e-05, "loss": 13.4492, "step": 80 }, { "epoch": 0.6521739130434783, "grad_norm": 4.154050350189209, "learning_rate": 9.629981024667933e-05, "loss": 16.0048, "step": 90 }, { "epoch": 0.7246376811594203, "grad_norm": 12.047609329223633, "learning_rate": 9.535104364326376e-05, "loss": 42.9753, "step": 100 }, { "epoch": 0.7971014492753623, "grad_norm": 4.798079490661621, "learning_rate": 9.44022770398482e-05, "loss": 17.2692, "step": 110 }, { "epoch": 0.8695652173913043, "grad_norm": 6.930491924285889, "learning_rate": 9.345351043643265e-05, "loss": 13.9401, "step": 120 }, { "epoch": 0.9420289855072463, "grad_norm": 7.896566867828369, "learning_rate": 9.250474383301708e-05, "loss": 16.1242, "step": 130 }, { "epoch": 1.0144927536231885, "grad_norm": 10.504218101501465, "learning_rate": 9.155597722960152e-05, "loss": 9.8417, "step": 140 }, { "epoch": 1.0869565217391304, "grad_norm": 9.594452857971191, "learning_rate": 9.060721062618596e-05, "loss": 12.6395, "step": 150 }, { "epoch": 1.1594202898550725, "grad_norm": 16.113290786743164, "learning_rate": 8.96584440227704e-05, "loss": 11.6141, "step": 160 }, { "epoch": 1.2318840579710144, "grad_norm": 10.904824256896973, "learning_rate": 8.870967741935484e-05, "loss": 15.8041, "step": 170 }, { "epoch": 1.3043478260869565, "grad_norm": 4.2903852462768555, "learning_rate": 8.776091081593929e-05, "loss": 10.5756, "step": 180 }, { "epoch": 1.3768115942028984, "grad_norm": 3.3363027572631836, "learning_rate": 8.681214421252373e-05, "loss": 12.2632, "step": 190 }, { "epoch": 1.4492753623188406, "grad_norm": 5.582556247711182, "learning_rate": 8.586337760910817e-05, "loss": 10.2026, "step": 200 }, { "epoch": 1.5217391304347827, "grad_norm": 3.090460777282715, "learning_rate": 8.49146110056926e-05, "loss": 9.8113, "step": 210 }, { "epoch": 1.5942028985507246, "grad_norm": 20.330148696899414, "learning_rate": 8.396584440227704e-05, "loss": 9.2478, "step": 220 }, { "epoch": 1.6666666666666665, "grad_norm": 16.418106079101562, "learning_rate": 8.301707779886149e-05, "loss": 11.377, "step": 230 }, { "epoch": 1.7391304347826086, "grad_norm": 1.9852267503738403, "learning_rate": 8.206831119544592e-05, "loss": 9.8674, "step": 240 }, { "epoch": 1.8115942028985508, "grad_norm": 4.2618489265441895, "learning_rate": 8.111954459203036e-05, "loss": 10.9124, "step": 250 }, { "epoch": 1.8840579710144927, "grad_norm": 5.334084510803223, "learning_rate": 8.017077798861481e-05, "loss": 29.9208, "step": 260 }, { "epoch": 1.9565217391304348, "grad_norm": 7.662145614624023, "learning_rate": 7.922201138519924e-05, "loss": 9.4491, "step": 270 }, { "epoch": 2.028985507246377, "grad_norm": 2.990098476409912, "learning_rate": 7.827324478178368e-05, "loss": 12.7051, "step": 280 }, { "epoch": 2.101449275362319, "grad_norm": 11.343267440795898, "learning_rate": 7.732447817836812e-05, "loss": 10.1094, "step": 290 }, { "epoch": 2.1739130434782608, "grad_norm": 4.076178550720215, "learning_rate": 7.637571157495257e-05, "loss": 10.8557, "step": 300 }, { "epoch": 2.246376811594203, "grad_norm": 3.965240240097046, "learning_rate": 7.542694497153701e-05, "loss": 10.2597, "step": 310 }, { "epoch": 2.318840579710145, "grad_norm": 16.37909507751465, "learning_rate": 7.447817836812145e-05, "loss": 9.0119, "step": 320 }, { "epoch": 2.391304347826087, "grad_norm": 3.6150219440460205, "learning_rate": 7.352941176470589e-05, "loss": 26.6018, "step": 330 }, { "epoch": 2.463768115942029, "grad_norm": 19.880943298339844, "learning_rate": 7.258064516129033e-05, "loss": 9.6189, "step": 340 }, { "epoch": 2.536231884057971, "grad_norm": 3.5458195209503174, "learning_rate": 7.163187855787477e-05, "loss": 8.4988, "step": 350 }, { "epoch": 2.608695652173913, "grad_norm": 13.963035583496094, "learning_rate": 7.06831119544592e-05, "loss": 11.8011, "step": 360 }, { "epoch": 2.681159420289855, "grad_norm": 19.048301696777344, "learning_rate": 6.973434535104365e-05, "loss": 10.9278, "step": 370 }, { "epoch": 2.753623188405797, "grad_norm": 4.992802143096924, "learning_rate": 6.878557874762808e-05, "loss": 8.919, "step": 380 }, { "epoch": 2.8260869565217392, "grad_norm": 11.363718032836914, "learning_rate": 6.783681214421252e-05, "loss": 13.1761, "step": 390 }, { "epoch": 2.898550724637681, "grad_norm": 5.374908924102783, "learning_rate": 6.688804554079696e-05, "loss": 10.3222, "step": 400 }, { "epoch": 2.971014492753623, "grad_norm": 8.949808120727539, "learning_rate": 6.59392789373814e-05, "loss": 10.7587, "step": 410 }, { "epoch": 3.0434782608695654, "grad_norm": 4.645139217376709, "learning_rate": 6.499051233396585e-05, "loss": 7.0093, "step": 420 }, { "epoch": 3.1159420289855073, "grad_norm": 1.7357378005981445, "learning_rate": 6.404174573055029e-05, "loss": 10.5791, "step": 430 }, { "epoch": 3.1884057971014492, "grad_norm": 6.632950305938721, "learning_rate": 6.309297912713473e-05, "loss": 8.7916, "step": 440 }, { "epoch": 3.260869565217391, "grad_norm": 3.3891420364379883, "learning_rate": 6.214421252371917e-05, "loss": 10.3499, "step": 450 }, { "epoch": 3.3333333333333335, "grad_norm": 10.14654541015625, "learning_rate": 6.119544592030361e-05, "loss": 8.5833, "step": 460 }, { "epoch": 3.4057971014492754, "grad_norm": 3.27583384513855, "learning_rate": 6.0246679316888046e-05, "loss": 28.3443, "step": 470 }, { "epoch": 3.4782608695652173, "grad_norm": 16.218280792236328, "learning_rate": 5.9297912713472494e-05, "loss": 11.3033, "step": 480 }, { "epoch": 3.550724637681159, "grad_norm": 5.851198196411133, "learning_rate": 5.834914611005693e-05, "loss": 7.5228, "step": 490 }, { "epoch": 3.6231884057971016, "grad_norm": 13.043642044067383, "learning_rate": 5.740037950664137e-05, "loss": 11.608, "step": 500 }, { "epoch": 3.6956521739130435, "grad_norm": 3.651071071624756, "learning_rate": 5.645161290322582e-05, "loss": 9.3775, "step": 510 }, { "epoch": 3.7681159420289854, "grad_norm": 3.946401357650757, "learning_rate": 5.550284629981025e-05, "loss": 9.9621, "step": 520 }, { "epoch": 3.8405797101449277, "grad_norm": 5.363879680633545, "learning_rate": 5.4554079696394686e-05, "loss": 7.2515, "step": 530 }, { "epoch": 3.9130434782608696, "grad_norm": 18.565500259399414, "learning_rate": 5.360531309297913e-05, "loss": 9.9015, "step": 540 }, { "epoch": 3.9855072463768115, "grad_norm": 2.068690299987793, "learning_rate": 5.2656546489563575e-05, "loss": 8.6521, "step": 550 }, { "epoch": 4.057971014492754, "grad_norm": 8.377117156982422, "learning_rate": 5.170777988614801e-05, "loss": 11.6352, "step": 560 }, { "epoch": 4.130434782608695, "grad_norm": 6.822054386138916, "learning_rate": 5.075901328273245e-05, "loss": 8.9412, "step": 570 }, { "epoch": 4.202898550724638, "grad_norm": 17.24335479736328, "learning_rate": 4.981024667931689e-05, "loss": 9.5539, "step": 580 }, { "epoch": 4.27536231884058, "grad_norm": 13.434685707092285, "learning_rate": 4.8861480075901326e-05, "loss": 10.8367, "step": 590 }, { "epoch": 4.3478260869565215, "grad_norm": 17.375398635864258, "learning_rate": 4.791271347248577e-05, "loss": 10.818, "step": 600 }, { "epoch": 4.420289855072464, "grad_norm": 4.53552770614624, "learning_rate": 4.6963946869070216e-05, "loss": 7.442, "step": 610 }, { "epoch": 4.492753623188406, "grad_norm": 4.74644136428833, "learning_rate": 4.601518026565465e-05, "loss": 9.6362, "step": 620 }, { "epoch": 4.565217391304348, "grad_norm": 3.9156153202056885, "learning_rate": 4.506641366223909e-05, "loss": 11.4203, "step": 630 }, { "epoch": 4.63768115942029, "grad_norm": 18.31324005126953, "learning_rate": 4.411764705882353e-05, "loss": 8.557, "step": 640 }, { "epoch": 4.710144927536232, "grad_norm": 4.924636363983154, "learning_rate": 4.3168880455407974e-05, "loss": 9.608, "step": 650 }, { "epoch": 4.782608695652174, "grad_norm": 5.3758769035339355, "learning_rate": 4.222011385199241e-05, "loss": 8.9764, "step": 660 }, { "epoch": 4.855072463768116, "grad_norm": 8.584567070007324, "learning_rate": 4.1271347248576856e-05, "loss": 9.621, "step": 670 }, { "epoch": 4.927536231884058, "grad_norm": 4.82368803024292, "learning_rate": 4.032258064516129e-05, "loss": 24.5753, "step": 680 }, { "epoch": 5.0, "grad_norm": 2.3411295413970947, "learning_rate": 3.937381404174573e-05, "loss": 7.9662, "step": 690 }, { "epoch": 5.072463768115942, "grad_norm": 3.99580979347229, "learning_rate": 3.842504743833017e-05, "loss": 6.8737, "step": 700 }, { "epoch": 5.144927536231884, "grad_norm": 24.173742294311523, "learning_rate": 3.7476280834914614e-05, "loss": 11.8038, "step": 710 }, { "epoch": 5.217391304347826, "grad_norm": 25.25533676147461, "learning_rate": 3.6527514231499055e-05, "loss": 11.9047, "step": 720 }, { "epoch": 5.2898550724637685, "grad_norm": 4.739432334899902, "learning_rate": 3.557874762808349e-05, "loss": 9.503, "step": 730 }, { "epoch": 5.36231884057971, "grad_norm": 7.874780178070068, "learning_rate": 3.462998102466794e-05, "loss": 10.949, "step": 740 }, { "epoch": 5.434782608695652, "grad_norm": 3.363119602203369, "learning_rate": 3.368121442125237e-05, "loss": 10.2591, "step": 750 }, { "epoch": 5.507246376811594, "grad_norm": 5.406454086303711, "learning_rate": 3.273244781783681e-05, "loss": 7.7785, "step": 760 }, { "epoch": 5.579710144927536, "grad_norm": 6.118616580963135, "learning_rate": 3.1783681214421254e-05, "loss": 7.2315, "step": 770 }, { "epoch": 5.6521739130434785, "grad_norm": 16.83717918395996, "learning_rate": 3.0834914611005695e-05, "loss": 7.6301, "step": 780 }, { "epoch": 5.72463768115942, "grad_norm": 11.110983848571777, "learning_rate": 2.9886148007590137e-05, "loss": 9.153, "step": 790 }, { "epoch": 5.797101449275362, "grad_norm": 8.291231155395508, "learning_rate": 2.8937381404174574e-05, "loss": 10.3227, "step": 800 }, { "epoch": 5.869565217391305, "grad_norm": 2.8294785022735596, "learning_rate": 2.7988614800759016e-05, "loss": 6.8461, "step": 810 }, { "epoch": 5.942028985507246, "grad_norm": 11.522751808166504, "learning_rate": 2.7039848197343453e-05, "loss": 25.7992, "step": 820 }, { "epoch": 6.0144927536231885, "grad_norm": 5.239048480987549, "learning_rate": 2.6091081593927898e-05, "loss": 10.1367, "step": 830 }, { "epoch": 6.086956521739131, "grad_norm": 14.244694709777832, "learning_rate": 2.5142314990512332e-05, "loss": 8.1876, "step": 840 }, { "epoch": 6.159420289855072, "grad_norm": 3.612367868423462, "learning_rate": 2.4193548387096777e-05, "loss": 9.5329, "step": 850 }, { "epoch": 6.231884057971015, "grad_norm": 4.118394374847412, "learning_rate": 2.3244781783681215e-05, "loss": 8.8621, "step": 860 }, { "epoch": 6.304347826086957, "grad_norm": 2.245436906814575, "learning_rate": 2.2296015180265656e-05, "loss": 7.363, "step": 870 }, { "epoch": 6.3768115942028984, "grad_norm": 7.4832658767700195, "learning_rate": 2.1347248576850097e-05, "loss": 8.4077, "step": 880 }, { "epoch": 6.449275362318841, "grad_norm": 20.146770477294922, "learning_rate": 2.0398481973434535e-05, "loss": 27.7444, "step": 890 }, { "epoch": 6.521739130434782, "grad_norm": 6.686460018157959, "learning_rate": 1.9449715370018976e-05, "loss": 9.9147, "step": 900 }, { "epoch": 6.594202898550725, "grad_norm": 11.729459762573242, "learning_rate": 1.8500948766603414e-05, "loss": 10.427, "step": 910 }, { "epoch": 6.666666666666667, "grad_norm": 5.6112518310546875, "learning_rate": 1.7552182163187855e-05, "loss": 8.1236, "step": 920 }, { "epoch": 6.739130434782608, "grad_norm": 21.59635353088379, "learning_rate": 1.66034155597723e-05, "loss": 8.0663, "step": 930 }, { "epoch": 6.811594202898551, "grad_norm": 7.840078353881836, "learning_rate": 1.5654648956356737e-05, "loss": 11.4745, "step": 940 }, { "epoch": 6.884057971014493, "grad_norm": 8.633024215698242, "learning_rate": 1.4705882352941177e-05, "loss": 8.9603, "step": 950 }, { "epoch": 6.956521739130435, "grad_norm": 2.9932422637939453, "learning_rate": 1.3757115749525618e-05, "loss": 8.8133, "step": 960 }, { "epoch": 7.028985507246377, "grad_norm": 4.314597129821777, "learning_rate": 1.2808349146110058e-05, "loss": 6.8162, "step": 970 }, { "epoch": 7.101449275362318, "grad_norm": 18.329147338867188, "learning_rate": 1.1859582542694497e-05, "loss": 11.9377, "step": 980 }, { "epoch": 7.173913043478261, "grad_norm": 7.561063289642334, "learning_rate": 1.0910815939278938e-05, "loss": 6.8045, "step": 990 }, { "epoch": 7.246376811594203, "grad_norm": 5.871479511260986, "learning_rate": 9.962049335863378e-06, "loss": 9.549, "step": 1000 }, { "epoch": 7.318840579710145, "grad_norm": 3.1535987854003906, "learning_rate": 9.013282732447819e-06, "loss": 6.7738, "step": 1010 }, { "epoch": 7.391304347826087, "grad_norm": 32.92097854614258, "learning_rate": 8.064516129032258e-06, "loss": 27.4036, "step": 1020 }, { "epoch": 7.463768115942029, "grad_norm": 7.275607109069824, "learning_rate": 7.115749525616698e-06, "loss": 9.3496, "step": 1030 }, { "epoch": 7.536231884057971, "grad_norm": 7.355132579803467, "learning_rate": 6.166982922201139e-06, "loss": 7.1842, "step": 1040 }, { "epoch": 7.608695652173913, "grad_norm": 11.539506912231445, "learning_rate": 5.218216318785579e-06, "loss": 10.1211, "step": 1050 }, { "epoch": 7.681159420289855, "grad_norm": 5.036221981048584, "learning_rate": 4.269449715370019e-06, "loss": 10.1483, "step": 1060 }, { "epoch": 7.753623188405797, "grad_norm": 24.608356475830078, "learning_rate": 3.320683111954459e-06, "loss": 9.3202, "step": 1070 }, { "epoch": 7.826086956521739, "grad_norm": 9.034127235412598, "learning_rate": 2.3719165085388996e-06, "loss": 9.5748, "step": 1080 }, { "epoch": 7.898550724637682, "grad_norm": 13.333020210266113, "learning_rate": 1.4231499051233397e-06, "loss": 8.6569, "step": 1090 }, { "epoch": 7.971014492753623, "grad_norm": 7.102665424346924, "learning_rate": 4.743833017077799e-07, "loss": 8.791, "step": 1100 } ], "logging_steps": 10, "max_steps": 1104, "num_input_tokens_seen": 0, "num_train_epochs": 8, "save_steps": 50, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 109892759371776.0, "train_batch_size": 4, "trial_name": null, "trial_params": null }