| { | |
| "best_global_step": null, | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 29.41176470588235, | |
| "eval_steps": 500, | |
| "global_step": 2000, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.14705882352941177, | |
| "grad_norm": 2.2869467735290527, | |
| "learning_rate": 1.9911764705882353e-05, | |
| "loss": 2.6394, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.29411764705882354, | |
| "grad_norm": 1.5338459014892578, | |
| "learning_rate": 1.981372549019608e-05, | |
| "loss": 2.6312, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.4411764705882353, | |
| "grad_norm": 2.0163090229034424, | |
| "learning_rate": 1.9715686274509805e-05, | |
| "loss": 2.6409, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.5882352941176471, | |
| "grad_norm": 3.02846097946167, | |
| "learning_rate": 1.961764705882353e-05, | |
| "loss": 2.6167, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.7352941176470589, | |
| "grad_norm": 4.507878303527832, | |
| "learning_rate": 1.9519607843137257e-05, | |
| "loss": 2.6231, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.8823529411764706, | |
| "grad_norm": 2.433847188949585, | |
| "learning_rate": 1.9421568627450982e-05, | |
| "loss": 2.6153, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "eval_loss": 2.579890012741089, | |
| "eval_runtime": 1.6326, | |
| "eval_samples_per_second": 82.691, | |
| "eval_steps_per_second": 10.413, | |
| "step": 68 | |
| }, | |
| { | |
| "epoch": 1.0294117647058822, | |
| "grad_norm": 3.4059886932373047, | |
| "learning_rate": 1.9323529411764706e-05, | |
| "loss": 2.5863, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 1.1764705882352942, | |
| "grad_norm": 2.305633068084717, | |
| "learning_rate": 1.9225490196078434e-05, | |
| "loss": 2.5677, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 1.3235294117647058, | |
| "grad_norm": 2.493675470352173, | |
| "learning_rate": 1.912745098039216e-05, | |
| "loss": 2.5061, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 1.4705882352941178, | |
| "grad_norm": 2.6469180583953857, | |
| "learning_rate": 1.9029411764705883e-05, | |
| "loss": 2.4657, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 1.6176470588235294, | |
| "grad_norm": 3.5626657009124756, | |
| "learning_rate": 1.8931372549019607e-05, | |
| "loss": 2.4594, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 1.7647058823529411, | |
| "grad_norm": 2.9186179637908936, | |
| "learning_rate": 1.8833333333333335e-05, | |
| "loss": 2.4393, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 1.9117647058823528, | |
| "grad_norm": 4.3771772384643555, | |
| "learning_rate": 1.873529411764706e-05, | |
| "loss": 2.3884, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 2.0, | |
| "eval_loss": 2.311964750289917, | |
| "eval_runtime": 1.6428, | |
| "eval_samples_per_second": 82.178, | |
| "eval_steps_per_second": 10.348, | |
| "step": 136 | |
| }, | |
| { | |
| "epoch": 2.0588235294117645, | |
| "grad_norm": 2.706939935684204, | |
| "learning_rate": 1.8637254901960787e-05, | |
| "loss": 2.3395, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 2.2058823529411766, | |
| "grad_norm": 3.1243748664855957, | |
| "learning_rate": 1.853921568627451e-05, | |
| "loss": 2.2711, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 2.3529411764705883, | |
| "grad_norm": 2.9195902347564697, | |
| "learning_rate": 1.844117647058824e-05, | |
| "loss": 2.2252, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 2.5, | |
| "grad_norm": 5.713516712188721, | |
| "learning_rate": 1.834313725490196e-05, | |
| "loss": 2.2211, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 2.6470588235294117, | |
| "grad_norm": 2.779804229736328, | |
| "learning_rate": 1.8245098039215688e-05, | |
| "loss": 2.1251, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 2.7941176470588234, | |
| "grad_norm": 2.9840755462646484, | |
| "learning_rate": 1.8147058823529412e-05, | |
| "loss": 2.1162, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 2.9411764705882355, | |
| "grad_norm": 3.6137640476226807, | |
| "learning_rate": 1.804901960784314e-05, | |
| "loss": 2.0859, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 3.0, | |
| "eval_loss": 1.982031226158142, | |
| "eval_runtime": 1.6147, | |
| "eval_samples_per_second": 83.606, | |
| "eval_steps_per_second": 10.528, | |
| "step": 204 | |
| }, | |
| { | |
| "epoch": 3.088235294117647, | |
| "grad_norm": 2.8977713584899902, | |
| "learning_rate": 1.7950980392156864e-05, | |
| "loss": 2.0281, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 3.235294117647059, | |
| "grad_norm": 3.248961925506592, | |
| "learning_rate": 1.7852941176470592e-05, | |
| "loss": 1.9454, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 3.3823529411764706, | |
| "grad_norm": 3.525045394897461, | |
| "learning_rate": 1.7754901960784313e-05, | |
| "loss": 1.8824, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 3.5294117647058822, | |
| "grad_norm": 3.95782470703125, | |
| "learning_rate": 1.765686274509804e-05, | |
| "loss": 1.8538, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 3.6764705882352944, | |
| "grad_norm": 4.282369613647461, | |
| "learning_rate": 1.7558823529411765e-05, | |
| "loss": 1.823, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 3.8235294117647056, | |
| "grad_norm": 6.753427505493164, | |
| "learning_rate": 1.7460784313725493e-05, | |
| "loss": 1.8116, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 3.9705882352941178, | |
| "grad_norm": 4.552963733673096, | |
| "learning_rate": 1.7362745098039217e-05, | |
| "loss": 1.7387, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 4.0, | |
| "eval_loss": 1.6489075422286987, | |
| "eval_runtime": 1.6483, | |
| "eval_samples_per_second": 81.902, | |
| "eval_steps_per_second": 10.314, | |
| "step": 272 | |
| }, | |
| { | |
| "epoch": 4.117647058823529, | |
| "grad_norm": 3.1702029705047607, | |
| "learning_rate": 1.7264705882352945e-05, | |
| "loss": 1.6522, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 4.264705882352941, | |
| "grad_norm": 4.5589494705200195, | |
| "learning_rate": 1.7166666666666666e-05, | |
| "loss": 1.6078, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 4.411764705882353, | |
| "grad_norm": 3.9019124507904053, | |
| "learning_rate": 1.7068627450980394e-05, | |
| "loss": 1.5599, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 4.5588235294117645, | |
| "grad_norm": 4.917637348175049, | |
| "learning_rate": 1.6970588235294118e-05, | |
| "loss": 1.5237, | |
| "step": 310 | |
| }, | |
| { | |
| "epoch": 4.705882352941177, | |
| "grad_norm": 3.638333320617676, | |
| "learning_rate": 1.6872549019607846e-05, | |
| "loss": 1.4842, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 4.852941176470588, | |
| "grad_norm": 3.488672971725464, | |
| "learning_rate": 1.677450980392157e-05, | |
| "loss": 1.4391, | |
| "step": 330 | |
| }, | |
| { | |
| "epoch": 5.0, | |
| "grad_norm": 8.806844711303711, | |
| "learning_rate": 1.6676470588235295e-05, | |
| "loss": 1.4165, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 5.0, | |
| "eval_loss": 1.3353697061538696, | |
| "eval_runtime": 1.6321, | |
| "eval_samples_per_second": 82.718, | |
| "eval_steps_per_second": 10.416, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 5.147058823529412, | |
| "grad_norm": 5.466882705688477, | |
| "learning_rate": 1.6578431372549022e-05, | |
| "loss": 1.3357, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 5.294117647058823, | |
| "grad_norm": 3.95070481300354, | |
| "learning_rate": 1.6480392156862747e-05, | |
| "loss": 1.3317, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 5.4411764705882355, | |
| "grad_norm": 3.2761778831481934, | |
| "learning_rate": 1.638235294117647e-05, | |
| "loss": 1.1954, | |
| "step": 370 | |
| }, | |
| { | |
| "epoch": 5.588235294117647, | |
| "grad_norm": 2.8257029056549072, | |
| "learning_rate": 1.62843137254902e-05, | |
| "loss": 1.1713, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 5.735294117647059, | |
| "grad_norm": 3.5745885372161865, | |
| "learning_rate": 1.6186274509803923e-05, | |
| "loss": 1.1767, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 5.882352941176471, | |
| "grad_norm": 3.625549077987671, | |
| "learning_rate": 1.6088235294117648e-05, | |
| "loss": 1.1028, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 6.0, | |
| "eval_loss": 1.0512332916259766, | |
| "eval_runtime": 1.6283, | |
| "eval_samples_per_second": 82.911, | |
| "eval_steps_per_second": 10.441, | |
| "step": 408 | |
| }, | |
| { | |
| "epoch": 6.029411764705882, | |
| "grad_norm": 3.7580277919769287, | |
| "learning_rate": 1.5990196078431375e-05, | |
| "loss": 1.0918, | |
| "step": 410 | |
| }, | |
| { | |
| "epoch": 6.176470588235294, | |
| "grad_norm": 5.0767292976379395, | |
| "learning_rate": 1.58921568627451e-05, | |
| "loss": 1.0799, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 6.323529411764706, | |
| "grad_norm": 3.362245798110962, | |
| "learning_rate": 1.5794117647058824e-05, | |
| "loss": 1.0246, | |
| "step": 430 | |
| }, | |
| { | |
| "epoch": 6.470588235294118, | |
| "grad_norm": 2.663010835647583, | |
| "learning_rate": 1.569607843137255e-05, | |
| "loss": 0.9335, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 6.617647058823529, | |
| "grad_norm": 2.7811036109924316, | |
| "learning_rate": 1.5598039215686276e-05, | |
| "loss": 0.9084, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 6.764705882352941, | |
| "grad_norm": 6.619537830352783, | |
| "learning_rate": 1.55e-05, | |
| "loss": 0.8652, | |
| "step": 460 | |
| }, | |
| { | |
| "epoch": 6.911764705882353, | |
| "grad_norm": 4.524352073669434, | |
| "learning_rate": 1.540196078431373e-05, | |
| "loss": 0.8706, | |
| "step": 470 | |
| }, | |
| { | |
| "epoch": 7.0, | |
| "eval_loss": 0.837912380695343, | |
| "eval_runtime": 1.6194, | |
| "eval_samples_per_second": 83.366, | |
| "eval_steps_per_second": 10.498, | |
| "step": 476 | |
| }, | |
| { | |
| "epoch": 7.0588235294117645, | |
| "grad_norm": 4.127895832061768, | |
| "learning_rate": 1.5303921568627453e-05, | |
| "loss": 0.7721, | |
| "step": 480 | |
| }, | |
| { | |
| "epoch": 7.205882352941177, | |
| "grad_norm": 2.9023818969726562, | |
| "learning_rate": 1.5205882352941177e-05, | |
| "loss": 0.7268, | |
| "step": 490 | |
| }, | |
| { | |
| "epoch": 7.352941176470588, | |
| "grad_norm": 3.2444231510162354, | |
| "learning_rate": 1.5107843137254903e-05, | |
| "loss": 0.7142, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 7.5, | |
| "grad_norm": 3.160628318786621, | |
| "learning_rate": 1.500980392156863e-05, | |
| "loss": 0.7019, | |
| "step": 510 | |
| }, | |
| { | |
| "epoch": 7.647058823529412, | |
| "grad_norm": 2.913454294204712, | |
| "learning_rate": 1.4911764705882354e-05, | |
| "loss": 0.6911, | |
| "step": 520 | |
| }, | |
| { | |
| "epoch": 7.794117647058823, | |
| "grad_norm": 2.9466731548309326, | |
| "learning_rate": 1.481372549019608e-05, | |
| "loss": 0.6639, | |
| "step": 530 | |
| }, | |
| { | |
| "epoch": 7.9411764705882355, | |
| "grad_norm": 2.8029139041900635, | |
| "learning_rate": 1.4715686274509806e-05, | |
| "loss": 0.5554, | |
| "step": 540 | |
| }, | |
| { | |
| "epoch": 8.0, | |
| "eval_loss": 0.5985532402992249, | |
| "eval_runtime": 1.6867, | |
| "eval_samples_per_second": 80.038, | |
| "eval_steps_per_second": 10.079, | |
| "step": 544 | |
| }, | |
| { | |
| "epoch": 8.088235294117647, | |
| "grad_norm": 2.267561912536621, | |
| "learning_rate": 1.461764705882353e-05, | |
| "loss": 0.5757, | |
| "step": 550 | |
| }, | |
| { | |
| "epoch": 8.235294117647058, | |
| "grad_norm": 2.1232762336730957, | |
| "learning_rate": 1.4519607843137256e-05, | |
| "loss": 0.5091, | |
| "step": 560 | |
| }, | |
| { | |
| "epoch": 8.382352941176471, | |
| "grad_norm": 3.035518169403076, | |
| "learning_rate": 1.442156862745098e-05, | |
| "loss": 0.5158, | |
| "step": 570 | |
| }, | |
| { | |
| "epoch": 8.529411764705882, | |
| "grad_norm": 2.085606336593628, | |
| "learning_rate": 1.4323529411764707e-05, | |
| "loss": 0.4673, | |
| "step": 580 | |
| }, | |
| { | |
| "epoch": 8.676470588235293, | |
| "grad_norm": 1.8189260959625244, | |
| "learning_rate": 1.4225490196078433e-05, | |
| "loss": 0.4604, | |
| "step": 590 | |
| }, | |
| { | |
| "epoch": 8.823529411764707, | |
| "grad_norm": 6.74431037902832, | |
| "learning_rate": 1.4127450980392159e-05, | |
| "loss": 0.4651, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 8.970588235294118, | |
| "grad_norm": 1.56160306930542, | |
| "learning_rate": 1.4029411764705883e-05, | |
| "loss": 0.4017, | |
| "step": 610 | |
| }, | |
| { | |
| "epoch": 9.0, | |
| "eval_loss": 0.40496423840522766, | |
| "eval_runtime": 1.6727, | |
| "eval_samples_per_second": 80.707, | |
| "eval_steps_per_second": 10.163, | |
| "step": 612 | |
| }, | |
| { | |
| "epoch": 9.117647058823529, | |
| "grad_norm": 1.9897135496139526, | |
| "learning_rate": 1.3931372549019608e-05, | |
| "loss": 0.3879, | |
| "step": 620 | |
| }, | |
| { | |
| "epoch": 9.264705882352942, | |
| "grad_norm": 1.6671721935272217, | |
| "learning_rate": 1.3833333333333334e-05, | |
| "loss": 0.3206, | |
| "step": 630 | |
| }, | |
| { | |
| "epoch": 9.411764705882353, | |
| "grad_norm": 1.8617277145385742, | |
| "learning_rate": 1.373529411764706e-05, | |
| "loss": 0.3565, | |
| "step": 640 | |
| }, | |
| { | |
| "epoch": 9.558823529411764, | |
| "grad_norm": 1.64924955368042, | |
| "learning_rate": 1.3637254901960786e-05, | |
| "loss": 0.3457, | |
| "step": 650 | |
| }, | |
| { | |
| "epoch": 9.705882352941176, | |
| "grad_norm": 1.5722205638885498, | |
| "learning_rate": 1.3539215686274512e-05, | |
| "loss": 0.3123, | |
| "step": 660 | |
| }, | |
| { | |
| "epoch": 9.852941176470589, | |
| "grad_norm": 2.3067121505737305, | |
| "learning_rate": 1.3441176470588238e-05, | |
| "loss": 0.3137, | |
| "step": 670 | |
| }, | |
| { | |
| "epoch": 10.0, | |
| "grad_norm": 4.091474533081055, | |
| "learning_rate": 1.334313725490196e-05, | |
| "loss": 0.2663, | |
| "step": 680 | |
| }, | |
| { | |
| "epoch": 10.0, | |
| "eval_loss": 0.29727470874786377, | |
| "eval_runtime": 1.6004, | |
| "eval_samples_per_second": 84.356, | |
| "eval_steps_per_second": 10.623, | |
| "step": 680 | |
| }, | |
| { | |
| "epoch": 10.147058823529411, | |
| "grad_norm": 1.217851161956787, | |
| "learning_rate": 1.3245098039215687e-05, | |
| "loss": 0.2133, | |
| "step": 690 | |
| }, | |
| { | |
| "epoch": 10.294117647058824, | |
| "grad_norm": 1.5889488458633423, | |
| "learning_rate": 1.3147058823529413e-05, | |
| "loss": 0.2299, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 10.441176470588236, | |
| "grad_norm": 1.2852109670639038, | |
| "learning_rate": 1.3049019607843139e-05, | |
| "loss": 0.1991, | |
| "step": 710 | |
| }, | |
| { | |
| "epoch": 10.588235294117647, | |
| "grad_norm": 1.1250885725021362, | |
| "learning_rate": 1.2950980392156865e-05, | |
| "loss": 0.2625, | |
| "step": 720 | |
| }, | |
| { | |
| "epoch": 10.735294117647058, | |
| "grad_norm": 0.8768565654754639, | |
| "learning_rate": 1.285294117647059e-05, | |
| "loss": 0.1657, | |
| "step": 730 | |
| }, | |
| { | |
| "epoch": 10.882352941176471, | |
| "grad_norm": 1.0275033712387085, | |
| "learning_rate": 1.2754901960784314e-05, | |
| "loss": 0.1814, | |
| "step": 740 | |
| }, | |
| { | |
| "epoch": 11.0, | |
| "eval_loss": 0.22045107185840607, | |
| "eval_runtime": 1.6056, | |
| "eval_samples_per_second": 84.083, | |
| "eval_steps_per_second": 10.588, | |
| "step": 748 | |
| }, | |
| { | |
| "epoch": 11.029411764705882, | |
| "grad_norm": 1.1377185583114624, | |
| "learning_rate": 1.265686274509804e-05, | |
| "loss": 0.2123, | |
| "step": 750 | |
| }, | |
| { | |
| "epoch": 11.176470588235293, | |
| "grad_norm": 0.6125518679618835, | |
| "learning_rate": 1.2558823529411766e-05, | |
| "loss": 0.145, | |
| "step": 760 | |
| }, | |
| { | |
| "epoch": 11.323529411764707, | |
| "grad_norm": 3.132908344268799, | |
| "learning_rate": 1.2460784313725492e-05, | |
| "loss": 0.1409, | |
| "step": 770 | |
| }, | |
| { | |
| "epoch": 11.470588235294118, | |
| "grad_norm": 7.12354850769043, | |
| "learning_rate": 1.2362745098039218e-05, | |
| "loss": 0.171, | |
| "step": 780 | |
| }, | |
| { | |
| "epoch": 11.617647058823529, | |
| "grad_norm": 0.9431573152542114, | |
| "learning_rate": 1.2264705882352944e-05, | |
| "loss": 0.1561, | |
| "step": 790 | |
| }, | |
| { | |
| "epoch": 11.764705882352942, | |
| "grad_norm": 0.5725237131118774, | |
| "learning_rate": 1.2166666666666667e-05, | |
| "loss": 0.1012, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 11.911764705882353, | |
| "grad_norm": 0.6124898791313171, | |
| "learning_rate": 1.2068627450980393e-05, | |
| "loss": 0.1072, | |
| "step": 810 | |
| }, | |
| { | |
| "epoch": 12.0, | |
| "eval_loss": 0.20114047825336456, | |
| "eval_runtime": 1.61, | |
| "eval_samples_per_second": 83.85, | |
| "eval_steps_per_second": 10.559, | |
| "step": 816 | |
| }, | |
| { | |
| "epoch": 12.058823529411764, | |
| "grad_norm": 0.6034418344497681, | |
| "learning_rate": 1.1970588235294119e-05, | |
| "loss": 0.1093, | |
| "step": 820 | |
| }, | |
| { | |
| "epoch": 12.205882352941176, | |
| "grad_norm": 0.6741713881492615, | |
| "learning_rate": 1.1872549019607845e-05, | |
| "loss": 0.0901, | |
| "step": 830 | |
| }, | |
| { | |
| "epoch": 12.352941176470589, | |
| "grad_norm": 0.7810522317886353, | |
| "learning_rate": 1.177450980392157e-05, | |
| "loss": 0.089, | |
| "step": 840 | |
| }, | |
| { | |
| "epoch": 12.5, | |
| "grad_norm": 0.6442411541938782, | |
| "learning_rate": 1.1676470588235295e-05, | |
| "loss": 0.1219, | |
| "step": 850 | |
| }, | |
| { | |
| "epoch": 12.647058823529411, | |
| "grad_norm": 3.0149757862091064, | |
| "learning_rate": 1.1578431372549021e-05, | |
| "loss": 0.0814, | |
| "step": 860 | |
| }, | |
| { | |
| "epoch": 12.794117647058824, | |
| "grad_norm": 0.46087804436683655, | |
| "learning_rate": 1.1480392156862746e-05, | |
| "loss": 0.1077, | |
| "step": 870 | |
| }, | |
| { | |
| "epoch": 12.941176470588236, | |
| "grad_norm": 0.56253981590271, | |
| "learning_rate": 1.1382352941176472e-05, | |
| "loss": 0.068, | |
| "step": 880 | |
| }, | |
| { | |
| "epoch": 13.0, | |
| "eval_loss": 0.1641794890165329, | |
| "eval_runtime": 1.6752, | |
| "eval_samples_per_second": 80.589, | |
| "eval_steps_per_second": 10.148, | |
| "step": 884 | |
| }, | |
| { | |
| "epoch": 13.088235294117647, | |
| "grad_norm": 0.6027768850326538, | |
| "learning_rate": 1.1284313725490198e-05, | |
| "loss": 0.0689, | |
| "step": 890 | |
| }, | |
| { | |
| "epoch": 13.235294117647058, | |
| "grad_norm": 0.369243860244751, | |
| "learning_rate": 1.1186274509803922e-05, | |
| "loss": 0.0613, | |
| "step": 900 | |
| }, | |
| { | |
| "epoch": 13.382352941176471, | |
| "grad_norm": 0.3756796419620514, | |
| "learning_rate": 1.1088235294117648e-05, | |
| "loss": 0.0599, | |
| "step": 910 | |
| }, | |
| { | |
| "epoch": 13.529411764705882, | |
| "grad_norm": 0.42388486862182617, | |
| "learning_rate": 1.0990196078431374e-05, | |
| "loss": 0.0561, | |
| "step": 920 | |
| }, | |
| { | |
| "epoch": 13.676470588235293, | |
| "grad_norm": 0.40439847111701965, | |
| "learning_rate": 1.0892156862745099e-05, | |
| "loss": 0.0542, | |
| "step": 930 | |
| }, | |
| { | |
| "epoch": 13.823529411764707, | |
| "grad_norm": 0.39990735054016113, | |
| "learning_rate": 1.0794117647058825e-05, | |
| "loss": 0.0672, | |
| "step": 940 | |
| }, | |
| { | |
| "epoch": 13.970588235294118, | |
| "grad_norm": 0.3704688251018524, | |
| "learning_rate": 1.0696078431372549e-05, | |
| "loss": 0.1051, | |
| "step": 950 | |
| }, | |
| { | |
| "epoch": 14.0, | |
| "eval_loss": 0.14415012300014496, | |
| "eval_runtime": 1.6471, | |
| "eval_samples_per_second": 81.961, | |
| "eval_steps_per_second": 10.321, | |
| "step": 952 | |
| }, | |
| { | |
| "epoch": 14.117647058823529, | |
| "grad_norm": 0.33351826667785645, | |
| "learning_rate": 1.0598039215686275e-05, | |
| "loss": 0.0492, | |
| "step": 960 | |
| }, | |
| { | |
| "epoch": 14.264705882352942, | |
| "grad_norm": 0.3827894926071167, | |
| "learning_rate": 1.0500000000000001e-05, | |
| "loss": 0.0484, | |
| "step": 970 | |
| }, | |
| { | |
| "epoch": 14.411764705882353, | |
| "grad_norm": 0.4303022623062134, | |
| "learning_rate": 1.0401960784313727e-05, | |
| "loss": 0.0461, | |
| "step": 980 | |
| }, | |
| { | |
| "epoch": 14.558823529411764, | |
| "grad_norm": 0.3589096665382385, | |
| "learning_rate": 1.0303921568627452e-05, | |
| "loss": 0.0468, | |
| "step": 990 | |
| }, | |
| { | |
| "epoch": 14.705882352941176, | |
| "grad_norm": 0.30145418643951416, | |
| "learning_rate": 1.0205882352941176e-05, | |
| "loss": 0.0942, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 14.852941176470589, | |
| "grad_norm": 0.3331129550933838, | |
| "learning_rate": 1.0107843137254902e-05, | |
| "loss": 0.0465, | |
| "step": 1010 | |
| }, | |
| { | |
| "epoch": 15.0, | |
| "grad_norm": 0.790107011795044, | |
| "learning_rate": 1.0009803921568628e-05, | |
| "loss": 0.0431, | |
| "step": 1020 | |
| }, | |
| { | |
| "epoch": 15.0, | |
| "eval_loss": 0.1515953093767166, | |
| "eval_runtime": 1.5961, | |
| "eval_samples_per_second": 84.583, | |
| "eval_steps_per_second": 10.651, | |
| "step": 1020 | |
| }, | |
| { | |
| "epoch": 15.147058823529411, | |
| "grad_norm": 0.4401082992553711, | |
| "learning_rate": 9.911764705882354e-06, | |
| "loss": 0.0511, | |
| "step": 1030 | |
| }, | |
| { | |
| "epoch": 15.294117647058824, | |
| "grad_norm": 0.3193589150905609, | |
| "learning_rate": 9.813725490196078e-06, | |
| "loss": 0.0369, | |
| "step": 1040 | |
| }, | |
| { | |
| "epoch": 15.441176470588236, | |
| "grad_norm": 0.33625075221061707, | |
| "learning_rate": 9.715686274509805e-06, | |
| "loss": 0.0363, | |
| "step": 1050 | |
| }, | |
| { | |
| "epoch": 15.588235294117647, | |
| "grad_norm": 0.18927061557769775, | |
| "learning_rate": 9.61764705882353e-06, | |
| "loss": 0.0349, | |
| "step": 1060 | |
| }, | |
| { | |
| "epoch": 15.735294117647058, | |
| "grad_norm": 0.28063905239105225, | |
| "learning_rate": 9.519607843137255e-06, | |
| "loss": 0.0355, | |
| "step": 1070 | |
| }, | |
| { | |
| "epoch": 15.882352941176471, | |
| "grad_norm": 0.31923410296440125, | |
| "learning_rate": 9.421568627450981e-06, | |
| "loss": 0.0949, | |
| "step": 1080 | |
| }, | |
| { | |
| "epoch": 16.0, | |
| "eval_loss": 0.16671152412891388, | |
| "eval_runtime": 1.6889, | |
| "eval_samples_per_second": 79.936, | |
| "eval_steps_per_second": 10.066, | |
| "step": 1088 | |
| }, | |
| { | |
| "epoch": 16.029411764705884, | |
| "grad_norm": 0.25026965141296387, | |
| "learning_rate": 9.323529411764707e-06, | |
| "loss": 0.0357, | |
| "step": 1090 | |
| }, | |
| { | |
| "epoch": 16.176470588235293, | |
| "grad_norm": 0.39554500579833984, | |
| "learning_rate": 9.225490196078433e-06, | |
| "loss": 0.0946, | |
| "step": 1100 | |
| }, | |
| { | |
| "epoch": 16.323529411764707, | |
| "grad_norm": 0.30123069882392883, | |
| "learning_rate": 9.127450980392158e-06, | |
| "loss": 0.0322, | |
| "step": 1110 | |
| }, | |
| { | |
| "epoch": 16.470588235294116, | |
| "grad_norm": 0.25656214356422424, | |
| "learning_rate": 9.029411764705884e-06, | |
| "loss": 0.0291, | |
| "step": 1120 | |
| }, | |
| { | |
| "epoch": 16.61764705882353, | |
| "grad_norm": 0.21169237792491913, | |
| "learning_rate": 8.93137254901961e-06, | |
| "loss": 0.029, | |
| "step": 1130 | |
| }, | |
| { | |
| "epoch": 16.764705882352942, | |
| "grad_norm": 0.22442790865898132, | |
| "learning_rate": 8.833333333333334e-06, | |
| "loss": 0.0316, | |
| "step": 1140 | |
| }, | |
| { | |
| "epoch": 16.91176470588235, | |
| "grad_norm": 0.20120474696159363, | |
| "learning_rate": 8.73529411764706e-06, | |
| "loss": 0.0287, | |
| "step": 1150 | |
| }, | |
| { | |
| "epoch": 17.0, | |
| "eval_loss": 0.13797470927238464, | |
| "eval_runtime": 1.6743, | |
| "eval_samples_per_second": 80.63, | |
| "eval_steps_per_second": 10.153, | |
| "step": 1156 | |
| }, | |
| { | |
| "epoch": 17.058823529411764, | |
| "grad_norm": 0.20405986905097961, | |
| "learning_rate": 8.637254901960786e-06, | |
| "loss": 0.0321, | |
| "step": 1160 | |
| }, | |
| { | |
| "epoch": 17.205882352941178, | |
| "grad_norm": 0.22055025398731232, | |
| "learning_rate": 8.53921568627451e-06, | |
| "loss": 0.027, | |
| "step": 1170 | |
| }, | |
| { | |
| "epoch": 17.352941176470587, | |
| "grad_norm": 0.18829534947872162, | |
| "learning_rate": 8.441176470588237e-06, | |
| "loss": 0.0769, | |
| "step": 1180 | |
| }, | |
| { | |
| "epoch": 17.5, | |
| "grad_norm": 0.7278969883918762, | |
| "learning_rate": 8.343137254901961e-06, | |
| "loss": 0.0276, | |
| "step": 1190 | |
| }, | |
| { | |
| "epoch": 17.647058823529413, | |
| "grad_norm": 0.22244372963905334, | |
| "learning_rate": 8.245098039215687e-06, | |
| "loss": 0.0263, | |
| "step": 1200 | |
| }, | |
| { | |
| "epoch": 17.794117647058822, | |
| "grad_norm": 0.1744055598974228, | |
| "learning_rate": 8.147058823529413e-06, | |
| "loss": 0.0241, | |
| "step": 1210 | |
| }, | |
| { | |
| "epoch": 17.941176470588236, | |
| "grad_norm": 0.16366326808929443, | |
| "learning_rate": 8.049019607843137e-06, | |
| "loss": 0.0246, | |
| "step": 1220 | |
| }, | |
| { | |
| "epoch": 18.0, | |
| "eval_loss": 0.15256382524967194, | |
| "eval_runtime": 1.6045, | |
| "eval_samples_per_second": 84.139, | |
| "eval_steps_per_second": 10.595, | |
| "step": 1224 | |
| }, | |
| { | |
| "epoch": 18.08823529411765, | |
| "grad_norm": 0.15204133093357086, | |
| "learning_rate": 7.950980392156864e-06, | |
| "loss": 0.0254, | |
| "step": 1230 | |
| }, | |
| { | |
| "epoch": 18.235294117647058, | |
| "grad_norm": 0.16595087945461273, | |
| "learning_rate": 7.85294117647059e-06, | |
| "loss": 0.0243, | |
| "step": 1240 | |
| }, | |
| { | |
| "epoch": 18.38235294117647, | |
| "grad_norm": 0.2022247016429901, | |
| "learning_rate": 7.754901960784314e-06, | |
| "loss": 0.0242, | |
| "step": 1250 | |
| }, | |
| { | |
| "epoch": 18.529411764705884, | |
| "grad_norm": 0.17330515384674072, | |
| "learning_rate": 7.65686274509804e-06, | |
| "loss": 0.024, | |
| "step": 1260 | |
| }, | |
| { | |
| "epoch": 18.676470588235293, | |
| "grad_norm": 0.20826362073421478, | |
| "learning_rate": 7.558823529411765e-06, | |
| "loss": 0.0801, | |
| "step": 1270 | |
| }, | |
| { | |
| "epoch": 18.823529411764707, | |
| "grad_norm": 0.2036994844675064, | |
| "learning_rate": 7.460784313725491e-06, | |
| "loss": 0.0232, | |
| "step": 1280 | |
| }, | |
| { | |
| "epoch": 18.970588235294116, | |
| "grad_norm": 0.1467086523771286, | |
| "learning_rate": 7.3627450980392165e-06, | |
| "loss": 0.0217, | |
| "step": 1290 | |
| }, | |
| { | |
| "epoch": 19.0, | |
| "eval_loss": 0.14522351324558258, | |
| "eval_runtime": 1.6184, | |
| "eval_samples_per_second": 83.417, | |
| "eval_steps_per_second": 10.504, | |
| "step": 1292 | |
| }, | |
| { | |
| "epoch": 19.11764705882353, | |
| "grad_norm": 0.21091414988040924, | |
| "learning_rate": 7.264705882352942e-06, | |
| "loss": 0.0217, | |
| "step": 1300 | |
| }, | |
| { | |
| "epoch": 19.264705882352942, | |
| "grad_norm": 0.16935530304908752, | |
| "learning_rate": 7.166666666666667e-06, | |
| "loss": 0.0742, | |
| "step": 1310 | |
| }, | |
| { | |
| "epoch": 19.41176470588235, | |
| "grad_norm": 0.19531255960464478, | |
| "learning_rate": 7.068627450980393e-06, | |
| "loss": 0.0203, | |
| "step": 1320 | |
| }, | |
| { | |
| "epoch": 19.558823529411764, | |
| "grad_norm": 0.18234974145889282, | |
| "learning_rate": 6.970588235294118e-06, | |
| "loss": 0.0539, | |
| "step": 1330 | |
| }, | |
| { | |
| "epoch": 19.705882352941178, | |
| "grad_norm": 0.1523897647857666, | |
| "learning_rate": 6.8725490196078434e-06, | |
| "loss": 0.0207, | |
| "step": 1340 | |
| }, | |
| { | |
| "epoch": 19.852941176470587, | |
| "grad_norm": 0.1887982189655304, | |
| "learning_rate": 6.7745098039215695e-06, | |
| "loss": 0.0205, | |
| "step": 1350 | |
| }, | |
| { | |
| "epoch": 20.0, | |
| "grad_norm": 1.2818535566329956, | |
| "learning_rate": 6.676470588235294e-06, | |
| "loss": 0.0216, | |
| "step": 1360 | |
| }, | |
| { | |
| "epoch": 20.0, | |
| "eval_loss": 0.15050244331359863, | |
| "eval_runtime": 1.6089, | |
| "eval_samples_per_second": 83.907, | |
| "eval_steps_per_second": 10.566, | |
| "step": 1360 | |
| }, | |
| { | |
| "epoch": 20.147058823529413, | |
| "grad_norm": 0.1478956639766693, | |
| "learning_rate": 6.57843137254902e-06, | |
| "loss": 0.0181, | |
| "step": 1370 | |
| }, | |
| { | |
| "epoch": 20.294117647058822, | |
| "grad_norm": 0.18312789499759674, | |
| "learning_rate": 6.480392156862746e-06, | |
| "loss": 0.1025, | |
| "step": 1380 | |
| }, | |
| { | |
| "epoch": 20.441176470588236, | |
| "grad_norm": 0.15513214468955994, | |
| "learning_rate": 6.38235294117647e-06, | |
| "loss": 0.0209, | |
| "step": 1390 | |
| }, | |
| { | |
| "epoch": 20.58823529411765, | |
| "grad_norm": 0.13837294280529022, | |
| "learning_rate": 6.2843137254901964e-06, | |
| "loss": 0.0199, | |
| "step": 1400 | |
| }, | |
| { | |
| "epoch": 20.735294117647058, | |
| "grad_norm": 0.1465756893157959, | |
| "learning_rate": 6.1862745098039225e-06, | |
| "loss": 0.0192, | |
| "step": 1410 | |
| }, | |
| { | |
| "epoch": 20.88235294117647, | |
| "grad_norm": 0.16790196299552917, | |
| "learning_rate": 6.088235294117647e-06, | |
| "loss": 0.0195, | |
| "step": 1420 | |
| }, | |
| { | |
| "epoch": 21.0, | |
| "eval_loss": 0.1510893851518631, | |
| "eval_runtime": 1.6277, | |
| "eval_samples_per_second": 82.937, | |
| "eval_steps_per_second": 10.444, | |
| "step": 1428 | |
| }, | |
| { | |
| "epoch": 21.029411764705884, | |
| "grad_norm": 0.29194769263267517, | |
| "learning_rate": 5.990196078431373e-06, | |
| "loss": 0.0181, | |
| "step": 1430 | |
| }, | |
| { | |
| "epoch": 21.176470588235293, | |
| "grad_norm": 0.16755153238773346, | |
| "learning_rate": 5.892156862745099e-06, | |
| "loss": 0.0186, | |
| "step": 1440 | |
| }, | |
| { | |
| "epoch": 21.323529411764707, | |
| "grad_norm": 0.19387616217136383, | |
| "learning_rate": 5.794117647058824e-06, | |
| "loss": 0.0184, | |
| "step": 1450 | |
| }, | |
| { | |
| "epoch": 21.470588235294116, | |
| "grad_norm": 0.15613146126270294, | |
| "learning_rate": 5.6960784313725494e-06, | |
| "loss": 0.0196, | |
| "step": 1460 | |
| }, | |
| { | |
| "epoch": 21.61764705882353, | |
| "grad_norm": 0.12751713395118713, | |
| "learning_rate": 5.598039215686275e-06, | |
| "loss": 0.0178, | |
| "step": 1470 | |
| }, | |
| { | |
| "epoch": 21.764705882352942, | |
| "grad_norm": 0.15414807200431824, | |
| "learning_rate": 5.500000000000001e-06, | |
| "loss": 0.0177, | |
| "step": 1480 | |
| }, | |
| { | |
| "epoch": 21.91176470588235, | |
| "grad_norm": 0.12436957657337189, | |
| "learning_rate": 5.401960784313726e-06, | |
| "loss": 0.0749, | |
| "step": 1490 | |
| }, | |
| { | |
| "epoch": 22.0, | |
| "eval_loss": 0.14262542128562927, | |
| "eval_runtime": 1.6896, | |
| "eval_samples_per_second": 79.901, | |
| "eval_steps_per_second": 10.062, | |
| "step": 1496 | |
| }, | |
| { | |
| "epoch": 22.058823529411764, | |
| "grad_norm": 0.1797289252281189, | |
| "learning_rate": 5.303921568627451e-06, | |
| "loss": 0.0163, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 22.205882352941178, | |
| "grad_norm": 0.15396994352340698, | |
| "learning_rate": 5.205882352941177e-06, | |
| "loss": 0.0164, | |
| "step": 1510 | |
| }, | |
| { | |
| "epoch": 22.352941176470587, | |
| "grad_norm": 0.43706995248794556, | |
| "learning_rate": 5.107843137254902e-06, | |
| "loss": 0.0173, | |
| "step": 1520 | |
| }, | |
| { | |
| "epoch": 22.5, | |
| "grad_norm": 0.13553638756275177, | |
| "learning_rate": 5.009803921568628e-06, | |
| "loss": 0.0701, | |
| "step": 1530 | |
| }, | |
| { | |
| "epoch": 22.647058823529413, | |
| "grad_norm": 0.14506298303604126, | |
| "learning_rate": 4.911764705882353e-06, | |
| "loss": 0.0158, | |
| "step": 1540 | |
| }, | |
| { | |
| "epoch": 22.794117647058822, | |
| "grad_norm": 0.25496020913124084, | |
| "learning_rate": 4.813725490196079e-06, | |
| "loss": 0.0152, | |
| "step": 1550 | |
| }, | |
| { | |
| "epoch": 22.941176470588236, | |
| "grad_norm": 0.16556541621685028, | |
| "learning_rate": 4.715686274509804e-06, | |
| "loss": 0.0157, | |
| "step": 1560 | |
| }, | |
| { | |
| "epoch": 23.0, | |
| "eval_loss": 0.1430669128894806, | |
| "eval_runtime": 1.6521, | |
| "eval_samples_per_second": 81.714, | |
| "eval_steps_per_second": 10.29, | |
| "step": 1564 | |
| }, | |
| { | |
| "epoch": 23.08823529411765, | |
| "grad_norm": 0.12428230047225952, | |
| "learning_rate": 4.61764705882353e-06, | |
| "loss": 0.0159, | |
| "step": 1570 | |
| }, | |
| { | |
| "epoch": 23.235294117647058, | |
| "grad_norm": 0.11920207738876343, | |
| "learning_rate": 4.519607843137255e-06, | |
| "loss": 0.0145, | |
| "step": 1580 | |
| }, | |
| { | |
| "epoch": 23.38235294117647, | |
| "grad_norm": 0.1940208524465561, | |
| "learning_rate": 4.421568627450981e-06, | |
| "loss": 0.0162, | |
| "step": 1590 | |
| }, | |
| { | |
| "epoch": 23.529411764705884, | |
| "grad_norm": 0.1417732536792755, | |
| "learning_rate": 4.323529411764707e-06, | |
| "loss": 0.0149, | |
| "step": 1600 | |
| }, | |
| { | |
| "epoch": 23.676470588235293, | |
| "grad_norm": 0.13574868440628052, | |
| "learning_rate": 4.225490196078432e-06, | |
| "loss": 0.0152, | |
| "step": 1610 | |
| }, | |
| { | |
| "epoch": 23.823529411764707, | |
| "grad_norm": 0.1438971906900406, | |
| "learning_rate": 4.127450980392157e-06, | |
| "loss": 0.0677, | |
| "step": 1620 | |
| }, | |
| { | |
| "epoch": 23.970588235294116, | |
| "grad_norm": 0.12435536831617355, | |
| "learning_rate": 4.029411764705883e-06, | |
| "loss": 0.0152, | |
| "step": 1630 | |
| }, | |
| { | |
| "epoch": 24.0, | |
| "eval_loss": 0.14459367096424103, | |
| "eval_runtime": 1.6409, | |
| "eval_samples_per_second": 82.271, | |
| "eval_steps_per_second": 10.36, | |
| "step": 1632 | |
| }, | |
| { | |
| "epoch": 24.11764705882353, | |
| "grad_norm": 0.15185730159282684, | |
| "learning_rate": 3.931372549019608e-06, | |
| "loss": 0.0146, | |
| "step": 1640 | |
| }, | |
| { | |
| "epoch": 24.264705882352942, | |
| "grad_norm": 0.14219020307064056, | |
| "learning_rate": 3.833333333333334e-06, | |
| "loss": 0.0141, | |
| "step": 1650 | |
| }, | |
| { | |
| "epoch": 24.41176470588235, | |
| "grad_norm": 0.1183333620429039, | |
| "learning_rate": 3.7352941176470593e-06, | |
| "loss": 0.0142, | |
| "step": 1660 | |
| }, | |
| { | |
| "epoch": 24.558823529411764, | |
| "grad_norm": 0.11505335569381714, | |
| "learning_rate": 3.6372549019607845e-06, | |
| "loss": 0.0601, | |
| "step": 1670 | |
| }, | |
| { | |
| "epoch": 24.705882352941178, | |
| "grad_norm": 0.1279689520597458, | |
| "learning_rate": 3.53921568627451e-06, | |
| "loss": 0.0149, | |
| "step": 1680 | |
| }, | |
| { | |
| "epoch": 24.852941176470587, | |
| "grad_norm": 0.09878543764352798, | |
| "learning_rate": 3.4411764705882358e-06, | |
| "loss": 0.0145, | |
| "step": 1690 | |
| }, | |
| { | |
| "epoch": 25.0, | |
| "grad_norm": 0.2872982919216156, | |
| "learning_rate": 3.343137254901961e-06, | |
| "loss": 0.0142, | |
| "step": 1700 | |
| }, | |
| { | |
| "epoch": 25.0, | |
| "eval_loss": 0.1434563398361206, | |
| "eval_runtime": 1.5906, | |
| "eval_samples_per_second": 84.875, | |
| "eval_steps_per_second": 10.688, | |
| "step": 1700 | |
| }, | |
| { | |
| "epoch": 25.147058823529413, | |
| "grad_norm": 13.370408058166504, | |
| "learning_rate": 3.2450980392156866e-06, | |
| "loss": 0.0515, | |
| "step": 1710 | |
| }, | |
| { | |
| "epoch": 25.294117647058822, | |
| "grad_norm": 0.11658412218093872, | |
| "learning_rate": 3.147058823529412e-06, | |
| "loss": 0.0136, | |
| "step": 1720 | |
| }, | |
| { | |
| "epoch": 25.441176470588236, | |
| "grad_norm": 0.082937091588974, | |
| "learning_rate": 3.049019607843138e-06, | |
| "loss": 0.0145, | |
| "step": 1730 | |
| }, | |
| { | |
| "epoch": 25.58823529411765, | |
| "grad_norm": 0.09096076339483261, | |
| "learning_rate": 2.950980392156863e-06, | |
| "loss": 0.0132, | |
| "step": 1740 | |
| }, | |
| { | |
| "epoch": 25.735294117647058, | |
| "grad_norm": 0.12208091467618942, | |
| "learning_rate": 2.8529411764705883e-06, | |
| "loss": 0.0144, | |
| "step": 1750 | |
| }, | |
| { | |
| "epoch": 25.88235294117647, | |
| "grad_norm": 0.11157109588384628, | |
| "learning_rate": 2.754901960784314e-06, | |
| "loss": 0.0134, | |
| "step": 1760 | |
| }, | |
| { | |
| "epoch": 26.0, | |
| "eval_loss": 0.14469991624355316, | |
| "eval_runtime": 1.6847, | |
| "eval_samples_per_second": 80.133, | |
| "eval_steps_per_second": 10.091, | |
| "step": 1768 | |
| }, | |
| { | |
| "epoch": 26.029411764705884, | |
| "grad_norm": 0.13175475597381592, | |
| "learning_rate": 2.6568627450980396e-06, | |
| "loss": 0.0139, | |
| "step": 1770 | |
| }, | |
| { | |
| "epoch": 26.176470588235293, | |
| "grad_norm": 0.12328551709651947, | |
| "learning_rate": 2.558823529411765e-06, | |
| "loss": 0.0127, | |
| "step": 1780 | |
| }, | |
| { | |
| "epoch": 26.323529411764707, | |
| "grad_norm": 0.08687178790569305, | |
| "learning_rate": 2.4607843137254905e-06, | |
| "loss": 0.0731, | |
| "step": 1790 | |
| }, | |
| { | |
| "epoch": 26.470588235294116, | |
| "grad_norm": 0.11597796529531479, | |
| "learning_rate": 2.3627450980392157e-06, | |
| "loss": 0.0135, | |
| "step": 1800 | |
| }, | |
| { | |
| "epoch": 26.61764705882353, | |
| "grad_norm": 0.09727944433689117, | |
| "learning_rate": 2.2647058823529413e-06, | |
| "loss": 0.014, | |
| "step": 1810 | |
| }, | |
| { | |
| "epoch": 26.764705882352942, | |
| "grad_norm": 0.1099948063492775, | |
| "learning_rate": 2.166666666666667e-06, | |
| "loss": 0.0139, | |
| "step": 1820 | |
| }, | |
| { | |
| "epoch": 26.91176470588235, | |
| "grad_norm": 0.1616813987493515, | |
| "learning_rate": 2.068627450980392e-06, | |
| "loss": 0.0137, | |
| "step": 1830 | |
| }, | |
| { | |
| "epoch": 27.0, | |
| "eval_loss": 0.15348146855831146, | |
| "eval_runtime": 1.6268, | |
| "eval_samples_per_second": 82.982, | |
| "eval_steps_per_second": 10.45, | |
| "step": 1836 | |
| }, | |
| { | |
| "epoch": 27.058823529411764, | |
| "grad_norm": 0.09702517092227936, | |
| "learning_rate": 1.970588235294118e-06, | |
| "loss": 0.013, | |
| "step": 1840 | |
| }, | |
| { | |
| "epoch": 27.205882352941178, | |
| "grad_norm": 0.4694797992706299, | |
| "learning_rate": 1.8725490196078432e-06, | |
| "loss": 0.014, | |
| "step": 1850 | |
| }, | |
| { | |
| "epoch": 27.352941176470587, | |
| "grad_norm": 0.11010562628507614, | |
| "learning_rate": 1.7745098039215689e-06, | |
| "loss": 0.0126, | |
| "step": 1860 | |
| }, | |
| { | |
| "epoch": 27.5, | |
| "grad_norm": 0.11507220566272736, | |
| "learning_rate": 1.676470588235294e-06, | |
| "loss": 0.0134, | |
| "step": 1870 | |
| }, | |
| { | |
| "epoch": 27.647058823529413, | |
| "grad_norm": 0.10667148232460022, | |
| "learning_rate": 1.5784313725490197e-06, | |
| "loss": 0.0703, | |
| "step": 1880 | |
| }, | |
| { | |
| "epoch": 27.794117647058822, | |
| "grad_norm": 0.09035525470972061, | |
| "learning_rate": 1.4803921568627452e-06, | |
| "loss": 0.0127, | |
| "step": 1890 | |
| }, | |
| { | |
| "epoch": 27.941176470588236, | |
| "grad_norm": 0.13801293075084686, | |
| "learning_rate": 1.3823529411764708e-06, | |
| "loss": 0.0127, | |
| "step": 1900 | |
| }, | |
| { | |
| "epoch": 28.0, | |
| "eval_loss": 0.151557058095932, | |
| "eval_runtime": 1.6327, | |
| "eval_samples_per_second": 82.687, | |
| "eval_steps_per_second": 10.412, | |
| "step": 1904 | |
| }, | |
| { | |
| "epoch": 28.08823529411765, | |
| "grad_norm": 0.12456930428743362, | |
| "learning_rate": 1.284313725490196e-06, | |
| "loss": 0.0121, | |
| "step": 1910 | |
| }, | |
| { | |
| "epoch": 28.235294117647058, | |
| "grad_norm": 0.1714206337928772, | |
| "learning_rate": 1.1862745098039217e-06, | |
| "loss": 0.0143, | |
| "step": 1920 | |
| }, | |
| { | |
| "epoch": 28.38235294117647, | |
| "grad_norm": 0.09358106553554535, | |
| "learning_rate": 1.088235294117647e-06, | |
| "loss": 0.0124, | |
| "step": 1930 | |
| }, | |
| { | |
| "epoch": 28.529411764705884, | |
| "grad_norm": 0.12065358459949493, | |
| "learning_rate": 9.901960784313725e-07, | |
| "loss": 0.013, | |
| "step": 1940 | |
| }, | |
| { | |
| "epoch": 28.676470588235293, | |
| "grad_norm": 0.10697101801633835, | |
| "learning_rate": 8.921568627450982e-07, | |
| "loss": 0.0128, | |
| "step": 1950 | |
| }, | |
| { | |
| "epoch": 28.823529411764707, | |
| "grad_norm": 0.11727521568536758, | |
| "learning_rate": 7.941176470588236e-07, | |
| "loss": 0.0121, | |
| "step": 1960 | |
| }, | |
| { | |
| "epoch": 28.970588235294116, | |
| "grad_norm": 0.10474637895822525, | |
| "learning_rate": 6.960784313725491e-07, | |
| "loss": 0.0548, | |
| "step": 1970 | |
| }, | |
| { | |
| "epoch": 29.0, | |
| "eval_loss": 0.15194356441497803, | |
| "eval_runtime": 1.6187, | |
| "eval_samples_per_second": 83.399, | |
| "eval_steps_per_second": 10.502, | |
| "step": 1972 | |
| }, | |
| { | |
| "epoch": 29.11764705882353, | |
| "grad_norm": 0.11856939643621445, | |
| "learning_rate": 5.980392156862745e-07, | |
| "loss": 0.0121, | |
| "step": 1980 | |
| }, | |
| { | |
| "epoch": 29.264705882352942, | |
| "grad_norm": 0.12045503407716751, | |
| "learning_rate": 5.000000000000001e-07, | |
| "loss": 0.0128, | |
| "step": 1990 | |
| }, | |
| { | |
| "epoch": 29.41176470588235, | |
| "grad_norm": 0.11332525312900543, | |
| "learning_rate": 4.0196078431372556e-07, | |
| "loss": 0.0119, | |
| "step": 2000 | |
| } | |
| ], | |
| "logging_steps": 10, | |
| "max_steps": 2040, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 30, | |
| "save_steps": 500, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": false | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 4156813161449472.0, | |
| "train_batch_size": 8, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |