| { | |
| "best_metric": 0.00020612729713320732, | |
| "best_model_checkpoint": "./ap_train_outputs_tobikoi/checkpoint-7830", | |
| "epoch": 150.0, | |
| "eval_steps": 500, | |
| "global_step": 8100, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 1.997530864197531e-05, | |
| "loss": 0.657, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 1.995061728395062e-05, | |
| "loss": 0.5358, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "learning_rate": 1.9925925925925928e-05, | |
| "loss": 0.5326, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "learning_rate": 1.9901234567901237e-05, | |
| "loss": 0.5757, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "learning_rate": 1.9876543209876546e-05, | |
| "loss": 0.4014, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "eval_accuracy": 0.7631578947368421, | |
| "eval_loss": 0.3551758825778961, | |
| "eval_runtime": 3.0112, | |
| "eval_samples_per_second": 25.239, | |
| "eval_steps_per_second": 3.321, | |
| "step": 54 | |
| }, | |
| { | |
| "epoch": 1.11, | |
| "learning_rate": 1.9851851851851855e-05, | |
| "loss": 0.3156, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 1.3, | |
| "learning_rate": 1.9827160493827163e-05, | |
| "loss": 0.3332, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 1.48, | |
| "learning_rate": 1.9802469135802472e-05, | |
| "loss": 0.2956, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 1.67, | |
| "learning_rate": 1.977777777777778e-05, | |
| "loss": 0.2752, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 1.85, | |
| "learning_rate": 1.9753086419753087e-05, | |
| "loss": 0.2253, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 2.0, | |
| "eval_accuracy": 0.9736842105263158, | |
| "eval_loss": 0.1712142825126648, | |
| "eval_runtime": 2.945, | |
| "eval_samples_per_second": 25.806, | |
| "eval_steps_per_second": 3.396, | |
| "step": 108 | |
| }, | |
| { | |
| "epoch": 2.04, | |
| "learning_rate": 1.9728395061728395e-05, | |
| "loss": 0.1685, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 2.22, | |
| "learning_rate": 1.9703703703703704e-05, | |
| "loss": 0.2022, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 2.41, | |
| "learning_rate": 1.9679012345679013e-05, | |
| "loss": 0.1469, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 2.59, | |
| "learning_rate": 1.9654320987654322e-05, | |
| "loss": 0.1177, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 2.78, | |
| "learning_rate": 1.962962962962963e-05, | |
| "loss": 0.081, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 2.96, | |
| "learning_rate": 1.960493827160494e-05, | |
| "loss": 0.0768, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 3.0, | |
| "eval_accuracy": 0.9868421052631579, | |
| "eval_loss": 0.07630854845046997, | |
| "eval_runtime": 3.0191, | |
| "eval_samples_per_second": 25.173, | |
| "eval_steps_per_second": 3.312, | |
| "step": 162 | |
| }, | |
| { | |
| "epoch": 3.15, | |
| "learning_rate": 1.958024691358025e-05, | |
| "loss": 0.1135, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 3.33, | |
| "learning_rate": 1.9555555555555557e-05, | |
| "loss": 0.0699, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 3.52, | |
| "learning_rate": 1.9530864197530866e-05, | |
| "loss": 0.0533, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 3.7, | |
| "learning_rate": 1.9506172839506175e-05, | |
| "loss": 0.0447, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 3.89, | |
| "learning_rate": 1.9481481481481484e-05, | |
| "loss": 0.0694, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 4.0, | |
| "eval_accuracy": 0.9868421052631579, | |
| "eval_loss": 0.061516497284173965, | |
| "eval_runtime": 2.9805, | |
| "eval_samples_per_second": 25.499, | |
| "eval_steps_per_second": 3.355, | |
| "step": 216 | |
| }, | |
| { | |
| "epoch": 4.07, | |
| "learning_rate": 1.9456790123456793e-05, | |
| "loss": 0.1095, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 4.26, | |
| "learning_rate": 1.94320987654321e-05, | |
| "loss": 0.0489, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 4.44, | |
| "learning_rate": 1.9407407407407407e-05, | |
| "loss": 0.0254, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 4.63, | |
| "learning_rate": 1.9382716049382716e-05, | |
| "loss": 0.0304, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 4.81, | |
| "learning_rate": 1.9358024691358025e-05, | |
| "loss": 0.0258, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 5.0, | |
| "learning_rate": 1.9333333333333333e-05, | |
| "loss": 0.0433, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 5.0, | |
| "eval_accuracy": 0.9868421052631579, | |
| "eval_loss": 0.05043227970600128, | |
| "eval_runtime": 3.0544, | |
| "eval_samples_per_second": 24.882, | |
| "eval_steps_per_second": 3.274, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 5.19, | |
| "learning_rate": 1.9308641975308646e-05, | |
| "loss": 0.0204, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 5.37, | |
| "learning_rate": 1.9283950617283955e-05, | |
| "loss": 0.0231, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 5.56, | |
| "learning_rate": 1.925925925925926e-05, | |
| "loss": 0.0463, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 5.74, | |
| "learning_rate": 1.923456790123457e-05, | |
| "loss": 0.0181, | |
| "step": 310 | |
| }, | |
| { | |
| "epoch": 5.93, | |
| "learning_rate": 1.9209876543209878e-05, | |
| "loss": 0.1045, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 6.0, | |
| "eval_accuracy": 0.9868421052631579, | |
| "eval_loss": 0.03228908032178879, | |
| "eval_runtime": 2.9647, | |
| "eval_samples_per_second": 25.635, | |
| "eval_steps_per_second": 3.373, | |
| "step": 324 | |
| }, | |
| { | |
| "epoch": 6.11, | |
| "learning_rate": 1.9185185185185186e-05, | |
| "loss": 0.0568, | |
| "step": 330 | |
| }, | |
| { | |
| "epoch": 6.3, | |
| "learning_rate": 1.9160493827160495e-05, | |
| "loss": 0.0373, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 6.48, | |
| "learning_rate": 1.9135802469135804e-05, | |
| "loss": 0.0157, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 6.67, | |
| "learning_rate": 1.9111111111111113e-05, | |
| "loss": 0.023, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 6.85, | |
| "learning_rate": 1.9086419753086422e-05, | |
| "loss": 0.0148, | |
| "step": 370 | |
| }, | |
| { | |
| "epoch": 7.0, | |
| "eval_accuracy": 0.9868421052631579, | |
| "eval_loss": 0.04355274513363838, | |
| "eval_runtime": 3.0476, | |
| "eval_samples_per_second": 24.938, | |
| "eval_steps_per_second": 3.281, | |
| "step": 378 | |
| }, | |
| { | |
| "epoch": 7.04, | |
| "learning_rate": 1.906172839506173e-05, | |
| "loss": 0.0364, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 7.22, | |
| "learning_rate": 1.903703703703704e-05, | |
| "loss": 0.0515, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 7.41, | |
| "learning_rate": 1.901234567901235e-05, | |
| "loss": 0.0271, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 7.59, | |
| "learning_rate": 1.8987654320987657e-05, | |
| "loss": 0.0134, | |
| "step": 410 | |
| }, | |
| { | |
| "epoch": 7.78, | |
| "learning_rate": 1.8962962962962966e-05, | |
| "loss": 0.0111, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 7.96, | |
| "learning_rate": 1.8938271604938275e-05, | |
| "loss": 0.0156, | |
| "step": 430 | |
| }, | |
| { | |
| "epoch": 8.0, | |
| "eval_accuracy": 0.9868421052631579, | |
| "eval_loss": 0.027104511857032776, | |
| "eval_runtime": 3.0509, | |
| "eval_samples_per_second": 24.911, | |
| "eval_steps_per_second": 3.278, | |
| "step": 432 | |
| }, | |
| { | |
| "epoch": 8.15, | |
| "learning_rate": 1.891358024691358e-05, | |
| "loss": 0.0141, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 8.33, | |
| "learning_rate": 1.888888888888889e-05, | |
| "loss": 0.0118, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 8.52, | |
| "learning_rate": 1.8864197530864198e-05, | |
| "loss": 0.0109, | |
| "step": 460 | |
| }, | |
| { | |
| "epoch": 8.7, | |
| "learning_rate": 1.8839506172839507e-05, | |
| "loss": 0.009, | |
| "step": 470 | |
| }, | |
| { | |
| "epoch": 8.89, | |
| "learning_rate": 1.8814814814814816e-05, | |
| "loss": 0.0109, | |
| "step": 480 | |
| }, | |
| { | |
| "epoch": 9.0, | |
| "eval_accuracy": 0.9868421052631579, | |
| "eval_loss": 0.05106474086642265, | |
| "eval_runtime": 3.0435, | |
| "eval_samples_per_second": 24.971, | |
| "eval_steps_per_second": 3.286, | |
| "step": 486 | |
| }, | |
| { | |
| "epoch": 9.07, | |
| "learning_rate": 1.8790123456790124e-05, | |
| "loss": 0.0117, | |
| "step": 490 | |
| }, | |
| { | |
| "epoch": 9.26, | |
| "learning_rate": 1.8765432098765433e-05, | |
| "loss": 0.0244, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 9.44, | |
| "learning_rate": 1.8740740740740742e-05, | |
| "loss": 0.1521, | |
| "step": 510 | |
| }, | |
| { | |
| "epoch": 9.63, | |
| "learning_rate": 1.871604938271605e-05, | |
| "loss": 0.0103, | |
| "step": 520 | |
| }, | |
| { | |
| "epoch": 9.81, | |
| "learning_rate": 1.869135802469136e-05, | |
| "loss": 0.0103, | |
| "step": 530 | |
| }, | |
| { | |
| "epoch": 10.0, | |
| "learning_rate": 1.866666666666667e-05, | |
| "loss": 0.0142, | |
| "step": 540 | |
| }, | |
| { | |
| "epoch": 10.0, | |
| "eval_accuracy": 0.9868421052631579, | |
| "eval_loss": 0.05626223608851433, | |
| "eval_runtime": 2.9574, | |
| "eval_samples_per_second": 25.698, | |
| "eval_steps_per_second": 3.381, | |
| "step": 540 | |
| }, | |
| { | |
| "epoch": 10.19, | |
| "learning_rate": 1.8641975308641977e-05, | |
| "loss": 0.008, | |
| "step": 550 | |
| }, | |
| { | |
| "epoch": 10.37, | |
| "learning_rate": 1.8617283950617286e-05, | |
| "loss": 0.0198, | |
| "step": 560 | |
| }, | |
| { | |
| "epoch": 10.56, | |
| "learning_rate": 1.8592592592592592e-05, | |
| "loss": 0.0093, | |
| "step": 570 | |
| }, | |
| { | |
| "epoch": 10.74, | |
| "learning_rate": 1.85679012345679e-05, | |
| "loss": 0.0084, | |
| "step": 580 | |
| }, | |
| { | |
| "epoch": 10.93, | |
| "learning_rate": 1.8543209876543213e-05, | |
| "loss": 0.0307, | |
| "step": 590 | |
| }, | |
| { | |
| "epoch": 11.0, | |
| "eval_accuracy": 0.9868421052631579, | |
| "eval_loss": 0.06329117715358734, | |
| "eval_runtime": 2.9789, | |
| "eval_samples_per_second": 25.513, | |
| "eval_steps_per_second": 3.357, | |
| "step": 594 | |
| }, | |
| { | |
| "epoch": 11.11, | |
| "learning_rate": 1.851851851851852e-05, | |
| "loss": 0.0108, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 11.3, | |
| "learning_rate": 1.849382716049383e-05, | |
| "loss": 0.0074, | |
| "step": 610 | |
| }, | |
| { | |
| "epoch": 11.48, | |
| "learning_rate": 1.846913580246914e-05, | |
| "loss": 0.0078, | |
| "step": 620 | |
| }, | |
| { | |
| "epoch": 11.67, | |
| "learning_rate": 1.8444444444444448e-05, | |
| "loss": 0.0068, | |
| "step": 630 | |
| }, | |
| { | |
| "epoch": 11.85, | |
| "learning_rate": 1.8419753086419754e-05, | |
| "loss": 0.0092, | |
| "step": 640 | |
| }, | |
| { | |
| "epoch": 12.0, | |
| "eval_accuracy": 0.9868421052631579, | |
| "eval_loss": 0.04296250268816948, | |
| "eval_runtime": 3.0671, | |
| "eval_samples_per_second": 24.779, | |
| "eval_steps_per_second": 3.26, | |
| "step": 648 | |
| }, | |
| { | |
| "epoch": 12.04, | |
| "learning_rate": 1.8395061728395062e-05, | |
| "loss": 0.092, | |
| "step": 650 | |
| }, | |
| { | |
| "epoch": 12.22, | |
| "learning_rate": 1.837037037037037e-05, | |
| "loss": 0.0068, | |
| "step": 660 | |
| }, | |
| { | |
| "epoch": 12.41, | |
| "learning_rate": 1.834567901234568e-05, | |
| "loss": 0.0508, | |
| "step": 670 | |
| }, | |
| { | |
| "epoch": 12.59, | |
| "learning_rate": 1.832098765432099e-05, | |
| "loss": 0.0167, | |
| "step": 680 | |
| }, | |
| { | |
| "epoch": 12.78, | |
| "learning_rate": 1.8296296296296298e-05, | |
| "loss": 0.0064, | |
| "step": 690 | |
| }, | |
| { | |
| "epoch": 12.96, | |
| "learning_rate": 1.8271604938271607e-05, | |
| "loss": 0.007, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 13.0, | |
| "eval_accuracy": 0.9868421052631579, | |
| "eval_loss": 0.050767723470926285, | |
| "eval_runtime": 2.9682, | |
| "eval_samples_per_second": 25.605, | |
| "eval_steps_per_second": 3.369, | |
| "step": 702 | |
| }, | |
| { | |
| "epoch": 13.15, | |
| "learning_rate": 1.8246913580246915e-05, | |
| "loss": 0.0063, | |
| "step": 710 | |
| }, | |
| { | |
| "epoch": 13.33, | |
| "learning_rate": 1.8222222222222224e-05, | |
| "loss": 0.006, | |
| "step": 720 | |
| }, | |
| { | |
| "epoch": 13.52, | |
| "learning_rate": 1.8197530864197533e-05, | |
| "loss": 0.0082, | |
| "step": 730 | |
| }, | |
| { | |
| "epoch": 13.7, | |
| "learning_rate": 1.8172839506172842e-05, | |
| "loss": 0.0056, | |
| "step": 740 | |
| }, | |
| { | |
| "epoch": 13.89, | |
| "learning_rate": 1.814814814814815e-05, | |
| "loss": 0.0059, | |
| "step": 750 | |
| }, | |
| { | |
| "epoch": 14.0, | |
| "eval_accuracy": 0.9868421052631579, | |
| "eval_loss": 0.05981181189417839, | |
| "eval_runtime": 2.9705, | |
| "eval_samples_per_second": 25.585, | |
| "eval_steps_per_second": 3.366, | |
| "step": 756 | |
| }, | |
| { | |
| "epoch": 14.07, | |
| "learning_rate": 1.812345679012346e-05, | |
| "loss": 0.0089, | |
| "step": 760 | |
| }, | |
| { | |
| "epoch": 14.26, | |
| "learning_rate": 1.8098765432098765e-05, | |
| "loss": 0.0058, | |
| "step": 770 | |
| }, | |
| { | |
| "epoch": 14.44, | |
| "learning_rate": 1.8074074074074074e-05, | |
| "loss": 0.0059, | |
| "step": 780 | |
| }, | |
| { | |
| "epoch": 14.63, | |
| "learning_rate": 1.8049382716049383e-05, | |
| "loss": 0.0434, | |
| "step": 790 | |
| }, | |
| { | |
| "epoch": 14.81, | |
| "learning_rate": 1.802469135802469e-05, | |
| "loss": 0.0054, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 15.0, | |
| "learning_rate": 1.8e-05, | |
| "loss": 0.0057, | |
| "step": 810 | |
| }, | |
| { | |
| "epoch": 15.0, | |
| "eval_accuracy": 0.9868421052631579, | |
| "eval_loss": 0.06393761932849884, | |
| "eval_runtime": 2.9181, | |
| "eval_samples_per_second": 26.044, | |
| "eval_steps_per_second": 3.427, | |
| "step": 810 | |
| }, | |
| { | |
| "epoch": 15.19, | |
| "learning_rate": 1.797530864197531e-05, | |
| "loss": 0.0052, | |
| "step": 820 | |
| }, | |
| { | |
| "epoch": 15.37, | |
| "learning_rate": 1.7950617283950618e-05, | |
| "loss": 0.0051, | |
| "step": 830 | |
| }, | |
| { | |
| "epoch": 15.56, | |
| "learning_rate": 1.7925925925925927e-05, | |
| "loss": 0.0249, | |
| "step": 840 | |
| }, | |
| { | |
| "epoch": 15.74, | |
| "learning_rate": 1.7901234567901236e-05, | |
| "loss": 0.0274, | |
| "step": 850 | |
| }, | |
| { | |
| "epoch": 15.93, | |
| "learning_rate": 1.7876543209876545e-05, | |
| "loss": 0.0513, | |
| "step": 860 | |
| }, | |
| { | |
| "epoch": 16.0, | |
| "eval_accuracy": 0.9868421052631579, | |
| "eval_loss": 0.05791584402322769, | |
| "eval_runtime": 2.9732, | |
| "eval_samples_per_second": 25.561, | |
| "eval_steps_per_second": 3.363, | |
| "step": 864 | |
| }, | |
| { | |
| "epoch": 16.11, | |
| "learning_rate": 1.7851851851851853e-05, | |
| "loss": 0.0049, | |
| "step": 870 | |
| }, | |
| { | |
| "epoch": 16.3, | |
| "learning_rate": 1.7827160493827162e-05, | |
| "loss": 0.0048, | |
| "step": 880 | |
| }, | |
| { | |
| "epoch": 16.48, | |
| "learning_rate": 1.780246913580247e-05, | |
| "loss": 0.0133, | |
| "step": 890 | |
| }, | |
| { | |
| "epoch": 16.67, | |
| "learning_rate": 1.7777777777777777e-05, | |
| "loss": 0.0101, | |
| "step": 900 | |
| }, | |
| { | |
| "epoch": 16.85, | |
| "learning_rate": 1.775308641975309e-05, | |
| "loss": 0.0259, | |
| "step": 910 | |
| }, | |
| { | |
| "epoch": 17.0, | |
| "eval_accuracy": 0.9868421052631579, | |
| "eval_loss": 0.07065781950950623, | |
| "eval_runtime": 2.9923, | |
| "eval_samples_per_second": 25.399, | |
| "eval_steps_per_second": 3.342, | |
| "step": 918 | |
| }, | |
| { | |
| "epoch": 17.04, | |
| "learning_rate": 1.7728395061728398e-05, | |
| "loss": 0.0122, | |
| "step": 920 | |
| }, | |
| { | |
| "epoch": 17.22, | |
| "learning_rate": 1.7703703703703706e-05, | |
| "loss": 0.0056, | |
| "step": 930 | |
| }, | |
| { | |
| "epoch": 17.41, | |
| "learning_rate": 1.7679012345679015e-05, | |
| "loss": 0.037, | |
| "step": 940 | |
| }, | |
| { | |
| "epoch": 17.59, | |
| "learning_rate": 1.7654320987654324e-05, | |
| "loss": 0.0046, | |
| "step": 950 | |
| }, | |
| { | |
| "epoch": 17.78, | |
| "learning_rate": 1.7629629629629633e-05, | |
| "loss": 0.0041, | |
| "step": 960 | |
| }, | |
| { | |
| "epoch": 17.96, | |
| "learning_rate": 1.760493827160494e-05, | |
| "loss": 0.0111, | |
| "step": 970 | |
| }, | |
| { | |
| "epoch": 18.0, | |
| "eval_accuracy": 0.9868421052631579, | |
| "eval_loss": 0.061096400022506714, | |
| "eval_runtime": 2.9964, | |
| "eval_samples_per_second": 25.364, | |
| "eval_steps_per_second": 3.337, | |
| "step": 972 | |
| }, | |
| { | |
| "epoch": 18.15, | |
| "learning_rate": 1.7580246913580247e-05, | |
| "loss": 0.0044, | |
| "step": 980 | |
| }, | |
| { | |
| "epoch": 18.33, | |
| "learning_rate": 1.7555555555555556e-05, | |
| "loss": 0.0612, | |
| "step": 990 | |
| }, | |
| { | |
| "epoch": 18.52, | |
| "learning_rate": 1.7530864197530865e-05, | |
| "loss": 0.0389, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 18.7, | |
| "learning_rate": 1.7506172839506174e-05, | |
| "loss": 0.0707, | |
| "step": 1010 | |
| }, | |
| { | |
| "epoch": 18.89, | |
| "learning_rate": 1.7481481481481483e-05, | |
| "loss": 0.014, | |
| "step": 1020 | |
| }, | |
| { | |
| "epoch": 19.0, | |
| "eval_accuracy": 0.9868421052631579, | |
| "eval_loss": 0.06202290579676628, | |
| "eval_runtime": 3.0244, | |
| "eval_samples_per_second": 25.129, | |
| "eval_steps_per_second": 3.306, | |
| "step": 1026 | |
| }, | |
| { | |
| "epoch": 19.07, | |
| "learning_rate": 1.745679012345679e-05, | |
| "loss": 0.0042, | |
| "step": 1030 | |
| }, | |
| { | |
| "epoch": 19.26, | |
| "learning_rate": 1.74320987654321e-05, | |
| "loss": 0.0052, | |
| "step": 1040 | |
| }, | |
| { | |
| "epoch": 19.44, | |
| "learning_rate": 1.740740740740741e-05, | |
| "loss": 0.0505, | |
| "step": 1050 | |
| }, | |
| { | |
| "epoch": 19.63, | |
| "learning_rate": 1.7382716049382718e-05, | |
| "loss": 0.0042, | |
| "step": 1060 | |
| }, | |
| { | |
| "epoch": 19.81, | |
| "learning_rate": 1.7358024691358027e-05, | |
| "loss": 0.0038, | |
| "step": 1070 | |
| }, | |
| { | |
| "epoch": 20.0, | |
| "learning_rate": 1.7333333333333336e-05, | |
| "loss": 0.004, | |
| "step": 1080 | |
| }, | |
| { | |
| "epoch": 20.0, | |
| "eval_accuracy": 1.0, | |
| "eval_loss": 0.005762842949479818, | |
| "eval_runtime": 3.075, | |
| "eval_samples_per_second": 24.715, | |
| "eval_steps_per_second": 3.252, | |
| "step": 1080 | |
| }, | |
| { | |
| "epoch": 20.19, | |
| "learning_rate": 1.7308641975308644e-05, | |
| "loss": 0.0039, | |
| "step": 1090 | |
| }, | |
| { | |
| "epoch": 20.37, | |
| "learning_rate": 1.728395061728395e-05, | |
| "loss": 0.0037, | |
| "step": 1100 | |
| }, | |
| { | |
| "epoch": 20.56, | |
| "learning_rate": 1.725925925925926e-05, | |
| "loss": 0.0037, | |
| "step": 1110 | |
| }, | |
| { | |
| "epoch": 20.74, | |
| "learning_rate": 1.7234567901234568e-05, | |
| "loss": 0.0054, | |
| "step": 1120 | |
| }, | |
| { | |
| "epoch": 20.93, | |
| "learning_rate": 1.7209876543209876e-05, | |
| "loss": 0.0036, | |
| "step": 1130 | |
| }, | |
| { | |
| "epoch": 21.0, | |
| "eval_accuracy": 1.0, | |
| "eval_loss": 0.004358741920441389, | |
| "eval_runtime": 3.0437, | |
| "eval_samples_per_second": 24.97, | |
| "eval_steps_per_second": 3.286, | |
| "step": 1134 | |
| }, | |
| { | |
| "epoch": 21.11, | |
| "learning_rate": 1.7185185185185185e-05, | |
| "loss": 0.0039, | |
| "step": 1140 | |
| }, | |
| { | |
| "epoch": 21.3, | |
| "learning_rate": 1.7160493827160498e-05, | |
| "loss": 0.0418, | |
| "step": 1150 | |
| }, | |
| { | |
| "epoch": 21.48, | |
| "learning_rate": 1.7135802469135806e-05, | |
| "loss": 0.0104, | |
| "step": 1160 | |
| }, | |
| { | |
| "epoch": 21.67, | |
| "learning_rate": 1.7111111111111112e-05, | |
| "loss": 0.1184, | |
| "step": 1170 | |
| }, | |
| { | |
| "epoch": 21.85, | |
| "learning_rate": 1.708641975308642e-05, | |
| "loss": 0.0545, | |
| "step": 1180 | |
| }, | |
| { | |
| "epoch": 22.0, | |
| "eval_accuracy": 1.0, | |
| "eval_loss": 0.011370152235031128, | |
| "eval_runtime": 2.9933, | |
| "eval_samples_per_second": 25.39, | |
| "eval_steps_per_second": 3.341, | |
| "step": 1188 | |
| }, | |
| { | |
| "epoch": 22.04, | |
| "learning_rate": 1.706172839506173e-05, | |
| "loss": 0.0295, | |
| "step": 1190 | |
| }, | |
| { | |
| "epoch": 22.22, | |
| "learning_rate": 1.7037037037037038e-05, | |
| "loss": 0.0633, | |
| "step": 1200 | |
| }, | |
| { | |
| "epoch": 22.41, | |
| "learning_rate": 1.7012345679012347e-05, | |
| "loss": 0.007, | |
| "step": 1210 | |
| }, | |
| { | |
| "epoch": 22.59, | |
| "learning_rate": 1.6987654320987656e-05, | |
| "loss": 0.0092, | |
| "step": 1220 | |
| }, | |
| { | |
| "epoch": 22.78, | |
| "learning_rate": 1.6962962962962965e-05, | |
| "loss": 0.0176, | |
| "step": 1230 | |
| }, | |
| { | |
| "epoch": 22.96, | |
| "learning_rate": 1.6938271604938274e-05, | |
| "loss": 0.0131, | |
| "step": 1240 | |
| }, | |
| { | |
| "epoch": 23.0, | |
| "eval_accuracy": 0.9868421052631579, | |
| "eval_loss": 0.06211669743061066, | |
| "eval_runtime": 3.0535, | |
| "eval_samples_per_second": 24.889, | |
| "eval_steps_per_second": 3.275, | |
| "step": 1242 | |
| }, | |
| { | |
| "epoch": 23.15, | |
| "learning_rate": 1.6913580246913582e-05, | |
| "loss": 0.4038, | |
| "step": 1250 | |
| }, | |
| { | |
| "epoch": 23.33, | |
| "learning_rate": 1.688888888888889e-05, | |
| "loss": 0.0658, | |
| "step": 1260 | |
| }, | |
| { | |
| "epoch": 23.52, | |
| "learning_rate": 1.68641975308642e-05, | |
| "loss": 0.0316, | |
| "step": 1270 | |
| }, | |
| { | |
| "epoch": 23.7, | |
| "learning_rate": 1.683950617283951e-05, | |
| "loss": 0.0705, | |
| "step": 1280 | |
| }, | |
| { | |
| "epoch": 23.89, | |
| "learning_rate": 1.6814814814814818e-05, | |
| "loss": 0.0651, | |
| "step": 1290 | |
| }, | |
| { | |
| "epoch": 24.0, | |
| "eval_accuracy": 0.9868421052631579, | |
| "eval_loss": 0.06920474767684937, | |
| "eval_runtime": 3.011, | |
| "eval_samples_per_second": 25.241, | |
| "eval_steps_per_second": 3.321, | |
| "step": 1296 | |
| }, | |
| { | |
| "epoch": 24.07, | |
| "learning_rate": 1.6790123456790123e-05, | |
| "loss": 0.0035, | |
| "step": 1300 | |
| }, | |
| { | |
| "epoch": 24.26, | |
| "learning_rate": 1.6765432098765432e-05, | |
| "loss": 0.0147, | |
| "step": 1310 | |
| }, | |
| { | |
| "epoch": 24.44, | |
| "learning_rate": 1.674074074074074e-05, | |
| "loss": 0.0045, | |
| "step": 1320 | |
| }, | |
| { | |
| "epoch": 24.63, | |
| "learning_rate": 1.671604938271605e-05, | |
| "loss": 0.0375, | |
| "step": 1330 | |
| }, | |
| { | |
| "epoch": 24.81, | |
| "learning_rate": 1.669135802469136e-05, | |
| "loss": 0.0278, | |
| "step": 1340 | |
| }, | |
| { | |
| "epoch": 25.0, | |
| "learning_rate": 1.6666666666666667e-05, | |
| "loss": 0.0047, | |
| "step": 1350 | |
| }, | |
| { | |
| "epoch": 25.0, | |
| "eval_accuracy": 1.0, | |
| "eval_loss": 0.003432216588407755, | |
| "eval_runtime": 2.9828, | |
| "eval_samples_per_second": 25.48, | |
| "eval_steps_per_second": 3.353, | |
| "step": 1350 | |
| }, | |
| { | |
| "epoch": 25.19, | |
| "learning_rate": 1.6641975308641976e-05, | |
| "loss": 0.071, | |
| "step": 1360 | |
| }, | |
| { | |
| "epoch": 25.37, | |
| "learning_rate": 1.6617283950617285e-05, | |
| "loss": 0.003, | |
| "step": 1370 | |
| }, | |
| { | |
| "epoch": 25.56, | |
| "learning_rate": 1.6592592592592594e-05, | |
| "loss": 0.0045, | |
| "step": 1380 | |
| }, | |
| { | |
| "epoch": 25.74, | |
| "learning_rate": 1.6567901234567903e-05, | |
| "loss": 0.0033, | |
| "step": 1390 | |
| }, | |
| { | |
| "epoch": 25.93, | |
| "learning_rate": 1.654320987654321e-05, | |
| "loss": 0.0374, | |
| "step": 1400 | |
| }, | |
| { | |
| "epoch": 26.0, | |
| "eval_accuracy": 1.0, | |
| "eval_loss": 0.003061347408220172, | |
| "eval_runtime": 2.9183, | |
| "eval_samples_per_second": 26.042, | |
| "eval_steps_per_second": 3.427, | |
| "step": 1404 | |
| }, | |
| { | |
| "epoch": 26.11, | |
| "learning_rate": 1.651851851851852e-05, | |
| "loss": 0.0077, | |
| "step": 1410 | |
| }, | |
| { | |
| "epoch": 26.3, | |
| "learning_rate": 1.649382716049383e-05, | |
| "loss": 0.0047, | |
| "step": 1420 | |
| }, | |
| { | |
| "epoch": 26.48, | |
| "learning_rate": 1.6469135802469135e-05, | |
| "loss": 0.0088, | |
| "step": 1430 | |
| }, | |
| { | |
| "epoch": 26.67, | |
| "learning_rate": 1.6444444444444444e-05, | |
| "loss": 0.0043, | |
| "step": 1440 | |
| }, | |
| { | |
| "epoch": 26.85, | |
| "learning_rate": 1.6419753086419752e-05, | |
| "loss": 0.0482, | |
| "step": 1450 | |
| }, | |
| { | |
| "epoch": 27.0, | |
| "eval_accuracy": 1.0, | |
| "eval_loss": 0.004511620849370956, | |
| "eval_runtime": 2.9675, | |
| "eval_samples_per_second": 25.611, | |
| "eval_steps_per_second": 3.37, | |
| "step": 1458 | |
| }, | |
| { | |
| "epoch": 27.04, | |
| "learning_rate": 1.6395061728395065e-05, | |
| "loss": 0.0029, | |
| "step": 1460 | |
| }, | |
| { | |
| "epoch": 27.22, | |
| "learning_rate": 1.6370370370370374e-05, | |
| "loss": 0.0028, | |
| "step": 1470 | |
| }, | |
| { | |
| "epoch": 27.41, | |
| "learning_rate": 1.6345679012345682e-05, | |
| "loss": 0.0027, | |
| "step": 1480 | |
| }, | |
| { | |
| "epoch": 27.59, | |
| "learning_rate": 1.632098765432099e-05, | |
| "loss": 0.0027, | |
| "step": 1490 | |
| }, | |
| { | |
| "epoch": 27.78, | |
| "learning_rate": 1.6296296296296297e-05, | |
| "loss": 0.0028, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 27.96, | |
| "learning_rate": 1.6271604938271605e-05, | |
| "loss": 0.0026, | |
| "step": 1510 | |
| }, | |
| { | |
| "epoch": 28.0, | |
| "eval_accuracy": 1.0, | |
| "eval_loss": 0.0028049517422914505, | |
| "eval_runtime": 2.9403, | |
| "eval_samples_per_second": 25.848, | |
| "eval_steps_per_second": 3.401, | |
| "step": 1512 | |
| }, | |
| { | |
| "epoch": 28.15, | |
| "learning_rate": 1.6246913580246914e-05, | |
| "loss": 0.0025, | |
| "step": 1520 | |
| }, | |
| { | |
| "epoch": 28.33, | |
| "learning_rate": 1.6222222222222223e-05, | |
| "loss": 0.0028, | |
| "step": 1530 | |
| }, | |
| { | |
| "epoch": 28.52, | |
| "learning_rate": 1.6197530864197532e-05, | |
| "loss": 0.0025, | |
| "step": 1540 | |
| }, | |
| { | |
| "epoch": 28.7, | |
| "learning_rate": 1.617283950617284e-05, | |
| "loss": 0.0024, | |
| "step": 1550 | |
| }, | |
| { | |
| "epoch": 28.89, | |
| "learning_rate": 1.614814814814815e-05, | |
| "loss": 0.0038, | |
| "step": 1560 | |
| }, | |
| { | |
| "epoch": 29.0, | |
| "eval_accuracy": 1.0, | |
| "eval_loss": 0.0024565160274505615, | |
| "eval_runtime": 3.0445, | |
| "eval_samples_per_second": 24.963, | |
| "eval_steps_per_second": 3.285, | |
| "step": 1566 | |
| }, | |
| { | |
| "epoch": 29.07, | |
| "learning_rate": 1.612345679012346e-05, | |
| "loss": 0.0023, | |
| "step": 1570 | |
| }, | |
| { | |
| "epoch": 29.26, | |
| "learning_rate": 1.6098765432098767e-05, | |
| "loss": 0.0024, | |
| "step": 1580 | |
| }, | |
| { | |
| "epoch": 29.44, | |
| "learning_rate": 1.6074074074074076e-05, | |
| "loss": 0.0293, | |
| "step": 1590 | |
| }, | |
| { | |
| "epoch": 29.63, | |
| "learning_rate": 1.6049382716049385e-05, | |
| "loss": 0.0024, | |
| "step": 1600 | |
| }, | |
| { | |
| "epoch": 29.81, | |
| "learning_rate": 1.6024691358024694e-05, | |
| "loss": 0.0023, | |
| "step": 1610 | |
| }, | |
| { | |
| "epoch": 30.0, | |
| "learning_rate": 1.6000000000000003e-05, | |
| "loss": 0.0027, | |
| "step": 1620 | |
| }, | |
| { | |
| "epoch": 30.0, | |
| "eval_accuracy": 1.0, | |
| "eval_loss": 0.0023441906087100506, | |
| "eval_runtime": 2.9794, | |
| "eval_samples_per_second": 25.508, | |
| "eval_steps_per_second": 3.356, | |
| "step": 1620 | |
| }, | |
| { | |
| "epoch": 30.19, | |
| "learning_rate": 1.5975308641975308e-05, | |
| "loss": 0.0022, | |
| "step": 1630 | |
| }, | |
| { | |
| "epoch": 30.37, | |
| "learning_rate": 1.5950617283950617e-05, | |
| "loss": 0.0023, | |
| "step": 1640 | |
| }, | |
| { | |
| "epoch": 30.56, | |
| "learning_rate": 1.5925925925925926e-05, | |
| "loss": 0.0715, | |
| "step": 1650 | |
| }, | |
| { | |
| "epoch": 30.74, | |
| "learning_rate": 1.5901234567901235e-05, | |
| "loss": 0.0029, | |
| "step": 1660 | |
| }, | |
| { | |
| "epoch": 30.93, | |
| "learning_rate": 1.5876543209876543e-05, | |
| "loss": 0.0145, | |
| "step": 1670 | |
| }, | |
| { | |
| "epoch": 31.0, | |
| "eval_accuracy": 0.9868421052631579, | |
| "eval_loss": 0.06983212381601334, | |
| "eval_runtime": 3.0396, | |
| "eval_samples_per_second": 25.003, | |
| "eval_steps_per_second": 3.29, | |
| "step": 1674 | |
| }, | |
| { | |
| "epoch": 31.11, | |
| "learning_rate": 1.5851851851851852e-05, | |
| "loss": 0.0025, | |
| "step": 1680 | |
| }, | |
| { | |
| "epoch": 31.3, | |
| "learning_rate": 1.582716049382716e-05, | |
| "loss": 0.0025, | |
| "step": 1690 | |
| }, | |
| { | |
| "epoch": 31.48, | |
| "learning_rate": 1.580246913580247e-05, | |
| "loss": 0.0655, | |
| "step": 1700 | |
| }, | |
| { | |
| "epoch": 31.67, | |
| "learning_rate": 1.577777777777778e-05, | |
| "loss": 0.0022, | |
| "step": 1710 | |
| }, | |
| { | |
| "epoch": 31.85, | |
| "learning_rate": 1.5753086419753088e-05, | |
| "loss": 0.0022, | |
| "step": 1720 | |
| }, | |
| { | |
| "epoch": 32.0, | |
| "eval_accuracy": 0.9868421052631579, | |
| "eval_loss": 0.025487201288342476, | |
| "eval_runtime": 3.0285, | |
| "eval_samples_per_second": 25.095, | |
| "eval_steps_per_second": 3.302, | |
| "step": 1728 | |
| }, | |
| { | |
| "epoch": 32.04, | |
| "learning_rate": 1.5728395061728396e-05, | |
| "loss": 0.0932, | |
| "step": 1730 | |
| }, | |
| { | |
| "epoch": 32.22, | |
| "learning_rate": 1.5703703703703705e-05, | |
| "loss": 0.0023, | |
| "step": 1740 | |
| }, | |
| { | |
| "epoch": 32.41, | |
| "learning_rate": 1.5679012345679014e-05, | |
| "loss": 0.0404, | |
| "step": 1750 | |
| }, | |
| { | |
| "epoch": 32.59, | |
| "learning_rate": 1.565432098765432e-05, | |
| "loss": 0.0065, | |
| "step": 1760 | |
| }, | |
| { | |
| "epoch": 32.78, | |
| "learning_rate": 1.5629629629629632e-05, | |
| "loss": 0.0058, | |
| "step": 1770 | |
| }, | |
| { | |
| "epoch": 32.96, | |
| "learning_rate": 1.560493827160494e-05, | |
| "loss": 0.0025, | |
| "step": 1780 | |
| }, | |
| { | |
| "epoch": 33.0, | |
| "eval_accuracy": 1.0, | |
| "eval_loss": 0.009514715522527695, | |
| "eval_runtime": 3.0612, | |
| "eval_samples_per_second": 24.827, | |
| "eval_steps_per_second": 3.267, | |
| "step": 1782 | |
| }, | |
| { | |
| "epoch": 33.15, | |
| "learning_rate": 1.558024691358025e-05, | |
| "loss": 0.002, | |
| "step": 1790 | |
| }, | |
| { | |
| "epoch": 33.33, | |
| "learning_rate": 1.555555555555556e-05, | |
| "loss": 0.0021, | |
| "step": 1800 | |
| }, | |
| { | |
| "epoch": 33.52, | |
| "learning_rate": 1.5530864197530867e-05, | |
| "loss": 0.0824, | |
| "step": 1810 | |
| }, | |
| { | |
| "epoch": 33.7, | |
| "learning_rate": 1.5506172839506176e-05, | |
| "loss": 0.0024, | |
| "step": 1820 | |
| }, | |
| { | |
| "epoch": 33.89, | |
| "learning_rate": 1.548148148148148e-05, | |
| "loss": 0.0022, | |
| "step": 1830 | |
| }, | |
| { | |
| "epoch": 34.0, | |
| "eval_accuracy": 0.9868421052631579, | |
| "eval_loss": 0.07251408696174622, | |
| "eval_runtime": 3.1052, | |
| "eval_samples_per_second": 24.475, | |
| "eval_steps_per_second": 3.22, | |
| "step": 1836 | |
| }, | |
| { | |
| "epoch": 34.07, | |
| "learning_rate": 1.545679012345679e-05, | |
| "loss": 0.0021, | |
| "step": 1840 | |
| }, | |
| { | |
| "epoch": 34.26, | |
| "learning_rate": 1.54320987654321e-05, | |
| "loss": 0.0022, | |
| "step": 1850 | |
| }, | |
| { | |
| "epoch": 34.44, | |
| "learning_rate": 1.5407407407407408e-05, | |
| "loss": 0.0035, | |
| "step": 1860 | |
| }, | |
| { | |
| "epoch": 34.63, | |
| "learning_rate": 1.5382716049382717e-05, | |
| "loss": 0.0021, | |
| "step": 1870 | |
| }, | |
| { | |
| "epoch": 34.81, | |
| "learning_rate": 1.5358024691358026e-05, | |
| "loss": 0.0019, | |
| "step": 1880 | |
| }, | |
| { | |
| "epoch": 35.0, | |
| "learning_rate": 1.5333333333333334e-05, | |
| "loss": 0.0019, | |
| "step": 1890 | |
| }, | |
| { | |
| "epoch": 35.0, | |
| "eval_accuracy": 0.9868421052631579, | |
| "eval_loss": 0.05920861288905144, | |
| "eval_runtime": 3.006, | |
| "eval_samples_per_second": 25.283, | |
| "eval_steps_per_second": 3.327, | |
| "step": 1890 | |
| }, | |
| { | |
| "epoch": 35.19, | |
| "learning_rate": 1.5308641975308643e-05, | |
| "loss": 0.002, | |
| "step": 1900 | |
| }, | |
| { | |
| "epoch": 35.37, | |
| "learning_rate": 1.5283950617283952e-05, | |
| "loss": 0.0019, | |
| "step": 1910 | |
| }, | |
| { | |
| "epoch": 35.56, | |
| "learning_rate": 1.525925925925926e-05, | |
| "loss": 0.0073, | |
| "step": 1920 | |
| }, | |
| { | |
| "epoch": 35.74, | |
| "learning_rate": 1.523456790123457e-05, | |
| "loss": 0.0018, | |
| "step": 1930 | |
| }, | |
| { | |
| "epoch": 35.93, | |
| "learning_rate": 1.5209876543209879e-05, | |
| "loss": 0.0159, | |
| "step": 1940 | |
| }, | |
| { | |
| "epoch": 36.0, | |
| "eval_accuracy": 0.9868421052631579, | |
| "eval_loss": 0.0747406929731369, | |
| "eval_runtime": 3.0538, | |
| "eval_samples_per_second": 24.887, | |
| "eval_steps_per_second": 3.275, | |
| "step": 1944 | |
| }, | |
| { | |
| "epoch": 36.11, | |
| "learning_rate": 1.5185185185185187e-05, | |
| "loss": 0.0019, | |
| "step": 1950 | |
| }, | |
| { | |
| "epoch": 36.3, | |
| "learning_rate": 1.5160493827160495e-05, | |
| "loss": 0.0028, | |
| "step": 1960 | |
| }, | |
| { | |
| "epoch": 36.48, | |
| "learning_rate": 1.5135802469135803e-05, | |
| "loss": 0.0191, | |
| "step": 1970 | |
| }, | |
| { | |
| "epoch": 36.67, | |
| "learning_rate": 1.5111111111111112e-05, | |
| "loss": 0.0018, | |
| "step": 1980 | |
| }, | |
| { | |
| "epoch": 36.85, | |
| "learning_rate": 1.5086419753086421e-05, | |
| "loss": 0.0018, | |
| "step": 1990 | |
| }, | |
| { | |
| "epoch": 37.0, | |
| "eval_accuracy": 0.9868421052631579, | |
| "eval_loss": 0.02436704747378826, | |
| "eval_runtime": 3.1044, | |
| "eval_samples_per_second": 24.482, | |
| "eval_steps_per_second": 3.221, | |
| "step": 1998 | |
| }, | |
| { | |
| "epoch": 37.04, | |
| "learning_rate": 1.506172839506173e-05, | |
| "loss": 0.0019, | |
| "step": 2000 | |
| }, | |
| { | |
| "epoch": 37.22, | |
| "learning_rate": 1.5037037037037039e-05, | |
| "loss": 0.0018, | |
| "step": 2010 | |
| }, | |
| { | |
| "epoch": 37.41, | |
| "learning_rate": 1.5012345679012348e-05, | |
| "loss": 0.0025, | |
| "step": 2020 | |
| }, | |
| { | |
| "epoch": 37.59, | |
| "learning_rate": 1.4987654320987655e-05, | |
| "loss": 0.0017, | |
| "step": 2030 | |
| }, | |
| { | |
| "epoch": 37.78, | |
| "learning_rate": 1.4962962962962964e-05, | |
| "loss": 0.0017, | |
| "step": 2040 | |
| }, | |
| { | |
| "epoch": 37.96, | |
| "learning_rate": 1.4938271604938272e-05, | |
| "loss": 0.0016, | |
| "step": 2050 | |
| }, | |
| { | |
| "epoch": 38.0, | |
| "eval_accuracy": 1.0, | |
| "eval_loss": 0.0018851511413231492, | |
| "eval_runtime": 3.0108, | |
| "eval_samples_per_second": 25.242, | |
| "eval_steps_per_second": 3.321, | |
| "step": 2052 | |
| }, | |
| { | |
| "epoch": 38.15, | |
| "learning_rate": 1.4913580246913581e-05, | |
| "loss": 0.0815, | |
| "step": 2060 | |
| }, | |
| { | |
| "epoch": 38.33, | |
| "learning_rate": 1.488888888888889e-05, | |
| "loss": 0.03, | |
| "step": 2070 | |
| }, | |
| { | |
| "epoch": 38.52, | |
| "learning_rate": 1.4864197530864199e-05, | |
| "loss": 0.0017, | |
| "step": 2080 | |
| }, | |
| { | |
| "epoch": 38.7, | |
| "learning_rate": 1.4839506172839508e-05, | |
| "loss": 0.0021, | |
| "step": 2090 | |
| }, | |
| { | |
| "epoch": 38.89, | |
| "learning_rate": 1.4814814814814815e-05, | |
| "loss": 0.0017, | |
| "step": 2100 | |
| }, | |
| { | |
| "epoch": 39.0, | |
| "eval_accuracy": 1.0, | |
| "eval_loss": 0.0017620852449908853, | |
| "eval_runtime": 3.0074, | |
| "eval_samples_per_second": 25.271, | |
| "eval_steps_per_second": 3.325, | |
| "step": 2106 | |
| }, | |
| { | |
| "epoch": 39.07, | |
| "learning_rate": 1.4790123456790124e-05, | |
| "loss": 0.0016, | |
| "step": 2110 | |
| }, | |
| { | |
| "epoch": 39.26, | |
| "learning_rate": 1.4765432098765433e-05, | |
| "loss": 0.0293, | |
| "step": 2120 | |
| }, | |
| { | |
| "epoch": 39.44, | |
| "learning_rate": 1.4740740740740741e-05, | |
| "loss": 0.0715, | |
| "step": 2130 | |
| }, | |
| { | |
| "epoch": 39.63, | |
| "learning_rate": 1.471604938271605e-05, | |
| "loss": 0.0355, | |
| "step": 2140 | |
| }, | |
| { | |
| "epoch": 39.81, | |
| "learning_rate": 1.469135802469136e-05, | |
| "loss": 0.0334, | |
| "step": 2150 | |
| }, | |
| { | |
| "epoch": 40.0, | |
| "learning_rate": 1.4666666666666666e-05, | |
| "loss": 0.053, | |
| "step": 2160 | |
| }, | |
| { | |
| "epoch": 40.0, | |
| "eval_accuracy": 1.0, | |
| "eval_loss": 0.0022870118264108896, | |
| "eval_runtime": 3.0403, | |
| "eval_samples_per_second": 24.997, | |
| "eval_steps_per_second": 3.289, | |
| "step": 2160 | |
| }, | |
| { | |
| "epoch": 40.19, | |
| "learning_rate": 1.4641975308641975e-05, | |
| "loss": 0.0024, | |
| "step": 2170 | |
| }, | |
| { | |
| "epoch": 40.37, | |
| "learning_rate": 1.4617283950617284e-05, | |
| "loss": 0.0018, | |
| "step": 2180 | |
| }, | |
| { | |
| "epoch": 40.56, | |
| "learning_rate": 1.4592592592592594e-05, | |
| "loss": 0.0016, | |
| "step": 2190 | |
| }, | |
| { | |
| "epoch": 40.74, | |
| "learning_rate": 1.4567901234567903e-05, | |
| "loss": 0.0064, | |
| "step": 2200 | |
| }, | |
| { | |
| "epoch": 40.93, | |
| "learning_rate": 1.4543209876543212e-05, | |
| "loss": 0.0016, | |
| "step": 2210 | |
| }, | |
| { | |
| "epoch": 41.0, | |
| "eval_accuracy": 1.0, | |
| "eval_loss": 0.006081284489482641, | |
| "eval_runtime": 3.0753, | |
| "eval_samples_per_second": 24.713, | |
| "eval_steps_per_second": 3.252, | |
| "step": 2214 | |
| }, | |
| { | |
| "epoch": 41.11, | |
| "learning_rate": 1.4518518518518521e-05, | |
| "loss": 0.0771, | |
| "step": 2220 | |
| }, | |
| { | |
| "epoch": 41.3, | |
| "learning_rate": 1.4493827160493828e-05, | |
| "loss": 0.0015, | |
| "step": 2230 | |
| }, | |
| { | |
| "epoch": 41.48, | |
| "learning_rate": 1.4469135802469137e-05, | |
| "loss": 0.0016, | |
| "step": 2240 | |
| }, | |
| { | |
| "epoch": 41.67, | |
| "learning_rate": 1.4444444444444446e-05, | |
| "loss": 0.0015, | |
| "step": 2250 | |
| }, | |
| { | |
| "epoch": 41.85, | |
| "learning_rate": 1.4419753086419755e-05, | |
| "loss": 0.0015, | |
| "step": 2260 | |
| }, | |
| { | |
| "epoch": 42.0, | |
| "eval_accuracy": 1.0, | |
| "eval_loss": 0.010232136584818363, | |
| "eval_runtime": 3.0562, | |
| "eval_samples_per_second": 24.868, | |
| "eval_steps_per_second": 3.272, | |
| "step": 2268 | |
| }, | |
| { | |
| "epoch": 42.04, | |
| "learning_rate": 1.4395061728395063e-05, | |
| "loss": 0.0015, | |
| "step": 2270 | |
| }, | |
| { | |
| "epoch": 42.22, | |
| "learning_rate": 1.4370370370370372e-05, | |
| "loss": 0.0016, | |
| "step": 2280 | |
| }, | |
| { | |
| "epoch": 42.41, | |
| "learning_rate": 1.4345679012345681e-05, | |
| "loss": 0.0015, | |
| "step": 2290 | |
| }, | |
| { | |
| "epoch": 42.59, | |
| "learning_rate": 1.4320987654320988e-05, | |
| "loss": 0.0014, | |
| "step": 2300 | |
| }, | |
| { | |
| "epoch": 42.78, | |
| "learning_rate": 1.4296296296296297e-05, | |
| "loss": 0.0014, | |
| "step": 2310 | |
| }, | |
| { | |
| "epoch": 42.96, | |
| "learning_rate": 1.4271604938271606e-05, | |
| "loss": 0.0015, | |
| "step": 2320 | |
| }, | |
| { | |
| "epoch": 43.0, | |
| "eval_accuracy": 1.0, | |
| "eval_loss": 0.0018882902804762125, | |
| "eval_runtime": 3.0482, | |
| "eval_samples_per_second": 24.932, | |
| "eval_steps_per_second": 3.281, | |
| "step": 2322 | |
| }, | |
| { | |
| "epoch": 43.15, | |
| "learning_rate": 1.4246913580246915e-05, | |
| "loss": 0.0083, | |
| "step": 2330 | |
| }, | |
| { | |
| "epoch": 43.33, | |
| "learning_rate": 1.4222222222222224e-05, | |
| "loss": 0.0014, | |
| "step": 2340 | |
| }, | |
| { | |
| "epoch": 43.52, | |
| "learning_rate": 1.4197530864197532e-05, | |
| "loss": 0.002, | |
| "step": 2350 | |
| }, | |
| { | |
| "epoch": 43.7, | |
| "learning_rate": 1.417283950617284e-05, | |
| "loss": 0.0014, | |
| "step": 2360 | |
| }, | |
| { | |
| "epoch": 43.89, | |
| "learning_rate": 1.4148148148148148e-05, | |
| "loss": 0.0015, | |
| "step": 2370 | |
| }, | |
| { | |
| "epoch": 44.0, | |
| "eval_accuracy": 1.0, | |
| "eval_loss": 0.006185897625982761, | |
| "eval_runtime": 3.0918, | |
| "eval_samples_per_second": 24.581, | |
| "eval_steps_per_second": 3.234, | |
| "step": 2376 | |
| }, | |
| { | |
| "epoch": 44.07, | |
| "learning_rate": 1.4123456790123457e-05, | |
| "loss": 0.0014, | |
| "step": 2380 | |
| }, | |
| { | |
| "epoch": 44.26, | |
| "learning_rate": 1.4098765432098766e-05, | |
| "loss": 0.1322, | |
| "step": 2390 | |
| }, | |
| { | |
| "epoch": 44.44, | |
| "learning_rate": 1.4074074074074075e-05, | |
| "loss": 0.0126, | |
| "step": 2400 | |
| }, | |
| { | |
| "epoch": 44.63, | |
| "learning_rate": 1.4049382716049384e-05, | |
| "loss": 0.0013, | |
| "step": 2410 | |
| }, | |
| { | |
| "epoch": 44.81, | |
| "learning_rate": 1.4024691358024694e-05, | |
| "loss": 0.0062, | |
| "step": 2420 | |
| }, | |
| { | |
| "epoch": 45.0, | |
| "learning_rate": 1.4e-05, | |
| "loss": 0.0014, | |
| "step": 2430 | |
| }, | |
| { | |
| "epoch": 45.0, | |
| "eval_accuracy": 1.0, | |
| "eval_loss": 0.001433242461644113, | |
| "eval_runtime": 3.0606, | |
| "eval_samples_per_second": 24.832, | |
| "eval_steps_per_second": 3.267, | |
| "step": 2430 | |
| }, | |
| { | |
| "epoch": 45.19, | |
| "learning_rate": 1.3975308641975309e-05, | |
| "loss": 0.0013, | |
| "step": 2440 | |
| }, | |
| { | |
| "epoch": 45.37, | |
| "learning_rate": 1.3950617283950617e-05, | |
| "loss": 0.0013, | |
| "step": 2450 | |
| }, | |
| { | |
| "epoch": 45.56, | |
| "learning_rate": 1.3925925925925928e-05, | |
| "loss": 0.0013, | |
| "step": 2460 | |
| }, | |
| { | |
| "epoch": 45.74, | |
| "learning_rate": 1.3901234567901237e-05, | |
| "loss": 0.0014, | |
| "step": 2470 | |
| }, | |
| { | |
| "epoch": 45.93, | |
| "learning_rate": 1.3876543209876546e-05, | |
| "loss": 0.0015, | |
| "step": 2480 | |
| }, | |
| { | |
| "epoch": 46.0, | |
| "eval_accuracy": 1.0, | |
| "eval_loss": 0.0014952166238799691, | |
| "eval_runtime": 3.073, | |
| "eval_samples_per_second": 24.731, | |
| "eval_steps_per_second": 3.254, | |
| "step": 2484 | |
| }, | |
| { | |
| "epoch": 46.11, | |
| "learning_rate": 1.3851851851851851e-05, | |
| "loss": 0.0012, | |
| "step": 2490 | |
| }, | |
| { | |
| "epoch": 46.3, | |
| "learning_rate": 1.3827160493827162e-05, | |
| "loss": 0.0015, | |
| "step": 2500 | |
| }, | |
| { | |
| "epoch": 46.48, | |
| "learning_rate": 1.380246913580247e-05, | |
| "loss": 0.0014, | |
| "step": 2510 | |
| }, | |
| { | |
| "epoch": 46.67, | |
| "learning_rate": 1.377777777777778e-05, | |
| "loss": 0.0013, | |
| "step": 2520 | |
| }, | |
| { | |
| "epoch": 46.85, | |
| "learning_rate": 1.3753086419753088e-05, | |
| "loss": 0.0013, | |
| "step": 2530 | |
| }, | |
| { | |
| "epoch": 47.0, | |
| "eval_accuracy": 0.9868421052631579, | |
| "eval_loss": 0.06722357869148254, | |
| "eval_runtime": 3.0794, | |
| "eval_samples_per_second": 24.68, | |
| "eval_steps_per_second": 3.247, | |
| "step": 2538 | |
| }, | |
| { | |
| "epoch": 47.04, | |
| "learning_rate": 1.3728395061728397e-05, | |
| "loss": 0.0013, | |
| "step": 2540 | |
| }, | |
| { | |
| "epoch": 47.22, | |
| "learning_rate": 1.3703703703703706e-05, | |
| "loss": 0.0724, | |
| "step": 2550 | |
| }, | |
| { | |
| "epoch": 47.41, | |
| "learning_rate": 1.3679012345679013e-05, | |
| "loss": 0.0012, | |
| "step": 2560 | |
| }, | |
| { | |
| "epoch": 47.59, | |
| "learning_rate": 1.3654320987654322e-05, | |
| "loss": 0.0012, | |
| "step": 2570 | |
| }, | |
| { | |
| "epoch": 47.78, | |
| "learning_rate": 1.362962962962963e-05, | |
| "loss": 0.0012, | |
| "step": 2580 | |
| }, | |
| { | |
| "epoch": 47.96, | |
| "learning_rate": 1.360493827160494e-05, | |
| "loss": 0.0012, | |
| "step": 2590 | |
| }, | |
| { | |
| "epoch": 48.0, | |
| "eval_accuracy": 1.0, | |
| "eval_loss": 0.0014614396495744586, | |
| "eval_runtime": 3.0742, | |
| "eval_samples_per_second": 24.722, | |
| "eval_steps_per_second": 3.253, | |
| "step": 2592 | |
| }, | |
| { | |
| "epoch": 48.15, | |
| "learning_rate": 1.3580246913580248e-05, | |
| "loss": 0.0012, | |
| "step": 2600 | |
| }, | |
| { | |
| "epoch": 48.33, | |
| "learning_rate": 1.3555555555555557e-05, | |
| "loss": 0.0012, | |
| "step": 2610 | |
| }, | |
| { | |
| "epoch": 48.52, | |
| "learning_rate": 1.3530864197530866e-05, | |
| "loss": 0.0014, | |
| "step": 2620 | |
| }, | |
| { | |
| "epoch": 48.7, | |
| "learning_rate": 1.3506172839506173e-05, | |
| "loss": 0.018, | |
| "step": 2630 | |
| }, | |
| { | |
| "epoch": 48.89, | |
| "learning_rate": 1.3481481481481482e-05, | |
| "loss": 0.0012, | |
| "step": 2640 | |
| }, | |
| { | |
| "epoch": 49.0, | |
| "eval_accuracy": 0.9868421052631579, | |
| "eval_loss": 0.06998171657323837, | |
| "eval_runtime": 3.0377, | |
| "eval_samples_per_second": 25.019, | |
| "eval_steps_per_second": 3.292, | |
| "step": 2646 | |
| }, | |
| { | |
| "epoch": 49.07, | |
| "learning_rate": 1.345679012345679e-05, | |
| "loss": 0.0488, | |
| "step": 2650 | |
| }, | |
| { | |
| "epoch": 49.26, | |
| "learning_rate": 1.34320987654321e-05, | |
| "loss": 0.0267, | |
| "step": 2660 | |
| }, | |
| { | |
| "epoch": 49.44, | |
| "learning_rate": 1.3407407407407408e-05, | |
| "loss": 0.0012, | |
| "step": 2670 | |
| }, | |
| { | |
| "epoch": 49.63, | |
| "learning_rate": 1.3382716049382717e-05, | |
| "loss": 0.0012, | |
| "step": 2680 | |
| }, | |
| { | |
| "epoch": 49.81, | |
| "learning_rate": 1.3358024691358024e-05, | |
| "loss": 0.0012, | |
| "step": 2690 | |
| }, | |
| { | |
| "epoch": 50.0, | |
| "learning_rate": 1.3333333333333333e-05, | |
| "loss": 0.0012, | |
| "step": 2700 | |
| }, | |
| { | |
| "epoch": 50.0, | |
| "eval_accuracy": 0.9868421052631579, | |
| "eval_loss": 0.057919345796108246, | |
| "eval_runtime": 3.0575, | |
| "eval_samples_per_second": 24.857, | |
| "eval_steps_per_second": 3.271, | |
| "step": 2700 | |
| }, | |
| { | |
| "epoch": 50.19, | |
| "learning_rate": 1.3308641975308642e-05, | |
| "loss": 0.0016, | |
| "step": 2710 | |
| }, | |
| { | |
| "epoch": 50.37, | |
| "learning_rate": 1.3283950617283951e-05, | |
| "loss": 0.0011, | |
| "step": 2720 | |
| }, | |
| { | |
| "epoch": 50.56, | |
| "learning_rate": 1.325925925925926e-05, | |
| "loss": 0.0011, | |
| "step": 2730 | |
| }, | |
| { | |
| "epoch": 50.74, | |
| "learning_rate": 1.323456790123457e-05, | |
| "loss": 0.0013, | |
| "step": 2740 | |
| }, | |
| { | |
| "epoch": 50.93, | |
| "learning_rate": 1.320987654320988e-05, | |
| "loss": 0.0011, | |
| "step": 2750 | |
| }, | |
| { | |
| "epoch": 51.0, | |
| "eval_accuracy": 0.9868421052631579, | |
| "eval_loss": 0.05705736204981804, | |
| "eval_runtime": 3.0295, | |
| "eval_samples_per_second": 25.087, | |
| "eval_steps_per_second": 3.301, | |
| "step": 2754 | |
| }, | |
| { | |
| "epoch": 51.11, | |
| "learning_rate": 1.3185185185185185e-05, | |
| "loss": 0.0011, | |
| "step": 2760 | |
| }, | |
| { | |
| "epoch": 51.3, | |
| "learning_rate": 1.3160493827160493e-05, | |
| "loss": 0.0011, | |
| "step": 2770 | |
| }, | |
| { | |
| "epoch": 51.48, | |
| "learning_rate": 1.3135802469135804e-05, | |
| "loss": 0.0011, | |
| "step": 2780 | |
| }, | |
| { | |
| "epoch": 51.67, | |
| "learning_rate": 1.3111111111111113e-05, | |
| "loss": 0.0011, | |
| "step": 2790 | |
| }, | |
| { | |
| "epoch": 51.85, | |
| "learning_rate": 1.3086419753086422e-05, | |
| "loss": 0.001, | |
| "step": 2800 | |
| }, | |
| { | |
| "epoch": 52.0, | |
| "eval_accuracy": 0.9868421052631579, | |
| "eval_loss": 0.0670037791132927, | |
| "eval_runtime": 3.067, | |
| "eval_samples_per_second": 24.78, | |
| "eval_steps_per_second": 3.261, | |
| "step": 2808 | |
| }, | |
| { | |
| "epoch": 52.04, | |
| "learning_rate": 1.306172839506173e-05, | |
| "loss": 0.001, | |
| "step": 2810 | |
| }, | |
| { | |
| "epoch": 52.22, | |
| "learning_rate": 1.303703703703704e-05, | |
| "loss": 0.0011, | |
| "step": 2820 | |
| }, | |
| { | |
| "epoch": 52.41, | |
| "learning_rate": 1.3012345679012346e-05, | |
| "loss": 0.001, | |
| "step": 2830 | |
| }, | |
| { | |
| "epoch": 52.59, | |
| "learning_rate": 1.2987654320987655e-05, | |
| "loss": 0.001, | |
| "step": 2840 | |
| }, | |
| { | |
| "epoch": 52.78, | |
| "learning_rate": 1.2962962962962964e-05, | |
| "loss": 0.0012, | |
| "step": 2850 | |
| }, | |
| { | |
| "epoch": 52.96, | |
| "learning_rate": 1.2938271604938273e-05, | |
| "loss": 0.001, | |
| "step": 2860 | |
| }, | |
| { | |
| "epoch": 53.0, | |
| "eval_accuracy": 0.9868421052631579, | |
| "eval_loss": 0.07296162843704224, | |
| "eval_runtime": 3.0567, | |
| "eval_samples_per_second": 24.863, | |
| "eval_steps_per_second": 3.271, | |
| "step": 2862 | |
| }, | |
| { | |
| "epoch": 53.15, | |
| "learning_rate": 1.2913580246913582e-05, | |
| "loss": 0.001, | |
| "step": 2870 | |
| }, | |
| { | |
| "epoch": 53.33, | |
| "learning_rate": 1.288888888888889e-05, | |
| "loss": 0.001, | |
| "step": 2880 | |
| }, | |
| { | |
| "epoch": 53.52, | |
| "learning_rate": 1.2864197530864198e-05, | |
| "loss": 0.001, | |
| "step": 2890 | |
| }, | |
| { | |
| "epoch": 53.7, | |
| "learning_rate": 1.2839506172839507e-05, | |
| "loss": 0.057, | |
| "step": 2900 | |
| }, | |
| { | |
| "epoch": 53.89, | |
| "learning_rate": 1.2814814814814815e-05, | |
| "loss": 0.0013, | |
| "step": 2910 | |
| }, | |
| { | |
| "epoch": 54.0, | |
| "eval_accuracy": 0.9868421052631579, | |
| "eval_loss": 0.013546849600970745, | |
| "eval_runtime": 2.9939, | |
| "eval_samples_per_second": 25.385, | |
| "eval_steps_per_second": 3.34, | |
| "step": 2916 | |
| }, | |
| { | |
| "epoch": 54.07, | |
| "learning_rate": 1.2790123456790124e-05, | |
| "loss": 0.001, | |
| "step": 2920 | |
| }, | |
| { | |
| "epoch": 54.26, | |
| "learning_rate": 1.2765432098765433e-05, | |
| "loss": 0.0012, | |
| "step": 2930 | |
| }, | |
| { | |
| "epoch": 54.44, | |
| "learning_rate": 1.2740740740740742e-05, | |
| "loss": 0.0011, | |
| "step": 2940 | |
| }, | |
| { | |
| "epoch": 54.63, | |
| "learning_rate": 1.271604938271605e-05, | |
| "loss": 0.0009, | |
| "step": 2950 | |
| }, | |
| { | |
| "epoch": 54.81, | |
| "learning_rate": 1.2691358024691358e-05, | |
| "loss": 0.001, | |
| "step": 2960 | |
| }, | |
| { | |
| "epoch": 55.0, | |
| "learning_rate": 1.2666666666666667e-05, | |
| "loss": 0.001, | |
| "step": 2970 | |
| }, | |
| { | |
| "epoch": 55.0, | |
| "eval_accuracy": 0.9868421052631579, | |
| "eval_loss": 0.08355703204870224, | |
| "eval_runtime": 3.0103, | |
| "eval_samples_per_second": 25.247, | |
| "eval_steps_per_second": 3.322, | |
| "step": 2970 | |
| }, | |
| { | |
| "epoch": 55.19, | |
| "learning_rate": 1.2641975308641976e-05, | |
| "loss": 0.0009, | |
| "step": 2980 | |
| }, | |
| { | |
| "epoch": 55.37, | |
| "learning_rate": 1.2617283950617284e-05, | |
| "loss": 0.0476, | |
| "step": 2990 | |
| }, | |
| { | |
| "epoch": 55.56, | |
| "learning_rate": 1.2592592592592593e-05, | |
| "loss": 0.0009, | |
| "step": 3000 | |
| }, | |
| { | |
| "epoch": 55.74, | |
| "learning_rate": 1.2567901234567904e-05, | |
| "loss": 0.001, | |
| "step": 3010 | |
| }, | |
| { | |
| "epoch": 55.93, | |
| "learning_rate": 1.2543209876543213e-05, | |
| "loss": 0.0009, | |
| "step": 3020 | |
| }, | |
| { | |
| "epoch": 56.0, | |
| "eval_accuracy": 1.0, | |
| "eval_loss": 0.0010229795007035136, | |
| "eval_runtime": 3.0546, | |
| "eval_samples_per_second": 24.881, | |
| "eval_steps_per_second": 3.274, | |
| "step": 3024 | |
| }, | |
| { | |
| "epoch": 56.11, | |
| "learning_rate": 1.2518518518518518e-05, | |
| "loss": 0.0616, | |
| "step": 3030 | |
| }, | |
| { | |
| "epoch": 56.3, | |
| "learning_rate": 1.2493827160493827e-05, | |
| "loss": 0.0185, | |
| "step": 3040 | |
| }, | |
| { | |
| "epoch": 56.48, | |
| "learning_rate": 1.2469135802469137e-05, | |
| "loss": 0.0011, | |
| "step": 3050 | |
| }, | |
| { | |
| "epoch": 56.67, | |
| "learning_rate": 1.2444444444444446e-05, | |
| "loss": 0.001, | |
| "step": 3060 | |
| }, | |
| { | |
| "epoch": 56.85, | |
| "learning_rate": 1.2419753086419755e-05, | |
| "loss": 0.0009, | |
| "step": 3070 | |
| }, | |
| { | |
| "epoch": 57.0, | |
| "eval_accuracy": 0.9868421052631579, | |
| "eval_loss": 0.012184691615402699, | |
| "eval_runtime": 2.9929, | |
| "eval_samples_per_second": 25.393, | |
| "eval_steps_per_second": 3.341, | |
| "step": 3078 | |
| }, | |
| { | |
| "epoch": 57.04, | |
| "learning_rate": 1.2395061728395064e-05, | |
| "loss": 0.0009, | |
| "step": 3080 | |
| }, | |
| { | |
| "epoch": 57.22, | |
| "learning_rate": 1.2370370370370371e-05, | |
| "loss": 0.0009, | |
| "step": 3090 | |
| }, | |
| { | |
| "epoch": 57.41, | |
| "learning_rate": 1.234567901234568e-05, | |
| "loss": 0.0719, | |
| "step": 3100 | |
| }, | |
| { | |
| "epoch": 57.59, | |
| "learning_rate": 1.2320987654320989e-05, | |
| "loss": 0.0009, | |
| "step": 3110 | |
| }, | |
| { | |
| "epoch": 57.78, | |
| "learning_rate": 1.2296296296296298e-05, | |
| "loss": 0.0009, | |
| "step": 3120 | |
| }, | |
| { | |
| "epoch": 57.96, | |
| "learning_rate": 1.2271604938271606e-05, | |
| "loss": 0.001, | |
| "step": 3130 | |
| }, | |
| { | |
| "epoch": 58.0, | |
| "eval_accuracy": 0.9868421052631579, | |
| "eval_loss": 0.010523747652769089, | |
| "eval_runtime": 3.0537, | |
| "eval_samples_per_second": 24.887, | |
| "eval_steps_per_second": 3.275, | |
| "step": 3132 | |
| }, | |
| { | |
| "epoch": 58.15, | |
| "learning_rate": 1.2246913580246915e-05, | |
| "loss": 0.0124, | |
| "step": 3140 | |
| }, | |
| { | |
| "epoch": 58.33, | |
| "learning_rate": 1.2222222222222224e-05, | |
| "loss": 0.0255, | |
| "step": 3150 | |
| }, | |
| { | |
| "epoch": 58.52, | |
| "learning_rate": 1.2197530864197531e-05, | |
| "loss": 0.0778, | |
| "step": 3160 | |
| }, | |
| { | |
| "epoch": 58.7, | |
| "learning_rate": 1.217283950617284e-05, | |
| "loss": 0.0009, | |
| "step": 3170 | |
| }, | |
| { | |
| "epoch": 58.89, | |
| "learning_rate": 1.2148148148148149e-05, | |
| "loss": 0.0017, | |
| "step": 3180 | |
| }, | |
| { | |
| "epoch": 59.0, | |
| "eval_accuracy": 1.0, | |
| "eval_loss": 0.007372140884399414, | |
| "eval_runtime": 2.9827, | |
| "eval_samples_per_second": 25.481, | |
| "eval_steps_per_second": 3.353, | |
| "step": 3186 | |
| }, | |
| { | |
| "epoch": 59.07, | |
| "learning_rate": 1.2123456790123458e-05, | |
| "loss": 0.0009, | |
| "step": 3190 | |
| }, | |
| { | |
| "epoch": 59.26, | |
| "learning_rate": 1.2098765432098767e-05, | |
| "loss": 0.0009, | |
| "step": 3200 | |
| }, | |
| { | |
| "epoch": 59.44, | |
| "learning_rate": 1.2074074074074075e-05, | |
| "loss": 0.0009, | |
| "step": 3210 | |
| }, | |
| { | |
| "epoch": 59.63, | |
| "learning_rate": 1.2049382716049384e-05, | |
| "loss": 0.0009, | |
| "step": 3220 | |
| }, | |
| { | |
| "epoch": 59.81, | |
| "learning_rate": 1.2024691358024691e-05, | |
| "loss": 0.0009, | |
| "step": 3230 | |
| }, | |
| { | |
| "epoch": 60.0, | |
| "learning_rate": 1.2e-05, | |
| "loss": 0.0009, | |
| "step": 3240 | |
| }, | |
| { | |
| "epoch": 60.0, | |
| "eval_accuracy": 1.0, | |
| "eval_loss": 0.001023871824145317, | |
| "eval_runtime": 3.0602, | |
| "eval_samples_per_second": 24.835, | |
| "eval_steps_per_second": 3.268, | |
| "step": 3240 | |
| }, | |
| { | |
| "epoch": 60.19, | |
| "learning_rate": 1.1975308641975309e-05, | |
| "loss": 0.0009, | |
| "step": 3250 | |
| }, | |
| { | |
| "epoch": 60.37, | |
| "learning_rate": 1.1950617283950618e-05, | |
| "loss": 0.0008, | |
| "step": 3260 | |
| }, | |
| { | |
| "epoch": 60.56, | |
| "learning_rate": 1.1925925925925927e-05, | |
| "loss": 0.0009, | |
| "step": 3270 | |
| }, | |
| { | |
| "epoch": 60.74, | |
| "learning_rate": 1.1901234567901236e-05, | |
| "loss": 0.0008, | |
| "step": 3280 | |
| }, | |
| { | |
| "epoch": 60.93, | |
| "learning_rate": 1.1876543209876543e-05, | |
| "loss": 0.0009, | |
| "step": 3290 | |
| }, | |
| { | |
| "epoch": 61.0, | |
| "eval_accuracy": 1.0, | |
| "eval_loss": 0.0009232511511072516, | |
| "eval_runtime": 3.0495, | |
| "eval_samples_per_second": 24.922, | |
| "eval_steps_per_second": 3.279, | |
| "step": 3294 | |
| }, | |
| { | |
| "epoch": 61.11, | |
| "learning_rate": 1.1851851851851852e-05, | |
| "loss": 0.0009, | |
| "step": 3300 | |
| }, | |
| { | |
| "epoch": 61.3, | |
| "learning_rate": 1.182716049382716e-05, | |
| "loss": 0.0008, | |
| "step": 3310 | |
| }, | |
| { | |
| "epoch": 61.48, | |
| "learning_rate": 1.180246913580247e-05, | |
| "loss": 0.0008, | |
| "step": 3320 | |
| }, | |
| { | |
| "epoch": 61.67, | |
| "learning_rate": 1.177777777777778e-05, | |
| "loss": 0.0008, | |
| "step": 3330 | |
| }, | |
| { | |
| "epoch": 61.85, | |
| "learning_rate": 1.1753086419753089e-05, | |
| "loss": 0.0381, | |
| "step": 3340 | |
| }, | |
| { | |
| "epoch": 62.0, | |
| "eval_accuracy": 1.0, | |
| "eval_loss": 0.0020482551772147417, | |
| "eval_runtime": 3.0331, | |
| "eval_samples_per_second": 25.057, | |
| "eval_steps_per_second": 3.297, | |
| "step": 3348 | |
| }, | |
| { | |
| "epoch": 62.04, | |
| "learning_rate": 1.1728395061728398e-05, | |
| "loss": 0.0008, | |
| "step": 3350 | |
| }, | |
| { | |
| "epoch": 62.22, | |
| "learning_rate": 1.1703703703703703e-05, | |
| "loss": 0.0963, | |
| "step": 3360 | |
| }, | |
| { | |
| "epoch": 62.41, | |
| "learning_rate": 1.1679012345679013e-05, | |
| "loss": 0.0008, | |
| "step": 3370 | |
| }, | |
| { | |
| "epoch": 62.59, | |
| "learning_rate": 1.1654320987654322e-05, | |
| "loss": 0.0009, | |
| "step": 3380 | |
| }, | |
| { | |
| "epoch": 62.78, | |
| "learning_rate": 1.1629629629629631e-05, | |
| "loss": 0.0012, | |
| "step": 3390 | |
| }, | |
| { | |
| "epoch": 62.96, | |
| "learning_rate": 1.160493827160494e-05, | |
| "loss": 0.0008, | |
| "step": 3400 | |
| }, | |
| { | |
| "epoch": 63.0, | |
| "eval_accuracy": 1.0, | |
| "eval_loss": 0.0008004654664546251, | |
| "eval_runtime": 2.9607, | |
| "eval_samples_per_second": 25.669, | |
| "eval_steps_per_second": 3.378, | |
| "step": 3402 | |
| }, | |
| { | |
| "epoch": 63.15, | |
| "learning_rate": 1.1580246913580249e-05, | |
| "loss": 0.0008, | |
| "step": 3410 | |
| }, | |
| { | |
| "epoch": 63.33, | |
| "learning_rate": 1.1555555555555556e-05, | |
| "loss": 0.0008, | |
| "step": 3420 | |
| }, | |
| { | |
| "epoch": 63.52, | |
| "learning_rate": 1.1530864197530865e-05, | |
| "loss": 0.0008, | |
| "step": 3430 | |
| }, | |
| { | |
| "epoch": 63.7, | |
| "learning_rate": 1.1506172839506174e-05, | |
| "loss": 0.0007, | |
| "step": 3440 | |
| }, | |
| { | |
| "epoch": 63.89, | |
| "learning_rate": 1.1481481481481482e-05, | |
| "loss": 0.0099, | |
| "step": 3450 | |
| }, | |
| { | |
| "epoch": 64.0, | |
| "eval_accuracy": 1.0, | |
| "eval_loss": 0.0007782263564877212, | |
| "eval_runtime": 2.9249, | |
| "eval_samples_per_second": 25.984, | |
| "eval_steps_per_second": 3.419, | |
| "step": 3456 | |
| }, | |
| { | |
| "epoch": 64.07, | |
| "learning_rate": 1.1456790123456791e-05, | |
| "loss": 0.2271, | |
| "step": 3460 | |
| }, | |
| { | |
| "epoch": 64.26, | |
| "learning_rate": 1.14320987654321e-05, | |
| "loss": 0.0007, | |
| "step": 3470 | |
| }, | |
| { | |
| "epoch": 64.44, | |
| "learning_rate": 1.1407407407407409e-05, | |
| "loss": 0.0008, | |
| "step": 3480 | |
| }, | |
| { | |
| "epoch": 64.63, | |
| "learning_rate": 1.1382716049382716e-05, | |
| "loss": 0.0007, | |
| "step": 3490 | |
| }, | |
| { | |
| "epoch": 64.81, | |
| "learning_rate": 1.1358024691358025e-05, | |
| "loss": 0.0008, | |
| "step": 3500 | |
| }, | |
| { | |
| "epoch": 65.0, | |
| "learning_rate": 1.1333333333333334e-05, | |
| "loss": 0.0007, | |
| "step": 3510 | |
| }, | |
| { | |
| "epoch": 65.0, | |
| "eval_accuracy": 0.9868421052631579, | |
| "eval_loss": 0.07570363581180573, | |
| "eval_runtime": 3.0318, | |
| "eval_samples_per_second": 25.068, | |
| "eval_steps_per_second": 3.298, | |
| "step": 3510 | |
| }, | |
| { | |
| "epoch": 65.19, | |
| "learning_rate": 1.1308641975308643e-05, | |
| "loss": 0.0009, | |
| "step": 3520 | |
| }, | |
| { | |
| "epoch": 65.37, | |
| "learning_rate": 1.1283950617283951e-05, | |
| "loss": 0.0009, | |
| "step": 3530 | |
| }, | |
| { | |
| "epoch": 65.56, | |
| "learning_rate": 1.125925925925926e-05, | |
| "loss": 0.0007, | |
| "step": 3540 | |
| }, | |
| { | |
| "epoch": 65.74, | |
| "learning_rate": 1.123456790123457e-05, | |
| "loss": 0.0007, | |
| "step": 3550 | |
| }, | |
| { | |
| "epoch": 65.93, | |
| "learning_rate": 1.1209876543209876e-05, | |
| "loss": 0.0008, | |
| "step": 3560 | |
| }, | |
| { | |
| "epoch": 66.0, | |
| "eval_accuracy": 0.9868421052631579, | |
| "eval_loss": 0.07635856419801712, | |
| "eval_runtime": 3.1106, | |
| "eval_samples_per_second": 24.432, | |
| "eval_steps_per_second": 3.215, | |
| "step": 3564 | |
| }, | |
| { | |
| "epoch": 66.11, | |
| "learning_rate": 1.1185185185185185e-05, | |
| "loss": 0.0752, | |
| "step": 3570 | |
| }, | |
| { | |
| "epoch": 66.3, | |
| "learning_rate": 1.1160493827160494e-05, | |
| "loss": 0.0008, | |
| "step": 3580 | |
| }, | |
| { | |
| "epoch": 66.48, | |
| "learning_rate": 1.1135802469135803e-05, | |
| "loss": 0.0788, | |
| "step": 3590 | |
| }, | |
| { | |
| "epoch": 66.67, | |
| "learning_rate": 1.1111111111111113e-05, | |
| "loss": 0.0007, | |
| "step": 3600 | |
| }, | |
| { | |
| "epoch": 66.85, | |
| "learning_rate": 1.1086419753086422e-05, | |
| "loss": 0.0007, | |
| "step": 3610 | |
| }, | |
| { | |
| "epoch": 67.0, | |
| "eval_accuracy": 0.9736842105263158, | |
| "eval_loss": 0.12568235397338867, | |
| "eval_runtime": 2.964, | |
| "eval_samples_per_second": 25.641, | |
| "eval_steps_per_second": 3.374, | |
| "step": 3618 | |
| }, | |
| { | |
| "epoch": 67.04, | |
| "learning_rate": 1.1061728395061728e-05, | |
| "loss": 0.001, | |
| "step": 3620 | |
| }, | |
| { | |
| "epoch": 67.22, | |
| "learning_rate": 1.1037037037037036e-05, | |
| "loss": 0.0007, | |
| "step": 3630 | |
| }, | |
| { | |
| "epoch": 67.41, | |
| "learning_rate": 1.1012345679012347e-05, | |
| "loss": 0.0011, | |
| "step": 3640 | |
| }, | |
| { | |
| "epoch": 67.59, | |
| "learning_rate": 1.0987654320987656e-05, | |
| "loss": 0.0007, | |
| "step": 3650 | |
| }, | |
| { | |
| "epoch": 67.78, | |
| "learning_rate": 1.0962962962962965e-05, | |
| "loss": 0.0007, | |
| "step": 3660 | |
| }, | |
| { | |
| "epoch": 67.96, | |
| "learning_rate": 1.0938271604938273e-05, | |
| "loss": 0.0007, | |
| "step": 3670 | |
| }, | |
| { | |
| "epoch": 68.0, | |
| "eval_accuracy": 0.9868421052631579, | |
| "eval_loss": 0.009799230843782425, | |
| "eval_runtime": 3.0392, | |
| "eval_samples_per_second": 25.007, | |
| "eval_steps_per_second": 3.29, | |
| "step": 3672 | |
| }, | |
| { | |
| "epoch": 68.15, | |
| "learning_rate": 1.0913580246913582e-05, | |
| "loss": 0.0841, | |
| "step": 3680 | |
| }, | |
| { | |
| "epoch": 68.33, | |
| "learning_rate": 1.088888888888889e-05, | |
| "loss": 0.0007, | |
| "step": 3690 | |
| }, | |
| { | |
| "epoch": 68.52, | |
| "learning_rate": 1.0864197530864198e-05, | |
| "loss": 0.0007, | |
| "step": 3700 | |
| }, | |
| { | |
| "epoch": 68.7, | |
| "learning_rate": 1.0839506172839507e-05, | |
| "loss": 0.0382, | |
| "step": 3710 | |
| }, | |
| { | |
| "epoch": 68.89, | |
| "learning_rate": 1.0814814814814816e-05, | |
| "loss": 0.0736, | |
| "step": 3720 | |
| }, | |
| { | |
| "epoch": 69.0, | |
| "eval_accuracy": 1.0, | |
| "eval_loss": 0.0007917585899122059, | |
| "eval_runtime": 3.0348, | |
| "eval_samples_per_second": 25.043, | |
| "eval_steps_per_second": 3.295, | |
| "step": 3726 | |
| }, | |
| { | |
| "epoch": 69.07, | |
| "learning_rate": 1.0790123456790125e-05, | |
| "loss": 0.0007, | |
| "step": 3730 | |
| }, | |
| { | |
| "epoch": 69.26, | |
| "learning_rate": 1.0765432098765434e-05, | |
| "loss": 0.0007, | |
| "step": 3740 | |
| }, | |
| { | |
| "epoch": 69.44, | |
| "learning_rate": 1.0740740740740742e-05, | |
| "loss": 0.0007, | |
| "step": 3750 | |
| }, | |
| { | |
| "epoch": 69.63, | |
| "learning_rate": 1.071604938271605e-05, | |
| "loss": 0.0008, | |
| "step": 3760 | |
| }, | |
| { | |
| "epoch": 69.81, | |
| "learning_rate": 1.0691358024691358e-05, | |
| "loss": 0.0014, | |
| "step": 3770 | |
| }, | |
| { | |
| "epoch": 70.0, | |
| "learning_rate": 1.0666666666666667e-05, | |
| "loss": 0.0007, | |
| "step": 3780 | |
| }, | |
| { | |
| "epoch": 70.0, | |
| "eval_accuracy": 0.9868421052631579, | |
| "eval_loss": 0.060479313135147095, | |
| "eval_runtime": 3.0758, | |
| "eval_samples_per_second": 24.709, | |
| "eval_steps_per_second": 3.251, | |
| "step": 3780 | |
| }, | |
| { | |
| "epoch": 70.19, | |
| "learning_rate": 1.0641975308641976e-05, | |
| "loss": 0.026, | |
| "step": 3790 | |
| }, | |
| { | |
| "epoch": 70.37, | |
| "learning_rate": 1.0617283950617285e-05, | |
| "loss": 0.0022, | |
| "step": 3800 | |
| }, | |
| { | |
| "epoch": 70.56, | |
| "learning_rate": 1.0592592592592594e-05, | |
| "loss": 0.0007, | |
| "step": 3810 | |
| }, | |
| { | |
| "epoch": 70.74, | |
| "learning_rate": 1.0567901234567901e-05, | |
| "loss": 0.0067, | |
| "step": 3820 | |
| }, | |
| { | |
| "epoch": 70.93, | |
| "learning_rate": 1.054320987654321e-05, | |
| "loss": 0.0006, | |
| "step": 3830 | |
| }, | |
| { | |
| "epoch": 71.0, | |
| "eval_accuracy": 1.0, | |
| "eval_loss": 0.0011679435847327113, | |
| "eval_runtime": 3.0887, | |
| "eval_samples_per_second": 24.606, | |
| "eval_steps_per_second": 3.238, | |
| "step": 3834 | |
| }, | |
| { | |
| "epoch": 71.11, | |
| "learning_rate": 1.0518518518518519e-05, | |
| "loss": 0.0007, | |
| "step": 3840 | |
| }, | |
| { | |
| "epoch": 71.3, | |
| "learning_rate": 1.0493827160493827e-05, | |
| "loss": 0.0007, | |
| "step": 3850 | |
| }, | |
| { | |
| "epoch": 71.48, | |
| "learning_rate": 1.0469135802469136e-05, | |
| "loss": 0.0591, | |
| "step": 3860 | |
| }, | |
| { | |
| "epoch": 71.67, | |
| "learning_rate": 1.0444444444444445e-05, | |
| "loss": 0.0737, | |
| "step": 3870 | |
| }, | |
| { | |
| "epoch": 71.85, | |
| "learning_rate": 1.0419753086419756e-05, | |
| "loss": 0.001, | |
| "step": 3880 | |
| }, | |
| { | |
| "epoch": 72.0, | |
| "eval_accuracy": 0.9736842105263158, | |
| "eval_loss": 0.16656266152858734, | |
| "eval_runtime": 3.0805, | |
| "eval_samples_per_second": 24.672, | |
| "eval_steps_per_second": 3.246, | |
| "step": 3888 | |
| }, | |
| { | |
| "epoch": 72.04, | |
| "learning_rate": 1.0395061728395061e-05, | |
| "loss": 0.0008, | |
| "step": 3890 | |
| }, | |
| { | |
| "epoch": 72.22, | |
| "learning_rate": 1.037037037037037e-05, | |
| "loss": 0.0015, | |
| "step": 3900 | |
| }, | |
| { | |
| "epoch": 72.41, | |
| "learning_rate": 1.0345679012345679e-05, | |
| "loss": 0.0065, | |
| "step": 3910 | |
| }, | |
| { | |
| "epoch": 72.59, | |
| "learning_rate": 1.032098765432099e-05, | |
| "loss": 0.0007, | |
| "step": 3920 | |
| }, | |
| { | |
| "epoch": 72.78, | |
| "learning_rate": 1.0296296296296298e-05, | |
| "loss": 0.0008, | |
| "step": 3930 | |
| }, | |
| { | |
| "epoch": 72.96, | |
| "learning_rate": 1.0271604938271607e-05, | |
| "loss": 0.0042, | |
| "step": 3940 | |
| }, | |
| { | |
| "epoch": 73.0, | |
| "eval_accuracy": 1.0, | |
| "eval_loss": 0.0007003291393630207, | |
| "eval_runtime": 3.0627, | |
| "eval_samples_per_second": 24.815, | |
| "eval_steps_per_second": 3.265, | |
| "step": 3942 | |
| }, | |
| { | |
| "epoch": 73.15, | |
| "learning_rate": 1.0246913580246916e-05, | |
| "loss": 0.0007, | |
| "step": 3950 | |
| }, | |
| { | |
| "epoch": 73.33, | |
| "learning_rate": 1.0222222222222223e-05, | |
| "loss": 0.0141, | |
| "step": 3960 | |
| }, | |
| { | |
| "epoch": 73.52, | |
| "learning_rate": 1.0197530864197532e-05, | |
| "loss": 0.0006, | |
| "step": 3970 | |
| }, | |
| { | |
| "epoch": 73.7, | |
| "learning_rate": 1.017283950617284e-05, | |
| "loss": 0.0006, | |
| "step": 3980 | |
| }, | |
| { | |
| "epoch": 73.89, | |
| "learning_rate": 1.014814814814815e-05, | |
| "loss": 0.0006, | |
| "step": 3990 | |
| }, | |
| { | |
| "epoch": 74.0, | |
| "eval_accuracy": 1.0, | |
| "eval_loss": 0.0006768173188902438, | |
| "eval_runtime": 3.0459, | |
| "eval_samples_per_second": 24.952, | |
| "eval_steps_per_second": 3.283, | |
| "step": 3996 | |
| }, | |
| { | |
| "epoch": 74.07, | |
| "learning_rate": 1.0123456790123458e-05, | |
| "loss": 0.0007, | |
| "step": 4000 | |
| }, | |
| { | |
| "epoch": 74.26, | |
| "learning_rate": 1.0098765432098767e-05, | |
| "loss": 0.0006, | |
| "step": 4010 | |
| }, | |
| { | |
| "epoch": 74.44, | |
| "learning_rate": 1.0074074074074074e-05, | |
| "loss": 0.0006, | |
| "step": 4020 | |
| }, | |
| { | |
| "epoch": 74.63, | |
| "learning_rate": 1.0049382716049383e-05, | |
| "loss": 0.0006, | |
| "step": 4030 | |
| }, | |
| { | |
| "epoch": 74.81, | |
| "learning_rate": 1.0024691358024692e-05, | |
| "loss": 0.0006, | |
| "step": 4040 | |
| }, | |
| { | |
| "epoch": 75.0, | |
| "learning_rate": 1e-05, | |
| "loss": 0.0007, | |
| "step": 4050 | |
| }, | |
| { | |
| "epoch": 75.0, | |
| "eval_accuracy": 1.0, | |
| "eval_loss": 0.0006513882544822991, | |
| "eval_runtime": 3.0655, | |
| "eval_samples_per_second": 24.792, | |
| "eval_steps_per_second": 3.262, | |
| "step": 4050 | |
| }, | |
| { | |
| "epoch": 75.19, | |
| "learning_rate": 9.97530864197531e-06, | |
| "loss": 0.001, | |
| "step": 4060 | |
| }, | |
| { | |
| "epoch": 75.37, | |
| "learning_rate": 9.950617283950618e-06, | |
| "loss": 0.0006, | |
| "step": 4070 | |
| }, | |
| { | |
| "epoch": 75.56, | |
| "learning_rate": 9.925925925925927e-06, | |
| "loss": 0.0006, | |
| "step": 4080 | |
| }, | |
| { | |
| "epoch": 75.74, | |
| "learning_rate": 9.901234567901236e-06, | |
| "loss": 0.0006, | |
| "step": 4090 | |
| }, | |
| { | |
| "epoch": 75.93, | |
| "learning_rate": 9.876543209876543e-06, | |
| "loss": 0.0006, | |
| "step": 4100 | |
| }, | |
| { | |
| "epoch": 76.0, | |
| "eval_accuracy": 0.9868421052631579, | |
| "eval_loss": 0.03311971202492714, | |
| "eval_runtime": 2.9918, | |
| "eval_samples_per_second": 25.403, | |
| "eval_steps_per_second": 3.343, | |
| "step": 4104 | |
| }, | |
| { | |
| "epoch": 76.11, | |
| "learning_rate": 9.851851851851852e-06, | |
| "loss": 0.0006, | |
| "step": 4110 | |
| }, | |
| { | |
| "epoch": 76.3, | |
| "learning_rate": 9.827160493827161e-06, | |
| "loss": 0.0006, | |
| "step": 4120 | |
| }, | |
| { | |
| "epoch": 76.48, | |
| "learning_rate": 9.80246913580247e-06, | |
| "loss": 0.0006, | |
| "step": 4130 | |
| }, | |
| { | |
| "epoch": 76.67, | |
| "learning_rate": 9.777777777777779e-06, | |
| "loss": 0.0006, | |
| "step": 4140 | |
| }, | |
| { | |
| "epoch": 76.85, | |
| "learning_rate": 9.753086419753087e-06, | |
| "loss": 0.0006, | |
| "step": 4150 | |
| }, | |
| { | |
| "epoch": 77.0, | |
| "eval_accuracy": 0.9868421052631579, | |
| "eval_loss": 0.01688443496823311, | |
| "eval_runtime": 3.0085, | |
| "eval_samples_per_second": 25.262, | |
| "eval_steps_per_second": 3.324, | |
| "step": 4158 | |
| }, | |
| { | |
| "epoch": 77.04, | |
| "learning_rate": 9.728395061728396e-06, | |
| "loss": 0.0005, | |
| "step": 4160 | |
| }, | |
| { | |
| "epoch": 77.22, | |
| "learning_rate": 9.703703703703703e-06, | |
| "loss": 0.0006, | |
| "step": 4170 | |
| }, | |
| { | |
| "epoch": 77.41, | |
| "learning_rate": 9.679012345679012e-06, | |
| "loss": 0.0005, | |
| "step": 4180 | |
| }, | |
| { | |
| "epoch": 77.59, | |
| "learning_rate": 9.654320987654323e-06, | |
| "loss": 0.0005, | |
| "step": 4190 | |
| }, | |
| { | |
| "epoch": 77.78, | |
| "learning_rate": 9.62962962962963e-06, | |
| "loss": 0.0005, | |
| "step": 4200 | |
| }, | |
| { | |
| "epoch": 77.96, | |
| "learning_rate": 9.604938271604939e-06, | |
| "loss": 0.0345, | |
| "step": 4210 | |
| }, | |
| { | |
| "epoch": 78.0, | |
| "eval_accuracy": 1.0, | |
| "eval_loss": 0.0006252001621760428, | |
| "eval_runtime": 3.0532, | |
| "eval_samples_per_second": 24.892, | |
| "eval_steps_per_second": 3.275, | |
| "step": 4212 | |
| }, | |
| { | |
| "epoch": 78.15, | |
| "learning_rate": 9.580246913580248e-06, | |
| "loss": 0.0769, | |
| "step": 4220 | |
| }, | |
| { | |
| "epoch": 78.33, | |
| "learning_rate": 9.555555555555556e-06, | |
| "loss": 0.0006, | |
| "step": 4230 | |
| }, | |
| { | |
| "epoch": 78.52, | |
| "learning_rate": 9.530864197530865e-06, | |
| "loss": 0.0006, | |
| "step": 4240 | |
| }, | |
| { | |
| "epoch": 78.7, | |
| "learning_rate": 9.506172839506174e-06, | |
| "loss": 0.0035, | |
| "step": 4250 | |
| }, | |
| { | |
| "epoch": 78.89, | |
| "learning_rate": 9.481481481481483e-06, | |
| "loss": 0.0005, | |
| "step": 4260 | |
| }, | |
| { | |
| "epoch": 79.0, | |
| "eval_accuracy": 0.9868421052631579, | |
| "eval_loss": 0.07624808698892593, | |
| "eval_runtime": 3.0922, | |
| "eval_samples_per_second": 24.578, | |
| "eval_steps_per_second": 3.234, | |
| "step": 4266 | |
| }, | |
| { | |
| "epoch": 79.07, | |
| "learning_rate": 9.45679012345679e-06, | |
| "loss": 0.0011, | |
| "step": 4270 | |
| }, | |
| { | |
| "epoch": 79.26, | |
| "learning_rate": 9.432098765432099e-06, | |
| "loss": 0.0005, | |
| "step": 4280 | |
| }, | |
| { | |
| "epoch": 79.44, | |
| "learning_rate": 9.407407407407408e-06, | |
| "loss": 0.0005, | |
| "step": 4290 | |
| }, | |
| { | |
| "epoch": 79.63, | |
| "learning_rate": 9.382716049382717e-06, | |
| "loss": 0.0005, | |
| "step": 4300 | |
| }, | |
| { | |
| "epoch": 79.81, | |
| "learning_rate": 9.358024691358025e-06, | |
| "loss": 0.0006, | |
| "step": 4310 | |
| }, | |
| { | |
| "epoch": 80.0, | |
| "learning_rate": 9.333333333333334e-06, | |
| "loss": 0.0005, | |
| "step": 4320 | |
| }, | |
| { | |
| "epoch": 80.0, | |
| "eval_accuracy": 1.0, | |
| "eval_loss": 0.0006611607386730611, | |
| "eval_runtime": 3.0596, | |
| "eval_samples_per_second": 24.84, | |
| "eval_steps_per_second": 3.268, | |
| "step": 4320 | |
| }, | |
| { | |
| "epoch": 80.19, | |
| "learning_rate": 9.308641975308643e-06, | |
| "loss": 0.0005, | |
| "step": 4330 | |
| }, | |
| { | |
| "epoch": 80.37, | |
| "learning_rate": 9.28395061728395e-06, | |
| "loss": 0.0027, | |
| "step": 4340 | |
| }, | |
| { | |
| "epoch": 80.56, | |
| "learning_rate": 9.25925925925926e-06, | |
| "loss": 0.0005, | |
| "step": 4350 | |
| }, | |
| { | |
| "epoch": 80.74, | |
| "learning_rate": 9.23456790123457e-06, | |
| "loss": 0.0006, | |
| "step": 4360 | |
| }, | |
| { | |
| "epoch": 80.93, | |
| "learning_rate": 9.209876543209877e-06, | |
| "loss": 0.0005, | |
| "step": 4370 | |
| }, | |
| { | |
| "epoch": 81.0, | |
| "eval_accuracy": 1.0, | |
| "eval_loss": 0.0005299604381434619, | |
| "eval_runtime": 3.0798, | |
| "eval_samples_per_second": 24.677, | |
| "eval_steps_per_second": 3.247, | |
| "step": 4374 | |
| }, | |
| { | |
| "epoch": 81.11, | |
| "learning_rate": 9.185185185185186e-06, | |
| "loss": 0.0005, | |
| "step": 4380 | |
| }, | |
| { | |
| "epoch": 81.3, | |
| "learning_rate": 9.160493827160494e-06, | |
| "loss": 0.0008, | |
| "step": 4390 | |
| }, | |
| { | |
| "epoch": 81.48, | |
| "learning_rate": 9.135802469135803e-06, | |
| "loss": 0.0163, | |
| "step": 4400 | |
| }, | |
| { | |
| "epoch": 81.67, | |
| "learning_rate": 9.111111111111112e-06, | |
| "loss": 0.0005, | |
| "step": 4410 | |
| }, | |
| { | |
| "epoch": 81.85, | |
| "learning_rate": 9.086419753086421e-06, | |
| "loss": 0.0005, | |
| "step": 4420 | |
| }, | |
| { | |
| "epoch": 82.0, | |
| "eval_accuracy": 1.0, | |
| "eval_loss": 0.000567257811781019, | |
| "eval_runtime": 3.0454, | |
| "eval_samples_per_second": 24.956, | |
| "eval_steps_per_second": 3.284, | |
| "step": 4428 | |
| }, | |
| { | |
| "epoch": 82.04, | |
| "learning_rate": 9.06172839506173e-06, | |
| "loss": 0.0005, | |
| "step": 4430 | |
| }, | |
| { | |
| "epoch": 82.22, | |
| "learning_rate": 9.037037037037037e-06, | |
| "loss": 0.0005, | |
| "step": 4440 | |
| }, | |
| { | |
| "epoch": 82.41, | |
| "learning_rate": 9.012345679012346e-06, | |
| "loss": 0.0005, | |
| "step": 4450 | |
| }, | |
| { | |
| "epoch": 82.59, | |
| "learning_rate": 8.987654320987655e-06, | |
| "loss": 0.0005, | |
| "step": 4460 | |
| }, | |
| { | |
| "epoch": 82.78, | |
| "learning_rate": 8.962962962962963e-06, | |
| "loss": 0.0007, | |
| "step": 4470 | |
| }, | |
| { | |
| "epoch": 82.96, | |
| "learning_rate": 8.938271604938272e-06, | |
| "loss": 0.0005, | |
| "step": 4480 | |
| }, | |
| { | |
| "epoch": 83.0, | |
| "eval_accuracy": 1.0, | |
| "eval_loss": 0.00053614896023646, | |
| "eval_runtime": 3.1232, | |
| "eval_samples_per_second": 24.334, | |
| "eval_steps_per_second": 3.202, | |
| "step": 4482 | |
| }, | |
| { | |
| "epoch": 83.15, | |
| "learning_rate": 8.913580246913581e-06, | |
| "loss": 0.0005, | |
| "step": 4490 | |
| }, | |
| { | |
| "epoch": 83.33, | |
| "learning_rate": 8.888888888888888e-06, | |
| "loss": 0.0005, | |
| "step": 4500 | |
| }, | |
| { | |
| "epoch": 83.52, | |
| "learning_rate": 8.864197530864199e-06, | |
| "loss": 0.0005, | |
| "step": 4510 | |
| }, | |
| { | |
| "epoch": 83.7, | |
| "learning_rate": 8.839506172839508e-06, | |
| "loss": 0.0005, | |
| "step": 4520 | |
| }, | |
| { | |
| "epoch": 83.89, | |
| "learning_rate": 8.814814814814817e-06, | |
| "loss": 0.0005, | |
| "step": 4530 | |
| }, | |
| { | |
| "epoch": 84.0, | |
| "eval_accuracy": 1.0, | |
| "eval_loss": 0.0005005816929042339, | |
| "eval_runtime": 3.0191, | |
| "eval_samples_per_second": 25.173, | |
| "eval_steps_per_second": 3.312, | |
| "step": 4536 | |
| }, | |
| { | |
| "epoch": 84.07, | |
| "learning_rate": 8.790123456790124e-06, | |
| "loss": 0.0005, | |
| "step": 4540 | |
| }, | |
| { | |
| "epoch": 84.26, | |
| "learning_rate": 8.765432098765432e-06, | |
| "loss": 0.0005, | |
| "step": 4550 | |
| }, | |
| { | |
| "epoch": 84.44, | |
| "learning_rate": 8.740740740740741e-06, | |
| "loss": 0.0042, | |
| "step": 4560 | |
| }, | |
| { | |
| "epoch": 84.63, | |
| "learning_rate": 8.71604938271605e-06, | |
| "loss": 0.0004, | |
| "step": 4570 | |
| }, | |
| { | |
| "epoch": 84.81, | |
| "learning_rate": 8.691358024691359e-06, | |
| "loss": 0.0074, | |
| "step": 4580 | |
| }, | |
| { | |
| "epoch": 85.0, | |
| "learning_rate": 8.666666666666668e-06, | |
| "loss": 0.0047, | |
| "step": 4590 | |
| }, | |
| { | |
| "epoch": 85.0, | |
| "eval_accuracy": 1.0, | |
| "eval_loss": 0.0006987836095504463, | |
| "eval_runtime": 3.0927, | |
| "eval_samples_per_second": 24.574, | |
| "eval_steps_per_second": 3.233, | |
| "step": 4590 | |
| }, | |
| { | |
| "epoch": 85.19, | |
| "learning_rate": 8.641975308641975e-06, | |
| "loss": 0.0005, | |
| "step": 4600 | |
| }, | |
| { | |
| "epoch": 85.37, | |
| "learning_rate": 8.617283950617284e-06, | |
| "loss": 0.0004, | |
| "step": 4610 | |
| }, | |
| { | |
| "epoch": 85.56, | |
| "learning_rate": 8.592592592592593e-06, | |
| "loss": 0.0005, | |
| "step": 4620 | |
| }, | |
| { | |
| "epoch": 85.74, | |
| "learning_rate": 8.567901234567903e-06, | |
| "loss": 0.06, | |
| "step": 4630 | |
| }, | |
| { | |
| "epoch": 85.93, | |
| "learning_rate": 8.54320987654321e-06, | |
| "loss": 0.0005, | |
| "step": 4640 | |
| }, | |
| { | |
| "epoch": 86.0, | |
| "eval_accuracy": 1.0, | |
| "eval_loss": 0.00046328367898240685, | |
| "eval_runtime": 3.024, | |
| "eval_samples_per_second": 25.132, | |
| "eval_steps_per_second": 3.307, | |
| "step": 4644 | |
| }, | |
| { | |
| "epoch": 86.11, | |
| "learning_rate": 8.518518518518519e-06, | |
| "loss": 0.0004, | |
| "step": 4650 | |
| }, | |
| { | |
| "epoch": 86.3, | |
| "learning_rate": 8.493827160493828e-06, | |
| "loss": 0.0004, | |
| "step": 4660 | |
| }, | |
| { | |
| "epoch": 86.48, | |
| "learning_rate": 8.469135802469137e-06, | |
| "loss": 0.0005, | |
| "step": 4670 | |
| }, | |
| { | |
| "epoch": 86.67, | |
| "learning_rate": 8.444444444444446e-06, | |
| "loss": 0.0005, | |
| "step": 4680 | |
| }, | |
| { | |
| "epoch": 86.85, | |
| "learning_rate": 8.419753086419754e-06, | |
| "loss": 0.0005, | |
| "step": 4690 | |
| }, | |
| { | |
| "epoch": 87.0, | |
| "eval_accuracy": 1.0, | |
| "eval_loss": 0.0005323333316482604, | |
| "eval_runtime": 3.0578, | |
| "eval_samples_per_second": 24.855, | |
| "eval_steps_per_second": 3.27, | |
| "step": 4698 | |
| }, | |
| { | |
| "epoch": 87.04, | |
| "learning_rate": 8.395061728395062e-06, | |
| "loss": 0.0007, | |
| "step": 4700 | |
| }, | |
| { | |
| "epoch": 87.22, | |
| "learning_rate": 8.37037037037037e-06, | |
| "loss": 0.0006, | |
| "step": 4710 | |
| }, | |
| { | |
| "epoch": 87.41, | |
| "learning_rate": 8.34567901234568e-06, | |
| "loss": 0.0004, | |
| "step": 4720 | |
| }, | |
| { | |
| "epoch": 87.59, | |
| "learning_rate": 8.320987654320988e-06, | |
| "loss": 0.0553, | |
| "step": 4730 | |
| }, | |
| { | |
| "epoch": 87.78, | |
| "learning_rate": 8.296296296296297e-06, | |
| "loss": 0.0005, | |
| "step": 4740 | |
| }, | |
| { | |
| "epoch": 87.96, | |
| "learning_rate": 8.271604938271606e-06, | |
| "loss": 0.0004, | |
| "step": 4750 | |
| }, | |
| { | |
| "epoch": 88.0, | |
| "eval_accuracy": 1.0, | |
| "eval_loss": 0.0004464307858143002, | |
| "eval_runtime": 3.0308, | |
| "eval_samples_per_second": 25.076, | |
| "eval_steps_per_second": 3.299, | |
| "step": 4752 | |
| }, | |
| { | |
| "epoch": 88.15, | |
| "learning_rate": 8.246913580246915e-06, | |
| "loss": 0.0004, | |
| "step": 4760 | |
| }, | |
| { | |
| "epoch": 88.33, | |
| "learning_rate": 8.222222222222222e-06, | |
| "loss": 0.0004, | |
| "step": 4770 | |
| }, | |
| { | |
| "epoch": 88.52, | |
| "learning_rate": 8.197530864197532e-06, | |
| "loss": 0.0004, | |
| "step": 4780 | |
| }, | |
| { | |
| "epoch": 88.7, | |
| "learning_rate": 8.172839506172841e-06, | |
| "loss": 0.0004, | |
| "step": 4790 | |
| }, | |
| { | |
| "epoch": 88.89, | |
| "learning_rate": 8.148148148148148e-06, | |
| "loss": 0.0004, | |
| "step": 4800 | |
| }, | |
| { | |
| "epoch": 89.0, | |
| "eval_accuracy": 1.0, | |
| "eval_loss": 0.0004430774424690753, | |
| "eval_runtime": 3.0165, | |
| "eval_samples_per_second": 25.195, | |
| "eval_steps_per_second": 3.315, | |
| "step": 4806 | |
| }, | |
| { | |
| "epoch": 89.07, | |
| "learning_rate": 8.123456790123457e-06, | |
| "loss": 0.0004, | |
| "step": 4810 | |
| }, | |
| { | |
| "epoch": 89.26, | |
| "learning_rate": 8.098765432098766e-06, | |
| "loss": 0.0004, | |
| "step": 4820 | |
| }, | |
| { | |
| "epoch": 89.44, | |
| "learning_rate": 8.074074074074075e-06, | |
| "loss": 0.0609, | |
| "step": 4830 | |
| }, | |
| { | |
| "epoch": 89.63, | |
| "learning_rate": 8.049382716049384e-06, | |
| "loss": 0.0007, | |
| "step": 4840 | |
| }, | |
| { | |
| "epoch": 89.81, | |
| "learning_rate": 8.024691358024692e-06, | |
| "loss": 0.0004, | |
| "step": 4850 | |
| }, | |
| { | |
| "epoch": 90.0, | |
| "learning_rate": 8.000000000000001e-06, | |
| "loss": 0.0005, | |
| "step": 4860 | |
| }, | |
| { | |
| "epoch": 90.0, | |
| "eval_accuracy": 1.0, | |
| "eval_loss": 0.00045739198685623705, | |
| "eval_runtime": 3.0947, | |
| "eval_samples_per_second": 24.558, | |
| "eval_steps_per_second": 3.231, | |
| "step": 4860 | |
| }, | |
| { | |
| "epoch": 90.19, | |
| "learning_rate": 7.975308641975308e-06, | |
| "loss": 0.0004, | |
| "step": 4870 | |
| }, | |
| { | |
| "epoch": 90.37, | |
| "learning_rate": 7.950617283950617e-06, | |
| "loss": 0.0564, | |
| "step": 4880 | |
| }, | |
| { | |
| "epoch": 90.56, | |
| "learning_rate": 7.925925925925926e-06, | |
| "loss": 0.0004, | |
| "step": 4890 | |
| }, | |
| { | |
| "epoch": 90.74, | |
| "learning_rate": 7.901234567901235e-06, | |
| "loss": 0.0004, | |
| "step": 4900 | |
| }, | |
| { | |
| "epoch": 90.93, | |
| "learning_rate": 7.876543209876544e-06, | |
| "loss": 0.0004, | |
| "step": 4910 | |
| }, | |
| { | |
| "epoch": 91.0, | |
| "eval_accuracy": 1.0, | |
| "eval_loss": 0.0004953287425450981, | |
| "eval_runtime": 3.047, | |
| "eval_samples_per_second": 24.942, | |
| "eval_steps_per_second": 3.282, | |
| "step": 4914 | |
| }, | |
| { | |
| "epoch": 91.11, | |
| "learning_rate": 7.851851851851853e-06, | |
| "loss": 0.0009, | |
| "step": 4920 | |
| }, | |
| { | |
| "epoch": 91.3, | |
| "learning_rate": 7.82716049382716e-06, | |
| "loss": 0.0004, | |
| "step": 4930 | |
| }, | |
| { | |
| "epoch": 91.48, | |
| "learning_rate": 7.80246913580247e-06, | |
| "loss": 0.0004, | |
| "step": 4940 | |
| }, | |
| { | |
| "epoch": 91.67, | |
| "learning_rate": 7.77777777777778e-06, | |
| "loss": 0.0004, | |
| "step": 4950 | |
| }, | |
| { | |
| "epoch": 91.85, | |
| "learning_rate": 7.753086419753088e-06, | |
| "loss": 0.0067, | |
| "step": 4960 | |
| }, | |
| { | |
| "epoch": 92.0, | |
| "eval_accuracy": 1.0, | |
| "eval_loss": 0.00039551357622258365, | |
| "eval_runtime": 2.9279, | |
| "eval_samples_per_second": 25.957, | |
| "eval_steps_per_second": 3.415, | |
| "step": 4968 | |
| }, | |
| { | |
| "epoch": 92.04, | |
| "learning_rate": 7.728395061728395e-06, | |
| "loss": 0.0008, | |
| "step": 4970 | |
| }, | |
| { | |
| "epoch": 92.22, | |
| "learning_rate": 7.703703703703704e-06, | |
| "loss": 0.0004, | |
| "step": 4980 | |
| }, | |
| { | |
| "epoch": 92.41, | |
| "learning_rate": 7.679012345679013e-06, | |
| "loss": 0.0004, | |
| "step": 4990 | |
| }, | |
| { | |
| "epoch": 92.59, | |
| "learning_rate": 7.654320987654322e-06, | |
| "loss": 0.0004, | |
| "step": 5000 | |
| }, | |
| { | |
| "epoch": 92.78, | |
| "learning_rate": 7.62962962962963e-06, | |
| "loss": 0.0004, | |
| "step": 5010 | |
| }, | |
| { | |
| "epoch": 92.96, | |
| "learning_rate": 7.604938271604939e-06, | |
| "loss": 0.0004, | |
| "step": 5020 | |
| }, | |
| { | |
| "epoch": 93.0, | |
| "eval_accuracy": 1.0, | |
| "eval_loss": 0.0003839218115899712, | |
| "eval_runtime": 3.0585, | |
| "eval_samples_per_second": 24.849, | |
| "eval_steps_per_second": 3.27, | |
| "step": 5022 | |
| }, | |
| { | |
| "epoch": 93.15, | |
| "learning_rate": 7.580246913580247e-06, | |
| "loss": 0.0004, | |
| "step": 5030 | |
| }, | |
| { | |
| "epoch": 93.33, | |
| "learning_rate": 7.555555555555556e-06, | |
| "loss": 0.0433, | |
| "step": 5040 | |
| }, | |
| { | |
| "epoch": 93.52, | |
| "learning_rate": 7.530864197530865e-06, | |
| "loss": 0.0092, | |
| "step": 5050 | |
| }, | |
| { | |
| "epoch": 93.7, | |
| "learning_rate": 7.506172839506174e-06, | |
| "loss": 0.0004, | |
| "step": 5060 | |
| }, | |
| { | |
| "epoch": 93.89, | |
| "learning_rate": 7.481481481481482e-06, | |
| "loss": 0.0004, | |
| "step": 5070 | |
| }, | |
| { | |
| "epoch": 94.0, | |
| "eval_accuracy": 1.0, | |
| "eval_loss": 0.0003877544077113271, | |
| "eval_runtime": 3.07, | |
| "eval_samples_per_second": 24.756, | |
| "eval_steps_per_second": 3.257, | |
| "step": 5076 | |
| }, | |
| { | |
| "epoch": 94.07, | |
| "learning_rate": 7.456790123456791e-06, | |
| "loss": 0.0004, | |
| "step": 5080 | |
| }, | |
| { | |
| "epoch": 94.26, | |
| "learning_rate": 7.4320987654320995e-06, | |
| "loss": 0.0004, | |
| "step": 5090 | |
| }, | |
| { | |
| "epoch": 94.44, | |
| "learning_rate": 7.4074074074074075e-06, | |
| "loss": 0.0004, | |
| "step": 5100 | |
| }, | |
| { | |
| "epoch": 94.63, | |
| "learning_rate": 7.382716049382716e-06, | |
| "loss": 0.0004, | |
| "step": 5110 | |
| }, | |
| { | |
| "epoch": 94.81, | |
| "learning_rate": 7.358024691358025e-06, | |
| "loss": 0.0035, | |
| "step": 5120 | |
| }, | |
| { | |
| "epoch": 95.0, | |
| "learning_rate": 7.333333333333333e-06, | |
| "loss": 0.0004, | |
| "step": 5130 | |
| }, | |
| { | |
| "epoch": 95.0, | |
| "eval_accuracy": 1.0, | |
| "eval_loss": 0.000382541969884187, | |
| "eval_runtime": 3.0815, | |
| "eval_samples_per_second": 24.663, | |
| "eval_steps_per_second": 3.245, | |
| "step": 5130 | |
| }, | |
| { | |
| "epoch": 95.19, | |
| "learning_rate": 7.308641975308642e-06, | |
| "loss": 0.0004, | |
| "step": 5140 | |
| }, | |
| { | |
| "epoch": 95.37, | |
| "learning_rate": 7.283950617283952e-06, | |
| "loss": 0.0004, | |
| "step": 5150 | |
| }, | |
| { | |
| "epoch": 95.56, | |
| "learning_rate": 7.2592592592592605e-06, | |
| "loss": 0.0004, | |
| "step": 5160 | |
| }, | |
| { | |
| "epoch": 95.74, | |
| "learning_rate": 7.2345679012345685e-06, | |
| "loss": 0.0004, | |
| "step": 5170 | |
| }, | |
| { | |
| "epoch": 95.93, | |
| "learning_rate": 7.209876543209877e-06, | |
| "loss": 0.0004, | |
| "step": 5180 | |
| }, | |
| { | |
| "epoch": 96.0, | |
| "eval_accuracy": 1.0, | |
| "eval_loss": 0.00039759272476658225, | |
| "eval_runtime": 3.0665, | |
| "eval_samples_per_second": 24.784, | |
| "eval_steps_per_second": 3.261, | |
| "step": 5184 | |
| }, | |
| { | |
| "epoch": 96.11, | |
| "learning_rate": 7.185185185185186e-06, | |
| "loss": 0.0004, | |
| "step": 5190 | |
| }, | |
| { | |
| "epoch": 96.3, | |
| "learning_rate": 7.160493827160494e-06, | |
| "loss": 0.0004, | |
| "step": 5200 | |
| }, | |
| { | |
| "epoch": 96.48, | |
| "learning_rate": 7.135802469135803e-06, | |
| "loss": 0.0004, | |
| "step": 5210 | |
| }, | |
| { | |
| "epoch": 96.67, | |
| "learning_rate": 7.111111111111112e-06, | |
| "loss": 0.0003, | |
| "step": 5220 | |
| }, | |
| { | |
| "epoch": 96.85, | |
| "learning_rate": 7.08641975308642e-06, | |
| "loss": 0.0004, | |
| "step": 5230 | |
| }, | |
| { | |
| "epoch": 97.0, | |
| "eval_accuracy": 1.0, | |
| "eval_loss": 0.00038783540367148817, | |
| "eval_runtime": 3.0953, | |
| "eval_samples_per_second": 24.553, | |
| "eval_steps_per_second": 3.231, | |
| "step": 5238 | |
| }, | |
| { | |
| "epoch": 97.04, | |
| "learning_rate": 7.061728395061729e-06, | |
| "loss": 0.0003, | |
| "step": 5240 | |
| }, | |
| { | |
| "epoch": 97.22, | |
| "learning_rate": 7.0370370370370375e-06, | |
| "loss": 0.0003, | |
| "step": 5250 | |
| }, | |
| { | |
| "epoch": 97.41, | |
| "learning_rate": 7.012345679012347e-06, | |
| "loss": 0.0004, | |
| "step": 5260 | |
| }, | |
| { | |
| "epoch": 97.59, | |
| "learning_rate": 6.987654320987654e-06, | |
| "loss": 0.0004, | |
| "step": 5270 | |
| }, | |
| { | |
| "epoch": 97.78, | |
| "learning_rate": 6.962962962962964e-06, | |
| "loss": 0.0004, | |
| "step": 5280 | |
| }, | |
| { | |
| "epoch": 97.96, | |
| "learning_rate": 6.938271604938273e-06, | |
| "loss": 0.0004, | |
| "step": 5290 | |
| }, | |
| { | |
| "epoch": 98.0, | |
| "eval_accuracy": 1.0, | |
| "eval_loss": 0.0003907608625013381, | |
| "eval_runtime": 3.0948, | |
| "eval_samples_per_second": 24.557, | |
| "eval_steps_per_second": 3.231, | |
| "step": 5292 | |
| }, | |
| { | |
| "epoch": 98.15, | |
| "learning_rate": 6.913580246913581e-06, | |
| "loss": 0.0003, | |
| "step": 5300 | |
| }, | |
| { | |
| "epoch": 98.33, | |
| "learning_rate": 6.88888888888889e-06, | |
| "loss": 0.0003, | |
| "step": 5310 | |
| }, | |
| { | |
| "epoch": 98.52, | |
| "learning_rate": 6.8641975308641985e-06, | |
| "loss": 0.0004, | |
| "step": 5320 | |
| }, | |
| { | |
| "epoch": 98.7, | |
| "learning_rate": 6.8395061728395065e-06, | |
| "loss": 0.0003, | |
| "step": 5330 | |
| }, | |
| { | |
| "epoch": 98.89, | |
| "learning_rate": 6.814814814814815e-06, | |
| "loss": 0.0003, | |
| "step": 5340 | |
| }, | |
| { | |
| "epoch": 99.0, | |
| "eval_accuracy": 1.0, | |
| "eval_loss": 0.0004004727234132588, | |
| "eval_runtime": 3.0696, | |
| "eval_samples_per_second": 24.759, | |
| "eval_steps_per_second": 3.258, | |
| "step": 5346 | |
| }, | |
| { | |
| "epoch": 99.07, | |
| "learning_rate": 6.790123456790124e-06, | |
| "loss": 0.0004, | |
| "step": 5350 | |
| }, | |
| { | |
| "epoch": 99.26, | |
| "learning_rate": 6.765432098765433e-06, | |
| "loss": 0.0003, | |
| "step": 5360 | |
| }, | |
| { | |
| "epoch": 99.44, | |
| "learning_rate": 6.740740740740741e-06, | |
| "loss": 0.0003, | |
| "step": 5370 | |
| }, | |
| { | |
| "epoch": 99.63, | |
| "learning_rate": 6.71604938271605e-06, | |
| "loss": 0.0003, | |
| "step": 5380 | |
| }, | |
| { | |
| "epoch": 99.81, | |
| "learning_rate": 6.691358024691359e-06, | |
| "loss": 0.0017, | |
| "step": 5390 | |
| }, | |
| { | |
| "epoch": 100.0, | |
| "learning_rate": 6.666666666666667e-06, | |
| "loss": 0.0003, | |
| "step": 5400 | |
| }, | |
| { | |
| "epoch": 100.0, | |
| "eval_accuracy": 1.0, | |
| "eval_loss": 0.00034208386205136776, | |
| "eval_runtime": 3.0385, | |
| "eval_samples_per_second": 25.012, | |
| "eval_steps_per_second": 3.291, | |
| "step": 5400 | |
| }, | |
| { | |
| "epoch": 100.19, | |
| "learning_rate": 6.6419753086419755e-06, | |
| "loss": 0.0003, | |
| "step": 5410 | |
| }, | |
| { | |
| "epoch": 100.37, | |
| "learning_rate": 6.617283950617285e-06, | |
| "loss": 0.0003, | |
| "step": 5420 | |
| }, | |
| { | |
| "epoch": 100.56, | |
| "learning_rate": 6.592592592592592e-06, | |
| "loss": 0.0003, | |
| "step": 5430 | |
| }, | |
| { | |
| "epoch": 100.74, | |
| "learning_rate": 6.567901234567902e-06, | |
| "loss": 0.0034, | |
| "step": 5440 | |
| }, | |
| { | |
| "epoch": 100.93, | |
| "learning_rate": 6.543209876543211e-06, | |
| "loss": 0.0003, | |
| "step": 5450 | |
| }, | |
| { | |
| "epoch": 101.0, | |
| "eval_accuracy": 1.0, | |
| "eval_loss": 0.00037925393553450704, | |
| "eval_runtime": 2.9647, | |
| "eval_samples_per_second": 25.635, | |
| "eval_steps_per_second": 3.373, | |
| "step": 5454 | |
| }, | |
| { | |
| "epoch": 101.11, | |
| "learning_rate": 6.51851851851852e-06, | |
| "loss": 0.0003, | |
| "step": 5460 | |
| }, | |
| { | |
| "epoch": 101.3, | |
| "learning_rate": 6.493827160493828e-06, | |
| "loss": 0.0922, | |
| "step": 5470 | |
| }, | |
| { | |
| "epoch": 101.48, | |
| "learning_rate": 6.4691358024691365e-06, | |
| "loss": 0.0003, | |
| "step": 5480 | |
| }, | |
| { | |
| "epoch": 101.67, | |
| "learning_rate": 6.444444444444445e-06, | |
| "loss": 0.0003, | |
| "step": 5490 | |
| }, | |
| { | |
| "epoch": 101.85, | |
| "learning_rate": 6.419753086419753e-06, | |
| "loss": 0.0004, | |
| "step": 5500 | |
| }, | |
| { | |
| "epoch": 102.0, | |
| "eval_accuracy": 1.0, | |
| "eval_loss": 0.00046697567449882627, | |
| "eval_runtime": 2.9966, | |
| "eval_samples_per_second": 25.362, | |
| "eval_steps_per_second": 3.337, | |
| "step": 5508 | |
| }, | |
| { | |
| "epoch": 102.04, | |
| "learning_rate": 6.395061728395062e-06, | |
| "loss": 0.0004, | |
| "step": 5510 | |
| }, | |
| { | |
| "epoch": 102.22, | |
| "learning_rate": 6.370370370370371e-06, | |
| "loss": 0.0003, | |
| "step": 5520 | |
| }, | |
| { | |
| "epoch": 102.41, | |
| "learning_rate": 6.345679012345679e-06, | |
| "loss": 0.0003, | |
| "step": 5530 | |
| }, | |
| { | |
| "epoch": 102.59, | |
| "learning_rate": 6.320987654320988e-06, | |
| "loss": 0.0003, | |
| "step": 5540 | |
| }, | |
| { | |
| "epoch": 102.78, | |
| "learning_rate": 6.296296296296297e-06, | |
| "loss": 0.0003, | |
| "step": 5550 | |
| }, | |
| { | |
| "epoch": 102.96, | |
| "learning_rate": 6.271604938271606e-06, | |
| "loss": 0.0004, | |
| "step": 5560 | |
| }, | |
| { | |
| "epoch": 103.0, | |
| "eval_accuracy": 1.0, | |
| "eval_loss": 0.00045954357483424246, | |
| "eval_runtime": 2.9804, | |
| "eval_samples_per_second": 25.5, | |
| "eval_steps_per_second": 3.355, | |
| "step": 5562 | |
| }, | |
| { | |
| "epoch": 103.15, | |
| "learning_rate": 6.2469135802469135e-06, | |
| "loss": 0.0003, | |
| "step": 5570 | |
| }, | |
| { | |
| "epoch": 103.33, | |
| "learning_rate": 6.222222222222223e-06, | |
| "loss": 0.0003, | |
| "step": 5580 | |
| }, | |
| { | |
| "epoch": 103.52, | |
| "learning_rate": 6.197530864197532e-06, | |
| "loss": 0.0003, | |
| "step": 5590 | |
| }, | |
| { | |
| "epoch": 103.7, | |
| "learning_rate": 6.17283950617284e-06, | |
| "loss": 0.0003, | |
| "step": 5600 | |
| }, | |
| { | |
| "epoch": 103.89, | |
| "learning_rate": 6.148148148148149e-06, | |
| "loss": 0.0004, | |
| "step": 5610 | |
| }, | |
| { | |
| "epoch": 104.0, | |
| "eval_accuracy": 1.0, | |
| "eval_loss": 0.0004455417220015079, | |
| "eval_runtime": 3.0351, | |
| "eval_samples_per_second": 25.04, | |
| "eval_steps_per_second": 3.295, | |
| "step": 5616 | |
| }, | |
| { | |
| "epoch": 104.07, | |
| "learning_rate": 6.123456790123458e-06, | |
| "loss": 0.0003, | |
| "step": 5620 | |
| }, | |
| { | |
| "epoch": 104.26, | |
| "learning_rate": 6.098765432098766e-06, | |
| "loss": 0.0003, | |
| "step": 5630 | |
| }, | |
| { | |
| "epoch": 104.44, | |
| "learning_rate": 6.0740740740740745e-06, | |
| "loss": 0.0241, | |
| "step": 5640 | |
| }, | |
| { | |
| "epoch": 104.63, | |
| "learning_rate": 6.049382716049383e-06, | |
| "loss": 0.0003, | |
| "step": 5650 | |
| }, | |
| { | |
| "epoch": 104.81, | |
| "learning_rate": 6.024691358024692e-06, | |
| "loss": 0.0003, | |
| "step": 5660 | |
| }, | |
| { | |
| "epoch": 105.0, | |
| "learning_rate": 6e-06, | |
| "loss": 0.0006, | |
| "step": 5670 | |
| }, | |
| { | |
| "epoch": 105.0, | |
| "eval_accuracy": 1.0, | |
| "eval_loss": 0.0003152435238007456, | |
| "eval_runtime": 3.0259, | |
| "eval_samples_per_second": 25.117, | |
| "eval_steps_per_second": 3.305, | |
| "step": 5670 | |
| }, | |
| { | |
| "epoch": 105.19, | |
| "learning_rate": 5.975308641975309e-06, | |
| "loss": 0.0003, | |
| "step": 5680 | |
| }, | |
| { | |
| "epoch": 105.37, | |
| "learning_rate": 5.950617283950618e-06, | |
| "loss": 0.0003, | |
| "step": 5690 | |
| }, | |
| { | |
| "epoch": 105.56, | |
| "learning_rate": 5.925925925925926e-06, | |
| "loss": 0.0003, | |
| "step": 5700 | |
| }, | |
| { | |
| "epoch": 105.74, | |
| "learning_rate": 5.901234567901235e-06, | |
| "loss": 0.0003, | |
| "step": 5710 | |
| }, | |
| { | |
| "epoch": 105.93, | |
| "learning_rate": 5.876543209876544e-06, | |
| "loss": 0.0005, | |
| "step": 5720 | |
| }, | |
| { | |
| "epoch": 106.0, | |
| "eval_accuracy": 1.0, | |
| "eval_loss": 0.00031142152147367597, | |
| "eval_runtime": 3.045, | |
| "eval_samples_per_second": 24.959, | |
| "eval_steps_per_second": 3.284, | |
| "step": 5724 | |
| }, | |
| { | |
| "epoch": 106.11, | |
| "learning_rate": 5.8518518518518515e-06, | |
| "loss": 0.0003, | |
| "step": 5730 | |
| }, | |
| { | |
| "epoch": 106.3, | |
| "learning_rate": 5.827160493827161e-06, | |
| "loss": 0.0003, | |
| "step": 5740 | |
| }, | |
| { | |
| "epoch": 106.48, | |
| "learning_rate": 5.80246913580247e-06, | |
| "loss": 0.0003, | |
| "step": 5750 | |
| }, | |
| { | |
| "epoch": 106.67, | |
| "learning_rate": 5.777777777777778e-06, | |
| "loss": 0.0032, | |
| "step": 5760 | |
| }, | |
| { | |
| "epoch": 106.85, | |
| "learning_rate": 5.753086419753087e-06, | |
| "loss": 0.0003, | |
| "step": 5770 | |
| }, | |
| { | |
| "epoch": 107.0, | |
| "eval_accuracy": 1.0, | |
| "eval_loss": 0.0003071220126003027, | |
| "eval_runtime": 3.0131, | |
| "eval_samples_per_second": 25.224, | |
| "eval_steps_per_second": 3.319, | |
| "step": 5778 | |
| }, | |
| { | |
| "epoch": 107.04, | |
| "learning_rate": 5.728395061728396e-06, | |
| "loss": 0.0003, | |
| "step": 5780 | |
| }, | |
| { | |
| "epoch": 107.22, | |
| "learning_rate": 5.7037037037037045e-06, | |
| "loss": 0.0033, | |
| "step": 5790 | |
| }, | |
| { | |
| "epoch": 107.41, | |
| "learning_rate": 5.6790123456790125e-06, | |
| "loss": 0.0003, | |
| "step": 5800 | |
| }, | |
| { | |
| "epoch": 107.59, | |
| "learning_rate": 5.654320987654321e-06, | |
| "loss": 0.0003, | |
| "step": 5810 | |
| }, | |
| { | |
| "epoch": 107.78, | |
| "learning_rate": 5.62962962962963e-06, | |
| "loss": 0.0003, | |
| "step": 5820 | |
| }, | |
| { | |
| "epoch": 107.96, | |
| "learning_rate": 5.604938271604938e-06, | |
| "loss": 0.0003, | |
| "step": 5830 | |
| }, | |
| { | |
| "epoch": 108.0, | |
| "eval_accuracy": 1.0, | |
| "eval_loss": 0.0002943011641036719, | |
| "eval_runtime": 3.044, | |
| "eval_samples_per_second": 24.967, | |
| "eval_steps_per_second": 3.285, | |
| "step": 5832 | |
| }, | |
| { | |
| "epoch": 108.15, | |
| "learning_rate": 5.580246913580247e-06, | |
| "loss": 0.0003, | |
| "step": 5840 | |
| }, | |
| { | |
| "epoch": 108.33, | |
| "learning_rate": 5.555555555555557e-06, | |
| "loss": 0.0004, | |
| "step": 5850 | |
| }, | |
| { | |
| "epoch": 108.52, | |
| "learning_rate": 5.530864197530864e-06, | |
| "loss": 0.0003, | |
| "step": 5860 | |
| }, | |
| { | |
| "epoch": 108.7, | |
| "learning_rate": 5.5061728395061735e-06, | |
| "loss": 0.0003, | |
| "step": 5870 | |
| }, | |
| { | |
| "epoch": 108.89, | |
| "learning_rate": 5.481481481481482e-06, | |
| "loss": 0.0003, | |
| "step": 5880 | |
| }, | |
| { | |
| "epoch": 109.0, | |
| "eval_accuracy": 1.0, | |
| "eval_loss": 0.00028869544621557, | |
| "eval_runtime": 3.0963, | |
| "eval_samples_per_second": 24.546, | |
| "eval_steps_per_second": 3.23, | |
| "step": 5886 | |
| }, | |
| { | |
| "epoch": 109.07, | |
| "learning_rate": 5.456790123456791e-06, | |
| "loss": 0.0003, | |
| "step": 5890 | |
| }, | |
| { | |
| "epoch": 109.26, | |
| "learning_rate": 5.432098765432099e-06, | |
| "loss": 0.0003, | |
| "step": 5900 | |
| }, | |
| { | |
| "epoch": 109.44, | |
| "learning_rate": 5.407407407407408e-06, | |
| "loss": 0.0003, | |
| "step": 5910 | |
| }, | |
| { | |
| "epoch": 109.63, | |
| "learning_rate": 5.382716049382717e-06, | |
| "loss": 0.0003, | |
| "step": 5920 | |
| }, | |
| { | |
| "epoch": 109.81, | |
| "learning_rate": 5.358024691358025e-06, | |
| "loss": 0.0003, | |
| "step": 5930 | |
| }, | |
| { | |
| "epoch": 110.0, | |
| "learning_rate": 5.333333333333334e-06, | |
| "loss": 0.0003, | |
| "step": 5940 | |
| }, | |
| { | |
| "epoch": 110.0, | |
| "eval_accuracy": 1.0, | |
| "eval_loss": 0.0002840006782207638, | |
| "eval_runtime": 2.9775, | |
| "eval_samples_per_second": 25.525, | |
| "eval_steps_per_second": 3.359, | |
| "step": 5940 | |
| }, | |
| { | |
| "epoch": 110.19, | |
| "learning_rate": 5.3086419753086425e-06, | |
| "loss": 0.0004, | |
| "step": 5950 | |
| }, | |
| { | |
| "epoch": 110.37, | |
| "learning_rate": 5.2839506172839505e-06, | |
| "loss": 0.0003, | |
| "step": 5960 | |
| }, | |
| { | |
| "epoch": 110.56, | |
| "learning_rate": 5.259259259259259e-06, | |
| "loss": 0.0003, | |
| "step": 5970 | |
| }, | |
| { | |
| "epoch": 110.74, | |
| "learning_rate": 5.234567901234568e-06, | |
| "loss": 0.0003, | |
| "step": 5980 | |
| }, | |
| { | |
| "epoch": 110.93, | |
| "learning_rate": 5.209876543209878e-06, | |
| "loss": 0.0003, | |
| "step": 5990 | |
| }, | |
| { | |
| "epoch": 111.0, | |
| "eval_accuracy": 1.0, | |
| "eval_loss": 0.0002830446173902601, | |
| "eval_runtime": 3.079, | |
| "eval_samples_per_second": 24.683, | |
| "eval_steps_per_second": 3.248, | |
| "step": 5994 | |
| }, | |
| { | |
| "epoch": 111.11, | |
| "learning_rate": 5.185185185185185e-06, | |
| "loss": 0.0003, | |
| "step": 6000 | |
| }, | |
| { | |
| "epoch": 111.3, | |
| "learning_rate": 5.160493827160495e-06, | |
| "loss": 0.0003, | |
| "step": 6010 | |
| }, | |
| { | |
| "epoch": 111.48, | |
| "learning_rate": 5.1358024691358035e-06, | |
| "loss": 0.0016, | |
| "step": 6020 | |
| }, | |
| { | |
| "epoch": 111.67, | |
| "learning_rate": 5.1111111111111115e-06, | |
| "loss": 0.0072, | |
| "step": 6030 | |
| }, | |
| { | |
| "epoch": 111.85, | |
| "learning_rate": 5.08641975308642e-06, | |
| "loss": 0.0003, | |
| "step": 6040 | |
| }, | |
| { | |
| "epoch": 112.0, | |
| "eval_accuracy": 1.0, | |
| "eval_loss": 0.00027556964778341353, | |
| "eval_runtime": 3.0731, | |
| "eval_samples_per_second": 24.731, | |
| "eval_steps_per_second": 3.254, | |
| "step": 6048 | |
| }, | |
| { | |
| "epoch": 112.04, | |
| "learning_rate": 5.061728395061729e-06, | |
| "loss": 0.0005, | |
| "step": 6050 | |
| }, | |
| { | |
| "epoch": 112.22, | |
| "learning_rate": 5.037037037037037e-06, | |
| "loss": 0.0003, | |
| "step": 6060 | |
| }, | |
| { | |
| "epoch": 112.41, | |
| "learning_rate": 5.012345679012346e-06, | |
| "loss": 0.0003, | |
| "step": 6070 | |
| }, | |
| { | |
| "epoch": 112.59, | |
| "learning_rate": 4.987654320987655e-06, | |
| "loss": 0.0003, | |
| "step": 6080 | |
| }, | |
| { | |
| "epoch": 112.78, | |
| "learning_rate": 4.962962962962964e-06, | |
| "loss": 0.0003, | |
| "step": 6090 | |
| }, | |
| { | |
| "epoch": 112.96, | |
| "learning_rate": 4.938271604938272e-06, | |
| "loss": 0.0003, | |
| "step": 6100 | |
| }, | |
| { | |
| "epoch": 113.0, | |
| "eval_accuracy": 1.0, | |
| "eval_loss": 0.00026999213150702417, | |
| "eval_runtime": 3.0277, | |
| "eval_samples_per_second": 25.102, | |
| "eval_steps_per_second": 3.303, | |
| "step": 6102 | |
| }, | |
| { | |
| "epoch": 113.15, | |
| "learning_rate": 4.9135802469135805e-06, | |
| "loss": 0.0003, | |
| "step": 6110 | |
| }, | |
| { | |
| "epoch": 113.33, | |
| "learning_rate": 4.888888888888889e-06, | |
| "loss": 0.0003, | |
| "step": 6120 | |
| }, | |
| { | |
| "epoch": 113.52, | |
| "learning_rate": 4.864197530864198e-06, | |
| "loss": 0.0003, | |
| "step": 6130 | |
| }, | |
| { | |
| "epoch": 113.7, | |
| "learning_rate": 4.839506172839506e-06, | |
| "loss": 0.0728, | |
| "step": 6140 | |
| }, | |
| { | |
| "epoch": 113.89, | |
| "learning_rate": 4.814814814814815e-06, | |
| "loss": 0.0003, | |
| "step": 6150 | |
| }, | |
| { | |
| "epoch": 114.0, | |
| "eval_accuracy": 1.0, | |
| "eval_loss": 0.00028843365726061165, | |
| "eval_runtime": 2.9883, | |
| "eval_samples_per_second": 25.433, | |
| "eval_steps_per_second": 3.346, | |
| "step": 6156 | |
| }, | |
| { | |
| "epoch": 114.07, | |
| "learning_rate": 4.790123456790124e-06, | |
| "loss": 0.0003, | |
| "step": 6160 | |
| }, | |
| { | |
| "epoch": 114.26, | |
| "learning_rate": 4.765432098765433e-06, | |
| "loss": 0.0003, | |
| "step": 6170 | |
| }, | |
| { | |
| "epoch": 114.44, | |
| "learning_rate": 4.7407407407407415e-06, | |
| "loss": 0.0003, | |
| "step": 6180 | |
| }, | |
| { | |
| "epoch": 114.63, | |
| "learning_rate": 4.7160493827160495e-06, | |
| "loss": 0.0003, | |
| "step": 6190 | |
| }, | |
| { | |
| "epoch": 114.81, | |
| "learning_rate": 4.691358024691358e-06, | |
| "loss": 0.0003, | |
| "step": 6200 | |
| }, | |
| { | |
| "epoch": 115.0, | |
| "learning_rate": 4.666666666666667e-06, | |
| "loss": 0.0003, | |
| "step": 6210 | |
| }, | |
| { | |
| "epoch": 115.0, | |
| "eval_accuracy": 1.0, | |
| "eval_loss": 0.0002904118155129254, | |
| "eval_runtime": 2.9961, | |
| "eval_samples_per_second": 25.367, | |
| "eval_steps_per_second": 3.338, | |
| "step": 6210 | |
| }, | |
| { | |
| "epoch": 115.19, | |
| "learning_rate": 4.641975308641975e-06, | |
| "loss": 0.0003, | |
| "step": 6220 | |
| }, | |
| { | |
| "epoch": 115.37, | |
| "learning_rate": 4.617283950617285e-06, | |
| "loss": 0.0003, | |
| "step": 6230 | |
| }, | |
| { | |
| "epoch": 115.56, | |
| "learning_rate": 4.592592592592593e-06, | |
| "loss": 0.0003, | |
| "step": 6240 | |
| }, | |
| { | |
| "epoch": 115.74, | |
| "learning_rate": 4.567901234567902e-06, | |
| "loss": 0.0003, | |
| "step": 6250 | |
| }, | |
| { | |
| "epoch": 115.93, | |
| "learning_rate": 4.5432098765432105e-06, | |
| "loss": 0.0003, | |
| "step": 6260 | |
| }, | |
| { | |
| "epoch": 116.0, | |
| "eval_accuracy": 1.0, | |
| "eval_loss": 0.0002848489675670862, | |
| "eval_runtime": 2.9812, | |
| "eval_samples_per_second": 25.493, | |
| "eval_steps_per_second": 3.354, | |
| "step": 6264 | |
| }, | |
| { | |
| "epoch": 116.11, | |
| "learning_rate": 4.5185185185185185e-06, | |
| "loss": 0.0003, | |
| "step": 6270 | |
| }, | |
| { | |
| "epoch": 116.3, | |
| "learning_rate": 4.493827160493827e-06, | |
| "loss": 0.0012, | |
| "step": 6280 | |
| }, | |
| { | |
| "epoch": 116.48, | |
| "learning_rate": 4.469135802469136e-06, | |
| "loss": 0.0003, | |
| "step": 6290 | |
| }, | |
| { | |
| "epoch": 116.67, | |
| "learning_rate": 4.444444444444444e-06, | |
| "loss": 0.0003, | |
| "step": 6300 | |
| }, | |
| { | |
| "epoch": 116.85, | |
| "learning_rate": 4.419753086419754e-06, | |
| "loss": 0.0003, | |
| "step": 6310 | |
| }, | |
| { | |
| "epoch": 117.0, | |
| "eval_accuracy": 1.0, | |
| "eval_loss": 0.00026769962278194726, | |
| "eval_runtime": 2.9936, | |
| "eval_samples_per_second": 25.388, | |
| "eval_steps_per_second": 3.34, | |
| "step": 6318 | |
| }, | |
| { | |
| "epoch": 117.04, | |
| "learning_rate": 4.395061728395062e-06, | |
| "loss": 0.0003, | |
| "step": 6320 | |
| }, | |
| { | |
| "epoch": 117.22, | |
| "learning_rate": 4.370370370370371e-06, | |
| "loss": 0.0002, | |
| "step": 6330 | |
| }, | |
| { | |
| "epoch": 117.41, | |
| "learning_rate": 4.3456790123456795e-06, | |
| "loss": 0.0003, | |
| "step": 6340 | |
| }, | |
| { | |
| "epoch": 117.59, | |
| "learning_rate": 4.3209876543209875e-06, | |
| "loss": 0.0002, | |
| "step": 6350 | |
| }, | |
| { | |
| "epoch": 117.78, | |
| "learning_rate": 4.296296296296296e-06, | |
| "loss": 0.0003, | |
| "step": 6360 | |
| }, | |
| { | |
| "epoch": 117.96, | |
| "learning_rate": 4.271604938271605e-06, | |
| "loss": 0.0003, | |
| "step": 6370 | |
| }, | |
| { | |
| "epoch": 118.0, | |
| "eval_accuracy": 1.0, | |
| "eval_loss": 0.00026247106143273413, | |
| "eval_runtime": 3.0703, | |
| "eval_samples_per_second": 24.754, | |
| "eval_steps_per_second": 3.257, | |
| "step": 6372 | |
| }, | |
| { | |
| "epoch": 118.15, | |
| "learning_rate": 4.246913580246914e-06, | |
| "loss": 0.0002, | |
| "step": 6380 | |
| }, | |
| { | |
| "epoch": 118.33, | |
| "learning_rate": 4.222222222222223e-06, | |
| "loss": 0.0493, | |
| "step": 6390 | |
| }, | |
| { | |
| "epoch": 118.52, | |
| "learning_rate": 4.197530864197531e-06, | |
| "loss": 0.0003, | |
| "step": 6400 | |
| }, | |
| { | |
| "epoch": 118.7, | |
| "learning_rate": 4.17283950617284e-06, | |
| "loss": 0.0003, | |
| "step": 6410 | |
| }, | |
| { | |
| "epoch": 118.89, | |
| "learning_rate": 4.1481481481481485e-06, | |
| "loss": 0.0002, | |
| "step": 6420 | |
| }, | |
| { | |
| "epoch": 119.0, | |
| "eval_accuracy": 1.0, | |
| "eval_loss": 0.00024906551698222756, | |
| "eval_runtime": 3.0184, | |
| "eval_samples_per_second": 25.179, | |
| "eval_steps_per_second": 3.313, | |
| "step": 6426 | |
| }, | |
| { | |
| "epoch": 119.07, | |
| "learning_rate": 4.123456790123457e-06, | |
| "loss": 0.0003, | |
| "step": 6430 | |
| }, | |
| { | |
| "epoch": 119.26, | |
| "learning_rate": 4.098765432098766e-06, | |
| "loss": 0.0804, | |
| "step": 6440 | |
| }, | |
| { | |
| "epoch": 119.44, | |
| "learning_rate": 4.074074074074074e-06, | |
| "loss": 0.0002, | |
| "step": 6450 | |
| }, | |
| { | |
| "epoch": 119.63, | |
| "learning_rate": 4.049382716049383e-06, | |
| "loss": 0.0005, | |
| "step": 6460 | |
| }, | |
| { | |
| "epoch": 119.81, | |
| "learning_rate": 4.024691358024692e-06, | |
| "loss": 0.0003, | |
| "step": 6470 | |
| }, | |
| { | |
| "epoch": 120.0, | |
| "learning_rate": 4.000000000000001e-06, | |
| "loss": 0.0002, | |
| "step": 6480 | |
| }, | |
| { | |
| "epoch": 120.0, | |
| "eval_accuracy": 1.0, | |
| "eval_loss": 0.0002486244193278253, | |
| "eval_runtime": 2.9769, | |
| "eval_samples_per_second": 25.53, | |
| "eval_steps_per_second": 3.359, | |
| "step": 6480 | |
| }, | |
| { | |
| "epoch": 120.19, | |
| "learning_rate": 3.975308641975309e-06, | |
| "loss": 0.0002, | |
| "step": 6490 | |
| }, | |
| { | |
| "epoch": 120.37, | |
| "learning_rate": 3.9506172839506175e-06, | |
| "loss": 0.0248, | |
| "step": 6500 | |
| }, | |
| { | |
| "epoch": 120.56, | |
| "learning_rate": 3.925925925925926e-06, | |
| "loss": 0.0018, | |
| "step": 6510 | |
| }, | |
| { | |
| "epoch": 120.74, | |
| "learning_rate": 3.901234567901235e-06, | |
| "loss": 0.0002, | |
| "step": 6520 | |
| }, | |
| { | |
| "epoch": 120.93, | |
| "learning_rate": 3.876543209876544e-06, | |
| "loss": 0.0002, | |
| "step": 6530 | |
| }, | |
| { | |
| "epoch": 121.0, | |
| "eval_accuracy": 1.0, | |
| "eval_loss": 0.00024542087339796126, | |
| "eval_runtime": 3.0455, | |
| "eval_samples_per_second": 24.955, | |
| "eval_steps_per_second": 3.283, | |
| "step": 6534 | |
| }, | |
| { | |
| "epoch": 121.11, | |
| "learning_rate": 3.851851851851852e-06, | |
| "loss": 0.0003, | |
| "step": 6540 | |
| }, | |
| { | |
| "epoch": 121.3, | |
| "learning_rate": 3.827160493827161e-06, | |
| "loss": 0.0002, | |
| "step": 6550 | |
| }, | |
| { | |
| "epoch": 121.48, | |
| "learning_rate": 3.8024691358024697e-06, | |
| "loss": 0.0002, | |
| "step": 6560 | |
| }, | |
| { | |
| "epoch": 121.67, | |
| "learning_rate": 3.777777777777778e-06, | |
| "loss": 0.0003, | |
| "step": 6570 | |
| }, | |
| { | |
| "epoch": 121.85, | |
| "learning_rate": 3.753086419753087e-06, | |
| "loss": 0.0003, | |
| "step": 6580 | |
| }, | |
| { | |
| "epoch": 122.0, | |
| "eval_accuracy": 1.0, | |
| "eval_loss": 0.0002419725788058713, | |
| "eval_runtime": 3.1297, | |
| "eval_samples_per_second": 24.283, | |
| "eval_steps_per_second": 3.195, | |
| "step": 6588 | |
| }, | |
| { | |
| "epoch": 122.04, | |
| "learning_rate": 3.7283950617283953e-06, | |
| "loss": 0.0002, | |
| "step": 6590 | |
| }, | |
| { | |
| "epoch": 122.22, | |
| "learning_rate": 3.7037037037037037e-06, | |
| "loss": 0.0004, | |
| "step": 6600 | |
| }, | |
| { | |
| "epoch": 122.41, | |
| "learning_rate": 3.6790123456790126e-06, | |
| "loss": 0.0002, | |
| "step": 6610 | |
| }, | |
| { | |
| "epoch": 122.59, | |
| "learning_rate": 3.654320987654321e-06, | |
| "loss": 0.0002, | |
| "step": 6620 | |
| }, | |
| { | |
| "epoch": 122.78, | |
| "learning_rate": 3.6296296296296302e-06, | |
| "loss": 0.0221, | |
| "step": 6630 | |
| }, | |
| { | |
| "epoch": 122.96, | |
| "learning_rate": 3.6049382716049387e-06, | |
| "loss": 0.0002, | |
| "step": 6640 | |
| }, | |
| { | |
| "epoch": 123.0, | |
| "eval_accuracy": 1.0, | |
| "eval_loss": 0.00023867662821430713, | |
| "eval_runtime": 3.0953, | |
| "eval_samples_per_second": 24.553, | |
| "eval_steps_per_second": 3.231, | |
| "step": 6642 | |
| }, | |
| { | |
| "epoch": 123.15, | |
| "learning_rate": 3.580246913580247e-06, | |
| "loss": 0.0002, | |
| "step": 6650 | |
| }, | |
| { | |
| "epoch": 123.33, | |
| "learning_rate": 3.555555555555556e-06, | |
| "loss": 0.0002, | |
| "step": 6660 | |
| }, | |
| { | |
| "epoch": 123.52, | |
| "learning_rate": 3.5308641975308643e-06, | |
| "loss": 0.0002, | |
| "step": 6670 | |
| }, | |
| { | |
| "epoch": 123.7, | |
| "learning_rate": 3.5061728395061736e-06, | |
| "loss": 0.0002, | |
| "step": 6680 | |
| }, | |
| { | |
| "epoch": 123.89, | |
| "learning_rate": 3.481481481481482e-06, | |
| "loss": 0.0002, | |
| "step": 6690 | |
| }, | |
| { | |
| "epoch": 124.0, | |
| "eval_accuracy": 1.0, | |
| "eval_loss": 0.00023590722412336618, | |
| "eval_runtime": 3.1057, | |
| "eval_samples_per_second": 24.471, | |
| "eval_steps_per_second": 3.22, | |
| "step": 6696 | |
| }, | |
| { | |
| "epoch": 124.07, | |
| "learning_rate": 3.4567901234567904e-06, | |
| "loss": 0.0002, | |
| "step": 6700 | |
| }, | |
| { | |
| "epoch": 124.26, | |
| "learning_rate": 3.4320987654320992e-06, | |
| "loss": 0.0002, | |
| "step": 6710 | |
| }, | |
| { | |
| "epoch": 124.44, | |
| "learning_rate": 3.4074074074074077e-06, | |
| "loss": 0.0014, | |
| "step": 6720 | |
| }, | |
| { | |
| "epoch": 124.63, | |
| "learning_rate": 3.3827160493827165e-06, | |
| "loss": 0.0003, | |
| "step": 6730 | |
| }, | |
| { | |
| "epoch": 124.81, | |
| "learning_rate": 3.358024691358025e-06, | |
| "loss": 0.0002, | |
| "step": 6740 | |
| }, | |
| { | |
| "epoch": 125.0, | |
| "learning_rate": 3.3333333333333333e-06, | |
| "loss": 0.0002, | |
| "step": 6750 | |
| }, | |
| { | |
| "epoch": 125.0, | |
| "eval_accuracy": 1.0, | |
| "eval_loss": 0.00023197440896183252, | |
| "eval_runtime": 3.1053, | |
| "eval_samples_per_second": 24.474, | |
| "eval_steps_per_second": 3.22, | |
| "step": 6750 | |
| }, | |
| { | |
| "epoch": 125.19, | |
| "learning_rate": 3.3086419753086426e-06, | |
| "loss": 0.0002, | |
| "step": 6760 | |
| }, | |
| { | |
| "epoch": 125.37, | |
| "learning_rate": 3.283950617283951e-06, | |
| "loss": 0.0002, | |
| "step": 6770 | |
| }, | |
| { | |
| "epoch": 125.56, | |
| "learning_rate": 3.25925925925926e-06, | |
| "loss": 0.0002, | |
| "step": 6780 | |
| }, | |
| { | |
| "epoch": 125.74, | |
| "learning_rate": 3.2345679012345682e-06, | |
| "loss": 0.0003, | |
| "step": 6790 | |
| }, | |
| { | |
| "epoch": 125.93, | |
| "learning_rate": 3.2098765432098767e-06, | |
| "loss": 0.0002, | |
| "step": 6800 | |
| }, | |
| { | |
| "epoch": 126.0, | |
| "eval_accuracy": 1.0, | |
| "eval_loss": 0.0002292316930834204, | |
| "eval_runtime": 3.1458, | |
| "eval_samples_per_second": 24.159, | |
| "eval_steps_per_second": 3.179, | |
| "step": 6804 | |
| }, | |
| { | |
| "epoch": 126.11, | |
| "learning_rate": 3.1851851851851855e-06, | |
| "loss": 0.0002, | |
| "step": 6810 | |
| }, | |
| { | |
| "epoch": 126.3, | |
| "learning_rate": 3.160493827160494e-06, | |
| "loss": 0.0002, | |
| "step": 6820 | |
| }, | |
| { | |
| "epoch": 126.48, | |
| "learning_rate": 3.135802469135803e-06, | |
| "loss": 0.0003, | |
| "step": 6830 | |
| }, | |
| { | |
| "epoch": 126.67, | |
| "learning_rate": 3.1111111111111116e-06, | |
| "loss": 0.0002, | |
| "step": 6840 | |
| }, | |
| { | |
| "epoch": 126.85, | |
| "learning_rate": 3.08641975308642e-06, | |
| "loss": 0.0712, | |
| "step": 6850 | |
| }, | |
| { | |
| "epoch": 127.0, | |
| "eval_accuracy": 1.0, | |
| "eval_loss": 0.00022830172383692116, | |
| "eval_runtime": 3.0547, | |
| "eval_samples_per_second": 24.879, | |
| "eval_steps_per_second": 3.274, | |
| "step": 6858 | |
| }, | |
| { | |
| "epoch": 127.04, | |
| "learning_rate": 3.061728395061729e-06, | |
| "loss": 0.0002, | |
| "step": 6860 | |
| }, | |
| { | |
| "epoch": 127.22, | |
| "learning_rate": 3.0370370370370372e-06, | |
| "loss": 0.0002, | |
| "step": 6870 | |
| }, | |
| { | |
| "epoch": 127.41, | |
| "learning_rate": 3.012345679012346e-06, | |
| "loss": 0.0002, | |
| "step": 6880 | |
| }, | |
| { | |
| "epoch": 127.59, | |
| "learning_rate": 2.9876543209876545e-06, | |
| "loss": 0.0002, | |
| "step": 6890 | |
| }, | |
| { | |
| "epoch": 127.78, | |
| "learning_rate": 2.962962962962963e-06, | |
| "loss": 0.0002, | |
| "step": 6900 | |
| }, | |
| { | |
| "epoch": 127.96, | |
| "learning_rate": 2.938271604938272e-06, | |
| "loss": 0.0002, | |
| "step": 6910 | |
| }, | |
| { | |
| "epoch": 128.0, | |
| "eval_accuracy": 1.0, | |
| "eval_loss": 0.00022887364320922643, | |
| "eval_runtime": 3.0745, | |
| "eval_samples_per_second": 24.72, | |
| "eval_steps_per_second": 3.253, | |
| "step": 6912 | |
| }, | |
| { | |
| "epoch": 128.15, | |
| "learning_rate": 2.9135802469135806e-06, | |
| "loss": 0.0002, | |
| "step": 6920 | |
| }, | |
| { | |
| "epoch": 128.33, | |
| "learning_rate": 2.888888888888889e-06, | |
| "loss": 0.0002, | |
| "step": 6930 | |
| }, | |
| { | |
| "epoch": 128.52, | |
| "learning_rate": 2.864197530864198e-06, | |
| "loss": 0.0002, | |
| "step": 6940 | |
| }, | |
| { | |
| "epoch": 128.7, | |
| "learning_rate": 2.8395061728395062e-06, | |
| "loss": 0.0002, | |
| "step": 6950 | |
| }, | |
| { | |
| "epoch": 128.89, | |
| "learning_rate": 2.814814814814815e-06, | |
| "loss": 0.0002, | |
| "step": 6960 | |
| }, | |
| { | |
| "epoch": 129.0, | |
| "eval_accuracy": 1.0, | |
| "eval_loss": 0.0002265323855681345, | |
| "eval_runtime": 3.0539, | |
| "eval_samples_per_second": 24.886, | |
| "eval_steps_per_second": 3.274, | |
| "step": 6966 | |
| }, | |
| { | |
| "epoch": 129.07, | |
| "learning_rate": 2.7901234567901235e-06, | |
| "loss": 0.0002, | |
| "step": 6970 | |
| }, | |
| { | |
| "epoch": 129.26, | |
| "learning_rate": 2.765432098765432e-06, | |
| "loss": 0.0002, | |
| "step": 6980 | |
| }, | |
| { | |
| "epoch": 129.44, | |
| "learning_rate": 2.740740740740741e-06, | |
| "loss": 0.0002, | |
| "step": 6990 | |
| }, | |
| { | |
| "epoch": 129.63, | |
| "learning_rate": 2.7160493827160496e-06, | |
| "loss": 0.0002, | |
| "step": 7000 | |
| }, | |
| { | |
| "epoch": 129.81, | |
| "learning_rate": 2.6913580246913584e-06, | |
| "loss": 0.0002, | |
| "step": 7010 | |
| }, | |
| { | |
| "epoch": 130.0, | |
| "learning_rate": 2.666666666666667e-06, | |
| "loss": 0.0002, | |
| "step": 7020 | |
| }, | |
| { | |
| "epoch": 130.0, | |
| "eval_accuracy": 1.0, | |
| "eval_loss": 0.00022419424203690141, | |
| "eval_runtime": 3.1051, | |
| "eval_samples_per_second": 24.476, | |
| "eval_steps_per_second": 3.221, | |
| "step": 7020 | |
| }, | |
| { | |
| "epoch": 130.19, | |
| "learning_rate": 2.6419753086419752e-06, | |
| "loss": 0.081, | |
| "step": 7030 | |
| }, | |
| { | |
| "epoch": 130.37, | |
| "learning_rate": 2.617283950617284e-06, | |
| "loss": 0.0002, | |
| "step": 7040 | |
| }, | |
| { | |
| "epoch": 130.56, | |
| "learning_rate": 2.5925925925925925e-06, | |
| "loss": 0.0002, | |
| "step": 7050 | |
| }, | |
| { | |
| "epoch": 130.74, | |
| "learning_rate": 2.5679012345679018e-06, | |
| "loss": 0.0002, | |
| "step": 7060 | |
| }, | |
| { | |
| "epoch": 130.93, | |
| "learning_rate": 2.54320987654321e-06, | |
| "loss": 0.0002, | |
| "step": 7070 | |
| }, | |
| { | |
| "epoch": 131.0, | |
| "eval_accuracy": 1.0, | |
| "eval_loss": 0.00022435725259128958, | |
| "eval_runtime": 3.0635, | |
| "eval_samples_per_second": 24.808, | |
| "eval_steps_per_second": 3.264, | |
| "step": 7074 | |
| }, | |
| { | |
| "epoch": 131.11, | |
| "learning_rate": 2.5185185185185186e-06, | |
| "loss": 0.0002, | |
| "step": 7080 | |
| }, | |
| { | |
| "epoch": 131.3, | |
| "learning_rate": 2.4938271604938274e-06, | |
| "loss": 0.0002, | |
| "step": 7090 | |
| }, | |
| { | |
| "epoch": 131.48, | |
| "learning_rate": 2.469135802469136e-06, | |
| "loss": 0.0002, | |
| "step": 7100 | |
| }, | |
| { | |
| "epoch": 131.67, | |
| "learning_rate": 2.4444444444444447e-06, | |
| "loss": 0.0002, | |
| "step": 7110 | |
| }, | |
| { | |
| "epoch": 131.85, | |
| "learning_rate": 2.419753086419753e-06, | |
| "loss": 0.0002, | |
| "step": 7120 | |
| }, | |
| { | |
| "epoch": 132.0, | |
| "eval_accuracy": 1.0, | |
| "eval_loss": 0.00022238449309952557, | |
| "eval_runtime": 3.099, | |
| "eval_samples_per_second": 24.524, | |
| "eval_steps_per_second": 3.227, | |
| "step": 7128 | |
| }, | |
| { | |
| "epoch": 132.04, | |
| "learning_rate": 2.395061728395062e-06, | |
| "loss": 0.0002, | |
| "step": 7130 | |
| }, | |
| { | |
| "epoch": 132.22, | |
| "learning_rate": 2.3703703703703707e-06, | |
| "loss": 0.0002, | |
| "step": 7140 | |
| }, | |
| { | |
| "epoch": 132.41, | |
| "learning_rate": 2.345679012345679e-06, | |
| "loss": 0.0002, | |
| "step": 7150 | |
| }, | |
| { | |
| "epoch": 132.59, | |
| "learning_rate": 2.3209876543209876e-06, | |
| "loss": 0.0002, | |
| "step": 7160 | |
| }, | |
| { | |
| "epoch": 132.78, | |
| "learning_rate": 2.2962962962962964e-06, | |
| "loss": 0.0002, | |
| "step": 7170 | |
| }, | |
| { | |
| "epoch": 132.96, | |
| "learning_rate": 2.2716049382716052e-06, | |
| "loss": 0.0002, | |
| "step": 7180 | |
| }, | |
| { | |
| "epoch": 133.0, | |
| "eval_accuracy": 1.0, | |
| "eval_loss": 0.00022056240413803607, | |
| "eval_runtime": 2.9993, | |
| "eval_samples_per_second": 25.34, | |
| "eval_steps_per_second": 3.334, | |
| "step": 7182 | |
| }, | |
| { | |
| "epoch": 133.15, | |
| "learning_rate": 2.2469135802469137e-06, | |
| "loss": 0.0002, | |
| "step": 7190 | |
| }, | |
| { | |
| "epoch": 133.33, | |
| "learning_rate": 2.222222222222222e-06, | |
| "loss": 0.0002, | |
| "step": 7200 | |
| }, | |
| { | |
| "epoch": 133.52, | |
| "learning_rate": 2.197530864197531e-06, | |
| "loss": 0.0002, | |
| "step": 7210 | |
| }, | |
| { | |
| "epoch": 133.7, | |
| "learning_rate": 2.1728395061728397e-06, | |
| "loss": 0.0002, | |
| "step": 7220 | |
| }, | |
| { | |
| "epoch": 133.89, | |
| "learning_rate": 2.148148148148148e-06, | |
| "loss": 0.0002, | |
| "step": 7230 | |
| }, | |
| { | |
| "epoch": 134.0, | |
| "eval_accuracy": 1.0, | |
| "eval_loss": 0.00021875585662201047, | |
| "eval_runtime": 3.0391, | |
| "eval_samples_per_second": 25.007, | |
| "eval_steps_per_second": 3.29, | |
| "step": 7236 | |
| }, | |
| { | |
| "epoch": 134.07, | |
| "learning_rate": 2.123456790123457e-06, | |
| "loss": 0.0002, | |
| "step": 7240 | |
| }, | |
| { | |
| "epoch": 134.26, | |
| "learning_rate": 2.0987654320987654e-06, | |
| "loss": 0.0002, | |
| "step": 7250 | |
| }, | |
| { | |
| "epoch": 134.44, | |
| "learning_rate": 2.0740740740740742e-06, | |
| "loss": 0.0002, | |
| "step": 7260 | |
| }, | |
| { | |
| "epoch": 134.63, | |
| "learning_rate": 2.049382716049383e-06, | |
| "loss": 0.0002, | |
| "step": 7270 | |
| }, | |
| { | |
| "epoch": 134.81, | |
| "learning_rate": 2.0246913580246915e-06, | |
| "loss": 0.0002, | |
| "step": 7280 | |
| }, | |
| { | |
| "epoch": 135.0, | |
| "learning_rate": 2.0000000000000003e-06, | |
| "loss": 0.0002, | |
| "step": 7290 | |
| }, | |
| { | |
| "epoch": 135.0, | |
| "eval_accuracy": 1.0, | |
| "eval_loss": 0.0002170340158045292, | |
| "eval_runtime": 3.0595, | |
| "eval_samples_per_second": 24.84, | |
| "eval_steps_per_second": 3.268, | |
| "step": 7290 | |
| }, | |
| { | |
| "epoch": 135.19, | |
| "learning_rate": 1.9753086419753087e-06, | |
| "loss": 0.0002, | |
| "step": 7300 | |
| }, | |
| { | |
| "epoch": 135.37, | |
| "learning_rate": 1.9506172839506176e-06, | |
| "loss": 0.0002, | |
| "step": 7310 | |
| }, | |
| { | |
| "epoch": 135.56, | |
| "learning_rate": 1.925925925925926e-06, | |
| "loss": 0.0002, | |
| "step": 7320 | |
| }, | |
| { | |
| "epoch": 135.74, | |
| "learning_rate": 1.9012345679012348e-06, | |
| "loss": 0.0002, | |
| "step": 7330 | |
| }, | |
| { | |
| "epoch": 135.93, | |
| "learning_rate": 1.8765432098765435e-06, | |
| "loss": 0.0003, | |
| "step": 7340 | |
| }, | |
| { | |
| "epoch": 136.0, | |
| "eval_accuracy": 1.0, | |
| "eval_loss": 0.00021536705025937408, | |
| "eval_runtime": 3.0723, | |
| "eval_samples_per_second": 24.737, | |
| "eval_steps_per_second": 3.255, | |
| "step": 7344 | |
| }, | |
| { | |
| "epoch": 136.11, | |
| "learning_rate": 1.8518518518518519e-06, | |
| "loss": 0.0002, | |
| "step": 7350 | |
| }, | |
| { | |
| "epoch": 136.3, | |
| "learning_rate": 1.8271604938271605e-06, | |
| "loss": 0.0002, | |
| "step": 7360 | |
| }, | |
| { | |
| "epoch": 136.48, | |
| "learning_rate": 1.8024691358024693e-06, | |
| "loss": 0.0002, | |
| "step": 7370 | |
| }, | |
| { | |
| "epoch": 136.67, | |
| "learning_rate": 1.777777777777778e-06, | |
| "loss": 0.0002, | |
| "step": 7380 | |
| }, | |
| { | |
| "epoch": 136.85, | |
| "learning_rate": 1.7530864197530868e-06, | |
| "loss": 0.0002, | |
| "step": 7390 | |
| }, | |
| { | |
| "epoch": 137.0, | |
| "eval_accuracy": 1.0, | |
| "eval_loss": 0.00021387105516623706, | |
| "eval_runtime": 3.1056, | |
| "eval_samples_per_second": 24.472, | |
| "eval_steps_per_second": 3.22, | |
| "step": 7398 | |
| }, | |
| { | |
| "epoch": 137.04, | |
| "learning_rate": 1.7283950617283952e-06, | |
| "loss": 0.0002, | |
| "step": 7400 | |
| }, | |
| { | |
| "epoch": 137.22, | |
| "learning_rate": 1.7037037037037038e-06, | |
| "loss": 0.0002, | |
| "step": 7410 | |
| }, | |
| { | |
| "epoch": 137.41, | |
| "learning_rate": 1.6790123456790125e-06, | |
| "loss": 0.0002, | |
| "step": 7420 | |
| }, | |
| { | |
| "epoch": 137.59, | |
| "learning_rate": 1.6543209876543213e-06, | |
| "loss": 0.0002, | |
| "step": 7430 | |
| }, | |
| { | |
| "epoch": 137.78, | |
| "learning_rate": 1.62962962962963e-06, | |
| "loss": 0.0002, | |
| "step": 7440 | |
| }, | |
| { | |
| "epoch": 137.96, | |
| "learning_rate": 1.6049382716049383e-06, | |
| "loss": 0.0002, | |
| "step": 7450 | |
| }, | |
| { | |
| "epoch": 138.0, | |
| "eval_accuracy": 1.0, | |
| "eval_loss": 0.00021247855329420418, | |
| "eval_runtime": 3.0565, | |
| "eval_samples_per_second": 24.865, | |
| "eval_steps_per_second": 3.272, | |
| "step": 7452 | |
| }, | |
| { | |
| "epoch": 138.15, | |
| "learning_rate": 1.580246913580247e-06, | |
| "loss": 0.0002, | |
| "step": 7460 | |
| }, | |
| { | |
| "epoch": 138.33, | |
| "learning_rate": 1.5555555555555558e-06, | |
| "loss": 0.0002, | |
| "step": 7470 | |
| }, | |
| { | |
| "epoch": 138.52, | |
| "learning_rate": 1.5308641975308644e-06, | |
| "loss": 0.0002, | |
| "step": 7480 | |
| }, | |
| { | |
| "epoch": 138.7, | |
| "learning_rate": 1.506172839506173e-06, | |
| "loss": 0.0002, | |
| "step": 7490 | |
| }, | |
| { | |
| "epoch": 138.89, | |
| "learning_rate": 1.4814814814814815e-06, | |
| "loss": 0.0028, | |
| "step": 7500 | |
| }, | |
| { | |
| "epoch": 139.0, | |
| "eval_accuracy": 1.0, | |
| "eval_loss": 0.0002113178779836744, | |
| "eval_runtime": 3.0392, | |
| "eval_samples_per_second": 25.007, | |
| "eval_steps_per_second": 3.29, | |
| "step": 7506 | |
| }, | |
| { | |
| "epoch": 139.07, | |
| "learning_rate": 1.4567901234567903e-06, | |
| "loss": 0.0002, | |
| "step": 7510 | |
| }, | |
| { | |
| "epoch": 139.26, | |
| "learning_rate": 1.432098765432099e-06, | |
| "loss": 0.0002, | |
| "step": 7520 | |
| }, | |
| { | |
| "epoch": 139.44, | |
| "learning_rate": 1.4074074074074075e-06, | |
| "loss": 0.0002, | |
| "step": 7530 | |
| }, | |
| { | |
| "epoch": 139.63, | |
| "learning_rate": 1.382716049382716e-06, | |
| "loss": 0.0002, | |
| "step": 7540 | |
| }, | |
| { | |
| "epoch": 139.81, | |
| "learning_rate": 1.3580246913580248e-06, | |
| "loss": 0.0002, | |
| "step": 7550 | |
| }, | |
| { | |
| "epoch": 140.0, | |
| "learning_rate": 1.3333333333333334e-06, | |
| "loss": 0.0006, | |
| "step": 7560 | |
| }, | |
| { | |
| "epoch": 140.0, | |
| "eval_accuracy": 1.0, | |
| "eval_loss": 0.000210436453926377, | |
| "eval_runtime": 3.0044, | |
| "eval_samples_per_second": 25.296, | |
| "eval_steps_per_second": 3.328, | |
| "step": 7560 | |
| }, | |
| { | |
| "epoch": 140.19, | |
| "learning_rate": 1.308641975308642e-06, | |
| "loss": 0.0002, | |
| "step": 7570 | |
| }, | |
| { | |
| "epoch": 140.37, | |
| "learning_rate": 1.2839506172839509e-06, | |
| "loss": 0.0002, | |
| "step": 7580 | |
| }, | |
| { | |
| "epoch": 140.56, | |
| "learning_rate": 1.2592592592592593e-06, | |
| "loss": 0.0002, | |
| "step": 7590 | |
| }, | |
| { | |
| "epoch": 140.74, | |
| "learning_rate": 1.234567901234568e-06, | |
| "loss": 0.0002, | |
| "step": 7600 | |
| }, | |
| { | |
| "epoch": 140.93, | |
| "learning_rate": 1.2098765432098765e-06, | |
| "loss": 0.0002, | |
| "step": 7610 | |
| }, | |
| { | |
| "epoch": 141.0, | |
| "eval_accuracy": 1.0, | |
| "eval_loss": 0.00020950206089764833, | |
| "eval_runtime": 3.0835, | |
| "eval_samples_per_second": 24.647, | |
| "eval_steps_per_second": 3.243, | |
| "step": 7614 | |
| }, | |
| { | |
| "epoch": 141.11, | |
| "learning_rate": 1.1851851851851854e-06, | |
| "loss": 0.0002, | |
| "step": 7620 | |
| }, | |
| { | |
| "epoch": 141.3, | |
| "learning_rate": 1.1604938271604938e-06, | |
| "loss": 0.0002, | |
| "step": 7630 | |
| }, | |
| { | |
| "epoch": 141.48, | |
| "learning_rate": 1.1358024691358026e-06, | |
| "loss": 0.0002, | |
| "step": 7640 | |
| }, | |
| { | |
| "epoch": 141.67, | |
| "learning_rate": 1.111111111111111e-06, | |
| "loss": 0.0002, | |
| "step": 7650 | |
| }, | |
| { | |
| "epoch": 141.85, | |
| "learning_rate": 1.0864197530864199e-06, | |
| "loss": 0.0002, | |
| "step": 7660 | |
| }, | |
| { | |
| "epoch": 142.0, | |
| "eval_accuracy": 1.0, | |
| "eval_loss": 0.00020854859030805528, | |
| "eval_runtime": 3.1029, | |
| "eval_samples_per_second": 24.493, | |
| "eval_steps_per_second": 3.223, | |
| "step": 7668 | |
| }, | |
| { | |
| "epoch": 142.04, | |
| "learning_rate": 1.0617283950617285e-06, | |
| "loss": 0.0002, | |
| "step": 7670 | |
| }, | |
| { | |
| "epoch": 142.22, | |
| "learning_rate": 1.0370370370370371e-06, | |
| "loss": 0.0002, | |
| "step": 7680 | |
| }, | |
| { | |
| "epoch": 142.41, | |
| "learning_rate": 1.0123456790123457e-06, | |
| "loss": 0.0002, | |
| "step": 7690 | |
| }, | |
| { | |
| "epoch": 142.59, | |
| "learning_rate": 9.876543209876544e-07, | |
| "loss": 0.0002, | |
| "step": 7700 | |
| }, | |
| { | |
| "epoch": 142.78, | |
| "learning_rate": 9.62962962962963e-07, | |
| "loss": 0.0002, | |
| "step": 7710 | |
| }, | |
| { | |
| "epoch": 142.96, | |
| "learning_rate": 9.382716049382717e-07, | |
| "loss": 0.0004, | |
| "step": 7720 | |
| }, | |
| { | |
| "epoch": 143.0, | |
| "eval_accuracy": 1.0, | |
| "eval_loss": 0.00020763746579177678, | |
| "eval_runtime": 3.038, | |
| "eval_samples_per_second": 25.017, | |
| "eval_steps_per_second": 3.292, | |
| "step": 7722 | |
| }, | |
| { | |
| "epoch": 143.15, | |
| "learning_rate": 9.135802469135802e-07, | |
| "loss": 0.0002, | |
| "step": 7730 | |
| }, | |
| { | |
| "epoch": 143.33, | |
| "learning_rate": 8.88888888888889e-07, | |
| "loss": 0.0002, | |
| "step": 7740 | |
| }, | |
| { | |
| "epoch": 143.52, | |
| "learning_rate": 8.641975308641976e-07, | |
| "loss": 0.0002, | |
| "step": 7750 | |
| }, | |
| { | |
| "epoch": 143.7, | |
| "learning_rate": 8.395061728395062e-07, | |
| "loss": 0.0002, | |
| "step": 7760 | |
| }, | |
| { | |
| "epoch": 143.89, | |
| "learning_rate": 8.14814814814815e-07, | |
| "loss": 0.0002, | |
| "step": 7770 | |
| }, | |
| { | |
| "epoch": 144.0, | |
| "eval_accuracy": 1.0, | |
| "eval_loss": 0.00020678593136835843, | |
| "eval_runtime": 3.0151, | |
| "eval_samples_per_second": 25.206, | |
| "eval_steps_per_second": 3.317, | |
| "step": 7776 | |
| }, | |
| { | |
| "epoch": 144.07, | |
| "learning_rate": 7.901234567901235e-07, | |
| "loss": 0.0002, | |
| "step": 7780 | |
| }, | |
| { | |
| "epoch": 144.26, | |
| "learning_rate": 7.654320987654322e-07, | |
| "loss": 0.0002, | |
| "step": 7790 | |
| }, | |
| { | |
| "epoch": 144.44, | |
| "learning_rate": 7.407407407407407e-07, | |
| "loss": 0.0002, | |
| "step": 7800 | |
| }, | |
| { | |
| "epoch": 144.63, | |
| "learning_rate": 7.160493827160495e-07, | |
| "loss": 0.0002, | |
| "step": 7810 | |
| }, | |
| { | |
| "epoch": 144.81, | |
| "learning_rate": 6.91358024691358e-07, | |
| "loss": 0.0002, | |
| "step": 7820 | |
| }, | |
| { | |
| "epoch": 145.0, | |
| "learning_rate": 6.666666666666667e-07, | |
| "loss": 0.0002, | |
| "step": 7830 | |
| }, | |
| { | |
| "epoch": 145.0, | |
| "eval_accuracy": 1.0, | |
| "eval_loss": 0.00020612729713320732, | |
| "eval_runtime": 3.0352, | |
| "eval_samples_per_second": 25.04, | |
| "eval_steps_per_second": 3.295, | |
| "step": 7830 | |
| }, | |
| { | |
| "epoch": 145.19, | |
| "learning_rate": 6.419753086419754e-07, | |
| "loss": 0.0002, | |
| "step": 7840 | |
| }, | |
| { | |
| "epoch": 145.37, | |
| "learning_rate": 6.17283950617284e-07, | |
| "loss": 0.0002, | |
| "step": 7850 | |
| }, | |
| { | |
| "epoch": 145.56, | |
| "learning_rate": 5.925925925925927e-07, | |
| "loss": 0.0002, | |
| "step": 7860 | |
| }, | |
| { | |
| "epoch": 145.74, | |
| "learning_rate": 5.679012345679013e-07, | |
| "loss": 0.0002, | |
| "step": 7870 | |
| }, | |
| { | |
| "epoch": 145.93, | |
| "learning_rate": 5.432098765432099e-07, | |
| "loss": 0.1028, | |
| "step": 7880 | |
| }, | |
| { | |
| "epoch": 146.0, | |
| "eval_accuracy": 1.0, | |
| "eval_loss": 0.00020691766985692084, | |
| "eval_runtime": 2.9867, | |
| "eval_samples_per_second": 25.446, | |
| "eval_steps_per_second": 3.348, | |
| "step": 7884 | |
| }, | |
| { | |
| "epoch": 146.11, | |
| "learning_rate": 5.185185185185186e-07, | |
| "loss": 0.0002, | |
| "step": 7890 | |
| }, | |
| { | |
| "epoch": 146.3, | |
| "learning_rate": 4.938271604938272e-07, | |
| "loss": 0.0002, | |
| "step": 7900 | |
| }, | |
| { | |
| "epoch": 146.48, | |
| "learning_rate": 4.6913580246913586e-07, | |
| "loss": 0.0002, | |
| "step": 7910 | |
| }, | |
| { | |
| "epoch": 146.67, | |
| "learning_rate": 4.444444444444445e-07, | |
| "loss": 0.0002, | |
| "step": 7920 | |
| }, | |
| { | |
| "epoch": 146.85, | |
| "learning_rate": 4.197530864197531e-07, | |
| "loss": 0.0002, | |
| "step": 7930 | |
| }, | |
| { | |
| "epoch": 147.0, | |
| "eval_accuracy": 1.0, | |
| "eval_loss": 0.00020708548254333436, | |
| "eval_runtime": 3.038, | |
| "eval_samples_per_second": 25.017, | |
| "eval_steps_per_second": 3.292, | |
| "step": 7938 | |
| }, | |
| { | |
| "epoch": 147.04, | |
| "learning_rate": 3.9506172839506174e-07, | |
| "loss": 0.0002, | |
| "step": 7940 | |
| }, | |
| { | |
| "epoch": 147.22, | |
| "learning_rate": 3.7037037037037036e-07, | |
| "loss": 0.0002, | |
| "step": 7950 | |
| }, | |
| { | |
| "epoch": 147.41, | |
| "learning_rate": 3.45679012345679e-07, | |
| "loss": 0.0002, | |
| "step": 7960 | |
| }, | |
| { | |
| "epoch": 147.59, | |
| "learning_rate": 3.209876543209877e-07, | |
| "loss": 0.0002, | |
| "step": 7970 | |
| }, | |
| { | |
| "epoch": 147.78, | |
| "learning_rate": 2.9629629629629634e-07, | |
| "loss": 0.0002, | |
| "step": 7980 | |
| }, | |
| { | |
| "epoch": 147.96, | |
| "learning_rate": 2.7160493827160497e-07, | |
| "loss": 0.0002, | |
| "step": 7990 | |
| }, | |
| { | |
| "epoch": 148.0, | |
| "eval_accuracy": 1.0, | |
| "eval_loss": 0.0002069223701255396, | |
| "eval_runtime": 3.0851, | |
| "eval_samples_per_second": 24.635, | |
| "eval_steps_per_second": 3.241, | |
| "step": 7992 | |
| }, | |
| { | |
| "epoch": 148.15, | |
| "learning_rate": 2.469135802469136e-07, | |
| "loss": 0.0002, | |
| "step": 8000 | |
| }, | |
| { | |
| "epoch": 148.33, | |
| "learning_rate": 2.2222222222222224e-07, | |
| "loss": 0.0002, | |
| "step": 8010 | |
| }, | |
| { | |
| "epoch": 148.52, | |
| "learning_rate": 1.9753086419753087e-07, | |
| "loss": 0.0002, | |
| "step": 8020 | |
| }, | |
| { | |
| "epoch": 148.7, | |
| "learning_rate": 1.728395061728395e-07, | |
| "loss": 0.0002, | |
| "step": 8030 | |
| }, | |
| { | |
| "epoch": 148.89, | |
| "learning_rate": 1.4814814814814817e-07, | |
| "loss": 0.0002, | |
| "step": 8040 | |
| }, | |
| { | |
| "epoch": 149.0, | |
| "eval_accuracy": 1.0, | |
| "eval_loss": 0.00020683615002781153, | |
| "eval_runtime": 3.076, | |
| "eval_samples_per_second": 24.708, | |
| "eval_steps_per_second": 3.251, | |
| "step": 8046 | |
| }, | |
| { | |
| "epoch": 149.07, | |
| "learning_rate": 1.234567901234568e-07, | |
| "loss": 0.0002, | |
| "step": 8050 | |
| }, | |
| { | |
| "epoch": 149.26, | |
| "learning_rate": 9.876543209876543e-08, | |
| "loss": 0.0002, | |
| "step": 8060 | |
| }, | |
| { | |
| "epoch": 149.44, | |
| "learning_rate": 7.407407407407409e-08, | |
| "loss": 0.0002, | |
| "step": 8070 | |
| }, | |
| { | |
| "epoch": 149.63, | |
| "learning_rate": 4.938271604938272e-08, | |
| "loss": 0.0002, | |
| "step": 8080 | |
| }, | |
| { | |
| "epoch": 149.81, | |
| "learning_rate": 2.469135802469136e-08, | |
| "loss": 0.0002, | |
| "step": 8090 | |
| }, | |
| { | |
| "epoch": 150.0, | |
| "learning_rate": 0.0, | |
| "loss": 0.0002, | |
| "step": 8100 | |
| }, | |
| { | |
| "epoch": 150.0, | |
| "eval_accuracy": 1.0, | |
| "eval_loss": 0.00020679850422311574, | |
| "eval_runtime": 3.0877, | |
| "eval_samples_per_second": 24.614, | |
| "eval_steps_per_second": 3.239, | |
| "step": 8100 | |
| }, | |
| { | |
| "epoch": 150.0, | |
| "step": 8100, | |
| "total_flos": 4.94011433793024e+18, | |
| "train_loss": 0.0, | |
| "train_runtime": 0.0622, | |
| "train_samples_per_second": 1025154.241, | |
| "train_steps_per_second": 130254.892 | |
| } | |
| ], | |
| "logging_steps": 10, | |
| "max_steps": 8100, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 150, | |
| "save_steps": 500, | |
| "total_flos": 4.94011433793024e+18, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |