| { | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 199.99750623441398, | |
| "global_step": 10000, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 6e-06, | |
| "loss": 1.0901, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 1.2e-05, | |
| "loss": 1.0728, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "learning_rate": 1.8e-05, | |
| "loss": 1.0838, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "learning_rate": 2.4e-05, | |
| "loss": 1.0591, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "learning_rate": 3e-05, | |
| "loss": 1.0419, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 1.2, | |
| "learning_rate": 3.6e-05, | |
| "loss": 1.0973, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 1.4, | |
| "learning_rate": 4.2e-05, | |
| "loss": 1.07, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 1.6, | |
| "learning_rate": 4.8e-05, | |
| "loss": 1.0816, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 1.8, | |
| "learning_rate": 5.4000000000000005e-05, | |
| "loss": 1.0822, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 2.0, | |
| "learning_rate": 6e-05, | |
| "loss": 1.0825, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 2.2, | |
| "learning_rate": 5.9877551020408165e-05, | |
| "loss": 1.1134, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 2.4, | |
| "learning_rate": 5.975510204081633e-05, | |
| "loss": 1.1079, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 2.6, | |
| "learning_rate": 5.963265306122449e-05, | |
| "loss": 1.1017, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 2.8, | |
| "learning_rate": 5.9510204081632654e-05, | |
| "loss": 1.1001, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 3.0, | |
| "learning_rate": 5.938775510204082e-05, | |
| "loss": 1.0689, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 3.2, | |
| "learning_rate": 5.926530612244898e-05, | |
| "loss": 1.1165, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 3.4, | |
| "learning_rate": 5.914285714285715e-05, | |
| "loss": 1.0812, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 3.6, | |
| "learning_rate": 5.9020408163265306e-05, | |
| "loss": 1.0963, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 3.8, | |
| "learning_rate": 5.889795918367347e-05, | |
| "loss": 1.0757, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 4.0, | |
| "learning_rate": 5.877551020408163e-05, | |
| "loss": 1.0759, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 4.2, | |
| "learning_rate": 5.8653061224489795e-05, | |
| "loss": 1.1081, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 4.4, | |
| "learning_rate": 5.8530612244897965e-05, | |
| "loss": 1.0893, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 4.6, | |
| "learning_rate": 5.840816326530613e-05, | |
| "loss": 1.0849, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 4.8, | |
| "learning_rate": 5.8285714285714284e-05, | |
| "loss": 1.0753, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 5.0, | |
| "learning_rate": 5.816326530612245e-05, | |
| "loss": 1.0611, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 5.2, | |
| "learning_rate": 5.804081632653062e-05, | |
| "loss": 1.0977, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 5.4, | |
| "learning_rate": 5.791836734693878e-05, | |
| "loss": 1.0702, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 5.6, | |
| "learning_rate": 5.779591836734694e-05, | |
| "loss": 1.0886, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 5.8, | |
| "learning_rate": 5.76734693877551e-05, | |
| "loss": 1.073, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 6.0, | |
| "learning_rate": 5.755102040816326e-05, | |
| "loss": 1.0663, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 6.2, | |
| "learning_rate": 5.742857142857143e-05, | |
| "loss": 1.0989, | |
| "step": 310 | |
| }, | |
| { | |
| "epoch": 6.4, | |
| "learning_rate": 5.7306122448979596e-05, | |
| "loss": 1.0773, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 6.6, | |
| "learning_rate": 5.718367346938776e-05, | |
| "loss": 1.0682, | |
| "step": 330 | |
| }, | |
| { | |
| "epoch": 6.8, | |
| "learning_rate": 5.7061224489795915e-05, | |
| "loss": 1.0759, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 7.0, | |
| "learning_rate": 5.6938775510204085e-05, | |
| "loss": 1.0497, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 7.2, | |
| "learning_rate": 5.681632653061225e-05, | |
| "loss": 1.0923, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 7.4, | |
| "learning_rate": 5.669387755102041e-05, | |
| "loss": 1.0503, | |
| "step": 370 | |
| }, | |
| { | |
| "epoch": 7.6, | |
| "learning_rate": 5.6571428571428574e-05, | |
| "loss": 1.0539, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 7.8, | |
| "learning_rate": 5.644897959183674e-05, | |
| "loss": 1.0693, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 8.0, | |
| "learning_rate": 5.63265306122449e-05, | |
| "loss": 1.0338, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 8.2, | |
| "learning_rate": 5.620408163265306e-05, | |
| "loss": 1.0612, | |
| "step": 410 | |
| }, | |
| { | |
| "epoch": 8.4, | |
| "learning_rate": 5.6081632653061226e-05, | |
| "loss": 1.0647, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 8.6, | |
| "learning_rate": 5.595918367346939e-05, | |
| "loss": 1.064, | |
| "step": 430 | |
| }, | |
| { | |
| "epoch": 8.8, | |
| "learning_rate": 5.583673469387755e-05, | |
| "loss": 1.0591, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 9.0, | |
| "learning_rate": 5.5714285714285715e-05, | |
| "loss": 1.0207, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 9.2, | |
| "learning_rate": 5.559183673469388e-05, | |
| "loss": 1.0733, | |
| "step": 460 | |
| }, | |
| { | |
| "epoch": 9.4, | |
| "learning_rate": 5.546938775510204e-05, | |
| "loss": 1.0442, | |
| "step": 470 | |
| }, | |
| { | |
| "epoch": 9.6, | |
| "learning_rate": 5.5346938775510204e-05, | |
| "loss": 1.0697, | |
| "step": 480 | |
| }, | |
| { | |
| "epoch": 9.8, | |
| "learning_rate": 5.5224489795918374e-05, | |
| "loss": 1.0628, | |
| "step": 490 | |
| }, | |
| { | |
| "epoch": 10.0, | |
| "learning_rate": 5.510204081632653e-05, | |
| "loss": 1.0344, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 10.0, | |
| "eval_loss": 0.40431472659111023, | |
| "eval_runtime": 247.681, | |
| "eval_samples_per_second": 21.043, | |
| "eval_steps_per_second": 1.316, | |
| "eval_wer": 0.451111573905347, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 10.2, | |
| "learning_rate": 5.4979591836734694e-05, | |
| "loss": 1.0725, | |
| "step": 510 | |
| }, | |
| { | |
| "epoch": 10.4, | |
| "learning_rate": 5.4857142857142857e-05, | |
| "loss": 1.0514, | |
| "step": 520 | |
| }, | |
| { | |
| "epoch": 10.6, | |
| "learning_rate": 5.473469387755102e-05, | |
| "loss": 1.0477, | |
| "step": 530 | |
| }, | |
| { | |
| "epoch": 10.8, | |
| "learning_rate": 5.461224489795919e-05, | |
| "loss": 1.0551, | |
| "step": 540 | |
| }, | |
| { | |
| "epoch": 11.0, | |
| "learning_rate": 5.4489795918367346e-05, | |
| "loss": 1.0322, | |
| "step": 550 | |
| }, | |
| { | |
| "epoch": 11.2, | |
| "learning_rate": 5.436734693877551e-05, | |
| "loss": 1.0435, | |
| "step": 560 | |
| }, | |
| { | |
| "epoch": 11.4, | |
| "learning_rate": 5.424489795918367e-05, | |
| "loss": 1.0594, | |
| "step": 570 | |
| }, | |
| { | |
| "epoch": 11.6, | |
| "learning_rate": 5.412244897959184e-05, | |
| "loss": 1.0419, | |
| "step": 580 | |
| }, | |
| { | |
| "epoch": 11.8, | |
| "learning_rate": 5.4000000000000005e-05, | |
| "loss": 1.0291, | |
| "step": 590 | |
| }, | |
| { | |
| "epoch": 12.0, | |
| "learning_rate": 5.387755102040817e-05, | |
| "loss": 1.0314, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 12.2, | |
| "learning_rate": 5.3755102040816324e-05, | |
| "loss": 1.0598, | |
| "step": 610 | |
| }, | |
| { | |
| "epoch": 12.4, | |
| "learning_rate": 5.363265306122449e-05, | |
| "loss": 1.0412, | |
| "step": 620 | |
| }, | |
| { | |
| "epoch": 12.6, | |
| "learning_rate": 5.351020408163266e-05, | |
| "loss": 1.0483, | |
| "step": 630 | |
| }, | |
| { | |
| "epoch": 12.8, | |
| "learning_rate": 5.338775510204082e-05, | |
| "loss": 1.0448, | |
| "step": 640 | |
| }, | |
| { | |
| "epoch": 13.0, | |
| "learning_rate": 5.326530612244898e-05, | |
| "loss": 0.9938, | |
| "step": 650 | |
| }, | |
| { | |
| "epoch": 13.2, | |
| "learning_rate": 5.314285714285714e-05, | |
| "loss": 1.053, | |
| "step": 660 | |
| }, | |
| { | |
| "epoch": 13.4, | |
| "learning_rate": 5.302040816326531e-05, | |
| "loss": 1.0178, | |
| "step": 670 | |
| }, | |
| { | |
| "epoch": 13.6, | |
| "learning_rate": 5.289795918367347e-05, | |
| "loss": 1.0174, | |
| "step": 680 | |
| }, | |
| { | |
| "epoch": 13.8, | |
| "learning_rate": 5.2775510204081635e-05, | |
| "loss": 1.0231, | |
| "step": 690 | |
| }, | |
| { | |
| "epoch": 14.0, | |
| "learning_rate": 5.26530612244898e-05, | |
| "loss": 1.0033, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 14.2, | |
| "learning_rate": 5.2530612244897955e-05, | |
| "loss": 1.0444, | |
| "step": 710 | |
| }, | |
| { | |
| "epoch": 14.4, | |
| "learning_rate": 5.2408163265306124e-05, | |
| "loss": 0.9893, | |
| "step": 720 | |
| }, | |
| { | |
| "epoch": 14.6, | |
| "learning_rate": 5.228571428571429e-05, | |
| "loss": 1.0216, | |
| "step": 730 | |
| }, | |
| { | |
| "epoch": 14.8, | |
| "learning_rate": 5.216326530612245e-05, | |
| "loss": 1.0247, | |
| "step": 740 | |
| }, | |
| { | |
| "epoch": 15.0, | |
| "learning_rate": 5.2040816326530614e-05, | |
| "loss": 1.0022, | |
| "step": 750 | |
| }, | |
| { | |
| "epoch": 15.2, | |
| "learning_rate": 5.191836734693878e-05, | |
| "loss": 1.032, | |
| "step": 760 | |
| }, | |
| { | |
| "epoch": 15.4, | |
| "learning_rate": 5.179591836734694e-05, | |
| "loss": 0.9963, | |
| "step": 770 | |
| }, | |
| { | |
| "epoch": 15.6, | |
| "learning_rate": 5.16734693877551e-05, | |
| "loss": 0.9952, | |
| "step": 780 | |
| }, | |
| { | |
| "epoch": 15.8, | |
| "learning_rate": 5.1551020408163266e-05, | |
| "loss": 1.0021, | |
| "step": 790 | |
| }, | |
| { | |
| "epoch": 16.0, | |
| "learning_rate": 5.142857142857143e-05, | |
| "loss": 1.0134, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 16.2, | |
| "learning_rate": 5.13061224489796e-05, | |
| "loss": 1.0319, | |
| "step": 810 | |
| }, | |
| { | |
| "epoch": 16.4, | |
| "learning_rate": 5.1183673469387755e-05, | |
| "loss": 0.9963, | |
| "step": 820 | |
| }, | |
| { | |
| "epoch": 16.6, | |
| "learning_rate": 5.106122448979592e-05, | |
| "loss": 0.9966, | |
| "step": 830 | |
| }, | |
| { | |
| "epoch": 16.8, | |
| "learning_rate": 5.093877551020408e-05, | |
| "loss": 1.0023, | |
| "step": 840 | |
| }, | |
| { | |
| "epoch": 17.0, | |
| "learning_rate": 5.081632653061225e-05, | |
| "loss": 0.9739, | |
| "step": 850 | |
| }, | |
| { | |
| "epoch": 17.2, | |
| "learning_rate": 5.0693877551020414e-05, | |
| "loss": 1.0182, | |
| "step": 860 | |
| }, | |
| { | |
| "epoch": 17.4, | |
| "learning_rate": 5.057142857142857e-05, | |
| "loss": 0.9884, | |
| "step": 870 | |
| }, | |
| { | |
| "epoch": 17.6, | |
| "learning_rate": 5.044897959183673e-05, | |
| "loss": 0.9623, | |
| "step": 880 | |
| }, | |
| { | |
| "epoch": 17.8, | |
| "learning_rate": 5.0326530612244896e-05, | |
| "loss": 1.0061, | |
| "step": 890 | |
| }, | |
| { | |
| "epoch": 18.0, | |
| "learning_rate": 5.0204081632653066e-05, | |
| "loss": 0.9684, | |
| "step": 900 | |
| }, | |
| { | |
| "epoch": 18.2, | |
| "learning_rate": 5.008163265306123e-05, | |
| "loss": 1.0066, | |
| "step": 910 | |
| }, | |
| { | |
| "epoch": 18.4, | |
| "learning_rate": 4.9959183673469386e-05, | |
| "loss": 0.9779, | |
| "step": 920 | |
| }, | |
| { | |
| "epoch": 18.6, | |
| "learning_rate": 4.983673469387755e-05, | |
| "loss": 0.9784, | |
| "step": 930 | |
| }, | |
| { | |
| "epoch": 18.8, | |
| "learning_rate": 4.971428571428572e-05, | |
| "loss": 0.9869, | |
| "step": 940 | |
| }, | |
| { | |
| "epoch": 19.0, | |
| "learning_rate": 4.959183673469388e-05, | |
| "loss": 0.9561, | |
| "step": 950 | |
| }, | |
| { | |
| "epoch": 19.2, | |
| "learning_rate": 4.9469387755102045e-05, | |
| "loss": 1.0003, | |
| "step": 960 | |
| }, | |
| { | |
| "epoch": 19.4, | |
| "learning_rate": 4.93469387755102e-05, | |
| "loss": 0.9838, | |
| "step": 970 | |
| }, | |
| { | |
| "epoch": 19.6, | |
| "learning_rate": 4.9224489795918364e-05, | |
| "loss": 0.9969, | |
| "step": 980 | |
| }, | |
| { | |
| "epoch": 19.8, | |
| "learning_rate": 4.9102040816326534e-05, | |
| "loss": 1.0037, | |
| "step": 990 | |
| }, | |
| { | |
| "epoch": 20.0, | |
| "learning_rate": 4.89795918367347e-05, | |
| "loss": 0.9651, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 20.0, | |
| "eval_loss": 0.3793290853500366, | |
| "eval_runtime": 244.7376, | |
| "eval_samples_per_second": 21.296, | |
| "eval_steps_per_second": 1.332, | |
| "eval_wer": 0.4159421044410892, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 20.2, | |
| "learning_rate": 4.885714285714286e-05, | |
| "loss": 1.0027, | |
| "step": 1010 | |
| }, | |
| { | |
| "epoch": 20.4, | |
| "learning_rate": 4.873469387755102e-05, | |
| "loss": 0.9598, | |
| "step": 1020 | |
| }, | |
| { | |
| "epoch": 20.6, | |
| "learning_rate": 4.862448979591836e-05, | |
| "loss": 0.979, | |
| "step": 1030 | |
| }, | |
| { | |
| "epoch": 20.8, | |
| "learning_rate": 4.850204081632653e-05, | |
| "loss": 0.9886, | |
| "step": 1040 | |
| }, | |
| { | |
| "epoch": 21.0, | |
| "learning_rate": 4.8379591836734696e-05, | |
| "loss": 0.9584, | |
| "step": 1050 | |
| }, | |
| { | |
| "epoch": 21.2, | |
| "learning_rate": 4.825714285714286e-05, | |
| "loss": 0.984, | |
| "step": 1060 | |
| }, | |
| { | |
| "epoch": 21.4, | |
| "learning_rate": 4.813469387755102e-05, | |
| "loss": 0.9586, | |
| "step": 1070 | |
| }, | |
| { | |
| "epoch": 21.6, | |
| "learning_rate": 4.8012244897959185e-05, | |
| "loss": 0.9669, | |
| "step": 1080 | |
| }, | |
| { | |
| "epoch": 21.8, | |
| "learning_rate": 4.788979591836735e-05, | |
| "loss": 0.9629, | |
| "step": 1090 | |
| }, | |
| { | |
| "epoch": 22.0, | |
| "learning_rate": 4.776734693877551e-05, | |
| "loss": 0.9468, | |
| "step": 1100 | |
| }, | |
| { | |
| "epoch": 22.2, | |
| "learning_rate": 4.7644897959183674e-05, | |
| "loss": 1.0055, | |
| "step": 1110 | |
| }, | |
| { | |
| "epoch": 22.4, | |
| "learning_rate": 4.752244897959184e-05, | |
| "loss": 0.9698, | |
| "step": 1120 | |
| }, | |
| { | |
| "epoch": 22.6, | |
| "learning_rate": 4.74e-05, | |
| "loss": 0.9557, | |
| "step": 1130 | |
| }, | |
| { | |
| "epoch": 22.8, | |
| "learning_rate": 4.727755102040816e-05, | |
| "loss": 0.9648, | |
| "step": 1140 | |
| }, | |
| { | |
| "epoch": 23.0, | |
| "learning_rate": 4.7155102040816326e-05, | |
| "loss": 0.9681, | |
| "step": 1150 | |
| }, | |
| { | |
| "epoch": 23.2, | |
| "learning_rate": 4.703265306122449e-05, | |
| "loss": 0.995, | |
| "step": 1160 | |
| }, | |
| { | |
| "epoch": 23.4, | |
| "learning_rate": 4.691020408163266e-05, | |
| "loss": 0.9415, | |
| "step": 1170 | |
| }, | |
| { | |
| "epoch": 23.6, | |
| "learning_rate": 4.678775510204082e-05, | |
| "loss": 0.955, | |
| "step": 1180 | |
| }, | |
| { | |
| "epoch": 23.8, | |
| "learning_rate": 4.666530612244898e-05, | |
| "loss": 0.9647, | |
| "step": 1190 | |
| }, | |
| { | |
| "epoch": 24.0, | |
| "learning_rate": 4.654285714285714e-05, | |
| "loss": 0.9572, | |
| "step": 1200 | |
| }, | |
| { | |
| "epoch": 24.2, | |
| "learning_rate": 4.6420408163265305e-05, | |
| "loss": 1.0043, | |
| "step": 1210 | |
| }, | |
| { | |
| "epoch": 24.4, | |
| "learning_rate": 4.6297959183673474e-05, | |
| "loss": 0.9545, | |
| "step": 1220 | |
| }, | |
| { | |
| "epoch": 24.6, | |
| "learning_rate": 4.617551020408164e-05, | |
| "loss": 0.9603, | |
| "step": 1230 | |
| }, | |
| { | |
| "epoch": 24.8, | |
| "learning_rate": 4.6053061224489794e-05, | |
| "loss": 0.9563, | |
| "step": 1240 | |
| }, | |
| { | |
| "epoch": 25.0, | |
| "learning_rate": 4.593061224489796e-05, | |
| "loss": 0.939, | |
| "step": 1250 | |
| }, | |
| { | |
| "epoch": 25.2, | |
| "learning_rate": 4.580816326530613e-05, | |
| "loss": 0.9755, | |
| "step": 1260 | |
| }, | |
| { | |
| "epoch": 25.4, | |
| "learning_rate": 4.568571428571429e-05, | |
| "loss": 0.9619, | |
| "step": 1270 | |
| }, | |
| { | |
| "epoch": 25.6, | |
| "learning_rate": 4.556326530612245e-05, | |
| "loss": 0.951, | |
| "step": 1280 | |
| }, | |
| { | |
| "epoch": 25.8, | |
| "learning_rate": 4.544081632653061e-05, | |
| "loss": 0.9442, | |
| "step": 1290 | |
| }, | |
| { | |
| "epoch": 26.0, | |
| "learning_rate": 4.531836734693877e-05, | |
| "loss": 0.9448, | |
| "step": 1300 | |
| }, | |
| { | |
| "epoch": 26.2, | |
| "learning_rate": 4.519591836734694e-05, | |
| "loss": 0.984, | |
| "step": 1310 | |
| }, | |
| { | |
| "epoch": 26.4, | |
| "learning_rate": 4.5073469387755105e-05, | |
| "loss": 0.9407, | |
| "step": 1320 | |
| }, | |
| { | |
| "epoch": 26.6, | |
| "learning_rate": 4.495102040816327e-05, | |
| "loss": 0.9506, | |
| "step": 1330 | |
| }, | |
| { | |
| "epoch": 26.8, | |
| "learning_rate": 4.4828571428571424e-05, | |
| "loss": 0.9393, | |
| "step": 1340 | |
| }, | |
| { | |
| "epoch": 27.0, | |
| "learning_rate": 4.4706122448979594e-05, | |
| "loss": 0.9513, | |
| "step": 1350 | |
| }, | |
| { | |
| "epoch": 27.2, | |
| "learning_rate": 4.458367346938776e-05, | |
| "loss": 0.9636, | |
| "step": 1360 | |
| }, | |
| { | |
| "epoch": 27.4, | |
| "learning_rate": 4.446122448979592e-05, | |
| "loss": 0.9441, | |
| "step": 1370 | |
| }, | |
| { | |
| "epoch": 27.6, | |
| "learning_rate": 4.433877551020408e-05, | |
| "loss": 0.9476, | |
| "step": 1380 | |
| }, | |
| { | |
| "epoch": 27.8, | |
| "learning_rate": 4.4216326530612246e-05, | |
| "loss": 0.9438, | |
| "step": 1390 | |
| }, | |
| { | |
| "epoch": 28.0, | |
| "learning_rate": 4.409387755102041e-05, | |
| "loss": 0.9376, | |
| "step": 1400 | |
| }, | |
| { | |
| "epoch": 28.2, | |
| "learning_rate": 4.397142857142857e-05, | |
| "loss": 0.9699, | |
| "step": 1410 | |
| }, | |
| { | |
| "epoch": 28.4, | |
| "learning_rate": 4.3848979591836736e-05, | |
| "loss": 0.9719, | |
| "step": 1420 | |
| }, | |
| { | |
| "epoch": 28.6, | |
| "learning_rate": 4.37265306122449e-05, | |
| "loss": 0.9481, | |
| "step": 1430 | |
| }, | |
| { | |
| "epoch": 28.8, | |
| "learning_rate": 4.360408163265307e-05, | |
| "loss": 0.9495, | |
| "step": 1440 | |
| }, | |
| { | |
| "epoch": 29.0, | |
| "learning_rate": 4.3481632653061225e-05, | |
| "loss": 0.9208, | |
| "step": 1450 | |
| }, | |
| { | |
| "epoch": 29.2, | |
| "learning_rate": 4.335918367346939e-05, | |
| "loss": 0.9725, | |
| "step": 1460 | |
| }, | |
| { | |
| "epoch": 29.4, | |
| "learning_rate": 4.323673469387755e-05, | |
| "loss": 0.9317, | |
| "step": 1470 | |
| }, | |
| { | |
| "epoch": 29.6, | |
| "learning_rate": 4.3114285714285714e-05, | |
| "loss": 0.9402, | |
| "step": 1480 | |
| }, | |
| { | |
| "epoch": 29.8, | |
| "learning_rate": 4.2991836734693884e-05, | |
| "loss": 0.9431, | |
| "step": 1490 | |
| }, | |
| { | |
| "epoch": 30.0, | |
| "learning_rate": 4.286938775510204e-05, | |
| "loss": 0.9125, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 30.0, | |
| "eval_loss": 0.3755570352077484, | |
| "eval_runtime": 246.7652, | |
| "eval_samples_per_second": 21.121, | |
| "eval_steps_per_second": 1.321, | |
| "eval_wer": 0.4046441401572343, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 30.2, | |
| "learning_rate": 4.27469387755102e-05, | |
| "loss": 0.9531, | |
| "step": 1510 | |
| }, | |
| { | |
| "epoch": 30.4, | |
| "learning_rate": 4.2624489795918366e-05, | |
| "loss": 0.9172, | |
| "step": 1520 | |
| }, | |
| { | |
| "epoch": 30.6, | |
| "learning_rate": 4.250204081632653e-05, | |
| "loss": 0.9438, | |
| "step": 1530 | |
| }, | |
| { | |
| "epoch": 30.8, | |
| "learning_rate": 4.23795918367347e-05, | |
| "loss": 0.9405, | |
| "step": 1540 | |
| }, | |
| { | |
| "epoch": 31.0, | |
| "learning_rate": 4.2257142857142855e-05, | |
| "loss": 0.9209, | |
| "step": 1550 | |
| }, | |
| { | |
| "epoch": 31.2, | |
| "learning_rate": 4.213469387755102e-05, | |
| "loss": 0.9827, | |
| "step": 1560 | |
| }, | |
| { | |
| "epoch": 31.4, | |
| "learning_rate": 4.201224489795918e-05, | |
| "loss": 0.9163, | |
| "step": 1570 | |
| }, | |
| { | |
| "epoch": 31.6, | |
| "learning_rate": 4.188979591836735e-05, | |
| "loss": 0.9175, | |
| "step": 1580 | |
| }, | |
| { | |
| "epoch": 31.8, | |
| "learning_rate": 4.1767346938775514e-05, | |
| "loss": 0.9288, | |
| "step": 1590 | |
| }, | |
| { | |
| "epoch": 32.0, | |
| "learning_rate": 4.164489795918368e-05, | |
| "loss": 0.8951, | |
| "step": 1600 | |
| }, | |
| { | |
| "epoch": 32.2, | |
| "learning_rate": 4.1522448979591834e-05, | |
| "loss": 0.9582, | |
| "step": 1610 | |
| }, | |
| { | |
| "epoch": 32.4, | |
| "learning_rate": 4.14e-05, | |
| "loss": 0.9183, | |
| "step": 1620 | |
| }, | |
| { | |
| "epoch": 32.6, | |
| "learning_rate": 4.1277551020408166e-05, | |
| "loss": 0.9113, | |
| "step": 1630 | |
| }, | |
| { | |
| "epoch": 32.8, | |
| "learning_rate": 4.115510204081633e-05, | |
| "loss": 0.9305, | |
| "step": 1640 | |
| }, | |
| { | |
| "epoch": 33.0, | |
| "learning_rate": 4.103265306122449e-05, | |
| "loss": 0.9159, | |
| "step": 1650 | |
| }, | |
| { | |
| "epoch": 33.2, | |
| "learning_rate": 4.091020408163265e-05, | |
| "loss": 0.9706, | |
| "step": 1660 | |
| }, | |
| { | |
| "epoch": 33.4, | |
| "learning_rate": 4.078775510204082e-05, | |
| "loss": 0.9359, | |
| "step": 1670 | |
| }, | |
| { | |
| "epoch": 33.6, | |
| "learning_rate": 4.066530612244898e-05, | |
| "loss": 0.916, | |
| "step": 1680 | |
| }, | |
| { | |
| "epoch": 33.8, | |
| "learning_rate": 4.0542857142857145e-05, | |
| "loss": 0.9333, | |
| "step": 1690 | |
| }, | |
| { | |
| "epoch": 34.0, | |
| "learning_rate": 4.042040816326531e-05, | |
| "loss": 0.892, | |
| "step": 1700 | |
| }, | |
| { | |
| "epoch": 34.2, | |
| "learning_rate": 4.0297959183673464e-05, | |
| "loss": 0.9556, | |
| "step": 1710 | |
| }, | |
| { | |
| "epoch": 34.4, | |
| "learning_rate": 4.0175510204081634e-05, | |
| "loss": 0.8906, | |
| "step": 1720 | |
| }, | |
| { | |
| "epoch": 34.6, | |
| "learning_rate": 4.00530612244898e-05, | |
| "loss": 0.9183, | |
| "step": 1730 | |
| }, | |
| { | |
| "epoch": 34.8, | |
| "learning_rate": 3.993061224489796e-05, | |
| "loss": 0.9262, | |
| "step": 1740 | |
| }, | |
| { | |
| "epoch": 35.0, | |
| "learning_rate": 3.980816326530612e-05, | |
| "loss": 0.9276, | |
| "step": 1750 | |
| }, | |
| { | |
| "epoch": 35.2, | |
| "learning_rate": 3.968571428571429e-05, | |
| "loss": 0.9383, | |
| "step": 1760 | |
| }, | |
| { | |
| "epoch": 35.4, | |
| "learning_rate": 3.956326530612245e-05, | |
| "loss": 0.9247, | |
| "step": 1770 | |
| }, | |
| { | |
| "epoch": 35.6, | |
| "learning_rate": 3.944081632653061e-05, | |
| "loss": 0.9339, | |
| "step": 1780 | |
| }, | |
| { | |
| "epoch": 35.8, | |
| "learning_rate": 3.9318367346938775e-05, | |
| "loss": 0.8878, | |
| "step": 1790 | |
| }, | |
| { | |
| "epoch": 36.0, | |
| "learning_rate": 3.919591836734694e-05, | |
| "loss": 0.9035, | |
| "step": 1800 | |
| }, | |
| { | |
| "epoch": 36.2, | |
| "learning_rate": 3.907346938775511e-05, | |
| "loss": 0.9587, | |
| "step": 1810 | |
| }, | |
| { | |
| "epoch": 36.4, | |
| "learning_rate": 3.8951020408163265e-05, | |
| "loss": 0.8989, | |
| "step": 1820 | |
| }, | |
| { | |
| "epoch": 36.6, | |
| "learning_rate": 3.882857142857143e-05, | |
| "loss": 0.922, | |
| "step": 1830 | |
| }, | |
| { | |
| "epoch": 36.8, | |
| "learning_rate": 3.870612244897959e-05, | |
| "loss": 0.9278, | |
| "step": 1840 | |
| }, | |
| { | |
| "epoch": 37.0, | |
| "learning_rate": 3.858367346938776e-05, | |
| "loss": 0.8979, | |
| "step": 1850 | |
| }, | |
| { | |
| "epoch": 37.2, | |
| "learning_rate": 3.8461224489795924e-05, | |
| "loss": 0.9302, | |
| "step": 1860 | |
| }, | |
| { | |
| "epoch": 37.4, | |
| "learning_rate": 3.833877551020408e-05, | |
| "loss": 0.9078, | |
| "step": 1870 | |
| }, | |
| { | |
| "epoch": 37.6, | |
| "learning_rate": 3.821632653061224e-05, | |
| "loss": 0.919, | |
| "step": 1880 | |
| }, | |
| { | |
| "epoch": 37.8, | |
| "learning_rate": 3.8093877551020406e-05, | |
| "loss": 0.9152, | |
| "step": 1890 | |
| }, | |
| { | |
| "epoch": 38.0, | |
| "learning_rate": 3.7971428571428576e-05, | |
| "loss": 0.8914, | |
| "step": 1900 | |
| }, | |
| { | |
| "epoch": 38.2, | |
| "learning_rate": 3.784897959183674e-05, | |
| "loss": 0.9297, | |
| "step": 1910 | |
| }, | |
| { | |
| "epoch": 38.4, | |
| "learning_rate": 3.7726530612244895e-05, | |
| "loss": 0.9091, | |
| "step": 1920 | |
| }, | |
| { | |
| "epoch": 38.6, | |
| "learning_rate": 3.760408163265306e-05, | |
| "loss": 0.9085, | |
| "step": 1930 | |
| }, | |
| { | |
| "epoch": 38.8, | |
| "learning_rate": 3.748163265306123e-05, | |
| "loss": 0.9078, | |
| "step": 1940 | |
| }, | |
| { | |
| "epoch": 39.0, | |
| "learning_rate": 3.735918367346939e-05, | |
| "loss": 0.8804, | |
| "step": 1950 | |
| }, | |
| { | |
| "epoch": 39.2, | |
| "learning_rate": 3.7236734693877554e-05, | |
| "loss": 0.9208, | |
| "step": 1960 | |
| }, | |
| { | |
| "epoch": 39.4, | |
| "learning_rate": 3.711428571428572e-05, | |
| "loss": 0.8786, | |
| "step": 1970 | |
| }, | |
| { | |
| "epoch": 39.6, | |
| "learning_rate": 3.6991836734693873e-05, | |
| "loss": 0.8991, | |
| "step": 1980 | |
| }, | |
| { | |
| "epoch": 39.8, | |
| "learning_rate": 3.686938775510204e-05, | |
| "loss": 0.9153, | |
| "step": 1990 | |
| }, | |
| { | |
| "epoch": 40.0, | |
| "learning_rate": 3.6746938775510206e-05, | |
| "loss": 0.8831, | |
| "step": 2000 | |
| }, | |
| { | |
| "epoch": 40.0, | |
| "eval_loss": 0.3649730086326599, | |
| "eval_runtime": 244.6059, | |
| "eval_samples_per_second": 21.308, | |
| "eval_steps_per_second": 1.333, | |
| "eval_wer": 0.3875930650283751, | |
| "step": 2000 | |
| }, | |
| { | |
| "epoch": 40.2, | |
| "learning_rate": 3.662448979591837e-05, | |
| "loss": 0.8995, | |
| "step": 2010 | |
| }, | |
| { | |
| "epoch": 40.4, | |
| "learning_rate": 3.650204081632653e-05, | |
| "loss": 0.91, | |
| "step": 2020 | |
| }, | |
| { | |
| "epoch": 40.6, | |
| "learning_rate": 3.6379591836734695e-05, | |
| "loss": 0.9105, | |
| "step": 2030 | |
| }, | |
| { | |
| "epoch": 40.8, | |
| "learning_rate": 3.625714285714286e-05, | |
| "loss": 0.897, | |
| "step": 2040 | |
| }, | |
| { | |
| "epoch": 41.0, | |
| "learning_rate": 3.613469387755102e-05, | |
| "loss": 0.8869, | |
| "step": 2050 | |
| }, | |
| { | |
| "epoch": 41.2, | |
| "learning_rate": 3.6012244897959185e-05, | |
| "loss": 0.9227, | |
| "step": 2060 | |
| }, | |
| { | |
| "epoch": 41.4, | |
| "learning_rate": 3.588979591836735e-05, | |
| "loss": 0.8919, | |
| "step": 2070 | |
| }, | |
| { | |
| "epoch": 41.6, | |
| "learning_rate": 3.576734693877551e-05, | |
| "loss": 0.8926, | |
| "step": 2080 | |
| }, | |
| { | |
| "epoch": 41.8, | |
| "learning_rate": 3.5644897959183674e-05, | |
| "loss": 0.8912, | |
| "step": 2090 | |
| }, | |
| { | |
| "epoch": 42.0, | |
| "learning_rate": 3.552244897959184e-05, | |
| "loss": 0.8841, | |
| "step": 2100 | |
| }, | |
| { | |
| "epoch": 42.2, | |
| "learning_rate": 3.54e-05, | |
| "loss": 0.9165, | |
| "step": 2110 | |
| }, | |
| { | |
| "epoch": 42.4, | |
| "learning_rate": 3.527755102040817e-05, | |
| "loss": 0.8981, | |
| "step": 2120 | |
| }, | |
| { | |
| "epoch": 42.6, | |
| "learning_rate": 3.5155102040816326e-05, | |
| "loss": 0.9079, | |
| "step": 2130 | |
| }, | |
| { | |
| "epoch": 42.8, | |
| "learning_rate": 3.503265306122449e-05, | |
| "loss": 0.9017, | |
| "step": 2140 | |
| }, | |
| { | |
| "epoch": 43.0, | |
| "learning_rate": 3.491020408163265e-05, | |
| "loss": 0.8534, | |
| "step": 2150 | |
| }, | |
| { | |
| "epoch": 43.2, | |
| "learning_rate": 3.4787755102040815e-05, | |
| "loss": 0.9108, | |
| "step": 2160 | |
| }, | |
| { | |
| "epoch": 43.4, | |
| "learning_rate": 3.4665306122448985e-05, | |
| "loss": 0.8763, | |
| "step": 2170 | |
| }, | |
| { | |
| "epoch": 43.6, | |
| "learning_rate": 3.454285714285715e-05, | |
| "loss": 0.903, | |
| "step": 2180 | |
| }, | |
| { | |
| "epoch": 43.8, | |
| "learning_rate": 3.4420408163265304e-05, | |
| "loss": 0.8842, | |
| "step": 2190 | |
| }, | |
| { | |
| "epoch": 44.0, | |
| "learning_rate": 3.429795918367347e-05, | |
| "loss": 0.8646, | |
| "step": 2200 | |
| }, | |
| { | |
| "epoch": 44.2, | |
| "learning_rate": 3.417551020408163e-05, | |
| "loss": 0.921, | |
| "step": 2210 | |
| }, | |
| { | |
| "epoch": 44.4, | |
| "learning_rate": 3.40530612244898e-05, | |
| "loss": 0.8833, | |
| "step": 2220 | |
| }, | |
| { | |
| "epoch": 44.6, | |
| "learning_rate": 3.393061224489796e-05, | |
| "loss": 0.8826, | |
| "step": 2230 | |
| }, | |
| { | |
| "epoch": 44.8, | |
| "learning_rate": 3.380816326530612e-05, | |
| "loss": 0.8892, | |
| "step": 2240 | |
| }, | |
| { | |
| "epoch": 45.0, | |
| "learning_rate": 3.368571428571428e-05, | |
| "loss": 0.8688, | |
| "step": 2250 | |
| }, | |
| { | |
| "epoch": 45.2, | |
| "learning_rate": 3.356326530612245e-05, | |
| "loss": 0.92, | |
| "step": 2260 | |
| }, | |
| { | |
| "epoch": 45.4, | |
| "learning_rate": 3.3440816326530616e-05, | |
| "loss": 0.8728, | |
| "step": 2270 | |
| }, | |
| { | |
| "epoch": 45.6, | |
| "learning_rate": 3.331836734693878e-05, | |
| "loss": 0.8921, | |
| "step": 2280 | |
| }, | |
| { | |
| "epoch": 45.8, | |
| "learning_rate": 3.3195918367346935e-05, | |
| "loss": 0.9218, | |
| "step": 2290 | |
| }, | |
| { | |
| "epoch": 46.0, | |
| "learning_rate": 3.30734693877551e-05, | |
| "loss": 0.8695, | |
| "step": 2300 | |
| }, | |
| { | |
| "epoch": 46.2, | |
| "learning_rate": 3.295102040816327e-05, | |
| "loss": 0.9205, | |
| "step": 2310 | |
| }, | |
| { | |
| "epoch": 46.4, | |
| "learning_rate": 3.282857142857143e-05, | |
| "loss": 0.8813, | |
| "step": 2320 | |
| }, | |
| { | |
| "epoch": 46.6, | |
| "learning_rate": 3.2706122448979594e-05, | |
| "loss": 0.881, | |
| "step": 2330 | |
| }, | |
| { | |
| "epoch": 46.8, | |
| "learning_rate": 3.258367346938776e-05, | |
| "loss": 0.8744, | |
| "step": 2340 | |
| }, | |
| { | |
| "epoch": 47.0, | |
| "learning_rate": 3.246122448979592e-05, | |
| "loss": 0.8419, | |
| "step": 2350 | |
| }, | |
| { | |
| "epoch": 47.2, | |
| "learning_rate": 3.233877551020408e-05, | |
| "loss": 0.8919, | |
| "step": 2360 | |
| }, | |
| { | |
| "epoch": 47.4, | |
| "learning_rate": 3.2216326530612246e-05, | |
| "loss": 0.8741, | |
| "step": 2370 | |
| }, | |
| { | |
| "epoch": 47.6, | |
| "learning_rate": 3.209387755102041e-05, | |
| "loss": 0.8904, | |
| "step": 2380 | |
| }, | |
| { | |
| "epoch": 47.8, | |
| "learning_rate": 3.197142857142857e-05, | |
| "loss": 0.8948, | |
| "step": 2390 | |
| }, | |
| { | |
| "epoch": 48.0, | |
| "learning_rate": 3.1848979591836735e-05, | |
| "loss": 0.8513, | |
| "step": 2400 | |
| }, | |
| { | |
| "epoch": 48.2, | |
| "learning_rate": 3.17265306122449e-05, | |
| "loss": 0.89, | |
| "step": 2410 | |
| }, | |
| { | |
| "epoch": 48.4, | |
| "learning_rate": 3.160408163265306e-05, | |
| "loss": 0.8818, | |
| "step": 2420 | |
| }, | |
| { | |
| "epoch": 48.6, | |
| "learning_rate": 3.1481632653061224e-05, | |
| "loss": 0.8772, | |
| "step": 2430 | |
| }, | |
| { | |
| "epoch": 48.8, | |
| "learning_rate": 3.1359183673469394e-05, | |
| "loss": 0.8851, | |
| "step": 2440 | |
| }, | |
| { | |
| "epoch": 49.0, | |
| "learning_rate": 3.123673469387755e-05, | |
| "loss": 0.8568, | |
| "step": 2450 | |
| }, | |
| { | |
| "epoch": 49.2, | |
| "learning_rate": 3.1114285714285714e-05, | |
| "loss": 0.8949, | |
| "step": 2460 | |
| }, | |
| { | |
| "epoch": 49.4, | |
| "learning_rate": 3.099183673469388e-05, | |
| "loss": 0.8748, | |
| "step": 2470 | |
| }, | |
| { | |
| "epoch": 49.6, | |
| "learning_rate": 3.086938775510204e-05, | |
| "loss": 0.861, | |
| "step": 2480 | |
| }, | |
| { | |
| "epoch": 49.8, | |
| "learning_rate": 3.074693877551021e-05, | |
| "loss": 0.8878, | |
| "step": 2490 | |
| }, | |
| { | |
| "epoch": 50.0, | |
| "learning_rate": 3.0624489795918366e-05, | |
| "loss": 0.8399, | |
| "step": 2500 | |
| }, | |
| { | |
| "epoch": 50.0, | |
| "eval_loss": 0.3605365753173828, | |
| "eval_runtime": 244.9656, | |
| "eval_samples_per_second": 21.276, | |
| "eval_steps_per_second": 1.331, | |
| "eval_wer": 0.377206226896444, | |
| "step": 2500 | |
| }, | |
| { | |
| "epoch": 50.2, | |
| "learning_rate": 3.0502040816326532e-05, | |
| "loss": 0.877, | |
| "step": 2510 | |
| }, | |
| { | |
| "epoch": 50.4, | |
| "learning_rate": 3.0379591836734692e-05, | |
| "loss": 0.8843, | |
| "step": 2520 | |
| }, | |
| { | |
| "epoch": 50.6, | |
| "learning_rate": 3.0257142857142862e-05, | |
| "loss": 0.8654, | |
| "step": 2530 | |
| }, | |
| { | |
| "epoch": 50.8, | |
| "learning_rate": 3.013469387755102e-05, | |
| "loss": 0.8465, | |
| "step": 2540 | |
| }, | |
| { | |
| "epoch": 51.0, | |
| "learning_rate": 3.0012244897959184e-05, | |
| "loss": 0.8448, | |
| "step": 2550 | |
| }, | |
| { | |
| "epoch": 51.2, | |
| "learning_rate": 2.9889795918367348e-05, | |
| "loss": 0.8887, | |
| "step": 2560 | |
| }, | |
| { | |
| "epoch": 51.4, | |
| "learning_rate": 2.976734693877551e-05, | |
| "loss": 0.8518, | |
| "step": 2570 | |
| }, | |
| { | |
| "epoch": 51.6, | |
| "learning_rate": 2.9644897959183674e-05, | |
| "loss": 0.887, | |
| "step": 2580 | |
| }, | |
| { | |
| "epoch": 51.8, | |
| "learning_rate": 2.952244897959184e-05, | |
| "loss": 0.8628, | |
| "step": 2590 | |
| }, | |
| { | |
| "epoch": 52.0, | |
| "learning_rate": 2.94e-05, | |
| "loss": 0.8681, | |
| "step": 2600 | |
| }, | |
| { | |
| "epoch": 52.2, | |
| "learning_rate": 2.9277551020408166e-05, | |
| "loss": 0.8941, | |
| "step": 2610 | |
| }, | |
| { | |
| "epoch": 52.4, | |
| "learning_rate": 2.9155102040816326e-05, | |
| "loss": 0.8754, | |
| "step": 2620 | |
| }, | |
| { | |
| "epoch": 52.6, | |
| "learning_rate": 2.903265306122449e-05, | |
| "loss": 0.8573, | |
| "step": 2630 | |
| }, | |
| { | |
| "epoch": 52.8, | |
| "learning_rate": 2.8910204081632655e-05, | |
| "loss": 0.8689, | |
| "step": 2640 | |
| }, | |
| { | |
| "epoch": 53.0, | |
| "learning_rate": 2.8787755102040815e-05, | |
| "loss": 0.8664, | |
| "step": 2650 | |
| }, | |
| { | |
| "epoch": 53.2, | |
| "learning_rate": 2.866530612244898e-05, | |
| "loss": 0.8923, | |
| "step": 2660 | |
| }, | |
| { | |
| "epoch": 53.4, | |
| "learning_rate": 2.854285714285714e-05, | |
| "loss": 0.8735, | |
| "step": 2670 | |
| }, | |
| { | |
| "epoch": 53.6, | |
| "learning_rate": 2.8420408163265308e-05, | |
| "loss": 0.8717, | |
| "step": 2680 | |
| }, | |
| { | |
| "epoch": 53.8, | |
| "learning_rate": 2.829795918367347e-05, | |
| "loss": 0.8558, | |
| "step": 2690 | |
| }, | |
| { | |
| "epoch": 54.0, | |
| "learning_rate": 2.8175510204081634e-05, | |
| "loss": 0.8226, | |
| "step": 2700 | |
| }, | |
| { | |
| "epoch": 54.2, | |
| "learning_rate": 2.8053061224489797e-05, | |
| "loss": 0.8947, | |
| "step": 2710 | |
| }, | |
| { | |
| "epoch": 54.4, | |
| "learning_rate": 2.793061224489796e-05, | |
| "loss": 0.8509, | |
| "step": 2720 | |
| }, | |
| { | |
| "epoch": 54.6, | |
| "learning_rate": 2.7808163265306123e-05, | |
| "loss": 0.8551, | |
| "step": 2730 | |
| }, | |
| { | |
| "epoch": 54.8, | |
| "learning_rate": 2.7685714285714286e-05, | |
| "loss": 0.8705, | |
| "step": 2740 | |
| }, | |
| { | |
| "epoch": 55.0, | |
| "learning_rate": 2.756326530612245e-05, | |
| "loss": 0.815, | |
| "step": 2750 | |
| }, | |
| { | |
| "epoch": 55.2, | |
| "learning_rate": 2.7440816326530612e-05, | |
| "loss": 0.8939, | |
| "step": 2760 | |
| }, | |
| { | |
| "epoch": 55.4, | |
| "learning_rate": 2.731836734693878e-05, | |
| "loss": 0.859, | |
| "step": 2770 | |
| }, | |
| { | |
| "epoch": 55.6, | |
| "learning_rate": 2.7195918367346938e-05, | |
| "loss": 0.8567, | |
| "step": 2780 | |
| }, | |
| { | |
| "epoch": 55.8, | |
| "learning_rate": 2.7073469387755105e-05, | |
| "loss": 0.8599, | |
| "step": 2790 | |
| }, | |
| { | |
| "epoch": 56.0, | |
| "learning_rate": 2.6951020408163268e-05, | |
| "loss": 0.8171, | |
| "step": 2800 | |
| }, | |
| { | |
| "epoch": 56.2, | |
| "learning_rate": 2.6828571428571427e-05, | |
| "loss": 0.896, | |
| "step": 2810 | |
| }, | |
| { | |
| "epoch": 56.4, | |
| "learning_rate": 2.6706122448979594e-05, | |
| "loss": 0.8554, | |
| "step": 2820 | |
| }, | |
| { | |
| "epoch": 56.6, | |
| "learning_rate": 2.6583673469387753e-05, | |
| "loss": 0.877, | |
| "step": 2830 | |
| }, | |
| { | |
| "epoch": 56.8, | |
| "learning_rate": 2.646122448979592e-05, | |
| "loss": 0.8576, | |
| "step": 2840 | |
| }, | |
| { | |
| "epoch": 57.0, | |
| "learning_rate": 2.6338775510204083e-05, | |
| "loss": 0.8321, | |
| "step": 2850 | |
| }, | |
| { | |
| "epoch": 57.2, | |
| "learning_rate": 2.6216326530612246e-05, | |
| "loss": 0.8606, | |
| "step": 2860 | |
| }, | |
| { | |
| "epoch": 57.4, | |
| "learning_rate": 2.609387755102041e-05, | |
| "loss": 0.8409, | |
| "step": 2870 | |
| }, | |
| { | |
| "epoch": 57.6, | |
| "learning_rate": 2.5971428571428575e-05, | |
| "loss": 0.8469, | |
| "step": 2880 | |
| }, | |
| { | |
| "epoch": 57.8, | |
| "learning_rate": 2.5848979591836735e-05, | |
| "loss": 0.8546, | |
| "step": 2890 | |
| }, | |
| { | |
| "epoch": 58.0, | |
| "learning_rate": 2.5726530612244898e-05, | |
| "loss": 0.829, | |
| "step": 2900 | |
| }, | |
| { | |
| "epoch": 58.2, | |
| "learning_rate": 2.560408163265306e-05, | |
| "loss": 0.8659, | |
| "step": 2910 | |
| }, | |
| { | |
| "epoch": 58.4, | |
| "learning_rate": 2.5481632653061224e-05, | |
| "loss": 0.877, | |
| "step": 2920 | |
| }, | |
| { | |
| "epoch": 58.6, | |
| "learning_rate": 2.535918367346939e-05, | |
| "loss": 0.8537, | |
| "step": 2930 | |
| }, | |
| { | |
| "epoch": 58.8, | |
| "learning_rate": 2.523673469387755e-05, | |
| "loss": 0.8489, | |
| "step": 2940 | |
| }, | |
| { | |
| "epoch": 59.0, | |
| "learning_rate": 2.5114285714285717e-05, | |
| "loss": 0.8184, | |
| "step": 2950 | |
| }, | |
| { | |
| "epoch": 59.2, | |
| "learning_rate": 2.4991836734693876e-05, | |
| "loss": 0.8597, | |
| "step": 2960 | |
| }, | |
| { | |
| "epoch": 59.4, | |
| "learning_rate": 2.486938775510204e-05, | |
| "loss": 0.8621, | |
| "step": 2970 | |
| }, | |
| { | |
| "epoch": 59.6, | |
| "learning_rate": 2.4746938775510206e-05, | |
| "loss": 0.8553, | |
| "step": 2980 | |
| }, | |
| { | |
| "epoch": 59.8, | |
| "learning_rate": 2.4624489795918366e-05, | |
| "loss": 0.8628, | |
| "step": 2990 | |
| }, | |
| { | |
| "epoch": 60.0, | |
| "learning_rate": 2.4502040816326532e-05, | |
| "loss": 0.819, | |
| "step": 3000 | |
| }, | |
| { | |
| "epoch": 60.0, | |
| "eval_loss": 0.3622128367424011, | |
| "eval_runtime": 245.6109, | |
| "eval_samples_per_second": 21.221, | |
| "eval_steps_per_second": 1.327, | |
| "eval_wer": 0.3714270838756703, | |
| "step": 3000 | |
| }, | |
| { | |
| "epoch": 60.2, | |
| "learning_rate": 2.4379591836734695e-05, | |
| "loss": 0.8644, | |
| "step": 3010 | |
| }, | |
| { | |
| "epoch": 60.4, | |
| "learning_rate": 2.4257142857142858e-05, | |
| "loss": 0.8434, | |
| "step": 3020 | |
| }, | |
| { | |
| "epoch": 60.6, | |
| "learning_rate": 2.413469387755102e-05, | |
| "loss": 0.8402, | |
| "step": 3030 | |
| }, | |
| { | |
| "epoch": 60.8, | |
| "learning_rate": 2.4012244897959184e-05, | |
| "loss": 0.8412, | |
| "step": 3040 | |
| }, | |
| { | |
| "epoch": 61.0, | |
| "learning_rate": 2.3889795918367347e-05, | |
| "loss": 0.7999, | |
| "step": 3050 | |
| }, | |
| { | |
| "epoch": 61.2, | |
| "learning_rate": 2.376734693877551e-05, | |
| "loss": 0.8662, | |
| "step": 3060 | |
| }, | |
| { | |
| "epoch": 61.4, | |
| "learning_rate": 2.3644897959183673e-05, | |
| "loss": 0.8329, | |
| "step": 3070 | |
| }, | |
| { | |
| "epoch": 61.6, | |
| "learning_rate": 2.3522448979591837e-05, | |
| "loss": 0.8458, | |
| "step": 3080 | |
| }, | |
| { | |
| "epoch": 61.8, | |
| "learning_rate": 2.3400000000000003e-05, | |
| "loss": 0.8423, | |
| "step": 3090 | |
| }, | |
| { | |
| "epoch": 62.0, | |
| "learning_rate": 2.3277551020408163e-05, | |
| "loss": 0.8163, | |
| "step": 3100 | |
| }, | |
| { | |
| "epoch": 62.2, | |
| "learning_rate": 2.315510204081633e-05, | |
| "loss": 0.8645, | |
| "step": 3110 | |
| }, | |
| { | |
| "epoch": 62.4, | |
| "learning_rate": 2.303265306122449e-05, | |
| "loss": 0.8425, | |
| "step": 3120 | |
| }, | |
| { | |
| "epoch": 62.6, | |
| "learning_rate": 2.2910204081632655e-05, | |
| "loss": 0.8474, | |
| "step": 3130 | |
| }, | |
| { | |
| "epoch": 62.8, | |
| "learning_rate": 2.2787755102040818e-05, | |
| "loss": 0.8289, | |
| "step": 3140 | |
| }, | |
| { | |
| "epoch": 63.0, | |
| "learning_rate": 2.2665306122448978e-05, | |
| "loss": 0.8156, | |
| "step": 3150 | |
| }, | |
| { | |
| "epoch": 63.2, | |
| "learning_rate": 2.2542857142857144e-05, | |
| "loss": 0.8886, | |
| "step": 3160 | |
| }, | |
| { | |
| "epoch": 63.4, | |
| "learning_rate": 2.243265306122449e-05, | |
| "loss": 0.8355, | |
| "step": 3170 | |
| }, | |
| { | |
| "epoch": 63.6, | |
| "learning_rate": 2.2310204081632654e-05, | |
| "loss": 0.8291, | |
| "step": 3180 | |
| }, | |
| { | |
| "epoch": 63.8, | |
| "learning_rate": 2.2187755102040817e-05, | |
| "loss": 0.8588, | |
| "step": 3190 | |
| }, | |
| { | |
| "epoch": 64.0, | |
| "learning_rate": 2.206530612244898e-05, | |
| "loss": 0.8194, | |
| "step": 3200 | |
| }, | |
| { | |
| "epoch": 64.2, | |
| "learning_rate": 2.1942857142857143e-05, | |
| "loss": 0.8688, | |
| "step": 3210 | |
| }, | |
| { | |
| "epoch": 64.4, | |
| "learning_rate": 2.1820408163265306e-05, | |
| "loss": 0.8319, | |
| "step": 3220 | |
| }, | |
| { | |
| "epoch": 64.6, | |
| "learning_rate": 2.169795918367347e-05, | |
| "loss": 0.8505, | |
| "step": 3230 | |
| }, | |
| { | |
| "epoch": 64.8, | |
| "learning_rate": 2.1575510204081632e-05, | |
| "loss": 0.8435, | |
| "step": 3240 | |
| }, | |
| { | |
| "epoch": 65.0, | |
| "learning_rate": 2.14530612244898e-05, | |
| "loss": 0.8047, | |
| "step": 3250 | |
| }, | |
| { | |
| "epoch": 65.2, | |
| "learning_rate": 2.133061224489796e-05, | |
| "loss": 0.8568, | |
| "step": 3260 | |
| }, | |
| { | |
| "epoch": 65.4, | |
| "learning_rate": 2.1208163265306125e-05, | |
| "loss": 0.8305, | |
| "step": 3270 | |
| }, | |
| { | |
| "epoch": 65.6, | |
| "learning_rate": 2.1085714285714285e-05, | |
| "loss": 0.8424, | |
| "step": 3280 | |
| }, | |
| { | |
| "epoch": 65.8, | |
| "learning_rate": 2.096326530612245e-05, | |
| "loss": 0.8319, | |
| "step": 3290 | |
| }, | |
| { | |
| "epoch": 66.0, | |
| "learning_rate": 2.0840816326530614e-05, | |
| "loss": 0.823, | |
| "step": 3300 | |
| }, | |
| { | |
| "epoch": 66.2, | |
| "learning_rate": 2.0718367346938774e-05, | |
| "loss": 0.8754, | |
| "step": 3310 | |
| }, | |
| { | |
| "epoch": 66.4, | |
| "learning_rate": 2.059591836734694e-05, | |
| "loss": 0.8253, | |
| "step": 3320 | |
| }, | |
| { | |
| "epoch": 66.6, | |
| "learning_rate": 2.04734693877551e-05, | |
| "loss": 0.8429, | |
| "step": 3330 | |
| }, | |
| { | |
| "epoch": 66.8, | |
| "learning_rate": 2.0351020408163266e-05, | |
| "loss": 0.8286, | |
| "step": 3340 | |
| }, | |
| { | |
| "epoch": 67.0, | |
| "learning_rate": 2.022857142857143e-05, | |
| "loss": 0.8149, | |
| "step": 3350 | |
| }, | |
| { | |
| "epoch": 67.2, | |
| "learning_rate": 2.0106122448979593e-05, | |
| "loss": 0.8683, | |
| "step": 3360 | |
| }, | |
| { | |
| "epoch": 67.4, | |
| "learning_rate": 1.9983673469387756e-05, | |
| "loss": 0.8284, | |
| "step": 3370 | |
| }, | |
| { | |
| "epoch": 67.6, | |
| "learning_rate": 1.986122448979592e-05, | |
| "loss": 0.8363, | |
| "step": 3380 | |
| }, | |
| { | |
| "epoch": 67.8, | |
| "learning_rate": 1.973877551020408e-05, | |
| "loss": 0.8364, | |
| "step": 3390 | |
| }, | |
| { | |
| "epoch": 68.0, | |
| "learning_rate": 1.9616326530612245e-05, | |
| "loss": 0.8126, | |
| "step": 3400 | |
| }, | |
| { | |
| "epoch": 68.2, | |
| "learning_rate": 1.9493877551020408e-05, | |
| "loss": 0.8599, | |
| "step": 3410 | |
| }, | |
| { | |
| "epoch": 68.4, | |
| "learning_rate": 1.937142857142857e-05, | |
| "loss": 0.8182, | |
| "step": 3420 | |
| }, | |
| { | |
| "epoch": 68.6, | |
| "learning_rate": 1.9248979591836737e-05, | |
| "loss": 0.8344, | |
| "step": 3430 | |
| }, | |
| { | |
| "epoch": 68.8, | |
| "learning_rate": 1.9126530612244897e-05, | |
| "loss": 0.8158, | |
| "step": 3440 | |
| }, | |
| { | |
| "epoch": 69.0, | |
| "learning_rate": 1.9004081632653063e-05, | |
| "loss": 0.8117, | |
| "step": 3450 | |
| }, | |
| { | |
| "epoch": 69.2, | |
| "learning_rate": 1.8881632653061226e-05, | |
| "loss": 0.8421, | |
| "step": 3460 | |
| }, | |
| { | |
| "epoch": 69.4, | |
| "learning_rate": 1.8759183673469386e-05, | |
| "loss": 0.8264, | |
| "step": 3470 | |
| }, | |
| { | |
| "epoch": 69.6, | |
| "learning_rate": 1.8636734693877553e-05, | |
| "loss": 0.8356, | |
| "step": 3480 | |
| }, | |
| { | |
| "epoch": 69.8, | |
| "learning_rate": 1.8514285714285712e-05, | |
| "loss": 0.8316, | |
| "step": 3490 | |
| }, | |
| { | |
| "epoch": 70.0, | |
| "learning_rate": 1.839183673469388e-05, | |
| "loss": 0.8029, | |
| "step": 3500 | |
| }, | |
| { | |
| "epoch": 70.0, | |
| "eval_loss": 0.3561089038848877, | |
| "eval_runtime": 245.8818, | |
| "eval_samples_per_second": 21.197, | |
| "eval_steps_per_second": 1.326, | |
| "eval_wer": 0.3663768417764357, | |
| "step": 3500 | |
| }, | |
| { | |
| "epoch": 70.2, | |
| "learning_rate": 3.934545454545455e-05, | |
| "loss": 0.8492, | |
| "step": 3510 | |
| }, | |
| { | |
| "epoch": 70.4, | |
| "learning_rate": 3.928484848484849e-05, | |
| "loss": 0.8477, | |
| "step": 3520 | |
| }, | |
| { | |
| "epoch": 70.6, | |
| "learning_rate": 3.9224242424242426e-05, | |
| "loss": 0.8657, | |
| "step": 3530 | |
| }, | |
| { | |
| "epoch": 70.8, | |
| "learning_rate": 3.9163636363636364e-05, | |
| "loss": 0.8734, | |
| "step": 3540 | |
| }, | |
| { | |
| "epoch": 71.0, | |
| "learning_rate": 3.91030303030303e-05, | |
| "loss": 0.8092, | |
| "step": 3550 | |
| }, | |
| { | |
| "epoch": 71.2, | |
| "learning_rate": 3.904242424242424e-05, | |
| "loss": 0.8904, | |
| "step": 3560 | |
| }, | |
| { | |
| "epoch": 71.4, | |
| "learning_rate": 3.898181818181818e-05, | |
| "loss": 0.8585, | |
| "step": 3570 | |
| }, | |
| { | |
| "epoch": 71.6, | |
| "learning_rate": 3.892121212121212e-05, | |
| "loss": 0.8565, | |
| "step": 3580 | |
| }, | |
| { | |
| "epoch": 71.8, | |
| "learning_rate": 3.8860606060606056e-05, | |
| "loss": 0.8823, | |
| "step": 3590 | |
| }, | |
| { | |
| "epoch": 72.0, | |
| "learning_rate": 3.8799999999999994e-05, | |
| "loss": 0.8282, | |
| "step": 3600 | |
| }, | |
| { | |
| "epoch": 72.2, | |
| "learning_rate": 3.8739393939393946e-05, | |
| "loss": 0.8659, | |
| "step": 3610 | |
| }, | |
| { | |
| "epoch": 72.4, | |
| "learning_rate": 3.8678787878787885e-05, | |
| "loss": 0.8746, | |
| "step": 3620 | |
| }, | |
| { | |
| "epoch": 72.6, | |
| "learning_rate": 3.861818181818182e-05, | |
| "loss": 0.8536, | |
| "step": 3630 | |
| }, | |
| { | |
| "epoch": 72.8, | |
| "learning_rate": 3.855757575757576e-05, | |
| "loss": 0.8535, | |
| "step": 3640 | |
| }, | |
| { | |
| "epoch": 73.0, | |
| "learning_rate": 3.84969696969697e-05, | |
| "loss": 0.8332, | |
| "step": 3650 | |
| }, | |
| { | |
| "epoch": 73.2, | |
| "learning_rate": 3.843636363636364e-05, | |
| "loss": 0.8779, | |
| "step": 3660 | |
| }, | |
| { | |
| "epoch": 73.4, | |
| "learning_rate": 3.8375757575757576e-05, | |
| "loss": 0.85, | |
| "step": 3670 | |
| }, | |
| { | |
| "epoch": 73.6, | |
| "learning_rate": 3.8315151515151515e-05, | |
| "loss": 0.8533, | |
| "step": 3680 | |
| }, | |
| { | |
| "epoch": 73.8, | |
| "learning_rate": 3.825454545454545e-05, | |
| "loss": 0.8542, | |
| "step": 3690 | |
| }, | |
| { | |
| "epoch": 74.0, | |
| "learning_rate": 3.819393939393939e-05, | |
| "loss": 0.8341, | |
| "step": 3700 | |
| }, | |
| { | |
| "epoch": 74.2, | |
| "learning_rate": 3.8133333333333336e-05, | |
| "loss": 0.8832, | |
| "step": 3710 | |
| }, | |
| { | |
| "epoch": 74.4, | |
| "learning_rate": 3.8072727272727275e-05, | |
| "loss": 0.8531, | |
| "step": 3720 | |
| }, | |
| { | |
| "epoch": 74.6, | |
| "learning_rate": 3.801212121212121e-05, | |
| "loss": 0.8413, | |
| "step": 3730 | |
| }, | |
| { | |
| "epoch": 74.8, | |
| "learning_rate": 3.795151515151515e-05, | |
| "loss": 0.8696, | |
| "step": 3740 | |
| }, | |
| { | |
| "epoch": 75.0, | |
| "learning_rate": 3.789090909090909e-05, | |
| "loss": 0.8196, | |
| "step": 3750 | |
| }, | |
| { | |
| "epoch": 75.2, | |
| "learning_rate": 3.783030303030303e-05, | |
| "loss": 0.8754, | |
| "step": 3760 | |
| }, | |
| { | |
| "epoch": 75.4, | |
| "learning_rate": 3.7769696969696966e-05, | |
| "loss": 0.8501, | |
| "step": 3770 | |
| }, | |
| { | |
| "epoch": 75.6, | |
| "learning_rate": 3.770909090909091e-05, | |
| "loss": 0.8606, | |
| "step": 3780 | |
| }, | |
| { | |
| "epoch": 75.8, | |
| "learning_rate": 3.764848484848485e-05, | |
| "loss": 0.836, | |
| "step": 3790 | |
| }, | |
| { | |
| "epoch": 76.0, | |
| "learning_rate": 3.758787878787879e-05, | |
| "loss": 0.814, | |
| "step": 3800 | |
| }, | |
| { | |
| "epoch": 76.2, | |
| "learning_rate": 3.752727272727273e-05, | |
| "loss": 0.8881, | |
| "step": 3810 | |
| }, | |
| { | |
| "epoch": 76.4, | |
| "learning_rate": 3.746666666666667e-05, | |
| "loss": 0.8602, | |
| "step": 3820 | |
| }, | |
| { | |
| "epoch": 76.6, | |
| "learning_rate": 3.740606060606061e-05, | |
| "loss": 0.8663, | |
| "step": 3830 | |
| }, | |
| { | |
| "epoch": 76.8, | |
| "learning_rate": 3.734545454545455e-05, | |
| "loss": 0.8413, | |
| "step": 3840 | |
| }, | |
| { | |
| "epoch": 77.0, | |
| "learning_rate": 3.7284848484848487e-05, | |
| "loss": 0.8251, | |
| "step": 3850 | |
| }, | |
| { | |
| "epoch": 77.2, | |
| "learning_rate": 3.7224242424242425e-05, | |
| "loss": 0.8757, | |
| "step": 3860 | |
| }, | |
| { | |
| "epoch": 77.4, | |
| "learning_rate": 3.716363636363636e-05, | |
| "loss": 0.8527, | |
| "step": 3870 | |
| }, | |
| { | |
| "epoch": 77.6, | |
| "learning_rate": 3.71030303030303e-05, | |
| "loss": 0.8572, | |
| "step": 3880 | |
| }, | |
| { | |
| "epoch": 77.8, | |
| "learning_rate": 3.704848484848485e-05, | |
| "loss": 0.8577, | |
| "step": 3890 | |
| }, | |
| { | |
| "epoch": 78.0, | |
| "learning_rate": 3.698787878787879e-05, | |
| "loss": 0.8164, | |
| "step": 3900 | |
| }, | |
| { | |
| "epoch": 78.2, | |
| "learning_rate": 3.692727272727273e-05, | |
| "loss": 0.8743, | |
| "step": 3910 | |
| }, | |
| { | |
| "epoch": 78.4, | |
| "learning_rate": 3.686666666666667e-05, | |
| "loss": 0.8553, | |
| "step": 3920 | |
| }, | |
| { | |
| "epoch": 78.6, | |
| "learning_rate": 3.680606060606061e-05, | |
| "loss": 0.8113, | |
| "step": 3930 | |
| }, | |
| { | |
| "epoch": 78.8, | |
| "learning_rate": 3.674545454545455e-05, | |
| "loss": 0.867, | |
| "step": 3940 | |
| }, | |
| { | |
| "epoch": 79.0, | |
| "learning_rate": 3.6684848484848486e-05, | |
| "loss": 0.8343, | |
| "step": 3950 | |
| }, | |
| { | |
| "epoch": 79.2, | |
| "learning_rate": 3.6624242424242424e-05, | |
| "loss": 0.8653, | |
| "step": 3960 | |
| }, | |
| { | |
| "epoch": 79.4, | |
| "learning_rate": 3.656363636363636e-05, | |
| "loss": 0.8443, | |
| "step": 3970 | |
| }, | |
| { | |
| "epoch": 79.6, | |
| "learning_rate": 3.65030303030303e-05, | |
| "loss": 0.8488, | |
| "step": 3980 | |
| }, | |
| { | |
| "epoch": 79.8, | |
| "learning_rate": 3.644242424242424e-05, | |
| "loss": 0.8573, | |
| "step": 3990 | |
| }, | |
| { | |
| "epoch": 80.0, | |
| "learning_rate": 3.638181818181818e-05, | |
| "loss": 0.8104, | |
| "step": 4000 | |
| }, | |
| { | |
| "epoch": 80.0, | |
| "eval_loss": 0.35952043533325195, | |
| "eval_runtime": 247.7629, | |
| "eval_samples_per_second": 21.036, | |
| "eval_steps_per_second": 1.316, | |
| "eval_wer": 0.36596032696412767, | |
| "step": 4000 | |
| }, | |
| { | |
| "epoch": 80.2, | |
| "learning_rate": 3.632121212121212e-05, | |
| "loss": 0.8759, | |
| "step": 4010 | |
| }, | |
| { | |
| "epoch": 80.4, | |
| "learning_rate": 3.626060606060606e-05, | |
| "loss": 0.835, | |
| "step": 4020 | |
| }, | |
| { | |
| "epoch": 80.6, | |
| "learning_rate": 3.6200000000000006e-05, | |
| "loss": 0.8466, | |
| "step": 4030 | |
| }, | |
| { | |
| "epoch": 80.8, | |
| "learning_rate": 3.6139393939393944e-05, | |
| "loss": 0.8545, | |
| "step": 4040 | |
| }, | |
| { | |
| "epoch": 81.0, | |
| "learning_rate": 3.607878787878788e-05, | |
| "loss": 0.8075, | |
| "step": 4050 | |
| }, | |
| { | |
| "epoch": 81.2, | |
| "learning_rate": 3.601818181818182e-05, | |
| "loss": 0.8679, | |
| "step": 4060 | |
| }, | |
| { | |
| "epoch": 81.4, | |
| "learning_rate": 3.595757575757576e-05, | |
| "loss": 0.8332, | |
| "step": 4070 | |
| }, | |
| { | |
| "epoch": 81.6, | |
| "learning_rate": 3.58969696969697e-05, | |
| "loss": 0.85, | |
| "step": 4080 | |
| }, | |
| { | |
| "epoch": 81.8, | |
| "learning_rate": 3.5836363636363636e-05, | |
| "loss": 0.861, | |
| "step": 4090 | |
| }, | |
| { | |
| "epoch": 82.0, | |
| "learning_rate": 3.5775757575757574e-05, | |
| "loss": 0.8356, | |
| "step": 4100 | |
| }, | |
| { | |
| "epoch": 82.2, | |
| "learning_rate": 3.571515151515152e-05, | |
| "loss": 0.8513, | |
| "step": 4110 | |
| }, | |
| { | |
| "epoch": 82.4, | |
| "learning_rate": 3.565454545454546e-05, | |
| "loss": 0.8467, | |
| "step": 4120 | |
| }, | |
| { | |
| "epoch": 82.6, | |
| "learning_rate": 3.5593939393939396e-05, | |
| "loss": 0.8141, | |
| "step": 4130 | |
| }, | |
| { | |
| "epoch": 82.8, | |
| "learning_rate": 3.5533333333333334e-05, | |
| "loss": 0.8386, | |
| "step": 4140 | |
| }, | |
| { | |
| "epoch": 83.0, | |
| "learning_rate": 3.547272727272727e-05, | |
| "loss": 0.8169, | |
| "step": 4150 | |
| }, | |
| { | |
| "epoch": 83.2, | |
| "learning_rate": 3.541212121212121e-05, | |
| "loss": 0.8609, | |
| "step": 4160 | |
| }, | |
| { | |
| "epoch": 83.4, | |
| "learning_rate": 3.535151515151515e-05, | |
| "loss": 0.8391, | |
| "step": 4170 | |
| }, | |
| { | |
| "epoch": 83.6, | |
| "learning_rate": 3.529090909090909e-05, | |
| "loss": 0.8345, | |
| "step": 4180 | |
| }, | |
| { | |
| "epoch": 83.8, | |
| "learning_rate": 3.5230303030303026e-05, | |
| "loss": 0.8368, | |
| "step": 4190 | |
| }, | |
| { | |
| "epoch": 84.0, | |
| "learning_rate": 3.516969696969697e-05, | |
| "loss": 0.8145, | |
| "step": 4200 | |
| }, | |
| { | |
| "epoch": 84.2, | |
| "learning_rate": 3.5109090909090916e-05, | |
| "loss": 0.8446, | |
| "step": 4210 | |
| }, | |
| { | |
| "epoch": 84.4, | |
| "learning_rate": 3.5048484848484854e-05, | |
| "loss": 0.8176, | |
| "step": 4220 | |
| }, | |
| { | |
| "epoch": 84.6, | |
| "learning_rate": 3.498787878787879e-05, | |
| "loss": 0.8356, | |
| "step": 4230 | |
| }, | |
| { | |
| "epoch": 84.8, | |
| "learning_rate": 3.492727272727273e-05, | |
| "loss": 0.8251, | |
| "step": 4240 | |
| }, | |
| { | |
| "epoch": 85.0, | |
| "learning_rate": 3.486666666666667e-05, | |
| "loss": 0.8151, | |
| "step": 4250 | |
| }, | |
| { | |
| "epoch": 85.2, | |
| "learning_rate": 3.480606060606061e-05, | |
| "loss": 0.8549, | |
| "step": 4260 | |
| }, | |
| { | |
| "epoch": 85.4, | |
| "learning_rate": 3.4745454545454546e-05, | |
| "loss": 0.8515, | |
| "step": 4270 | |
| }, | |
| { | |
| "epoch": 85.6, | |
| "learning_rate": 3.4684848484848484e-05, | |
| "loss": 0.8255, | |
| "step": 4280 | |
| }, | |
| { | |
| "epoch": 85.8, | |
| "learning_rate": 3.462424242424242e-05, | |
| "loss": 0.8325, | |
| "step": 4290 | |
| }, | |
| { | |
| "epoch": 86.0, | |
| "learning_rate": 3.456363636363636e-05, | |
| "loss": 0.8162, | |
| "step": 4300 | |
| }, | |
| { | |
| "epoch": 86.2, | |
| "learning_rate": 3.4503030303030306e-05, | |
| "loss": 0.8628, | |
| "step": 4310 | |
| }, | |
| { | |
| "epoch": 86.4, | |
| "learning_rate": 3.4442424242424244e-05, | |
| "loss": 0.8125, | |
| "step": 4320 | |
| }, | |
| { | |
| "epoch": 86.6, | |
| "learning_rate": 3.438181818181818e-05, | |
| "loss": 0.8436, | |
| "step": 4330 | |
| }, | |
| { | |
| "epoch": 86.8, | |
| "learning_rate": 3.432121212121212e-05, | |
| "loss": 0.8343, | |
| "step": 4340 | |
| }, | |
| { | |
| "epoch": 87.0, | |
| "learning_rate": 3.426060606060606e-05, | |
| "loss": 0.8063, | |
| "step": 4350 | |
| }, | |
| { | |
| "epoch": 87.2, | |
| "learning_rate": 3.42e-05, | |
| "loss": 0.8492, | |
| "step": 4360 | |
| }, | |
| { | |
| "epoch": 87.4, | |
| "learning_rate": 3.4139393939393936e-05, | |
| "loss": 0.8393, | |
| "step": 4370 | |
| }, | |
| { | |
| "epoch": 87.6, | |
| "learning_rate": 3.407878787878788e-05, | |
| "loss": 0.827, | |
| "step": 4380 | |
| }, | |
| { | |
| "epoch": 87.8, | |
| "learning_rate": 3.401818181818182e-05, | |
| "loss": 0.8371, | |
| "step": 4390 | |
| }, | |
| { | |
| "epoch": 88.0, | |
| "learning_rate": 3.395757575757576e-05, | |
| "loss": 0.8047, | |
| "step": 4400 | |
| }, | |
| { | |
| "epoch": 88.2, | |
| "learning_rate": 3.3896969696969696e-05, | |
| "loss": 0.8506, | |
| "step": 4410 | |
| }, | |
| { | |
| "epoch": 88.4, | |
| "learning_rate": 3.383636363636364e-05, | |
| "loss": 0.8203, | |
| "step": 4420 | |
| }, | |
| { | |
| "epoch": 88.6, | |
| "learning_rate": 3.377575757575758e-05, | |
| "loss": 0.8127, | |
| "step": 4430 | |
| }, | |
| { | |
| "epoch": 88.8, | |
| "learning_rate": 3.371515151515152e-05, | |
| "loss": 0.8318, | |
| "step": 4440 | |
| }, | |
| { | |
| "epoch": 89.0, | |
| "learning_rate": 3.3654545454545456e-05, | |
| "loss": 0.801, | |
| "step": 4450 | |
| }, | |
| { | |
| "epoch": 89.2, | |
| "learning_rate": 3.3593939393939395e-05, | |
| "loss": 0.8433, | |
| "step": 4460 | |
| }, | |
| { | |
| "epoch": 89.4, | |
| "learning_rate": 3.353333333333333e-05, | |
| "loss": 0.8188, | |
| "step": 4470 | |
| }, | |
| { | |
| "epoch": 89.6, | |
| "learning_rate": 3.347272727272727e-05, | |
| "loss": 0.8296, | |
| "step": 4480 | |
| }, | |
| { | |
| "epoch": 89.8, | |
| "learning_rate": 3.341212121212121e-05, | |
| "loss": 0.8313, | |
| "step": 4490 | |
| }, | |
| { | |
| "epoch": 90.0, | |
| "learning_rate": 3.335151515151515e-05, | |
| "loss": 0.8118, | |
| "step": 4500 | |
| }, | |
| { | |
| "epoch": 90.0, | |
| "eval_loss": 0.3459583818912506, | |
| "eval_runtime": 243.3931, | |
| "eval_samples_per_second": 21.414, | |
| "eval_steps_per_second": 1.339, | |
| "eval_wer": 0.3591659290883532, | |
| "step": 4500 | |
| }, | |
| { | |
| "epoch": 90.2, | |
| "learning_rate": 3.3290909090909086e-05, | |
| "loss": 0.8586, | |
| "step": 4510 | |
| }, | |
| { | |
| "epoch": 90.4, | |
| "learning_rate": 3.323030303030303e-05, | |
| "loss": 0.8303, | |
| "step": 4520 | |
| }, | |
| { | |
| "epoch": 90.6, | |
| "learning_rate": 3.316969696969697e-05, | |
| "loss": 0.8199, | |
| "step": 4530 | |
| }, | |
| { | |
| "epoch": 90.8, | |
| "learning_rate": 3.310909090909091e-05, | |
| "loss": 0.8272, | |
| "step": 4540 | |
| }, | |
| { | |
| "epoch": 91.0, | |
| "learning_rate": 3.304848484848485e-05, | |
| "loss": 0.7925, | |
| "step": 4550 | |
| }, | |
| { | |
| "epoch": 91.2, | |
| "learning_rate": 3.298787878787879e-05, | |
| "loss": 0.8542, | |
| "step": 4560 | |
| }, | |
| { | |
| "epoch": 91.4, | |
| "learning_rate": 3.292727272727273e-05, | |
| "loss": 0.8222, | |
| "step": 4570 | |
| }, | |
| { | |
| "epoch": 91.6, | |
| "learning_rate": 3.286666666666667e-05, | |
| "loss": 0.7901, | |
| "step": 4580 | |
| }, | |
| { | |
| "epoch": 91.8, | |
| "learning_rate": 3.2806060606060607e-05, | |
| "loss": 0.8227, | |
| "step": 4590 | |
| }, | |
| { | |
| "epoch": 92.0, | |
| "learning_rate": 3.2745454545454545e-05, | |
| "loss": 0.7861, | |
| "step": 4600 | |
| }, | |
| { | |
| "epoch": 92.2, | |
| "learning_rate": 3.268484848484848e-05, | |
| "loss": 0.8382, | |
| "step": 4610 | |
| }, | |
| { | |
| "epoch": 92.4, | |
| "learning_rate": 3.262424242424243e-05, | |
| "loss": 0.8279, | |
| "step": 4620 | |
| }, | |
| { | |
| "epoch": 92.6, | |
| "learning_rate": 3.256363636363637e-05, | |
| "loss": 0.8148, | |
| "step": 4630 | |
| }, | |
| { | |
| "epoch": 92.8, | |
| "learning_rate": 3.2503030303030305e-05, | |
| "loss": 0.8054, | |
| "step": 4640 | |
| }, | |
| { | |
| "epoch": 93.0, | |
| "learning_rate": 3.244242424242424e-05, | |
| "loss": 0.8003, | |
| "step": 4650 | |
| }, | |
| { | |
| "epoch": 93.2, | |
| "learning_rate": 3.238181818181818e-05, | |
| "loss": 0.8252, | |
| "step": 4660 | |
| }, | |
| { | |
| "epoch": 93.4, | |
| "learning_rate": 3.232121212121212e-05, | |
| "loss": 0.8, | |
| "step": 4670 | |
| }, | |
| { | |
| "epoch": 93.6, | |
| "learning_rate": 3.226060606060606e-05, | |
| "loss": 0.811, | |
| "step": 4680 | |
| }, | |
| { | |
| "epoch": 93.8, | |
| "learning_rate": 3.22e-05, | |
| "loss": 0.8202, | |
| "step": 4690 | |
| }, | |
| { | |
| "epoch": 94.0, | |
| "learning_rate": 3.2139393939393935e-05, | |
| "loss": 0.7872, | |
| "step": 4700 | |
| }, | |
| { | |
| "epoch": 94.2, | |
| "learning_rate": 3.207878787878787e-05, | |
| "loss": 0.8567, | |
| "step": 4710 | |
| }, | |
| { | |
| "epoch": 94.4, | |
| "learning_rate": 3.2018181818181825e-05, | |
| "loss": 0.8226, | |
| "step": 4720 | |
| }, | |
| { | |
| "epoch": 94.6, | |
| "learning_rate": 3.1957575757575764e-05, | |
| "loss": 0.7977, | |
| "step": 4730 | |
| }, | |
| { | |
| "epoch": 94.8, | |
| "learning_rate": 3.18969696969697e-05, | |
| "loss": 0.8119, | |
| "step": 4740 | |
| }, | |
| { | |
| "epoch": 95.0, | |
| "learning_rate": 3.183636363636364e-05, | |
| "loss": 0.7972, | |
| "step": 4750 | |
| }, | |
| { | |
| "epoch": 95.2, | |
| "learning_rate": 3.177575757575758e-05, | |
| "loss": 0.8297, | |
| "step": 4760 | |
| }, | |
| { | |
| "epoch": 95.4, | |
| "learning_rate": 3.171515151515152e-05, | |
| "loss": 0.8184, | |
| "step": 4770 | |
| }, | |
| { | |
| "epoch": 95.6, | |
| "learning_rate": 3.1654545454545455e-05, | |
| "loss": 0.8319, | |
| "step": 4780 | |
| }, | |
| { | |
| "epoch": 95.8, | |
| "learning_rate": 3.1593939393939394e-05, | |
| "loss": 0.7927, | |
| "step": 4790 | |
| }, | |
| { | |
| "epoch": 96.0, | |
| "learning_rate": 3.153333333333333e-05, | |
| "loss": 0.781, | |
| "step": 4800 | |
| }, | |
| { | |
| "epoch": 96.2, | |
| "learning_rate": 3.147272727272727e-05, | |
| "loss": 0.8324, | |
| "step": 4810 | |
| }, | |
| { | |
| "epoch": 96.4, | |
| "learning_rate": 3.1412121212121215e-05, | |
| "loss": 0.8089, | |
| "step": 4820 | |
| }, | |
| { | |
| "epoch": 96.6, | |
| "learning_rate": 3.1351515151515154e-05, | |
| "loss": 0.8066, | |
| "step": 4830 | |
| }, | |
| { | |
| "epoch": 96.8, | |
| "learning_rate": 3.129090909090909e-05, | |
| "loss": 0.8038, | |
| "step": 4840 | |
| }, | |
| { | |
| "epoch": 97.0, | |
| "learning_rate": 3.123030303030303e-05, | |
| "loss": 0.7848, | |
| "step": 4850 | |
| }, | |
| { | |
| "epoch": 97.2, | |
| "learning_rate": 3.116969696969697e-05, | |
| "loss": 0.7951, | |
| "step": 4860 | |
| }, | |
| { | |
| "epoch": 97.4, | |
| "learning_rate": 3.110909090909091e-05, | |
| "loss": 0.8023, | |
| "step": 4870 | |
| }, | |
| { | |
| "epoch": 97.6, | |
| "learning_rate": 3.1048484848484845e-05, | |
| "loss": 0.8058, | |
| "step": 4880 | |
| }, | |
| { | |
| "epoch": 97.8, | |
| "learning_rate": 3.098787878787879e-05, | |
| "loss": 0.8003, | |
| "step": 4890 | |
| }, | |
| { | |
| "epoch": 98.0, | |
| "learning_rate": 3.092727272727273e-05, | |
| "loss": 0.7844, | |
| "step": 4900 | |
| }, | |
| { | |
| "epoch": 98.2, | |
| "learning_rate": 3.086666666666667e-05, | |
| "loss": 0.8175, | |
| "step": 4910 | |
| }, | |
| { | |
| "epoch": 98.4, | |
| "learning_rate": 3.0806060606060605e-05, | |
| "loss": 0.8186, | |
| "step": 4920 | |
| }, | |
| { | |
| "epoch": 98.6, | |
| "learning_rate": 3.074545454545455e-05, | |
| "loss": 0.8055, | |
| "step": 4930 | |
| }, | |
| { | |
| "epoch": 98.8, | |
| "learning_rate": 3.068484848484849e-05, | |
| "loss": 0.8004, | |
| "step": 4940 | |
| }, | |
| { | |
| "epoch": 99.0, | |
| "learning_rate": 3.062424242424243e-05, | |
| "loss": 0.7751, | |
| "step": 4950 | |
| }, | |
| { | |
| "epoch": 99.2, | |
| "learning_rate": 3.0563636363636365e-05, | |
| "loss": 0.8077, | |
| "step": 4960 | |
| }, | |
| { | |
| "epoch": 99.4, | |
| "learning_rate": 3.0503030303030304e-05, | |
| "loss": 0.7806, | |
| "step": 4970 | |
| }, | |
| { | |
| "epoch": 99.6, | |
| "learning_rate": 3.0442424242424242e-05, | |
| "loss": 0.8005, | |
| "step": 4980 | |
| }, | |
| { | |
| "epoch": 99.8, | |
| "learning_rate": 3.038181818181818e-05, | |
| "loss": 0.7966, | |
| "step": 4990 | |
| }, | |
| { | |
| "epoch": 100.0, | |
| "learning_rate": 3.0321212121212122e-05, | |
| "loss": 0.7831, | |
| "step": 5000 | |
| }, | |
| { | |
| "epoch": 100.0, | |
| "eval_loss": 0.35662171244621277, | |
| "eval_runtime": 244.6026, | |
| "eval_samples_per_second": 21.308, | |
| "eval_steps_per_second": 1.333, | |
| "eval_wer": 0.35934815431873796, | |
| "step": 5000 | |
| }, | |
| { | |
| "epoch": 100.2, | |
| "learning_rate": 3.026060606060606e-05, | |
| "loss": 0.8186, | |
| "step": 5010 | |
| }, | |
| { | |
| "epoch": 100.4, | |
| "learning_rate": 3.02e-05, | |
| "loss": 0.7988, | |
| "step": 5020 | |
| }, | |
| { | |
| "epoch": 100.6, | |
| "learning_rate": 3.0139393939393937e-05, | |
| "loss": 0.7912, | |
| "step": 5030 | |
| }, | |
| { | |
| "epoch": 100.8, | |
| "learning_rate": 3.0078787878787876e-05, | |
| "loss": 0.7922, | |
| "step": 5040 | |
| }, | |
| { | |
| "epoch": 101.0, | |
| "learning_rate": 3.0018181818181817e-05, | |
| "loss": 0.8013, | |
| "step": 5050 | |
| }, | |
| { | |
| "epoch": 101.2, | |
| "learning_rate": 2.995757575757576e-05, | |
| "loss": 0.8025, | |
| "step": 5060 | |
| }, | |
| { | |
| "epoch": 101.4, | |
| "learning_rate": 2.9896969696969697e-05, | |
| "loss": 0.8028, | |
| "step": 5070 | |
| }, | |
| { | |
| "epoch": 101.6, | |
| "learning_rate": 2.9836363636363636e-05, | |
| "loss": 0.8106, | |
| "step": 5080 | |
| }, | |
| { | |
| "epoch": 101.8, | |
| "learning_rate": 2.9775757575757574e-05, | |
| "loss": 0.8195, | |
| "step": 5090 | |
| }, | |
| { | |
| "epoch": 102.0, | |
| "learning_rate": 2.971515151515152e-05, | |
| "loss": 0.7637, | |
| "step": 5100 | |
| }, | |
| { | |
| "epoch": 102.2, | |
| "learning_rate": 2.9654545454545457e-05, | |
| "loss": 0.8283, | |
| "step": 5110 | |
| }, | |
| { | |
| "epoch": 102.4, | |
| "learning_rate": 2.9593939393939396e-05, | |
| "loss": 0.7809, | |
| "step": 5120 | |
| }, | |
| { | |
| "epoch": 102.6, | |
| "learning_rate": 2.9533333333333334e-05, | |
| "loss": 0.7841, | |
| "step": 5130 | |
| }, | |
| { | |
| "epoch": 102.8, | |
| "learning_rate": 2.9472727272727272e-05, | |
| "loss": 0.7974, | |
| "step": 5140 | |
| }, | |
| { | |
| "epoch": 103.0, | |
| "learning_rate": 2.9412121212121214e-05, | |
| "loss": 0.786, | |
| "step": 5150 | |
| }, | |
| { | |
| "epoch": 103.2, | |
| "learning_rate": 2.9351515151515152e-05, | |
| "loss": 0.8257, | |
| "step": 5160 | |
| }, | |
| { | |
| "epoch": 103.4, | |
| "learning_rate": 2.929090909090909e-05, | |
| "loss": 0.7662, | |
| "step": 5170 | |
| }, | |
| { | |
| "epoch": 103.6, | |
| "learning_rate": 2.923030303030303e-05, | |
| "loss": 0.7893, | |
| "step": 5180 | |
| }, | |
| { | |
| "epoch": 103.8, | |
| "learning_rate": 2.916969696969697e-05, | |
| "loss": 0.7976, | |
| "step": 5190 | |
| }, | |
| { | |
| "epoch": 104.0, | |
| "learning_rate": 2.910909090909091e-05, | |
| "loss": 0.7868, | |
| "step": 5200 | |
| }, | |
| { | |
| "epoch": 104.2, | |
| "learning_rate": 2.904848484848485e-05, | |
| "loss": 0.8219, | |
| "step": 5210 | |
| }, | |
| { | |
| "epoch": 104.4, | |
| "learning_rate": 2.898787878787879e-05, | |
| "loss": 0.7969, | |
| "step": 5220 | |
| }, | |
| { | |
| "epoch": 104.6, | |
| "learning_rate": 2.8927272727272728e-05, | |
| "loss": 0.7914, | |
| "step": 5230 | |
| }, | |
| { | |
| "epoch": 104.8, | |
| "learning_rate": 2.8866666666666666e-05, | |
| "loss": 0.7971, | |
| "step": 5240 | |
| }, | |
| { | |
| "epoch": 105.0, | |
| "learning_rate": 2.8806060606060604e-05, | |
| "loss": 0.7807, | |
| "step": 5250 | |
| }, | |
| { | |
| "epoch": 105.2, | |
| "learning_rate": 2.8745454545454546e-05, | |
| "loss": 0.803, | |
| "step": 5260 | |
| }, | |
| { | |
| "epoch": 105.4, | |
| "learning_rate": 2.8684848484848488e-05, | |
| "loss": 0.81, | |
| "step": 5270 | |
| }, | |
| { | |
| "epoch": 105.6, | |
| "learning_rate": 2.8624242424242426e-05, | |
| "loss": 0.7943, | |
| "step": 5280 | |
| }, | |
| { | |
| "epoch": 105.8, | |
| "learning_rate": 2.8563636363636364e-05, | |
| "loss": 0.7767, | |
| "step": 5290 | |
| }, | |
| { | |
| "epoch": 106.0, | |
| "learning_rate": 2.8503030303030303e-05, | |
| "loss": 0.7721, | |
| "step": 5300 | |
| }, | |
| { | |
| "epoch": 106.2, | |
| "learning_rate": 2.8442424242424244e-05, | |
| "loss": 0.7863, | |
| "step": 5310 | |
| }, | |
| { | |
| "epoch": 106.4, | |
| "learning_rate": 2.8381818181818183e-05, | |
| "loss": 0.7824, | |
| "step": 5320 | |
| }, | |
| { | |
| "epoch": 106.6, | |
| "learning_rate": 2.832121212121212e-05, | |
| "loss": 0.8002, | |
| "step": 5330 | |
| }, | |
| { | |
| "epoch": 106.8, | |
| "learning_rate": 2.826060606060606e-05, | |
| "loss": 0.7921, | |
| "step": 5340 | |
| }, | |
| { | |
| "epoch": 107.0, | |
| "learning_rate": 2.8199999999999998e-05, | |
| "loss": 0.7757, | |
| "step": 5350 | |
| }, | |
| { | |
| "epoch": 107.2, | |
| "learning_rate": 2.8139393939393943e-05, | |
| "loss": 0.8046, | |
| "step": 5360 | |
| }, | |
| { | |
| "epoch": 107.4, | |
| "learning_rate": 2.807878787878788e-05, | |
| "loss": 0.775, | |
| "step": 5370 | |
| }, | |
| { | |
| "epoch": 107.6, | |
| "learning_rate": 2.801818181818182e-05, | |
| "loss": 0.7975, | |
| "step": 5380 | |
| }, | |
| { | |
| "epoch": 107.8, | |
| "learning_rate": 2.7957575757575758e-05, | |
| "loss": 0.794, | |
| "step": 5390 | |
| }, | |
| { | |
| "epoch": 108.0, | |
| "learning_rate": 2.7896969696969696e-05, | |
| "loss": 0.7465, | |
| "step": 5400 | |
| }, | |
| { | |
| "epoch": 108.2, | |
| "learning_rate": 2.7836363636363638e-05, | |
| "loss": 0.8104, | |
| "step": 5410 | |
| }, | |
| { | |
| "epoch": 108.4, | |
| "learning_rate": 2.7775757575757576e-05, | |
| "loss": 0.7862, | |
| "step": 5420 | |
| }, | |
| { | |
| "epoch": 108.6, | |
| "learning_rate": 2.7715151515151514e-05, | |
| "loss": 0.7767, | |
| "step": 5430 | |
| }, | |
| { | |
| "epoch": 108.8, | |
| "learning_rate": 2.7654545454545456e-05, | |
| "loss": 0.7807, | |
| "step": 5440 | |
| }, | |
| { | |
| "epoch": 109.0, | |
| "learning_rate": 2.7593939393939395e-05, | |
| "loss": 0.7628, | |
| "step": 5450 | |
| }, | |
| { | |
| "epoch": 109.2, | |
| "learning_rate": 2.7533333333333336e-05, | |
| "loss": 0.8077, | |
| "step": 5460 | |
| }, | |
| { | |
| "epoch": 109.4, | |
| "learning_rate": 2.7472727272727275e-05, | |
| "loss": 0.8049, | |
| "step": 5470 | |
| }, | |
| { | |
| "epoch": 109.6, | |
| "learning_rate": 2.7412121212121213e-05, | |
| "loss": 0.777, | |
| "step": 5480 | |
| }, | |
| { | |
| "epoch": 109.8, | |
| "learning_rate": 2.735151515151515e-05, | |
| "loss": 0.7799, | |
| "step": 5490 | |
| }, | |
| { | |
| "epoch": 110.0, | |
| "learning_rate": 2.729090909090909e-05, | |
| "loss": 0.744, | |
| "step": 5500 | |
| }, | |
| { | |
| "epoch": 110.0, | |
| "eval_loss": 0.35784289240837097, | |
| "eval_runtime": 244.0889, | |
| "eval_samples_per_second": 21.353, | |
| "eval_steps_per_second": 1.336, | |
| "eval_wer": 0.35351694694642577, | |
| "step": 5500 | |
| }, | |
| { | |
| "epoch": 110.2, | |
| "learning_rate": 2.723030303030303e-05, | |
| "loss": 0.7943, | |
| "step": 5510 | |
| }, | |
| { | |
| "epoch": 110.4, | |
| "learning_rate": 2.716969696969697e-05, | |
| "loss": 0.7865, | |
| "step": 5520 | |
| }, | |
| { | |
| "epoch": 110.6, | |
| "learning_rate": 2.710909090909091e-05, | |
| "loss": 0.7929, | |
| "step": 5530 | |
| }, | |
| { | |
| "epoch": 110.8, | |
| "learning_rate": 2.704848484848485e-05, | |
| "loss": 0.7974, | |
| "step": 5540 | |
| }, | |
| { | |
| "epoch": 111.0, | |
| "learning_rate": 2.6987878787878788e-05, | |
| "loss": 0.765, | |
| "step": 5550 | |
| }, | |
| { | |
| "epoch": 111.2, | |
| "learning_rate": 2.692727272727273e-05, | |
| "loss": 0.7867, | |
| "step": 5560 | |
| }, | |
| { | |
| "epoch": 111.4, | |
| "learning_rate": 2.6866666666666668e-05, | |
| "loss": 0.7903, | |
| "step": 5570 | |
| }, | |
| { | |
| "epoch": 111.6, | |
| "learning_rate": 2.6806060606060606e-05, | |
| "loss": 0.7852, | |
| "step": 5580 | |
| }, | |
| { | |
| "epoch": 111.8, | |
| "learning_rate": 2.6745454545454545e-05, | |
| "loss": 0.7745, | |
| "step": 5590 | |
| }, | |
| { | |
| "epoch": 112.0, | |
| "learning_rate": 2.6684848484848483e-05, | |
| "loss": 0.7464, | |
| "step": 5600 | |
| }, | |
| { | |
| "epoch": 112.2, | |
| "learning_rate": 2.6624242424242428e-05, | |
| "loss": 0.8193, | |
| "step": 5610 | |
| }, | |
| { | |
| "epoch": 112.4, | |
| "learning_rate": 2.6563636363636366e-05, | |
| "loss": 0.7732, | |
| "step": 5620 | |
| }, | |
| { | |
| "epoch": 112.6, | |
| "learning_rate": 2.6503030303030305e-05, | |
| "loss": 0.7941, | |
| "step": 5630 | |
| }, | |
| { | |
| "epoch": 112.8, | |
| "learning_rate": 2.6442424242424243e-05, | |
| "loss": 0.7822, | |
| "step": 5640 | |
| }, | |
| { | |
| "epoch": 113.0, | |
| "learning_rate": 2.638181818181818e-05, | |
| "loss": 0.7487, | |
| "step": 5650 | |
| }, | |
| { | |
| "epoch": 113.2, | |
| "learning_rate": 2.632121212121212e-05, | |
| "loss": 0.7932, | |
| "step": 5660 | |
| }, | |
| { | |
| "epoch": 113.4, | |
| "learning_rate": 2.626060606060606e-05, | |
| "loss": 0.7677, | |
| "step": 5670 | |
| }, | |
| { | |
| "epoch": 113.6, | |
| "learning_rate": 2.62e-05, | |
| "loss": 0.7945, | |
| "step": 5680 | |
| }, | |
| { | |
| "epoch": 113.8, | |
| "learning_rate": 2.6139393939393938e-05, | |
| "loss": 0.781, | |
| "step": 5690 | |
| }, | |
| { | |
| "epoch": 114.0, | |
| "learning_rate": 2.607878787878788e-05, | |
| "loss": 0.7669, | |
| "step": 5700 | |
| }, | |
| { | |
| "epoch": 114.2, | |
| "learning_rate": 2.6018181818181818e-05, | |
| "loss": 0.7967, | |
| "step": 5710 | |
| }, | |
| { | |
| "epoch": 114.4, | |
| "learning_rate": 2.595757575757576e-05, | |
| "loss": 0.7637, | |
| "step": 5720 | |
| }, | |
| { | |
| "epoch": 114.6, | |
| "learning_rate": 2.58969696969697e-05, | |
| "loss": 0.766, | |
| "step": 5730 | |
| }, | |
| { | |
| "epoch": 114.8, | |
| "learning_rate": 2.5836363636363637e-05, | |
| "loss": 0.7611, | |
| "step": 5740 | |
| }, | |
| { | |
| "epoch": 115.0, | |
| "learning_rate": 2.5775757575757575e-05, | |
| "loss": 0.761, | |
| "step": 5750 | |
| }, | |
| { | |
| "epoch": 115.2, | |
| "learning_rate": 2.5715151515151513e-05, | |
| "loss": 0.7997, | |
| "step": 5760 | |
| }, | |
| { | |
| "epoch": 115.4, | |
| "learning_rate": 2.5654545454545455e-05, | |
| "loss": 0.773, | |
| "step": 5770 | |
| }, | |
| { | |
| "epoch": 115.6, | |
| "learning_rate": 2.5593939393939397e-05, | |
| "loss": 0.7727, | |
| "step": 5780 | |
| }, | |
| { | |
| "epoch": 115.8, | |
| "learning_rate": 2.5533333333333335e-05, | |
| "loss": 0.771, | |
| "step": 5790 | |
| }, | |
| { | |
| "epoch": 116.0, | |
| "learning_rate": 2.5472727272727273e-05, | |
| "loss": 0.7392, | |
| "step": 5800 | |
| }, | |
| { | |
| "epoch": 116.2, | |
| "learning_rate": 2.5412121212121212e-05, | |
| "loss": 0.795, | |
| "step": 5810 | |
| }, | |
| { | |
| "epoch": 116.4, | |
| "learning_rate": 2.5351515151515153e-05, | |
| "loss": 0.7673, | |
| "step": 5820 | |
| }, | |
| { | |
| "epoch": 116.6, | |
| "learning_rate": 2.5290909090909092e-05, | |
| "loss": 0.7486, | |
| "step": 5830 | |
| }, | |
| { | |
| "epoch": 116.8, | |
| "learning_rate": 2.523030303030303e-05, | |
| "loss": 0.78, | |
| "step": 5840 | |
| }, | |
| { | |
| "epoch": 117.0, | |
| "learning_rate": 2.516969696969697e-05, | |
| "loss": 0.7339, | |
| "step": 5850 | |
| }, | |
| { | |
| "epoch": 117.2, | |
| "learning_rate": 2.5109090909090907e-05, | |
| "loss": 0.8006, | |
| "step": 5860 | |
| }, | |
| { | |
| "epoch": 117.4, | |
| "learning_rate": 2.5048484848484852e-05, | |
| "loss": 0.7578, | |
| "step": 5870 | |
| }, | |
| { | |
| "epoch": 117.6, | |
| "learning_rate": 2.498787878787879e-05, | |
| "loss": 0.765, | |
| "step": 5880 | |
| }, | |
| { | |
| "epoch": 117.8, | |
| "learning_rate": 2.492727272727273e-05, | |
| "loss": 0.7907, | |
| "step": 5890 | |
| }, | |
| { | |
| "epoch": 118.0, | |
| "learning_rate": 2.4866666666666667e-05, | |
| "loss": 0.7417, | |
| "step": 5900 | |
| }, | |
| { | |
| "epoch": 118.2, | |
| "learning_rate": 2.4806060606060605e-05, | |
| "loss": 0.7902, | |
| "step": 5910 | |
| }, | |
| { | |
| "epoch": 118.4, | |
| "learning_rate": 2.4745454545454547e-05, | |
| "loss": 0.7617, | |
| "step": 5920 | |
| }, | |
| { | |
| "epoch": 118.6, | |
| "learning_rate": 2.4684848484848485e-05, | |
| "loss": 0.771, | |
| "step": 5930 | |
| }, | |
| { | |
| "epoch": 118.8, | |
| "learning_rate": 2.4624242424242424e-05, | |
| "loss": 0.7884, | |
| "step": 5940 | |
| }, | |
| { | |
| "epoch": 119.0, | |
| "learning_rate": 2.4563636363636365e-05, | |
| "loss": 0.7506, | |
| "step": 5950 | |
| }, | |
| { | |
| "epoch": 119.2, | |
| "learning_rate": 2.4503030303030304e-05, | |
| "loss": 0.7876, | |
| "step": 5960 | |
| }, | |
| { | |
| "epoch": 119.4, | |
| "learning_rate": 2.4442424242424245e-05, | |
| "loss": 0.7781, | |
| "step": 5970 | |
| }, | |
| { | |
| "epoch": 119.6, | |
| "learning_rate": 2.4381818181818184e-05, | |
| "loss": 0.7787, | |
| "step": 5980 | |
| }, | |
| { | |
| "epoch": 119.8, | |
| "learning_rate": 2.4321212121212122e-05, | |
| "loss": 0.7615, | |
| "step": 5990 | |
| }, | |
| { | |
| "epoch": 120.0, | |
| "learning_rate": 2.426060606060606e-05, | |
| "loss": 0.7388, | |
| "step": 6000 | |
| }, | |
| { | |
| "epoch": 120.0, | |
| "eval_loss": 0.353779673576355, | |
| "eval_runtime": 243.4641, | |
| "eval_samples_per_second": 21.408, | |
| "eval_steps_per_second": 1.339, | |
| "eval_wer": 0.3520331129275785, | |
| "step": 6000 | |
| }, | |
| { | |
| "epoch": 120.2, | |
| "learning_rate": 2.42e-05, | |
| "loss": 0.7812, | |
| "step": 6010 | |
| }, | |
| { | |
| "epoch": 120.4, | |
| "learning_rate": 2.413939393939394e-05, | |
| "loss": 0.7723, | |
| "step": 6020 | |
| }, | |
| { | |
| "epoch": 120.6, | |
| "learning_rate": 2.407878787878788e-05, | |
| "loss": 0.7685, | |
| "step": 6030 | |
| }, | |
| { | |
| "epoch": 120.8, | |
| "learning_rate": 2.401818181818182e-05, | |
| "loss": 0.7561, | |
| "step": 6040 | |
| }, | |
| { | |
| "epoch": 121.0, | |
| "learning_rate": 2.395757575757576e-05, | |
| "loss": 0.7497, | |
| "step": 6050 | |
| }, | |
| { | |
| "epoch": 121.2, | |
| "learning_rate": 2.3896969696969697e-05, | |
| "loss": 0.8065, | |
| "step": 6060 | |
| }, | |
| { | |
| "epoch": 121.4, | |
| "learning_rate": 2.383636363636364e-05, | |
| "loss": 0.7661, | |
| "step": 6070 | |
| }, | |
| { | |
| "epoch": 121.6, | |
| "learning_rate": 2.3775757575757577e-05, | |
| "loss": 0.7622, | |
| "step": 6080 | |
| }, | |
| { | |
| "epoch": 121.8, | |
| "learning_rate": 2.3715151515151516e-05, | |
| "loss": 0.7465, | |
| "step": 6090 | |
| }, | |
| { | |
| "epoch": 122.0, | |
| "learning_rate": 2.3654545454545454e-05, | |
| "loss": 0.7463, | |
| "step": 6100 | |
| }, | |
| { | |
| "epoch": 122.2, | |
| "learning_rate": 2.3593939393939392e-05, | |
| "loss": 0.7849, | |
| "step": 6110 | |
| }, | |
| { | |
| "epoch": 122.4, | |
| "learning_rate": 2.3533333333333337e-05, | |
| "loss": 0.7639, | |
| "step": 6120 | |
| }, | |
| { | |
| "epoch": 122.6, | |
| "learning_rate": 2.3472727272727276e-05, | |
| "loss": 0.7712, | |
| "step": 6130 | |
| }, | |
| { | |
| "epoch": 122.8, | |
| "learning_rate": 2.3412121212121214e-05, | |
| "loss": 0.7513, | |
| "step": 6140 | |
| }, | |
| { | |
| "epoch": 123.0, | |
| "learning_rate": 2.3351515151515152e-05, | |
| "loss": 0.7454, | |
| "step": 6150 | |
| }, | |
| { | |
| "epoch": 123.2, | |
| "learning_rate": 2.329090909090909e-05, | |
| "loss": 0.7733, | |
| "step": 6160 | |
| }, | |
| { | |
| "epoch": 123.4, | |
| "learning_rate": 2.323030303030303e-05, | |
| "loss": 0.7506, | |
| "step": 6170 | |
| }, | |
| { | |
| "epoch": 123.6, | |
| "learning_rate": 2.316969696969697e-05, | |
| "loss": 0.7561, | |
| "step": 6180 | |
| }, | |
| { | |
| "epoch": 123.8, | |
| "learning_rate": 2.310909090909091e-05, | |
| "loss": 0.7694, | |
| "step": 6190 | |
| }, | |
| { | |
| "epoch": 124.0, | |
| "learning_rate": 2.3048484848484847e-05, | |
| "loss": 0.7404, | |
| "step": 6200 | |
| }, | |
| { | |
| "epoch": 124.2, | |
| "learning_rate": 2.298787878787879e-05, | |
| "loss": 0.7819, | |
| "step": 6210 | |
| }, | |
| { | |
| "epoch": 124.4, | |
| "learning_rate": 2.2927272727272727e-05, | |
| "loss": 0.7499, | |
| "step": 6220 | |
| }, | |
| { | |
| "epoch": 124.6, | |
| "learning_rate": 2.286666666666667e-05, | |
| "loss": 0.756, | |
| "step": 6230 | |
| }, | |
| { | |
| "epoch": 124.8, | |
| "learning_rate": 2.2806060606060607e-05, | |
| "loss": 0.7716, | |
| "step": 6240 | |
| }, | |
| { | |
| "epoch": 125.0, | |
| "learning_rate": 2.2745454545454546e-05, | |
| "loss": 0.7427, | |
| "step": 6250 | |
| }, | |
| { | |
| "epoch": 125.2, | |
| "learning_rate": 2.2684848484848484e-05, | |
| "loss": 0.7816, | |
| "step": 6260 | |
| }, | |
| { | |
| "epoch": 125.4, | |
| "learning_rate": 2.2624242424242422e-05, | |
| "loss": 0.7623, | |
| "step": 6270 | |
| }, | |
| { | |
| "epoch": 125.6, | |
| "learning_rate": 2.2563636363636364e-05, | |
| "loss": 0.7561, | |
| "step": 6280 | |
| }, | |
| { | |
| "epoch": 125.8, | |
| "learning_rate": 2.2503030303030302e-05, | |
| "loss": 0.7409, | |
| "step": 6290 | |
| }, | |
| { | |
| "epoch": 126.0, | |
| "learning_rate": 2.2442424242424244e-05, | |
| "loss": 0.7169, | |
| "step": 6300 | |
| }, | |
| { | |
| "epoch": 126.2, | |
| "learning_rate": 2.2381818181818183e-05, | |
| "loss": 0.7972, | |
| "step": 6310 | |
| }, | |
| { | |
| "epoch": 126.4, | |
| "learning_rate": 2.232121212121212e-05, | |
| "loss": 0.7426, | |
| "step": 6320 | |
| }, | |
| { | |
| "epoch": 126.6, | |
| "learning_rate": 2.2260606060606063e-05, | |
| "loss": 0.7575, | |
| "step": 6330 | |
| }, | |
| { | |
| "epoch": 126.8, | |
| "learning_rate": 2.22e-05, | |
| "loss": 0.7446, | |
| "step": 6340 | |
| }, | |
| { | |
| "epoch": 127.0, | |
| "learning_rate": 2.213939393939394e-05, | |
| "loss": 0.7381, | |
| "step": 6350 | |
| }, | |
| { | |
| "epoch": 127.2, | |
| "learning_rate": 2.2078787878787878e-05, | |
| "loss": 0.7664, | |
| "step": 6360 | |
| }, | |
| { | |
| "epoch": 127.4, | |
| "learning_rate": 2.2018181818181816e-05, | |
| "loss": 0.7522, | |
| "step": 6370 | |
| }, | |
| { | |
| "epoch": 127.6, | |
| "learning_rate": 2.195757575757576e-05, | |
| "loss": 0.7696, | |
| "step": 6380 | |
| }, | |
| { | |
| "epoch": 127.8, | |
| "learning_rate": 2.18969696969697e-05, | |
| "loss": 0.7475, | |
| "step": 6390 | |
| }, | |
| { | |
| "epoch": 128.0, | |
| "learning_rate": 2.1836363636363638e-05, | |
| "loss": 0.7163, | |
| "step": 6400 | |
| }, | |
| { | |
| "epoch": 128.2, | |
| "learning_rate": 2.1775757575757576e-05, | |
| "loss": 0.7693, | |
| "step": 6410 | |
| }, | |
| { | |
| "epoch": 128.4, | |
| "learning_rate": 2.1715151515151514e-05, | |
| "loss": 0.7437, | |
| "step": 6420 | |
| }, | |
| { | |
| "epoch": 128.6, | |
| "learning_rate": 2.1654545454545456e-05, | |
| "loss": 0.7631, | |
| "step": 6430 | |
| }, | |
| { | |
| "epoch": 128.8, | |
| "learning_rate": 2.1593939393939394e-05, | |
| "loss": 0.7367, | |
| "step": 6440 | |
| }, | |
| { | |
| "epoch": 129.0, | |
| "learning_rate": 2.1533333333333333e-05, | |
| "loss": 0.7289, | |
| "step": 6450 | |
| }, | |
| { | |
| "epoch": 129.2, | |
| "learning_rate": 2.147272727272727e-05, | |
| "loss": 0.7635, | |
| "step": 6460 | |
| }, | |
| { | |
| "epoch": 129.4, | |
| "learning_rate": 2.1412121212121213e-05, | |
| "loss": 0.7493, | |
| "step": 6470 | |
| }, | |
| { | |
| "epoch": 129.6, | |
| "learning_rate": 2.1351515151515154e-05, | |
| "loss": 0.769, | |
| "step": 6480 | |
| }, | |
| { | |
| "epoch": 129.8, | |
| "learning_rate": 2.1290909090909093e-05, | |
| "loss": 0.7307, | |
| "step": 6490 | |
| }, | |
| { | |
| "epoch": 130.0, | |
| "learning_rate": 2.123030303030303e-05, | |
| "loss": 0.714, | |
| "step": 6500 | |
| }, | |
| { | |
| "epoch": 130.0, | |
| "eval_loss": 0.3682139217853546, | |
| "eval_runtime": 243.0361, | |
| "eval_samples_per_second": 21.445, | |
| "eval_steps_per_second": 1.341, | |
| "eval_wer": 0.3506013432602697, | |
| "step": 6500 | |
| }, | |
| { | |
| "epoch": 130.2, | |
| "learning_rate": 2.116969696969697e-05, | |
| "loss": 0.7817, | |
| "step": 6510 | |
| }, | |
| { | |
| "epoch": 130.4, | |
| "learning_rate": 2.1109090909090908e-05, | |
| "loss": 0.7611, | |
| "step": 6520 | |
| }, | |
| { | |
| "epoch": 130.6, | |
| "learning_rate": 2.104848484848485e-05, | |
| "loss": 0.7488, | |
| "step": 6530 | |
| }, | |
| { | |
| "epoch": 130.8, | |
| "learning_rate": 2.0987878787878788e-05, | |
| "loss": 0.7409, | |
| "step": 6540 | |
| }, | |
| { | |
| "epoch": 131.0, | |
| "learning_rate": 2.092727272727273e-05, | |
| "loss": 0.7171, | |
| "step": 6550 | |
| }, | |
| { | |
| "epoch": 131.2, | |
| "learning_rate": 2.0866666666666668e-05, | |
| "loss": 0.7772, | |
| "step": 6560 | |
| }, | |
| { | |
| "epoch": 131.4, | |
| "learning_rate": 2.0806060606060606e-05, | |
| "loss": 0.7418, | |
| "step": 6570 | |
| }, | |
| { | |
| "epoch": 131.6, | |
| "learning_rate": 2.0745454545454548e-05, | |
| "loss": 0.7403, | |
| "step": 6580 | |
| }, | |
| { | |
| "epoch": 131.8, | |
| "learning_rate": 2.0684848484848486e-05, | |
| "loss": 0.7406, | |
| "step": 6590 | |
| }, | |
| { | |
| "epoch": 132.0, | |
| "learning_rate": 2.0624242424242425e-05, | |
| "loss": 0.7281, | |
| "step": 6600 | |
| }, | |
| { | |
| "epoch": 132.2, | |
| "learning_rate": 2.0563636363636363e-05, | |
| "loss": 0.7583, | |
| "step": 6610 | |
| }, | |
| { | |
| "epoch": 132.4, | |
| "learning_rate": 2.05030303030303e-05, | |
| "loss": 0.7586, | |
| "step": 6620 | |
| }, | |
| { | |
| "epoch": 132.6, | |
| "learning_rate": 2.0442424242424243e-05, | |
| "loss": 0.7361, | |
| "step": 6630 | |
| }, | |
| { | |
| "epoch": 132.8, | |
| "learning_rate": 2.0381818181818185e-05, | |
| "loss": 0.7476, | |
| "step": 6640 | |
| }, | |
| { | |
| "epoch": 133.0, | |
| "learning_rate": 2.0321212121212123e-05, | |
| "loss": 0.7112, | |
| "step": 6650 | |
| }, | |
| { | |
| "epoch": 133.2, | |
| "learning_rate": 2.026060606060606e-05, | |
| "loss": 0.762, | |
| "step": 6660 | |
| }, | |
| { | |
| "epoch": 133.4, | |
| "learning_rate": 2.02e-05, | |
| "loss": 0.7406, | |
| "step": 6670 | |
| }, | |
| { | |
| "epoch": 133.6, | |
| "learning_rate": 2.0139393939393938e-05, | |
| "loss": 0.7421, | |
| "step": 6680 | |
| }, | |
| { | |
| "epoch": 133.8, | |
| "learning_rate": 2.007878787878788e-05, | |
| "loss": 0.7531, | |
| "step": 6690 | |
| }, | |
| { | |
| "epoch": 134.0, | |
| "learning_rate": 2.0018181818181818e-05, | |
| "loss": 0.7052, | |
| "step": 6700 | |
| }, | |
| { | |
| "epoch": 134.2, | |
| "learning_rate": 1.9957575757575756e-05, | |
| "loss": 0.7579, | |
| "step": 6710 | |
| }, | |
| { | |
| "epoch": 134.4, | |
| "learning_rate": 1.9896969696969698e-05, | |
| "loss": 0.7309, | |
| "step": 6720 | |
| }, | |
| { | |
| "epoch": 134.6, | |
| "learning_rate": 1.9836363636363636e-05, | |
| "loss": 0.7491, | |
| "step": 6730 | |
| }, | |
| { | |
| "epoch": 134.8, | |
| "learning_rate": 1.9775757575757578e-05, | |
| "loss": 0.7368, | |
| "step": 6740 | |
| }, | |
| { | |
| "epoch": 135.0, | |
| "learning_rate": 1.9715151515151517e-05, | |
| "loss": 0.6959, | |
| "step": 6750 | |
| }, | |
| { | |
| "epoch": 135.2, | |
| "learning_rate": 1.9654545454545455e-05, | |
| "loss": 0.7675, | |
| "step": 6760 | |
| }, | |
| { | |
| "epoch": 135.4, | |
| "learning_rate": 1.9593939393939393e-05, | |
| "loss": 0.7377, | |
| "step": 6770 | |
| }, | |
| { | |
| "epoch": 135.6, | |
| "learning_rate": 1.953333333333333e-05, | |
| "loss": 0.7589, | |
| "step": 6780 | |
| }, | |
| { | |
| "epoch": 135.8, | |
| "learning_rate": 1.9472727272727273e-05, | |
| "loss": 0.7398, | |
| "step": 6790 | |
| }, | |
| { | |
| "epoch": 136.0, | |
| "learning_rate": 1.941212121212121e-05, | |
| "loss": 0.7044, | |
| "step": 6800 | |
| }, | |
| { | |
| "epoch": 136.2, | |
| "learning_rate": 1.9351515151515153e-05, | |
| "loss": 0.759, | |
| "step": 6810 | |
| }, | |
| { | |
| "epoch": 136.4, | |
| "learning_rate": 1.929090909090909e-05, | |
| "loss": 0.7356, | |
| "step": 6820 | |
| }, | |
| { | |
| "epoch": 136.6, | |
| "learning_rate": 1.923030303030303e-05, | |
| "loss": 0.7181, | |
| "step": 6830 | |
| }, | |
| { | |
| "epoch": 136.8, | |
| "learning_rate": 1.916969696969697e-05, | |
| "loss": 0.7349, | |
| "step": 6840 | |
| }, | |
| { | |
| "epoch": 137.0, | |
| "learning_rate": 1.910909090909091e-05, | |
| "loss": 0.7004, | |
| "step": 6850 | |
| }, | |
| { | |
| "epoch": 137.2, | |
| "learning_rate": 1.904848484848485e-05, | |
| "loss": 0.7531, | |
| "step": 6860 | |
| }, | |
| { | |
| "epoch": 137.4, | |
| "learning_rate": 1.8987878787878787e-05, | |
| "loss": 0.749, | |
| "step": 6870 | |
| }, | |
| { | |
| "epoch": 137.6, | |
| "learning_rate": 1.8927272727272725e-05, | |
| "loss": 0.7306, | |
| "step": 6880 | |
| }, | |
| { | |
| "epoch": 137.8, | |
| "learning_rate": 1.886666666666667e-05, | |
| "loss": 0.7281, | |
| "step": 6890 | |
| }, | |
| { | |
| "epoch": 138.0, | |
| "learning_rate": 1.880606060606061e-05, | |
| "loss": 0.7322, | |
| "step": 6900 | |
| }, | |
| { | |
| "epoch": 138.2, | |
| "learning_rate": 1.8745454545454547e-05, | |
| "loss": 0.7754, | |
| "step": 6910 | |
| }, | |
| { | |
| "epoch": 138.4, | |
| "learning_rate": 1.8684848484848485e-05, | |
| "loss": 0.7235, | |
| "step": 6920 | |
| }, | |
| { | |
| "epoch": 138.6, | |
| "learning_rate": 1.8624242424242423e-05, | |
| "loss": 0.7408, | |
| "step": 6930 | |
| }, | |
| { | |
| "epoch": 138.8, | |
| "learning_rate": 1.8563636363636365e-05, | |
| "loss": 0.7376, | |
| "step": 6940 | |
| }, | |
| { | |
| "epoch": 139.0, | |
| "learning_rate": 1.8503030303030303e-05, | |
| "loss": 0.7067, | |
| "step": 6950 | |
| }, | |
| { | |
| "epoch": 139.2, | |
| "learning_rate": 1.8442424242424242e-05, | |
| "loss": 0.7594, | |
| "step": 6960 | |
| }, | |
| { | |
| "epoch": 139.4, | |
| "learning_rate": 1.838181818181818e-05, | |
| "loss": 0.7245, | |
| "step": 6970 | |
| }, | |
| { | |
| "epoch": 139.6, | |
| "learning_rate": 1.8321212121212122e-05, | |
| "loss": 0.7527, | |
| "step": 6980 | |
| }, | |
| { | |
| "epoch": 139.8, | |
| "learning_rate": 1.8260606060606064e-05, | |
| "loss": 0.7291, | |
| "step": 6990 | |
| }, | |
| { | |
| "epoch": 140.0, | |
| "learning_rate": 1.8200000000000002e-05, | |
| "loss": 0.7291, | |
| "step": 7000 | |
| }, | |
| { | |
| "epoch": 140.0, | |
| "eval_loss": 0.36252087354660034, | |
| "eval_runtime": 244.2417, | |
| "eval_samples_per_second": 21.34, | |
| "eval_steps_per_second": 1.335, | |
| "eval_wer": 0.3504972145571927, | |
| "step": 7000 | |
| }, | |
| { | |
| "epoch": 140.2, | |
| "learning_rate": 1.813939393939394e-05, | |
| "loss": 0.7564, | |
| "step": 7010 | |
| }, | |
| { | |
| "epoch": 140.4, | |
| "learning_rate": 1.807878787878788e-05, | |
| "loss": 0.7686, | |
| "step": 7020 | |
| }, | |
| { | |
| "epoch": 140.6, | |
| "learning_rate": 1.8018181818181817e-05, | |
| "loss": 0.7334, | |
| "step": 7030 | |
| }, | |
| { | |
| "epoch": 140.8, | |
| "learning_rate": 1.795757575757576e-05, | |
| "loss": 0.7459, | |
| "step": 7040 | |
| }, | |
| { | |
| "epoch": 141.0, | |
| "learning_rate": 1.7896969696969697e-05, | |
| "loss": 0.7179, | |
| "step": 7050 | |
| }, | |
| { | |
| "epoch": 141.2, | |
| "learning_rate": 1.783636363636364e-05, | |
| "loss": 0.7633, | |
| "step": 7060 | |
| }, | |
| { | |
| "epoch": 141.4, | |
| "learning_rate": 1.7775757575757577e-05, | |
| "loss": 0.7249, | |
| "step": 7070 | |
| }, | |
| { | |
| "epoch": 141.6, | |
| "learning_rate": 1.7715151515151515e-05, | |
| "loss": 0.726, | |
| "step": 7080 | |
| }, | |
| { | |
| "epoch": 141.8, | |
| "learning_rate": 1.7654545454545457e-05, | |
| "loss": 0.7401, | |
| "step": 7090 | |
| }, | |
| { | |
| "epoch": 142.0, | |
| "learning_rate": 1.7593939393939395e-05, | |
| "loss": 0.7075, | |
| "step": 7100 | |
| }, | |
| { | |
| "epoch": 142.2, | |
| "learning_rate": 1.7533333333333334e-05, | |
| "loss": 0.7549, | |
| "step": 7110 | |
| }, | |
| { | |
| "epoch": 142.4, | |
| "learning_rate": 1.7472727272727272e-05, | |
| "loss": 0.7279, | |
| "step": 7120 | |
| }, | |
| { | |
| "epoch": 142.6, | |
| "learning_rate": 1.741212121212121e-05, | |
| "loss": 0.7364, | |
| "step": 7130 | |
| }, | |
| { | |
| "epoch": 142.8, | |
| "learning_rate": 1.7351515151515152e-05, | |
| "loss": 0.7469, | |
| "step": 7140 | |
| }, | |
| { | |
| "epoch": 143.0, | |
| "learning_rate": 1.7290909090909094e-05, | |
| "loss": 0.7131, | |
| "step": 7150 | |
| }, | |
| { | |
| "epoch": 143.2, | |
| "learning_rate": 1.7230303030303032e-05, | |
| "loss": 0.751, | |
| "step": 7160 | |
| }, | |
| { | |
| "epoch": 143.4, | |
| "learning_rate": 1.716969696969697e-05, | |
| "loss": 0.7274, | |
| "step": 7170 | |
| }, | |
| { | |
| "epoch": 143.6, | |
| "learning_rate": 1.710909090909091e-05, | |
| "loss": 0.7217, | |
| "step": 7180 | |
| }, | |
| { | |
| "epoch": 143.8, | |
| "learning_rate": 1.7048484848484847e-05, | |
| "loss": 0.7416, | |
| "step": 7190 | |
| }, | |
| { | |
| "epoch": 144.0, | |
| "learning_rate": 1.698787878787879e-05, | |
| "loss": 0.6873, | |
| "step": 7200 | |
| }, | |
| { | |
| "epoch": 144.2, | |
| "learning_rate": 1.6927272727272727e-05, | |
| "loss": 0.7562, | |
| "step": 7210 | |
| }, | |
| { | |
| "epoch": 144.4, | |
| "learning_rate": 1.6866666666666666e-05, | |
| "loss": 0.7104, | |
| "step": 7220 | |
| }, | |
| { | |
| "epoch": 144.6, | |
| "learning_rate": 1.6806060606060607e-05, | |
| "loss": 0.7403, | |
| "step": 7230 | |
| }, | |
| { | |
| "epoch": 144.8, | |
| "learning_rate": 1.6745454545454546e-05, | |
| "loss": 0.7242, | |
| "step": 7240 | |
| }, | |
| { | |
| "epoch": 145.0, | |
| "learning_rate": 1.6684848484848487e-05, | |
| "loss": 0.7091, | |
| "step": 7250 | |
| }, | |
| { | |
| "epoch": 145.2, | |
| "learning_rate": 1.6624242424242426e-05, | |
| "loss": 0.7634, | |
| "step": 7260 | |
| }, | |
| { | |
| "epoch": 145.4, | |
| "learning_rate": 1.6563636363636364e-05, | |
| "loss": 0.7249, | |
| "step": 7270 | |
| }, | |
| { | |
| "epoch": 145.6, | |
| "learning_rate": 1.6503030303030302e-05, | |
| "loss": 0.7354, | |
| "step": 7280 | |
| }, | |
| { | |
| "epoch": 145.8, | |
| "learning_rate": 1.644242424242424e-05, | |
| "loss": 0.7336, | |
| "step": 7290 | |
| }, | |
| { | |
| "epoch": 146.0, | |
| "learning_rate": 1.6381818181818182e-05, | |
| "loss": 0.69, | |
| "step": 7300 | |
| }, | |
| { | |
| "epoch": 146.2, | |
| "learning_rate": 1.632121212121212e-05, | |
| "loss": 0.7587, | |
| "step": 7310 | |
| }, | |
| { | |
| "epoch": 146.4, | |
| "learning_rate": 1.6260606060606062e-05, | |
| "loss": 0.7295, | |
| "step": 7320 | |
| }, | |
| { | |
| "epoch": 146.6, | |
| "learning_rate": 1.62e-05, | |
| "loss": 0.733, | |
| "step": 7330 | |
| }, | |
| { | |
| "epoch": 146.8, | |
| "learning_rate": 1.613939393939394e-05, | |
| "loss": 0.7043, | |
| "step": 7340 | |
| }, | |
| { | |
| "epoch": 147.0, | |
| "learning_rate": 1.607878787878788e-05, | |
| "loss": 0.6902, | |
| "step": 7350 | |
| }, | |
| { | |
| "epoch": 147.2, | |
| "learning_rate": 1.601818181818182e-05, | |
| "loss": 0.7559, | |
| "step": 7360 | |
| }, | |
| { | |
| "epoch": 147.4, | |
| "learning_rate": 1.5957575757575757e-05, | |
| "loss": 0.733, | |
| "step": 7370 | |
| }, | |
| { | |
| "epoch": 147.6, | |
| "learning_rate": 1.5896969696969696e-05, | |
| "loss": 0.7107, | |
| "step": 7380 | |
| }, | |
| { | |
| "epoch": 147.8, | |
| "learning_rate": 1.5836363636363634e-05, | |
| "loss": 0.7186, | |
| "step": 7390 | |
| }, | |
| { | |
| "epoch": 148.0, | |
| "learning_rate": 1.577575757575758e-05, | |
| "loss": 0.6978, | |
| "step": 7400 | |
| }, | |
| { | |
| "epoch": 148.2, | |
| "learning_rate": 1.5715151515151518e-05, | |
| "loss": 0.7545, | |
| "step": 7410 | |
| }, | |
| { | |
| "epoch": 148.4, | |
| "learning_rate": 1.5654545454545456e-05, | |
| "loss": 0.7321, | |
| "step": 7420 | |
| }, | |
| { | |
| "epoch": 148.6, | |
| "learning_rate": 1.5593939393939394e-05, | |
| "loss": 0.7223, | |
| "step": 7430 | |
| }, | |
| { | |
| "epoch": 148.8, | |
| "learning_rate": 1.5533333333333333e-05, | |
| "loss": 0.7216, | |
| "step": 7440 | |
| }, | |
| { | |
| "epoch": 149.0, | |
| "learning_rate": 1.5472727272727274e-05, | |
| "loss": 0.7029, | |
| "step": 7450 | |
| }, | |
| { | |
| "epoch": 149.2, | |
| "learning_rate": 1.5412121212121213e-05, | |
| "loss": 0.7595, | |
| "step": 7460 | |
| }, | |
| { | |
| "epoch": 149.4, | |
| "learning_rate": 1.535151515151515e-05, | |
| "loss": 0.7406, | |
| "step": 7470 | |
| }, | |
| { | |
| "epoch": 149.6, | |
| "learning_rate": 1.529090909090909e-05, | |
| "loss": 0.7215, | |
| "step": 7480 | |
| }, | |
| { | |
| "epoch": 149.8, | |
| "learning_rate": 1.5230303030303033e-05, | |
| "loss": 0.7085, | |
| "step": 7490 | |
| }, | |
| { | |
| "epoch": 150.0, | |
| "learning_rate": 1.5169696969696971e-05, | |
| "loss": 0.697, | |
| "step": 7500 | |
| }, | |
| { | |
| "epoch": 150.0, | |
| "eval_loss": 0.36194926500320435, | |
| "eval_runtime": 244.3442, | |
| "eval_samples_per_second": 21.331, | |
| "eval_steps_per_second": 1.334, | |
| "eval_wer": 0.3479460613318061, | |
| "step": 7500 | |
| }, | |
| { | |
| "epoch": 150.2, | |
| "learning_rate": 1.510909090909091e-05, | |
| "loss": 0.7518, | |
| "step": 7510 | |
| }, | |
| { | |
| "epoch": 150.4, | |
| "learning_rate": 1.504848484848485e-05, | |
| "loss": 0.7236, | |
| "step": 7520 | |
| }, | |
| { | |
| "epoch": 150.6, | |
| "learning_rate": 1.4987878787878788e-05, | |
| "loss": 0.7105, | |
| "step": 7530 | |
| }, | |
| { | |
| "epoch": 150.8, | |
| "learning_rate": 1.4927272727272728e-05, | |
| "loss": 0.7035, | |
| "step": 7540 | |
| }, | |
| { | |
| "epoch": 151.0, | |
| "learning_rate": 1.4866666666666668e-05, | |
| "loss": 0.7026, | |
| "step": 7550 | |
| }, | |
| { | |
| "epoch": 151.2, | |
| "learning_rate": 1.4806060606060606e-05, | |
| "loss": 0.7266, | |
| "step": 7560 | |
| }, | |
| { | |
| "epoch": 151.4, | |
| "learning_rate": 1.4745454545454546e-05, | |
| "loss": 0.7391, | |
| "step": 7570 | |
| }, | |
| { | |
| "epoch": 151.6, | |
| "learning_rate": 1.4684848484848484e-05, | |
| "loss": 0.7343, | |
| "step": 7580 | |
| }, | |
| { | |
| "epoch": 151.8, | |
| "learning_rate": 1.4624242424242424e-05, | |
| "loss": 0.7262, | |
| "step": 7590 | |
| }, | |
| { | |
| "epoch": 152.0, | |
| "learning_rate": 1.4563636363636364e-05, | |
| "loss": 0.697, | |
| "step": 7600 | |
| }, | |
| { | |
| "epoch": 152.2, | |
| "learning_rate": 1.4503030303030303e-05, | |
| "loss": 0.7542, | |
| "step": 7610 | |
| }, | |
| { | |
| "epoch": 152.4, | |
| "learning_rate": 1.4442424242424243e-05, | |
| "loss": 0.716, | |
| "step": 7620 | |
| }, | |
| { | |
| "epoch": 152.6, | |
| "learning_rate": 1.4381818181818181e-05, | |
| "loss": 0.7246, | |
| "step": 7630 | |
| }, | |
| { | |
| "epoch": 152.8, | |
| "learning_rate": 1.4321212121212123e-05, | |
| "loss": 0.7253, | |
| "step": 7640 | |
| }, | |
| { | |
| "epoch": 153.0, | |
| "learning_rate": 1.4260606060606061e-05, | |
| "loss": 0.684, | |
| "step": 7650 | |
| }, | |
| { | |
| "epoch": 153.2, | |
| "learning_rate": 1.42e-05, | |
| "loss": 0.7596, | |
| "step": 7660 | |
| }, | |
| { | |
| "epoch": 153.4, | |
| "learning_rate": 1.413939393939394e-05, | |
| "loss": 0.7173, | |
| "step": 7670 | |
| }, | |
| { | |
| "epoch": 153.6, | |
| "learning_rate": 1.407878787878788e-05, | |
| "loss": 0.7008, | |
| "step": 7680 | |
| }, | |
| { | |
| "epoch": 153.8, | |
| "learning_rate": 1.401818181818182e-05, | |
| "loss": 0.7028, | |
| "step": 7690 | |
| }, | |
| { | |
| "epoch": 154.0, | |
| "learning_rate": 1.3957575757575758e-05, | |
| "loss": 0.6925, | |
| "step": 7700 | |
| }, | |
| { | |
| "epoch": 154.2, | |
| "learning_rate": 1.3896969696969696e-05, | |
| "loss": 0.7455, | |
| "step": 7710 | |
| }, | |
| { | |
| "epoch": 154.4, | |
| "learning_rate": 1.3836363636363636e-05, | |
| "loss": 0.7204, | |
| "step": 7720 | |
| }, | |
| { | |
| "epoch": 154.6, | |
| "learning_rate": 1.3775757575757576e-05, | |
| "loss": 0.7157, | |
| "step": 7730 | |
| }, | |
| { | |
| "epoch": 154.8, | |
| "learning_rate": 1.3715151515151516e-05, | |
| "loss": 0.7088, | |
| "step": 7740 | |
| }, | |
| { | |
| "epoch": 155.0, | |
| "learning_rate": 1.3654545454545455e-05, | |
| "loss": 0.7002, | |
| "step": 7750 | |
| }, | |
| { | |
| "epoch": 155.2, | |
| "learning_rate": 1.3593939393939393e-05, | |
| "loss": 0.7425, | |
| "step": 7760 | |
| }, | |
| { | |
| "epoch": 155.4, | |
| "learning_rate": 1.3533333333333335e-05, | |
| "loss": 0.737, | |
| "step": 7770 | |
| }, | |
| { | |
| "epoch": 155.6, | |
| "learning_rate": 1.3472727272727273e-05, | |
| "loss": 0.7224, | |
| "step": 7780 | |
| }, | |
| { | |
| "epoch": 155.8, | |
| "learning_rate": 1.3412121212121211e-05, | |
| "loss": 0.7324, | |
| "step": 7790 | |
| }, | |
| { | |
| "epoch": 156.0, | |
| "learning_rate": 1.3351515151515151e-05, | |
| "loss": 0.6667, | |
| "step": 7800 | |
| }, | |
| { | |
| "epoch": 156.2, | |
| "learning_rate": 1.3290909090909091e-05, | |
| "loss": 0.7378, | |
| "step": 7810 | |
| }, | |
| { | |
| "epoch": 156.4, | |
| "learning_rate": 1.3230303030303032e-05, | |
| "loss": 0.7136, | |
| "step": 7820 | |
| }, | |
| { | |
| "epoch": 156.6, | |
| "learning_rate": 1.316969696969697e-05, | |
| "loss": 0.7082, | |
| "step": 7830 | |
| }, | |
| { | |
| "epoch": 156.8, | |
| "learning_rate": 1.3109090909090908e-05, | |
| "loss": 0.6986, | |
| "step": 7840 | |
| }, | |
| { | |
| "epoch": 157.0, | |
| "learning_rate": 1.304848484848485e-05, | |
| "loss": 0.6915, | |
| "step": 7850 | |
| }, | |
| { | |
| "epoch": 157.2, | |
| "learning_rate": 1.2987878787878788e-05, | |
| "loss": 0.7195, | |
| "step": 7860 | |
| }, | |
| { | |
| "epoch": 157.4, | |
| "learning_rate": 1.2927272727272728e-05, | |
| "loss": 0.7176, | |
| "step": 7870 | |
| }, | |
| { | |
| "epoch": 157.6, | |
| "learning_rate": 1.2866666666666667e-05, | |
| "loss": 0.7083, | |
| "step": 7880 | |
| }, | |
| { | |
| "epoch": 157.8, | |
| "learning_rate": 1.2806060606060605e-05, | |
| "loss": 0.7098, | |
| "step": 7890 | |
| }, | |
| { | |
| "epoch": 158.0, | |
| "learning_rate": 1.2745454545454547e-05, | |
| "loss": 0.6857, | |
| "step": 7900 | |
| }, | |
| { | |
| "epoch": 158.2, | |
| "learning_rate": 1.2684848484848485e-05, | |
| "loss": 0.7282, | |
| "step": 7910 | |
| }, | |
| { | |
| "epoch": 158.4, | |
| "learning_rate": 1.2624242424242425e-05, | |
| "loss": 0.7154, | |
| "step": 7920 | |
| }, | |
| { | |
| "epoch": 158.6, | |
| "learning_rate": 1.2563636363636363e-05, | |
| "loss": 0.706, | |
| "step": 7930 | |
| }, | |
| { | |
| "epoch": 158.8, | |
| "learning_rate": 1.2503030303030303e-05, | |
| "loss": 0.722, | |
| "step": 7940 | |
| }, | |
| { | |
| "epoch": 159.0, | |
| "learning_rate": 1.2442424242424243e-05, | |
| "loss": 0.6904, | |
| "step": 7950 | |
| }, | |
| { | |
| "epoch": 159.2, | |
| "learning_rate": 1.2381818181818182e-05, | |
| "loss": 0.7324, | |
| "step": 7960 | |
| }, | |
| { | |
| "epoch": 159.4, | |
| "learning_rate": 1.2321212121212122e-05, | |
| "loss": 0.723, | |
| "step": 7970 | |
| }, | |
| { | |
| "epoch": 159.6, | |
| "learning_rate": 1.2260606060606062e-05, | |
| "loss": 0.7141, | |
| "step": 7980 | |
| }, | |
| { | |
| "epoch": 159.8, | |
| "learning_rate": 1.22e-05, | |
| "loss": 0.7177, | |
| "step": 7990 | |
| }, | |
| { | |
| "epoch": 160.0, | |
| "learning_rate": 1.213939393939394e-05, | |
| "loss": 0.6811, | |
| "step": 8000 | |
| }, | |
| { | |
| "epoch": 160.0, | |
| "eval_loss": 0.36307966709136963, | |
| "eval_runtime": 242.5601, | |
| "eval_samples_per_second": 21.487, | |
| "eval_steps_per_second": 1.344, | |
| "eval_wer": 0.3440412349664185, | |
| "step": 8000 | |
| }, | |
| { | |
| "epoch": 160.2, | |
| "learning_rate": 1.2078787878787878e-05, | |
| "loss": 0.7477, | |
| "step": 8010 | |
| }, | |
| { | |
| "epoch": 160.4, | |
| "learning_rate": 1.2018181818181818e-05, | |
| "loss": 0.7098, | |
| "step": 8020 | |
| }, | |
| { | |
| "epoch": 160.6, | |
| "learning_rate": 1.1957575757575758e-05, | |
| "loss": 0.6994, | |
| "step": 8030 | |
| }, | |
| { | |
| "epoch": 160.8, | |
| "learning_rate": 1.1896969696969697e-05, | |
| "loss": 0.7192, | |
| "step": 8040 | |
| }, | |
| { | |
| "epoch": 161.0, | |
| "learning_rate": 1.1836363636363637e-05, | |
| "loss": 0.6757, | |
| "step": 8050 | |
| }, | |
| { | |
| "epoch": 161.2, | |
| "learning_rate": 1.1775757575757575e-05, | |
| "loss": 0.7521, | |
| "step": 8060 | |
| }, | |
| { | |
| "epoch": 161.4, | |
| "learning_rate": 1.1715151515151515e-05, | |
| "loss": 0.6944, | |
| "step": 8070 | |
| }, | |
| { | |
| "epoch": 161.6, | |
| "learning_rate": 1.1654545454545455e-05, | |
| "loss": 0.7216, | |
| "step": 8080 | |
| }, | |
| { | |
| "epoch": 161.8, | |
| "learning_rate": 1.1593939393939394e-05, | |
| "loss": 0.7034, | |
| "step": 8090 | |
| }, | |
| { | |
| "epoch": 162.0, | |
| "learning_rate": 1.1533333333333334e-05, | |
| "loss": 0.6685, | |
| "step": 8100 | |
| }, | |
| { | |
| "epoch": 162.2, | |
| "learning_rate": 1.1478787878787879e-05, | |
| "loss": 0.7395, | |
| "step": 8110 | |
| }, | |
| { | |
| "epoch": 162.4, | |
| "learning_rate": 1.1418181818181819e-05, | |
| "loss": 0.7139, | |
| "step": 8120 | |
| }, | |
| { | |
| "epoch": 162.6, | |
| "learning_rate": 1.1357575757575757e-05, | |
| "loss": 0.6898, | |
| "step": 8130 | |
| }, | |
| { | |
| "epoch": 162.8, | |
| "learning_rate": 1.1296969696969697e-05, | |
| "loss": 0.7013, | |
| "step": 8140 | |
| }, | |
| { | |
| "epoch": 163.0, | |
| "learning_rate": 1.1236363636363638e-05, | |
| "loss": 0.6784, | |
| "step": 8150 | |
| }, | |
| { | |
| "epoch": 163.2, | |
| "learning_rate": 1.1175757575757576e-05, | |
| "loss": 0.7114, | |
| "step": 8160 | |
| }, | |
| { | |
| "epoch": 163.4, | |
| "learning_rate": 1.1115151515151516e-05, | |
| "loss": 0.7094, | |
| "step": 8170 | |
| }, | |
| { | |
| "epoch": 163.6, | |
| "learning_rate": 1.1054545454545454e-05, | |
| "loss": 0.7064, | |
| "step": 8180 | |
| }, | |
| { | |
| "epoch": 163.8, | |
| "learning_rate": 1.0993939393939394e-05, | |
| "loss": 0.6946, | |
| "step": 8190 | |
| }, | |
| { | |
| "epoch": 164.0, | |
| "learning_rate": 1.0933333333333334e-05, | |
| "loss": 0.6638, | |
| "step": 8200 | |
| }, | |
| { | |
| "epoch": 164.2, | |
| "learning_rate": 1.0872727272727273e-05, | |
| "loss": 0.7257, | |
| "step": 8210 | |
| }, | |
| { | |
| "epoch": 164.4, | |
| "learning_rate": 1.0812121212121213e-05, | |
| "loss": 0.7131, | |
| "step": 8220 | |
| }, | |
| { | |
| "epoch": 164.6, | |
| "learning_rate": 1.0751515151515151e-05, | |
| "loss": 0.7032, | |
| "step": 8230 | |
| }, | |
| { | |
| "epoch": 164.8, | |
| "learning_rate": 1.0690909090909091e-05, | |
| "loss": 0.707, | |
| "step": 8240 | |
| }, | |
| { | |
| "epoch": 165.0, | |
| "learning_rate": 1.0630303030303031e-05, | |
| "loss": 0.6797, | |
| "step": 8250 | |
| }, | |
| { | |
| "epoch": 165.2, | |
| "learning_rate": 1.056969696969697e-05, | |
| "loss": 0.7334, | |
| "step": 8260 | |
| }, | |
| { | |
| "epoch": 165.4, | |
| "learning_rate": 1.050909090909091e-05, | |
| "loss": 0.7339, | |
| "step": 8270 | |
| }, | |
| { | |
| "epoch": 165.6, | |
| "learning_rate": 1.044848484848485e-05, | |
| "loss": 0.7148, | |
| "step": 8280 | |
| }, | |
| { | |
| "epoch": 165.8, | |
| "learning_rate": 1.0387878787878788e-05, | |
| "loss": 0.7074, | |
| "step": 8290 | |
| }, | |
| { | |
| "epoch": 166.0, | |
| "learning_rate": 1.0327272727272728e-05, | |
| "loss": 0.6874, | |
| "step": 8300 | |
| }, | |
| { | |
| "epoch": 166.2, | |
| "learning_rate": 1.0266666666666666e-05, | |
| "loss": 0.7472, | |
| "step": 8310 | |
| }, | |
| { | |
| "epoch": 166.4, | |
| "learning_rate": 1.0206060606060608e-05, | |
| "loss": 0.711, | |
| "step": 8320 | |
| }, | |
| { | |
| "epoch": 166.6, | |
| "learning_rate": 1.0145454545454546e-05, | |
| "loss": 0.707, | |
| "step": 8330 | |
| }, | |
| { | |
| "epoch": 166.8, | |
| "learning_rate": 1.0084848484848484e-05, | |
| "loss": 0.7023, | |
| "step": 8340 | |
| }, | |
| { | |
| "epoch": 167.0, | |
| "learning_rate": 1.0024242424242424e-05, | |
| "loss": 0.6812, | |
| "step": 8350 | |
| }, | |
| { | |
| "epoch": 167.2, | |
| "learning_rate": 9.963636363636363e-06, | |
| "loss": 0.7549, | |
| "step": 8360 | |
| }, | |
| { | |
| "epoch": 167.4, | |
| "learning_rate": 9.903030303030305e-06, | |
| "loss": 0.69, | |
| "step": 8370 | |
| }, | |
| { | |
| "epoch": 167.6, | |
| "learning_rate": 9.842424242424243e-06, | |
| "loss": 0.6981, | |
| "step": 8380 | |
| }, | |
| { | |
| "epoch": 167.8, | |
| "learning_rate": 9.781818181818181e-06, | |
| "loss": 0.7245, | |
| "step": 8390 | |
| }, | |
| { | |
| "epoch": 168.0, | |
| "learning_rate": 9.721212121212121e-06, | |
| "loss": 0.6736, | |
| "step": 8400 | |
| }, | |
| { | |
| "epoch": 168.2, | |
| "learning_rate": 9.660606060606061e-06, | |
| "loss": 0.7403, | |
| "step": 8410 | |
| }, | |
| { | |
| "epoch": 168.4, | |
| "learning_rate": 9.600000000000001e-06, | |
| "loss": 0.68, | |
| "step": 8420 | |
| }, | |
| { | |
| "epoch": 168.6, | |
| "learning_rate": 9.53939393939394e-06, | |
| "loss": 0.7077, | |
| "step": 8430 | |
| }, | |
| { | |
| "epoch": 168.8, | |
| "learning_rate": 9.478787878787878e-06, | |
| "loss": 0.6924, | |
| "step": 8440 | |
| }, | |
| { | |
| "epoch": 169.0, | |
| "learning_rate": 9.41818181818182e-06, | |
| "loss": 0.679, | |
| "step": 8450 | |
| }, | |
| { | |
| "epoch": 169.2, | |
| "learning_rate": 9.357575757575758e-06, | |
| "loss": 0.7259, | |
| "step": 8460 | |
| }, | |
| { | |
| "epoch": 169.4, | |
| "learning_rate": 9.296969696969696e-06, | |
| "loss": 0.7021, | |
| "step": 8470 | |
| }, | |
| { | |
| "epoch": 169.6, | |
| "learning_rate": 9.236363636363636e-06, | |
| "loss": 0.6986, | |
| "step": 8480 | |
| }, | |
| { | |
| "epoch": 169.8, | |
| "learning_rate": 9.175757575757576e-06, | |
| "loss": 0.6945, | |
| "step": 8490 | |
| }, | |
| { | |
| "epoch": 170.0, | |
| "learning_rate": 9.115151515151516e-06, | |
| "loss": 0.6841, | |
| "step": 8500 | |
| }, | |
| { | |
| "epoch": 170.0, | |
| "eval_loss": 0.3671566843986511, | |
| "eval_runtime": 243.7376, | |
| "eval_samples_per_second": 21.384, | |
| "eval_steps_per_second": 1.338, | |
| "eval_wer": 0.3460457125006508, | |
| "step": 8500 | |
| }, | |
| { | |
| "epoch": 170.2, | |
| "learning_rate": 9.054545454545455e-06, | |
| "loss": 0.7282, | |
| "step": 8510 | |
| }, | |
| { | |
| "epoch": 170.4, | |
| "learning_rate": 8.993939393939393e-06, | |
| "loss": 0.6899, | |
| "step": 8520 | |
| }, | |
| { | |
| "epoch": 170.6, | |
| "learning_rate": 8.933333333333333e-06, | |
| "loss": 0.7181, | |
| "step": 8530 | |
| }, | |
| { | |
| "epoch": 170.8, | |
| "learning_rate": 8.872727272727273e-06, | |
| "loss": 0.6958, | |
| "step": 8540 | |
| }, | |
| { | |
| "epoch": 171.0, | |
| "learning_rate": 8.812121212121213e-06, | |
| "loss": 0.6931, | |
| "step": 8550 | |
| }, | |
| { | |
| "epoch": 171.2, | |
| "learning_rate": 8.751515151515151e-06, | |
| "loss": 0.7227, | |
| "step": 8560 | |
| }, | |
| { | |
| "epoch": 171.4, | |
| "learning_rate": 8.69090909090909e-06, | |
| "loss": 0.7281, | |
| "step": 8570 | |
| }, | |
| { | |
| "epoch": 171.6, | |
| "learning_rate": 8.630303030303032e-06, | |
| "loss": 0.7049, | |
| "step": 8580 | |
| }, | |
| { | |
| "epoch": 171.8, | |
| "learning_rate": 8.56969696969697e-06, | |
| "loss": 0.7046, | |
| "step": 8590 | |
| }, | |
| { | |
| "epoch": 172.0, | |
| "learning_rate": 8.50909090909091e-06, | |
| "loss": 0.661, | |
| "step": 8600 | |
| }, | |
| { | |
| "epoch": 172.2, | |
| "learning_rate": 8.448484848484848e-06, | |
| "loss": 0.723, | |
| "step": 8610 | |
| }, | |
| { | |
| "epoch": 172.4, | |
| "learning_rate": 8.387878787878788e-06, | |
| "loss": 0.7063, | |
| "step": 8620 | |
| }, | |
| { | |
| "epoch": 172.6, | |
| "learning_rate": 8.327272727272728e-06, | |
| "loss": 0.7094, | |
| "step": 8630 | |
| }, | |
| { | |
| "epoch": 172.8, | |
| "learning_rate": 8.266666666666667e-06, | |
| "loss": 0.7137, | |
| "step": 8640 | |
| }, | |
| { | |
| "epoch": 173.0, | |
| "learning_rate": 8.206060606060607e-06, | |
| "loss": 0.6544, | |
| "step": 8650 | |
| }, | |
| { | |
| "epoch": 173.2, | |
| "learning_rate": 8.145454545454547e-06, | |
| "loss": 0.7202, | |
| "step": 8660 | |
| }, | |
| { | |
| "epoch": 173.4, | |
| "learning_rate": 8.084848484848485e-06, | |
| "loss": 0.6964, | |
| "step": 8670 | |
| }, | |
| { | |
| "epoch": 173.6, | |
| "learning_rate": 8.024242424242425e-06, | |
| "loss": 0.7093, | |
| "step": 8680 | |
| }, | |
| { | |
| "epoch": 173.8, | |
| "learning_rate": 7.963636363636363e-06, | |
| "loss": 0.6938, | |
| "step": 8690 | |
| }, | |
| { | |
| "epoch": 174.0, | |
| "learning_rate": 7.903030303030303e-06, | |
| "loss": 0.6669, | |
| "step": 8700 | |
| }, | |
| { | |
| "epoch": 174.2, | |
| "learning_rate": 7.842424242424243e-06, | |
| "loss": 0.7225, | |
| "step": 8710 | |
| }, | |
| { | |
| "epoch": 174.4, | |
| "learning_rate": 7.781818181818182e-06, | |
| "loss": 0.7163, | |
| "step": 8720 | |
| }, | |
| { | |
| "epoch": 174.6, | |
| "learning_rate": 7.721212121212122e-06, | |
| "loss": 0.6901, | |
| "step": 8730 | |
| }, | |
| { | |
| "epoch": 174.8, | |
| "learning_rate": 7.66060606060606e-06, | |
| "loss": 0.6937, | |
| "step": 8740 | |
| }, | |
| { | |
| "epoch": 175.0, | |
| "learning_rate": 7.600000000000001e-06, | |
| "loss": 0.6705, | |
| "step": 8750 | |
| }, | |
| { | |
| "epoch": 175.2, | |
| "learning_rate": 7.53939393939394e-06, | |
| "loss": 0.7241, | |
| "step": 8760 | |
| }, | |
| { | |
| "epoch": 175.4, | |
| "learning_rate": 7.4787878787878784e-06, | |
| "loss": 0.6951, | |
| "step": 8770 | |
| }, | |
| { | |
| "epoch": 175.6, | |
| "learning_rate": 7.4181818181818185e-06, | |
| "loss": 0.6993, | |
| "step": 8780 | |
| }, | |
| { | |
| "epoch": 175.8, | |
| "learning_rate": 7.357575757575758e-06, | |
| "loss": 0.7076, | |
| "step": 8790 | |
| }, | |
| { | |
| "epoch": 176.0, | |
| "learning_rate": 7.296969696969698e-06, | |
| "loss": 0.6692, | |
| "step": 8800 | |
| }, | |
| { | |
| "epoch": 176.2, | |
| "learning_rate": 7.236363636363637e-06, | |
| "loss": 0.7284, | |
| "step": 8810 | |
| }, | |
| { | |
| "epoch": 176.4, | |
| "learning_rate": 7.175757575757576e-06, | |
| "loss": 0.6959, | |
| "step": 8820 | |
| }, | |
| { | |
| "epoch": 176.6, | |
| "learning_rate": 7.115151515151515e-06, | |
| "loss": 0.678, | |
| "step": 8830 | |
| }, | |
| { | |
| "epoch": 176.8, | |
| "learning_rate": 7.054545454545454e-06, | |
| "loss": 0.7012, | |
| "step": 8840 | |
| }, | |
| { | |
| "epoch": 177.0, | |
| "learning_rate": 6.993939393939394e-06, | |
| "loss": 0.6797, | |
| "step": 8850 | |
| }, | |
| { | |
| "epoch": 177.2, | |
| "learning_rate": 6.933333333333334e-06, | |
| "loss": 0.7238, | |
| "step": 8860 | |
| }, | |
| { | |
| "epoch": 177.4, | |
| "learning_rate": 6.872727272727273e-06, | |
| "loss": 0.6986, | |
| "step": 8870 | |
| }, | |
| { | |
| "epoch": 177.6, | |
| "learning_rate": 6.812121212121212e-06, | |
| "loss": 0.7004, | |
| "step": 8880 | |
| }, | |
| { | |
| "epoch": 177.8, | |
| "learning_rate": 6.751515151515152e-06, | |
| "loss": 0.6681, | |
| "step": 8890 | |
| }, | |
| { | |
| "epoch": 178.0, | |
| "learning_rate": 6.690909090909091e-06, | |
| "loss": 0.6817, | |
| "step": 8900 | |
| }, | |
| { | |
| "epoch": 178.2, | |
| "learning_rate": 6.63030303030303e-06, | |
| "loss": 0.7241, | |
| "step": 8910 | |
| }, | |
| { | |
| "epoch": 178.4, | |
| "learning_rate": 6.5696969696969695e-06, | |
| "loss": 0.7102, | |
| "step": 8920 | |
| }, | |
| { | |
| "epoch": 178.6, | |
| "learning_rate": 6.5090909090909095e-06, | |
| "loss": 0.6847, | |
| "step": 8930 | |
| }, | |
| { | |
| "epoch": 178.8, | |
| "learning_rate": 6.448484848484849e-06, | |
| "loss": 0.7004, | |
| "step": 8940 | |
| }, | |
| { | |
| "epoch": 179.0, | |
| "learning_rate": 6.387878787878789e-06, | |
| "loss": 0.6661, | |
| "step": 8950 | |
| }, | |
| { | |
| "epoch": 179.2, | |
| "learning_rate": 6.327272727272727e-06, | |
| "loss": 0.7154, | |
| "step": 8960 | |
| }, | |
| { | |
| "epoch": 179.4, | |
| "learning_rate": 6.266666666666667e-06, | |
| "loss": 0.69, | |
| "step": 8970 | |
| }, | |
| { | |
| "epoch": 179.6, | |
| "learning_rate": 6.206060606060606e-06, | |
| "loss": 0.6799, | |
| "step": 8980 | |
| }, | |
| { | |
| "epoch": 179.8, | |
| "learning_rate": 6.1454545454545454e-06, | |
| "loss": 0.6946, | |
| "step": 8990 | |
| }, | |
| { | |
| "epoch": 180.0, | |
| "learning_rate": 6.0848484848484855e-06, | |
| "loss": 0.6616, | |
| "step": 9000 | |
| }, | |
| { | |
| "epoch": 180.0, | |
| "eval_loss": 0.36765044927597046, | |
| "eval_runtime": 243.8885, | |
| "eval_samples_per_second": 21.37, | |
| "eval_steps_per_second": 1.337, | |
| "eval_wer": 0.34099547040141615, | |
| "step": 9000 | |
| }, | |
| { | |
| "epoch": 180.2, | |
| "learning_rate": 6.024242424242424e-06, | |
| "loss": 0.718, | |
| "step": 9010 | |
| }, | |
| { | |
| "epoch": 180.4, | |
| "learning_rate": 5.963636363636364e-06, | |
| "loss": 0.7075, | |
| "step": 9020 | |
| }, | |
| { | |
| "epoch": 180.6, | |
| "learning_rate": 5.903030303030303e-06, | |
| "loss": 0.7133, | |
| "step": 9030 | |
| }, | |
| { | |
| "epoch": 180.8, | |
| "learning_rate": 5.842424242424243e-06, | |
| "loss": 0.698, | |
| "step": 9040 | |
| }, | |
| { | |
| "epoch": 181.0, | |
| "learning_rate": 5.781818181818181e-06, | |
| "loss": 0.6749, | |
| "step": 9050 | |
| }, | |
| { | |
| "epoch": 181.2, | |
| "learning_rate": 5.721212121212121e-06, | |
| "loss": 0.7197, | |
| "step": 9060 | |
| }, | |
| { | |
| "epoch": 181.4, | |
| "learning_rate": 5.6606060606060606e-06, | |
| "loss": 0.7206, | |
| "step": 9070 | |
| }, | |
| { | |
| "epoch": 181.6, | |
| "learning_rate": 5.600000000000001e-06, | |
| "loss": 0.6867, | |
| "step": 9080 | |
| }, | |
| { | |
| "epoch": 181.8, | |
| "learning_rate": 5.53939393939394e-06, | |
| "loss": 0.6846, | |
| "step": 9090 | |
| }, | |
| { | |
| "epoch": 182.0, | |
| "learning_rate": 5.478787878787879e-06, | |
| "loss": 0.6683, | |
| "step": 9100 | |
| }, | |
| { | |
| "epoch": 182.2, | |
| "learning_rate": 5.418181818181818e-06, | |
| "loss": 0.7129, | |
| "step": 9110 | |
| }, | |
| { | |
| "epoch": 182.4, | |
| "learning_rate": 5.357575757575758e-06, | |
| "loss": 0.6931, | |
| "step": 9120 | |
| }, | |
| { | |
| "epoch": 182.6, | |
| "learning_rate": 5.296969696969697e-06, | |
| "loss": 0.6892, | |
| "step": 9130 | |
| }, | |
| { | |
| "epoch": 182.8, | |
| "learning_rate": 5.2363636363636365e-06, | |
| "loss": 0.6927, | |
| "step": 9140 | |
| }, | |
| { | |
| "epoch": 183.0, | |
| "learning_rate": 5.175757575757576e-06, | |
| "loss": 0.6858, | |
| "step": 9150 | |
| }, | |
| { | |
| "epoch": 183.2, | |
| "learning_rate": 5.115151515151515e-06, | |
| "loss": 0.726, | |
| "step": 9160 | |
| }, | |
| { | |
| "epoch": 183.4, | |
| "learning_rate": 5.054545454545455e-06, | |
| "loss": 0.6908, | |
| "step": 9170 | |
| }, | |
| { | |
| "epoch": 183.6, | |
| "learning_rate": 4.993939393939394e-06, | |
| "loss": 0.7123, | |
| "step": 9180 | |
| }, | |
| { | |
| "epoch": 183.8, | |
| "learning_rate": 4.933333333333333e-06, | |
| "loss": 0.7041, | |
| "step": 9190 | |
| }, | |
| { | |
| "epoch": 184.0, | |
| "learning_rate": 4.8727272727272724e-06, | |
| "loss": 0.6502, | |
| "step": 9200 | |
| }, | |
| { | |
| "epoch": 184.2, | |
| "learning_rate": 4.8121212121212125e-06, | |
| "loss": 0.7302, | |
| "step": 9210 | |
| }, | |
| { | |
| "epoch": 184.4, | |
| "learning_rate": 4.751515151515152e-06, | |
| "loss": 0.6944, | |
| "step": 9220 | |
| }, | |
| { | |
| "epoch": 184.6, | |
| "learning_rate": 4.690909090909092e-06, | |
| "loss": 0.6774, | |
| "step": 9230 | |
| }, | |
| { | |
| "epoch": 184.8, | |
| "learning_rate": 4.63030303030303e-06, | |
| "loss": 0.6884, | |
| "step": 9240 | |
| }, | |
| { | |
| "epoch": 185.0, | |
| "learning_rate": 4.56969696969697e-06, | |
| "loss": 0.6506, | |
| "step": 9250 | |
| }, | |
| { | |
| "epoch": 185.2, | |
| "learning_rate": 4.509090909090909e-06, | |
| "loss": 0.7361, | |
| "step": 9260 | |
| }, | |
| { | |
| "epoch": 185.4, | |
| "learning_rate": 4.448484848484849e-06, | |
| "loss": 0.6978, | |
| "step": 9270 | |
| }, | |
| { | |
| "epoch": 185.6, | |
| "learning_rate": 4.3878787878787876e-06, | |
| "loss": 0.68, | |
| "step": 9280 | |
| }, | |
| { | |
| "epoch": 185.8, | |
| "learning_rate": 4.327272727272728e-06, | |
| "loss": 0.6859, | |
| "step": 9290 | |
| }, | |
| { | |
| "epoch": 186.0, | |
| "learning_rate": 4.266666666666667e-06, | |
| "loss": 0.6716, | |
| "step": 9300 | |
| }, | |
| { | |
| "epoch": 186.2, | |
| "learning_rate": 4.206060606060606e-06, | |
| "loss": 0.6995, | |
| "step": 9310 | |
| }, | |
| { | |
| "epoch": 186.4, | |
| "learning_rate": 4.145454545454546e-06, | |
| "loss": 0.7181, | |
| "step": 9320 | |
| }, | |
| { | |
| "epoch": 186.6, | |
| "learning_rate": 4.084848484848484e-06, | |
| "loss": 0.7069, | |
| "step": 9330 | |
| }, | |
| { | |
| "epoch": 186.8, | |
| "learning_rate": 4.024242424242424e-06, | |
| "loss": 0.7083, | |
| "step": 9340 | |
| }, | |
| { | |
| "epoch": 187.0, | |
| "learning_rate": 3.9636363636363635e-06, | |
| "loss": 0.662, | |
| "step": 9350 | |
| }, | |
| { | |
| "epoch": 187.2, | |
| "learning_rate": 3.9030303030303035e-06, | |
| "loss": 0.7155, | |
| "step": 9360 | |
| }, | |
| { | |
| "epoch": 187.4, | |
| "learning_rate": 3.842424242424243e-06, | |
| "loss": 0.7061, | |
| "step": 9370 | |
| }, | |
| { | |
| "epoch": 187.6, | |
| "learning_rate": 3.7818181818181823e-06, | |
| "loss": 0.6878, | |
| "step": 9380 | |
| }, | |
| { | |
| "epoch": 187.8, | |
| "learning_rate": 3.721212121212121e-06, | |
| "loss": 0.687, | |
| "step": 9390 | |
| }, | |
| { | |
| "epoch": 188.0, | |
| "learning_rate": 3.6606060606060607e-06, | |
| "loss": 0.6586, | |
| "step": 9400 | |
| }, | |
| { | |
| "epoch": 188.2, | |
| "learning_rate": 3.6e-06, | |
| "loss": 0.6959, | |
| "step": 9410 | |
| }, | |
| { | |
| "epoch": 188.4, | |
| "learning_rate": 3.5393939393939394e-06, | |
| "loss": 0.698, | |
| "step": 9420 | |
| }, | |
| { | |
| "epoch": 188.6, | |
| "learning_rate": 3.4787878787878786e-06, | |
| "loss": 0.6989, | |
| "step": 9430 | |
| }, | |
| { | |
| "epoch": 188.8, | |
| "learning_rate": 3.4181818181818182e-06, | |
| "loss": 0.6927, | |
| "step": 9440 | |
| }, | |
| { | |
| "epoch": 189.0, | |
| "learning_rate": 3.357575757575758e-06, | |
| "loss": 0.6649, | |
| "step": 9450 | |
| }, | |
| { | |
| "epoch": 189.2, | |
| "learning_rate": 3.296969696969697e-06, | |
| "loss": 0.715, | |
| "step": 9460 | |
| }, | |
| { | |
| "epoch": 189.4, | |
| "learning_rate": 3.2363636363636366e-06, | |
| "loss": 0.694, | |
| "step": 9470 | |
| }, | |
| { | |
| "epoch": 189.6, | |
| "learning_rate": 3.1757575757575758e-06, | |
| "loss": 0.7162, | |
| "step": 9480 | |
| }, | |
| { | |
| "epoch": 189.8, | |
| "learning_rate": 3.1151515151515154e-06, | |
| "loss": 0.685, | |
| "step": 9490 | |
| }, | |
| { | |
| "epoch": 190.0, | |
| "learning_rate": 3.0545454545454546e-06, | |
| "loss": 0.6471, | |
| "step": 9500 | |
| }, | |
| { | |
| "epoch": 190.0, | |
| "eval_loss": 0.37067291140556335, | |
| "eval_runtime": 243.2712, | |
| "eval_samples_per_second": 21.425, | |
| "eval_steps_per_second": 1.34, | |
| "eval_wer": 0.3420367574321862, | |
| "step": 9500 | |
| }, | |
| { | |
| "epoch": 190.2, | |
| "learning_rate": 2.993939393939394e-06, | |
| "loss": 0.7389, | |
| "step": 9510 | |
| }, | |
| { | |
| "epoch": 190.4, | |
| "learning_rate": 2.9333333333333338e-06, | |
| "loss": 0.6763, | |
| "step": 9520 | |
| }, | |
| { | |
| "epoch": 190.6, | |
| "learning_rate": 2.872727272727273e-06, | |
| "loss": 0.6837, | |
| "step": 9530 | |
| }, | |
| { | |
| "epoch": 190.8, | |
| "learning_rate": 2.8121212121212125e-06, | |
| "loss": 0.6793, | |
| "step": 9540 | |
| }, | |
| { | |
| "epoch": 191.0, | |
| "learning_rate": 2.7515151515151513e-06, | |
| "loss": 0.6621, | |
| "step": 9550 | |
| }, | |
| { | |
| "epoch": 191.2, | |
| "learning_rate": 2.690909090909091e-06, | |
| "loss": 0.7213, | |
| "step": 9560 | |
| }, | |
| { | |
| "epoch": 191.4, | |
| "learning_rate": 2.63030303030303e-06, | |
| "loss": 0.6993, | |
| "step": 9570 | |
| }, | |
| { | |
| "epoch": 191.6, | |
| "learning_rate": 2.5696969696969697e-06, | |
| "loss": 0.6682, | |
| "step": 9580 | |
| }, | |
| { | |
| "epoch": 191.8, | |
| "learning_rate": 2.5090909090909093e-06, | |
| "loss": 0.6759, | |
| "step": 9590 | |
| }, | |
| { | |
| "epoch": 192.0, | |
| "learning_rate": 2.4484848484848485e-06, | |
| "loss": 0.6718, | |
| "step": 9600 | |
| }, | |
| { | |
| "epoch": 192.2, | |
| "learning_rate": 2.387878787878788e-06, | |
| "loss": 0.7276, | |
| "step": 9610 | |
| }, | |
| { | |
| "epoch": 192.4, | |
| "learning_rate": 2.3272727272727272e-06, | |
| "loss": 0.6857, | |
| "step": 9620 | |
| }, | |
| { | |
| "epoch": 192.6, | |
| "learning_rate": 2.266666666666667e-06, | |
| "loss": 0.6914, | |
| "step": 9630 | |
| }, | |
| { | |
| "epoch": 192.8, | |
| "learning_rate": 2.206060606060606e-06, | |
| "loss": 0.7176, | |
| "step": 9640 | |
| }, | |
| { | |
| "epoch": 193.0, | |
| "learning_rate": 2.1454545454545456e-06, | |
| "loss": 0.6491, | |
| "step": 9650 | |
| }, | |
| { | |
| "epoch": 193.2, | |
| "learning_rate": 2.084848484848485e-06, | |
| "loss": 0.6961, | |
| "step": 9660 | |
| }, | |
| { | |
| "epoch": 193.4, | |
| "learning_rate": 2.0242424242424244e-06, | |
| "loss": 0.7066, | |
| "step": 9670 | |
| }, | |
| { | |
| "epoch": 193.6, | |
| "learning_rate": 1.963636363636364e-06, | |
| "loss": 0.6874, | |
| "step": 9680 | |
| }, | |
| { | |
| "epoch": 193.8, | |
| "learning_rate": 1.9030303030303032e-06, | |
| "loss": 0.7069, | |
| "step": 9690 | |
| }, | |
| { | |
| "epoch": 194.0, | |
| "learning_rate": 1.8424242424242424e-06, | |
| "loss": 0.6636, | |
| "step": 9700 | |
| }, | |
| { | |
| "epoch": 194.2, | |
| "learning_rate": 1.781818181818182e-06, | |
| "loss": 0.7159, | |
| "step": 9710 | |
| }, | |
| { | |
| "epoch": 194.4, | |
| "learning_rate": 1.7212121212121214e-06, | |
| "loss": 0.6926, | |
| "step": 9720 | |
| }, | |
| { | |
| "epoch": 194.6, | |
| "learning_rate": 1.6606060606060607e-06, | |
| "loss": 0.6831, | |
| "step": 9730 | |
| }, | |
| { | |
| "epoch": 194.8, | |
| "learning_rate": 1.6000000000000001e-06, | |
| "loss": 0.6727, | |
| "step": 9740 | |
| }, | |
| { | |
| "epoch": 195.0, | |
| "learning_rate": 1.5393939393939393e-06, | |
| "loss": 0.6714, | |
| "step": 9750 | |
| }, | |
| { | |
| "epoch": 195.2, | |
| "learning_rate": 1.4787878787878787e-06, | |
| "loss": 0.7056, | |
| "step": 9760 | |
| }, | |
| { | |
| "epoch": 195.4, | |
| "learning_rate": 1.418181818181818e-06, | |
| "loss": 0.68, | |
| "step": 9770 | |
| }, | |
| { | |
| "epoch": 195.6, | |
| "learning_rate": 1.3575757575757577e-06, | |
| "loss": 0.6768, | |
| "step": 9780 | |
| }, | |
| { | |
| "epoch": 195.8, | |
| "learning_rate": 1.296969696969697e-06, | |
| "loss": 0.6806, | |
| "step": 9790 | |
| }, | |
| { | |
| "epoch": 196.0, | |
| "learning_rate": 1.2363636363636365e-06, | |
| "loss": 0.6654, | |
| "step": 9800 | |
| }, | |
| { | |
| "epoch": 196.2, | |
| "learning_rate": 1.1757575757575759e-06, | |
| "loss": 0.7139, | |
| "step": 9810 | |
| }, | |
| { | |
| "epoch": 196.4, | |
| "learning_rate": 1.1151515151515153e-06, | |
| "loss": 0.6718, | |
| "step": 9820 | |
| }, | |
| { | |
| "epoch": 196.6, | |
| "learning_rate": 1.0545454545454544e-06, | |
| "loss": 0.6785, | |
| "step": 9830 | |
| }, | |
| { | |
| "epoch": 196.8, | |
| "learning_rate": 9.939393939393938e-07, | |
| "loss": 0.6821, | |
| "step": 9840 | |
| }, | |
| { | |
| "epoch": 197.0, | |
| "learning_rate": 9.333333333333333e-07, | |
| "loss": 0.6711, | |
| "step": 9850 | |
| }, | |
| { | |
| "epoch": 197.2, | |
| "learning_rate": 8.727272727272727e-07, | |
| "loss": 0.7221, | |
| "step": 9860 | |
| }, | |
| { | |
| "epoch": 197.4, | |
| "learning_rate": 8.121212121212122e-07, | |
| "loss": 0.6933, | |
| "step": 9870 | |
| }, | |
| { | |
| "epoch": 197.6, | |
| "learning_rate": 7.515151515151516e-07, | |
| "loss": 0.6855, | |
| "step": 9880 | |
| }, | |
| { | |
| "epoch": 197.8, | |
| "learning_rate": 6.909090909090909e-07, | |
| "loss": 0.6902, | |
| "step": 9890 | |
| }, | |
| { | |
| "epoch": 198.0, | |
| "learning_rate": 6.303030303030303e-07, | |
| "loss": 0.6686, | |
| "step": 9900 | |
| }, | |
| { | |
| "epoch": 198.2, | |
| "learning_rate": 5.696969696969698e-07, | |
| "loss": 0.7282, | |
| "step": 9910 | |
| }, | |
| { | |
| "epoch": 198.4, | |
| "learning_rate": 5.090909090909092e-07, | |
| "loss": 0.6848, | |
| "step": 9920 | |
| }, | |
| { | |
| "epoch": 198.6, | |
| "learning_rate": 4.484848484848485e-07, | |
| "loss": 0.6939, | |
| "step": 9930 | |
| }, | |
| { | |
| "epoch": 198.8, | |
| "learning_rate": 3.878787878787879e-07, | |
| "loss": 0.6961, | |
| "step": 9940 | |
| }, | |
| { | |
| "epoch": 199.0, | |
| "learning_rate": 3.3333333333333335e-07, | |
| "loss": 0.6564, | |
| "step": 9950 | |
| }, | |
| { | |
| "epoch": 199.2, | |
| "learning_rate": 2.7272727272727274e-07, | |
| "loss": 0.7132, | |
| "step": 9960 | |
| }, | |
| { | |
| "epoch": 199.4, | |
| "learning_rate": 2.1212121212121213e-07, | |
| "loss": 0.6813, | |
| "step": 9970 | |
| }, | |
| { | |
| "epoch": 199.6, | |
| "learning_rate": 1.5151515151515152e-07, | |
| "loss": 0.6865, | |
| "step": 9980 | |
| }, | |
| { | |
| "epoch": 199.8, | |
| "learning_rate": 9.090909090909091e-08, | |
| "loss": 0.6773, | |
| "step": 9990 | |
| }, | |
| { | |
| "epoch": 200.0, | |
| "learning_rate": 3.0303030303030305e-08, | |
| "loss": 0.6759, | |
| "step": 10000 | |
| }, | |
| { | |
| "epoch": 200.0, | |
| "eval_loss": 0.37058258056640625, | |
| "eval_runtime": 244.1231, | |
| "eval_samples_per_second": 21.35, | |
| "eval_steps_per_second": 1.335, | |
| "eval_wer": 0.3420888217837247, | |
| "step": 10000 | |
| }, | |
| { | |
| "epoch": 200.0, | |
| "step": 10000, | |
| "total_flos": 3.2791120983682476e+20, | |
| "train_loss": 0.48956193776130674, | |
| "train_runtime": 78626.4431, | |
| "train_samples_per_second": 32.574, | |
| "train_steps_per_second": 0.127 | |
| } | |
| ], | |
| "max_steps": 10000, | |
| "num_train_epochs": 200, | |
| "total_flos": 3.2791120983682476e+20, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |