| { | |
| "best_metric": 0.9261042886093471, | |
| "best_model_checkpoint": "videomae-surf-analytics-runpod7/checkpoint-8463", | |
| "epoch": 5.002391304347826, | |
| "eval_steps": 500, | |
| "global_step": 8857, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.0004347826086956522, | |
| "grad_norm": 10.079814910888672, | |
| "learning_rate": 2.173913043478261e-07, | |
| "loss": 1.3951, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.0008695652173913044, | |
| "grad_norm": 9.322056770324707, | |
| "learning_rate": 4.347826086956522e-07, | |
| "loss": 1.3872, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.0013043478260869566, | |
| "grad_norm": 6.732794761657715, | |
| "learning_rate": 6.521739130434782e-07, | |
| "loss": 1.3724, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.0017391304347826088, | |
| "grad_norm": 12.304842948913574, | |
| "learning_rate": 8.695652173913044e-07, | |
| "loss": 1.3522, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.002173913043478261, | |
| "grad_norm": 8.10595703125, | |
| "learning_rate": 1.0869565217391306e-06, | |
| "loss": 1.3382, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.0026086956521739132, | |
| "grad_norm": 6.890604019165039, | |
| "learning_rate": 1.3043478260869564e-06, | |
| "loss": 1.3565, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.003043478260869565, | |
| "grad_norm": 11.639082908630371, | |
| "learning_rate": 1.5217391304347827e-06, | |
| "loss": 1.2803, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 0.0034782608695652175, | |
| "grad_norm": 9.563755989074707, | |
| "learning_rate": 1.7391304347826088e-06, | |
| "loss": 1.2934, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.00391304347826087, | |
| "grad_norm": 5.474796295166016, | |
| "learning_rate": 1.956521739130435e-06, | |
| "loss": 1.2803, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 0.004043478260869565, | |
| "eval_accuracy": 0.4426229508196721, | |
| "eval_f1": 0.28327868852459015, | |
| "eval_loss": 1.2739324569702148, | |
| "eval_runtime": 99.1883, | |
| "eval_samples_per_second": 1.23, | |
| "eval_steps_per_second": 0.161, | |
| "step": 93 | |
| }, | |
| { | |
| "epoch": 1.000304347826087, | |
| "grad_norm": 4.489353179931641, | |
| "learning_rate": 2.173913043478261e-06, | |
| "loss": 1.2057, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 1.0007391304347826, | |
| "grad_norm": 4.060410976409912, | |
| "learning_rate": 2.391304347826087e-06, | |
| "loss": 1.2485, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 1.0011739130434782, | |
| "grad_norm": 6.812442302703857, | |
| "learning_rate": 2.608695652173913e-06, | |
| "loss": 1.1412, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 1.001608695652174, | |
| "grad_norm": 6.631100177764893, | |
| "learning_rate": 2.8260869565217393e-06, | |
| "loss": 1.1666, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 1.0020434782608696, | |
| "grad_norm": 5.362147331237793, | |
| "learning_rate": 3.0434782608695654e-06, | |
| "loss": 1.1275, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 1.0024782608695653, | |
| "grad_norm": 6.68129301071167, | |
| "learning_rate": 3.2608695652173914e-06, | |
| "loss": 1.1033, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 1.002913043478261, | |
| "grad_norm": 13.178942680358887, | |
| "learning_rate": 3.4782608695652175e-06, | |
| "loss": 1.1418, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 1.0033478260869566, | |
| "grad_norm": 13.012588500976562, | |
| "learning_rate": 3.695652173913044e-06, | |
| "loss": 1.265, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 1.0037826086956523, | |
| "grad_norm": 5.870707988739014, | |
| "learning_rate": 3.91304347826087e-06, | |
| "loss": 1.0844, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 1.0040434782608696, | |
| "eval_accuracy": 0.45081967213114754, | |
| "eval_f1": 0.285264425303325, | |
| "eval_loss": 1.186285376548767, | |
| "eval_runtime": 74.0993, | |
| "eval_samples_per_second": 1.646, | |
| "eval_steps_per_second": 0.216, | |
| "step": 186 | |
| }, | |
| { | |
| "epoch": 2.0001739130434784, | |
| "grad_norm": 6.740423679351807, | |
| "learning_rate": 4.130434782608695e-06, | |
| "loss": 1.1175, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 2.000608695652174, | |
| "grad_norm": 8.124707221984863, | |
| "learning_rate": 4.347826086956522e-06, | |
| "loss": 1.0575, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 2.0010434782608697, | |
| "grad_norm": 7.8553266525268555, | |
| "learning_rate": 4.565217391304348e-06, | |
| "loss": 1.0517, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 2.001478260869565, | |
| "grad_norm": 9.870964050292969, | |
| "learning_rate": 4.782608695652174e-06, | |
| "loss": 1.1031, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 2.001913043478261, | |
| "grad_norm": 8.00858211517334, | |
| "learning_rate": 5e-06, | |
| "loss": 0.9564, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 2.0023478260869565, | |
| "grad_norm": 8.030350685119629, | |
| "learning_rate": 5.217391304347826e-06, | |
| "loss": 1.0014, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 2.0027826086956524, | |
| "grad_norm": 6.9377522468566895, | |
| "learning_rate": 5.4347826086956525e-06, | |
| "loss": 0.9769, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 2.003217391304348, | |
| "grad_norm": 6.949226379394531, | |
| "learning_rate": 5.652173913043479e-06, | |
| "loss": 1.0181, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 2.0036521739130433, | |
| "grad_norm": 8.180449485778809, | |
| "learning_rate": 5.869565217391305e-06, | |
| "loss": 0.9224, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 2.0040434782608694, | |
| "eval_accuracy": 0.5245901639344263, | |
| "eval_f1": 0.4366849047835448, | |
| "eval_loss": 1.0750652551651, | |
| "eval_runtime": 69.4086, | |
| "eval_samples_per_second": 1.758, | |
| "eval_steps_per_second": 0.231, | |
| "step": 279 | |
| }, | |
| { | |
| "epoch": 3.0000434782608694, | |
| "grad_norm": 6.714990139007568, | |
| "learning_rate": 6.086956521739131e-06, | |
| "loss": 0.9395, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 3.0004782608695653, | |
| "grad_norm": 9.589141845703125, | |
| "learning_rate": 6.304347826086957e-06, | |
| "loss": 0.9219, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 3.0009130434782607, | |
| "grad_norm": 17.355304718017578, | |
| "learning_rate": 6.521739130434783e-06, | |
| "loss": 0.764, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 3.0013478260869566, | |
| "grad_norm": 14.70441722869873, | |
| "learning_rate": 6.739130434782609e-06, | |
| "loss": 0.9942, | |
| "step": 310 | |
| }, | |
| { | |
| "epoch": 3.001782608695652, | |
| "grad_norm": 8.38525104522705, | |
| "learning_rate": 6.956521739130435e-06, | |
| "loss": 0.9428, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 3.002217391304348, | |
| "grad_norm": 9.241342544555664, | |
| "learning_rate": 7.173913043478261e-06, | |
| "loss": 0.8697, | |
| "step": 330 | |
| }, | |
| { | |
| "epoch": 3.0026521739130434, | |
| "grad_norm": 20.59735107421875, | |
| "learning_rate": 7.391304347826088e-06, | |
| "loss": 0.7409, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 3.0030869565217393, | |
| "grad_norm": 9.710801124572754, | |
| "learning_rate": 7.608695652173914e-06, | |
| "loss": 0.8818, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 3.0035217391304347, | |
| "grad_norm": 8.560890197753906, | |
| "learning_rate": 7.82608695652174e-06, | |
| "loss": 0.8719, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 3.0039565217391306, | |
| "grad_norm": 6.683012962341309, | |
| "learning_rate": 8.043478260869565e-06, | |
| "loss": 0.7226, | |
| "step": 370 | |
| }, | |
| { | |
| "epoch": 3.0040434782608694, | |
| "eval_accuracy": 0.5655737704918032, | |
| "eval_f1": 0.4909767680959837, | |
| "eval_loss": 1.003716230392456, | |
| "eval_runtime": 71.8801, | |
| "eval_samples_per_second": 1.697, | |
| "eval_steps_per_second": 0.223, | |
| "step": 372 | |
| }, | |
| { | |
| "epoch": 4.000347826086957, | |
| "grad_norm": 10.244637489318848, | |
| "learning_rate": 8.26086956521739e-06, | |
| "loss": 0.678, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 4.000782608695652, | |
| "grad_norm": 19.61251449584961, | |
| "learning_rate": 8.478260869565217e-06, | |
| "loss": 0.6654, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 4.001217391304348, | |
| "grad_norm": 15.357687950134277, | |
| "learning_rate": 8.695652173913044e-06, | |
| "loss": 0.7421, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 4.001652173913043, | |
| "grad_norm": 10.771465301513672, | |
| "learning_rate": 8.91304347826087e-06, | |
| "loss": 0.5394, | |
| "step": 410 | |
| }, | |
| { | |
| "epoch": 4.002086956521739, | |
| "grad_norm": 26.172359466552734, | |
| "learning_rate": 9.130434782608697e-06, | |
| "loss": 0.7181, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 4.002521739130435, | |
| "grad_norm": 24.619651794433594, | |
| "learning_rate": 9.347826086956523e-06, | |
| "loss": 0.605, | |
| "step": 430 | |
| }, | |
| { | |
| "epoch": 4.00295652173913, | |
| "grad_norm": 23.711095809936523, | |
| "learning_rate": 9.565217391304349e-06, | |
| "loss": 0.7164, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 4.003391304347826, | |
| "grad_norm": 11.760732650756836, | |
| "learning_rate": 9.782608695652175e-06, | |
| "loss": 0.5269, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 4.003826086956522, | |
| "grad_norm": 21.210412979125977, | |
| "learning_rate": 1e-05, | |
| "loss": 0.7025, | |
| "step": 460 | |
| }, | |
| { | |
| "epoch": 4.004043478260869, | |
| "eval_accuracy": 0.5983606557377049, | |
| "eval_f1": 0.5695439501100179, | |
| "eval_loss": 0.9121292233467102, | |
| "eval_runtime": 78.8043, | |
| "eval_samples_per_second": 1.548, | |
| "eval_steps_per_second": 0.203, | |
| "step": 465 | |
| }, | |
| { | |
| "epoch": 5.000217391304348, | |
| "grad_norm": 6.867463111877441, | |
| "learning_rate": 1.0217391304347827e-05, | |
| "loss": 0.4847, | |
| "step": 470 | |
| }, | |
| { | |
| "epoch": 5.000652173913044, | |
| "grad_norm": 3.2302322387695312, | |
| "learning_rate": 1.0434782608695651e-05, | |
| "loss": 0.3333, | |
| "step": 480 | |
| }, | |
| { | |
| "epoch": 5.001086956521739, | |
| "grad_norm": 21.763275146484375, | |
| "learning_rate": 1.0652173913043479e-05, | |
| "loss": 0.494, | |
| "step": 490 | |
| }, | |
| { | |
| "epoch": 5.0015217391304345, | |
| "grad_norm": 13.490067481994629, | |
| "learning_rate": 1.0869565217391305e-05, | |
| "loss": 0.3955, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 5.001956521739131, | |
| "grad_norm": 20.92444610595703, | |
| "learning_rate": 1.1086956521739131e-05, | |
| "loss": 0.627, | |
| "step": 510 | |
| }, | |
| { | |
| "epoch": 5.002391304347826, | |
| "grad_norm": 23.297204971313477, | |
| "learning_rate": 1.1304347826086957e-05, | |
| "loss": 0.5445, | |
| "step": 520 | |
| }, | |
| { | |
| "epoch": 5.002826086956522, | |
| "grad_norm": 30.925203323364258, | |
| "learning_rate": 1.1521739130434783e-05, | |
| "loss": 0.4394, | |
| "step": 530 | |
| }, | |
| { | |
| "epoch": 5.003260869565217, | |
| "grad_norm": 35.236183166503906, | |
| "learning_rate": 1.173913043478261e-05, | |
| "loss": 0.6552, | |
| "step": 540 | |
| }, | |
| { | |
| "epoch": 5.003695652173913, | |
| "grad_norm": 4.368950843811035, | |
| "learning_rate": 1.1956521739130435e-05, | |
| "loss": 0.5447, | |
| "step": 550 | |
| }, | |
| { | |
| "epoch": 5.004043478260869, | |
| "eval_accuracy": 0.6065573770491803, | |
| "eval_f1": 0.5667313624233182, | |
| "eval_loss": 0.9521207809448242, | |
| "eval_runtime": 75.9971, | |
| "eval_samples_per_second": 1.605, | |
| "eval_steps_per_second": 0.211, | |
| "step": 558 | |
| }, | |
| { | |
| "epoch": 6.000086956521739, | |
| "grad_norm": 6.621151447296143, | |
| "learning_rate": 1.2173913043478261e-05, | |
| "loss": 0.4113, | |
| "step": 560 | |
| }, | |
| { | |
| "epoch": 6.000521739130435, | |
| "grad_norm": 20.69222640991211, | |
| "learning_rate": 1.2391304347826088e-05, | |
| "loss": 0.3668, | |
| "step": 570 | |
| }, | |
| { | |
| "epoch": 6.0009565217391305, | |
| "grad_norm": 11.611593246459961, | |
| "learning_rate": 1.2608695652173914e-05, | |
| "loss": 0.3163, | |
| "step": 580 | |
| }, | |
| { | |
| "epoch": 6.001391304347826, | |
| "grad_norm": 24.46592140197754, | |
| "learning_rate": 1.2826086956521741e-05, | |
| "loss": 0.4861, | |
| "step": 590 | |
| }, | |
| { | |
| "epoch": 6.001826086956521, | |
| "grad_norm": 4.075656890869141, | |
| "learning_rate": 1.3043478260869566e-05, | |
| "loss": 0.2725, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 6.002260869565218, | |
| "grad_norm": 46.16673278808594, | |
| "learning_rate": 1.3260869565217394e-05, | |
| "loss": 0.5276, | |
| "step": 610 | |
| }, | |
| { | |
| "epoch": 6.002695652173913, | |
| "grad_norm": 25.51651382446289, | |
| "learning_rate": 1.3478260869565218e-05, | |
| "loss": 0.3851, | |
| "step": 620 | |
| }, | |
| { | |
| "epoch": 6.003130434782609, | |
| "grad_norm": 13.07934284210205, | |
| "learning_rate": 1.3695652173913042e-05, | |
| "loss": 0.4227, | |
| "step": 630 | |
| }, | |
| { | |
| "epoch": 6.003565217391304, | |
| "grad_norm": 48.23696517944336, | |
| "learning_rate": 1.391304347826087e-05, | |
| "loss": 0.4125, | |
| "step": 640 | |
| }, | |
| { | |
| "epoch": 6.004, | |
| "grad_norm": 14.747217178344727, | |
| "learning_rate": 1.4130434782608694e-05, | |
| "loss": 0.4752, | |
| "step": 650 | |
| }, | |
| { | |
| "epoch": 6.004043478260869, | |
| "eval_accuracy": 0.7131147540983607, | |
| "eval_f1": 0.6802356113831525, | |
| "eval_loss": 0.8963390588760376, | |
| "eval_runtime": 68.2476, | |
| "eval_samples_per_second": 1.788, | |
| "eval_steps_per_second": 0.234, | |
| "step": 651 | |
| }, | |
| { | |
| "epoch": 7.0003913043478265, | |
| "grad_norm": 19.88422393798828, | |
| "learning_rate": 1.4347826086956522e-05, | |
| "loss": 0.2777, | |
| "step": 660 | |
| }, | |
| { | |
| "epoch": 7.000826086956522, | |
| "grad_norm": 13.482111930847168, | |
| "learning_rate": 1.4565217391304348e-05, | |
| "loss": 0.4554, | |
| "step": 670 | |
| }, | |
| { | |
| "epoch": 7.001260869565217, | |
| "grad_norm": 11.722796440124512, | |
| "learning_rate": 1.4782608695652176e-05, | |
| "loss": 0.2043, | |
| "step": 680 | |
| }, | |
| { | |
| "epoch": 7.001695652173913, | |
| "grad_norm": 16.32537269592285, | |
| "learning_rate": 1.5e-05, | |
| "loss": 0.1863, | |
| "step": 690 | |
| }, | |
| { | |
| "epoch": 7.002130434782608, | |
| "grad_norm": 3.987351179122925, | |
| "learning_rate": 1.5217391304347828e-05, | |
| "loss": 0.1437, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 7.002565217391305, | |
| "grad_norm": 1.536703109741211, | |
| "learning_rate": 1.5434782608695654e-05, | |
| "loss": 0.2714, | |
| "step": 710 | |
| }, | |
| { | |
| "epoch": 7.003, | |
| "grad_norm": 31.207611083984375, | |
| "learning_rate": 1.565217391304348e-05, | |
| "loss": 0.5409, | |
| "step": 720 | |
| }, | |
| { | |
| "epoch": 7.003434782608696, | |
| "grad_norm": 10.564144134521484, | |
| "learning_rate": 1.5869565217391306e-05, | |
| "loss": 0.347, | |
| "step": 730 | |
| }, | |
| { | |
| "epoch": 7.003869565217391, | |
| "grad_norm": 15.847654342651367, | |
| "learning_rate": 1.608695652173913e-05, | |
| "loss": 0.3256, | |
| "step": 740 | |
| }, | |
| { | |
| "epoch": 7.004043478260869, | |
| "eval_accuracy": 0.7213114754098361, | |
| "eval_f1": 0.7152792066865421, | |
| "eval_loss": 0.7381582260131836, | |
| "eval_runtime": 85.386, | |
| "eval_samples_per_second": 1.429, | |
| "eval_steps_per_second": 0.187, | |
| "step": 744 | |
| }, | |
| { | |
| "epoch": 8.000260869565217, | |
| "grad_norm": 4.257699489593506, | |
| "learning_rate": 1.630434782608696e-05, | |
| "loss": 0.1054, | |
| "step": 750 | |
| }, | |
| { | |
| "epoch": 8.000695652173913, | |
| "grad_norm": 2.503180742263794, | |
| "learning_rate": 1.652173913043478e-05, | |
| "loss": 0.182, | |
| "step": 760 | |
| }, | |
| { | |
| "epoch": 8.001130434782608, | |
| "grad_norm": 13.347521781921387, | |
| "learning_rate": 1.673913043478261e-05, | |
| "loss": 0.1134, | |
| "step": 770 | |
| }, | |
| { | |
| "epoch": 8.001565217391304, | |
| "grad_norm": 7.592967510223389, | |
| "learning_rate": 1.6956521739130433e-05, | |
| "loss": 0.2086, | |
| "step": 780 | |
| }, | |
| { | |
| "epoch": 8.002, | |
| "grad_norm": 9.522619247436523, | |
| "learning_rate": 1.7173913043478263e-05, | |
| "loss": 0.2546, | |
| "step": 790 | |
| }, | |
| { | |
| "epoch": 8.002434782608695, | |
| "grad_norm": 22.83403778076172, | |
| "learning_rate": 1.739130434782609e-05, | |
| "loss": 0.4554, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 8.002869565217392, | |
| "grad_norm": 25.185588836669922, | |
| "learning_rate": 1.7608695652173915e-05, | |
| "loss": 0.3155, | |
| "step": 810 | |
| }, | |
| { | |
| "epoch": 8.003304347826086, | |
| "grad_norm": 4.784695625305176, | |
| "learning_rate": 1.782608695652174e-05, | |
| "loss": 0.2442, | |
| "step": 820 | |
| }, | |
| { | |
| "epoch": 8.003739130434782, | |
| "grad_norm": 49.901519775390625, | |
| "learning_rate": 1.8043478260869567e-05, | |
| "loss": 0.1971, | |
| "step": 830 | |
| }, | |
| { | |
| "epoch": 8.00404347826087, | |
| "eval_accuracy": 0.6885245901639344, | |
| "eval_f1": 0.652401417303907, | |
| "eval_loss": 1.1315979957580566, | |
| "eval_runtime": 75.6618, | |
| "eval_samples_per_second": 1.612, | |
| "eval_steps_per_second": 0.211, | |
| "step": 837 | |
| }, | |
| { | |
| "epoch": 9.000130434782609, | |
| "grad_norm": 11.014081001281738, | |
| "learning_rate": 1.8260869565217393e-05, | |
| "loss": 0.2153, | |
| "step": 840 | |
| }, | |
| { | |
| "epoch": 9.000565217391305, | |
| "grad_norm": 3.1919617652893066, | |
| "learning_rate": 1.8478260869565216e-05, | |
| "loss": 0.143, | |
| "step": 850 | |
| }, | |
| { | |
| "epoch": 9.001, | |
| "grad_norm": 50.34938049316406, | |
| "learning_rate": 1.8695652173913045e-05, | |
| "loss": 0.1954, | |
| "step": 860 | |
| }, | |
| { | |
| "epoch": 9.001434782608696, | |
| "grad_norm": 42.44783401489258, | |
| "learning_rate": 1.8913043478260868e-05, | |
| "loss": 0.2195, | |
| "step": 870 | |
| }, | |
| { | |
| "epoch": 9.001869565217392, | |
| "grad_norm": 27.059328079223633, | |
| "learning_rate": 1.9130434782608697e-05, | |
| "loss": 0.1425, | |
| "step": 880 | |
| }, | |
| { | |
| "epoch": 9.002304347826087, | |
| "grad_norm": 58.7094841003418, | |
| "learning_rate": 1.9347826086956523e-05, | |
| "loss": 0.3012, | |
| "step": 890 | |
| }, | |
| { | |
| "epoch": 9.002739130434783, | |
| "grad_norm": 20.4552001953125, | |
| "learning_rate": 1.956521739130435e-05, | |
| "loss": 0.445, | |
| "step": 900 | |
| }, | |
| { | |
| "epoch": 9.003173913043478, | |
| "grad_norm": 0.22613933682441711, | |
| "learning_rate": 1.9782608695652176e-05, | |
| "loss": 0.3575, | |
| "step": 910 | |
| }, | |
| { | |
| "epoch": 9.003608695652174, | |
| "grad_norm": 24.949710845947266, | |
| "learning_rate": 2e-05, | |
| "loss": 0.2652, | |
| "step": 920 | |
| }, | |
| { | |
| "epoch": 9.00404347826087, | |
| "grad_norm": 26.43189811706543, | |
| "learning_rate": 2.0217391304347828e-05, | |
| "loss": 0.5144, | |
| "step": 930 | |
| }, | |
| { | |
| "epoch": 9.00404347826087, | |
| "eval_accuracy": 0.7459016393442623, | |
| "eval_f1": 0.7200532157059749, | |
| "eval_loss": 0.8559982180595398, | |
| "eval_runtime": 77.4935, | |
| "eval_samples_per_second": 1.574, | |
| "eval_steps_per_second": 0.206, | |
| "step": 930 | |
| }, | |
| { | |
| "epoch": 10.000434782608696, | |
| "grad_norm": 1.3086036443710327, | |
| "learning_rate": 2.0434782608695654e-05, | |
| "loss": 0.2612, | |
| "step": 940 | |
| }, | |
| { | |
| "epoch": 10.000869565217391, | |
| "grad_norm": 8.613253593444824, | |
| "learning_rate": 2.065217391304348e-05, | |
| "loss": 0.0451, | |
| "step": 950 | |
| }, | |
| { | |
| "epoch": 10.001304347826087, | |
| "grad_norm": 0.1866026222705841, | |
| "learning_rate": 2.0869565217391303e-05, | |
| "loss": 0.1809, | |
| "step": 960 | |
| }, | |
| { | |
| "epoch": 10.001739130434782, | |
| "grad_norm": 5.5913825035095215, | |
| "learning_rate": 2.1086956521739132e-05, | |
| "loss": 0.2038, | |
| "step": 970 | |
| }, | |
| { | |
| "epoch": 10.002173913043478, | |
| "grad_norm": 0.5955354571342468, | |
| "learning_rate": 2.1304347826086958e-05, | |
| "loss": 0.2476, | |
| "step": 980 | |
| }, | |
| { | |
| "epoch": 10.002608695652174, | |
| "grad_norm": 2.7775936126708984, | |
| "learning_rate": 2.1521739130434784e-05, | |
| "loss": 0.0573, | |
| "step": 990 | |
| }, | |
| { | |
| "epoch": 10.003043478260869, | |
| "grad_norm": 0.2440994530916214, | |
| "learning_rate": 2.173913043478261e-05, | |
| "loss": 0.0826, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 10.003478260869565, | |
| "grad_norm": 0.09992615133523941, | |
| "learning_rate": 2.1956521739130436e-05, | |
| "loss": 0.2261, | |
| "step": 1010 | |
| }, | |
| { | |
| "epoch": 10.003913043478262, | |
| "grad_norm": 62.008052825927734, | |
| "learning_rate": 2.2173913043478262e-05, | |
| "loss": 0.2414, | |
| "step": 1020 | |
| }, | |
| { | |
| "epoch": 10.00404347826087, | |
| "eval_accuracy": 0.7704918032786885, | |
| "eval_f1": 0.7633172613307619, | |
| "eval_loss": 0.86429762840271, | |
| "eval_runtime": 71.2234, | |
| "eval_samples_per_second": 1.713, | |
| "eval_steps_per_second": 0.225, | |
| "step": 1023 | |
| }, | |
| { | |
| "epoch": 11.000304347826088, | |
| "grad_norm": 10.307206153869629, | |
| "learning_rate": 2.239130434782609e-05, | |
| "loss": 0.2326, | |
| "step": 1030 | |
| }, | |
| { | |
| "epoch": 11.000739130434782, | |
| "grad_norm": 0.6196430921554565, | |
| "learning_rate": 2.2608695652173914e-05, | |
| "loss": 0.0381, | |
| "step": 1040 | |
| }, | |
| { | |
| "epoch": 11.001173913043479, | |
| "grad_norm": 40.027503967285156, | |
| "learning_rate": 2.282608695652174e-05, | |
| "loss": 0.1084, | |
| "step": 1050 | |
| }, | |
| { | |
| "epoch": 11.001608695652173, | |
| "grad_norm": 14.861847877502441, | |
| "learning_rate": 2.3043478260869567e-05, | |
| "loss": 0.092, | |
| "step": 1060 | |
| }, | |
| { | |
| "epoch": 11.00204347826087, | |
| "grad_norm": 4.125478744506836, | |
| "learning_rate": 2.3260869565217393e-05, | |
| "loss": 0.1853, | |
| "step": 1070 | |
| }, | |
| { | |
| "epoch": 11.002478260869566, | |
| "grad_norm": 6.733792781829834, | |
| "learning_rate": 2.347826086956522e-05, | |
| "loss": 0.3253, | |
| "step": 1080 | |
| }, | |
| { | |
| "epoch": 11.00291304347826, | |
| "grad_norm": 78.53083038330078, | |
| "learning_rate": 2.3695652173913045e-05, | |
| "loss": 0.3464, | |
| "step": 1090 | |
| }, | |
| { | |
| "epoch": 11.003347826086957, | |
| "grad_norm": 23.05046844482422, | |
| "learning_rate": 2.391304347826087e-05, | |
| "loss": 0.194, | |
| "step": 1100 | |
| }, | |
| { | |
| "epoch": 11.003782608695651, | |
| "grad_norm": 0.8519552946090698, | |
| "learning_rate": 2.4130434782608697e-05, | |
| "loss": 0.0972, | |
| "step": 1110 | |
| }, | |
| { | |
| "epoch": 11.00404347826087, | |
| "eval_accuracy": 0.6967213114754098, | |
| "eval_f1": 0.6836640572123361, | |
| "eval_loss": 1.2529149055480957, | |
| "eval_runtime": 76.6021, | |
| "eval_samples_per_second": 1.593, | |
| "eval_steps_per_second": 0.209, | |
| "step": 1116 | |
| }, | |
| { | |
| "epoch": 12.000173913043477, | |
| "grad_norm": 0.48925498127937317, | |
| "learning_rate": 2.4347826086956523e-05, | |
| "loss": 0.0541, | |
| "step": 1120 | |
| }, | |
| { | |
| "epoch": 12.000608695652174, | |
| "grad_norm": 0.18016597628593445, | |
| "learning_rate": 2.456521739130435e-05, | |
| "loss": 0.0193, | |
| "step": 1130 | |
| }, | |
| { | |
| "epoch": 12.00104347826087, | |
| "grad_norm": 61.650970458984375, | |
| "learning_rate": 2.4782608695652175e-05, | |
| "loss": 0.1264, | |
| "step": 1140 | |
| }, | |
| { | |
| "epoch": 12.001478260869565, | |
| "grad_norm": 0.5460256934165955, | |
| "learning_rate": 2.5e-05, | |
| "loss": 0.1422, | |
| "step": 1150 | |
| }, | |
| { | |
| "epoch": 12.001913043478261, | |
| "grad_norm": 1.2666577100753784, | |
| "learning_rate": 2.5217391304347827e-05, | |
| "loss": 0.166, | |
| "step": 1160 | |
| }, | |
| { | |
| "epoch": 12.002347826086957, | |
| "grad_norm": 38.8865966796875, | |
| "learning_rate": 2.543478260869565e-05, | |
| "loss": 0.2384, | |
| "step": 1170 | |
| }, | |
| { | |
| "epoch": 12.002782608695652, | |
| "grad_norm": 7.161799907684326, | |
| "learning_rate": 2.5652173913043483e-05, | |
| "loss": 0.1389, | |
| "step": 1180 | |
| }, | |
| { | |
| "epoch": 12.003217391304348, | |
| "grad_norm": 31.820159912109375, | |
| "learning_rate": 2.5869565217391305e-05, | |
| "loss": 0.3968, | |
| "step": 1190 | |
| }, | |
| { | |
| "epoch": 12.003652173913043, | |
| "grad_norm": 18.057905197143555, | |
| "learning_rate": 2.608695652173913e-05, | |
| "loss": 0.3371, | |
| "step": 1200 | |
| }, | |
| { | |
| "epoch": 12.00404347826087, | |
| "eval_accuracy": 0.7213114754098361, | |
| "eval_f1": 0.7123017501384928, | |
| "eval_loss": 1.0838274955749512, | |
| "eval_runtime": 75.3169, | |
| "eval_samples_per_second": 1.62, | |
| "eval_steps_per_second": 0.212, | |
| "step": 1209 | |
| }, | |
| { | |
| "epoch": 13.000043478260869, | |
| "grad_norm": 50.66958236694336, | |
| "learning_rate": 2.6304347826086954e-05, | |
| "loss": 0.4032, | |
| "step": 1210 | |
| }, | |
| { | |
| "epoch": 13.000478260869565, | |
| "grad_norm": 4.794472694396973, | |
| "learning_rate": 2.6521739130434787e-05, | |
| "loss": 0.169, | |
| "step": 1220 | |
| }, | |
| { | |
| "epoch": 13.000913043478262, | |
| "grad_norm": 3.5523149967193604, | |
| "learning_rate": 2.673913043478261e-05, | |
| "loss": 0.0234, | |
| "step": 1230 | |
| }, | |
| { | |
| "epoch": 13.001347826086956, | |
| "grad_norm": 0.033308062702417374, | |
| "learning_rate": 2.6956521739130436e-05, | |
| "loss": 0.0155, | |
| "step": 1240 | |
| }, | |
| { | |
| "epoch": 13.001782608695652, | |
| "grad_norm": 2.988459587097168, | |
| "learning_rate": 2.7173913043478262e-05, | |
| "loss": 0.0085, | |
| "step": 1250 | |
| }, | |
| { | |
| "epoch": 13.002217391304347, | |
| "grad_norm": 78.12406921386719, | |
| "learning_rate": 2.7391304347826085e-05, | |
| "loss": 0.2082, | |
| "step": 1260 | |
| }, | |
| { | |
| "epoch": 13.002652173913043, | |
| "grad_norm": 23.308473587036133, | |
| "learning_rate": 2.7608695652173917e-05, | |
| "loss": 0.0724, | |
| "step": 1270 | |
| }, | |
| { | |
| "epoch": 13.00308695652174, | |
| "grad_norm": 0.23175761103630066, | |
| "learning_rate": 2.782608695652174e-05, | |
| "loss": 0.3893, | |
| "step": 1280 | |
| }, | |
| { | |
| "epoch": 13.003521739130434, | |
| "grad_norm": 5.151690483093262, | |
| "learning_rate": 2.8043478260869566e-05, | |
| "loss": 0.1066, | |
| "step": 1290 | |
| }, | |
| { | |
| "epoch": 13.00395652173913, | |
| "grad_norm": 0.07266629487276077, | |
| "learning_rate": 2.826086956521739e-05, | |
| "loss": 0.1133, | |
| "step": 1300 | |
| }, | |
| { | |
| "epoch": 13.00404347826087, | |
| "eval_accuracy": 0.7295081967213115, | |
| "eval_f1": 0.7128166457266807, | |
| "eval_loss": 1.2047704458236694, | |
| "eval_runtime": 70.4266, | |
| "eval_samples_per_second": 1.732, | |
| "eval_steps_per_second": 0.227, | |
| "step": 1302 | |
| }, | |
| { | |
| "epoch": 14.000347826086957, | |
| "grad_norm": 25.666845321655273, | |
| "learning_rate": 2.847826086956522e-05, | |
| "loss": 0.2143, | |
| "step": 1310 | |
| }, | |
| { | |
| "epoch": 14.000782608695653, | |
| "grad_norm": 0.48047909140586853, | |
| "learning_rate": 2.8695652173913044e-05, | |
| "loss": 0.111, | |
| "step": 1320 | |
| }, | |
| { | |
| "epoch": 14.001217391304348, | |
| "grad_norm": 1.6062359809875488, | |
| "learning_rate": 2.891304347826087e-05, | |
| "loss": 0.1202, | |
| "step": 1330 | |
| }, | |
| { | |
| "epoch": 14.001652173913044, | |
| "grad_norm": 46.08817672729492, | |
| "learning_rate": 2.9130434782608696e-05, | |
| "loss": 0.3498, | |
| "step": 1340 | |
| }, | |
| { | |
| "epoch": 14.002086956521739, | |
| "grad_norm": 0.1339997947216034, | |
| "learning_rate": 2.9347826086956526e-05, | |
| "loss": 0.366, | |
| "step": 1350 | |
| }, | |
| { | |
| "epoch": 14.002521739130435, | |
| "grad_norm": 49.66046905517578, | |
| "learning_rate": 2.9565217391304352e-05, | |
| "loss": 0.2443, | |
| "step": 1360 | |
| }, | |
| { | |
| "epoch": 14.002956521739131, | |
| "grad_norm": 0.571479082107544, | |
| "learning_rate": 2.9782608695652175e-05, | |
| "loss": 0.0953, | |
| "step": 1370 | |
| }, | |
| { | |
| "epoch": 14.003391304347826, | |
| "grad_norm": 0.11195419728755951, | |
| "learning_rate": 3e-05, | |
| "loss": 0.0104, | |
| "step": 1380 | |
| }, | |
| { | |
| "epoch": 14.003826086956522, | |
| "grad_norm": 0.2546157240867615, | |
| "learning_rate": 3.0217391304347827e-05, | |
| "loss": 0.1091, | |
| "step": 1390 | |
| }, | |
| { | |
| "epoch": 14.00404347826087, | |
| "eval_accuracy": 0.7868852459016393, | |
| "eval_f1": 0.772974027172348, | |
| "eval_loss": 1.0909240245819092, | |
| "eval_runtime": 78.9629, | |
| "eval_samples_per_second": 1.545, | |
| "eval_steps_per_second": 0.203, | |
| "step": 1395 | |
| }, | |
| { | |
| "epoch": 15.000217391304348, | |
| "grad_norm": 24.337543487548828, | |
| "learning_rate": 3.0434782608695656e-05, | |
| "loss": 0.1526, | |
| "step": 1400 | |
| }, | |
| { | |
| "epoch": 15.000652173913043, | |
| "grad_norm": 0.14164993166923523, | |
| "learning_rate": 3.065217391304348e-05, | |
| "loss": 0.0855, | |
| "step": 1410 | |
| }, | |
| { | |
| "epoch": 15.001086956521739, | |
| "grad_norm": 0.027599574998021126, | |
| "learning_rate": 3.086956521739131e-05, | |
| "loss": 0.1999, | |
| "step": 1420 | |
| }, | |
| { | |
| "epoch": 15.001521739130435, | |
| "grad_norm": 0.1311451494693756, | |
| "learning_rate": 3.108695652173913e-05, | |
| "loss": 0.1114, | |
| "step": 1430 | |
| }, | |
| { | |
| "epoch": 15.00195652173913, | |
| "grad_norm": 0.4929811358451843, | |
| "learning_rate": 3.130434782608696e-05, | |
| "loss": 0.269, | |
| "step": 1440 | |
| }, | |
| { | |
| "epoch": 15.002391304347826, | |
| "grad_norm": 73.08831024169922, | |
| "learning_rate": 3.152173913043479e-05, | |
| "loss": 0.149, | |
| "step": 1450 | |
| }, | |
| { | |
| "epoch": 15.00282608695652, | |
| "grad_norm": 0.11047495901584625, | |
| "learning_rate": 3.173913043478261e-05, | |
| "loss": 0.224, | |
| "step": 1460 | |
| }, | |
| { | |
| "epoch": 15.003260869565217, | |
| "grad_norm": 0.08704402297735214, | |
| "learning_rate": 3.195652173913043e-05, | |
| "loss": 0.2719, | |
| "step": 1470 | |
| }, | |
| { | |
| "epoch": 15.003695652173914, | |
| "grad_norm": 0.2108268141746521, | |
| "learning_rate": 3.217391304347826e-05, | |
| "loss": 0.0642, | |
| "step": 1480 | |
| }, | |
| { | |
| "epoch": 15.00404347826087, | |
| "eval_accuracy": 0.8278688524590164, | |
| "eval_f1": 0.8271381313811401, | |
| "eval_loss": 0.8966262936592102, | |
| "eval_runtime": 75.9595, | |
| "eval_samples_per_second": 1.606, | |
| "eval_steps_per_second": 0.211, | |
| "step": 1488 | |
| }, | |
| { | |
| "epoch": 16.000086956521738, | |
| "grad_norm": 0.016246311366558075, | |
| "learning_rate": 3.239130434782609e-05, | |
| "loss": 0.057, | |
| "step": 1490 | |
| }, | |
| { | |
| "epoch": 16.000521739130434, | |
| "grad_norm": 0.024972323328256607, | |
| "learning_rate": 3.260869565217392e-05, | |
| "loss": 0.0157, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 16.00095652173913, | |
| "grad_norm": 0.2207304686307907, | |
| "learning_rate": 3.282608695652174e-05, | |
| "loss": 0.0099, | |
| "step": 1510 | |
| }, | |
| { | |
| "epoch": 16.001391304347827, | |
| "grad_norm": 0.05993523448705673, | |
| "learning_rate": 3.304347826086956e-05, | |
| "loss": 0.2193, | |
| "step": 1520 | |
| }, | |
| { | |
| "epoch": 16.001826086956523, | |
| "grad_norm": 1.4549189805984497, | |
| "learning_rate": 3.3260869565217395e-05, | |
| "loss": 0.2023, | |
| "step": 1530 | |
| }, | |
| { | |
| "epoch": 16.002260869565216, | |
| "grad_norm": 0.024738606065511703, | |
| "learning_rate": 3.347826086956522e-05, | |
| "loss": 0.2535, | |
| "step": 1540 | |
| }, | |
| { | |
| "epoch": 16.002695652173912, | |
| "grad_norm": 43.513267517089844, | |
| "learning_rate": 3.369565217391305e-05, | |
| "loss": 0.1875, | |
| "step": 1550 | |
| }, | |
| { | |
| "epoch": 16.00313043478261, | |
| "grad_norm": 6.210598945617676, | |
| "learning_rate": 3.3913043478260867e-05, | |
| "loss": 0.0425, | |
| "step": 1560 | |
| }, | |
| { | |
| "epoch": 16.003565217391305, | |
| "grad_norm": 0.13376347720623016, | |
| "learning_rate": 3.413043478260869e-05, | |
| "loss": 0.1634, | |
| "step": 1570 | |
| }, | |
| { | |
| "epoch": 16.004, | |
| "grad_norm": 66.50971984863281, | |
| "learning_rate": 3.4347826086956526e-05, | |
| "loss": 0.1608, | |
| "step": 1580 | |
| }, | |
| { | |
| "epoch": 16.00404347826087, | |
| "eval_accuracy": 0.7950819672131147, | |
| "eval_f1": 0.7918367666938827, | |
| "eval_loss": 0.7916240096092224, | |
| "eval_runtime": 68.1076, | |
| "eval_samples_per_second": 1.791, | |
| "eval_steps_per_second": 0.235, | |
| "step": 1581 | |
| }, | |
| { | |
| "epoch": 17.000391304347826, | |
| "grad_norm": 52.1787223815918, | |
| "learning_rate": 3.456521739130435e-05, | |
| "loss": 0.0693, | |
| "step": 1590 | |
| }, | |
| { | |
| "epoch": 17.000826086956522, | |
| "grad_norm": 0.028341565281152725, | |
| "learning_rate": 3.478260869565218e-05, | |
| "loss": 0.168, | |
| "step": 1600 | |
| }, | |
| { | |
| "epoch": 17.00126086956522, | |
| "grad_norm": 0.06816719472408295, | |
| "learning_rate": 3.5e-05, | |
| "loss": 0.1966, | |
| "step": 1610 | |
| }, | |
| { | |
| "epoch": 17.001695652173915, | |
| "grad_norm": 0.06487961858510971, | |
| "learning_rate": 3.521739130434783e-05, | |
| "loss": 0.2377, | |
| "step": 1620 | |
| }, | |
| { | |
| "epoch": 17.002130434782607, | |
| "grad_norm": 1.7196918725967407, | |
| "learning_rate": 3.5434782608695656e-05, | |
| "loss": 0.0641, | |
| "step": 1630 | |
| }, | |
| { | |
| "epoch": 17.002565217391304, | |
| "grad_norm": 1.357792615890503, | |
| "learning_rate": 3.565217391304348e-05, | |
| "loss": 0.1772, | |
| "step": 1640 | |
| }, | |
| { | |
| "epoch": 17.003, | |
| "grad_norm": 48.746490478515625, | |
| "learning_rate": 3.58695652173913e-05, | |
| "loss": 0.2143, | |
| "step": 1650 | |
| }, | |
| { | |
| "epoch": 17.003434782608696, | |
| "grad_norm": 0.2187844216823578, | |
| "learning_rate": 3.6086956521739134e-05, | |
| "loss": 0.2618, | |
| "step": 1660 | |
| }, | |
| { | |
| "epoch": 17.003869565217393, | |
| "grad_norm": 27.896387100219727, | |
| "learning_rate": 3.630434782608696e-05, | |
| "loss": 0.0274, | |
| "step": 1670 | |
| }, | |
| { | |
| "epoch": 17.00404347826087, | |
| "eval_accuracy": 0.8360655737704918, | |
| "eval_f1": 0.8350123233242202, | |
| "eval_loss": 0.7110782265663147, | |
| "eval_runtime": 69.761, | |
| "eval_samples_per_second": 1.749, | |
| "eval_steps_per_second": 0.229, | |
| "step": 1674 | |
| }, | |
| { | |
| "epoch": 18.000260869565217, | |
| "grad_norm": 0.09103922545909882, | |
| "learning_rate": 3.6521739130434786e-05, | |
| "loss": 0.0427, | |
| "step": 1680 | |
| }, | |
| { | |
| "epoch": 18.000695652173913, | |
| "grad_norm": 0.02076614834368229, | |
| "learning_rate": 3.673913043478261e-05, | |
| "loss": 0.011, | |
| "step": 1690 | |
| }, | |
| { | |
| "epoch": 18.00113043478261, | |
| "grad_norm": 1.5786951780319214, | |
| "learning_rate": 3.695652173913043e-05, | |
| "loss": 0.0256, | |
| "step": 1700 | |
| }, | |
| { | |
| "epoch": 18.001565217391306, | |
| "grad_norm": 0.04359204322099686, | |
| "learning_rate": 3.7173913043478264e-05, | |
| "loss": 0.1148, | |
| "step": 1710 | |
| }, | |
| { | |
| "epoch": 18.002, | |
| "grad_norm": 0.9095045924186707, | |
| "learning_rate": 3.739130434782609e-05, | |
| "loss": 0.3361, | |
| "step": 1720 | |
| }, | |
| { | |
| "epoch": 18.002434782608695, | |
| "grad_norm": 69.62400817871094, | |
| "learning_rate": 3.7608695652173917e-05, | |
| "loss": 0.2205, | |
| "step": 1730 | |
| }, | |
| { | |
| "epoch": 18.00286956521739, | |
| "grad_norm": 30.217559814453125, | |
| "learning_rate": 3.7826086956521736e-05, | |
| "loss": 0.1105, | |
| "step": 1740 | |
| }, | |
| { | |
| "epoch": 18.003304347826088, | |
| "grad_norm": 0.47550705075263977, | |
| "learning_rate": 3.804347826086957e-05, | |
| "loss": 0.1594, | |
| "step": 1750 | |
| }, | |
| { | |
| "epoch": 18.003739130434784, | |
| "grad_norm": 68.50760650634766, | |
| "learning_rate": 3.8260869565217395e-05, | |
| "loss": 0.2955, | |
| "step": 1760 | |
| }, | |
| { | |
| "epoch": 18.00404347826087, | |
| "eval_accuracy": 0.7950819672131147, | |
| "eval_f1": 0.7968699869172604, | |
| "eval_loss": 0.9873971343040466, | |
| "eval_runtime": 68.4296, | |
| "eval_samples_per_second": 1.783, | |
| "eval_steps_per_second": 0.234, | |
| "step": 1767 | |
| }, | |
| { | |
| "epoch": 19.00013043478261, | |
| "grad_norm": 6.308517932891846, | |
| "learning_rate": 3.847826086956522e-05, | |
| "loss": 0.4724, | |
| "step": 1770 | |
| }, | |
| { | |
| "epoch": 19.000565217391305, | |
| "grad_norm": 30.987382888793945, | |
| "learning_rate": 3.869565217391305e-05, | |
| "loss": 0.2195, | |
| "step": 1780 | |
| }, | |
| { | |
| "epoch": 19.001, | |
| "grad_norm": 0.02650444395840168, | |
| "learning_rate": 3.8913043478260866e-05, | |
| "loss": 0.003, | |
| "step": 1790 | |
| }, | |
| { | |
| "epoch": 19.001434782608694, | |
| "grad_norm": 0.2237127423286438, | |
| "learning_rate": 3.91304347826087e-05, | |
| "loss": 0.2009, | |
| "step": 1800 | |
| }, | |
| { | |
| "epoch": 19.00186956521739, | |
| "grad_norm": 0.05298967659473419, | |
| "learning_rate": 3.9347826086956525e-05, | |
| "loss": 0.1038, | |
| "step": 1810 | |
| }, | |
| { | |
| "epoch": 19.002304347826087, | |
| "grad_norm": 62.364654541015625, | |
| "learning_rate": 3.956521739130435e-05, | |
| "loss": 0.0839, | |
| "step": 1820 | |
| }, | |
| { | |
| "epoch": 19.002739130434783, | |
| "grad_norm": 0.050453417003154755, | |
| "learning_rate": 3.978260869565217e-05, | |
| "loss": 0.0338, | |
| "step": 1830 | |
| }, | |
| { | |
| "epoch": 19.00317391304348, | |
| "grad_norm": 31.080060958862305, | |
| "learning_rate": 4e-05, | |
| "loss": 0.0454, | |
| "step": 1840 | |
| }, | |
| { | |
| "epoch": 19.003608695652172, | |
| "grad_norm": 62.11554718017578, | |
| "learning_rate": 4.021739130434783e-05, | |
| "loss": 0.1027, | |
| "step": 1850 | |
| }, | |
| { | |
| "epoch": 19.00404347826087, | |
| "grad_norm": 0.8470473885536194, | |
| "learning_rate": 4.0434782608695655e-05, | |
| "loss": 0.3371, | |
| "step": 1860 | |
| }, | |
| { | |
| "epoch": 19.00404347826087, | |
| "eval_accuracy": 0.7377049180327869, | |
| "eval_f1": 0.725067323090087, | |
| "eval_loss": 1.4002968072891235, | |
| "eval_runtime": 82.0023, | |
| "eval_samples_per_second": 1.488, | |
| "eval_steps_per_second": 0.195, | |
| "step": 1860 | |
| }, | |
| { | |
| "epoch": 20.000434782608696, | |
| "grad_norm": 0.7842612862586975, | |
| "learning_rate": 4.065217391304348e-05, | |
| "loss": 0.2897, | |
| "step": 1870 | |
| }, | |
| { | |
| "epoch": 20.000869565217393, | |
| "grad_norm": 45.9306526184082, | |
| "learning_rate": 4.086956521739131e-05, | |
| "loss": 0.4222, | |
| "step": 1880 | |
| }, | |
| { | |
| "epoch": 20.001304347826085, | |
| "grad_norm": 0.01409738790243864, | |
| "learning_rate": 4.1086956521739134e-05, | |
| "loss": 0.2394, | |
| "step": 1890 | |
| }, | |
| { | |
| "epoch": 20.001739130434782, | |
| "grad_norm": 0.834911048412323, | |
| "learning_rate": 4.130434782608696e-05, | |
| "loss": 0.0976, | |
| "step": 1900 | |
| }, | |
| { | |
| "epoch": 20.002173913043478, | |
| "grad_norm": 1.8054020404815674, | |
| "learning_rate": 4.1521739130434786e-05, | |
| "loss": 0.1883, | |
| "step": 1910 | |
| }, | |
| { | |
| "epoch": 20.002608695652174, | |
| "grad_norm": 0.029708171263337135, | |
| "learning_rate": 4.1739130434782605e-05, | |
| "loss": 0.2925, | |
| "step": 1920 | |
| }, | |
| { | |
| "epoch": 20.00304347826087, | |
| "grad_norm": 0.054145630449056625, | |
| "learning_rate": 4.195652173913044e-05, | |
| "loss": 0.4088, | |
| "step": 1930 | |
| }, | |
| { | |
| "epoch": 20.003478260869564, | |
| "grad_norm": 12.527750015258789, | |
| "learning_rate": 4.2173913043478264e-05, | |
| "loss": 0.3098, | |
| "step": 1940 | |
| }, | |
| { | |
| "epoch": 20.00391304347826, | |
| "grad_norm": 0.0675775408744812, | |
| "learning_rate": 4.239130434782609e-05, | |
| "loss": 0.1977, | |
| "step": 1950 | |
| }, | |
| { | |
| "epoch": 20.00404347826087, | |
| "eval_accuracy": 0.8278688524590164, | |
| "eval_f1": 0.8285237447031244, | |
| "eval_loss": 0.729884147644043, | |
| "eval_runtime": 78.9289, | |
| "eval_samples_per_second": 1.546, | |
| "eval_steps_per_second": 0.203, | |
| "step": 1953 | |
| }, | |
| { | |
| "epoch": 21.000304347826088, | |
| "grad_norm": 0.1842467486858368, | |
| "learning_rate": 4.2608695652173916e-05, | |
| "loss": 0.0141, | |
| "step": 1960 | |
| }, | |
| { | |
| "epoch": 21.000739130434784, | |
| "grad_norm": 35.704742431640625, | |
| "learning_rate": 4.282608695652174e-05, | |
| "loss": 0.1253, | |
| "step": 1970 | |
| }, | |
| { | |
| "epoch": 21.001173913043477, | |
| "grad_norm": 25.044652938842773, | |
| "learning_rate": 4.304347826086957e-05, | |
| "loss": 0.2175, | |
| "step": 1980 | |
| }, | |
| { | |
| "epoch": 21.001608695652173, | |
| "grad_norm": 9.421120643615723, | |
| "learning_rate": 4.3260869565217394e-05, | |
| "loss": 0.0545, | |
| "step": 1990 | |
| }, | |
| { | |
| "epoch": 21.00204347826087, | |
| "grad_norm": 0.5154250264167786, | |
| "learning_rate": 4.347826086956522e-05, | |
| "loss": 0.3332, | |
| "step": 2000 | |
| }, | |
| { | |
| "epoch": 21.002478260869566, | |
| "grad_norm": 0.18789710104465485, | |
| "learning_rate": 4.3695652173913046e-05, | |
| "loss": 0.4334, | |
| "step": 2010 | |
| }, | |
| { | |
| "epoch": 21.002913043478262, | |
| "grad_norm": 0.4829055666923523, | |
| "learning_rate": 4.391304347826087e-05, | |
| "loss": 0.2281, | |
| "step": 2020 | |
| }, | |
| { | |
| "epoch": 21.003347826086955, | |
| "grad_norm": 35.912315368652344, | |
| "learning_rate": 4.41304347826087e-05, | |
| "loss": 0.3521, | |
| "step": 2030 | |
| }, | |
| { | |
| "epoch": 21.00378260869565, | |
| "grad_norm": 0.1704353392124176, | |
| "learning_rate": 4.4347826086956525e-05, | |
| "loss": 0.1155, | |
| "step": 2040 | |
| }, | |
| { | |
| "epoch": 21.00404347826087, | |
| "eval_accuracy": 0.8524590163934426, | |
| "eval_f1": 0.8524527538533494, | |
| "eval_loss": 0.7051700353622437, | |
| "eval_runtime": 79.7194, | |
| "eval_samples_per_second": 1.53, | |
| "eval_steps_per_second": 0.201, | |
| "step": 2046 | |
| }, | |
| { | |
| "epoch": 22.00017391304348, | |
| "grad_norm": 0.10614888370037079, | |
| "learning_rate": 4.456521739130435e-05, | |
| "loss": 0.1375, | |
| "step": 2050 | |
| }, | |
| { | |
| "epoch": 22.000608695652176, | |
| "grad_norm": 0.028057333081960678, | |
| "learning_rate": 4.478260869565218e-05, | |
| "loss": 0.0373, | |
| "step": 2060 | |
| }, | |
| { | |
| "epoch": 22.00104347826087, | |
| "grad_norm": 81.05805206298828, | |
| "learning_rate": 4.5e-05, | |
| "loss": 0.0963, | |
| "step": 2070 | |
| }, | |
| { | |
| "epoch": 22.001478260869565, | |
| "grad_norm": 15.839224815368652, | |
| "learning_rate": 4.521739130434783e-05, | |
| "loss": 0.1109, | |
| "step": 2080 | |
| }, | |
| { | |
| "epoch": 22.00191304347826, | |
| "grad_norm": 17.285728454589844, | |
| "learning_rate": 4.5434782608695655e-05, | |
| "loss": 0.086, | |
| "step": 2090 | |
| }, | |
| { | |
| "epoch": 22.002347826086957, | |
| "grad_norm": 0.011755176819860935, | |
| "learning_rate": 4.565217391304348e-05, | |
| "loss": 0.005, | |
| "step": 2100 | |
| }, | |
| { | |
| "epoch": 22.002782608695654, | |
| "grad_norm": 0.04565587639808655, | |
| "learning_rate": 4.586956521739131e-05, | |
| "loss": 0.2197, | |
| "step": 2110 | |
| }, | |
| { | |
| "epoch": 22.003217391304347, | |
| "grad_norm": 0.1595648229122162, | |
| "learning_rate": 4.608695652173913e-05, | |
| "loss": 0.2304, | |
| "step": 2120 | |
| }, | |
| { | |
| "epoch": 22.003652173913043, | |
| "grad_norm": 0.09167420864105225, | |
| "learning_rate": 4.630434782608696e-05, | |
| "loss": 0.0134, | |
| "step": 2130 | |
| }, | |
| { | |
| "epoch": 22.00404347826087, | |
| "eval_accuracy": 0.8770491803278688, | |
| "eval_f1": 0.8773614363778297, | |
| "eval_loss": 0.7146317958831787, | |
| "eval_runtime": 78.4014, | |
| "eval_samples_per_second": 1.556, | |
| "eval_steps_per_second": 0.204, | |
| "step": 2139 | |
| }, | |
| { | |
| "epoch": 23.00004347826087, | |
| "grad_norm": 0.05441008135676384, | |
| "learning_rate": 4.6521739130434785e-05, | |
| "loss": 0.1697, | |
| "step": 2140 | |
| }, | |
| { | |
| "epoch": 23.000478260869563, | |
| "grad_norm": 63.973304748535156, | |
| "learning_rate": 4.673913043478261e-05, | |
| "loss": 0.3334, | |
| "step": 2150 | |
| }, | |
| { | |
| "epoch": 23.00091304347826, | |
| "grad_norm": 27.009883880615234, | |
| "learning_rate": 4.695652173913044e-05, | |
| "loss": 0.1678, | |
| "step": 2160 | |
| }, | |
| { | |
| "epoch": 23.001347826086956, | |
| "grad_norm": 0.037991661578416824, | |
| "learning_rate": 4.7173913043478264e-05, | |
| "loss": 0.1246, | |
| "step": 2170 | |
| }, | |
| { | |
| "epoch": 23.001782608695652, | |
| "grad_norm": 0.040783315896987915, | |
| "learning_rate": 4.739130434782609e-05, | |
| "loss": 0.0249, | |
| "step": 2180 | |
| }, | |
| { | |
| "epoch": 23.00221739130435, | |
| "grad_norm": 18.87450408935547, | |
| "learning_rate": 4.7608695652173916e-05, | |
| "loss": 0.1833, | |
| "step": 2190 | |
| }, | |
| { | |
| "epoch": 23.002652173913045, | |
| "grad_norm": 0.08050432801246643, | |
| "learning_rate": 4.782608695652174e-05, | |
| "loss": 0.2014, | |
| "step": 2200 | |
| }, | |
| { | |
| "epoch": 23.003086956521738, | |
| "grad_norm": 0.021646007895469666, | |
| "learning_rate": 4.804347826086957e-05, | |
| "loss": 0.1677, | |
| "step": 2210 | |
| }, | |
| { | |
| "epoch": 23.003521739130434, | |
| "grad_norm": 33.0234375, | |
| "learning_rate": 4.8260869565217394e-05, | |
| "loss": 0.2297, | |
| "step": 2220 | |
| }, | |
| { | |
| "epoch": 23.00395652173913, | |
| "grad_norm": 0.19216619431972504, | |
| "learning_rate": 4.847826086956522e-05, | |
| "loss": 0.1011, | |
| "step": 2230 | |
| }, | |
| { | |
| "epoch": 23.00404347826087, | |
| "eval_accuracy": 0.8278688524590164, | |
| "eval_f1": 0.8276948820520866, | |
| "eval_loss": 0.8542459607124329, | |
| "eval_runtime": 70.3843, | |
| "eval_samples_per_second": 1.733, | |
| "eval_steps_per_second": 0.227, | |
| "step": 2232 | |
| }, | |
| { | |
| "epoch": 24.000347826086955, | |
| "grad_norm": 19.964126586914062, | |
| "learning_rate": 4.8695652173913046e-05, | |
| "loss": 0.395, | |
| "step": 2240 | |
| }, | |
| { | |
| "epoch": 24.00078260869565, | |
| "grad_norm": 1.8568761348724365, | |
| "learning_rate": 4.891304347826087e-05, | |
| "loss": 0.2026, | |
| "step": 2250 | |
| }, | |
| { | |
| "epoch": 24.001217391304348, | |
| "grad_norm": 29.76133918762207, | |
| "learning_rate": 4.91304347826087e-05, | |
| "loss": 0.1282, | |
| "step": 2260 | |
| }, | |
| { | |
| "epoch": 24.001652173913044, | |
| "grad_norm": 0.4110662043094635, | |
| "learning_rate": 4.9347826086956524e-05, | |
| "loss": 0.0744, | |
| "step": 2270 | |
| }, | |
| { | |
| "epoch": 24.00208695652174, | |
| "grad_norm": 0.05301794409751892, | |
| "learning_rate": 4.956521739130435e-05, | |
| "loss": 0.1104, | |
| "step": 2280 | |
| }, | |
| { | |
| "epoch": 24.002521739130433, | |
| "grad_norm": 100.25152587890625, | |
| "learning_rate": 4.9782608695652176e-05, | |
| "loss": 0.0886, | |
| "step": 2290 | |
| }, | |
| { | |
| "epoch": 24.00295652173913, | |
| "grad_norm": 4.144306182861328, | |
| "learning_rate": 5e-05, | |
| "loss": 0.4572, | |
| "step": 2300 | |
| }, | |
| { | |
| "epoch": 24.003391304347826, | |
| "grad_norm": 84.53608703613281, | |
| "learning_rate": 4.997584541062802e-05, | |
| "loss": 0.4282, | |
| "step": 2310 | |
| }, | |
| { | |
| "epoch": 24.003826086956522, | |
| "grad_norm": 1.2512717247009277, | |
| "learning_rate": 4.9951690821256045e-05, | |
| "loss": 0.0834, | |
| "step": 2320 | |
| }, | |
| { | |
| "epoch": 24.00404347826087, | |
| "eval_accuracy": 0.7131147540983607, | |
| "eval_f1": 0.71432835887995, | |
| "eval_loss": 1.6143077611923218, | |
| "eval_runtime": 78.2242, | |
| "eval_samples_per_second": 1.56, | |
| "eval_steps_per_second": 0.205, | |
| "step": 2325 | |
| }, | |
| { | |
| "epoch": 25.000217391304346, | |
| "grad_norm": 0.014534312300384045, | |
| "learning_rate": 4.9927536231884056e-05, | |
| "loss": 0.0735, | |
| "step": 2330 | |
| }, | |
| { | |
| "epoch": 25.000652173913043, | |
| "grad_norm": 2.242990493774414, | |
| "learning_rate": 4.990338164251208e-05, | |
| "loss": 0.0942, | |
| "step": 2340 | |
| }, | |
| { | |
| "epoch": 25.00108695652174, | |
| "grad_norm": 0.02222440391778946, | |
| "learning_rate": 4.98792270531401e-05, | |
| "loss": 0.0775, | |
| "step": 2350 | |
| }, | |
| { | |
| "epoch": 25.001521739130435, | |
| "grad_norm": 0.014810199849307537, | |
| "learning_rate": 4.985507246376812e-05, | |
| "loss": 0.1185, | |
| "step": 2360 | |
| }, | |
| { | |
| "epoch": 25.00195652173913, | |
| "grad_norm": 53.17665100097656, | |
| "learning_rate": 4.983091787439614e-05, | |
| "loss": 0.5145, | |
| "step": 2370 | |
| }, | |
| { | |
| "epoch": 25.002391304347825, | |
| "grad_norm": 0.023323476314544678, | |
| "learning_rate": 4.980676328502415e-05, | |
| "loss": 0.0293, | |
| "step": 2380 | |
| }, | |
| { | |
| "epoch": 25.00282608695652, | |
| "grad_norm": 1.3914483785629272, | |
| "learning_rate": 4.9782608695652176e-05, | |
| "loss": 0.1876, | |
| "step": 2390 | |
| }, | |
| { | |
| "epoch": 25.003260869565217, | |
| "grad_norm": 0.20789887011051178, | |
| "learning_rate": 4.9758454106280194e-05, | |
| "loss": 0.1446, | |
| "step": 2400 | |
| }, | |
| { | |
| "epoch": 25.003695652173914, | |
| "grad_norm": 6.419387340545654, | |
| "learning_rate": 4.973429951690822e-05, | |
| "loss": 0.3776, | |
| "step": 2410 | |
| }, | |
| { | |
| "epoch": 25.00404347826087, | |
| "eval_accuracy": 0.7868852459016393, | |
| "eval_f1": 0.7833783048095553, | |
| "eval_loss": 1.0066475868225098, | |
| "eval_runtime": 75.7936, | |
| "eval_samples_per_second": 1.61, | |
| "eval_steps_per_second": 0.211, | |
| "step": 2418 | |
| }, | |
| { | |
| "epoch": 26.000086956521738, | |
| "grad_norm": 0.11995876580476761, | |
| "learning_rate": 4.9710144927536237e-05, | |
| "loss": 0.1378, | |
| "step": 2420 | |
| }, | |
| { | |
| "epoch": 26.000521739130434, | |
| "grad_norm": 0.10857793688774109, | |
| "learning_rate": 4.9685990338164254e-05, | |
| "loss": 0.0582, | |
| "step": 2430 | |
| }, | |
| { | |
| "epoch": 26.00095652173913, | |
| "grad_norm": 3.5968430042266846, | |
| "learning_rate": 4.966183574879227e-05, | |
| "loss": 0.0596, | |
| "step": 2440 | |
| }, | |
| { | |
| "epoch": 26.001391304347827, | |
| "grad_norm": 0.012462865561246872, | |
| "learning_rate": 4.963768115942029e-05, | |
| "loss": 0.036, | |
| "step": 2450 | |
| }, | |
| { | |
| "epoch": 26.001826086956523, | |
| "grad_norm": 88.11896514892578, | |
| "learning_rate": 4.9613526570048315e-05, | |
| "loss": 0.1124, | |
| "step": 2460 | |
| }, | |
| { | |
| "epoch": 26.002260869565216, | |
| "grad_norm": 0.010418935678899288, | |
| "learning_rate": 4.958937198067633e-05, | |
| "loss": 0.2141, | |
| "step": 2470 | |
| }, | |
| { | |
| "epoch": 26.002695652173912, | |
| "grad_norm": 0.8756917119026184, | |
| "learning_rate": 4.956521739130435e-05, | |
| "loss": 0.0481, | |
| "step": 2480 | |
| }, | |
| { | |
| "epoch": 26.00313043478261, | |
| "grad_norm": 0.014706839807331562, | |
| "learning_rate": 4.954106280193237e-05, | |
| "loss": 0.0433, | |
| "step": 2490 | |
| }, | |
| { | |
| "epoch": 26.003565217391305, | |
| "grad_norm": 33.63422393798828, | |
| "learning_rate": 4.9516908212560386e-05, | |
| "loss": 0.1711, | |
| "step": 2500 | |
| }, | |
| { | |
| "epoch": 26.004, | |
| "grad_norm": 8.496404647827148, | |
| "learning_rate": 4.949275362318841e-05, | |
| "loss": 0.0745, | |
| "step": 2510 | |
| }, | |
| { | |
| "epoch": 26.00404347826087, | |
| "eval_accuracy": 0.8442622950819673, | |
| "eval_f1": 0.8448557228692234, | |
| "eval_loss": 0.794857382774353, | |
| "eval_runtime": 80.5247, | |
| "eval_samples_per_second": 1.515, | |
| "eval_steps_per_second": 0.199, | |
| "step": 2511 | |
| }, | |
| { | |
| "epoch": 27.000391304347826, | |
| "grad_norm": 30.50170135498047, | |
| "learning_rate": 4.946859903381643e-05, | |
| "loss": 0.0833, | |
| "step": 2520 | |
| }, | |
| { | |
| "epoch": 27.000826086956522, | |
| "grad_norm": 0.010464085265994072, | |
| "learning_rate": 4.9444444444444446e-05, | |
| "loss": 0.0171, | |
| "step": 2530 | |
| }, | |
| { | |
| "epoch": 27.00126086956522, | |
| "grad_norm": 0.013533636927604675, | |
| "learning_rate": 4.9420289855072464e-05, | |
| "loss": 0.1177, | |
| "step": 2540 | |
| }, | |
| { | |
| "epoch": 27.001695652173915, | |
| "grad_norm": 0.06204245612025261, | |
| "learning_rate": 4.939613526570048e-05, | |
| "loss": 0.0623, | |
| "step": 2550 | |
| }, | |
| { | |
| "epoch": 27.002130434782607, | |
| "grad_norm": 0.007505207788199186, | |
| "learning_rate": 4.9371980676328506e-05, | |
| "loss": 0.0204, | |
| "step": 2560 | |
| }, | |
| { | |
| "epoch": 27.002565217391304, | |
| "grad_norm": 0.03377118706703186, | |
| "learning_rate": 4.9347826086956524e-05, | |
| "loss": 0.0159, | |
| "step": 2570 | |
| }, | |
| { | |
| "epoch": 27.003, | |
| "grad_norm": 0.009533570148050785, | |
| "learning_rate": 4.932367149758454e-05, | |
| "loss": 0.2513, | |
| "step": 2580 | |
| }, | |
| { | |
| "epoch": 27.003434782608696, | |
| "grad_norm": 0.007220500614494085, | |
| "learning_rate": 4.929951690821256e-05, | |
| "loss": 0.0012, | |
| "step": 2590 | |
| }, | |
| { | |
| "epoch": 27.003869565217393, | |
| "grad_norm": 0.008370584808290005, | |
| "learning_rate": 4.9275362318840584e-05, | |
| "loss": 0.0012, | |
| "step": 2600 | |
| }, | |
| { | |
| "epoch": 27.00404347826087, | |
| "eval_accuracy": 0.8524590163934426, | |
| "eval_f1": 0.8519739016181949, | |
| "eval_loss": 0.8901496529579163, | |
| "eval_runtime": 78.3188, | |
| "eval_samples_per_second": 1.558, | |
| "eval_steps_per_second": 0.204, | |
| "step": 2604 | |
| }, | |
| { | |
| "epoch": 28.000260869565217, | |
| "grad_norm": 10.66296672821045, | |
| "learning_rate": 4.92512077294686e-05, | |
| "loss": 0.0684, | |
| "step": 2610 | |
| }, | |
| { | |
| "epoch": 28.000695652173913, | |
| "grad_norm": 0.016988499090075493, | |
| "learning_rate": 4.922705314009662e-05, | |
| "loss": 0.0504, | |
| "step": 2620 | |
| }, | |
| { | |
| "epoch": 28.00113043478261, | |
| "grad_norm": 11.786551475524902, | |
| "learning_rate": 4.920289855072464e-05, | |
| "loss": 0.1493, | |
| "step": 2630 | |
| }, | |
| { | |
| "epoch": 28.001565217391306, | |
| "grad_norm": 0.027907541021704674, | |
| "learning_rate": 4.9178743961352656e-05, | |
| "loss": 0.0218, | |
| "step": 2640 | |
| }, | |
| { | |
| "epoch": 28.002, | |
| "grad_norm": 0.006081975065171719, | |
| "learning_rate": 4.915458937198068e-05, | |
| "loss": 0.0492, | |
| "step": 2650 | |
| }, | |
| { | |
| "epoch": 28.002434782608695, | |
| "grad_norm": 0.011573576368391514, | |
| "learning_rate": 4.91304347826087e-05, | |
| "loss": 0.0006, | |
| "step": 2660 | |
| }, | |
| { | |
| "epoch": 28.00286956521739, | |
| "grad_norm": 6.206803321838379, | |
| "learning_rate": 4.910628019323672e-05, | |
| "loss": 0.1155, | |
| "step": 2670 | |
| }, | |
| { | |
| "epoch": 28.003304347826088, | |
| "grad_norm": 0.031189128756523132, | |
| "learning_rate": 4.9082125603864734e-05, | |
| "loss": 0.0613, | |
| "step": 2680 | |
| }, | |
| { | |
| "epoch": 28.003739130434784, | |
| "grad_norm": 0.06879521161317825, | |
| "learning_rate": 4.905797101449275e-05, | |
| "loss": 0.0433, | |
| "step": 2690 | |
| }, | |
| { | |
| "epoch": 28.00404347826087, | |
| "eval_accuracy": 0.8360655737704918, | |
| "eval_f1": 0.8360363588217401, | |
| "eval_loss": 0.9370259642601013, | |
| "eval_runtime": 75.9173, | |
| "eval_samples_per_second": 1.607, | |
| "eval_steps_per_second": 0.211, | |
| "step": 2697 | |
| }, | |
| { | |
| "epoch": 29.00013043478261, | |
| "grad_norm": 0.008160380646586418, | |
| "learning_rate": 4.9033816425120776e-05, | |
| "loss": 0.0016, | |
| "step": 2700 | |
| }, | |
| { | |
| "epoch": 29.000565217391305, | |
| "grad_norm": 0.0150864627212286, | |
| "learning_rate": 4.9009661835748794e-05, | |
| "loss": 0.0037, | |
| "step": 2710 | |
| }, | |
| { | |
| "epoch": 29.001, | |
| "grad_norm": 6.826006889343262, | |
| "learning_rate": 4.898550724637682e-05, | |
| "loss": 0.0263, | |
| "step": 2720 | |
| }, | |
| { | |
| "epoch": 29.001434782608694, | |
| "grad_norm": 0.010020852088928223, | |
| "learning_rate": 4.8961352657004836e-05, | |
| "loss": 0.0041, | |
| "step": 2730 | |
| }, | |
| { | |
| "epoch": 29.00186956521739, | |
| "grad_norm": 0.0070611899718642235, | |
| "learning_rate": 4.893719806763285e-05, | |
| "loss": 0.058, | |
| "step": 2740 | |
| }, | |
| { | |
| "epoch": 29.002304347826087, | |
| "grad_norm": 0.006664105225354433, | |
| "learning_rate": 4.891304347826087e-05, | |
| "loss": 0.0311, | |
| "step": 2750 | |
| }, | |
| { | |
| "epoch": 29.002739130434783, | |
| "grad_norm": 0.011251281015574932, | |
| "learning_rate": 4.888888888888889e-05, | |
| "loss": 0.0891, | |
| "step": 2760 | |
| }, | |
| { | |
| "epoch": 29.00317391304348, | |
| "grad_norm": 0.0483035147190094, | |
| "learning_rate": 4.8864734299516914e-05, | |
| "loss": 0.0673, | |
| "step": 2770 | |
| }, | |
| { | |
| "epoch": 29.003608695652172, | |
| "grad_norm": 0.01135986763983965, | |
| "learning_rate": 4.884057971014493e-05, | |
| "loss": 0.1729, | |
| "step": 2780 | |
| }, | |
| { | |
| "epoch": 29.00404347826087, | |
| "grad_norm": 0.04626130685210228, | |
| "learning_rate": 4.881642512077295e-05, | |
| "loss": 0.0708, | |
| "step": 2790 | |
| }, | |
| { | |
| "epoch": 29.00404347826087, | |
| "eval_accuracy": 0.8114754098360656, | |
| "eval_f1": 0.8062107421774495, | |
| "eval_loss": 1.0806177854537964, | |
| "eval_runtime": 75.2291, | |
| "eval_samples_per_second": 1.622, | |
| "eval_steps_per_second": 0.213, | |
| "step": 2790 | |
| }, | |
| { | |
| "epoch": 30.000434782608696, | |
| "grad_norm": 0.009849685244262218, | |
| "learning_rate": 4.879227053140097e-05, | |
| "loss": 0.0661, | |
| "step": 2800 | |
| }, | |
| { | |
| "epoch": 30.000869565217393, | |
| "grad_norm": 0.009964972734451294, | |
| "learning_rate": 4.8768115942028986e-05, | |
| "loss": 0.1059, | |
| "step": 2810 | |
| }, | |
| { | |
| "epoch": 30.001304347826085, | |
| "grad_norm": 0.016774578019976616, | |
| "learning_rate": 4.874396135265701e-05, | |
| "loss": 0.0907, | |
| "step": 2820 | |
| }, | |
| { | |
| "epoch": 30.001739130434782, | |
| "grad_norm": 0.040092673152685165, | |
| "learning_rate": 4.871980676328503e-05, | |
| "loss": 0.0298, | |
| "step": 2830 | |
| }, | |
| { | |
| "epoch": 30.002173913043478, | |
| "grad_norm": 0.12641261518001556, | |
| "learning_rate": 4.8695652173913046e-05, | |
| "loss": 0.0006, | |
| "step": 2840 | |
| }, | |
| { | |
| "epoch": 30.002608695652174, | |
| "grad_norm": 0.009654521942138672, | |
| "learning_rate": 4.8671497584541064e-05, | |
| "loss": 0.0342, | |
| "step": 2850 | |
| }, | |
| { | |
| "epoch": 30.00304347826087, | |
| "grad_norm": 5.2498040199279785, | |
| "learning_rate": 4.864734299516908e-05, | |
| "loss": 0.249, | |
| "step": 2860 | |
| }, | |
| { | |
| "epoch": 30.003478260869564, | |
| "grad_norm": 0.485901802778244, | |
| "learning_rate": 4.8623188405797106e-05, | |
| "loss": 0.0121, | |
| "step": 2870 | |
| }, | |
| { | |
| "epoch": 30.00391304347826, | |
| "grad_norm": 0.030848659574985504, | |
| "learning_rate": 4.8599033816425124e-05, | |
| "loss": 0.0335, | |
| "step": 2880 | |
| }, | |
| { | |
| "epoch": 30.00404347826087, | |
| "eval_accuracy": 0.819672131147541, | |
| "eval_f1": 0.8112946016110332, | |
| "eval_loss": 1.1136860847473145, | |
| "eval_runtime": 74.2045, | |
| "eval_samples_per_second": 1.644, | |
| "eval_steps_per_second": 0.216, | |
| "step": 2883 | |
| }, | |
| { | |
| "epoch": 31.000304347826088, | |
| "grad_norm": 2.0275466442108154, | |
| "learning_rate": 4.857487922705314e-05, | |
| "loss": 0.0145, | |
| "step": 2890 | |
| }, | |
| { | |
| "epoch": 31.000739130434784, | |
| "grad_norm": 10.900162696838379, | |
| "learning_rate": 4.855072463768116e-05, | |
| "loss": 0.0999, | |
| "step": 2900 | |
| }, | |
| { | |
| "epoch": 31.001173913043477, | |
| "grad_norm": 0.12738248705863953, | |
| "learning_rate": 4.8526570048309184e-05, | |
| "loss": 0.0026, | |
| "step": 2910 | |
| }, | |
| { | |
| "epoch": 31.001608695652173, | |
| "grad_norm": 29.855815887451172, | |
| "learning_rate": 4.85024154589372e-05, | |
| "loss": 0.2289, | |
| "step": 2920 | |
| }, | |
| { | |
| "epoch": 31.00204347826087, | |
| "grad_norm": 0.00635722279548645, | |
| "learning_rate": 4.847826086956522e-05, | |
| "loss": 0.0039, | |
| "step": 2930 | |
| }, | |
| { | |
| "epoch": 31.002478260869566, | |
| "grad_norm": 0.007796896155923605, | |
| "learning_rate": 4.845410628019324e-05, | |
| "loss": 0.1027, | |
| "step": 2940 | |
| }, | |
| { | |
| "epoch": 31.002913043478262, | |
| "grad_norm": 0.011087978258728981, | |
| "learning_rate": 4.8429951690821256e-05, | |
| "loss": 0.1863, | |
| "step": 2950 | |
| }, | |
| { | |
| "epoch": 31.003347826086955, | |
| "grad_norm": 0.16981664299964905, | |
| "learning_rate": 4.840579710144928e-05, | |
| "loss": 0.1017, | |
| "step": 2960 | |
| }, | |
| { | |
| "epoch": 31.00378260869565, | |
| "grad_norm": 0.005533864255994558, | |
| "learning_rate": 4.83816425120773e-05, | |
| "loss": 0.0571, | |
| "step": 2970 | |
| }, | |
| { | |
| "epoch": 31.00404347826087, | |
| "eval_accuracy": 0.7540983606557377, | |
| "eval_f1": 0.7606644114840837, | |
| "eval_loss": 1.6491549015045166, | |
| "eval_runtime": 72.4132, | |
| "eval_samples_per_second": 1.685, | |
| "eval_steps_per_second": 0.221, | |
| "step": 2976 | |
| }, | |
| { | |
| "epoch": 32.000173913043476, | |
| "grad_norm": 4.148243427276611, | |
| "learning_rate": 4.8357487922705316e-05, | |
| "loss": 0.2991, | |
| "step": 2980 | |
| }, | |
| { | |
| "epoch": 32.000608695652176, | |
| "grad_norm": 0.16210371255874634, | |
| "learning_rate": 4.8333333333333334e-05, | |
| "loss": 0.0832, | |
| "step": 2990 | |
| }, | |
| { | |
| "epoch": 32.00104347826087, | |
| "grad_norm": 0.11444944888353348, | |
| "learning_rate": 4.830917874396135e-05, | |
| "loss": 0.131, | |
| "step": 3000 | |
| }, | |
| { | |
| "epoch": 32.00147826086957, | |
| "grad_norm": 20.465253829956055, | |
| "learning_rate": 4.8285024154589376e-05, | |
| "loss": 0.0559, | |
| "step": 3010 | |
| }, | |
| { | |
| "epoch": 32.00191304347826, | |
| "grad_norm": 0.008609439246356487, | |
| "learning_rate": 4.8260869565217394e-05, | |
| "loss": 0.1028, | |
| "step": 3020 | |
| }, | |
| { | |
| "epoch": 32.002347826086954, | |
| "grad_norm": 0.005297601222991943, | |
| "learning_rate": 4.823671497584542e-05, | |
| "loss": 0.0911, | |
| "step": 3030 | |
| }, | |
| { | |
| "epoch": 32.002782608695654, | |
| "grad_norm": 0.009243907406926155, | |
| "learning_rate": 4.821256038647343e-05, | |
| "loss": 0.4324, | |
| "step": 3040 | |
| }, | |
| { | |
| "epoch": 32.00321739130435, | |
| "grad_norm": 0.07699494808912277, | |
| "learning_rate": 4.818840579710145e-05, | |
| "loss": 0.1528, | |
| "step": 3050 | |
| }, | |
| { | |
| "epoch": 32.003652173913046, | |
| "grad_norm": 0.5034440159797668, | |
| "learning_rate": 4.816425120772947e-05, | |
| "loss": 0.0965, | |
| "step": 3060 | |
| }, | |
| { | |
| "epoch": 32.00404347826087, | |
| "eval_accuracy": 0.8770491803278688, | |
| "eval_f1": 0.877946484907584, | |
| "eval_loss": 0.7584701180458069, | |
| "eval_runtime": 68.4338, | |
| "eval_samples_per_second": 1.783, | |
| "eval_steps_per_second": 0.234, | |
| "step": 3069 | |
| }, | |
| { | |
| "epoch": 4.347826086956522e-05, | |
| "grad_norm": 0.03973979130387306, | |
| "learning_rate": 4.814009661835749e-05, | |
| "loss": 0.0007, | |
| "step": 3070 | |
| }, | |
| { | |
| "epoch": 0.0004782608695652174, | |
| "grad_norm": 0.011583452112972736, | |
| "learning_rate": 4.8115942028985514e-05, | |
| "loss": 0.0774, | |
| "step": 3080 | |
| }, | |
| { | |
| "epoch": 0.0009130434782608696, | |
| "grad_norm": 0.13716067373752594, | |
| "learning_rate": 4.8091787439613525e-05, | |
| "loss": 0.0874, | |
| "step": 3090 | |
| }, | |
| { | |
| "epoch": 0.0013478260869565217, | |
| "grad_norm": 52.338008880615234, | |
| "learning_rate": 4.806763285024155e-05, | |
| "loss": 0.0513, | |
| "step": 3100 | |
| }, | |
| { | |
| "epoch": 0.0017826086956521739, | |
| "grad_norm": 14.776463508605957, | |
| "learning_rate": 4.804347826086957e-05, | |
| "loss": 0.1144, | |
| "step": 3110 | |
| }, | |
| { | |
| "epoch": 0.002217391304347826, | |
| "grad_norm": 0.02545876055955887, | |
| "learning_rate": 4.8019323671497586e-05, | |
| "loss": 0.1638, | |
| "step": 3120 | |
| }, | |
| { | |
| "epoch": 0.0026521739130434784, | |
| "grad_norm": 0.31789082288742065, | |
| "learning_rate": 4.799516908212561e-05, | |
| "loss": 0.1676, | |
| "step": 3130 | |
| }, | |
| { | |
| "epoch": 0.0030869565217391303, | |
| "grad_norm": 2.230436325073242, | |
| "learning_rate": 4.797101449275362e-05, | |
| "loss": 0.0492, | |
| "step": 3140 | |
| }, | |
| { | |
| "epoch": 0.0035217391304347826, | |
| "grad_norm": 0.008552751503884792, | |
| "learning_rate": 4.7946859903381646e-05, | |
| "loss": 0.1069, | |
| "step": 3150 | |
| }, | |
| { | |
| "epoch": 0.003956521739130435, | |
| "grad_norm": 0.006410131696611643, | |
| "learning_rate": 4.7922705314009664e-05, | |
| "loss": 0.0019, | |
| "step": 3160 | |
| }, | |
| { | |
| "epoch": 0.004043478260869565, | |
| "eval_accuracy": 0.8688524590163934, | |
| "eval_f1": 0.8677470707315438, | |
| "eval_loss": 0.8008872866630554, | |
| "eval_runtime": 75.7619, | |
| "eval_samples_per_second": 1.61, | |
| "eval_steps_per_second": 0.211, | |
| "step": 3162 | |
| }, | |
| { | |
| "epoch": 1.0003478260869565, | |
| "grad_norm": 0.0073925890028476715, | |
| "learning_rate": 4.789855072463768e-05, | |
| "loss": 0.0007, | |
| "step": 3170 | |
| }, | |
| { | |
| "epoch": 1.0007826086956522, | |
| "grad_norm": 0.009194503538310528, | |
| "learning_rate": 4.7874396135265706e-05, | |
| "loss": 0.0828, | |
| "step": 3180 | |
| }, | |
| { | |
| "epoch": 1.0012173913043478, | |
| "grad_norm": 0.01020804513245821, | |
| "learning_rate": 4.785024154589372e-05, | |
| "loss": 0.0298, | |
| "step": 3190 | |
| }, | |
| { | |
| "epoch": 1.0016521739130435, | |
| "grad_norm": 46.36565017700195, | |
| "learning_rate": 4.782608695652174e-05, | |
| "loss": 0.0086, | |
| "step": 3200 | |
| }, | |
| { | |
| "epoch": 1.0020869565217392, | |
| "grad_norm": 0.007284100167453289, | |
| "learning_rate": 4.780193236714976e-05, | |
| "loss": 0.1071, | |
| "step": 3210 | |
| }, | |
| { | |
| "epoch": 1.0025217391304349, | |
| "grad_norm": 3.947695016860962, | |
| "learning_rate": 4.7777777777777784e-05, | |
| "loss": 0.0892, | |
| "step": 3220 | |
| }, | |
| { | |
| "epoch": 1.0029565217391305, | |
| "grad_norm": 49.097679138183594, | |
| "learning_rate": 4.77536231884058e-05, | |
| "loss": 0.1349, | |
| "step": 3230 | |
| }, | |
| { | |
| "epoch": 1.0033913043478262, | |
| "grad_norm": 0.18935832381248474, | |
| "learning_rate": 4.772946859903382e-05, | |
| "loss": 0.062, | |
| "step": 3240 | |
| }, | |
| { | |
| "epoch": 1.0038260869565216, | |
| "grad_norm": 2.587529182434082, | |
| "learning_rate": 4.770531400966184e-05, | |
| "loss": 0.0724, | |
| "step": 3250 | |
| }, | |
| { | |
| "epoch": 1.0040434782608696, | |
| "eval_accuracy": 0.8770491803278688, | |
| "eval_f1": 0.8773590184659386, | |
| "eval_loss": 0.7587926983833313, | |
| "eval_runtime": 73.9063, | |
| "eval_samples_per_second": 1.651, | |
| "eval_steps_per_second": 0.216, | |
| "step": 3255 | |
| }, | |
| { | |
| "epoch": 2.0002173913043477, | |
| "grad_norm": 0.12760405242443085, | |
| "learning_rate": 4.7681159420289855e-05, | |
| "loss": 0.0737, | |
| "step": 3260 | |
| }, | |
| { | |
| "epoch": 2.0006521739130436, | |
| "grad_norm": 0.626366376876831, | |
| "learning_rate": 4.765700483091788e-05, | |
| "loss": 0.04, | |
| "step": 3270 | |
| }, | |
| { | |
| "epoch": 2.001086956521739, | |
| "grad_norm": 0.5861339569091797, | |
| "learning_rate": 4.76328502415459e-05, | |
| "loss": 0.0014, | |
| "step": 3280 | |
| }, | |
| { | |
| "epoch": 2.001521739130435, | |
| "grad_norm": 0.016094276681542397, | |
| "learning_rate": 4.7608695652173916e-05, | |
| "loss": 0.0006, | |
| "step": 3290 | |
| }, | |
| { | |
| "epoch": 2.0019565217391304, | |
| "grad_norm": 0.03283320367336273, | |
| "learning_rate": 4.7584541062801933e-05, | |
| "loss": 0.0004, | |
| "step": 3300 | |
| }, | |
| { | |
| "epoch": 2.0023913043478263, | |
| "grad_norm": 0.007951553910970688, | |
| "learning_rate": 4.756038647342995e-05, | |
| "loss": 0.0869, | |
| "step": 3310 | |
| }, | |
| { | |
| "epoch": 2.0028260869565218, | |
| "grad_norm": 78.94805145263672, | |
| "learning_rate": 4.7536231884057976e-05, | |
| "loss": 0.0195, | |
| "step": 3320 | |
| }, | |
| { | |
| "epoch": 2.003260869565217, | |
| "grad_norm": 0.00863903108984232, | |
| "learning_rate": 4.7512077294685994e-05, | |
| "loss": 0.0031, | |
| "step": 3330 | |
| }, | |
| { | |
| "epoch": 2.003695652173913, | |
| "grad_norm": 0.010919756256043911, | |
| "learning_rate": 4.748792270531401e-05, | |
| "loss": 0.0539, | |
| "step": 3340 | |
| }, | |
| { | |
| "epoch": 2.0040434782608694, | |
| "eval_accuracy": 0.8852459016393442, | |
| "eval_f1": 0.8845549202267897, | |
| "eval_loss": 0.784960150718689, | |
| "eval_runtime": 71.8592, | |
| "eval_samples_per_second": 1.698, | |
| "eval_steps_per_second": 0.223, | |
| "step": 3348 | |
| }, | |
| { | |
| "epoch": 3.000086956521739, | |
| "grad_norm": 2.7393057346343994, | |
| "learning_rate": 4.746376811594203e-05, | |
| "loss": 0.0407, | |
| "step": 3350 | |
| }, | |
| { | |
| "epoch": 3.0005217391304346, | |
| "grad_norm": 0.034644246101379395, | |
| "learning_rate": 4.743961352657005e-05, | |
| "loss": 0.0011, | |
| "step": 3360 | |
| }, | |
| { | |
| "epoch": 3.0009565217391305, | |
| "grad_norm": 0.042398907244205475, | |
| "learning_rate": 4.741545893719807e-05, | |
| "loss": 0.0005, | |
| "step": 3370 | |
| }, | |
| { | |
| "epoch": 3.001391304347826, | |
| "grad_norm": 0.005558915436267853, | |
| "learning_rate": 4.739130434782609e-05, | |
| "loss": 0.0004, | |
| "step": 3380 | |
| }, | |
| { | |
| "epoch": 3.001826086956522, | |
| "grad_norm": 0.0052604591473937035, | |
| "learning_rate": 4.736714975845411e-05, | |
| "loss": 0.0087, | |
| "step": 3390 | |
| }, | |
| { | |
| "epoch": 3.0022608695652173, | |
| "grad_norm": 0.0026659213472157717, | |
| "learning_rate": 4.7342995169082125e-05, | |
| "loss": 0.0153, | |
| "step": 3400 | |
| }, | |
| { | |
| "epoch": 3.002695652173913, | |
| "grad_norm": 0.005264274310320616, | |
| "learning_rate": 4.731884057971015e-05, | |
| "loss": 0.0119, | |
| "step": 3410 | |
| }, | |
| { | |
| "epoch": 3.0031304347826087, | |
| "grad_norm": 0.007553906179964542, | |
| "learning_rate": 4.729468599033817e-05, | |
| "loss": 0.0531, | |
| "step": 3420 | |
| }, | |
| { | |
| "epoch": 3.0035652173913046, | |
| "grad_norm": 2.2136473655700684, | |
| "learning_rate": 4.7270531400966185e-05, | |
| "loss": 0.0227, | |
| "step": 3430 | |
| }, | |
| { | |
| "epoch": 3.004, | |
| "grad_norm": 0.020709699019789696, | |
| "learning_rate": 4.72463768115942e-05, | |
| "loss": 0.1208, | |
| "step": 3440 | |
| }, | |
| { | |
| "epoch": 3.0040434782608694, | |
| "eval_accuracy": 0.8852459016393442, | |
| "eval_f1": 0.8853364802271643, | |
| "eval_loss": 0.7922825813293457, | |
| "eval_runtime": 76.251, | |
| "eval_samples_per_second": 1.6, | |
| "eval_steps_per_second": 0.21, | |
| "step": 3441 | |
| }, | |
| { | |
| "epoch": 4.0003913043478265, | |
| "grad_norm": 0.029359057545661926, | |
| "learning_rate": 4.722222222222222e-05, | |
| "loss": 0.0104, | |
| "step": 3450 | |
| }, | |
| { | |
| "epoch": 4.000826086956522, | |
| "grad_norm": 0.0061401985585689545, | |
| "learning_rate": 4.7198067632850246e-05, | |
| "loss": 0.0329, | |
| "step": 3460 | |
| }, | |
| { | |
| "epoch": 4.001260869565217, | |
| "grad_norm": 0.004194505978375673, | |
| "learning_rate": 4.7173913043478264e-05, | |
| "loss": 0.0005, | |
| "step": 3470 | |
| }, | |
| { | |
| "epoch": 4.001695652173913, | |
| "grad_norm": 0.016542797908186913, | |
| "learning_rate": 4.714975845410628e-05, | |
| "loss": 0.0005, | |
| "step": 3480 | |
| }, | |
| { | |
| "epoch": 4.002130434782608, | |
| "grad_norm": 0.009181750938296318, | |
| "learning_rate": 4.71256038647343e-05, | |
| "loss": 0.0004, | |
| "step": 3490 | |
| }, | |
| { | |
| "epoch": 4.002565217391305, | |
| "grad_norm": 0.0038063046522438526, | |
| "learning_rate": 4.710144927536232e-05, | |
| "loss": 0.0102, | |
| "step": 3500 | |
| }, | |
| { | |
| "epoch": 4.003, | |
| "grad_norm": 0.012156417593359947, | |
| "learning_rate": 4.707729468599034e-05, | |
| "loss": 0.0101, | |
| "step": 3510 | |
| }, | |
| { | |
| "epoch": 4.003434782608696, | |
| "grad_norm": 0.00890531949698925, | |
| "learning_rate": 4.705314009661836e-05, | |
| "loss": 0.0217, | |
| "step": 3520 | |
| }, | |
| { | |
| "epoch": 4.003869565217391, | |
| "grad_norm": 0.004342174157500267, | |
| "learning_rate": 4.7028985507246384e-05, | |
| "loss": 0.0467, | |
| "step": 3530 | |
| }, | |
| { | |
| "epoch": 4.004043478260869, | |
| "eval_accuracy": 0.860655737704918, | |
| "eval_f1": 0.8612302252359274, | |
| "eval_loss": 1.0240317583084106, | |
| "eval_runtime": 63.4449, | |
| "eval_samples_per_second": 1.923, | |
| "eval_steps_per_second": 0.252, | |
| "step": 3534 | |
| }, | |
| { | |
| "epoch": 5.000260869565217, | |
| "grad_norm": 0.025819310918450356, | |
| "learning_rate": 4.7004830917874395e-05, | |
| "loss": 0.0004, | |
| "step": 3540 | |
| }, | |
| { | |
| "epoch": 5.000695652173913, | |
| "grad_norm": 0.010917126201093197, | |
| "learning_rate": 4.698067632850241e-05, | |
| "loss": 0.0163, | |
| "step": 3550 | |
| }, | |
| { | |
| "epoch": 5.001130434782609, | |
| "grad_norm": 0.0050573805347085, | |
| "learning_rate": 4.695652173913044e-05, | |
| "loss": 0.2046, | |
| "step": 3560 | |
| }, | |
| { | |
| "epoch": 5.001565217391304, | |
| "grad_norm": 22.131898880004883, | |
| "learning_rate": 4.6932367149758455e-05, | |
| "loss": 0.0945, | |
| "step": 3570 | |
| }, | |
| { | |
| "epoch": 5.002, | |
| "grad_norm": 0.013935157097876072, | |
| "learning_rate": 4.690821256038648e-05, | |
| "loss": 0.0026, | |
| "step": 3580 | |
| }, | |
| { | |
| "epoch": 5.002434782608695, | |
| "grad_norm": 0.006178060546517372, | |
| "learning_rate": 4.68840579710145e-05, | |
| "loss": 0.0018, | |
| "step": 3590 | |
| }, | |
| { | |
| "epoch": 5.002869565217392, | |
| "grad_norm": 0.006322579458355904, | |
| "learning_rate": 4.6859903381642516e-05, | |
| "loss": 0.0309, | |
| "step": 3600 | |
| }, | |
| { | |
| "epoch": 5.003304347826087, | |
| "grad_norm": 0.006985423155128956, | |
| "learning_rate": 4.683574879227053e-05, | |
| "loss": 0.0285, | |
| "step": 3610 | |
| }, | |
| { | |
| "epoch": 5.0037391304347825, | |
| "grad_norm": 0.004971032030880451, | |
| "learning_rate": 4.681159420289855e-05, | |
| "loss": 0.0764, | |
| "step": 3620 | |
| }, | |
| { | |
| "epoch": 5.004043478260869, | |
| "eval_accuracy": 0.8852459016393442, | |
| "eval_f1": 0.8863337496985733, | |
| "eval_loss": 0.7494808435440063, | |
| "eval_runtime": 68.5855, | |
| "eval_samples_per_second": 1.779, | |
| "eval_steps_per_second": 0.233, | |
| "step": 3627 | |
| }, | |
| { | |
| "epoch": 6.0001304347826085, | |
| "grad_norm": 0.26816919445991516, | |
| "learning_rate": 4.6787439613526576e-05, | |
| "loss": 0.0008, | |
| "step": 3630 | |
| }, | |
| { | |
| "epoch": 6.000565217391304, | |
| "grad_norm": 0.011856114491820335, | |
| "learning_rate": 4.6763285024154594e-05, | |
| "loss": 0.08, | |
| "step": 3640 | |
| }, | |
| { | |
| "epoch": 6.001, | |
| "grad_norm": 0.006870527286082506, | |
| "learning_rate": 4.673913043478261e-05, | |
| "loss": 0.0005, | |
| "step": 3650 | |
| }, | |
| { | |
| "epoch": 6.001434782608696, | |
| "grad_norm": 0.20965741574764252, | |
| "learning_rate": 4.671497584541063e-05, | |
| "loss": 0.1102, | |
| "step": 3660 | |
| }, | |
| { | |
| "epoch": 6.001869565217391, | |
| "grad_norm": 0.018843989819288254, | |
| "learning_rate": 4.669082125603865e-05, | |
| "loss": 0.0741, | |
| "step": 3670 | |
| }, | |
| { | |
| "epoch": 6.002304347826087, | |
| "grad_norm": 97.60299682617188, | |
| "learning_rate": 4.666666666666667e-05, | |
| "loss": 0.049, | |
| "step": 3680 | |
| }, | |
| { | |
| "epoch": 6.002739130434783, | |
| "grad_norm": 0.006012632045894861, | |
| "learning_rate": 4.664251207729469e-05, | |
| "loss": 0.0635, | |
| "step": 3690 | |
| }, | |
| { | |
| "epoch": 6.0031739130434785, | |
| "grad_norm": 10.848343849182129, | |
| "learning_rate": 4.661835748792271e-05, | |
| "loss": 0.0984, | |
| "step": 3700 | |
| }, | |
| { | |
| "epoch": 6.003608695652174, | |
| "grad_norm": 0.004343053791671991, | |
| "learning_rate": 4.6594202898550725e-05, | |
| "loss": 0.0887, | |
| "step": 3710 | |
| }, | |
| { | |
| "epoch": 6.004043478260869, | |
| "grad_norm": 0.006757623050361872, | |
| "learning_rate": 4.657004830917875e-05, | |
| "loss": 0.0006, | |
| "step": 3720 | |
| }, | |
| { | |
| "epoch": 6.004043478260869, | |
| "eval_accuracy": 0.8360655737704918, | |
| "eval_f1": 0.83469500668468, | |
| "eval_loss": 0.9808096289634705, | |
| "eval_runtime": 64.2569, | |
| "eval_samples_per_second": 1.899, | |
| "eval_steps_per_second": 0.249, | |
| "step": 3720 | |
| }, | |
| { | |
| "epoch": 7.0004347826086954, | |
| "grad_norm": 0.014435753226280212, | |
| "learning_rate": 4.654589371980677e-05, | |
| "loss": 0.0005, | |
| "step": 3730 | |
| }, | |
| { | |
| "epoch": 7.000869565217391, | |
| "grad_norm": 0.0036905386950820684, | |
| "learning_rate": 4.6521739130434785e-05, | |
| "loss": 0.0005, | |
| "step": 3740 | |
| }, | |
| { | |
| "epoch": 7.001304347826087, | |
| "grad_norm": 0.005206993781030178, | |
| "learning_rate": 4.64975845410628e-05, | |
| "loss": 0.0003, | |
| "step": 3750 | |
| }, | |
| { | |
| "epoch": 7.001739130434783, | |
| "grad_norm": 0.015529815107584, | |
| "learning_rate": 4.647342995169082e-05, | |
| "loss": 0.0177, | |
| "step": 3760 | |
| }, | |
| { | |
| "epoch": 7.002173913043478, | |
| "grad_norm": 0.02242157608270645, | |
| "learning_rate": 4.6449275362318846e-05, | |
| "loss": 0.0003, | |
| "step": 3770 | |
| }, | |
| { | |
| "epoch": 7.002608695652174, | |
| "grad_norm": 0.004425575491040945, | |
| "learning_rate": 4.642512077294686e-05, | |
| "loss": 0.0123, | |
| "step": 3780 | |
| }, | |
| { | |
| "epoch": 7.00304347826087, | |
| "grad_norm": 0.0038235390093177557, | |
| "learning_rate": 4.640096618357488e-05, | |
| "loss": 0.0101, | |
| "step": 3790 | |
| }, | |
| { | |
| "epoch": 7.003478260869565, | |
| "grad_norm": 0.051797546446323395, | |
| "learning_rate": 4.63768115942029e-05, | |
| "loss": 0.0724, | |
| "step": 3800 | |
| }, | |
| { | |
| "epoch": 7.003913043478261, | |
| "grad_norm": 0.021821072325110435, | |
| "learning_rate": 4.635265700483092e-05, | |
| "loss": 0.1386, | |
| "step": 3810 | |
| }, | |
| { | |
| "epoch": 7.004043478260869, | |
| "eval_accuracy": 0.8770491803278688, | |
| "eval_f1": 0.8770793838312594, | |
| "eval_loss": 0.7652427554130554, | |
| "eval_runtime": 71.3373, | |
| "eval_samples_per_second": 1.71, | |
| "eval_steps_per_second": 0.224, | |
| "step": 3813 | |
| }, | |
| { | |
| "epoch": 8.000304347826088, | |
| "grad_norm": 0.0032370842527598143, | |
| "learning_rate": 4.632850241545894e-05, | |
| "loss": 0.0129, | |
| "step": 3820 | |
| }, | |
| { | |
| "epoch": 8.000739130434782, | |
| "grad_norm": 1.7167001962661743, | |
| "learning_rate": 4.630434782608696e-05, | |
| "loss": 0.021, | |
| "step": 3830 | |
| }, | |
| { | |
| "epoch": 8.001173913043479, | |
| "grad_norm": 0.003859913209453225, | |
| "learning_rate": 4.628019323671498e-05, | |
| "loss": 0.0015, | |
| "step": 3840 | |
| }, | |
| { | |
| "epoch": 8.001608695652173, | |
| "grad_norm": 0.0031530093401670456, | |
| "learning_rate": 4.6256038647342995e-05, | |
| "loss": 0.0003, | |
| "step": 3850 | |
| }, | |
| { | |
| "epoch": 8.00204347826087, | |
| "grad_norm": 0.007489124778658152, | |
| "learning_rate": 4.623188405797101e-05, | |
| "loss": 0.0081, | |
| "step": 3860 | |
| }, | |
| { | |
| "epoch": 8.002478260869566, | |
| "grad_norm": 2.5608975887298584, | |
| "learning_rate": 4.620772946859904e-05, | |
| "loss": 0.1232, | |
| "step": 3870 | |
| }, | |
| { | |
| "epoch": 8.00291304347826, | |
| "grad_norm": 63.46572494506836, | |
| "learning_rate": 4.6183574879227055e-05, | |
| "loss": 0.0507, | |
| "step": 3880 | |
| }, | |
| { | |
| "epoch": 8.003347826086957, | |
| "grad_norm": 124.65734100341797, | |
| "learning_rate": 4.615942028985508e-05, | |
| "loss": 0.1479, | |
| "step": 3890 | |
| }, | |
| { | |
| "epoch": 8.003782608695651, | |
| "grad_norm": 57.03828430175781, | |
| "learning_rate": 4.613526570048309e-05, | |
| "loss": 0.1791, | |
| "step": 3900 | |
| }, | |
| { | |
| "epoch": 8.00404347826087, | |
| "eval_accuracy": 0.8360655737704918, | |
| "eval_f1": 0.8328396085819532, | |
| "eval_loss": 1.0916365385055542, | |
| "eval_runtime": 76.5479, | |
| "eval_samples_per_second": 1.594, | |
| "eval_steps_per_second": 0.209, | |
| "step": 3906 | |
| }, | |
| { | |
| "epoch": 9.000173913043477, | |
| "grad_norm": 0.017659667879343033, | |
| "learning_rate": 4.6111111111111115e-05, | |
| "loss": 0.0024, | |
| "step": 3910 | |
| }, | |
| { | |
| "epoch": 9.000608695652174, | |
| "grad_norm": 0.004569427575916052, | |
| "learning_rate": 4.608695652173913e-05, | |
| "loss": 0.1475, | |
| "step": 3920 | |
| }, | |
| { | |
| "epoch": 9.00104347826087, | |
| "grad_norm": 0.4839405417442322, | |
| "learning_rate": 4.606280193236715e-05, | |
| "loss": 0.0576, | |
| "step": 3930 | |
| }, | |
| { | |
| "epoch": 9.001478260869565, | |
| "grad_norm": 0.005257844924926758, | |
| "learning_rate": 4.6038647342995176e-05, | |
| "loss": 0.0811, | |
| "step": 3940 | |
| }, | |
| { | |
| "epoch": 9.001913043478261, | |
| "grad_norm": 0.005183045752346516, | |
| "learning_rate": 4.601449275362319e-05, | |
| "loss": 0.2326, | |
| "step": 3950 | |
| }, | |
| { | |
| "epoch": 9.002347826086957, | |
| "grad_norm": 0.20619826018810272, | |
| "learning_rate": 4.599033816425121e-05, | |
| "loss": 0.084, | |
| "step": 3960 | |
| }, | |
| { | |
| "epoch": 9.002782608695652, | |
| "grad_norm": 0.10698182135820389, | |
| "learning_rate": 4.596618357487923e-05, | |
| "loss": 0.0156, | |
| "step": 3970 | |
| }, | |
| { | |
| "epoch": 9.003217391304348, | |
| "grad_norm": 0.004087397828698158, | |
| "learning_rate": 4.594202898550725e-05, | |
| "loss": 0.0312, | |
| "step": 3980 | |
| }, | |
| { | |
| "epoch": 9.003652173913043, | |
| "grad_norm": 0.004443726502358913, | |
| "learning_rate": 4.591787439613527e-05, | |
| "loss": 0.0169, | |
| "step": 3990 | |
| }, | |
| { | |
| "epoch": 9.00404347826087, | |
| "eval_accuracy": 0.8852459016393442, | |
| "eval_f1": 0.8848651800896989, | |
| "eval_loss": 0.6546049118041992, | |
| "eval_runtime": 74.6104, | |
| "eval_samples_per_second": 1.635, | |
| "eval_steps_per_second": 0.214, | |
| "step": 3999 | |
| }, | |
| { | |
| "epoch": 10.000043478260869, | |
| "grad_norm": 0.004015814512968063, | |
| "learning_rate": 4.589371980676328e-05, | |
| "loss": 0.086, | |
| "step": 4000 | |
| }, | |
| { | |
| "epoch": 10.000478260869565, | |
| "grad_norm": 0.0028713741339743137, | |
| "learning_rate": 4.586956521739131e-05, | |
| "loss": 0.0137, | |
| "step": 4010 | |
| }, | |
| { | |
| "epoch": 10.000913043478262, | |
| "grad_norm": 0.0047510224394500256, | |
| "learning_rate": 4.5845410628019325e-05, | |
| "loss": 0.0065, | |
| "step": 4020 | |
| }, | |
| { | |
| "epoch": 10.001347826086956, | |
| "grad_norm": 0.0044855643063783646, | |
| "learning_rate": 4.582125603864735e-05, | |
| "loss": 0.0504, | |
| "step": 4030 | |
| }, | |
| { | |
| "epoch": 10.001782608695652, | |
| "grad_norm": 0.07487785071134567, | |
| "learning_rate": 4.579710144927537e-05, | |
| "loss": 0.0837, | |
| "step": 4040 | |
| }, | |
| { | |
| "epoch": 10.002217391304347, | |
| "grad_norm": 0.025662241503596306, | |
| "learning_rate": 4.577294685990338e-05, | |
| "loss": 0.0422, | |
| "step": 4050 | |
| }, | |
| { | |
| "epoch": 10.002652173913043, | |
| "grad_norm": 3.2152881622314453, | |
| "learning_rate": 4.57487922705314e-05, | |
| "loss": 0.0009, | |
| "step": 4060 | |
| }, | |
| { | |
| "epoch": 10.00308695652174, | |
| "grad_norm": 0.009620684199035168, | |
| "learning_rate": 4.572463768115942e-05, | |
| "loss": 0.0598, | |
| "step": 4070 | |
| }, | |
| { | |
| "epoch": 10.003521739130434, | |
| "grad_norm": 0.0028383173048496246, | |
| "learning_rate": 4.5700483091787445e-05, | |
| "loss": 0.1004, | |
| "step": 4080 | |
| }, | |
| { | |
| "epoch": 10.00395652173913, | |
| "grad_norm": 0.23549234867095947, | |
| "learning_rate": 4.567632850241546e-05, | |
| "loss": 0.0228, | |
| "step": 4090 | |
| }, | |
| { | |
| "epoch": 10.00404347826087, | |
| "eval_accuracy": 0.8524590163934426, | |
| "eval_f1": 0.8534775589608828, | |
| "eval_loss": 0.9656055569648743, | |
| "eval_runtime": 77.7343, | |
| "eval_samples_per_second": 1.569, | |
| "eval_steps_per_second": 0.206, | |
| "step": 4092 | |
| }, | |
| { | |
| "epoch": 11.000347826086957, | |
| "grad_norm": 0.001953911269083619, | |
| "learning_rate": 4.565217391304348e-05, | |
| "loss": 0.0454, | |
| "step": 4100 | |
| }, | |
| { | |
| "epoch": 11.000782608695653, | |
| "grad_norm": 0.09437835961580276, | |
| "learning_rate": 4.56280193236715e-05, | |
| "loss": 0.0004, | |
| "step": 4110 | |
| }, | |
| { | |
| "epoch": 11.001217391304348, | |
| "grad_norm": 0.009714662097394466, | |
| "learning_rate": 4.560386473429952e-05, | |
| "loss": 0.0106, | |
| "step": 4120 | |
| }, | |
| { | |
| "epoch": 11.001652173913044, | |
| "grad_norm": 0.002876744605600834, | |
| "learning_rate": 4.557971014492754e-05, | |
| "loss": 0.0918, | |
| "step": 4130 | |
| }, | |
| { | |
| "epoch": 11.002086956521739, | |
| "grad_norm": 0.0034505492076277733, | |
| "learning_rate": 4.555555555555556e-05, | |
| "loss": 0.0166, | |
| "step": 4140 | |
| }, | |
| { | |
| "epoch": 11.002521739130435, | |
| "grad_norm": 0.0048155467957258224, | |
| "learning_rate": 4.553140096618358e-05, | |
| "loss": 0.3045, | |
| "step": 4150 | |
| }, | |
| { | |
| "epoch": 11.002956521739131, | |
| "grad_norm": 1.3779292106628418, | |
| "learning_rate": 4.5507246376811595e-05, | |
| "loss": 0.0084, | |
| "step": 4160 | |
| }, | |
| { | |
| "epoch": 11.003391304347826, | |
| "grad_norm": 0.0718054547905922, | |
| "learning_rate": 4.548309178743961e-05, | |
| "loss": 0.2187, | |
| "step": 4170 | |
| }, | |
| { | |
| "epoch": 11.003826086956522, | |
| "grad_norm": 0.044493287801742554, | |
| "learning_rate": 4.545893719806764e-05, | |
| "loss": 0.0307, | |
| "step": 4180 | |
| }, | |
| { | |
| "epoch": 11.00404347826087, | |
| "eval_accuracy": 0.7950819672131147, | |
| "eval_f1": 0.7963983516128617, | |
| "eval_loss": 1.334099292755127, | |
| "eval_runtime": 79.1262, | |
| "eval_samples_per_second": 1.542, | |
| "eval_steps_per_second": 0.202, | |
| "step": 4185 | |
| }, | |
| { | |
| "epoch": 12.000217391304348, | |
| "grad_norm": 12.314984321594238, | |
| "learning_rate": 4.5434782608695655e-05, | |
| "loss": 0.2449, | |
| "step": 4190 | |
| }, | |
| { | |
| "epoch": 12.000652173913043, | |
| "grad_norm": 0.4841020107269287, | |
| "learning_rate": 4.541062801932367e-05, | |
| "loss": 0.0824, | |
| "step": 4200 | |
| }, | |
| { | |
| "epoch": 12.001086956521739, | |
| "grad_norm": 31.790552139282227, | |
| "learning_rate": 4.538647342995169e-05, | |
| "loss": 0.0731, | |
| "step": 4210 | |
| }, | |
| { | |
| "epoch": 12.001521739130435, | |
| "grad_norm": 0.0036116011906415224, | |
| "learning_rate": 4.5362318840579715e-05, | |
| "loss": 0.0276, | |
| "step": 4220 | |
| }, | |
| { | |
| "epoch": 12.00195652173913, | |
| "grad_norm": 0.016497021540999413, | |
| "learning_rate": 4.533816425120773e-05, | |
| "loss": 0.096, | |
| "step": 4230 | |
| }, | |
| { | |
| "epoch": 12.002391304347826, | |
| "grad_norm": 0.016652707010507584, | |
| "learning_rate": 4.531400966183575e-05, | |
| "loss": 0.4404, | |
| "step": 4240 | |
| }, | |
| { | |
| "epoch": 12.00282608695652, | |
| "grad_norm": 0.03973345831036568, | |
| "learning_rate": 4.528985507246377e-05, | |
| "loss": 0.0343, | |
| "step": 4250 | |
| }, | |
| { | |
| "epoch": 12.003260869565217, | |
| "grad_norm": 133.8014373779297, | |
| "learning_rate": 4.5265700483091786e-05, | |
| "loss": 0.497, | |
| "step": 4260 | |
| }, | |
| { | |
| "epoch": 12.003695652173914, | |
| "grad_norm": 2.6162045001983643, | |
| "learning_rate": 4.524154589371981e-05, | |
| "loss": 0.5106, | |
| "step": 4270 | |
| }, | |
| { | |
| "epoch": 12.00404347826087, | |
| "eval_accuracy": 0.7950819672131147, | |
| "eval_f1": 0.7876538622910378, | |
| "eval_loss": 1.2499514818191528, | |
| "eval_runtime": 68.1294, | |
| "eval_samples_per_second": 1.791, | |
| "eval_steps_per_second": 0.235, | |
| "step": 4278 | |
| }, | |
| { | |
| "epoch": 13.00008695652174, | |
| "grad_norm": 41.279598236083984, | |
| "learning_rate": 4.521739130434783e-05, | |
| "loss": 0.1115, | |
| "step": 4280 | |
| }, | |
| { | |
| "epoch": 13.000521739130434, | |
| "grad_norm": 1.6680902242660522, | |
| "learning_rate": 4.519323671497585e-05, | |
| "loss": 0.018, | |
| "step": 4290 | |
| }, | |
| { | |
| "epoch": 13.00095652173913, | |
| "grad_norm": 0.009661109186708927, | |
| "learning_rate": 4.5169082125603865e-05, | |
| "loss": 0.1721, | |
| "step": 4300 | |
| }, | |
| { | |
| "epoch": 13.001391304347827, | |
| "grad_norm": 0.02459769882261753, | |
| "learning_rate": 4.514492753623188e-05, | |
| "loss": 0.0452, | |
| "step": 4310 | |
| }, | |
| { | |
| "epoch": 13.001826086956521, | |
| "grad_norm": 0.026296913623809814, | |
| "learning_rate": 4.512077294685991e-05, | |
| "loss": 0.1139, | |
| "step": 4320 | |
| }, | |
| { | |
| "epoch": 13.002260869565218, | |
| "grad_norm": 0.31147295236587524, | |
| "learning_rate": 4.5096618357487925e-05, | |
| "loss": 0.0273, | |
| "step": 4330 | |
| }, | |
| { | |
| "epoch": 13.002695652173912, | |
| "grad_norm": 12.396149635314941, | |
| "learning_rate": 4.507246376811595e-05, | |
| "loss": 0.3046, | |
| "step": 4340 | |
| }, | |
| { | |
| "epoch": 13.003130434782609, | |
| "grad_norm": 0.010565920732915401, | |
| "learning_rate": 4.504830917874396e-05, | |
| "loss": 0.0204, | |
| "step": 4350 | |
| }, | |
| { | |
| "epoch": 13.003565217391305, | |
| "grad_norm": 0.004811236634850502, | |
| "learning_rate": 4.502415458937198e-05, | |
| "loss": 0.0167, | |
| "step": 4360 | |
| }, | |
| { | |
| "epoch": 13.004, | |
| "grad_norm": 0.008650255389511585, | |
| "learning_rate": 4.5e-05, | |
| "loss": 0.0108, | |
| "step": 4370 | |
| }, | |
| { | |
| "epoch": 13.00404347826087, | |
| "eval_accuracy": 0.8524590163934426, | |
| "eval_f1": 0.8495038417270663, | |
| "eval_loss": 0.8027352690696716, | |
| "eval_runtime": 94.8246, | |
| "eval_samples_per_second": 1.287, | |
| "eval_steps_per_second": 0.169, | |
| "step": 4371 | |
| }, | |
| { | |
| "epoch": 14.000391304347826, | |
| "grad_norm": 2.1783447265625, | |
| "learning_rate": 4.497584541062802e-05, | |
| "loss": 0.0578, | |
| "step": 4380 | |
| }, | |
| { | |
| "epoch": 14.000826086956522, | |
| "grad_norm": 0.008437794633209705, | |
| "learning_rate": 4.4951690821256045e-05, | |
| "loss": 0.0542, | |
| "step": 4390 | |
| }, | |
| { | |
| "epoch": 14.001260869565217, | |
| "grad_norm": 0.006100878585129976, | |
| "learning_rate": 4.492753623188406e-05, | |
| "loss": 0.0004, | |
| "step": 4400 | |
| }, | |
| { | |
| "epoch": 14.001695652173913, | |
| "grad_norm": 0.005631875246763229, | |
| "learning_rate": 4.4903381642512074e-05, | |
| "loss": 0.0932, | |
| "step": 4410 | |
| }, | |
| { | |
| "epoch": 14.00213043478261, | |
| "grad_norm": 52.9527587890625, | |
| "learning_rate": 4.48792270531401e-05, | |
| "loss": 0.1454, | |
| "step": 4420 | |
| }, | |
| { | |
| "epoch": 14.002565217391304, | |
| "grad_norm": 0.0049645500257611275, | |
| "learning_rate": 4.4855072463768117e-05, | |
| "loss": 0.0007, | |
| "step": 4430 | |
| }, | |
| { | |
| "epoch": 14.003, | |
| "grad_norm": 0.005069791804999113, | |
| "learning_rate": 4.483091787439614e-05, | |
| "loss": 0.0104, | |
| "step": 4440 | |
| }, | |
| { | |
| "epoch": 14.003434782608696, | |
| "grad_norm": 0.06516723334789276, | |
| "learning_rate": 4.480676328502416e-05, | |
| "loss": 0.1654, | |
| "step": 4450 | |
| }, | |
| { | |
| "epoch": 14.003869565217391, | |
| "grad_norm": 0.01129752118140459, | |
| "learning_rate": 4.478260869565218e-05, | |
| "loss": 0.0873, | |
| "step": 4460 | |
| }, | |
| { | |
| "epoch": 14.00404347826087, | |
| "eval_accuracy": 0.9016393442622951, | |
| "eval_f1": 0.9017991182732418, | |
| "eval_loss": 0.680787980556488, | |
| "eval_runtime": 78.5271, | |
| "eval_samples_per_second": 1.554, | |
| "eval_steps_per_second": 0.204, | |
| "step": 4464 | |
| }, | |
| { | |
| "epoch": 15.000260869565217, | |
| "grad_norm": 0.0042943209409713745, | |
| "learning_rate": 4.4758454106280195e-05, | |
| "loss": 0.0165, | |
| "step": 4470 | |
| }, | |
| { | |
| "epoch": 15.000695652173913, | |
| "grad_norm": 0.0025938786566257477, | |
| "learning_rate": 4.473429951690821e-05, | |
| "loss": 0.0602, | |
| "step": 4480 | |
| }, | |
| { | |
| "epoch": 15.001130434782608, | |
| "grad_norm": 0.02354370430111885, | |
| "learning_rate": 4.471014492753624e-05, | |
| "loss": 0.0003, | |
| "step": 4490 | |
| }, | |
| { | |
| "epoch": 15.001565217391304, | |
| "grad_norm": 0.0034618491772562265, | |
| "learning_rate": 4.4685990338164255e-05, | |
| "loss": 0.0157, | |
| "step": 4500 | |
| }, | |
| { | |
| "epoch": 15.002, | |
| "grad_norm": 0.002460119780153036, | |
| "learning_rate": 4.466183574879227e-05, | |
| "loss": 0.0834, | |
| "step": 4510 | |
| }, | |
| { | |
| "epoch": 15.002434782608695, | |
| "grad_norm": 0.008845089003443718, | |
| "learning_rate": 4.463768115942029e-05, | |
| "loss": 0.1572, | |
| "step": 4520 | |
| }, | |
| { | |
| "epoch": 15.002869565217392, | |
| "grad_norm": 0.004269142635166645, | |
| "learning_rate": 4.4613526570048315e-05, | |
| "loss": 0.3163, | |
| "step": 4530 | |
| }, | |
| { | |
| "epoch": 15.003304347826086, | |
| "grad_norm": 0.04755193367600441, | |
| "learning_rate": 4.458937198067633e-05, | |
| "loss": 0.1112, | |
| "step": 4540 | |
| }, | |
| { | |
| "epoch": 15.003739130434782, | |
| "grad_norm": 0.022059939801692963, | |
| "learning_rate": 4.456521739130435e-05, | |
| "loss": 0.0149, | |
| "step": 4550 | |
| }, | |
| { | |
| "epoch": 15.00404347826087, | |
| "eval_accuracy": 0.8852459016393442, | |
| "eval_f1": 0.8867104809727759, | |
| "eval_loss": 0.6811984777450562, | |
| "eval_runtime": 79.1497, | |
| "eval_samples_per_second": 1.541, | |
| "eval_steps_per_second": 0.202, | |
| "step": 4557 | |
| }, | |
| { | |
| "epoch": 16.00013043478261, | |
| "grad_norm": 0.004382013343274593, | |
| "learning_rate": 4.454106280193237e-05, | |
| "loss": 0.0005, | |
| "step": 4560 | |
| }, | |
| { | |
| "epoch": 16.000565217391305, | |
| "grad_norm": 0.007960804738104343, | |
| "learning_rate": 4.4516908212560386e-05, | |
| "loss": 0.0917, | |
| "step": 4570 | |
| }, | |
| { | |
| "epoch": 16.001, | |
| "grad_norm": 22.968700408935547, | |
| "learning_rate": 4.449275362318841e-05, | |
| "loss": 0.1009, | |
| "step": 4580 | |
| }, | |
| { | |
| "epoch": 16.001434782608694, | |
| "grad_norm": 0.004100767895579338, | |
| "learning_rate": 4.446859903381643e-05, | |
| "loss": 0.006, | |
| "step": 4590 | |
| }, | |
| { | |
| "epoch": 16.00186956521739, | |
| "grad_norm": 41.614383697509766, | |
| "learning_rate": 4.4444444444444447e-05, | |
| "loss": 0.0491, | |
| "step": 4600 | |
| }, | |
| { | |
| "epoch": 16.002304347826087, | |
| "grad_norm": 75.32587432861328, | |
| "learning_rate": 4.4420289855072464e-05, | |
| "loss": 0.2419, | |
| "step": 4610 | |
| }, | |
| { | |
| "epoch": 16.002739130434783, | |
| "grad_norm": 0.010296987369656563, | |
| "learning_rate": 4.439613526570048e-05, | |
| "loss": 0.0824, | |
| "step": 4620 | |
| }, | |
| { | |
| "epoch": 16.00317391304348, | |
| "grad_norm": 0.010856476612389088, | |
| "learning_rate": 4.437198067632851e-05, | |
| "loss": 0.1039, | |
| "step": 4630 | |
| }, | |
| { | |
| "epoch": 16.003608695652172, | |
| "grad_norm": 0.015950683504343033, | |
| "learning_rate": 4.4347826086956525e-05, | |
| "loss": 0.0357, | |
| "step": 4640 | |
| }, | |
| { | |
| "epoch": 16.00404347826087, | |
| "grad_norm": 0.006432201247662306, | |
| "learning_rate": 4.432367149758454e-05, | |
| "loss": 0.0146, | |
| "step": 4650 | |
| }, | |
| { | |
| "epoch": 16.00404347826087, | |
| "eval_accuracy": 0.9016393442622951, | |
| "eval_f1": 0.9015790225488998, | |
| "eval_loss": 0.6003494262695312, | |
| "eval_runtime": 64.7244, | |
| "eval_samples_per_second": 1.885, | |
| "eval_steps_per_second": 0.247, | |
| "step": 4650 | |
| }, | |
| { | |
| "epoch": 17.000434782608696, | |
| "grad_norm": 0.01721499115228653, | |
| "learning_rate": 4.429951690821256e-05, | |
| "loss": 0.0075, | |
| "step": 4660 | |
| }, | |
| { | |
| "epoch": 17.000869565217393, | |
| "grad_norm": 0.010429782792925835, | |
| "learning_rate": 4.427536231884058e-05, | |
| "loss": 0.0034, | |
| "step": 4670 | |
| }, | |
| { | |
| "epoch": 17.001304347826085, | |
| "grad_norm": 4.780106544494629, | |
| "learning_rate": 4.42512077294686e-05, | |
| "loss": 0.0837, | |
| "step": 4680 | |
| }, | |
| { | |
| "epoch": 17.001739130434782, | |
| "grad_norm": 0.015756214037537575, | |
| "learning_rate": 4.422705314009662e-05, | |
| "loss": 0.0114, | |
| "step": 4690 | |
| }, | |
| { | |
| "epoch": 17.002173913043478, | |
| "grad_norm": 0.3344515860080719, | |
| "learning_rate": 4.4202898550724645e-05, | |
| "loss": 0.0015, | |
| "step": 4700 | |
| }, | |
| { | |
| "epoch": 17.002608695652174, | |
| "grad_norm": 2.3182766437530518, | |
| "learning_rate": 4.4178743961352656e-05, | |
| "loss": 0.0464, | |
| "step": 4710 | |
| }, | |
| { | |
| "epoch": 17.00304347826087, | |
| "grad_norm": 0.008929559960961342, | |
| "learning_rate": 4.4154589371980674e-05, | |
| "loss": 0.0879, | |
| "step": 4720 | |
| }, | |
| { | |
| "epoch": 17.003478260869564, | |
| "grad_norm": 0.004621615167707205, | |
| "learning_rate": 4.41304347826087e-05, | |
| "loss": 0.0379, | |
| "step": 4730 | |
| }, | |
| { | |
| "epoch": 17.00391304347826, | |
| "grad_norm": 0.012686613015830517, | |
| "learning_rate": 4.4106280193236716e-05, | |
| "loss": 0.0994, | |
| "step": 4740 | |
| }, | |
| { | |
| "epoch": 17.00404347826087, | |
| "eval_accuracy": 0.860655737704918, | |
| "eval_f1": 0.8632714330882123, | |
| "eval_loss": 0.8046178221702576, | |
| "eval_runtime": 85.1421, | |
| "eval_samples_per_second": 1.433, | |
| "eval_steps_per_second": 0.188, | |
| "step": 4743 | |
| }, | |
| { | |
| "epoch": 18.000304347826088, | |
| "grad_norm": 0.003794121090322733, | |
| "learning_rate": 4.408212560386474e-05, | |
| "loss": 0.0121, | |
| "step": 4750 | |
| }, | |
| { | |
| "epoch": 18.000739130434784, | |
| "grad_norm": 0.002671042224392295, | |
| "learning_rate": 4.405797101449275e-05, | |
| "loss": 0.0112, | |
| "step": 4760 | |
| }, | |
| { | |
| "epoch": 18.001173913043477, | |
| "grad_norm": 0.0030148469377309084, | |
| "learning_rate": 4.403381642512078e-05, | |
| "loss": 0.0222, | |
| "step": 4770 | |
| }, | |
| { | |
| "epoch": 18.001608695652173, | |
| "grad_norm": 0.004704237449914217, | |
| "learning_rate": 4.4009661835748794e-05, | |
| "loss": 0.152, | |
| "step": 4780 | |
| }, | |
| { | |
| "epoch": 18.00204347826087, | |
| "grad_norm": 0.003907118458300829, | |
| "learning_rate": 4.398550724637681e-05, | |
| "loss": 0.0005, | |
| "step": 4790 | |
| }, | |
| { | |
| "epoch": 18.002478260869566, | |
| "grad_norm": 0.0029985394794493914, | |
| "learning_rate": 4.396135265700484e-05, | |
| "loss": 0.0003, | |
| "step": 4800 | |
| }, | |
| { | |
| "epoch": 18.002913043478262, | |
| "grad_norm": 2.0193307399749756, | |
| "learning_rate": 4.393719806763285e-05, | |
| "loss": 0.1093, | |
| "step": 4810 | |
| }, | |
| { | |
| "epoch": 18.003347826086955, | |
| "grad_norm": 0.007247489411383867, | |
| "learning_rate": 4.391304347826087e-05, | |
| "loss": 0.0969, | |
| "step": 4820 | |
| }, | |
| { | |
| "epoch": 18.00378260869565, | |
| "grad_norm": 35.1983642578125, | |
| "learning_rate": 4.388888888888889e-05, | |
| "loss": 0.1715, | |
| "step": 4830 | |
| }, | |
| { | |
| "epoch": 18.00404347826087, | |
| "eval_accuracy": 0.8852459016393442, | |
| "eval_f1": 0.8853454924681606, | |
| "eval_loss": 0.5631609559059143, | |
| "eval_runtime": 81.4165, | |
| "eval_samples_per_second": 1.498, | |
| "eval_steps_per_second": 0.197, | |
| "step": 4836 | |
| }, | |
| { | |
| "epoch": 19.00017391304348, | |
| "grad_norm": 0.24528348445892334, | |
| "learning_rate": 4.3864734299516915e-05, | |
| "loss": 0.0645, | |
| "step": 4840 | |
| }, | |
| { | |
| "epoch": 19.000608695652176, | |
| "grad_norm": 0.018249379470944405, | |
| "learning_rate": 4.384057971014493e-05, | |
| "loss": 0.0146, | |
| "step": 4850 | |
| }, | |
| { | |
| "epoch": 19.00104347826087, | |
| "grad_norm": 17.096920013427734, | |
| "learning_rate": 4.3816425120772944e-05, | |
| "loss": 0.0072, | |
| "step": 4860 | |
| }, | |
| { | |
| "epoch": 19.001478260869565, | |
| "grad_norm": 0.0039555905386805534, | |
| "learning_rate": 4.379227053140097e-05, | |
| "loss": 0.0008, | |
| "step": 4870 | |
| }, | |
| { | |
| "epoch": 19.00191304347826, | |
| "grad_norm": 0.050936609506607056, | |
| "learning_rate": 4.3768115942028986e-05, | |
| "loss": 0.0004, | |
| "step": 4880 | |
| }, | |
| { | |
| "epoch": 19.002347826086957, | |
| "grad_norm": 0.0030154080595821142, | |
| "learning_rate": 4.374396135265701e-05, | |
| "loss": 0.003, | |
| "step": 4890 | |
| }, | |
| { | |
| "epoch": 19.002782608695654, | |
| "grad_norm": 0.003285268321633339, | |
| "learning_rate": 4.371980676328503e-05, | |
| "loss": 0.0207, | |
| "step": 4900 | |
| }, | |
| { | |
| "epoch": 19.003217391304347, | |
| "grad_norm": 0.0029449728317558765, | |
| "learning_rate": 4.3695652173913046e-05, | |
| "loss": 0.0097, | |
| "step": 4910 | |
| }, | |
| { | |
| "epoch": 19.003652173913043, | |
| "grad_norm": 0.003369753947481513, | |
| "learning_rate": 4.3671497584541064e-05, | |
| "loss": 0.0176, | |
| "step": 4920 | |
| }, | |
| { | |
| "epoch": 19.00404347826087, | |
| "eval_accuracy": 0.8934426229508197, | |
| "eval_f1": 0.8942685404457407, | |
| "eval_loss": 0.5749318599700928, | |
| "eval_runtime": 93.0357, | |
| "eval_samples_per_second": 1.311, | |
| "eval_steps_per_second": 0.172, | |
| "step": 4929 | |
| }, | |
| { | |
| "epoch": 20.00004347826087, | |
| "grad_norm": 0.03470902889966965, | |
| "learning_rate": 4.364734299516908e-05, | |
| "loss": 0.018, | |
| "step": 4930 | |
| }, | |
| { | |
| "epoch": 20.000478260869563, | |
| "grad_norm": 0.003587034996598959, | |
| "learning_rate": 4.362318840579711e-05, | |
| "loss": 0.0015, | |
| "step": 4940 | |
| }, | |
| { | |
| "epoch": 20.00091304347826, | |
| "grad_norm": 0.005226669833064079, | |
| "learning_rate": 4.3599033816425124e-05, | |
| "loss": 0.0002, | |
| "step": 4950 | |
| }, | |
| { | |
| "epoch": 20.001347826086956, | |
| "grad_norm": 0.0028250955510884523, | |
| "learning_rate": 4.357487922705314e-05, | |
| "loss": 0.0108, | |
| "step": 4960 | |
| }, | |
| { | |
| "epoch": 20.001782608695652, | |
| "grad_norm": 0.002752603031694889, | |
| "learning_rate": 4.355072463768116e-05, | |
| "loss": 0.0002, | |
| "step": 4970 | |
| }, | |
| { | |
| "epoch": 20.00221739130435, | |
| "grad_norm": 0.003101842477917671, | |
| "learning_rate": 4.352657004830918e-05, | |
| "loss": 0.0122, | |
| "step": 4980 | |
| }, | |
| { | |
| "epoch": 20.002652173913045, | |
| "grad_norm": 1.1627310514450073, | |
| "learning_rate": 4.35024154589372e-05, | |
| "loss": 0.0164, | |
| "step": 4990 | |
| }, | |
| { | |
| "epoch": 20.003086956521738, | |
| "grad_norm": 0.0022625508718192577, | |
| "learning_rate": 4.347826086956522e-05, | |
| "loss": 0.0273, | |
| "step": 5000 | |
| }, | |
| { | |
| "epoch": 20.003521739130434, | |
| "grad_norm": 0.0464153066277504, | |
| "learning_rate": 4.345410628019324e-05, | |
| "loss": 0.0002, | |
| "step": 5010 | |
| }, | |
| { | |
| "epoch": 20.00395652173913, | |
| "grad_norm": 0.0038320214953273535, | |
| "learning_rate": 4.3429951690821256e-05, | |
| "loss": 0.0002, | |
| "step": 5020 | |
| }, | |
| { | |
| "epoch": 20.00404347826087, | |
| "eval_accuracy": 0.9180327868852459, | |
| "eval_f1": 0.9177113694429735, | |
| "eval_loss": 0.4604657292366028, | |
| "eval_runtime": 88.8434, | |
| "eval_samples_per_second": 1.373, | |
| "eval_steps_per_second": 0.18, | |
| "step": 5022 | |
| }, | |
| { | |
| "epoch": 21.000347826086955, | |
| "grad_norm": 41.3768310546875, | |
| "learning_rate": 4.3405797101449274e-05, | |
| "loss": 0.1367, | |
| "step": 5030 | |
| }, | |
| { | |
| "epoch": 21.00078260869565, | |
| "grad_norm": 0.003354653250426054, | |
| "learning_rate": 4.33816425120773e-05, | |
| "loss": 0.0002, | |
| "step": 5040 | |
| }, | |
| { | |
| "epoch": 21.001217391304348, | |
| "grad_norm": 0.0035812542773783207, | |
| "learning_rate": 4.3357487922705316e-05, | |
| "loss": 0.0102, | |
| "step": 5050 | |
| }, | |
| { | |
| "epoch": 21.001652173913044, | |
| "grad_norm": 0.0032342160120606422, | |
| "learning_rate": 4.3333333333333334e-05, | |
| "loss": 0.0198, | |
| "step": 5060 | |
| }, | |
| { | |
| "epoch": 21.00208695652174, | |
| "grad_norm": 0.0021403057035058737, | |
| "learning_rate": 4.330917874396135e-05, | |
| "loss": 0.0005, | |
| "step": 5070 | |
| }, | |
| { | |
| "epoch": 21.002521739130433, | |
| "grad_norm": 1.876338005065918, | |
| "learning_rate": 4.3285024154589376e-05, | |
| "loss": 0.02, | |
| "step": 5080 | |
| }, | |
| { | |
| "epoch": 21.00295652173913, | |
| "grad_norm": 0.0027030061464756727, | |
| "learning_rate": 4.3260869565217394e-05, | |
| "loss": 0.0136, | |
| "step": 5090 | |
| }, | |
| { | |
| "epoch": 21.003391304347826, | |
| "grad_norm": 0.0033732217270880938, | |
| "learning_rate": 4.323671497584541e-05, | |
| "loss": 0.1237, | |
| "step": 5100 | |
| }, | |
| { | |
| "epoch": 21.003826086956522, | |
| "grad_norm": 0.004835965111851692, | |
| "learning_rate": 4.321256038647343e-05, | |
| "loss": 0.0002, | |
| "step": 5110 | |
| }, | |
| { | |
| "epoch": 21.00404347826087, | |
| "eval_accuracy": 0.8852459016393442, | |
| "eval_f1": 0.885472695453409, | |
| "eval_loss": 0.6781102418899536, | |
| "eval_runtime": 84.1219, | |
| "eval_samples_per_second": 1.45, | |
| "eval_steps_per_second": 0.19, | |
| "step": 5115 | |
| }, | |
| { | |
| "epoch": 0.0002173913043478261, | |
| "grad_norm": 0.019052619114518166, | |
| "learning_rate": 4.318840579710145e-05, | |
| "loss": 0.0002, | |
| "step": 5120 | |
| }, | |
| { | |
| "epoch": 0.0006521739130434783, | |
| "grad_norm": 0.008575594052672386, | |
| "learning_rate": 4.316425120772947e-05, | |
| "loss": 0.0003, | |
| "step": 5130 | |
| }, | |
| { | |
| "epoch": 0.0010869565217391304, | |
| "grad_norm": 0.0024844331201165915, | |
| "learning_rate": 4.314009661835749e-05, | |
| "loss": 0.0002, | |
| "step": 5140 | |
| }, | |
| { | |
| "epoch": 0.0015217391304347826, | |
| "grad_norm": 0.003250346751883626, | |
| "learning_rate": 4.3115942028985515e-05, | |
| "loss": 0.005, | |
| "step": 5150 | |
| }, | |
| { | |
| "epoch": 0.001956521739130435, | |
| "grad_norm": 0.016990575939416885, | |
| "learning_rate": 4.3091787439613526e-05, | |
| "loss": 0.0092, | |
| "step": 5160 | |
| }, | |
| { | |
| "epoch": 0.002391304347826087, | |
| "grad_norm": 0.009739736095070839, | |
| "learning_rate": 4.3067632850241544e-05, | |
| "loss": 0.0109, | |
| "step": 5170 | |
| }, | |
| { | |
| "epoch": 0.002826086956521739, | |
| "grad_norm": 0.00987186562269926, | |
| "learning_rate": 4.304347826086957e-05, | |
| "loss": 0.0241, | |
| "step": 5180 | |
| }, | |
| { | |
| "epoch": 0.003260869565217391, | |
| "grad_norm": 0.0024082553572952747, | |
| "learning_rate": 4.3019323671497586e-05, | |
| "loss": 0.0089, | |
| "step": 5190 | |
| }, | |
| { | |
| "epoch": 0.0036956521739130435, | |
| "grad_norm": 0.018611734732985497, | |
| "learning_rate": 4.299516908212561e-05, | |
| "loss": 0.0447, | |
| "step": 5200 | |
| }, | |
| { | |
| "epoch": 0.004043478260869565, | |
| "eval_accuracy": 0.8852459016393442, | |
| "eval_f1": 0.884764867622235, | |
| "eval_loss": 0.8523582816123962, | |
| "eval_runtime": 79.2349, | |
| "eval_samples_per_second": 1.54, | |
| "eval_steps_per_second": 0.202, | |
| "step": 5208 | |
| }, | |
| { | |
| "epoch": 1.0000869565217392, | |
| "grad_norm": 0.0037880963645875454, | |
| "learning_rate": 4.297101449275363e-05, | |
| "loss": 0.0003, | |
| "step": 5210 | |
| }, | |
| { | |
| "epoch": 1.0005217391304349, | |
| "grad_norm": 0.002207667101174593, | |
| "learning_rate": 4.294685990338164e-05, | |
| "loss": 0.0002, | |
| "step": 5220 | |
| }, | |
| { | |
| "epoch": 1.0009565217391305, | |
| "grad_norm": 0.003156916704028845, | |
| "learning_rate": 4.2922705314009664e-05, | |
| "loss": 0.0118, | |
| "step": 5230 | |
| }, | |
| { | |
| "epoch": 1.0013913043478262, | |
| "grad_norm": 0.0022045038640499115, | |
| "learning_rate": 4.289855072463768e-05, | |
| "loss": 0.0133, | |
| "step": 5240 | |
| }, | |
| { | |
| "epoch": 1.0018260869565216, | |
| "grad_norm": 0.00293562444858253, | |
| "learning_rate": 4.2874396135265707e-05, | |
| "loss": 0.0092, | |
| "step": 5250 | |
| }, | |
| { | |
| "epoch": 1.0022608695652173, | |
| "grad_norm": 0.0026238299906253815, | |
| "learning_rate": 4.2850241545893724e-05, | |
| "loss": 0.0002, | |
| "step": 5260 | |
| }, | |
| { | |
| "epoch": 1.002695652173913, | |
| "grad_norm": 0.002277608262374997, | |
| "learning_rate": 4.282608695652174e-05, | |
| "loss": 0.0002, | |
| "step": 5270 | |
| }, | |
| { | |
| "epoch": 1.0031304347826087, | |
| "grad_norm": 0.002632198855280876, | |
| "learning_rate": 4.280193236714976e-05, | |
| "loss": 0.0244, | |
| "step": 5280 | |
| }, | |
| { | |
| "epoch": 1.0035652173913043, | |
| "grad_norm": 0.0037972743157297373, | |
| "learning_rate": 4.277777777777778e-05, | |
| "loss": 0.0152, | |
| "step": 5290 | |
| }, | |
| { | |
| "epoch": 1.004, | |
| "grad_norm": 0.002417921321466565, | |
| "learning_rate": 4.27536231884058e-05, | |
| "loss": 0.0002, | |
| "step": 5300 | |
| }, | |
| { | |
| "epoch": 1.0040434782608696, | |
| "eval_accuracy": 0.9180327868852459, | |
| "eval_f1": 0.9179627294381393, | |
| "eval_loss": 0.7144317030906677, | |
| "eval_runtime": 62.0825, | |
| "eval_samples_per_second": 1.965, | |
| "eval_steps_per_second": 0.258, | |
| "step": 5301 | |
| }, | |
| { | |
| "epoch": 2.000391304347826, | |
| "grad_norm": 0.0018766126595437527, | |
| "learning_rate": 4.272946859903382e-05, | |
| "loss": 0.0002, | |
| "step": 5310 | |
| }, | |
| { | |
| "epoch": 2.0008260869565215, | |
| "grad_norm": 0.006529956124722958, | |
| "learning_rate": 4.270531400966184e-05, | |
| "loss": 0.0178, | |
| "step": 5320 | |
| }, | |
| { | |
| "epoch": 2.0012608695652174, | |
| "grad_norm": 0.002090044552460313, | |
| "learning_rate": 4.2681159420289856e-05, | |
| "loss": 0.0005, | |
| "step": 5330 | |
| }, | |
| { | |
| "epoch": 2.001695652173913, | |
| "grad_norm": 0.004191617481410503, | |
| "learning_rate": 4.2657004830917874e-05, | |
| "loss": 0.0002, | |
| "step": 5340 | |
| }, | |
| { | |
| "epoch": 2.0021304347826088, | |
| "grad_norm": 0.0032148694153875113, | |
| "learning_rate": 4.26328502415459e-05, | |
| "loss": 0.2159, | |
| "step": 5350 | |
| }, | |
| { | |
| "epoch": 2.002565217391304, | |
| "grad_norm": 0.0023662829771637917, | |
| "learning_rate": 4.2608695652173916e-05, | |
| "loss": 0.0307, | |
| "step": 5360 | |
| }, | |
| { | |
| "epoch": 2.003, | |
| "grad_norm": 0.002081876154989004, | |
| "learning_rate": 4.2584541062801934e-05, | |
| "loss": 0.0354, | |
| "step": 5370 | |
| }, | |
| { | |
| "epoch": 2.0034347826086956, | |
| "grad_norm": 0.002491084625944495, | |
| "learning_rate": 4.256038647342995e-05, | |
| "loss": 0.0328, | |
| "step": 5380 | |
| }, | |
| { | |
| "epoch": 2.0038695652173915, | |
| "grad_norm": 0.0021801667753607035, | |
| "learning_rate": 4.2536231884057976e-05, | |
| "loss": 0.0047, | |
| "step": 5390 | |
| }, | |
| { | |
| "epoch": 2.0040434782608694, | |
| "eval_accuracy": 0.8114754098360656, | |
| "eval_f1": 0.8120128926178732, | |
| "eval_loss": 1.1922979354858398, | |
| "eval_runtime": 71.8577, | |
| "eval_samples_per_second": 1.698, | |
| "eval_steps_per_second": 0.223, | |
| "step": 5394 | |
| }, | |
| { | |
| "epoch": 3.0002608695652175, | |
| "grad_norm": 0.004960095044225454, | |
| "learning_rate": 4.2512077294685994e-05, | |
| "loss": 0.2896, | |
| "step": 5400 | |
| }, | |
| { | |
| "epoch": 3.000695652173913, | |
| "grad_norm": 0.2444988638162613, | |
| "learning_rate": 4.248792270531401e-05, | |
| "loss": 0.1418, | |
| "step": 5410 | |
| }, | |
| { | |
| "epoch": 3.001130434782609, | |
| "grad_norm": 0.33257365226745605, | |
| "learning_rate": 4.246376811594203e-05, | |
| "loss": 0.0054, | |
| "step": 5420 | |
| }, | |
| { | |
| "epoch": 3.0015652173913043, | |
| "grad_norm": 0.028887081891298294, | |
| "learning_rate": 4.243961352657005e-05, | |
| "loss": 0.0013, | |
| "step": 5430 | |
| }, | |
| { | |
| "epoch": 3.002, | |
| "grad_norm": 0.00819989200681448, | |
| "learning_rate": 4.241545893719807e-05, | |
| "loss": 0.0003, | |
| "step": 5440 | |
| }, | |
| { | |
| "epoch": 3.0024347826086957, | |
| "grad_norm": 0.005405559204518795, | |
| "learning_rate": 4.239130434782609e-05, | |
| "loss": 0.0362, | |
| "step": 5450 | |
| }, | |
| { | |
| "epoch": 3.002869565217391, | |
| "grad_norm": 0.005664165131747723, | |
| "learning_rate": 4.236714975845411e-05, | |
| "loss": 0.1342, | |
| "step": 5460 | |
| }, | |
| { | |
| "epoch": 3.003304347826087, | |
| "grad_norm": 0.012249620631337166, | |
| "learning_rate": 4.2342995169082126e-05, | |
| "loss": 0.0077, | |
| "step": 5470 | |
| }, | |
| { | |
| "epoch": 3.0037391304347825, | |
| "grad_norm": 0.010283100418746471, | |
| "learning_rate": 4.2318840579710143e-05, | |
| "loss": 0.0298, | |
| "step": 5480 | |
| }, | |
| { | |
| "epoch": 3.0040434782608694, | |
| "eval_accuracy": 0.860655737704918, | |
| "eval_f1": 0.8604635470785563, | |
| "eval_loss": 0.9435864090919495, | |
| "eval_runtime": 62.0517, | |
| "eval_samples_per_second": 1.966, | |
| "eval_steps_per_second": 0.258, | |
| "step": 5487 | |
| }, | |
| { | |
| "epoch": 4.0001304347826085, | |
| "grad_norm": 1.3064463138580322, | |
| "learning_rate": 4.229468599033817e-05, | |
| "loss": 0.0081, | |
| "step": 5490 | |
| }, | |
| { | |
| "epoch": 4.000565217391304, | |
| "grad_norm": 0.008303687907755375, | |
| "learning_rate": 4.2270531400966186e-05, | |
| "loss": 0.0299, | |
| "step": 5500 | |
| }, | |
| { | |
| "epoch": 4.001, | |
| "grad_norm": 0.008156073279678822, | |
| "learning_rate": 4.224637681159421e-05, | |
| "loss": 0.0087, | |
| "step": 5510 | |
| }, | |
| { | |
| "epoch": 4.001434782608696, | |
| "grad_norm": 0.003047351958230138, | |
| "learning_rate": 4.222222222222222e-05, | |
| "loss": 0.0008, | |
| "step": 5520 | |
| }, | |
| { | |
| "epoch": 4.001869565217391, | |
| "grad_norm": 0.0039140209555625916, | |
| "learning_rate": 4.219806763285024e-05, | |
| "loss": 0.0339, | |
| "step": 5530 | |
| }, | |
| { | |
| "epoch": 4.002304347826087, | |
| "grad_norm": 0.015015006996691227, | |
| "learning_rate": 4.2173913043478264e-05, | |
| "loss": 0.0004, | |
| "step": 5540 | |
| }, | |
| { | |
| "epoch": 4.002739130434783, | |
| "grad_norm": 0.00695813400670886, | |
| "learning_rate": 4.214975845410628e-05, | |
| "loss": 0.0869, | |
| "step": 5550 | |
| }, | |
| { | |
| "epoch": 4.0031739130434785, | |
| "grad_norm": 180.4748077392578, | |
| "learning_rate": 4.2125603864734306e-05, | |
| "loss": 0.0392, | |
| "step": 5560 | |
| }, | |
| { | |
| "epoch": 4.003608695652174, | |
| "grad_norm": 0.003250324632972479, | |
| "learning_rate": 4.210144927536232e-05, | |
| "loss": 0.0073, | |
| "step": 5570 | |
| }, | |
| { | |
| "epoch": 4.004043478260869, | |
| "grad_norm": 0.002948058070614934, | |
| "learning_rate": 4.207729468599034e-05, | |
| "loss": 0.0244, | |
| "step": 5580 | |
| }, | |
| { | |
| "epoch": 4.004043478260869, | |
| "eval_accuracy": 0.8770491803278688, | |
| "eval_f1": 0.8768068873079699, | |
| "eval_loss": 0.8491593599319458, | |
| "eval_runtime": 63.8759, | |
| "eval_samples_per_second": 1.91, | |
| "eval_steps_per_second": 0.25, | |
| "step": 5580 | |
| }, | |
| { | |
| "epoch": 5.0004347826086954, | |
| "grad_norm": 1.8423364162445068, | |
| "learning_rate": 4.205314009661836e-05, | |
| "loss": 0.055, | |
| "step": 5590 | |
| }, | |
| { | |
| "epoch": 5.000869565217391, | |
| "grad_norm": 0.005326602142304182, | |
| "learning_rate": 4.202898550724638e-05, | |
| "loss": 0.0004, | |
| "step": 5600 | |
| }, | |
| { | |
| "epoch": 5.001304347826087, | |
| "grad_norm": 0.0026577089447528124, | |
| "learning_rate": 4.20048309178744e-05, | |
| "loss": 0.0002, | |
| "step": 5610 | |
| }, | |
| { | |
| "epoch": 5.001739130434783, | |
| "grad_norm": 0.002887171460315585, | |
| "learning_rate": 4.198067632850241e-05, | |
| "loss": 0.0037, | |
| "step": 5620 | |
| }, | |
| { | |
| "epoch": 5.002173913043478, | |
| "grad_norm": 0.0028708172030746937, | |
| "learning_rate": 4.195652173913044e-05, | |
| "loss": 0.0003, | |
| "step": 5630 | |
| }, | |
| { | |
| "epoch": 5.002608695652174, | |
| "grad_norm": 0.3837863504886627, | |
| "learning_rate": 4.1932367149758456e-05, | |
| "loss": 0.0188, | |
| "step": 5640 | |
| }, | |
| { | |
| "epoch": 5.00304347826087, | |
| "grad_norm": 0.0025473798159509897, | |
| "learning_rate": 4.1908212560386474e-05, | |
| "loss": 0.0156, | |
| "step": 5650 | |
| }, | |
| { | |
| "epoch": 5.003478260869565, | |
| "grad_norm": 0.001516752177849412, | |
| "learning_rate": 4.18840579710145e-05, | |
| "loss": 0.0002, | |
| "step": 5660 | |
| }, | |
| { | |
| "epoch": 5.003913043478261, | |
| "grad_norm": 0.0037251082248985767, | |
| "learning_rate": 4.185990338164251e-05, | |
| "loss": 0.1255, | |
| "step": 5670 | |
| }, | |
| { | |
| "epoch": 5.004043478260869, | |
| "eval_accuracy": 0.8934426229508197, | |
| "eval_f1": 0.893593266973388, | |
| "eval_loss": 0.8068659901618958, | |
| "eval_runtime": 71.6747, | |
| "eval_samples_per_second": 1.702, | |
| "eval_steps_per_second": 0.223, | |
| "step": 5673 | |
| }, | |
| { | |
| "epoch": 6.000304347826087, | |
| "grad_norm": 0.0021381359547376633, | |
| "learning_rate": 4.1835748792270534e-05, | |
| "loss": 0.0126, | |
| "step": 5680 | |
| }, | |
| { | |
| "epoch": 6.000739130434782, | |
| "grad_norm": 0.0025412938557565212, | |
| "learning_rate": 4.181159420289855e-05, | |
| "loss": 0.0066, | |
| "step": 5690 | |
| }, | |
| { | |
| "epoch": 6.001173913043479, | |
| "grad_norm": 1.9342097043991089, | |
| "learning_rate": 4.1787439613526576e-05, | |
| "loss": 0.0237, | |
| "step": 5700 | |
| }, | |
| { | |
| "epoch": 6.001608695652174, | |
| "grad_norm": 0.003369554877281189, | |
| "learning_rate": 4.1763285024154594e-05, | |
| "loss": 0.0775, | |
| "step": 5710 | |
| }, | |
| { | |
| "epoch": 6.00204347826087, | |
| "grad_norm": 0.0038899381179362535, | |
| "learning_rate": 4.1739130434782605e-05, | |
| "loss": 0.0123, | |
| "step": 5720 | |
| }, | |
| { | |
| "epoch": 6.002478260869565, | |
| "grad_norm": 0.002373365219682455, | |
| "learning_rate": 4.171497584541063e-05, | |
| "loss": 0.0174, | |
| "step": 5730 | |
| }, | |
| { | |
| "epoch": 6.0029130434782605, | |
| "grad_norm": 0.03260812163352966, | |
| "learning_rate": 4.169082125603865e-05, | |
| "loss": 0.0002, | |
| "step": 5740 | |
| }, | |
| { | |
| "epoch": 6.003347826086957, | |
| "grad_norm": 39.89988708496094, | |
| "learning_rate": 4.166666666666667e-05, | |
| "loss": 0.021, | |
| "step": 5750 | |
| }, | |
| { | |
| "epoch": 6.003782608695652, | |
| "grad_norm": 0.002141584875062108, | |
| "learning_rate": 4.164251207729469e-05, | |
| "loss": 0.0002, | |
| "step": 5760 | |
| }, | |
| { | |
| "epoch": 6.004043478260869, | |
| "eval_accuracy": 0.8688524590163934, | |
| "eval_f1": 0.8688620747444277, | |
| "eval_loss": 0.9789605736732483, | |
| "eval_runtime": 61.9663, | |
| "eval_samples_per_second": 1.969, | |
| "eval_steps_per_second": 0.258, | |
| "step": 5766 | |
| }, | |
| { | |
| "epoch": 7.000173913043478, | |
| "grad_norm": 0.0036677520256489515, | |
| "learning_rate": 4.161835748792271e-05, | |
| "loss": 0.0002, | |
| "step": 5770 | |
| }, | |
| { | |
| "epoch": 7.000608695652174, | |
| "grad_norm": 0.0025961585342884064, | |
| "learning_rate": 4.1594202898550726e-05, | |
| "loss": 0.0003, | |
| "step": 5780 | |
| }, | |
| { | |
| "epoch": 7.001043478260869, | |
| "grad_norm": 0.003315119771286845, | |
| "learning_rate": 4.157004830917874e-05, | |
| "loss": 0.0002, | |
| "step": 5790 | |
| }, | |
| { | |
| "epoch": 7.001478260869566, | |
| "grad_norm": 0.015529535710811615, | |
| "learning_rate": 4.154589371980677e-05, | |
| "loss": 0.0002, | |
| "step": 5800 | |
| }, | |
| { | |
| "epoch": 7.001913043478261, | |
| "grad_norm": 0.0024377082008868456, | |
| "learning_rate": 4.1521739130434786e-05, | |
| "loss": 0.0154, | |
| "step": 5810 | |
| }, | |
| { | |
| "epoch": 7.0023478260869565, | |
| "grad_norm": 0.002965396037325263, | |
| "learning_rate": 4.1497584541062804e-05, | |
| "loss": 0.0002, | |
| "step": 5820 | |
| }, | |
| { | |
| "epoch": 7.002782608695652, | |
| "grad_norm": 0.0019290961790829897, | |
| "learning_rate": 4.147342995169082e-05, | |
| "loss": 0.0131, | |
| "step": 5830 | |
| }, | |
| { | |
| "epoch": 7.003217391304347, | |
| "grad_norm": 0.0026897243224084377, | |
| "learning_rate": 4.144927536231884e-05, | |
| "loss": 0.0002, | |
| "step": 5840 | |
| }, | |
| { | |
| "epoch": 7.003652173913044, | |
| "grad_norm": 0.0020241059828549623, | |
| "learning_rate": 4.1425120772946864e-05, | |
| "loss": 0.0225, | |
| "step": 5850 | |
| }, | |
| { | |
| "epoch": 7.004043478260869, | |
| "eval_accuracy": 0.8934426229508197, | |
| "eval_f1": 0.8935846750223277, | |
| "eval_loss": 0.83548903465271, | |
| "eval_runtime": 69.454, | |
| "eval_samples_per_second": 1.757, | |
| "eval_steps_per_second": 0.23, | |
| "step": 5859 | |
| }, | |
| { | |
| "epoch": 8.000043478260869, | |
| "grad_norm": 0.001890950370579958, | |
| "learning_rate": 4.140096618357488e-05, | |
| "loss": 0.1447, | |
| "step": 5860 | |
| }, | |
| { | |
| "epoch": 8.000478260869565, | |
| "grad_norm": 0.006349007599055767, | |
| "learning_rate": 4.13768115942029e-05, | |
| "loss": 0.1237, | |
| "step": 5870 | |
| }, | |
| { | |
| "epoch": 8.000913043478262, | |
| "grad_norm": 1.3416682481765747, | |
| "learning_rate": 4.135265700483092e-05, | |
| "loss": 0.0205, | |
| "step": 5880 | |
| }, | |
| { | |
| "epoch": 8.001347826086956, | |
| "grad_norm": 0.01573265716433525, | |
| "learning_rate": 4.132850241545894e-05, | |
| "loss": 0.0002, | |
| "step": 5890 | |
| }, | |
| { | |
| "epoch": 8.001782608695652, | |
| "grad_norm": 0.9573756456375122, | |
| "learning_rate": 4.130434782608696e-05, | |
| "loss": 0.0054, | |
| "step": 5900 | |
| }, | |
| { | |
| "epoch": 8.002217391304347, | |
| "grad_norm": 0.002642852021381259, | |
| "learning_rate": 4.128019323671498e-05, | |
| "loss": 0.0002, | |
| "step": 5910 | |
| }, | |
| { | |
| "epoch": 8.002652173913043, | |
| "grad_norm": 0.03945804014801979, | |
| "learning_rate": 4.1256038647342995e-05, | |
| "loss": 0.0368, | |
| "step": 5920 | |
| }, | |
| { | |
| "epoch": 8.00308695652174, | |
| "grad_norm": 0.0025761283468455076, | |
| "learning_rate": 4.123188405797101e-05, | |
| "loss": 0.0001, | |
| "step": 5930 | |
| }, | |
| { | |
| "epoch": 8.003521739130434, | |
| "grad_norm": 0.004890764597803354, | |
| "learning_rate": 4.120772946859904e-05, | |
| "loss": 0.0002, | |
| "step": 5940 | |
| }, | |
| { | |
| "epoch": 8.00395652173913, | |
| "grad_norm": 0.0019304666202515364, | |
| "learning_rate": 4.1183574879227056e-05, | |
| "loss": 0.0002, | |
| "step": 5950 | |
| }, | |
| { | |
| "epoch": 8.00404347826087, | |
| "eval_accuracy": 0.8278688524590164, | |
| "eval_f1": 0.827731349212412, | |
| "eval_loss": 1.2655301094055176, | |
| "eval_runtime": 71.2771, | |
| "eval_samples_per_second": 1.712, | |
| "eval_steps_per_second": 0.224, | |
| "step": 5952 | |
| }, | |
| { | |
| "epoch": 9.000347826086957, | |
| "grad_norm": 0.0022518360055983067, | |
| "learning_rate": 4.115942028985507e-05, | |
| "loss": 0.0055, | |
| "step": 5960 | |
| }, | |
| { | |
| "epoch": 9.000782608695653, | |
| "grad_norm": 0.002156842267140746, | |
| "learning_rate": 4.113526570048309e-05, | |
| "loss": 0.0135, | |
| "step": 5970 | |
| }, | |
| { | |
| "epoch": 9.001217391304348, | |
| "grad_norm": 0.006422512233257294, | |
| "learning_rate": 4.111111111111111e-05, | |
| "loss": 0.0061, | |
| "step": 5980 | |
| }, | |
| { | |
| "epoch": 9.001652173913044, | |
| "grad_norm": 0.0017242766916751862, | |
| "learning_rate": 4.1086956521739134e-05, | |
| "loss": 0.0002, | |
| "step": 5990 | |
| }, | |
| { | |
| "epoch": 9.002086956521739, | |
| "grad_norm": 0.0017920706886798143, | |
| "learning_rate": 4.106280193236715e-05, | |
| "loss": 0.0001, | |
| "step": 6000 | |
| }, | |
| { | |
| "epoch": 9.002521739130435, | |
| "grad_norm": 0.0021232902072370052, | |
| "learning_rate": 4.1038647342995176e-05, | |
| "loss": 0.0001, | |
| "step": 6010 | |
| }, | |
| { | |
| "epoch": 9.002956521739131, | |
| "grad_norm": 0.001960981637239456, | |
| "learning_rate": 4.101449275362319e-05, | |
| "loss": 0.0002, | |
| "step": 6020 | |
| }, | |
| { | |
| "epoch": 9.003391304347826, | |
| "grad_norm": 0.0018294297624379396, | |
| "learning_rate": 4.0990338164251205e-05, | |
| "loss": 0.0181, | |
| "step": 6030 | |
| }, | |
| { | |
| "epoch": 9.003826086956522, | |
| "grad_norm": 0.002742977347224951, | |
| "learning_rate": 4.096618357487923e-05, | |
| "loss": 0.0153, | |
| "step": 6040 | |
| }, | |
| { | |
| "epoch": 9.00404347826087, | |
| "eval_accuracy": 0.8278688524590164, | |
| "eval_f1": 0.827731349212412, | |
| "eval_loss": 1.2649469375610352, | |
| "eval_runtime": 75.5533, | |
| "eval_samples_per_second": 1.615, | |
| "eval_steps_per_second": 0.212, | |
| "step": 6045 | |
| }, | |
| { | |
| "epoch": 10.000217391304348, | |
| "grad_norm": 0.001949600875377655, | |
| "learning_rate": 4.094202898550725e-05, | |
| "loss": 0.0073, | |
| "step": 6050 | |
| }, | |
| { | |
| "epoch": 10.000652173913043, | |
| "grad_norm": 0.0014879930531606078, | |
| "learning_rate": 4.091787439613527e-05, | |
| "loss": 0.0192, | |
| "step": 6060 | |
| }, | |
| { | |
| "epoch": 10.001086956521739, | |
| "grad_norm": 0.0032951009925454855, | |
| "learning_rate": 4.089371980676329e-05, | |
| "loss": 0.0266, | |
| "step": 6070 | |
| }, | |
| { | |
| "epoch": 10.001521739130435, | |
| "grad_norm": 0.0022343776654452085, | |
| "learning_rate": 4.086956521739131e-05, | |
| "loss": 0.0002, | |
| "step": 6080 | |
| }, | |
| { | |
| "epoch": 10.00195652173913, | |
| "grad_norm": 0.0020602564327418804, | |
| "learning_rate": 4.0845410628019325e-05, | |
| "loss": 0.0175, | |
| "step": 6090 | |
| }, | |
| { | |
| "epoch": 10.002391304347826, | |
| "grad_norm": 0.001147582079283893, | |
| "learning_rate": 4.082125603864734e-05, | |
| "loss": 0.0001, | |
| "step": 6100 | |
| }, | |
| { | |
| "epoch": 10.00282608695652, | |
| "grad_norm": 0.00242516677826643, | |
| "learning_rate": 4.079710144927537e-05, | |
| "loss": 0.0002, | |
| "step": 6110 | |
| }, | |
| { | |
| "epoch": 10.003260869565217, | |
| "grad_norm": 0.24815811216831207, | |
| "learning_rate": 4.0772946859903386e-05, | |
| "loss": 0.0488, | |
| "step": 6120 | |
| }, | |
| { | |
| "epoch": 10.003695652173914, | |
| "grad_norm": 0.0025010844692587852, | |
| "learning_rate": 4.0748792270531403e-05, | |
| "loss": 0.0002, | |
| "step": 6130 | |
| }, | |
| { | |
| "epoch": 10.00404347826087, | |
| "eval_accuracy": 0.8524590163934426, | |
| "eval_f1": 0.8526902679666354, | |
| "eval_loss": 1.144180417060852, | |
| "eval_runtime": 72.1738, | |
| "eval_samples_per_second": 1.69, | |
| "eval_steps_per_second": 0.222, | |
| "step": 6138 | |
| }, | |
| { | |
| "epoch": 11.00008695652174, | |
| "grad_norm": 1.7668468952178955, | |
| "learning_rate": 4.072463768115942e-05, | |
| "loss": 0.1222, | |
| "step": 6140 | |
| }, | |
| { | |
| "epoch": 11.000521739130434, | |
| "grad_norm": 32.476314544677734, | |
| "learning_rate": 4.070048309178744e-05, | |
| "loss": 0.0861, | |
| "step": 6150 | |
| }, | |
| { | |
| "epoch": 11.00095652173913, | |
| "grad_norm": 0.011918846517801285, | |
| "learning_rate": 4.0676328502415464e-05, | |
| "loss": 0.0074, | |
| "step": 6160 | |
| }, | |
| { | |
| "epoch": 11.001391304347827, | |
| "grad_norm": 0.0021514969412237406, | |
| "learning_rate": 4.065217391304348e-05, | |
| "loss": 0.1079, | |
| "step": 6170 | |
| }, | |
| { | |
| "epoch": 11.001826086956521, | |
| "grad_norm": 0.006829989142715931, | |
| "learning_rate": 4.06280193236715e-05, | |
| "loss": 0.0009, | |
| "step": 6180 | |
| }, | |
| { | |
| "epoch": 11.002260869565218, | |
| "grad_norm": 0.00387940788641572, | |
| "learning_rate": 4.060386473429952e-05, | |
| "loss": 0.0529, | |
| "step": 6190 | |
| }, | |
| { | |
| "epoch": 11.002695652173912, | |
| "grad_norm": 0.001859953161329031, | |
| "learning_rate": 4.057971014492754e-05, | |
| "loss": 0.1167, | |
| "step": 6200 | |
| }, | |
| { | |
| "epoch": 11.003130434782609, | |
| "grad_norm": 5.892602443695068, | |
| "learning_rate": 4.055555555555556e-05, | |
| "loss": 0.0117, | |
| "step": 6210 | |
| }, | |
| { | |
| "epoch": 11.003565217391305, | |
| "grad_norm": 0.0074269757606089115, | |
| "learning_rate": 4.053140096618358e-05, | |
| "loss": 0.0002, | |
| "step": 6220 | |
| }, | |
| { | |
| "epoch": 11.004, | |
| "grad_norm": 33.7471809387207, | |
| "learning_rate": 4.0507246376811595e-05, | |
| "loss": 0.1902, | |
| "step": 6230 | |
| }, | |
| { | |
| "epoch": 11.00404347826087, | |
| "eval_accuracy": 0.8524590163934426, | |
| "eval_f1": 0.8515994026118366, | |
| "eval_loss": 1.0411537885665894, | |
| "eval_runtime": 66.8169, | |
| "eval_samples_per_second": 1.826, | |
| "eval_steps_per_second": 0.239, | |
| "step": 6231 | |
| }, | |
| { | |
| "epoch": 12.000391304347826, | |
| "grad_norm": 0.05204038694500923, | |
| "learning_rate": 4.048309178743961e-05, | |
| "loss": 0.0344, | |
| "step": 6240 | |
| }, | |
| { | |
| "epoch": 12.000826086956522, | |
| "grad_norm": 1.3173763751983643, | |
| "learning_rate": 4.045893719806764e-05, | |
| "loss": 0.1976, | |
| "step": 6250 | |
| }, | |
| { | |
| "epoch": 12.001260869565217, | |
| "grad_norm": 0.0014710385585203767, | |
| "learning_rate": 4.0434782608695655e-05, | |
| "loss": 0.0094, | |
| "step": 6260 | |
| }, | |
| { | |
| "epoch": 12.001695652173913, | |
| "grad_norm": 0.0017820164794102311, | |
| "learning_rate": 4.041062801932367e-05, | |
| "loss": 0.0094, | |
| "step": 6270 | |
| }, | |
| { | |
| "epoch": 12.00213043478261, | |
| "grad_norm": 16.647350311279297, | |
| "learning_rate": 4.038647342995169e-05, | |
| "loss": 0.1017, | |
| "step": 6280 | |
| }, | |
| { | |
| "epoch": 12.002565217391304, | |
| "grad_norm": 9.301661491394043, | |
| "learning_rate": 4.036231884057971e-05, | |
| "loss": 0.1693, | |
| "step": 6290 | |
| }, | |
| { | |
| "epoch": 12.003, | |
| "grad_norm": 16.365354537963867, | |
| "learning_rate": 4.0338164251207733e-05, | |
| "loss": 0.1188, | |
| "step": 6300 | |
| }, | |
| { | |
| "epoch": 12.003434782608696, | |
| "grad_norm": 8.132902145385742, | |
| "learning_rate": 4.031400966183575e-05, | |
| "loss": 0.0691, | |
| "step": 6310 | |
| }, | |
| { | |
| "epoch": 12.003869565217391, | |
| "grad_norm": 0.017481571063399315, | |
| "learning_rate": 4.028985507246377e-05, | |
| "loss": 0.0622, | |
| "step": 6320 | |
| }, | |
| { | |
| "epoch": 12.00404347826087, | |
| "eval_accuracy": 0.8114754098360656, | |
| "eval_f1": 0.8078045336889167, | |
| "eval_loss": 1.100170612335205, | |
| "eval_runtime": 80.9477, | |
| "eval_samples_per_second": 1.507, | |
| "eval_steps_per_second": 0.198, | |
| "step": 6324 | |
| }, | |
| { | |
| "epoch": 13.000260869565217, | |
| "grad_norm": 62.808135986328125, | |
| "learning_rate": 4.026570048309179e-05, | |
| "loss": 0.0311, | |
| "step": 6330 | |
| }, | |
| { | |
| "epoch": 13.000695652173913, | |
| "grad_norm": 0.0037931206170469522, | |
| "learning_rate": 4.0241545893719805e-05, | |
| "loss": 0.0109, | |
| "step": 6340 | |
| }, | |
| { | |
| "epoch": 13.001130434782608, | |
| "grad_norm": 27.289682388305664, | |
| "learning_rate": 4.021739130434783e-05, | |
| "loss": 0.0662, | |
| "step": 6350 | |
| }, | |
| { | |
| "epoch": 13.001565217391304, | |
| "grad_norm": 0.018042033538222313, | |
| "learning_rate": 4.019323671497585e-05, | |
| "loss": 0.0022, | |
| "step": 6360 | |
| }, | |
| { | |
| "epoch": 13.002, | |
| "grad_norm": 0.010825947858393192, | |
| "learning_rate": 4.016908212560387e-05, | |
| "loss": 0.0002, | |
| "step": 6370 | |
| }, | |
| { | |
| "epoch": 13.002434782608695, | |
| "grad_norm": 61.149078369140625, | |
| "learning_rate": 4.014492753623188e-05, | |
| "loss": 0.2017, | |
| "step": 6380 | |
| }, | |
| { | |
| "epoch": 13.002869565217392, | |
| "grad_norm": 0.00819044467061758, | |
| "learning_rate": 4.012077294685991e-05, | |
| "loss": 0.0005, | |
| "step": 6390 | |
| }, | |
| { | |
| "epoch": 13.003304347826086, | |
| "grad_norm": 7.115654468536377, | |
| "learning_rate": 4.0096618357487925e-05, | |
| "loss": 0.0981, | |
| "step": 6400 | |
| }, | |
| { | |
| "epoch": 13.003739130434782, | |
| "grad_norm": 0.06668082624673843, | |
| "learning_rate": 4.007246376811594e-05, | |
| "loss": 0.1367, | |
| "step": 6410 | |
| }, | |
| { | |
| "epoch": 13.00404347826087, | |
| "eval_accuracy": 0.8442622950819673, | |
| "eval_f1": 0.8402451615360002, | |
| "eval_loss": 1.031615138053894, | |
| "eval_runtime": 71.7599, | |
| "eval_samples_per_second": 1.7, | |
| "eval_steps_per_second": 0.223, | |
| "step": 6417 | |
| }, | |
| { | |
| "epoch": 14.000130434782609, | |
| "grad_norm": 1.688230037689209, | |
| "learning_rate": 4.004830917874397e-05, | |
| "loss": 0.1085, | |
| "step": 6420 | |
| }, | |
| { | |
| "epoch": 14.000565217391305, | |
| "grad_norm": 0.002879357198253274, | |
| "learning_rate": 4.002415458937198e-05, | |
| "loss": 0.0148, | |
| "step": 6430 | |
| }, | |
| { | |
| "epoch": 14.001, | |
| "grad_norm": 0.005582920741289854, | |
| "learning_rate": 4e-05, | |
| "loss": 0.0532, | |
| "step": 6440 | |
| }, | |
| { | |
| "epoch": 14.001434782608696, | |
| "grad_norm": 0.01366361603140831, | |
| "learning_rate": 3.997584541062802e-05, | |
| "loss": 0.0036, | |
| "step": 6450 | |
| }, | |
| { | |
| "epoch": 14.001869565217392, | |
| "grad_norm": 0.009635832160711288, | |
| "learning_rate": 3.995169082125604e-05, | |
| "loss": 0.0027, | |
| "step": 6460 | |
| }, | |
| { | |
| "epoch": 14.002304347826087, | |
| "grad_norm": 6.955427646636963, | |
| "learning_rate": 3.9927536231884064e-05, | |
| "loss": 0.3633, | |
| "step": 6470 | |
| }, | |
| { | |
| "epoch": 14.002739130434783, | |
| "grad_norm": 25.112009048461914, | |
| "learning_rate": 3.9903381642512075e-05, | |
| "loss": 0.161, | |
| "step": 6480 | |
| }, | |
| { | |
| "epoch": 14.003173913043478, | |
| "grad_norm": 0.009927201084792614, | |
| "learning_rate": 3.98792270531401e-05, | |
| "loss": 0.0014, | |
| "step": 6490 | |
| }, | |
| { | |
| "epoch": 14.003608695652174, | |
| "grad_norm": 0.014570885337889194, | |
| "learning_rate": 3.985507246376812e-05, | |
| "loss": 0.02, | |
| "step": 6500 | |
| }, | |
| { | |
| "epoch": 14.00404347826087, | |
| "grad_norm": 1.4303860664367676, | |
| "learning_rate": 3.983091787439614e-05, | |
| "loss": 0.0165, | |
| "step": 6510 | |
| }, | |
| { | |
| "epoch": 14.00404347826087, | |
| "eval_accuracy": 0.819672131147541, | |
| "eval_f1": 0.82006315755217, | |
| "eval_loss": 1.0104899406433105, | |
| "eval_runtime": 79.2274, | |
| "eval_samples_per_second": 1.54, | |
| "eval_steps_per_second": 0.202, | |
| "step": 6510 | |
| }, | |
| { | |
| "epoch": 15.000434782608696, | |
| "grad_norm": 0.014703233726322651, | |
| "learning_rate": 3.980676328502416e-05, | |
| "loss": 0.1467, | |
| "step": 6520 | |
| }, | |
| { | |
| "epoch": 15.000869565217391, | |
| "grad_norm": 0.021894462406635284, | |
| "learning_rate": 3.978260869565217e-05, | |
| "loss": 0.1916, | |
| "step": 6530 | |
| }, | |
| { | |
| "epoch": 15.001304347826087, | |
| "grad_norm": 48.06509780883789, | |
| "learning_rate": 3.9758454106280195e-05, | |
| "loss": 0.1599, | |
| "step": 6540 | |
| }, | |
| { | |
| "epoch": 15.001739130434782, | |
| "grad_norm": 26.76535987854004, | |
| "learning_rate": 3.973429951690821e-05, | |
| "loss": 0.5953, | |
| "step": 6550 | |
| }, | |
| { | |
| "epoch": 15.002173913043478, | |
| "grad_norm": 9.520294189453125, | |
| "learning_rate": 3.971014492753624e-05, | |
| "loss": 0.1125, | |
| "step": 6560 | |
| }, | |
| { | |
| "epoch": 15.002608695652174, | |
| "grad_norm": 1.8650518655776978, | |
| "learning_rate": 3.9685990338164255e-05, | |
| "loss": 0.0568, | |
| "step": 6570 | |
| }, | |
| { | |
| "epoch": 15.003043478260869, | |
| "grad_norm": 0.007623857352882624, | |
| "learning_rate": 3.966183574879227e-05, | |
| "loss": 0.0069, | |
| "step": 6580 | |
| }, | |
| { | |
| "epoch": 15.003478260869565, | |
| "grad_norm": 0.0030073132365942, | |
| "learning_rate": 3.963768115942029e-05, | |
| "loss": 0.0094, | |
| "step": 6590 | |
| }, | |
| { | |
| "epoch": 15.003913043478262, | |
| "grad_norm": 32.13124084472656, | |
| "learning_rate": 3.961352657004831e-05, | |
| "loss": 0.166, | |
| "step": 6600 | |
| }, | |
| { | |
| "epoch": 15.00404347826087, | |
| "eval_accuracy": 0.8688524590163934, | |
| "eval_f1": 0.8678181563427466, | |
| "eval_loss": 0.8523256778717041, | |
| "eval_runtime": 75.8405, | |
| "eval_samples_per_second": 1.609, | |
| "eval_steps_per_second": 0.211, | |
| "step": 6603 | |
| }, | |
| { | |
| "epoch": 16.000304347826088, | |
| "grad_norm": 0.024344058707356453, | |
| "learning_rate": 3.958937198067633e-05, | |
| "loss": 0.0005, | |
| "step": 6610 | |
| }, | |
| { | |
| "epoch": 16.000739130434784, | |
| "grad_norm": 0.008748427033424377, | |
| "learning_rate": 3.956521739130435e-05, | |
| "loss": 0.2414, | |
| "step": 6620 | |
| }, | |
| { | |
| "epoch": 16.001173913043477, | |
| "grad_norm": 26.980012893676758, | |
| "learning_rate": 3.954106280193237e-05, | |
| "loss": 0.1468, | |
| "step": 6630 | |
| }, | |
| { | |
| "epoch": 16.001608695652173, | |
| "grad_norm": 0.027768012136220932, | |
| "learning_rate": 3.951690821256039e-05, | |
| "loss": 0.001, | |
| "step": 6640 | |
| }, | |
| { | |
| "epoch": 16.00204347826087, | |
| "grad_norm": 0.05727599933743477, | |
| "learning_rate": 3.9492753623188405e-05, | |
| "loss": 0.0021, | |
| "step": 6650 | |
| }, | |
| { | |
| "epoch": 16.002478260869566, | |
| "grad_norm": 0.025578580796718597, | |
| "learning_rate": 3.946859903381643e-05, | |
| "loss": 0.0005, | |
| "step": 6660 | |
| }, | |
| { | |
| "epoch": 16.002913043478262, | |
| "grad_norm": 34.755123138427734, | |
| "learning_rate": 3.944444444444445e-05, | |
| "loss": 0.0842, | |
| "step": 6670 | |
| }, | |
| { | |
| "epoch": 16.003347826086955, | |
| "grad_norm": 0.7231902480125427, | |
| "learning_rate": 3.9420289855072465e-05, | |
| "loss": 0.0845, | |
| "step": 6680 | |
| }, | |
| { | |
| "epoch": 16.00378260869565, | |
| "grad_norm": 0.006564811337739229, | |
| "learning_rate": 3.939613526570048e-05, | |
| "loss": 0.0898, | |
| "step": 6690 | |
| }, | |
| { | |
| "epoch": 16.00404347826087, | |
| "eval_accuracy": 0.9016393442622951, | |
| "eval_f1": 0.9015288061359007, | |
| "eval_loss": 0.6139894127845764, | |
| "eval_runtime": 69.4275, | |
| "eval_samples_per_second": 1.757, | |
| "eval_steps_per_second": 0.23, | |
| "step": 6696 | |
| }, | |
| { | |
| "epoch": 17.00017391304348, | |
| "grad_norm": 0.0037501673214137554, | |
| "learning_rate": 3.937198067632851e-05, | |
| "loss": 0.0117, | |
| "step": 6700 | |
| }, | |
| { | |
| "epoch": 17.000608695652176, | |
| "grad_norm": 0.00506971450522542, | |
| "learning_rate": 3.9347826086956525e-05, | |
| "loss": 0.0186, | |
| "step": 6710 | |
| }, | |
| { | |
| "epoch": 17.00104347826087, | |
| "grad_norm": 0.003616392845287919, | |
| "learning_rate": 3.932367149758454e-05, | |
| "loss": 0.0002, | |
| "step": 6720 | |
| }, | |
| { | |
| "epoch": 17.001478260869565, | |
| "grad_norm": 0.002944325562566519, | |
| "learning_rate": 3.929951690821256e-05, | |
| "loss": 0.0869, | |
| "step": 6730 | |
| }, | |
| { | |
| "epoch": 17.00191304347826, | |
| "grad_norm": 0.018326986581087112, | |
| "learning_rate": 3.927536231884058e-05, | |
| "loss": 0.0846, | |
| "step": 6740 | |
| }, | |
| { | |
| "epoch": 17.002347826086957, | |
| "grad_norm": 28.93506622314453, | |
| "learning_rate": 3.92512077294686e-05, | |
| "loss": 0.0637, | |
| "step": 6750 | |
| }, | |
| { | |
| "epoch": 17.002782608695654, | |
| "grad_norm": 0.0321519635617733, | |
| "learning_rate": 3.922705314009662e-05, | |
| "loss": 0.103, | |
| "step": 6760 | |
| }, | |
| { | |
| "epoch": 17.003217391304347, | |
| "grad_norm": 0.00872319471091032, | |
| "learning_rate": 3.920289855072464e-05, | |
| "loss": 0.1741, | |
| "step": 6770 | |
| }, | |
| { | |
| "epoch": 17.003652173913043, | |
| "grad_norm": 0.005549349822103977, | |
| "learning_rate": 3.9178743961352657e-05, | |
| "loss": 0.0384, | |
| "step": 6780 | |
| }, | |
| { | |
| "epoch": 17.00404347826087, | |
| "eval_accuracy": 0.8852459016393442, | |
| "eval_f1": 0.8832644966100266, | |
| "eval_loss": 0.6065811514854431, | |
| "eval_runtime": 73.4098, | |
| "eval_samples_per_second": 1.662, | |
| "eval_steps_per_second": 0.218, | |
| "step": 6789 | |
| }, | |
| { | |
| "epoch": 18.00004347826087, | |
| "grad_norm": 0.0036791262682527304, | |
| "learning_rate": 3.9154589371980674e-05, | |
| "loss": 0.0881, | |
| "step": 6790 | |
| }, | |
| { | |
| "epoch": 18.000478260869563, | |
| "grad_norm": 0.02601425163447857, | |
| "learning_rate": 3.91304347826087e-05, | |
| "loss": 0.0205, | |
| "step": 6800 | |
| }, | |
| { | |
| "epoch": 18.00091304347826, | |
| "grad_norm": 1.111623764038086, | |
| "learning_rate": 3.910628019323672e-05, | |
| "loss": 0.0746, | |
| "step": 6810 | |
| }, | |
| { | |
| "epoch": 18.001347826086956, | |
| "grad_norm": 0.08039382100105286, | |
| "learning_rate": 3.908212560386474e-05, | |
| "loss": 0.0017, | |
| "step": 6820 | |
| }, | |
| { | |
| "epoch": 18.001782608695652, | |
| "grad_norm": 0.0041095721535384655, | |
| "learning_rate": 3.905797101449275e-05, | |
| "loss": 0.0002, | |
| "step": 6830 | |
| }, | |
| { | |
| "epoch": 18.00221739130435, | |
| "grad_norm": 0.02693447656929493, | |
| "learning_rate": 3.903381642512077e-05, | |
| "loss": 0.0433, | |
| "step": 6840 | |
| }, | |
| { | |
| "epoch": 18.002652173913045, | |
| "grad_norm": 14.348990440368652, | |
| "learning_rate": 3.9009661835748795e-05, | |
| "loss": 0.0977, | |
| "step": 6850 | |
| }, | |
| { | |
| "epoch": 18.003086956521738, | |
| "grad_norm": 0.0028455532155930996, | |
| "learning_rate": 3.898550724637681e-05, | |
| "loss": 0.0241, | |
| "step": 6860 | |
| }, | |
| { | |
| "epoch": 18.003521739130434, | |
| "grad_norm": 0.03243542090058327, | |
| "learning_rate": 3.896135265700484e-05, | |
| "loss": 0.0003, | |
| "step": 6870 | |
| }, | |
| { | |
| "epoch": 18.00395652173913, | |
| "grad_norm": 0.002338533289730549, | |
| "learning_rate": 3.8937198067632855e-05, | |
| "loss": 0.0206, | |
| "step": 6880 | |
| }, | |
| { | |
| "epoch": 18.00404347826087, | |
| "eval_accuracy": 0.9098360655737705, | |
| "eval_f1": 0.9110839096946348, | |
| "eval_loss": 0.5883862972259521, | |
| "eval_runtime": 78.6246, | |
| "eval_samples_per_second": 1.552, | |
| "eval_steps_per_second": 0.203, | |
| "step": 6882 | |
| }, | |
| { | |
| "epoch": 0.00034782608695652176, | |
| "grad_norm": 0.007827730849385262, | |
| "learning_rate": 3.8913043478260866e-05, | |
| "loss": 0.0003, | |
| "step": 6890 | |
| }, | |
| { | |
| "epoch": 0.000782608695652174, | |
| "grad_norm": 0.004137350246310234, | |
| "learning_rate": 3.888888888888889e-05, | |
| "loss": 0.0008, | |
| "step": 6900 | |
| }, | |
| { | |
| "epoch": 0.001217391304347826, | |
| "grad_norm": 0.0026433924213051796, | |
| "learning_rate": 3.886473429951691e-05, | |
| "loss": 0.0002, | |
| "step": 6910 | |
| }, | |
| { | |
| "epoch": 0.0016521739130434783, | |
| "grad_norm": 0.001748176058754325, | |
| "learning_rate": 3.884057971014493e-05, | |
| "loss": 0.0018, | |
| "step": 6920 | |
| }, | |
| { | |
| "epoch": 0.0020869565217391303, | |
| "grad_norm": 0.009172141551971436, | |
| "learning_rate": 3.881642512077295e-05, | |
| "loss": 0.0194, | |
| "step": 6930 | |
| }, | |
| { | |
| "epoch": 0.0025217391304347826, | |
| "grad_norm": 0.002410092856734991, | |
| "learning_rate": 3.879227053140097e-05, | |
| "loss": 0.0002, | |
| "step": 6940 | |
| }, | |
| { | |
| "epoch": 0.002956521739130435, | |
| "grad_norm": 0.003946972079575062, | |
| "learning_rate": 3.876811594202899e-05, | |
| "loss": 0.017, | |
| "step": 6950 | |
| }, | |
| { | |
| "epoch": 0.003391304347826087, | |
| "grad_norm": 14.169075965881348, | |
| "learning_rate": 3.8743961352657004e-05, | |
| "loss": 0.0713, | |
| "step": 6960 | |
| }, | |
| { | |
| "epoch": 0.0038260869565217392, | |
| "grad_norm": 0.0025988216511905193, | |
| "learning_rate": 3.871980676328503e-05, | |
| "loss": 0.0009, | |
| "step": 6970 | |
| }, | |
| { | |
| "epoch": 0.004043478260869565, | |
| "eval_accuracy": 0.9016393442622951, | |
| "eval_f1": 0.9011749047870193, | |
| "eval_loss": 0.6782234907150269, | |
| "eval_runtime": 88.3548, | |
| "eval_samples_per_second": 1.381, | |
| "eval_steps_per_second": 0.181, | |
| "step": 6975 | |
| }, | |
| { | |
| "epoch": 1.0002173913043477, | |
| "grad_norm": 0.5530427694320679, | |
| "learning_rate": 3.869565217391305e-05, | |
| "loss": 0.0023, | |
| "step": 6980 | |
| }, | |
| { | |
| "epoch": 1.0006521739130434, | |
| "grad_norm": 0.007848013192415237, | |
| "learning_rate": 3.8671497584541065e-05, | |
| "loss": 0.0118, | |
| "step": 6990 | |
| }, | |
| { | |
| "epoch": 1.001086956521739, | |
| "grad_norm": 0.001887777354568243, | |
| "learning_rate": 3.864734299516908e-05, | |
| "loss": 0.0071, | |
| "step": 7000 | |
| }, | |
| { | |
| "epoch": 1.0015217391304347, | |
| "grad_norm": 0.003241229336708784, | |
| "learning_rate": 3.862318840579711e-05, | |
| "loss": 0.1008, | |
| "step": 7010 | |
| }, | |
| { | |
| "epoch": 1.0019565217391304, | |
| "grad_norm": 0.0033748249989002943, | |
| "learning_rate": 3.8599033816425125e-05, | |
| "loss": 0.0148, | |
| "step": 7020 | |
| }, | |
| { | |
| "epoch": 1.002391304347826, | |
| "grad_norm": 0.0038462060038000345, | |
| "learning_rate": 3.857487922705314e-05, | |
| "loss": 0.0002, | |
| "step": 7030 | |
| }, | |
| { | |
| "epoch": 1.0028260869565218, | |
| "grad_norm": 0.0021478806156665087, | |
| "learning_rate": 3.855072463768116e-05, | |
| "loss": 0.0003, | |
| "step": 7040 | |
| }, | |
| { | |
| "epoch": 1.0032608695652174, | |
| "grad_norm": 0.006713178474456072, | |
| "learning_rate": 3.852657004830918e-05, | |
| "loss": 0.0368, | |
| "step": 7050 | |
| }, | |
| { | |
| "epoch": 1.003695652173913, | |
| "grad_norm": 13.128242492675781, | |
| "learning_rate": 3.85024154589372e-05, | |
| "loss": 0.088, | |
| "step": 7060 | |
| }, | |
| { | |
| "epoch": 1.0040434782608696, | |
| "eval_accuracy": 0.8360655737704918, | |
| "eval_f1": 0.8393668317566647, | |
| "eval_loss": 0.8946360945701599, | |
| "eval_runtime": 84.9121, | |
| "eval_samples_per_second": 1.437, | |
| "eval_steps_per_second": 0.188, | |
| "step": 7068 | |
| }, | |
| { | |
| "epoch": 2.000086956521739, | |
| "grad_norm": 0.02620559372007847, | |
| "learning_rate": 3.847826086956522e-05, | |
| "loss": 0.0704, | |
| "step": 7070 | |
| }, | |
| { | |
| "epoch": 2.0005217391304346, | |
| "grad_norm": 0.0025947631802409887, | |
| "learning_rate": 3.845410628019324e-05, | |
| "loss": 0.0662, | |
| "step": 7080 | |
| }, | |
| { | |
| "epoch": 2.0009565217391305, | |
| "grad_norm": 0.007165535818785429, | |
| "learning_rate": 3.8429951690821256e-05, | |
| "loss": 0.0202, | |
| "step": 7090 | |
| }, | |
| { | |
| "epoch": 2.001391304347826, | |
| "grad_norm": 0.0037356368266046047, | |
| "learning_rate": 3.8405797101449274e-05, | |
| "loss": 0.2776, | |
| "step": 7100 | |
| }, | |
| { | |
| "epoch": 2.001826086956522, | |
| "grad_norm": 0.0042053996585309505, | |
| "learning_rate": 3.83816425120773e-05, | |
| "loss": 0.0702, | |
| "step": 7110 | |
| }, | |
| { | |
| "epoch": 2.0022608695652173, | |
| "grad_norm": 0.0028894192073494196, | |
| "learning_rate": 3.835748792270532e-05, | |
| "loss": 0.0299, | |
| "step": 7120 | |
| }, | |
| { | |
| "epoch": 2.002695652173913, | |
| "grad_norm": 0.002774425782263279, | |
| "learning_rate": 3.8333333333333334e-05, | |
| "loss": 0.0154, | |
| "step": 7130 | |
| }, | |
| { | |
| "epoch": 2.0031304347826087, | |
| "grad_norm": 0.0035831418354064226, | |
| "learning_rate": 3.830917874396135e-05, | |
| "loss": 0.0061, | |
| "step": 7140 | |
| }, | |
| { | |
| "epoch": 2.0035652173913046, | |
| "grad_norm": 0.0023740469478070736, | |
| "learning_rate": 3.828502415458937e-05, | |
| "loss": 0.0294, | |
| "step": 7150 | |
| }, | |
| { | |
| "epoch": 2.004, | |
| "grad_norm": 0.007240784354507923, | |
| "learning_rate": 3.8260869565217395e-05, | |
| "loss": 0.0002, | |
| "step": 7160 | |
| }, | |
| { | |
| "epoch": 2.0040434782608694, | |
| "eval_accuracy": 0.9016393442622951, | |
| "eval_f1": 0.9007072239914696, | |
| "eval_loss": 0.6232790946960449, | |
| "eval_runtime": 74.029, | |
| "eval_samples_per_second": 1.648, | |
| "eval_steps_per_second": 0.216, | |
| "step": 7161 | |
| }, | |
| { | |
| "epoch": 3.000391304347826, | |
| "grad_norm": 0.002939334139227867, | |
| "learning_rate": 3.823671497584541e-05, | |
| "loss": 0.0094, | |
| "step": 7170 | |
| }, | |
| { | |
| "epoch": 3.0008260869565215, | |
| "grad_norm": 0.003376334672793746, | |
| "learning_rate": 3.821256038647344e-05, | |
| "loss": 0.0002, | |
| "step": 7180 | |
| }, | |
| { | |
| "epoch": 3.0012608695652174, | |
| "grad_norm": 0.0019673022907227278, | |
| "learning_rate": 3.818840579710145e-05, | |
| "loss": 0.0002, | |
| "step": 7190 | |
| }, | |
| { | |
| "epoch": 3.001695652173913, | |
| "grad_norm": 0.002067855792120099, | |
| "learning_rate": 3.8164251207729466e-05, | |
| "loss": 0.0003, | |
| "step": 7200 | |
| }, | |
| { | |
| "epoch": 3.0021304347826088, | |
| "grad_norm": 0.0016294581582769752, | |
| "learning_rate": 3.814009661835749e-05, | |
| "loss": 0.0114, | |
| "step": 7210 | |
| }, | |
| { | |
| "epoch": 3.002565217391304, | |
| "grad_norm": 0.0030627830419689417, | |
| "learning_rate": 3.811594202898551e-05, | |
| "loss": 0.0105, | |
| "step": 7220 | |
| }, | |
| { | |
| "epoch": 3.003, | |
| "grad_norm": 0.0019431186374276876, | |
| "learning_rate": 3.809178743961353e-05, | |
| "loss": 0.0248, | |
| "step": 7230 | |
| }, | |
| { | |
| "epoch": 3.0034347826086956, | |
| "grad_norm": 0.005873087793588638, | |
| "learning_rate": 3.8067632850241544e-05, | |
| "loss": 0.0144, | |
| "step": 7240 | |
| }, | |
| { | |
| "epoch": 3.0038695652173915, | |
| "grad_norm": 0.005152291618287563, | |
| "learning_rate": 3.804347826086957e-05, | |
| "loss": 0.0118, | |
| "step": 7250 | |
| }, | |
| { | |
| "epoch": 3.0040434782608694, | |
| "eval_accuracy": 0.8524590163934426, | |
| "eval_f1": 0.8544169462202248, | |
| "eval_loss": 0.8239415884017944, | |
| "eval_runtime": 84.6909, | |
| "eval_samples_per_second": 1.441, | |
| "eval_steps_per_second": 0.189, | |
| "step": 7254 | |
| }, | |
| { | |
| "epoch": 4.000260869565217, | |
| "grad_norm": 0.002831781283020973, | |
| "learning_rate": 3.8019323671497586e-05, | |
| "loss": 0.0087, | |
| "step": 7260 | |
| }, | |
| { | |
| "epoch": 4.000695652173913, | |
| "grad_norm": 0.0033079604618251324, | |
| "learning_rate": 3.7995169082125604e-05, | |
| "loss": 0.0111, | |
| "step": 7270 | |
| }, | |
| { | |
| "epoch": 4.001130434782609, | |
| "grad_norm": 0.0020894971676170826, | |
| "learning_rate": 3.797101449275363e-05, | |
| "loss": 0.0001, | |
| "step": 7280 | |
| }, | |
| { | |
| "epoch": 4.001565217391304, | |
| "grad_norm": 0.00628332793712616, | |
| "learning_rate": 3.794685990338164e-05, | |
| "loss": 0.0008, | |
| "step": 7290 | |
| }, | |
| { | |
| "epoch": 4.002, | |
| "grad_norm": 0.0022865217179059982, | |
| "learning_rate": 3.7922705314009665e-05, | |
| "loss": 0.0003, | |
| "step": 7300 | |
| }, | |
| { | |
| "epoch": 4.002434782608695, | |
| "grad_norm": 0.003050363389775157, | |
| "learning_rate": 3.789855072463768e-05, | |
| "loss": 0.0001, | |
| "step": 7310 | |
| }, | |
| { | |
| "epoch": 4.002869565217392, | |
| "grad_norm": 0.0020351973362267017, | |
| "learning_rate": 3.787439613526571e-05, | |
| "loss": 0.0001, | |
| "step": 7320 | |
| }, | |
| { | |
| "epoch": 4.003304347826087, | |
| "grad_norm": 0.002275005681440234, | |
| "learning_rate": 3.7850241545893725e-05, | |
| "loss": 0.0282, | |
| "step": 7330 | |
| }, | |
| { | |
| "epoch": 4.0037391304347825, | |
| "grad_norm": 0.056021079421043396, | |
| "learning_rate": 3.7826086956521736e-05, | |
| "loss": 0.01, | |
| "step": 7340 | |
| }, | |
| { | |
| "epoch": 4.004043478260869, | |
| "eval_accuracy": 0.8770491803278688, | |
| "eval_f1": 0.875608828010926, | |
| "eval_loss": 0.8939600586891174, | |
| "eval_runtime": 73.7752, | |
| "eval_samples_per_second": 1.654, | |
| "eval_steps_per_second": 0.217, | |
| "step": 7347 | |
| }, | |
| { | |
| "epoch": 5.0001304347826085, | |
| "grad_norm": 0.0015969000523909926, | |
| "learning_rate": 3.780193236714976e-05, | |
| "loss": 0.0001, | |
| "step": 7350 | |
| }, | |
| { | |
| "epoch": 5.000565217391304, | |
| "grad_norm": 0.002829205244779587, | |
| "learning_rate": 3.777777777777778e-05, | |
| "loss": 0.0204, | |
| "step": 7360 | |
| }, | |
| { | |
| "epoch": 5.001, | |
| "grad_norm": 0.0035543395206332207, | |
| "learning_rate": 3.77536231884058e-05, | |
| "loss": 0.0002, | |
| "step": 7370 | |
| }, | |
| { | |
| "epoch": 5.001434782608696, | |
| "grad_norm": 0.0014324752846732736, | |
| "learning_rate": 3.772946859903382e-05, | |
| "loss": 0.0001, | |
| "step": 7380 | |
| }, | |
| { | |
| "epoch": 5.001869565217391, | |
| "grad_norm": 0.009829861111938953, | |
| "learning_rate": 3.770531400966184e-05, | |
| "loss": 0.0003, | |
| "step": 7390 | |
| }, | |
| { | |
| "epoch": 5.002304347826087, | |
| "grad_norm": 0.0021648143883794546, | |
| "learning_rate": 3.7681159420289856e-05, | |
| "loss": 0.0214, | |
| "step": 7400 | |
| }, | |
| { | |
| "epoch": 5.002739130434783, | |
| "grad_norm": 0.002256903564557433, | |
| "learning_rate": 3.7657004830917874e-05, | |
| "loss": 0.0179, | |
| "step": 7410 | |
| }, | |
| { | |
| "epoch": 5.0031739130434785, | |
| "grad_norm": 0.008477986790239811, | |
| "learning_rate": 3.76328502415459e-05, | |
| "loss": 0.0156, | |
| "step": 7420 | |
| }, | |
| { | |
| "epoch": 5.003608695652174, | |
| "grad_norm": 1.539802074432373, | |
| "learning_rate": 3.7608695652173917e-05, | |
| "loss": 0.0102, | |
| "step": 7430 | |
| }, | |
| { | |
| "epoch": 5.004043478260869, | |
| "grad_norm": 0.001704772817902267, | |
| "learning_rate": 3.7584541062801934e-05, | |
| "loss": 0.0922, | |
| "step": 7440 | |
| }, | |
| { | |
| "epoch": 5.004043478260869, | |
| "eval_accuracy": 0.8934426229508197, | |
| "eval_f1": 0.890414307004471, | |
| "eval_loss": 0.7101210951805115, | |
| "eval_runtime": 74.1203, | |
| "eval_samples_per_second": 1.646, | |
| "eval_steps_per_second": 0.216, | |
| "step": 7440 | |
| }, | |
| { | |
| "epoch": 6.0004347826086954, | |
| "grad_norm": 0.0019552623853087425, | |
| "learning_rate": 3.756038647342995e-05, | |
| "loss": 0.0084, | |
| "step": 7450 | |
| }, | |
| { | |
| "epoch": 6.000869565217391, | |
| "grad_norm": 0.014820483513176441, | |
| "learning_rate": 3.753623188405797e-05, | |
| "loss": 0.0081, | |
| "step": 7460 | |
| }, | |
| { | |
| "epoch": 6.001304347826087, | |
| "grad_norm": 0.0021731434389948845, | |
| "learning_rate": 3.7512077294685995e-05, | |
| "loss": 0.0126, | |
| "step": 7470 | |
| }, | |
| { | |
| "epoch": 6.001739130434783, | |
| "grad_norm": 0.0026017893105745316, | |
| "learning_rate": 3.748792270531401e-05, | |
| "loss": 0.0001, | |
| "step": 7480 | |
| }, | |
| { | |
| "epoch": 6.002173913043478, | |
| "grad_norm": 16.81439781188965, | |
| "learning_rate": 3.746376811594203e-05, | |
| "loss": 0.0209, | |
| "step": 7490 | |
| }, | |
| { | |
| "epoch": 6.002608695652174, | |
| "grad_norm": 56.37519454956055, | |
| "learning_rate": 3.743961352657005e-05, | |
| "loss": 0.0623, | |
| "step": 7500 | |
| }, | |
| { | |
| "epoch": 6.00304347826087, | |
| "grad_norm": 0.0025837034918367863, | |
| "learning_rate": 3.7415458937198066e-05, | |
| "loss": 0.0009, | |
| "step": 7510 | |
| }, | |
| { | |
| "epoch": 6.003478260869565, | |
| "grad_norm": 0.0026143037248402834, | |
| "learning_rate": 3.739130434782609e-05, | |
| "loss": 0.0022, | |
| "step": 7520 | |
| }, | |
| { | |
| "epoch": 6.003913043478261, | |
| "grad_norm": 0.0012122008483856916, | |
| "learning_rate": 3.736714975845411e-05, | |
| "loss": 0.0004, | |
| "step": 7530 | |
| }, | |
| { | |
| "epoch": 6.004043478260869, | |
| "eval_accuracy": 0.8852459016393442, | |
| "eval_f1": 0.8842763274645694, | |
| "eval_loss": 0.8199717998504639, | |
| "eval_runtime": 75.9271, | |
| "eval_samples_per_second": 1.607, | |
| "eval_steps_per_second": 0.211, | |
| "step": 7533 | |
| }, | |
| { | |
| "epoch": 7.000304347826087, | |
| "grad_norm": 0.0016479407204315066, | |
| "learning_rate": 3.7342995169082126e-05, | |
| "loss": 0.1146, | |
| "step": 7540 | |
| }, | |
| { | |
| "epoch": 7.000739130434782, | |
| "grad_norm": 0.006008865777403116, | |
| "learning_rate": 3.7318840579710144e-05, | |
| "loss": 0.0003, | |
| "step": 7550 | |
| }, | |
| { | |
| "epoch": 7.001173913043479, | |
| "grad_norm": 0.013335944153368473, | |
| "learning_rate": 3.729468599033817e-05, | |
| "loss": 0.0013, | |
| "step": 7560 | |
| }, | |
| { | |
| "epoch": 7.001608695652174, | |
| "grad_norm": 0.010156466625630856, | |
| "learning_rate": 3.7270531400966186e-05, | |
| "loss": 0.001, | |
| "step": 7570 | |
| }, | |
| { | |
| "epoch": 7.00204347826087, | |
| "grad_norm": 97.3908462524414, | |
| "learning_rate": 3.7246376811594204e-05, | |
| "loss": 0.1149, | |
| "step": 7580 | |
| }, | |
| { | |
| "epoch": 7.002478260869565, | |
| "grad_norm": 1.504181146621704, | |
| "learning_rate": 3.722222222222222e-05, | |
| "loss": 0.0102, | |
| "step": 7590 | |
| }, | |
| { | |
| "epoch": 7.0029130434782605, | |
| "grad_norm": 95.67487335205078, | |
| "learning_rate": 3.719806763285024e-05, | |
| "loss": 0.1669, | |
| "step": 7600 | |
| }, | |
| { | |
| "epoch": 7.003347826086957, | |
| "grad_norm": 1.208970069885254, | |
| "learning_rate": 3.7173913043478264e-05, | |
| "loss": 0.0376, | |
| "step": 7610 | |
| }, | |
| { | |
| "epoch": 7.003782608695652, | |
| "grad_norm": 0.00800069235265255, | |
| "learning_rate": 3.714975845410628e-05, | |
| "loss": 0.1436, | |
| "step": 7620 | |
| }, | |
| { | |
| "epoch": 7.004043478260869, | |
| "eval_accuracy": 0.8278688524590164, | |
| "eval_f1": 0.8165147652045807, | |
| "eval_loss": 1.2253801822662354, | |
| "eval_runtime": 75.6061, | |
| "eval_samples_per_second": 1.614, | |
| "eval_steps_per_second": 0.212, | |
| "step": 7626 | |
| }, | |
| { | |
| "epoch": 8.000173913043477, | |
| "grad_norm": 0.7469026446342468, | |
| "learning_rate": 3.71256038647343e-05, | |
| "loss": 0.1145, | |
| "step": 7630 | |
| }, | |
| { | |
| "epoch": 8.000608695652174, | |
| "grad_norm": 0.007953154854476452, | |
| "learning_rate": 3.710144927536232e-05, | |
| "loss": 0.1984, | |
| "step": 7640 | |
| }, | |
| { | |
| "epoch": 8.00104347826087, | |
| "grad_norm": 0.0036359576042741537, | |
| "learning_rate": 3.7077294685990336e-05, | |
| "loss": 0.0086, | |
| "step": 7650 | |
| }, | |
| { | |
| "epoch": 8.001478260869565, | |
| "grad_norm": 0.003452168544754386, | |
| "learning_rate": 3.705314009661836e-05, | |
| "loss": 0.0541, | |
| "step": 7660 | |
| }, | |
| { | |
| "epoch": 8.001913043478261, | |
| "grad_norm": 0.0031451256945729256, | |
| "learning_rate": 3.702898550724638e-05, | |
| "loss": 0.0092, | |
| "step": 7670 | |
| }, | |
| { | |
| "epoch": 8.002347826086957, | |
| "grad_norm": 0.003350870218127966, | |
| "learning_rate": 3.70048309178744e-05, | |
| "loss": 0.0102, | |
| "step": 7680 | |
| }, | |
| { | |
| "epoch": 8.002782608695652, | |
| "grad_norm": 99.97783660888672, | |
| "learning_rate": 3.698067632850242e-05, | |
| "loss": 0.1118, | |
| "step": 7690 | |
| }, | |
| { | |
| "epoch": 8.003217391304348, | |
| "grad_norm": 0.0027307348791509867, | |
| "learning_rate": 3.695652173913043e-05, | |
| "loss": 0.0789, | |
| "step": 7700 | |
| }, | |
| { | |
| "epoch": 8.003652173913043, | |
| "grad_norm": 0.002151375636458397, | |
| "learning_rate": 3.6932367149758456e-05, | |
| "loss": 0.0002, | |
| "step": 7710 | |
| }, | |
| { | |
| "epoch": 8.00404347826087, | |
| "eval_accuracy": 0.860655737704918, | |
| "eval_f1": 0.8544375353306953, | |
| "eval_loss": 0.8904465436935425, | |
| "eval_runtime": 74.2202, | |
| "eval_samples_per_second": 1.644, | |
| "eval_steps_per_second": 0.216, | |
| "step": 7719 | |
| }, | |
| { | |
| "epoch": 9.000043478260869, | |
| "grad_norm": 0.0031512887217104435, | |
| "learning_rate": 3.6908212560386474e-05, | |
| "loss": 0.0003, | |
| "step": 7720 | |
| }, | |
| { | |
| "epoch": 9.000478260869565, | |
| "grad_norm": 0.004215599969029427, | |
| "learning_rate": 3.68840579710145e-05, | |
| "loss": 0.0434, | |
| "step": 7730 | |
| }, | |
| { | |
| "epoch": 9.000913043478262, | |
| "grad_norm": 0.0027098949067294598, | |
| "learning_rate": 3.6859903381642516e-05, | |
| "loss": 0.0012, | |
| "step": 7740 | |
| }, | |
| { | |
| "epoch": 9.001347826086956, | |
| "grad_norm": 0.0018039434216916561, | |
| "learning_rate": 3.6835748792270534e-05, | |
| "loss": 0.0084, | |
| "step": 7750 | |
| }, | |
| { | |
| "epoch": 9.001782608695652, | |
| "grad_norm": 0.002414983231574297, | |
| "learning_rate": 3.681159420289855e-05, | |
| "loss": 0.0079, | |
| "step": 7760 | |
| }, | |
| { | |
| "epoch": 9.002217391304347, | |
| "grad_norm": 0.001327994978055358, | |
| "learning_rate": 3.678743961352657e-05, | |
| "loss": 0.0002, | |
| "step": 7770 | |
| }, | |
| { | |
| "epoch": 9.002652173913043, | |
| "grad_norm": 0.002577988663688302, | |
| "learning_rate": 3.6763285024154594e-05, | |
| "loss": 0.0805, | |
| "step": 7780 | |
| }, | |
| { | |
| "epoch": 9.00308695652174, | |
| "grad_norm": 0.002542161149904132, | |
| "learning_rate": 3.673913043478261e-05, | |
| "loss": 0.0002, | |
| "step": 7790 | |
| }, | |
| { | |
| "epoch": 9.003521739130434, | |
| "grad_norm": 0.002315797843039036, | |
| "learning_rate": 3.671497584541063e-05, | |
| "loss": 0.0132, | |
| "step": 7800 | |
| }, | |
| { | |
| "epoch": 9.00395652173913, | |
| "grad_norm": 12.634377479553223, | |
| "learning_rate": 3.669082125603865e-05, | |
| "loss": 0.1179, | |
| "step": 7810 | |
| }, | |
| { | |
| "epoch": 9.00404347826087, | |
| "eval_accuracy": 0.9016393442622951, | |
| "eval_f1": 0.9009970695079303, | |
| "eval_loss": 0.6947475671768188, | |
| "eval_runtime": 78.5379, | |
| "eval_samples_per_second": 1.553, | |
| "eval_steps_per_second": 0.204, | |
| "step": 7812 | |
| }, | |
| { | |
| "epoch": 10.000347826086957, | |
| "grad_norm": 0.004218494985252619, | |
| "learning_rate": 3.6666666666666666e-05, | |
| "loss": 0.0091, | |
| "step": 7820 | |
| }, | |
| { | |
| "epoch": 10.000782608695653, | |
| "grad_norm": 0.07844866067171097, | |
| "learning_rate": 3.664251207729469e-05, | |
| "loss": 0.0356, | |
| "step": 7830 | |
| }, | |
| { | |
| "epoch": 10.001217391304348, | |
| "grad_norm": 0.0030379490926861763, | |
| "learning_rate": 3.661835748792271e-05, | |
| "loss": 0.0255, | |
| "step": 7840 | |
| }, | |
| { | |
| "epoch": 10.001652173913044, | |
| "grad_norm": 0.01330595277249813, | |
| "learning_rate": 3.6594202898550726e-05, | |
| "loss": 0.0002, | |
| "step": 7850 | |
| }, | |
| { | |
| "epoch": 10.002086956521739, | |
| "grad_norm": 0.0024776251520961523, | |
| "learning_rate": 3.6570048309178744e-05, | |
| "loss": 0.0133, | |
| "step": 7860 | |
| }, | |
| { | |
| "epoch": 10.002521739130435, | |
| "grad_norm": 0.005490115378051996, | |
| "learning_rate": 3.654589371980677e-05, | |
| "loss": 0.0002, | |
| "step": 7870 | |
| }, | |
| { | |
| "epoch": 10.002956521739131, | |
| "grad_norm": 0.0017077559605240822, | |
| "learning_rate": 3.6521739130434786e-05, | |
| "loss": 0.0002, | |
| "step": 7880 | |
| }, | |
| { | |
| "epoch": 10.003391304347826, | |
| "grad_norm": 0.0023673507384955883, | |
| "learning_rate": 3.6497584541062804e-05, | |
| "loss": 0.0071, | |
| "step": 7890 | |
| }, | |
| { | |
| "epoch": 10.003826086956522, | |
| "grad_norm": 0.047657690942287445, | |
| "learning_rate": 3.647342995169082e-05, | |
| "loss": 0.0033, | |
| "step": 7900 | |
| }, | |
| { | |
| "epoch": 10.00404347826087, | |
| "eval_accuracy": 0.9016393442622951, | |
| "eval_f1": 0.9009244537168961, | |
| "eval_loss": 0.6869687438011169, | |
| "eval_runtime": 73.4337, | |
| "eval_samples_per_second": 1.661, | |
| "eval_steps_per_second": 0.218, | |
| "step": 7905 | |
| }, | |
| { | |
| "epoch": 11.000217391304348, | |
| "grad_norm": 0.0020451026502996683, | |
| "learning_rate": 3.644927536231884e-05, | |
| "loss": 0.009, | |
| "step": 7910 | |
| }, | |
| { | |
| "epoch": 11.000652173913043, | |
| "grad_norm": 0.03150273114442825, | |
| "learning_rate": 3.6425120772946864e-05, | |
| "loss": 0.0003, | |
| "step": 7920 | |
| }, | |
| { | |
| "epoch": 11.001086956521739, | |
| "grad_norm": 0.0021130284294486046, | |
| "learning_rate": 3.640096618357488e-05, | |
| "loss": 0.0102, | |
| "step": 7930 | |
| }, | |
| { | |
| "epoch": 11.001521739130435, | |
| "grad_norm": 0.0027534847613424063, | |
| "learning_rate": 3.63768115942029e-05, | |
| "loss": 0.0001, | |
| "step": 7940 | |
| }, | |
| { | |
| "epoch": 11.00195652173913, | |
| "grad_norm": 0.0018295743502676487, | |
| "learning_rate": 3.635265700483092e-05, | |
| "loss": 0.0022, | |
| "step": 7950 | |
| }, | |
| { | |
| "epoch": 11.002391304347826, | |
| "grad_norm": 0.001954080071300268, | |
| "learning_rate": 3.6328502415458936e-05, | |
| "loss": 0.0287, | |
| "step": 7960 | |
| }, | |
| { | |
| "epoch": 11.00282608695652, | |
| "grad_norm": 0.001990045653656125, | |
| "learning_rate": 3.630434782608696e-05, | |
| "loss": 0.0001, | |
| "step": 7970 | |
| }, | |
| { | |
| "epoch": 11.003260869565217, | |
| "grad_norm": 0.0021924017928540707, | |
| "learning_rate": 3.628019323671498e-05, | |
| "loss": 0.1093, | |
| "step": 7980 | |
| }, | |
| { | |
| "epoch": 11.003695652173914, | |
| "grad_norm": 0.0014019764494150877, | |
| "learning_rate": 3.6256038647343e-05, | |
| "loss": 0.0003, | |
| "step": 7990 | |
| }, | |
| { | |
| "epoch": 11.00404347826087, | |
| "eval_accuracy": 0.8852459016393442, | |
| "eval_f1": 0.8834037386396851, | |
| "eval_loss": 0.6801308393478394, | |
| "eval_runtime": 76.3228, | |
| "eval_samples_per_second": 1.598, | |
| "eval_steps_per_second": 0.21, | |
| "step": 7998 | |
| }, | |
| { | |
| "epoch": 12.00008695652174, | |
| "grad_norm": 0.0022465840447694063, | |
| "learning_rate": 3.6231884057971014e-05, | |
| "loss": 0.0086, | |
| "step": 8000 | |
| }, | |
| { | |
| "epoch": 12.000521739130434, | |
| "grad_norm": 0.009984511882066727, | |
| "learning_rate": 3.620772946859903e-05, | |
| "loss": 0.0171, | |
| "step": 8010 | |
| }, | |
| { | |
| "epoch": 12.00095652173913, | |
| "grad_norm": 0.0018829620676115155, | |
| "learning_rate": 3.6183574879227056e-05, | |
| "loss": 0.0082, | |
| "step": 8020 | |
| }, | |
| { | |
| "epoch": 12.001391304347827, | |
| "grad_norm": 0.0022015601862221956, | |
| "learning_rate": 3.6159420289855074e-05, | |
| "loss": 0.0018, | |
| "step": 8030 | |
| }, | |
| { | |
| "epoch": 12.001826086956521, | |
| "grad_norm": 0.0011975386878475547, | |
| "learning_rate": 3.61352657004831e-05, | |
| "loss": 0.0168, | |
| "step": 8040 | |
| }, | |
| { | |
| "epoch": 12.002260869565218, | |
| "grad_norm": 0.0013748151250183582, | |
| "learning_rate": 3.611111111111111e-05, | |
| "loss": 0.0003, | |
| "step": 8050 | |
| }, | |
| { | |
| "epoch": 12.002695652173912, | |
| "grad_norm": 0.0027548738289624453, | |
| "learning_rate": 3.6086956521739134e-05, | |
| "loss": 0.0103, | |
| "step": 8060 | |
| }, | |
| { | |
| "epoch": 12.003130434782609, | |
| "grad_norm": 0.001284057623706758, | |
| "learning_rate": 3.606280193236715e-05, | |
| "loss": 0.1057, | |
| "step": 8070 | |
| }, | |
| { | |
| "epoch": 12.003565217391305, | |
| "grad_norm": 0.0024062602315098047, | |
| "learning_rate": 3.603864734299517e-05, | |
| "loss": 0.0391, | |
| "step": 8080 | |
| }, | |
| { | |
| "epoch": 12.004, | |
| "grad_norm": 1.8473665714263916, | |
| "learning_rate": 3.6014492753623194e-05, | |
| "loss": 0.0566, | |
| "step": 8090 | |
| }, | |
| { | |
| "epoch": 12.00404347826087, | |
| "eval_accuracy": 0.8852459016393442, | |
| "eval_f1": 0.8826751117734725, | |
| "eval_loss": 0.7470003366470337, | |
| "eval_runtime": 76.8169, | |
| "eval_samples_per_second": 1.588, | |
| "eval_steps_per_second": 0.208, | |
| "step": 8091 | |
| }, | |
| { | |
| "epoch": 13.000391304347826, | |
| "grad_norm": 0.006008814089000225, | |
| "learning_rate": 3.5990338164251205e-05, | |
| "loss": 0.01, | |
| "step": 8100 | |
| }, | |
| { | |
| "epoch": 13.000826086956522, | |
| "grad_norm": 0.003681233851239085, | |
| "learning_rate": 3.596618357487923e-05, | |
| "loss": 0.0084, | |
| "step": 8110 | |
| }, | |
| { | |
| "epoch": 13.001260869565217, | |
| "grad_norm": 0.002310408977791667, | |
| "learning_rate": 3.594202898550725e-05, | |
| "loss": 0.1267, | |
| "step": 8120 | |
| }, | |
| { | |
| "epoch": 13.001695652173913, | |
| "grad_norm": 0.009503014385700226, | |
| "learning_rate": 3.5917874396135266e-05, | |
| "loss": 0.0066, | |
| "step": 8130 | |
| }, | |
| { | |
| "epoch": 13.00213043478261, | |
| "grad_norm": 0.009119764901697636, | |
| "learning_rate": 3.589371980676329e-05, | |
| "loss": 0.0108, | |
| "step": 8140 | |
| }, | |
| { | |
| "epoch": 13.002565217391304, | |
| "grad_norm": 0.00435229716822505, | |
| "learning_rate": 3.58695652173913e-05, | |
| "loss": 0.0003, | |
| "step": 8150 | |
| }, | |
| { | |
| "epoch": 13.003, | |
| "grad_norm": 0.0040861936286091805, | |
| "learning_rate": 3.5845410628019326e-05, | |
| "loss": 0.0109, | |
| "step": 8160 | |
| }, | |
| { | |
| "epoch": 13.003434782608696, | |
| "grad_norm": 0.0029671050142496824, | |
| "learning_rate": 3.5821256038647344e-05, | |
| "loss": 0.0003, | |
| "step": 8170 | |
| }, | |
| { | |
| "epoch": 13.003869565217391, | |
| "grad_norm": 0.0019406734500080347, | |
| "learning_rate": 3.579710144927537e-05, | |
| "loss": 0.0111, | |
| "step": 8180 | |
| }, | |
| { | |
| "epoch": 13.00404347826087, | |
| "eval_accuracy": 0.8770491803278688, | |
| "eval_f1": 0.8738998934662604, | |
| "eval_loss": 0.738959789276123, | |
| "eval_runtime": 77.6141, | |
| "eval_samples_per_second": 1.572, | |
| "eval_steps_per_second": 0.206, | |
| "step": 8184 | |
| }, | |
| { | |
| "epoch": 14.000260869565217, | |
| "grad_norm": 0.0053750560618937016, | |
| "learning_rate": 3.5772946859903386e-05, | |
| "loss": 0.0188, | |
| "step": 8190 | |
| }, | |
| { | |
| "epoch": 14.000695652173913, | |
| "grad_norm": 0.0030118555296212435, | |
| "learning_rate": 3.57487922705314e-05, | |
| "loss": 0.0217, | |
| "step": 8200 | |
| }, | |
| { | |
| "epoch": 14.001130434782608, | |
| "grad_norm": 0.0030062675941735506, | |
| "learning_rate": 3.572463768115942e-05, | |
| "loss": 0.0091, | |
| "step": 8210 | |
| }, | |
| { | |
| "epoch": 14.001565217391304, | |
| "grad_norm": 0.002661097561940551, | |
| "learning_rate": 3.570048309178744e-05, | |
| "loss": 0.0002, | |
| "step": 8220 | |
| }, | |
| { | |
| "epoch": 14.002, | |
| "grad_norm": 0.0018812001217156649, | |
| "learning_rate": 3.5676328502415464e-05, | |
| "loss": 0.0124, | |
| "step": 8230 | |
| }, | |
| { | |
| "epoch": 14.002434782608695, | |
| "grad_norm": 0.003940836526453495, | |
| "learning_rate": 3.565217391304348e-05, | |
| "loss": 0.0001, | |
| "step": 8240 | |
| }, | |
| { | |
| "epoch": 14.002869565217392, | |
| "grad_norm": 0.0017722449265420437, | |
| "learning_rate": 3.56280193236715e-05, | |
| "loss": 0.0001, | |
| "step": 8250 | |
| }, | |
| { | |
| "epoch": 14.003304347826086, | |
| "grad_norm": 0.0024237651377916336, | |
| "learning_rate": 3.560386473429952e-05, | |
| "loss": 0.0001, | |
| "step": 8260 | |
| }, | |
| { | |
| "epoch": 14.003739130434782, | |
| "grad_norm": 0.0013878210447728634, | |
| "learning_rate": 3.5579710144927535e-05, | |
| "loss": 0.0096, | |
| "step": 8270 | |
| }, | |
| { | |
| "epoch": 14.00404347826087, | |
| "eval_accuracy": 0.9262295081967213, | |
| "eval_f1": 0.9249199343461638, | |
| "eval_loss": 0.449712336063385, | |
| "eval_runtime": 71.2976, | |
| "eval_samples_per_second": 1.711, | |
| "eval_steps_per_second": 0.224, | |
| "step": 8277 | |
| }, | |
| { | |
| "epoch": 15.000130434782609, | |
| "grad_norm": 0.0015423373552039266, | |
| "learning_rate": 3.555555555555556e-05, | |
| "loss": 0.0096, | |
| "step": 8280 | |
| }, | |
| { | |
| "epoch": 15.000565217391305, | |
| "grad_norm": 0.0017125405138358474, | |
| "learning_rate": 3.553140096618358e-05, | |
| "loss": 0.0094, | |
| "step": 8290 | |
| }, | |
| { | |
| "epoch": 15.001, | |
| "grad_norm": 0.0017512089107185602, | |
| "learning_rate": 3.5507246376811596e-05, | |
| "loss": 0.0001, | |
| "step": 8300 | |
| }, | |
| { | |
| "epoch": 15.001434782608696, | |
| "grad_norm": 0.0016162145184352994, | |
| "learning_rate": 3.5483091787439613e-05, | |
| "loss": 0.0001, | |
| "step": 8310 | |
| }, | |
| { | |
| "epoch": 15.001869565217392, | |
| "grad_norm": 0.001804832834750414, | |
| "learning_rate": 3.545893719806763e-05, | |
| "loss": 0.0098, | |
| "step": 8320 | |
| }, | |
| { | |
| "epoch": 15.002304347826087, | |
| "grad_norm": 0.0017595053650438786, | |
| "learning_rate": 3.5434782608695656e-05, | |
| "loss": 0.0001, | |
| "step": 8330 | |
| }, | |
| { | |
| "epoch": 15.002739130434783, | |
| "grad_norm": 0.0015612575225532055, | |
| "learning_rate": 3.5410628019323674e-05, | |
| "loss": 0.0186, | |
| "step": 8340 | |
| }, | |
| { | |
| "epoch": 15.003173913043478, | |
| "grad_norm": 0.01624816469848156, | |
| "learning_rate": 3.538647342995169e-05, | |
| "loss": 0.0138, | |
| "step": 8350 | |
| }, | |
| { | |
| "epoch": 15.003608695652174, | |
| "grad_norm": 0.0014068948803469539, | |
| "learning_rate": 3.536231884057971e-05, | |
| "loss": 0.0308, | |
| "step": 8360 | |
| }, | |
| { | |
| "epoch": 15.00404347826087, | |
| "grad_norm": 0.003260946599766612, | |
| "learning_rate": 3.5338164251207734e-05, | |
| "loss": 0.0001, | |
| "step": 8370 | |
| }, | |
| { | |
| "epoch": 15.00404347826087, | |
| "eval_accuracy": 0.9262295081967213, | |
| "eval_f1": 0.9258667205923087, | |
| "eval_loss": 0.5235211253166199, | |
| "eval_runtime": 86.922, | |
| "eval_samples_per_second": 1.404, | |
| "eval_steps_per_second": 0.184, | |
| "step": 8370 | |
| }, | |
| { | |
| "epoch": 0.0010869565217391304, | |
| "grad_norm": 0.0017507346346974373, | |
| "learning_rate": 4.951690821256039e-06, | |
| "loss": 0.0001, | |
| "step": 8380 | |
| }, | |
| { | |
| "epoch": 0.002173913043478261, | |
| "grad_norm": 0.0017071341862902045, | |
| "learning_rate": 4.891304347826087e-06, | |
| "loss": 0.0001, | |
| "step": 8390 | |
| }, | |
| { | |
| "epoch": 0.003260869565217391, | |
| "grad_norm": 0.0015891263028606772, | |
| "learning_rate": 4.830917874396135e-06, | |
| "loss": 0.0001, | |
| "step": 8400 | |
| }, | |
| { | |
| "epoch": 0.004347826086956522, | |
| "grad_norm": 0.002104247687384486, | |
| "learning_rate": 4.770531400966183e-06, | |
| "loss": 0.0099, | |
| "step": 8410 | |
| }, | |
| { | |
| "epoch": 0.005434782608695652, | |
| "grad_norm": 0.0007910887943580747, | |
| "learning_rate": 4.710144927536232e-06, | |
| "loss": 0.008, | |
| "step": 8420 | |
| }, | |
| { | |
| "epoch": 0.006521739130434782, | |
| "grad_norm": 0.0013656615046784282, | |
| "learning_rate": 4.64975845410628e-06, | |
| "loss": 0.0101, | |
| "step": 8430 | |
| }, | |
| { | |
| "epoch": 0.007608695652173913, | |
| "grad_norm": 1.222392201423645, | |
| "learning_rate": 4.589371980676329e-06, | |
| "loss": 0.0146, | |
| "step": 8440 | |
| }, | |
| { | |
| "epoch": 0.008695652173913044, | |
| "grad_norm": 0.0019674005452543497, | |
| "learning_rate": 4.528985507246377e-06, | |
| "loss": 0.0108, | |
| "step": 8450 | |
| }, | |
| { | |
| "epoch": 0.009782608695652175, | |
| "grad_norm": 0.0012134442804381251, | |
| "learning_rate": 4.468599033816425e-06, | |
| "loss": 0.0002, | |
| "step": 8460 | |
| }, | |
| { | |
| "epoch": 0.010108695652173913, | |
| "eval_accuracy": 0.9262295081967213, | |
| "eval_f1": 0.9261042886093471, | |
| "eval_loss": 0.5340837240219116, | |
| "eval_runtime": 87.7431, | |
| "eval_samples_per_second": 1.39, | |
| "eval_steps_per_second": 0.182, | |
| "step": 8463 | |
| }, | |
| { | |
| "epoch": 1.0007608695652175, | |
| "grad_norm": 0.001567196799442172, | |
| "learning_rate": 4.408212560386473e-06, | |
| "loss": 0.0001, | |
| "step": 8470 | |
| }, | |
| { | |
| "epoch": 1.0018478260869565, | |
| "grad_norm": 0.0016586320707574487, | |
| "learning_rate": 4.347826086956522e-06, | |
| "loss": 0.0096, | |
| "step": 8480 | |
| }, | |
| { | |
| "epoch": 1.0029347826086956, | |
| "grad_norm": 0.0023030065931379795, | |
| "learning_rate": 4.28743961352657e-06, | |
| "loss": 0.0077, | |
| "step": 8490 | |
| }, | |
| { | |
| "epoch": 1.0040217391304347, | |
| "grad_norm": 0.002070185961201787, | |
| "learning_rate": 4.227053140096619e-06, | |
| "loss": 0.0001, | |
| "step": 8500 | |
| }, | |
| { | |
| "epoch": 1.005108695652174, | |
| "grad_norm": 0.0012701443629339337, | |
| "learning_rate": 4.166666666666667e-06, | |
| "loss": 0.0106, | |
| "step": 8510 | |
| }, | |
| { | |
| "epoch": 1.006195652173913, | |
| "grad_norm": 0.0018321698298677802, | |
| "learning_rate": 4.106280193236716e-06, | |
| "loss": 0.0001, | |
| "step": 8520 | |
| }, | |
| { | |
| "epoch": 1.0072826086956521, | |
| "grad_norm": 0.0011002240935340524, | |
| "learning_rate": 4.045893719806764e-06, | |
| "loss": 0.0002, | |
| "step": 8530 | |
| }, | |
| { | |
| "epoch": 1.0083695652173914, | |
| "grad_norm": 0.0006837508990429342, | |
| "learning_rate": 3.9855072463768115e-06, | |
| "loss": 0.0253, | |
| "step": 8540 | |
| }, | |
| { | |
| "epoch": 1.0094565217391305, | |
| "grad_norm": 0.0011974646477028728, | |
| "learning_rate": 3.92512077294686e-06, | |
| "loss": 0.0001, | |
| "step": 8550 | |
| }, | |
| { | |
| "epoch": 1.0101086956521739, | |
| "eval_accuracy": 0.9180327868852459, | |
| "eval_f1": 0.9178693897416935, | |
| "eval_loss": 0.5411162376403809, | |
| "eval_runtime": 71.3448, | |
| "eval_samples_per_second": 1.71, | |
| "eval_steps_per_second": 0.224, | |
| "step": 8556 | |
| }, | |
| { | |
| "epoch": 2.0004347826086954, | |
| "grad_norm": 0.00192980095744133, | |
| "learning_rate": 3.864734299516908e-06, | |
| "loss": 0.0747, | |
| "step": 8560 | |
| }, | |
| { | |
| "epoch": 2.001521739130435, | |
| "grad_norm": 0.0019794933032244444, | |
| "learning_rate": 3.804347826086957e-06, | |
| "loss": 0.0176, | |
| "step": 8570 | |
| }, | |
| { | |
| "epoch": 2.002608695652174, | |
| "grad_norm": 0.0013387734070420265, | |
| "learning_rate": 3.743961352657005e-06, | |
| "loss": 0.0028, | |
| "step": 8580 | |
| }, | |
| { | |
| "epoch": 2.003695652173913, | |
| "grad_norm": 0.0012565733632072806, | |
| "learning_rate": 3.683574879227053e-06, | |
| "loss": 0.0001, | |
| "step": 8590 | |
| }, | |
| { | |
| "epoch": 2.004782608695652, | |
| "grad_norm": 0.0024499129503965378, | |
| "learning_rate": 3.6231884057971017e-06, | |
| "loss": 0.0001, | |
| "step": 8600 | |
| }, | |
| { | |
| "epoch": 2.0058695652173912, | |
| "grad_norm": 0.0025332611985504627, | |
| "learning_rate": 3.5628019323671496e-06, | |
| "loss": 0.0001, | |
| "step": 8610 | |
| }, | |
| { | |
| "epoch": 2.0069565217391303, | |
| "grad_norm": 0.0016111568547785282, | |
| "learning_rate": 3.5024154589371984e-06, | |
| "loss": 0.0075, | |
| "step": 8620 | |
| }, | |
| { | |
| "epoch": 2.0080434782608694, | |
| "grad_norm": 0.0015177864115685225, | |
| "learning_rate": 3.4420289855072464e-06, | |
| "loss": 0.0068, | |
| "step": 8630 | |
| }, | |
| { | |
| "epoch": 2.009130434782609, | |
| "grad_norm": 0.002059792634099722, | |
| "learning_rate": 3.3816425120772947e-06, | |
| "loss": 0.0211, | |
| "step": 8640 | |
| }, | |
| { | |
| "epoch": 2.010108695652174, | |
| "eval_accuracy": 0.9180327868852459, | |
| "eval_f1": 0.9178693897416935, | |
| "eval_loss": 0.5265507698059082, | |
| "eval_runtime": 77.2441, | |
| "eval_samples_per_second": 1.579, | |
| "eval_steps_per_second": 0.207, | |
| "step": 8649 | |
| }, | |
| { | |
| "epoch": 3.000108695652174, | |
| "grad_norm": 0.0017185317119583488, | |
| "learning_rate": 3.3212560386473435e-06, | |
| "loss": 0.0001, | |
| "step": 8650 | |
| }, | |
| { | |
| "epoch": 3.001195652173913, | |
| "grad_norm": 0.0017326638335362077, | |
| "learning_rate": 3.2608695652173914e-06, | |
| "loss": 0.0079, | |
| "step": 8660 | |
| }, | |
| { | |
| "epoch": 3.002282608695652, | |
| "grad_norm": 0.002113594673573971, | |
| "learning_rate": 3.2004830917874402e-06, | |
| "loss": 0.0001, | |
| "step": 8670 | |
| }, | |
| { | |
| "epoch": 3.0033695652173913, | |
| "grad_norm": 0.0021217581816017628, | |
| "learning_rate": 3.140096618357488e-06, | |
| "loss": 0.0001, | |
| "step": 8680 | |
| }, | |
| { | |
| "epoch": 3.0044565217391304, | |
| "grad_norm": 0.0013151352759450674, | |
| "learning_rate": 3.0797101449275365e-06, | |
| "loss": 0.0001, | |
| "step": 8690 | |
| }, | |
| { | |
| "epoch": 3.0055434782608694, | |
| "grad_norm": 0.001686801202595234, | |
| "learning_rate": 3.0193236714975845e-06, | |
| "loss": 0.0098, | |
| "step": 8700 | |
| }, | |
| { | |
| "epoch": 3.0066304347826085, | |
| "grad_norm": 0.001321412273682654, | |
| "learning_rate": 2.958937198067633e-06, | |
| "loss": 0.0105, | |
| "step": 8710 | |
| }, | |
| { | |
| "epoch": 3.007717391304348, | |
| "grad_norm": 0.0011213988764211535, | |
| "learning_rate": 2.898550724637681e-06, | |
| "loss": 0.01, | |
| "step": 8720 | |
| }, | |
| { | |
| "epoch": 3.008804347826087, | |
| "grad_norm": 0.0020139431580901146, | |
| "learning_rate": 2.8381642512077295e-06, | |
| "loss": 0.0076, | |
| "step": 8730 | |
| }, | |
| { | |
| "epoch": 3.009891304347826, | |
| "grad_norm": 0.0018427580362185836, | |
| "learning_rate": 2.777777777777778e-06, | |
| "loss": 0.0077, | |
| "step": 8740 | |
| }, | |
| { | |
| "epoch": 3.010108695652174, | |
| "eval_accuracy": 0.9180327868852459, | |
| "eval_f1": 0.9178693897416935, | |
| "eval_loss": 0.5282787680625916, | |
| "eval_runtime": 69.3627, | |
| "eval_samples_per_second": 1.759, | |
| "eval_steps_per_second": 0.231, | |
| "step": 8742 | |
| }, | |
| { | |
| "epoch": 4.000869565217391, | |
| "grad_norm": 0.0015537792351096869, | |
| "learning_rate": 2.7173913043478263e-06, | |
| "loss": 0.0104, | |
| "step": 8750 | |
| }, | |
| { | |
| "epoch": 4.001956521739131, | |
| "grad_norm": 0.0009913562098518014, | |
| "learning_rate": 2.6570048309178746e-06, | |
| "loss": 0.008, | |
| "step": 8760 | |
| }, | |
| { | |
| "epoch": 4.00304347826087, | |
| "grad_norm": 0.0024362721014767885, | |
| "learning_rate": 2.596618357487923e-06, | |
| "loss": 0.0001, | |
| "step": 8770 | |
| }, | |
| { | |
| "epoch": 4.004130434782609, | |
| "grad_norm": 0.0016615530475974083, | |
| "learning_rate": 2.5362318840579714e-06, | |
| "loss": 0.0001, | |
| "step": 8780 | |
| }, | |
| { | |
| "epoch": 4.005217391304348, | |
| "grad_norm": 0.0013662370620295405, | |
| "learning_rate": 2.4758454106280193e-06, | |
| "loss": 0.0001, | |
| "step": 8790 | |
| }, | |
| { | |
| "epoch": 4.006304347826087, | |
| "grad_norm": 0.0018021605210378766, | |
| "learning_rate": 2.4154589371980677e-06, | |
| "loss": 0.0001, | |
| "step": 8800 | |
| }, | |
| { | |
| "epoch": 4.007391304347826, | |
| "grad_norm": 2.3952181339263916, | |
| "learning_rate": 2.355072463768116e-06, | |
| "loss": 0.0146, | |
| "step": 8810 | |
| }, | |
| { | |
| "epoch": 4.008478260869565, | |
| "grad_norm": 0.0014442868996411562, | |
| "learning_rate": 2.2946859903381644e-06, | |
| "loss": 0.0098, | |
| "step": 8820 | |
| }, | |
| { | |
| "epoch": 4.009565217391304, | |
| "grad_norm": 0.0016332181403413415, | |
| "learning_rate": 2.2342995169082123e-06, | |
| "loss": 0.0101, | |
| "step": 8830 | |
| }, | |
| { | |
| "epoch": 4.010108695652174, | |
| "eval_accuracy": 0.9180327868852459, | |
| "eval_f1": 0.9178693897416935, | |
| "eval_loss": 0.5279106497764587, | |
| "eval_runtime": 68.8675, | |
| "eval_samples_per_second": 1.772, | |
| "eval_steps_per_second": 0.232, | |
| "step": 8835 | |
| }, | |
| { | |
| "epoch": 5.0005434782608695, | |
| "grad_norm": 0.001660835579968989, | |
| "learning_rate": 2.173913043478261e-06, | |
| "loss": 0.0001, | |
| "step": 8840 | |
| }, | |
| { | |
| "epoch": 5.001630434782609, | |
| "grad_norm": 0.0013907772954553366, | |
| "learning_rate": 2.1135265700483095e-06, | |
| "loss": 0.015, | |
| "step": 8850 | |
| }, | |
| { | |
| "epoch": 5.002391304347826, | |
| "eval_accuracy": 0.8777292576419214, | |
| "eval_f1": 0.8759548256004502, | |
| "eval_loss": 0.9210804104804993, | |
| "eval_runtime": 135.0706, | |
| "eval_samples_per_second": 1.695, | |
| "eval_steps_per_second": 0.215, | |
| "step": 8857 | |
| }, | |
| { | |
| "epoch": 5.002391304347826, | |
| "eval_accuracy": 0.9180327868852459, | |
| "eval_f1": 0.9178693897416935, | |
| "eval_loss": 0.5277689099311829, | |
| "eval_runtime": 72.002, | |
| "eval_samples_per_second": 1.694, | |
| "eval_steps_per_second": 0.222, | |
| "step": 8857 | |
| }, | |
| { | |
| "epoch": 5.002391304347826, | |
| "eval_accuracy": 0.8777292576419214, | |
| "eval_f1": 0.8759548256004502, | |
| "eval_loss": 0.9210802912712097, | |
| "eval_runtime": 134.9933, | |
| "eval_samples_per_second": 1.696, | |
| "eval_steps_per_second": 0.215, | |
| "step": 8857 | |
| }, | |
| { | |
| "epoch": 5.002391304347826, | |
| "eval_accuracy": 0.9180327868852459, | |
| "eval_f1": 0.9178693897416935, | |
| "eval_loss": 0.5277688503265381, | |
| "eval_runtime": 73.3325, | |
| "eval_samples_per_second": 1.664, | |
| "eval_steps_per_second": 0.218, | |
| "step": 8857 | |
| } | |
| ], | |
| "logging_steps": 10, | |
| "max_steps": 9200, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 9223372036854775807, | |
| "save_steps": 500, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": false | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 8.786781173362655e+19, | |
| "train_batch_size": 8, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |