| { | |
| "best_global_step": 6000, | |
| "best_metric": 21.781282824708335, | |
| "best_model_checkpoint": "./SALAMA_C7/checkpoint-6000", | |
| "epoch": 2.44408229782033, | |
| "eval_steps": 2000, | |
| "global_step": 6000, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.008148299042574863, | |
| "grad_norm": 15.915007591247559, | |
| "learning_rate": 1.9000000000000002e-06, | |
| "loss": 1.0767, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.016296598085149726, | |
| "grad_norm": 10.589001655578613, | |
| "learning_rate": 3.900000000000001e-06, | |
| "loss": 0.9639, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.024444897127724588, | |
| "grad_norm": 11.770127296447754, | |
| "learning_rate": 5.9e-06, | |
| "loss": 0.9679, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.03259319617029945, | |
| "grad_norm": 11.00736141204834, | |
| "learning_rate": 7.9e-06, | |
| "loss": 0.9264, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.04074149521287431, | |
| "grad_norm": 8.219990730285645, | |
| "learning_rate": 9.9e-06, | |
| "loss": 0.7982, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.048889794255449176, | |
| "grad_norm": 7.805874824523926, | |
| "learning_rate": 9.973847212663455e-06, | |
| "loss": 0.7226, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 0.057038093298024034, | |
| "grad_norm": 6.7659430503845215, | |
| "learning_rate": 9.946317962835514e-06, | |
| "loss": 0.7378, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 0.0651863923405989, | |
| "grad_norm": 6.5482282638549805, | |
| "learning_rate": 9.91878871300757e-06, | |
| "loss": 0.716, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 0.07333469138317376, | |
| "grad_norm": 6.729607105255127, | |
| "learning_rate": 9.891259463179629e-06, | |
| "loss": 0.7048, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 0.08148299042574862, | |
| "grad_norm": 8.517411231994629, | |
| "learning_rate": 9.863730213351686e-06, | |
| "loss": 0.7016, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.0896312894683235, | |
| "grad_norm": 6.785459041595459, | |
| "learning_rate": 9.836200963523744e-06, | |
| "loss": 0.7395, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 0.09777958851089835, | |
| "grad_norm": 6.437356472015381, | |
| "learning_rate": 9.808671713695803e-06, | |
| "loss": 0.6579, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 0.10592788755347321, | |
| "grad_norm": 6.3125762939453125, | |
| "learning_rate": 9.781142463867861e-06, | |
| "loss": 0.6969, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 0.11407618659604807, | |
| "grad_norm": 6.493209362030029, | |
| "learning_rate": 9.753613214039918e-06, | |
| "loss": 0.6338, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 0.12222448563862294, | |
| "grad_norm": 8.855860710144043, | |
| "learning_rate": 9.726083964211977e-06, | |
| "loss": 0.6657, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.1303727846811978, | |
| "grad_norm": 5.335328578948975, | |
| "learning_rate": 9.698554714384033e-06, | |
| "loss": 0.6906, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 0.13852108372377267, | |
| "grad_norm": 6.506489276885986, | |
| "learning_rate": 9.671025464556092e-06, | |
| "loss": 0.6485, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 0.14666938276634753, | |
| "grad_norm": 5.023345947265625, | |
| "learning_rate": 9.64349621472815e-06, | |
| "loss": 0.6291, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 0.15481768180892239, | |
| "grad_norm": 8.132065773010254, | |
| "learning_rate": 9.615966964900207e-06, | |
| "loss": 0.6553, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 0.16296598085149724, | |
| "grad_norm": 7.615749359130859, | |
| "learning_rate": 9.588437715072266e-06, | |
| "loss": 0.6635, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 0.1711142798940721, | |
| "grad_norm": 6.487279415130615, | |
| "learning_rate": 9.560908465244322e-06, | |
| "loss": 0.7554, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 0.179262578936647, | |
| "grad_norm": 6.409642219543457, | |
| "learning_rate": 9.53337921541638e-06, | |
| "loss": 0.5839, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 0.18741087797922185, | |
| "grad_norm": 5.239986419677734, | |
| "learning_rate": 9.505849965588438e-06, | |
| "loss": 0.7271, | |
| "step": 460 | |
| }, | |
| { | |
| "epoch": 0.1955591770217967, | |
| "grad_norm": 5.452706336975098, | |
| "learning_rate": 9.478320715760496e-06, | |
| "loss": 0.6758, | |
| "step": 480 | |
| }, | |
| { | |
| "epoch": 0.20370747606437156, | |
| "grad_norm": 4.827511787414551, | |
| "learning_rate": 9.450791465932555e-06, | |
| "loss": 0.6178, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.21185577510694642, | |
| "grad_norm": 6.320280075073242, | |
| "learning_rate": 9.423262216104611e-06, | |
| "loss": 0.6781, | |
| "step": 520 | |
| }, | |
| { | |
| "epoch": 0.22000407414952128, | |
| "grad_norm": 5.648658275604248, | |
| "learning_rate": 9.39573296627667e-06, | |
| "loss": 0.6448, | |
| "step": 540 | |
| }, | |
| { | |
| "epoch": 0.22815237319209614, | |
| "grad_norm": 6.806399822235107, | |
| "learning_rate": 9.368203716448727e-06, | |
| "loss": 0.6206, | |
| "step": 560 | |
| }, | |
| { | |
| "epoch": 0.23630067223467102, | |
| "grad_norm": 6.556358337402344, | |
| "learning_rate": 9.340674466620785e-06, | |
| "loss": 0.6105, | |
| "step": 580 | |
| }, | |
| { | |
| "epoch": 0.24444897127724588, | |
| "grad_norm": 5.21808385848999, | |
| "learning_rate": 9.313145216792842e-06, | |
| "loss": 0.651, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 0.2525972703198207, | |
| "grad_norm": 6.145657539367676, | |
| "learning_rate": 9.2856159669649e-06, | |
| "loss": 0.6504, | |
| "step": 620 | |
| }, | |
| { | |
| "epoch": 0.2607455693623956, | |
| "grad_norm": 4.201127529144287, | |
| "learning_rate": 9.258086717136959e-06, | |
| "loss": 0.5885, | |
| "step": 640 | |
| }, | |
| { | |
| "epoch": 0.2688938684049705, | |
| "grad_norm": 5.38640022277832, | |
| "learning_rate": 9.230557467309017e-06, | |
| "loss": 0.5888, | |
| "step": 660 | |
| }, | |
| { | |
| "epoch": 0.27704216744754534, | |
| "grad_norm": 5.217910289764404, | |
| "learning_rate": 9.203028217481074e-06, | |
| "loss": 0.5901, | |
| "step": 680 | |
| }, | |
| { | |
| "epoch": 0.2851904664901202, | |
| "grad_norm": 5.227973461151123, | |
| "learning_rate": 9.175498967653133e-06, | |
| "loss": 0.7178, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 0.29333876553269506, | |
| "grad_norm": 6.235045909881592, | |
| "learning_rate": 9.147969717825191e-06, | |
| "loss": 0.6711, | |
| "step": 720 | |
| }, | |
| { | |
| "epoch": 0.3014870645752699, | |
| "grad_norm": 5.588578224182129, | |
| "learning_rate": 9.120440467997248e-06, | |
| "loss": 0.522, | |
| "step": 740 | |
| }, | |
| { | |
| "epoch": 0.30963536361784477, | |
| "grad_norm": 5.49254035949707, | |
| "learning_rate": 9.092911218169306e-06, | |
| "loss": 0.5962, | |
| "step": 760 | |
| }, | |
| { | |
| "epoch": 0.31778366266041963, | |
| "grad_norm": 5.4190239906311035, | |
| "learning_rate": 9.065381968341363e-06, | |
| "loss": 0.5981, | |
| "step": 780 | |
| }, | |
| { | |
| "epoch": 0.3259319617029945, | |
| "grad_norm": 6.350463390350342, | |
| "learning_rate": 9.037852718513422e-06, | |
| "loss": 0.6114, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 0.33408026074556935, | |
| "grad_norm": 6.231777191162109, | |
| "learning_rate": 9.010323468685478e-06, | |
| "loss": 0.6729, | |
| "step": 820 | |
| }, | |
| { | |
| "epoch": 0.3422285597881442, | |
| "grad_norm": 6.590610027313232, | |
| "learning_rate": 8.982794218857537e-06, | |
| "loss": 0.6124, | |
| "step": 840 | |
| }, | |
| { | |
| "epoch": 0.35037685883071906, | |
| "grad_norm": 5.504190444946289, | |
| "learning_rate": 8.955264969029595e-06, | |
| "loss": 0.6555, | |
| "step": 860 | |
| }, | |
| { | |
| "epoch": 0.358525157873294, | |
| "grad_norm": 4.821358680725098, | |
| "learning_rate": 8.927735719201652e-06, | |
| "loss": 0.5743, | |
| "step": 880 | |
| }, | |
| { | |
| "epoch": 0.36667345691586883, | |
| "grad_norm": 6.006564140319824, | |
| "learning_rate": 8.90020646937371e-06, | |
| "loss": 0.6157, | |
| "step": 900 | |
| }, | |
| { | |
| "epoch": 0.3748217559584437, | |
| "grad_norm": 5.9578471183776855, | |
| "learning_rate": 8.872677219545767e-06, | |
| "loss": 0.6485, | |
| "step": 920 | |
| }, | |
| { | |
| "epoch": 0.38297005500101855, | |
| "grad_norm": 5.149762153625488, | |
| "learning_rate": 8.845147969717826e-06, | |
| "loss": 0.5951, | |
| "step": 940 | |
| }, | |
| { | |
| "epoch": 0.3911183540435934, | |
| "grad_norm": 6.143222332000732, | |
| "learning_rate": 8.817618719889883e-06, | |
| "loss": 0.5805, | |
| "step": 960 | |
| }, | |
| { | |
| "epoch": 0.39926665308616827, | |
| "grad_norm": 5.496342658996582, | |
| "learning_rate": 8.790089470061941e-06, | |
| "loss": 0.6158, | |
| "step": 980 | |
| }, | |
| { | |
| "epoch": 0.4074149521287431, | |
| "grad_norm": 5.14535665512085, | |
| "learning_rate": 8.762560220234e-06, | |
| "loss": 0.6036, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 0.415563251171318, | |
| "grad_norm": 5.73581600189209, | |
| "learning_rate": 8.735030970406058e-06, | |
| "loss": 0.5925, | |
| "step": 1020 | |
| }, | |
| { | |
| "epoch": 0.42371155021389284, | |
| "grad_norm": 4.455753326416016, | |
| "learning_rate": 8.707501720578115e-06, | |
| "loss": 0.5894, | |
| "step": 1040 | |
| }, | |
| { | |
| "epoch": 0.4318598492564677, | |
| "grad_norm": 6.366560935974121, | |
| "learning_rate": 8.679972470750173e-06, | |
| "loss": 0.6424, | |
| "step": 1060 | |
| }, | |
| { | |
| "epoch": 0.44000814829904256, | |
| "grad_norm": 5.11430025100708, | |
| "learning_rate": 8.652443220922232e-06, | |
| "loss": 0.5731, | |
| "step": 1080 | |
| }, | |
| { | |
| "epoch": 0.4481564473416174, | |
| "grad_norm": 6.0588812828063965, | |
| "learning_rate": 8.624913971094288e-06, | |
| "loss": 0.6257, | |
| "step": 1100 | |
| }, | |
| { | |
| "epoch": 0.4563047463841923, | |
| "grad_norm": 6.879603385925293, | |
| "learning_rate": 8.597384721266347e-06, | |
| "loss": 0.6427, | |
| "step": 1120 | |
| }, | |
| { | |
| "epoch": 0.4644530454267672, | |
| "grad_norm": 4.6673502922058105, | |
| "learning_rate": 8.569855471438404e-06, | |
| "loss": 0.5763, | |
| "step": 1140 | |
| }, | |
| { | |
| "epoch": 0.47260134446934204, | |
| "grad_norm": 5.316527843475342, | |
| "learning_rate": 8.542326221610462e-06, | |
| "loss": 0.6207, | |
| "step": 1160 | |
| }, | |
| { | |
| "epoch": 0.4807496435119169, | |
| "grad_norm": 5.4205780029296875, | |
| "learning_rate": 8.514796971782519e-06, | |
| "loss": 0.5435, | |
| "step": 1180 | |
| }, | |
| { | |
| "epoch": 0.48889794255449176, | |
| "grad_norm": 4.579510688781738, | |
| "learning_rate": 8.487267721954577e-06, | |
| "loss": 0.5937, | |
| "step": 1200 | |
| }, | |
| { | |
| "epoch": 0.4970462415970666, | |
| "grad_norm": 6.327101707458496, | |
| "learning_rate": 8.459738472126636e-06, | |
| "loss": 0.6186, | |
| "step": 1220 | |
| }, | |
| { | |
| "epoch": 0.5051945406396414, | |
| "grad_norm": 5.347973346710205, | |
| "learning_rate": 8.432209222298693e-06, | |
| "loss": 0.564, | |
| "step": 1240 | |
| }, | |
| { | |
| "epoch": 0.5133428396822164, | |
| "grad_norm": 5.44881010055542, | |
| "learning_rate": 8.404679972470751e-06, | |
| "loss": 0.5221, | |
| "step": 1260 | |
| }, | |
| { | |
| "epoch": 0.5214911387247912, | |
| "grad_norm": 5.802955150604248, | |
| "learning_rate": 8.377150722642808e-06, | |
| "loss": 0.5893, | |
| "step": 1280 | |
| }, | |
| { | |
| "epoch": 0.5296394377673661, | |
| "grad_norm": 5.931309223175049, | |
| "learning_rate": 8.349621472814866e-06, | |
| "loss": 0.6306, | |
| "step": 1300 | |
| }, | |
| { | |
| "epoch": 0.537787736809941, | |
| "grad_norm": 5.793541431427002, | |
| "learning_rate": 8.322092222986923e-06, | |
| "loss": 0.6199, | |
| "step": 1320 | |
| }, | |
| { | |
| "epoch": 0.5459360358525158, | |
| "grad_norm": 4.972778797149658, | |
| "learning_rate": 8.294562973158982e-06, | |
| "loss": 0.4771, | |
| "step": 1340 | |
| }, | |
| { | |
| "epoch": 0.5540843348950907, | |
| "grad_norm": 5.573855876922607, | |
| "learning_rate": 8.26703372333104e-06, | |
| "loss": 0.5054, | |
| "step": 1360 | |
| }, | |
| { | |
| "epoch": 0.5622326339376655, | |
| "grad_norm": 4.7720947265625, | |
| "learning_rate": 8.239504473503099e-06, | |
| "loss": 0.5599, | |
| "step": 1380 | |
| }, | |
| { | |
| "epoch": 0.5703809329802404, | |
| "grad_norm": 6.338388442993164, | |
| "learning_rate": 8.211975223675155e-06, | |
| "loss": 0.5612, | |
| "step": 1400 | |
| }, | |
| { | |
| "epoch": 0.5785292320228153, | |
| "grad_norm": 6.1034464836120605, | |
| "learning_rate": 8.184445973847214e-06, | |
| "loss": 0.5662, | |
| "step": 1420 | |
| }, | |
| { | |
| "epoch": 0.5866775310653901, | |
| "grad_norm": 5.423385143280029, | |
| "learning_rate": 8.15691672401927e-06, | |
| "loss": 0.6246, | |
| "step": 1440 | |
| }, | |
| { | |
| "epoch": 0.594825830107965, | |
| "grad_norm": 6.507946014404297, | |
| "learning_rate": 8.12938747419133e-06, | |
| "loss": 0.5879, | |
| "step": 1460 | |
| }, | |
| { | |
| "epoch": 0.6029741291505398, | |
| "grad_norm": 5.927179336547852, | |
| "learning_rate": 8.101858224363388e-06, | |
| "loss": 0.5479, | |
| "step": 1480 | |
| }, | |
| { | |
| "epoch": 0.6111224281931147, | |
| "grad_norm": 6.775645732879639, | |
| "learning_rate": 8.074328974535444e-06, | |
| "loss": 0.5809, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 0.6192707272356895, | |
| "grad_norm": 4.263314247131348, | |
| "learning_rate": 8.046799724707503e-06, | |
| "loss": 0.55, | |
| "step": 1520 | |
| }, | |
| { | |
| "epoch": 0.6274190262782644, | |
| "grad_norm": 4.671892166137695, | |
| "learning_rate": 8.01927047487956e-06, | |
| "loss": 0.529, | |
| "step": 1540 | |
| }, | |
| { | |
| "epoch": 0.6355673253208393, | |
| "grad_norm": 5.7395501136779785, | |
| "learning_rate": 7.991741225051618e-06, | |
| "loss": 0.5349, | |
| "step": 1560 | |
| }, | |
| { | |
| "epoch": 0.6437156243634141, | |
| "grad_norm": 5.88397216796875, | |
| "learning_rate": 7.964211975223675e-06, | |
| "loss": 0.5675, | |
| "step": 1580 | |
| }, | |
| { | |
| "epoch": 0.651863923405989, | |
| "grad_norm": 5.037132263183594, | |
| "learning_rate": 7.936682725395733e-06, | |
| "loss": 0.5838, | |
| "step": 1600 | |
| }, | |
| { | |
| "epoch": 0.6600122224485638, | |
| "grad_norm": 6.442205429077148, | |
| "learning_rate": 7.909153475567792e-06, | |
| "loss": 0.5275, | |
| "step": 1620 | |
| }, | |
| { | |
| "epoch": 0.6681605214911387, | |
| "grad_norm": 4.626034259796143, | |
| "learning_rate": 7.881624225739849e-06, | |
| "loss": 0.5251, | |
| "step": 1640 | |
| }, | |
| { | |
| "epoch": 0.6763088205337136, | |
| "grad_norm": 6.96089506149292, | |
| "learning_rate": 7.854094975911907e-06, | |
| "loss": 0.589, | |
| "step": 1660 | |
| }, | |
| { | |
| "epoch": 0.6844571195762884, | |
| "grad_norm": 5.303503036499023, | |
| "learning_rate": 7.826565726083964e-06, | |
| "loss": 0.6233, | |
| "step": 1680 | |
| }, | |
| { | |
| "epoch": 0.6926054186188633, | |
| "grad_norm": 6.015293121337891, | |
| "learning_rate": 7.799036476256022e-06, | |
| "loss": 0.5517, | |
| "step": 1700 | |
| }, | |
| { | |
| "epoch": 0.7007537176614381, | |
| "grad_norm": 5.348663806915283, | |
| "learning_rate": 7.77150722642808e-06, | |
| "loss": 0.5403, | |
| "step": 1720 | |
| }, | |
| { | |
| "epoch": 0.708902016704013, | |
| "grad_norm": 4.189513206481934, | |
| "learning_rate": 7.743977976600138e-06, | |
| "loss": 0.5817, | |
| "step": 1740 | |
| }, | |
| { | |
| "epoch": 0.717050315746588, | |
| "grad_norm": 7.446457386016846, | |
| "learning_rate": 7.716448726772196e-06, | |
| "loss": 0.5437, | |
| "step": 1760 | |
| }, | |
| { | |
| "epoch": 0.7251986147891628, | |
| "grad_norm": 4.759880542755127, | |
| "learning_rate": 7.688919476944255e-06, | |
| "loss": 0.5648, | |
| "step": 1780 | |
| }, | |
| { | |
| "epoch": 0.7333469138317377, | |
| "grad_norm": 5.6719584465026855, | |
| "learning_rate": 7.661390227116311e-06, | |
| "loss": 0.6118, | |
| "step": 1800 | |
| }, | |
| { | |
| "epoch": 0.7414952128743125, | |
| "grad_norm": 4.943815231323242, | |
| "learning_rate": 7.63386097728837e-06, | |
| "loss": 0.5326, | |
| "step": 1820 | |
| }, | |
| { | |
| "epoch": 0.7496435119168874, | |
| "grad_norm": 3.6254968643188477, | |
| "learning_rate": 7.6063317274604275e-06, | |
| "loss": 0.5087, | |
| "step": 1840 | |
| }, | |
| { | |
| "epoch": 0.7577918109594622, | |
| "grad_norm": 5.9594526290893555, | |
| "learning_rate": 7.578802477632485e-06, | |
| "loss": 0.532, | |
| "step": 1860 | |
| }, | |
| { | |
| "epoch": 0.7659401100020371, | |
| "grad_norm": 5.103460311889648, | |
| "learning_rate": 7.551273227804543e-06, | |
| "loss": 0.553, | |
| "step": 1880 | |
| }, | |
| { | |
| "epoch": 0.774088409044612, | |
| "grad_norm": 6.327749729156494, | |
| "learning_rate": 7.5237439779766004e-06, | |
| "loss": 0.5767, | |
| "step": 1900 | |
| }, | |
| { | |
| "epoch": 0.7822367080871868, | |
| "grad_norm": 5.8910088539123535, | |
| "learning_rate": 7.496214728148659e-06, | |
| "loss": 0.5948, | |
| "step": 1920 | |
| }, | |
| { | |
| "epoch": 0.7903850071297617, | |
| "grad_norm": 5.571481227874756, | |
| "learning_rate": 7.4686854783207165e-06, | |
| "loss": 0.5199, | |
| "step": 1940 | |
| }, | |
| { | |
| "epoch": 0.7985333061723365, | |
| "grad_norm": 4.894184589385986, | |
| "learning_rate": 7.441156228492774e-06, | |
| "loss": 0.5577, | |
| "step": 1960 | |
| }, | |
| { | |
| "epoch": 0.8066816052149114, | |
| "grad_norm": 5.402436256408691, | |
| "learning_rate": 7.413626978664832e-06, | |
| "loss": 0.5395, | |
| "step": 1980 | |
| }, | |
| { | |
| "epoch": 0.8148299042574862, | |
| "grad_norm": 5.212170600891113, | |
| "learning_rate": 7.3860977288368894e-06, | |
| "loss": 0.5319, | |
| "step": 2000 | |
| }, | |
| { | |
| "epoch": 0.8148299042574862, | |
| "eval_loss": 0.4680393934249878, | |
| "eval_runtime": 10851.5751, | |
| "eval_samples_per_second": 1.81, | |
| "eval_steps_per_second": 0.226, | |
| "eval_wer": 38.607193210094984, | |
| "step": 2000 | |
| }, | |
| { | |
| "epoch": 0.8229782033000611, | |
| "grad_norm": 5.673357963562012, | |
| "learning_rate": 7.358568479008947e-06, | |
| "loss": 0.5641, | |
| "step": 2020 | |
| }, | |
| { | |
| "epoch": 0.831126502342636, | |
| "grad_norm": 4.529426097869873, | |
| "learning_rate": 7.331039229181005e-06, | |
| "loss": 0.489, | |
| "step": 2040 | |
| }, | |
| { | |
| "epoch": 0.8392748013852108, | |
| "grad_norm": 4.871108055114746, | |
| "learning_rate": 7.303509979353063e-06, | |
| "loss": 0.5413, | |
| "step": 2060 | |
| }, | |
| { | |
| "epoch": 0.8474231004277857, | |
| "grad_norm": 4.407359600067139, | |
| "learning_rate": 7.275980729525121e-06, | |
| "loss": 0.4993, | |
| "step": 2080 | |
| }, | |
| { | |
| "epoch": 0.8555713994703605, | |
| "grad_norm": 4.965864181518555, | |
| "learning_rate": 7.248451479697178e-06, | |
| "loss": 0.5825, | |
| "step": 2100 | |
| }, | |
| { | |
| "epoch": 0.8637196985129354, | |
| "grad_norm": 6.825056552886963, | |
| "learning_rate": 7.220922229869237e-06, | |
| "loss": 0.5672, | |
| "step": 2120 | |
| }, | |
| { | |
| "epoch": 0.8718679975555103, | |
| "grad_norm": 5.5648040771484375, | |
| "learning_rate": 7.1933929800412945e-06, | |
| "loss": 0.5188, | |
| "step": 2140 | |
| }, | |
| { | |
| "epoch": 0.8800162965980851, | |
| "grad_norm": 6.5213799476623535, | |
| "learning_rate": 7.165863730213353e-06, | |
| "loss": 0.5806, | |
| "step": 2160 | |
| }, | |
| { | |
| "epoch": 0.88816459564066, | |
| "grad_norm": 5.0149993896484375, | |
| "learning_rate": 7.138334480385411e-06, | |
| "loss": 0.5729, | |
| "step": 2180 | |
| }, | |
| { | |
| "epoch": 0.8963128946832348, | |
| "grad_norm": 4.783459186553955, | |
| "learning_rate": 7.110805230557468e-06, | |
| "loss": 0.4991, | |
| "step": 2200 | |
| }, | |
| { | |
| "epoch": 0.9044611937258097, | |
| "grad_norm": 5.186647891998291, | |
| "learning_rate": 7.083275980729526e-06, | |
| "loss": 0.6553, | |
| "step": 2220 | |
| }, | |
| { | |
| "epoch": 0.9126094927683845, | |
| "grad_norm": 4.135760307312012, | |
| "learning_rate": 7.0557467309015835e-06, | |
| "loss": 0.5947, | |
| "step": 2240 | |
| }, | |
| { | |
| "epoch": 0.9207577918109595, | |
| "grad_norm": 5.20528507232666, | |
| "learning_rate": 7.028217481073641e-06, | |
| "loss": 0.5454, | |
| "step": 2260 | |
| }, | |
| { | |
| "epoch": 0.9289060908535344, | |
| "grad_norm": 5.948816299438477, | |
| "learning_rate": 7.000688231245699e-06, | |
| "loss": 0.5475, | |
| "step": 2280 | |
| }, | |
| { | |
| "epoch": 0.9370543898961092, | |
| "grad_norm": 5.592283248901367, | |
| "learning_rate": 6.973158981417757e-06, | |
| "loss": 0.5153, | |
| "step": 2300 | |
| }, | |
| { | |
| "epoch": 0.9452026889386841, | |
| "grad_norm": 5.162081241607666, | |
| "learning_rate": 6.945629731589815e-06, | |
| "loss": 0.4637, | |
| "step": 2320 | |
| }, | |
| { | |
| "epoch": 0.953350987981259, | |
| "grad_norm": 5.888420104980469, | |
| "learning_rate": 6.9181004817618725e-06, | |
| "loss": 0.4956, | |
| "step": 2340 | |
| }, | |
| { | |
| "epoch": 0.9614992870238338, | |
| "grad_norm": 6.354236602783203, | |
| "learning_rate": 6.89057123193393e-06, | |
| "loss": 0.4912, | |
| "step": 2360 | |
| }, | |
| { | |
| "epoch": 0.9696475860664087, | |
| "grad_norm": 4.269996166229248, | |
| "learning_rate": 6.863041982105988e-06, | |
| "loss": 0.489, | |
| "step": 2380 | |
| }, | |
| { | |
| "epoch": 0.9777958851089835, | |
| "grad_norm": 5.822803974151611, | |
| "learning_rate": 6.835512732278045e-06, | |
| "loss": 0.55, | |
| "step": 2400 | |
| }, | |
| { | |
| "epoch": 0.9859441841515584, | |
| "grad_norm": 4.092251777648926, | |
| "learning_rate": 6.807983482450103e-06, | |
| "loss": 0.5625, | |
| "step": 2420 | |
| }, | |
| { | |
| "epoch": 0.9940924831941332, | |
| "grad_norm": 6.759193420410156, | |
| "learning_rate": 6.7804542326221615e-06, | |
| "loss": 0.5363, | |
| "step": 2440 | |
| }, | |
| { | |
| "epoch": 1.0020370747606437, | |
| "grad_norm": 5.126402378082275, | |
| "learning_rate": 6.752924982794219e-06, | |
| "loss": 0.535, | |
| "step": 2460 | |
| }, | |
| { | |
| "epoch": 1.0101853738032185, | |
| "grad_norm": 5.261082172393799, | |
| "learning_rate": 6.725395732966277e-06, | |
| "loss": 0.4262, | |
| "step": 2480 | |
| }, | |
| { | |
| "epoch": 1.0183336728457935, | |
| "grad_norm": 4.407663345336914, | |
| "learning_rate": 6.697866483138335e-06, | |
| "loss": 0.4076, | |
| "step": 2500 | |
| }, | |
| { | |
| "epoch": 1.0264819718883682, | |
| "grad_norm": 5.785089492797852, | |
| "learning_rate": 6.670337233310394e-06, | |
| "loss": 0.4255, | |
| "step": 2520 | |
| }, | |
| { | |
| "epoch": 1.0346302709309432, | |
| "grad_norm": 4.659370422363281, | |
| "learning_rate": 6.642807983482451e-06, | |
| "loss": 0.4116, | |
| "step": 2540 | |
| }, | |
| { | |
| "epoch": 1.042778569973518, | |
| "grad_norm": 5.880095958709717, | |
| "learning_rate": 6.615278733654509e-06, | |
| "loss": 0.4943, | |
| "step": 2560 | |
| }, | |
| { | |
| "epoch": 1.050926869016093, | |
| "grad_norm": 4.94119930267334, | |
| "learning_rate": 6.587749483826567e-06, | |
| "loss": 0.3904, | |
| "step": 2580 | |
| }, | |
| { | |
| "epoch": 1.0590751680586679, | |
| "grad_norm": 4.008668422698975, | |
| "learning_rate": 6.560220233998624e-06, | |
| "loss": 0.403, | |
| "step": 2600 | |
| }, | |
| { | |
| "epoch": 1.0672234671012426, | |
| "grad_norm": 6.079473495483398, | |
| "learning_rate": 6.532690984170682e-06, | |
| "loss": 0.4833, | |
| "step": 2620 | |
| }, | |
| { | |
| "epoch": 1.0753717661438176, | |
| "grad_norm": 5.263147830963135, | |
| "learning_rate": 6.5051617343427395e-06, | |
| "loss": 0.4423, | |
| "step": 2640 | |
| }, | |
| { | |
| "epoch": 1.0835200651863923, | |
| "grad_norm": 4.1324543952941895, | |
| "learning_rate": 6.477632484514798e-06, | |
| "loss": 0.4631, | |
| "step": 2660 | |
| }, | |
| { | |
| "epoch": 1.0916683642289673, | |
| "grad_norm": 3.648576259613037, | |
| "learning_rate": 6.450103234686856e-06, | |
| "loss": 0.4002, | |
| "step": 2680 | |
| }, | |
| { | |
| "epoch": 1.099816663271542, | |
| "grad_norm": 5.458010673522949, | |
| "learning_rate": 6.422573984858913e-06, | |
| "loss": 0.4572, | |
| "step": 2700 | |
| }, | |
| { | |
| "epoch": 1.107964962314117, | |
| "grad_norm": 5.379840850830078, | |
| "learning_rate": 6.395044735030971e-06, | |
| "loss": 0.4283, | |
| "step": 2720 | |
| }, | |
| { | |
| "epoch": 1.1161132613566918, | |
| "grad_norm": 5.2092108726501465, | |
| "learning_rate": 6.3675154852030285e-06, | |
| "loss": 0.4384, | |
| "step": 2740 | |
| }, | |
| { | |
| "epoch": 1.1242615603992667, | |
| "grad_norm": 5.582503795623779, | |
| "learning_rate": 6.339986235375086e-06, | |
| "loss": 0.4202, | |
| "step": 2760 | |
| }, | |
| { | |
| "epoch": 1.1324098594418415, | |
| "grad_norm": 4.186822891235352, | |
| "learning_rate": 6.312456985547144e-06, | |
| "loss": 0.3972, | |
| "step": 2780 | |
| }, | |
| { | |
| "epoch": 1.1405581584844164, | |
| "grad_norm": 5.029235363006592, | |
| "learning_rate": 6.284927735719201e-06, | |
| "loss": 0.4368, | |
| "step": 2800 | |
| }, | |
| { | |
| "epoch": 1.1487064575269912, | |
| "grad_norm": 4.080201148986816, | |
| "learning_rate": 6.25739848589126e-06, | |
| "loss": 0.4561, | |
| "step": 2820 | |
| }, | |
| { | |
| "epoch": 1.1568547565695662, | |
| "grad_norm": 5.171266078948975, | |
| "learning_rate": 6.2298692360633175e-06, | |
| "loss": 0.4594, | |
| "step": 2840 | |
| }, | |
| { | |
| "epoch": 1.165003055612141, | |
| "grad_norm": 6.01788854598999, | |
| "learning_rate": 6.202339986235376e-06, | |
| "loss": 0.363, | |
| "step": 2860 | |
| }, | |
| { | |
| "epoch": 1.1731513546547159, | |
| "grad_norm": 5.374858379364014, | |
| "learning_rate": 6.1748107364074336e-06, | |
| "loss": 0.4782, | |
| "step": 2880 | |
| }, | |
| { | |
| "epoch": 1.1812996536972906, | |
| "grad_norm": 4.696728706359863, | |
| "learning_rate": 6.147281486579492e-06, | |
| "loss": 0.4598, | |
| "step": 2900 | |
| }, | |
| { | |
| "epoch": 1.1894479527398656, | |
| "grad_norm": 4.785463333129883, | |
| "learning_rate": 6.11975223675155e-06, | |
| "loss": 0.4502, | |
| "step": 2920 | |
| }, | |
| { | |
| "epoch": 1.1975962517824403, | |
| "grad_norm": 4.824402809143066, | |
| "learning_rate": 6.092222986923607e-06, | |
| "loss": 0.4216, | |
| "step": 2940 | |
| }, | |
| { | |
| "epoch": 1.2057445508250153, | |
| "grad_norm": 4.554792404174805, | |
| "learning_rate": 6.064693737095665e-06, | |
| "loss": 0.39, | |
| "step": 2960 | |
| }, | |
| { | |
| "epoch": 1.21389284986759, | |
| "grad_norm": 5.259772300720215, | |
| "learning_rate": 6.0371644872677226e-06, | |
| "loss": 0.4199, | |
| "step": 2980 | |
| }, | |
| { | |
| "epoch": 1.222041148910165, | |
| "grad_norm": 4.8776140213012695, | |
| "learning_rate": 6.00963523743978e-06, | |
| "loss": 0.4742, | |
| "step": 3000 | |
| }, | |
| { | |
| "epoch": 1.23018944795274, | |
| "grad_norm": 5.5702033042907715, | |
| "learning_rate": 5.982105987611838e-06, | |
| "loss": 0.3934, | |
| "step": 3020 | |
| }, | |
| { | |
| "epoch": 1.2383377469953147, | |
| "grad_norm": 5.325311660766602, | |
| "learning_rate": 5.954576737783896e-06, | |
| "loss": 0.3655, | |
| "step": 3040 | |
| }, | |
| { | |
| "epoch": 1.2464860460378895, | |
| "grad_norm": 3.9275801181793213, | |
| "learning_rate": 5.927047487955954e-06, | |
| "loss": 0.3965, | |
| "step": 3060 | |
| }, | |
| { | |
| "epoch": 1.2546343450804645, | |
| "grad_norm": 4.71364688873291, | |
| "learning_rate": 5.8995182381280116e-06, | |
| "loss": 0.4026, | |
| "step": 3080 | |
| }, | |
| { | |
| "epoch": 1.2627826441230394, | |
| "grad_norm": 4.507187843322754, | |
| "learning_rate": 5.871988988300069e-06, | |
| "loss": 0.4004, | |
| "step": 3100 | |
| }, | |
| { | |
| "epoch": 1.2709309431656142, | |
| "grad_norm": 4.939427375793457, | |
| "learning_rate": 5.844459738472127e-06, | |
| "loss": 0.407, | |
| "step": 3120 | |
| }, | |
| { | |
| "epoch": 1.279079242208189, | |
| "grad_norm": 5.642360210418701, | |
| "learning_rate": 5.8169304886441844e-06, | |
| "loss": 0.4273, | |
| "step": 3140 | |
| }, | |
| { | |
| "epoch": 1.2872275412507639, | |
| "grad_norm": 4.531342506408691, | |
| "learning_rate": 5.789401238816242e-06, | |
| "loss": 0.4455, | |
| "step": 3160 | |
| }, | |
| { | |
| "epoch": 1.2953758402933389, | |
| "grad_norm": 5.926376819610596, | |
| "learning_rate": 5.7618719889883005e-06, | |
| "loss": 0.4627, | |
| "step": 3180 | |
| }, | |
| { | |
| "epoch": 1.3035241393359136, | |
| "grad_norm": 3.818305730819702, | |
| "learning_rate": 5.734342739160358e-06, | |
| "loss": 0.3957, | |
| "step": 3200 | |
| }, | |
| { | |
| "epoch": 1.3116724383784886, | |
| "grad_norm": 4.200014114379883, | |
| "learning_rate": 5.706813489332416e-06, | |
| "loss": 0.4516, | |
| "step": 3220 | |
| }, | |
| { | |
| "epoch": 1.3198207374210633, | |
| "grad_norm": 5.380039691925049, | |
| "learning_rate": 5.679284239504474e-06, | |
| "loss": 0.4775, | |
| "step": 3240 | |
| }, | |
| { | |
| "epoch": 1.3279690364636383, | |
| "grad_norm": 6.382922172546387, | |
| "learning_rate": 5.651754989676532e-06, | |
| "loss": 0.3792, | |
| "step": 3260 | |
| }, | |
| { | |
| "epoch": 1.336117335506213, | |
| "grad_norm": 5.670558929443359, | |
| "learning_rate": 5.62422573984859e-06, | |
| "loss": 0.4301, | |
| "step": 3280 | |
| }, | |
| { | |
| "epoch": 1.344265634548788, | |
| "grad_norm": 4.901285171508789, | |
| "learning_rate": 5.596696490020648e-06, | |
| "loss": 0.4095, | |
| "step": 3300 | |
| }, | |
| { | |
| "epoch": 1.3524139335913627, | |
| "grad_norm": 5.272710800170898, | |
| "learning_rate": 5.569167240192706e-06, | |
| "loss": 0.3585, | |
| "step": 3320 | |
| }, | |
| { | |
| "epoch": 1.3605622326339377, | |
| "grad_norm": 5.442037582397461, | |
| "learning_rate": 5.541637990364763e-06, | |
| "loss": 0.4019, | |
| "step": 3340 | |
| }, | |
| { | |
| "epoch": 1.3687105316765125, | |
| "grad_norm": 4.817733287811279, | |
| "learning_rate": 5.514108740536821e-06, | |
| "loss": 0.368, | |
| "step": 3360 | |
| }, | |
| { | |
| "epoch": 1.3768588307190874, | |
| "grad_norm": 4.459897994995117, | |
| "learning_rate": 5.4865794907088785e-06, | |
| "loss": 0.3646, | |
| "step": 3380 | |
| }, | |
| { | |
| "epoch": 1.3850071297616622, | |
| "grad_norm": 5.927896976470947, | |
| "learning_rate": 5.459050240880936e-06, | |
| "loss": 0.4448, | |
| "step": 3400 | |
| }, | |
| { | |
| "epoch": 1.3931554288042372, | |
| "grad_norm": 5.564925193786621, | |
| "learning_rate": 5.431520991052995e-06, | |
| "loss": 0.3794, | |
| "step": 3420 | |
| }, | |
| { | |
| "epoch": 1.401303727846812, | |
| "grad_norm": 4.68287467956543, | |
| "learning_rate": 5.403991741225052e-06, | |
| "loss": 0.3931, | |
| "step": 3440 | |
| }, | |
| { | |
| "epoch": 1.4094520268893869, | |
| "grad_norm": 5.004433631896973, | |
| "learning_rate": 5.37646249139711e-06, | |
| "loss": 0.432, | |
| "step": 3460 | |
| }, | |
| { | |
| "epoch": 1.4176003259319616, | |
| "grad_norm": 5.418098449707031, | |
| "learning_rate": 5.3489332415691675e-06, | |
| "loss": 0.4533, | |
| "step": 3480 | |
| }, | |
| { | |
| "epoch": 1.4257486249745366, | |
| "grad_norm": 4.825730800628662, | |
| "learning_rate": 5.321403991741225e-06, | |
| "loss": 0.4159, | |
| "step": 3500 | |
| }, | |
| { | |
| "epoch": 1.4338969240171116, | |
| "grad_norm": 5.892017364501953, | |
| "learning_rate": 5.293874741913283e-06, | |
| "loss": 0.394, | |
| "step": 3520 | |
| }, | |
| { | |
| "epoch": 1.4420452230596863, | |
| "grad_norm": 5.600419521331787, | |
| "learning_rate": 5.26634549208534e-06, | |
| "loss": 0.4357, | |
| "step": 3540 | |
| }, | |
| { | |
| "epoch": 1.450193522102261, | |
| "grad_norm": 4.615939617156982, | |
| "learning_rate": 5.238816242257399e-06, | |
| "loss": 0.4029, | |
| "step": 3560 | |
| }, | |
| { | |
| "epoch": 1.458341821144836, | |
| "grad_norm": 4.384542465209961, | |
| "learning_rate": 5.2112869924294565e-06, | |
| "loss": 0.4042, | |
| "step": 3580 | |
| }, | |
| { | |
| "epoch": 1.466490120187411, | |
| "grad_norm": 5.981232643127441, | |
| "learning_rate": 5.183757742601514e-06, | |
| "loss": 0.3923, | |
| "step": 3600 | |
| }, | |
| { | |
| "epoch": 1.4746384192299857, | |
| "grad_norm": 5.077232837677002, | |
| "learning_rate": 5.156228492773573e-06, | |
| "loss": 0.4995, | |
| "step": 3620 | |
| }, | |
| { | |
| "epoch": 1.4827867182725605, | |
| "grad_norm": 5.000918865203857, | |
| "learning_rate": 5.12869924294563e-06, | |
| "loss": 0.4167, | |
| "step": 3640 | |
| }, | |
| { | |
| "epoch": 1.4909350173151354, | |
| "grad_norm": 2.953489303588867, | |
| "learning_rate": 5.101169993117689e-06, | |
| "loss": 0.3798, | |
| "step": 3660 | |
| }, | |
| { | |
| "epoch": 1.4990833163577104, | |
| "grad_norm": 5.880030632019043, | |
| "learning_rate": 5.073640743289746e-06, | |
| "loss": 0.4349, | |
| "step": 3680 | |
| }, | |
| { | |
| "epoch": 1.5072316154002852, | |
| "grad_norm": 5.726639270782471, | |
| "learning_rate": 5.046111493461804e-06, | |
| "loss": 0.3809, | |
| "step": 3700 | |
| }, | |
| { | |
| "epoch": 1.51537991444286, | |
| "grad_norm": 4.755875587463379, | |
| "learning_rate": 5.018582243633862e-06, | |
| "loss": 0.4345, | |
| "step": 3720 | |
| }, | |
| { | |
| "epoch": 1.5235282134854349, | |
| "grad_norm": 3.825904130935669, | |
| "learning_rate": 4.991052993805919e-06, | |
| "loss": 0.4304, | |
| "step": 3740 | |
| }, | |
| { | |
| "epoch": 1.5316765125280098, | |
| "grad_norm": 6.1705217361450195, | |
| "learning_rate": 4.963523743977977e-06, | |
| "loss": 0.4224, | |
| "step": 3760 | |
| }, | |
| { | |
| "epoch": 1.5398248115705846, | |
| "grad_norm": 5.098936557769775, | |
| "learning_rate": 4.9359944941500345e-06, | |
| "loss": 0.4254, | |
| "step": 3780 | |
| }, | |
| { | |
| "epoch": 1.5479731106131593, | |
| "grad_norm": 4.633610248565674, | |
| "learning_rate": 4.908465244322093e-06, | |
| "loss": 0.3542, | |
| "step": 3800 | |
| }, | |
| { | |
| "epoch": 1.5561214096557343, | |
| "grad_norm": 5.469082832336426, | |
| "learning_rate": 4.880935994494151e-06, | |
| "loss": 0.3577, | |
| "step": 3820 | |
| }, | |
| { | |
| "epoch": 1.5642697086983093, | |
| "grad_norm": 4.476110458374023, | |
| "learning_rate": 4.853406744666208e-06, | |
| "loss": 0.3931, | |
| "step": 3840 | |
| }, | |
| { | |
| "epoch": 1.5724180077408842, | |
| "grad_norm": 4.693641185760498, | |
| "learning_rate": 4.825877494838266e-06, | |
| "loss": 0.3793, | |
| "step": 3860 | |
| }, | |
| { | |
| "epoch": 1.580566306783459, | |
| "grad_norm": 5.025706768035889, | |
| "learning_rate": 4.7983482450103235e-06, | |
| "loss": 0.4152, | |
| "step": 3880 | |
| }, | |
| { | |
| "epoch": 1.5887146058260337, | |
| "grad_norm": 5.075003147125244, | |
| "learning_rate": 4.770818995182382e-06, | |
| "loss": 0.3813, | |
| "step": 3900 | |
| }, | |
| { | |
| "epoch": 1.5968629048686087, | |
| "grad_norm": 4.8126115798950195, | |
| "learning_rate": 4.74328974535444e-06, | |
| "loss": 0.408, | |
| "step": 3920 | |
| }, | |
| { | |
| "epoch": 1.6050112039111837, | |
| "grad_norm": 4.3969340324401855, | |
| "learning_rate": 4.715760495526497e-06, | |
| "loss": 0.3667, | |
| "step": 3940 | |
| }, | |
| { | |
| "epoch": 1.6131595029537584, | |
| "grad_norm": 4.514830589294434, | |
| "learning_rate": 4.688231245698555e-06, | |
| "loss": 0.4022, | |
| "step": 3960 | |
| }, | |
| { | |
| "epoch": 1.6213078019963332, | |
| "grad_norm": 5.9185309410095215, | |
| "learning_rate": 4.6607019958706125e-06, | |
| "loss": 0.4277, | |
| "step": 3980 | |
| }, | |
| { | |
| "epoch": 1.6294561010389081, | |
| "grad_norm": 4.446202754974365, | |
| "learning_rate": 4.633172746042671e-06, | |
| "loss": 0.4179, | |
| "step": 4000 | |
| }, | |
| { | |
| "epoch": 1.6294561010389081, | |
| "eval_loss": 0.3749663531780243, | |
| "eval_runtime": 13864.4556, | |
| "eval_samples_per_second": 1.416, | |
| "eval_steps_per_second": 0.177, | |
| "eval_wer": 60.07223542231883, | |
| "step": 4000 | |
| }, | |
| { | |
| "epoch": 1.6376044000814831, | |
| "grad_norm": 5.807653903961182, | |
| "learning_rate": 4.605643496214729e-06, | |
| "loss": 0.4408, | |
| "step": 4020 | |
| }, | |
| { | |
| "epoch": 1.6457526991240579, | |
| "grad_norm": 7.353831768035889, | |
| "learning_rate": 4.578114246386786e-06, | |
| "loss": 0.4145, | |
| "step": 4040 | |
| }, | |
| { | |
| "epoch": 1.6539009981666326, | |
| "grad_norm": 3.9637539386749268, | |
| "learning_rate": 4.550584996558844e-06, | |
| "loss": 0.4446, | |
| "step": 4060 | |
| }, | |
| { | |
| "epoch": 1.6620492972092076, | |
| "grad_norm": 5.770019054412842, | |
| "learning_rate": 4.523055746730902e-06, | |
| "loss": 0.428, | |
| "step": 4080 | |
| }, | |
| { | |
| "epoch": 1.6701975962517825, | |
| "grad_norm": 6.5047688484191895, | |
| "learning_rate": 4.49552649690296e-06, | |
| "loss": 0.4178, | |
| "step": 4100 | |
| }, | |
| { | |
| "epoch": 1.6783458952943573, | |
| "grad_norm": 6.010311603546143, | |
| "learning_rate": 4.467997247075018e-06, | |
| "loss": 0.4164, | |
| "step": 4120 | |
| }, | |
| { | |
| "epoch": 1.686494194336932, | |
| "grad_norm": 5.805934906005859, | |
| "learning_rate": 4.440467997247075e-06, | |
| "loss": 0.388, | |
| "step": 4140 | |
| }, | |
| { | |
| "epoch": 1.694642493379507, | |
| "grad_norm": 5.374007701873779, | |
| "learning_rate": 4.412938747419133e-06, | |
| "loss": 0.4259, | |
| "step": 4160 | |
| }, | |
| { | |
| "epoch": 1.702790792422082, | |
| "grad_norm": 5.766719341278076, | |
| "learning_rate": 4.385409497591191e-06, | |
| "loss": 0.4388, | |
| "step": 4180 | |
| }, | |
| { | |
| "epoch": 1.7109390914646567, | |
| "grad_norm": 3.3628714084625244, | |
| "learning_rate": 4.357880247763249e-06, | |
| "loss": 0.3568, | |
| "step": 4200 | |
| }, | |
| { | |
| "epoch": 1.7190873905072315, | |
| "grad_norm": 3.830449342727661, | |
| "learning_rate": 4.3303509979353066e-06, | |
| "loss": 0.391, | |
| "step": 4220 | |
| }, | |
| { | |
| "epoch": 1.7272356895498064, | |
| "grad_norm": 5.824588298797607, | |
| "learning_rate": 4.302821748107364e-06, | |
| "loss": 0.4416, | |
| "step": 4240 | |
| }, | |
| { | |
| "epoch": 1.7353839885923814, | |
| "grad_norm": 5.445583343505859, | |
| "learning_rate": 4.275292498279423e-06, | |
| "loss": 0.3935, | |
| "step": 4260 | |
| }, | |
| { | |
| "epoch": 1.7435322876349562, | |
| "grad_norm": 3.97123646736145, | |
| "learning_rate": 4.24776324845148e-06, | |
| "loss": 0.3596, | |
| "step": 4280 | |
| }, | |
| { | |
| "epoch": 1.751680586677531, | |
| "grad_norm": 3.1131341457366943, | |
| "learning_rate": 4.220233998623538e-06, | |
| "loss": 0.4498, | |
| "step": 4300 | |
| }, | |
| { | |
| "epoch": 1.7598288857201059, | |
| "grad_norm": 3.785813093185425, | |
| "learning_rate": 4.1927047487955956e-06, | |
| "loss": 0.389, | |
| "step": 4320 | |
| }, | |
| { | |
| "epoch": 1.7679771847626808, | |
| "grad_norm": 4.826812744140625, | |
| "learning_rate": 4.165175498967653e-06, | |
| "loss": 0.4157, | |
| "step": 4340 | |
| }, | |
| { | |
| "epoch": 1.7761254838052558, | |
| "grad_norm": 5.719616889953613, | |
| "learning_rate": 4.137646249139711e-06, | |
| "loss": 0.333, | |
| "step": 4360 | |
| }, | |
| { | |
| "epoch": 1.7842737828478306, | |
| "grad_norm": 5.590183734893799, | |
| "learning_rate": 4.110116999311769e-06, | |
| "loss": 0.4176, | |
| "step": 4380 | |
| }, | |
| { | |
| "epoch": 1.7924220818904053, | |
| "grad_norm": 7.032625675201416, | |
| "learning_rate": 4.082587749483827e-06, | |
| "loss": 0.3992, | |
| "step": 4400 | |
| }, | |
| { | |
| "epoch": 1.8005703809329803, | |
| "grad_norm": 5.545759677886963, | |
| "learning_rate": 4.0550584996558846e-06, | |
| "loss": 0.4189, | |
| "step": 4420 | |
| }, | |
| { | |
| "epoch": 1.8087186799755552, | |
| "grad_norm": 6.759115219116211, | |
| "learning_rate": 4.027529249827942e-06, | |
| "loss": 0.413, | |
| "step": 4440 | |
| }, | |
| { | |
| "epoch": 1.81686697901813, | |
| "grad_norm": 5.781651973724365, | |
| "learning_rate": 4.000000000000001e-06, | |
| "loss": 0.4273, | |
| "step": 4460 | |
| }, | |
| { | |
| "epoch": 1.8250152780607047, | |
| "grad_norm": 4.690906524658203, | |
| "learning_rate": 3.972470750172058e-06, | |
| "loss": 0.3943, | |
| "step": 4480 | |
| }, | |
| { | |
| "epoch": 1.8331635771032797, | |
| "grad_norm": 5.967350006103516, | |
| "learning_rate": 3.944941500344116e-06, | |
| "loss": 0.3424, | |
| "step": 4500 | |
| }, | |
| { | |
| "epoch": 1.8413118761458547, | |
| "grad_norm": 5.927072048187256, | |
| "learning_rate": 3.9174122505161736e-06, | |
| "loss": 0.3857, | |
| "step": 4520 | |
| }, | |
| { | |
| "epoch": 1.8494601751884294, | |
| "grad_norm": 3.1506669521331787, | |
| "learning_rate": 3.889883000688231e-06, | |
| "loss": 0.3751, | |
| "step": 4540 | |
| }, | |
| { | |
| "epoch": 1.8576084742310042, | |
| "grad_norm": 3.858947992324829, | |
| "learning_rate": 3.86235375086029e-06, | |
| "loss": 0.3735, | |
| "step": 4560 | |
| }, | |
| { | |
| "epoch": 1.8657567732735791, | |
| "grad_norm": 6.204343318939209, | |
| "learning_rate": 3.834824501032347e-06, | |
| "loss": 0.4406, | |
| "step": 4580 | |
| }, | |
| { | |
| "epoch": 1.873905072316154, | |
| "grad_norm": 5.989670276641846, | |
| "learning_rate": 3.807295251204405e-06, | |
| "loss": 0.4468, | |
| "step": 4600 | |
| }, | |
| { | |
| "epoch": 1.8820533713587289, | |
| "grad_norm": 5.673979759216309, | |
| "learning_rate": 3.7797660013764625e-06, | |
| "loss": 0.4066, | |
| "step": 4620 | |
| }, | |
| { | |
| "epoch": 1.8902016704013036, | |
| "grad_norm": 4.2719197273254395, | |
| "learning_rate": 3.752236751548521e-06, | |
| "loss": 0.4463, | |
| "step": 4640 | |
| }, | |
| { | |
| "epoch": 1.8983499694438786, | |
| "grad_norm": 4.865420818328857, | |
| "learning_rate": 3.7247075017205787e-06, | |
| "loss": 0.4532, | |
| "step": 4660 | |
| }, | |
| { | |
| "epoch": 1.9064982684864535, | |
| "grad_norm": 4.259848117828369, | |
| "learning_rate": 3.6971782518926363e-06, | |
| "loss": 0.462, | |
| "step": 4680 | |
| }, | |
| { | |
| "epoch": 1.9146465675290283, | |
| "grad_norm": 3.7624459266662598, | |
| "learning_rate": 3.669649002064694e-06, | |
| "loss": 0.4175, | |
| "step": 4700 | |
| }, | |
| { | |
| "epoch": 1.922794866571603, | |
| "grad_norm": 4.538738250732422, | |
| "learning_rate": 3.642119752236752e-06, | |
| "loss": 0.3381, | |
| "step": 4720 | |
| }, | |
| { | |
| "epoch": 1.930943165614178, | |
| "grad_norm": 3.968751907348633, | |
| "learning_rate": 3.6145905024088096e-06, | |
| "loss": 0.3636, | |
| "step": 4740 | |
| }, | |
| { | |
| "epoch": 1.939091464656753, | |
| "grad_norm": 4.724824905395508, | |
| "learning_rate": 3.5870612525808672e-06, | |
| "loss": 0.3928, | |
| "step": 4760 | |
| }, | |
| { | |
| "epoch": 1.9472397636993277, | |
| "grad_norm": 4.30732536315918, | |
| "learning_rate": 3.5595320027529253e-06, | |
| "loss": 0.4176, | |
| "step": 4780 | |
| }, | |
| { | |
| "epoch": 1.9553880627419025, | |
| "grad_norm": 5.967071533203125, | |
| "learning_rate": 3.532002752924983e-06, | |
| "loss": 0.3979, | |
| "step": 4800 | |
| }, | |
| { | |
| "epoch": 1.9635363617844774, | |
| "grad_norm": 5.40061092376709, | |
| "learning_rate": 3.504473503097041e-06, | |
| "loss": 0.4085, | |
| "step": 4820 | |
| }, | |
| { | |
| "epoch": 1.9716846608270524, | |
| "grad_norm": 6.015532970428467, | |
| "learning_rate": 3.476944253269099e-06, | |
| "loss": 0.3589, | |
| "step": 4840 | |
| }, | |
| { | |
| "epoch": 1.9798329598696274, | |
| "grad_norm": 4.844973087310791, | |
| "learning_rate": 3.4494150034411566e-06, | |
| "loss": 0.4034, | |
| "step": 4860 | |
| }, | |
| { | |
| "epoch": 1.9879812589122021, | |
| "grad_norm": 3.5679476261138916, | |
| "learning_rate": 3.4218857536132143e-06, | |
| "loss": 0.3458, | |
| "step": 4880 | |
| }, | |
| { | |
| "epoch": 1.9961295579547769, | |
| "grad_norm": 5.0959153175354, | |
| "learning_rate": 3.3943565037852723e-06, | |
| "loss": 0.3818, | |
| "step": 4900 | |
| }, | |
| { | |
| "epoch": 2.0040741495212875, | |
| "grad_norm": 4.379068851470947, | |
| "learning_rate": 3.36682725395733e-06, | |
| "loss": 0.3527, | |
| "step": 4920 | |
| }, | |
| { | |
| "epoch": 2.0122224485638625, | |
| "grad_norm": 4.350337982177734, | |
| "learning_rate": 3.3392980041293876e-06, | |
| "loss": 0.348, | |
| "step": 4940 | |
| }, | |
| { | |
| "epoch": 2.020370747606437, | |
| "grad_norm": 3.3158302307128906, | |
| "learning_rate": 3.311768754301445e-06, | |
| "loss": 0.2506, | |
| "step": 4960 | |
| }, | |
| { | |
| "epoch": 2.028519046649012, | |
| "grad_norm": 5.536191463470459, | |
| "learning_rate": 3.2842395044735033e-06, | |
| "loss": 0.3206, | |
| "step": 4980 | |
| }, | |
| { | |
| "epoch": 2.036667345691587, | |
| "grad_norm": 7.019426345825195, | |
| "learning_rate": 3.256710254645561e-06, | |
| "loss": 0.3724, | |
| "step": 5000 | |
| }, | |
| { | |
| "epoch": 2.044815644734162, | |
| "grad_norm": 4.025884628295898, | |
| "learning_rate": 3.2291810048176194e-06, | |
| "loss": 0.3062, | |
| "step": 5020 | |
| }, | |
| { | |
| "epoch": 2.0529639437767364, | |
| "grad_norm": 4.956037521362305, | |
| "learning_rate": 3.201651754989677e-06, | |
| "loss": 0.3124, | |
| "step": 5040 | |
| }, | |
| { | |
| "epoch": 2.0611122428193114, | |
| "grad_norm": 6.105999946594238, | |
| "learning_rate": 3.1741225051617346e-06, | |
| "loss": 0.3047, | |
| "step": 5060 | |
| }, | |
| { | |
| "epoch": 2.0692605418618863, | |
| "grad_norm": 5.956209182739258, | |
| "learning_rate": 3.1465932553337922e-06, | |
| "loss": 0.3221, | |
| "step": 5080 | |
| }, | |
| { | |
| "epoch": 2.0774088409044613, | |
| "grad_norm": 5.465188026428223, | |
| "learning_rate": 3.1190640055058503e-06, | |
| "loss": 0.2867, | |
| "step": 5100 | |
| }, | |
| { | |
| "epoch": 2.085557139947036, | |
| "grad_norm": 5.349979877471924, | |
| "learning_rate": 3.091534755677908e-06, | |
| "loss": 0.323, | |
| "step": 5120 | |
| }, | |
| { | |
| "epoch": 2.093705438989611, | |
| "grad_norm": 4.8694305419921875, | |
| "learning_rate": 3.0640055058499656e-06, | |
| "loss": 0.3486, | |
| "step": 5140 | |
| }, | |
| { | |
| "epoch": 2.101853738032186, | |
| "grad_norm": 4.890730857849121, | |
| "learning_rate": 3.0364762560220236e-06, | |
| "loss": 0.3654, | |
| "step": 5160 | |
| }, | |
| { | |
| "epoch": 2.1100020370747608, | |
| "grad_norm": 4.631341457366943, | |
| "learning_rate": 3.0089470061940812e-06, | |
| "loss": 0.2903, | |
| "step": 5180 | |
| }, | |
| { | |
| "epoch": 2.1181503361173357, | |
| "grad_norm": 3.8383870124816895, | |
| "learning_rate": 2.9814177563661393e-06, | |
| "loss": 0.2885, | |
| "step": 5200 | |
| }, | |
| { | |
| "epoch": 2.1262986351599102, | |
| "grad_norm": 4.418496131896973, | |
| "learning_rate": 2.9538885065381973e-06, | |
| "loss": 0.3318, | |
| "step": 5220 | |
| }, | |
| { | |
| "epoch": 2.134446934202485, | |
| "grad_norm": 5.5654401779174805, | |
| "learning_rate": 2.926359256710255e-06, | |
| "loss": 0.3022, | |
| "step": 5240 | |
| }, | |
| { | |
| "epoch": 2.14259523324506, | |
| "grad_norm": 3.4770777225494385, | |
| "learning_rate": 2.8988300068823126e-06, | |
| "loss": 0.3232, | |
| "step": 5260 | |
| }, | |
| { | |
| "epoch": 2.150743532287635, | |
| "grad_norm": 3.5590999126434326, | |
| "learning_rate": 2.8713007570543707e-06, | |
| "loss": 0.2824, | |
| "step": 5280 | |
| }, | |
| { | |
| "epoch": 2.1588918313302097, | |
| "grad_norm": 5.644620895385742, | |
| "learning_rate": 2.8437715072264283e-06, | |
| "loss": 0.3407, | |
| "step": 5300 | |
| }, | |
| { | |
| "epoch": 2.1670401303727846, | |
| "grad_norm": 4.308077335357666, | |
| "learning_rate": 2.816242257398486e-06, | |
| "loss": 0.2948, | |
| "step": 5320 | |
| }, | |
| { | |
| "epoch": 2.1751884294153596, | |
| "grad_norm": 4.851817607879639, | |
| "learning_rate": 2.7887130075705435e-06, | |
| "loss": 0.3457, | |
| "step": 5340 | |
| }, | |
| { | |
| "epoch": 2.1833367284579346, | |
| "grad_norm": 4.964145660400391, | |
| "learning_rate": 2.7611837577426016e-06, | |
| "loss": 0.318, | |
| "step": 5360 | |
| }, | |
| { | |
| "epoch": 2.191485027500509, | |
| "grad_norm": 3.544668197631836, | |
| "learning_rate": 2.7336545079146596e-06, | |
| "loss": 0.3017, | |
| "step": 5380 | |
| }, | |
| { | |
| "epoch": 2.199633326543084, | |
| "grad_norm": 5.297823905944824, | |
| "learning_rate": 2.7061252580867177e-06, | |
| "loss": 0.2955, | |
| "step": 5400 | |
| }, | |
| { | |
| "epoch": 2.207781625585659, | |
| "grad_norm": 4.36570930480957, | |
| "learning_rate": 2.6785960082587753e-06, | |
| "loss": 0.3451, | |
| "step": 5420 | |
| }, | |
| { | |
| "epoch": 2.215929924628234, | |
| "grad_norm": 4.8854475021362305, | |
| "learning_rate": 2.651066758430833e-06, | |
| "loss": 0.3115, | |
| "step": 5440 | |
| }, | |
| { | |
| "epoch": 2.2240782236708085, | |
| "grad_norm": 5.848690986633301, | |
| "learning_rate": 2.623537508602891e-06, | |
| "loss": 0.3196, | |
| "step": 5460 | |
| }, | |
| { | |
| "epoch": 2.2322265227133835, | |
| "grad_norm": 6.296275615692139, | |
| "learning_rate": 2.5960082587749486e-06, | |
| "loss": 0.3577, | |
| "step": 5480 | |
| }, | |
| { | |
| "epoch": 2.2403748217559585, | |
| "grad_norm": 6.340885162353516, | |
| "learning_rate": 2.5684790089470063e-06, | |
| "loss": 0.3347, | |
| "step": 5500 | |
| }, | |
| { | |
| "epoch": 2.2485231207985334, | |
| "grad_norm": 5.3483195304870605, | |
| "learning_rate": 2.540949759119064e-06, | |
| "loss": 0.2829, | |
| "step": 5520 | |
| }, | |
| { | |
| "epoch": 2.2566714198411084, | |
| "grad_norm": 3.8228771686553955, | |
| "learning_rate": 2.513420509291122e-06, | |
| "loss": 0.3002, | |
| "step": 5540 | |
| }, | |
| { | |
| "epoch": 2.264819718883683, | |
| "grad_norm": 5.000001907348633, | |
| "learning_rate": 2.48589125946318e-06, | |
| "loss": 0.2816, | |
| "step": 5560 | |
| }, | |
| { | |
| "epoch": 2.272968017926258, | |
| "grad_norm": 4.822083473205566, | |
| "learning_rate": 2.4583620096352376e-06, | |
| "loss": 0.2982, | |
| "step": 5580 | |
| }, | |
| { | |
| "epoch": 2.281116316968833, | |
| "grad_norm": 4.827035427093506, | |
| "learning_rate": 2.4308327598072953e-06, | |
| "loss": 0.3384, | |
| "step": 5600 | |
| }, | |
| { | |
| "epoch": 2.2892646160114074, | |
| "grad_norm": 4.659197807312012, | |
| "learning_rate": 2.4033035099793533e-06, | |
| "loss": 0.2845, | |
| "step": 5620 | |
| }, | |
| { | |
| "epoch": 2.2974129150539824, | |
| "grad_norm": 3.4446418285369873, | |
| "learning_rate": 2.375774260151411e-06, | |
| "loss": 0.3019, | |
| "step": 5640 | |
| }, | |
| { | |
| "epoch": 2.3055612140965573, | |
| "grad_norm": 4.726231575012207, | |
| "learning_rate": 2.348245010323469e-06, | |
| "loss": 0.3568, | |
| "step": 5660 | |
| }, | |
| { | |
| "epoch": 2.3137095131391323, | |
| "grad_norm": 5.017887115478516, | |
| "learning_rate": 2.3207157604955266e-06, | |
| "loss": 0.2857, | |
| "step": 5680 | |
| }, | |
| { | |
| "epoch": 2.3218578121817073, | |
| "grad_norm": 4.18927001953125, | |
| "learning_rate": 2.2931865106675847e-06, | |
| "loss": 0.3285, | |
| "step": 5700 | |
| }, | |
| { | |
| "epoch": 2.330006111224282, | |
| "grad_norm": 5.701695919036865, | |
| "learning_rate": 2.2656572608396423e-06, | |
| "loss": 0.3167, | |
| "step": 5720 | |
| }, | |
| { | |
| "epoch": 2.3381544102668568, | |
| "grad_norm": 3.6883463859558105, | |
| "learning_rate": 2.2381280110117e-06, | |
| "loss": 0.3102, | |
| "step": 5740 | |
| }, | |
| { | |
| "epoch": 2.3463027093094317, | |
| "grad_norm": 5.979183197021484, | |
| "learning_rate": 2.210598761183758e-06, | |
| "loss": 0.3405, | |
| "step": 5760 | |
| }, | |
| { | |
| "epoch": 2.3544510083520063, | |
| "grad_norm": 3.127885580062866, | |
| "learning_rate": 2.1830695113558156e-06, | |
| "loss": 0.3427, | |
| "step": 5780 | |
| }, | |
| { | |
| "epoch": 2.3625993073945812, | |
| "grad_norm": 4.842801094055176, | |
| "learning_rate": 2.1555402615278737e-06, | |
| "loss": 0.2836, | |
| "step": 5800 | |
| }, | |
| { | |
| "epoch": 2.370747606437156, | |
| "grad_norm": 4.311405658721924, | |
| "learning_rate": 2.1280110116999313e-06, | |
| "loss": 0.3005, | |
| "step": 5820 | |
| }, | |
| { | |
| "epoch": 2.378895905479731, | |
| "grad_norm": 4.682214260101318, | |
| "learning_rate": 2.1004817618719894e-06, | |
| "loss": 0.321, | |
| "step": 5840 | |
| }, | |
| { | |
| "epoch": 2.387044204522306, | |
| "grad_norm": 7.238677501678467, | |
| "learning_rate": 2.072952512044047e-06, | |
| "loss": 0.2943, | |
| "step": 5860 | |
| }, | |
| { | |
| "epoch": 2.3951925035648807, | |
| "grad_norm": 2.814465045928955, | |
| "learning_rate": 2.0454232622161046e-06, | |
| "loss": 0.3057, | |
| "step": 5880 | |
| }, | |
| { | |
| "epoch": 2.4033408026074556, | |
| "grad_norm": 5.485500812530518, | |
| "learning_rate": 2.0178940123881627e-06, | |
| "loss": 0.3054, | |
| "step": 5900 | |
| }, | |
| { | |
| "epoch": 2.4114891016500306, | |
| "grad_norm": 5.754581451416016, | |
| "learning_rate": 1.9903647625602203e-06, | |
| "loss": 0.314, | |
| "step": 5920 | |
| }, | |
| { | |
| "epoch": 2.4196374006926056, | |
| "grad_norm": 3.692474603652954, | |
| "learning_rate": 1.9628355127322783e-06, | |
| "loss": 0.3377, | |
| "step": 5940 | |
| }, | |
| { | |
| "epoch": 2.42778569973518, | |
| "grad_norm": 5.103245258331299, | |
| "learning_rate": 1.935306262904336e-06, | |
| "loss": 0.3204, | |
| "step": 5960 | |
| }, | |
| { | |
| "epoch": 2.435933998777755, | |
| "grad_norm": 3.2880680561065674, | |
| "learning_rate": 1.907777013076394e-06, | |
| "loss": 0.2942, | |
| "step": 5980 | |
| }, | |
| { | |
| "epoch": 2.44408229782033, | |
| "grad_norm": 4.239877700805664, | |
| "learning_rate": 1.8802477632484517e-06, | |
| "loss": 0.3616, | |
| "step": 6000 | |
| }, | |
| { | |
| "epoch": 2.44408229782033, | |
| "eval_loss": 0.3231365978717804, | |
| "eval_runtime": 9411.3014, | |
| "eval_samples_per_second": 2.086, | |
| "eval_steps_per_second": 0.261, | |
| "eval_wer": 21.781282824708335, | |
| "step": 6000 | |
| } | |
| ], | |
| "logging_steps": 20, | |
| "max_steps": 7365, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 3, | |
| "save_steps": 2000, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": false | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 5.53968534454272e+19, | |
| "train_batch_size": 16, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |