Baselhany's picture
Training in progress, step 42090, checkpoint
679c0bd verified
{
"best_global_step": 32400,
"best_metric": 0.19510757441791923,
"best_model_checkpoint": "./distil-whisper/checkpoint-32400",
"epoch": 29.996971316586496,
"eval_steps": 400,
"global_step": 42090,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.07126313914127917,
"grad_norm": 681.8389892578125,
"learning_rate": 1.7000000000000003e-05,
"loss": 1297.2127,
"step": 100
},
{
"epoch": 0.14252627828255834,
"grad_norm": 464.8807067871094,
"learning_rate": 3.7e-05,
"loss": 187.2825,
"step": 200
},
{
"epoch": 0.21378941742383753,
"grad_norm": 389.95465087890625,
"learning_rate": 5.6999999999999996e-05,
"loss": 79.5026,
"step": 300
},
{
"epoch": 0.2850525565651167,
"grad_norm": 271.1416931152344,
"learning_rate": 7.7e-05,
"loss": 46.3716,
"step": 400
},
{
"epoch": 0.2850525565651167,
"eval_loss": 0.16972535848617554,
"eval_runtime": 171.9129,
"eval_samples_per_second": 2.908,
"eval_steps_per_second": 0.366,
"eval_wer": 0.6097848511641615,
"step": 400
},
{
"epoch": 0.3563156957063959,
"grad_norm": 200.9510955810547,
"learning_rate": 9.7e-05,
"loss": 32.0457,
"step": 500
},
{
"epoch": 0.42757883484767506,
"grad_norm": 148.45339965820312,
"learning_rate": 9.937176644493719e-05,
"loss": 24.0656,
"step": 600
},
{
"epoch": 0.4988419739889542,
"grad_norm": 204.2472686767578,
"learning_rate": 9.863266814486328e-05,
"loss": 19.3495,
"step": 700
},
{
"epoch": 0.5701051131302334,
"grad_norm": 113.892578125,
"learning_rate": 9.789356984478936e-05,
"loss": 16.3556,
"step": 800
},
{
"epoch": 0.5701051131302334,
"eval_loss": 0.13547594845294952,
"eval_runtime": 162.518,
"eval_samples_per_second": 3.077,
"eval_steps_per_second": 0.388,
"eval_wer": 0.3555850279988211,
"step": 800
},
{
"epoch": 0.6413682522715125,
"grad_norm": 103.74024963378906,
"learning_rate": 9.715447154471545e-05,
"loss": 14.24,
"step": 900
},
{
"epoch": 0.7126313914127917,
"grad_norm": 181.4265594482422,
"learning_rate": 9.641537324464154e-05,
"loss": 12.9121,
"step": 1000
},
{
"epoch": 0.7838945305540709,
"grad_norm": 116.5692138671875,
"learning_rate": 9.567627494456764e-05,
"loss": 11.8352,
"step": 1100
},
{
"epoch": 0.8551576696953501,
"grad_norm": 106.76929473876953,
"learning_rate": 9.493717664449371e-05,
"loss": 11.9327,
"step": 1200
},
{
"epoch": 0.8551576696953501,
"eval_loss": 0.12299185246229172,
"eval_runtime": 159.9552,
"eval_samples_per_second": 3.126,
"eval_steps_per_second": 0.394,
"eval_wer": 0.30002947244326555,
"step": 1200
},
{
"epoch": 0.9264208088366293,
"grad_norm": 148.32879638671875,
"learning_rate": 9.419807834441981e-05,
"loss": 10.2177,
"step": 1300
},
{
"epoch": 0.9976839479779084,
"grad_norm": 166.2821807861328,
"learning_rate": 9.34589800443459e-05,
"loss": 10.1695,
"step": 1400
},
{
"epoch": 1.068412613575628,
"grad_norm": 108.40921020507812,
"learning_rate": 9.2719881744272e-05,
"loss": 7.8838,
"step": 1500
},
{
"epoch": 1.1396757527169072,
"grad_norm": 81.83760833740234,
"learning_rate": 9.198078344419808e-05,
"loss": 8.1222,
"step": 1600
},
{
"epoch": 1.1396757527169072,
"eval_loss": 0.11962544173002243,
"eval_runtime": 154.6688,
"eval_samples_per_second": 3.233,
"eval_steps_per_second": 0.407,
"eval_wer": 0.2543471853816681,
"step": 1600
},
{
"epoch": 1.2109388918581863,
"grad_norm": 66.1691665649414,
"learning_rate": 9.124168514412418e-05,
"loss": 7.4228,
"step": 1700
},
{
"epoch": 1.2822020309994655,
"grad_norm": 98.77230072021484,
"learning_rate": 9.050258684405026e-05,
"loss": 7.4755,
"step": 1800
},
{
"epoch": 1.3534651701407447,
"grad_norm": 94.88159942626953,
"learning_rate": 8.976348854397635e-05,
"loss": 7.2464,
"step": 1900
},
{
"epoch": 1.424728309282024,
"grad_norm": 78.41780090332031,
"learning_rate": 8.902439024390244e-05,
"loss": 6.2775,
"step": 2000
},
{
"epoch": 1.424728309282024,
"eval_loss": 0.11653754115104675,
"eval_runtime": 157.609,
"eval_samples_per_second": 3.172,
"eval_steps_per_second": 0.4,
"eval_wer": 0.26186265841438255,
"step": 2000
},
{
"epoch": 1.495991448423303,
"grad_norm": 68.5005874633789,
"learning_rate": 8.828529194382853e-05,
"loss": 6.5,
"step": 2100
},
{
"epoch": 1.5672545875645822,
"grad_norm": 123.60826873779297,
"learning_rate": 8.754619364375463e-05,
"loss": 6.7454,
"step": 2200
},
{
"epoch": 1.6385177267058615,
"grad_norm": 60.07472229003906,
"learning_rate": 8.680709534368072e-05,
"loss": 6.4664,
"step": 2300
},
{
"epoch": 1.7097808658471405,
"grad_norm": 91.09046173095703,
"learning_rate": 8.60679970436068e-05,
"loss": 5.6861,
"step": 2400
},
{
"epoch": 1.7097808658471405,
"eval_loss": 0.11434157937765121,
"eval_runtime": 159.2078,
"eval_samples_per_second": 3.141,
"eval_steps_per_second": 0.396,
"eval_wer": 0.23902151488358384,
"step": 2400
},
{
"epoch": 1.7810440049884197,
"grad_norm": 53.910728454589844,
"learning_rate": 8.532889874353289e-05,
"loss": 5.7159,
"step": 2500
},
{
"epoch": 1.852307144129699,
"grad_norm": 47.61832046508789,
"learning_rate": 8.458980044345899e-05,
"loss": 5.5957,
"step": 2600
},
{
"epoch": 1.923570283270978,
"grad_norm": 71.0946044921875,
"learning_rate": 8.385070214338508e-05,
"loss": 5.6279,
"step": 2700
},
{
"epoch": 1.9948334224122573,
"grad_norm": 81.29457092285156,
"learning_rate": 8.311160384331117e-05,
"loss": 5.238,
"step": 2800
},
{
"epoch": 1.9948334224122573,
"eval_loss": 0.11149635910987854,
"eval_runtime": 159.9177,
"eval_samples_per_second": 3.127,
"eval_steps_per_second": 0.394,
"eval_wer": 0.23460064839375183,
"step": 2800
},
{
"epoch": 2.065562088009977,
"grad_norm": 42.97408676147461,
"learning_rate": 8.237250554323725e-05,
"loss": 4.1653,
"step": 2900
},
{
"epoch": 2.136825227151256,
"grad_norm": 86.63339233398438,
"learning_rate": 8.163340724316334e-05,
"loss": 4.1343,
"step": 3000
},
{
"epoch": 2.208088366292535,
"grad_norm": 87.8547134399414,
"learning_rate": 8.089430894308944e-05,
"loss": 4.69,
"step": 3100
},
{
"epoch": 2.2793515054338145,
"grad_norm": 41.078346252441406,
"learning_rate": 8.015521064301552e-05,
"loss": 4.5097,
"step": 3200
},
{
"epoch": 2.2793515054338145,
"eval_loss": 0.11067745089530945,
"eval_runtime": 156.5697,
"eval_samples_per_second": 3.193,
"eval_steps_per_second": 0.402,
"eval_wer": 0.2256115531977601,
"step": 3200
},
{
"epoch": 2.3506146445750935,
"grad_norm": 56.07571792602539,
"learning_rate": 7.941611234294162e-05,
"loss": 4.2569,
"step": 3300
},
{
"epoch": 2.4218777837163725,
"grad_norm": 80.53390502929688,
"learning_rate": 7.86770140428677e-05,
"loss": 4.3998,
"step": 3400
},
{
"epoch": 2.493140922857652,
"grad_norm": 70.70401763916016,
"learning_rate": 7.79379157427938e-05,
"loss": 4.1977,
"step": 3500
},
{
"epoch": 2.564404061998931,
"grad_norm": 62.703758239746094,
"learning_rate": 7.719881744271988e-05,
"loss": 3.9677,
"step": 3600
},
{
"epoch": 2.564404061998931,
"eval_loss": 0.10945500433444977,
"eval_runtime": 159.3631,
"eval_samples_per_second": 3.137,
"eval_steps_per_second": 0.395,
"eval_wer": 0.22620100206307103,
"step": 3600
},
{
"epoch": 2.63566720114021,
"grad_norm": 40.595977783203125,
"learning_rate": 7.645971914264598e-05,
"loss": 3.9478,
"step": 3700
},
{
"epoch": 2.7069303402814895,
"grad_norm": 43.14921188354492,
"learning_rate": 7.572062084257207e-05,
"loss": 3.9157,
"step": 3800
},
{
"epoch": 2.7781934794227685,
"grad_norm": 53.32358169555664,
"learning_rate": 7.498152254249816e-05,
"loss": 3.877,
"step": 3900
},
{
"epoch": 2.849456618564048,
"grad_norm": 53.401611328125,
"learning_rate": 7.424242424242424e-05,
"loss": 3.8998,
"step": 4000
},
{
"epoch": 2.849456618564048,
"eval_loss": 0.1084858775138855,
"eval_runtime": 161.4334,
"eval_samples_per_second": 3.097,
"eval_steps_per_second": 0.39,
"eval_wer": 0.2300324196875921,
"step": 4000
},
{
"epoch": 2.920719757705327,
"grad_norm": 52.240577697753906,
"learning_rate": 7.350332594235033e-05,
"loss": 3.7483,
"step": 4100
},
{
"epoch": 2.991982896846606,
"grad_norm": 89.89251708984375,
"learning_rate": 7.276422764227643e-05,
"loss": 3.8055,
"step": 4200
},
{
"epoch": 3.0627115624443255,
"grad_norm": 44.852535247802734,
"learning_rate": 7.202512934220252e-05,
"loss": 3.1335,
"step": 4300
},
{
"epoch": 3.133974701585605,
"grad_norm": 51.19050598144531,
"learning_rate": 7.12860310421286e-05,
"loss": 3.3351,
"step": 4400
},
{
"epoch": 3.133974701585605,
"eval_loss": 0.10667955130338669,
"eval_runtime": 156.6885,
"eval_samples_per_second": 3.191,
"eval_steps_per_second": 0.402,
"eval_wer": 0.21396993810786913,
"step": 4400
},
{
"epoch": 3.205237840726884,
"grad_norm": 85.72518157958984,
"learning_rate": 7.05469327420547e-05,
"loss": 3.2473,
"step": 4500
},
{
"epoch": 3.276500979868163,
"grad_norm": 53.99726867675781,
"learning_rate": 6.98078344419808e-05,
"loss": 3.0099,
"step": 4600
},
{
"epoch": 3.3477641190094425,
"grad_norm": 61.77521514892578,
"learning_rate": 6.906873614190688e-05,
"loss": 3.1004,
"step": 4700
},
{
"epoch": 3.4190272581507215,
"grad_norm": 49.96204376220703,
"learning_rate": 6.832963784183297e-05,
"loss": 3.1317,
"step": 4800
},
{
"epoch": 3.4190272581507215,
"eval_loss": 0.1066509559750557,
"eval_runtime": 157.6866,
"eval_samples_per_second": 3.171,
"eval_steps_per_second": 0.4,
"eval_wer": 0.21986442676097848,
"step": 4800
},
{
"epoch": 3.4902903972920005,
"grad_norm": 60.96116638183594,
"learning_rate": 6.759053954175906e-05,
"loss": 3.1707,
"step": 4900
},
{
"epoch": 3.56155353643328,
"grad_norm": 62.14186477661133,
"learning_rate": 6.685144124168514e-05,
"loss": 3.1594,
"step": 5000
},
{
"epoch": 3.632816675574559,
"grad_norm": 35.275325775146484,
"learning_rate": 6.611234294161123e-05,
"loss": 3.1857,
"step": 5100
},
{
"epoch": 3.704079814715838,
"grad_norm": 54.71980285644531,
"learning_rate": 6.537324464153732e-05,
"loss": 2.9814,
"step": 5200
},
{
"epoch": 3.704079814715838,
"eval_loss": 0.10459103435277939,
"eval_runtime": 155.7856,
"eval_samples_per_second": 3.21,
"eval_steps_per_second": 0.404,
"eval_wer": 0.2119068670792809,
"step": 5200
},
{
"epoch": 3.7753429538571175,
"grad_norm": 44.45451354980469,
"learning_rate": 6.463414634146342e-05,
"loss": 2.9051,
"step": 5300
},
{
"epoch": 3.8466060929983965,
"grad_norm": 32.420108795166016,
"learning_rate": 6.389504804138951e-05,
"loss": 3.2812,
"step": 5400
},
{
"epoch": 3.917869232139676,
"grad_norm": 56.81290817260742,
"learning_rate": 6.31559497413156e-05,
"loss": 3.1616,
"step": 5500
},
{
"epoch": 3.989132371280955,
"grad_norm": 47.525577545166016,
"learning_rate": 6.241685144124168e-05,
"loss": 3.167,
"step": 5600
},
{
"epoch": 3.989132371280955,
"eval_loss": 0.103938028216362,
"eval_runtime": 153.9307,
"eval_samples_per_second": 3.248,
"eval_steps_per_second": 0.409,
"eval_wer": 0.21043324491600354,
"step": 5600
},
{
"epoch": 4.059861036878675,
"grad_norm": 41.089988708496094,
"learning_rate": 6.167775314116778e-05,
"loss": 2.7796,
"step": 5700
},
{
"epoch": 4.131124176019954,
"grad_norm": 28.174148559570312,
"learning_rate": 6.0938654841093865e-05,
"loss": 2.4943,
"step": 5800
},
{
"epoch": 4.202387315161233,
"grad_norm": 73.63423156738281,
"learning_rate": 6.0199556541019966e-05,
"loss": 2.5883,
"step": 5900
},
{
"epoch": 4.273650454302512,
"grad_norm": 35.365089416503906,
"learning_rate": 5.9460458240946046e-05,
"loss": 2.498,
"step": 6000
},
{
"epoch": 4.273650454302512,
"eval_loss": 0.1066313236951828,
"eval_runtime": 154.5122,
"eval_samples_per_second": 3.236,
"eval_steps_per_second": 0.408,
"eval_wer": 0.2176539935160625,
"step": 6000
},
{
"epoch": 4.344913593443791,
"grad_norm": 43.693443298339844,
"learning_rate": 5.8721359940872134e-05,
"loss": 2.5415,
"step": 6100
},
{
"epoch": 4.41617673258507,
"grad_norm": 55.82042694091797,
"learning_rate": 5.798226164079823e-05,
"loss": 2.5464,
"step": 6200
},
{
"epoch": 4.48743987172635,
"grad_norm": 37.502227783203125,
"learning_rate": 5.7243163340724315e-05,
"loss": 2.534,
"step": 6300
},
{
"epoch": 4.558703010867629,
"grad_norm": 36.125301361083984,
"learning_rate": 5.650406504065041e-05,
"loss": 2.8372,
"step": 6400
},
{
"epoch": 4.558703010867629,
"eval_loss": 0.10223670303821564,
"eval_runtime": 155.3604,
"eval_samples_per_second": 3.218,
"eval_steps_per_second": 0.406,
"eval_wer": 0.20984379605069262,
"step": 6400
},
{
"epoch": 4.629966150008908,
"grad_norm": 48.96213912963867,
"learning_rate": 5.57649667405765e-05,
"loss": 2.6121,
"step": 6500
},
{
"epoch": 4.701229289150187,
"grad_norm": 59.73322677612305,
"learning_rate": 5.502586844050259e-05,
"loss": 2.7291,
"step": 6600
},
{
"epoch": 4.772492428291466,
"grad_norm": 120.08707427978516,
"learning_rate": 5.428677014042868e-05,
"loss": 2.7742,
"step": 6700
},
{
"epoch": 4.843755567432745,
"grad_norm": 46.421714782714844,
"learning_rate": 5.354767184035477e-05,
"loss": 2.5573,
"step": 6800
},
{
"epoch": 4.843755567432745,
"eval_loss": 0.10277628153562546,
"eval_runtime": 157.4825,
"eval_samples_per_second": 3.175,
"eval_steps_per_second": 0.4,
"eval_wer": 0.21809608016504567,
"step": 6800
},
{
"epoch": 4.915018706574025,
"grad_norm": 50.61531448364258,
"learning_rate": 5.280857354028086e-05,
"loss": 2.3244,
"step": 6900
},
{
"epoch": 4.986281845715304,
"grad_norm": 39.10790252685547,
"learning_rate": 5.2069475240206955e-05,
"loss": 2.5899,
"step": 7000
},
{
"epoch": 5.057010511313023,
"grad_norm": 36.1613883972168,
"learning_rate": 5.133037694013304e-05,
"loss": 2.2914,
"step": 7100
},
{
"epoch": 5.128273650454302,
"grad_norm": 41.908348083496094,
"learning_rate": 5.059127864005912e-05,
"loss": 2.3309,
"step": 7200
},
{
"epoch": 5.128273650454302,
"eval_loss": 0.10057391971349716,
"eval_runtime": 155.7627,
"eval_samples_per_second": 3.21,
"eval_steps_per_second": 0.404,
"eval_wer": 0.20910698496905394,
"step": 7200
},
{
"epoch": 5.199536789595582,
"grad_norm": 73.74039459228516,
"learning_rate": 4.9852180339985224e-05,
"loss": 2.3553,
"step": 7300
},
{
"epoch": 5.270799928736861,
"grad_norm": 30.432615280151367,
"learning_rate": 4.911308203991131e-05,
"loss": 2.2731,
"step": 7400
},
{
"epoch": 5.34206306787814,
"grad_norm": 43.42770767211914,
"learning_rate": 4.8373983739837406e-05,
"loss": 2.1843,
"step": 7500
},
{
"epoch": 5.413326207019419,
"grad_norm": 36.634033203125,
"learning_rate": 4.7634885439763486e-05,
"loss": 2.2589,
"step": 7600
},
{
"epoch": 5.413326207019419,
"eval_loss": 0.10152015835046768,
"eval_runtime": 157.1024,
"eval_samples_per_second": 3.183,
"eval_steps_per_second": 0.401,
"eval_wer": 0.20999115826702033,
"step": 7600
},
{
"epoch": 5.484589346160698,
"grad_norm": 44.67591094970703,
"learning_rate": 4.689578713968958e-05,
"loss": 2.3519,
"step": 7700
},
{
"epoch": 5.555852485301978,
"grad_norm": 34.04768753051758,
"learning_rate": 4.615668883961567e-05,
"loss": 1.9867,
"step": 7800
},
{
"epoch": 5.627115624443257,
"grad_norm": 29.320768356323242,
"learning_rate": 4.541759053954176e-05,
"loss": 2.1691,
"step": 7900
},
{
"epoch": 5.698378763584536,
"grad_norm": 35.776451110839844,
"learning_rate": 4.467849223946785e-05,
"loss": 2.1409,
"step": 8000
},
{
"epoch": 5.698378763584536,
"eval_loss": 0.10244826227426529,
"eval_runtime": 154.1587,
"eval_samples_per_second": 3.243,
"eval_steps_per_second": 0.409,
"eval_wer": 0.20645446507515472,
"step": 8000
},
{
"epoch": 5.769641902725815,
"grad_norm": 48.27608108520508,
"learning_rate": 4.3939393939393944e-05,
"loss": 2.1876,
"step": 8100
},
{
"epoch": 5.840905041867094,
"grad_norm": 35.93488311767578,
"learning_rate": 4.320029563932003e-05,
"loss": 2.2169,
"step": 8200
},
{
"epoch": 5.912168181008373,
"grad_norm": 23.94822120666504,
"learning_rate": 4.2461197339246126e-05,
"loss": 2.1963,
"step": 8300
},
{
"epoch": 5.983431320149653,
"grad_norm": 28.82554817199707,
"learning_rate": 4.172209903917221e-05,
"loss": 2.1048,
"step": 8400
},
{
"epoch": 5.983431320149653,
"eval_loss": 0.09920582175254822,
"eval_runtime": 158.0225,
"eval_samples_per_second": 3.164,
"eval_steps_per_second": 0.399,
"eval_wer": 0.2138225758915414,
"step": 8400
},
{
"epoch": 6.054159985747372,
"grad_norm": 30.941600799560547,
"learning_rate": 4.09830007390983e-05,
"loss": 2.1247,
"step": 8500
},
{
"epoch": 6.125423124888651,
"grad_norm": 30.994831085205078,
"learning_rate": 4.0243902439024395e-05,
"loss": 1.9593,
"step": 8600
},
{
"epoch": 6.196686264029931,
"grad_norm": 34.325279235839844,
"learning_rate": 3.950480413895048e-05,
"loss": 1.9629,
"step": 8700
},
{
"epoch": 6.26794940317121,
"grad_norm": 24.892839431762695,
"learning_rate": 3.876570583887657e-05,
"loss": 1.8826,
"step": 8800
},
{
"epoch": 6.26794940317121,
"eval_loss": 0.09866400063037872,
"eval_runtime": 157.1762,
"eval_samples_per_second": 3.181,
"eval_steps_per_second": 0.401,
"eval_wer": 0.2116121426466254,
"step": 8800
},
{
"epoch": 6.339212542312489,
"grad_norm": 30.01650047302246,
"learning_rate": 3.8026607538802664e-05,
"loss": 1.9085,
"step": 8900
},
{
"epoch": 6.410475681453768,
"grad_norm": 45.52370071411133,
"learning_rate": 3.728750923872875e-05,
"loss": 1.9014,
"step": 9000
},
{
"epoch": 6.481738820595047,
"grad_norm": 26.402570724487305,
"learning_rate": 3.6548410938654846e-05,
"loss": 1.824,
"step": 9100
},
{
"epoch": 6.553001959736326,
"grad_norm": 43.61695861816406,
"learning_rate": 3.580931263858093e-05,
"loss": 1.8778,
"step": 9200
},
{
"epoch": 6.553001959736326,
"eval_loss": 0.09877074509859085,
"eval_runtime": 155.4125,
"eval_samples_per_second": 3.217,
"eval_steps_per_second": 0.405,
"eval_wer": 0.20733863837312114,
"step": 9200
},
{
"epoch": 6.624265098877606,
"grad_norm": 40.971221923828125,
"learning_rate": 3.507021433850702e-05,
"loss": 1.9757,
"step": 9300
},
{
"epoch": 6.695528238018885,
"grad_norm": 39.76367950439453,
"learning_rate": 3.4331116038433115e-05,
"loss": 1.9012,
"step": 9400
},
{
"epoch": 6.766791377160164,
"grad_norm": 39.237640380859375,
"learning_rate": 3.35920177383592e-05,
"loss": 1.9141,
"step": 9500
},
{
"epoch": 6.838054516301443,
"grad_norm": 31.298519134521484,
"learning_rate": 3.2852919438285297e-05,
"loss": 2.0199,
"step": 9600
},
{
"epoch": 6.838054516301443,
"eval_loss": 0.09809073060750961,
"eval_runtime": 156.8243,
"eval_samples_per_second": 3.188,
"eval_steps_per_second": 0.402,
"eval_wer": 0.2045387562628942,
"step": 9600
},
{
"epoch": 6.909317655442722,
"grad_norm": 59.62895202636719,
"learning_rate": 3.2113821138211384e-05,
"loss": 1.9717,
"step": 9700
},
{
"epoch": 6.980580794584001,
"grad_norm": 26.735105514526367,
"learning_rate": 3.137472283813747e-05,
"loss": 1.9839,
"step": 9800
},
{
"epoch": 7.051309460181721,
"grad_norm": 39.49419021606445,
"learning_rate": 3.063562453806356e-05,
"loss": 1.7422,
"step": 9900
},
{
"epoch": 7.122572599323,
"grad_norm": 36.06403732299805,
"learning_rate": 2.9896526237989653e-05,
"loss": 1.7238,
"step": 10000
},
{
"epoch": 7.122572599323,
"eval_loss": 0.09965246915817261,
"eval_runtime": 156.0402,
"eval_samples_per_second": 3.204,
"eval_steps_per_second": 0.404,
"eval_wer": 0.20218096080165046,
"step": 10000
},
{
"epoch": 7.193835738464279,
"grad_norm": 34.01579284667969,
"learning_rate": 2.9157427937915744e-05,
"loss": 1.8574,
"step": 10100
},
{
"epoch": 7.265098877605558,
"grad_norm": 23.857589721679688,
"learning_rate": 2.8418329637841835e-05,
"loss": 1.7262,
"step": 10200
},
{
"epoch": 7.336362016746838,
"grad_norm": 22.911731719970703,
"learning_rate": 2.7679231337767926e-05,
"loss": 1.803,
"step": 10300
},
{
"epoch": 7.407625155888117,
"grad_norm": 40.41181564331055,
"learning_rate": 2.6940133037694017e-05,
"loss": 1.8087,
"step": 10400
},
{
"epoch": 7.407625155888117,
"eval_loss": 0.09829334914684296,
"eval_runtime": 156.8822,
"eval_samples_per_second": 3.187,
"eval_steps_per_second": 0.402,
"eval_wer": 0.2036545829649278,
"step": 10400
},
{
"epoch": 7.483876714769286,
"grad_norm": 34.382659912109375,
"learning_rate": 2.6201034737620107e-05,
"loss": 1.8111,
"step": 10500
},
{
"epoch": 7.555139853910565,
"grad_norm": 33.14358901977539,
"learning_rate": 2.5461936437546198e-05,
"loss": 1.7896,
"step": 10600
},
{
"epoch": 7.626402993051844,
"grad_norm": 39.97333526611328,
"learning_rate": 2.4722838137472286e-05,
"loss": 1.7157,
"step": 10700
},
{
"epoch": 7.697666132193123,
"grad_norm": 33.9200553894043,
"learning_rate": 2.3983739837398377e-05,
"loss": 1.7075,
"step": 10800
},
{
"epoch": 7.697666132193123,
"eval_loss": 0.09847129136323929,
"eval_runtime": 148.952,
"eval_samples_per_second": 3.357,
"eval_steps_per_second": 0.423,
"eval_wer": 0.2058650162098438,
"step": 10800
},
{
"epoch": 7.768929271334402,
"grad_norm": 22.087848663330078,
"learning_rate": 2.3244641537324467e-05,
"loss": 1.6991,
"step": 10900
},
{
"epoch": 7.840192410475682,
"grad_norm": 42.86452102661133,
"learning_rate": 2.2505543237250555e-05,
"loss": 1.7294,
"step": 11000
},
{
"epoch": 7.911455549616961,
"grad_norm": 29.147266387939453,
"learning_rate": 2.1766444937176646e-05,
"loss": 1.7738,
"step": 11100
},
{
"epoch": 7.98271868875824,
"grad_norm": 38.3438835144043,
"learning_rate": 2.1027346637102737e-05,
"loss": 1.7072,
"step": 11200
},
{
"epoch": 7.98271868875824,
"eval_loss": 0.09768818318843842,
"eval_runtime": 146.2719,
"eval_samples_per_second": 3.418,
"eval_steps_per_second": 0.431,
"eval_wer": 0.20615974064249926,
"step": 11200
},
{
"epoch": 8.054159985747372,
"grad_norm": 25.453853607177734,
"learning_rate": 2.0288248337028824e-05,
"loss": 1.5749,
"step": 11300
},
{
"epoch": 8.125423124888652,
"grad_norm": 21.901262283325195,
"learning_rate": 1.9549150036954915e-05,
"loss": 1.588,
"step": 11400
},
{
"epoch": 8.19668626402993,
"grad_norm": 47.3566780090332,
"learning_rate": 1.8810051736881006e-05,
"loss": 1.6455,
"step": 11500
},
{
"epoch": 8.26794940317121,
"grad_norm": 31.993080139160156,
"learning_rate": 1.8070953436807093e-05,
"loss": 1.5864,
"step": 11600
},
{
"epoch": 8.26794940317121,
"eval_loss": 0.09770301729440689,
"eval_runtime": 145.9667,
"eval_samples_per_second": 3.425,
"eval_steps_per_second": 0.432,
"eval_wer": 0.20660182729148246,
"step": 11600
},
{
"epoch": 8.339212542312488,
"grad_norm": 22.330097198486328,
"learning_rate": 1.7331855136733184e-05,
"loss": 1.6879,
"step": 11700
},
{
"epoch": 8.410475681453768,
"grad_norm": 25.55733299255371,
"learning_rate": 1.6592756836659275e-05,
"loss": 1.525,
"step": 11800
},
{
"epoch": 8.481738820595048,
"grad_norm": 25.518842697143555,
"learning_rate": 1.5853658536585366e-05,
"loss": 1.6049,
"step": 11900
},
{
"epoch": 8.553001959736326,
"grad_norm": 23.22325897216797,
"learning_rate": 1.5114560236511458e-05,
"loss": 1.6869,
"step": 12000
},
{
"epoch": 8.553001959736326,
"eval_loss": 0.0971849113702774,
"eval_runtime": 146.1299,
"eval_samples_per_second": 3.422,
"eval_steps_per_second": 0.431,
"eval_wer": 0.2080754494547598,
"step": 12000
},
{
"epoch": 8.624265098877606,
"grad_norm": 26.13189125061035,
"learning_rate": 1.4375461936437547e-05,
"loss": 1.6066,
"step": 12100
},
{
"epoch": 8.695528238018884,
"grad_norm": 18.172691345214844,
"learning_rate": 1.3636363636363637e-05,
"loss": 1.4317,
"step": 12200
},
{
"epoch": 8.766791377160164,
"grad_norm": 28.6180419921875,
"learning_rate": 1.2897265336289727e-05,
"loss": 1.6208,
"step": 12300
},
{
"epoch": 8.838054516301444,
"grad_norm": 24.86948013305664,
"learning_rate": 1.2158167036215817e-05,
"loss": 1.7383,
"step": 12400
},
{
"epoch": 8.838054516301444,
"eval_loss": 0.09760728478431702,
"eval_runtime": 147.3722,
"eval_samples_per_second": 3.393,
"eval_steps_per_second": 0.427,
"eval_wer": 0.20409666961391099,
"step": 12400
},
{
"epoch": 8.909317655442722,
"grad_norm": 24.008527755737305,
"learning_rate": 1.1419068736141907e-05,
"loss": 1.5721,
"step": 12500
},
{
"epoch": 8.980580794584002,
"grad_norm": 19.964536666870117,
"learning_rate": 1.0679970436067997e-05,
"loss": 1.5477,
"step": 12600
},
{
"epoch": 9.051309460181722,
"grad_norm": 26.542261123657227,
"learning_rate": 9.940872135994087e-06,
"loss": 1.4216,
"step": 12700
},
{
"epoch": 9.122572599323,
"grad_norm": 16.520477294921875,
"learning_rate": 9.201773835920177e-06,
"loss": 1.4336,
"step": 12800
},
{
"epoch": 9.122572599323,
"eval_loss": 0.09696762263774872,
"eval_runtime": 145.4672,
"eval_samples_per_second": 3.437,
"eval_steps_per_second": 0.433,
"eval_wer": 0.2045387562628942,
"step": 12800
},
{
"epoch": 9.19383573846428,
"grad_norm": 21.41847801208496,
"learning_rate": 8.462675535846267e-06,
"loss": 1.4627,
"step": 12900
},
{
"epoch": 9.265098877605558,
"grad_norm": 28.16568946838379,
"learning_rate": 7.723577235772358e-06,
"loss": 1.4259,
"step": 13000
},
{
"epoch": 9.336362016746838,
"grad_norm": 19.043367385864258,
"learning_rate": 6.984478935698447e-06,
"loss": 1.5803,
"step": 13100
},
{
"epoch": 9.407625155888116,
"grad_norm": 18.10884666442871,
"learning_rate": 6.245380635624538e-06,
"loss": 1.5429,
"step": 13200
},
{
"epoch": 9.407625155888116,
"eval_loss": 0.09686141461133957,
"eval_runtime": 144.8892,
"eval_samples_per_second": 3.451,
"eval_steps_per_second": 0.435,
"eval_wer": 0.20100206307102858,
"step": 13200
},
{
"epoch": 9.478888295029396,
"grad_norm": 19.601333618164062,
"learning_rate": 5.506282335550628e-06,
"loss": 1.6243,
"step": 13300
},
{
"epoch": 9.550151434170676,
"grad_norm": 26.340774536132812,
"learning_rate": 4.767184035476718e-06,
"loss": 1.4496,
"step": 13400
},
{
"epoch": 9.621414573311954,
"grad_norm": 19.866046905517578,
"learning_rate": 4.028085735402809e-06,
"loss": 1.5778,
"step": 13500
},
{
"epoch": 9.692677712453234,
"grad_norm": 16.61330223083496,
"learning_rate": 3.288987435328899e-06,
"loss": 1.5726,
"step": 13600
},
{
"epoch": 9.692677712453234,
"eval_loss": 0.09693081676959991,
"eval_runtime": 145.8875,
"eval_samples_per_second": 3.427,
"eval_steps_per_second": 0.432,
"eval_wer": 0.20837017388741527,
"step": 13600
},
{
"epoch": 9.763940851594512,
"grad_norm": 26.37179183959961,
"learning_rate": 2.549889135254989e-06,
"loss": 1.546,
"step": 13700
},
{
"epoch": 9.835203990735792,
"grad_norm": 17.592809677124023,
"learning_rate": 1.8107908351810791e-06,
"loss": 1.4815,
"step": 13800
},
{
"epoch": 9.906467129877072,
"grad_norm": 14.37030029296875,
"learning_rate": 1.0716925351071693e-06,
"loss": 1.414,
"step": 13900
},
{
"epoch": 9.97773026901835,
"grad_norm": 21.16905975341797,
"learning_rate": 3.3259423503325944e-07,
"loss": 1.4709,
"step": 14000
},
{
"epoch": 9.97773026901835,
"eval_loss": 0.097068190574646,
"eval_runtime": 145.6677,
"eval_samples_per_second": 3.432,
"eval_steps_per_second": 0.432,
"eval_wer": 0.20439139404656645,
"step": 14000
},
{
"epoch": 10.049884197398896,
"grad_norm": 27.002792358398438,
"learning_rate": 4.1822620016273397e-05,
"loss": 1.4666,
"step": 14100
},
{
"epoch": 10.121147336540174,
"grad_norm": 27.5255069732666,
"learning_rate": 4.139437283199863e-05,
"loss": 1.58,
"step": 14200
},
{
"epoch": 10.192410475681454,
"grad_norm": 29.90217399597168,
"learning_rate": 4.096612564772387e-05,
"loss": 1.7165,
"step": 14300
},
{
"epoch": 10.263673614822732,
"grad_norm": 35.381736755371094,
"learning_rate": 4.05378784634491e-05,
"loss": 1.5442,
"step": 14400
},
{
"epoch": 10.263673614822732,
"eval_loss": 0.09780226647853851,
"eval_runtime": 154.9465,
"eval_samples_per_second": 3.227,
"eval_steps_per_second": 0.407,
"eval_wer": 0.20881226053639848,
"step": 14400
},
{
"epoch": 10.334936753964012,
"grad_norm": 24.984546661376953,
"learning_rate": 4.010963127917434e-05,
"loss": 1.6025,
"step": 14500
},
{
"epoch": 10.406199893105292,
"grad_norm": 29.315156936645508,
"learning_rate": 3.968138409489958e-05,
"loss": 1.6566,
"step": 14600
},
{
"epoch": 10.47746303224657,
"grad_norm": 26.027830123901367,
"learning_rate": 3.925313691062482e-05,
"loss": 1.5891,
"step": 14700
},
{
"epoch": 10.54872617138785,
"grad_norm": 23.834056854248047,
"learning_rate": 3.882488972635005e-05,
"loss": 1.5764,
"step": 14800
},
{
"epoch": 10.54872617138785,
"eval_loss": 0.09848986566066742,
"eval_runtime": 151.1928,
"eval_samples_per_second": 3.307,
"eval_steps_per_second": 0.417,
"eval_wer": 0.215148835838491,
"step": 14800
},
{
"epoch": 10.619989310529128,
"grad_norm": 30.214248657226562,
"learning_rate": 3.839664254207529e-05,
"loss": 1.6469,
"step": 14900
},
{
"epoch": 10.691252449670408,
"grad_norm": 22.957365036010742,
"learning_rate": 3.796839535780052e-05,
"loss": 1.609,
"step": 15000
},
{
"epoch": 10.762515588811688,
"grad_norm": 24.754613876342773,
"learning_rate": 3.7540148173525766e-05,
"loss": 1.6223,
"step": 15100
},
{
"epoch": 10.833778727952966,
"grad_norm": 29.267169952392578,
"learning_rate": 3.7111900989251e-05,
"loss": 1.6821,
"step": 15200
},
{
"epoch": 10.833778727952966,
"eval_loss": 0.09700831025838852,
"eval_runtime": 149.5595,
"eval_samples_per_second": 3.343,
"eval_steps_per_second": 0.421,
"eval_wer": 0.20660182729148246,
"step": 15200
},
{
"epoch": 10.905041867094246,
"grad_norm": 23.508710861206055,
"learning_rate": 3.668365380497623e-05,
"loss": 1.6449,
"step": 15300
},
{
"epoch": 10.976305006235524,
"grad_norm": 19.81301498413086,
"learning_rate": 3.625540662070147e-05,
"loss": 1.6106,
"step": 15400
},
{
"epoch": 11.047033671833244,
"grad_norm": 27.057025909423828,
"learning_rate": 3.58271594364267e-05,
"loss": 1.6911,
"step": 15500
},
{
"epoch": 11.118296810974524,
"grad_norm": 24.53127098083496,
"learning_rate": 3.539891225215195e-05,
"loss": 1.6529,
"step": 15600
},
{
"epoch": 11.118296810974524,
"eval_loss": 0.09739436209201813,
"eval_runtime": 150.7387,
"eval_samples_per_second": 3.317,
"eval_steps_per_second": 0.418,
"eval_wer": 0.20822281167108753,
"step": 15600
},
{
"epoch": 11.189559950115802,
"grad_norm": 28.257545471191406,
"learning_rate": 3.497066506787718e-05,
"loss": 1.5513,
"step": 15700
},
{
"epoch": 11.260823089257082,
"grad_norm": 18.52756690979004,
"learning_rate": 3.454241788360242e-05,
"loss": 1.4921,
"step": 15800
},
{
"epoch": 11.33208622839836,
"grad_norm": 50.12236022949219,
"learning_rate": 3.411417069932765e-05,
"loss": 1.5063,
"step": 15900
},
{
"epoch": 11.40334936753964,
"grad_norm": 44.198509216308594,
"learning_rate": 3.368592351505289e-05,
"loss": 1.5455,
"step": 16000
},
{
"epoch": 11.40334936753964,
"eval_loss": 0.09711702167987823,
"eval_runtime": 147.0043,
"eval_samples_per_second": 3.401,
"eval_steps_per_second": 0.429,
"eval_wer": 0.20571765399351608,
"step": 16000
},
{
"epoch": 11.47461250668092,
"grad_norm": 29.74050521850586,
"learning_rate": 3.325767633077813e-05,
"loss": 1.6061,
"step": 16100
},
{
"epoch": 11.545875645822198,
"grad_norm": 28.548961639404297,
"learning_rate": 3.282942914650337e-05,
"loss": 1.5224,
"step": 16200
},
{
"epoch": 11.617138784963478,
"grad_norm": 20.90842628479004,
"learning_rate": 3.24011819622286e-05,
"loss": 1.491,
"step": 16300
},
{
"epoch": 11.688401924104756,
"grad_norm": 32.348304748535156,
"learning_rate": 3.197293477795384e-05,
"loss": 1.4845,
"step": 16400
},
{
"epoch": 11.688401924104756,
"eval_loss": 0.09728699922561646,
"eval_runtime": 150.28,
"eval_samples_per_second": 3.327,
"eval_steps_per_second": 0.419,
"eval_wer": 0.21396993810786913,
"step": 16400
},
{
"epoch": 11.759665063246036,
"grad_norm": 31.023475646972656,
"learning_rate": 3.154468759367907e-05,
"loss": 1.5757,
"step": 16500
},
{
"epoch": 11.830928202387316,
"grad_norm": 28.920454025268555,
"learning_rate": 3.111644040940431e-05,
"loss": 1.5098,
"step": 16600
},
{
"epoch": 11.902191341528594,
"grad_norm": 28.04616928100586,
"learning_rate": 3.068819322512955e-05,
"loss": 1.3697,
"step": 16700
},
{
"epoch": 11.973454480669874,
"grad_norm": 25.157943725585938,
"learning_rate": 3.025994604085478e-05,
"loss": 1.4953,
"step": 16800
},
{
"epoch": 11.973454480669874,
"eval_loss": 0.09595508873462677,
"eval_runtime": 147.9777,
"eval_samples_per_second": 3.379,
"eval_steps_per_second": 0.426,
"eval_wer": 0.20291777188328913,
"step": 16800
},
{
"epoch": 12.044183146267594,
"grad_norm": 33.86606216430664,
"learning_rate": 2.983169885658002e-05,
"loss": 1.4119,
"step": 16900
},
{
"epoch": 12.115446285408872,
"grad_norm": 22.206035614013672,
"learning_rate": 2.9403451672305256e-05,
"loss": 1.4145,
"step": 17000
},
{
"epoch": 12.186709424550152,
"grad_norm": 30.262418746948242,
"learning_rate": 2.8975204488030495e-05,
"loss": 1.4809,
"step": 17100
},
{
"epoch": 12.25797256369143,
"grad_norm": 16.270219802856445,
"learning_rate": 2.8546957303755727e-05,
"loss": 1.4349,
"step": 17200
},
{
"epoch": 12.25797256369143,
"eval_loss": 0.09577883034944534,
"eval_runtime": 146.7879,
"eval_samples_per_second": 3.406,
"eval_steps_per_second": 0.429,
"eval_wer": 0.20085470085470086,
"step": 17200
},
{
"epoch": 12.32923570283271,
"grad_norm": 19.906198501586914,
"learning_rate": 2.8118710119480966e-05,
"loss": 1.4291,
"step": 17300
},
{
"epoch": 12.400498841973988,
"grad_norm": 29.69014549255371,
"learning_rate": 2.7690462935206202e-05,
"loss": 1.3993,
"step": 17400
},
{
"epoch": 12.471761981115268,
"grad_norm": 21.653276443481445,
"learning_rate": 2.726221575093144e-05,
"loss": 1.4257,
"step": 17500
},
{
"epoch": 12.543025120256548,
"grad_norm": 19.317724227905273,
"learning_rate": 2.6833968566656676e-05,
"loss": 1.4104,
"step": 17600
},
{
"epoch": 12.543025120256548,
"eval_loss": 0.09735482186079025,
"eval_runtime": 147.616,
"eval_samples_per_second": 3.387,
"eval_steps_per_second": 0.427,
"eval_wer": 0.20247568523430592,
"step": 17600
},
{
"epoch": 12.614288259397826,
"grad_norm": 24.637189865112305,
"learning_rate": 2.640572138238191e-05,
"loss": 1.343,
"step": 17700
},
{
"epoch": 12.685551398539106,
"grad_norm": 24.483217239379883,
"learning_rate": 2.5977474198107148e-05,
"loss": 1.4593,
"step": 17800
},
{
"epoch": 12.756814537680384,
"grad_norm": 17.71087074279785,
"learning_rate": 2.5549227013832383e-05,
"loss": 1.4749,
"step": 17900
},
{
"epoch": 12.828077676821664,
"grad_norm": 22.82439613342285,
"learning_rate": 2.5120979829557622e-05,
"loss": 1.5073,
"step": 18000
},
{
"epoch": 12.828077676821664,
"eval_loss": 0.09532631188631058,
"eval_runtime": 149.2979,
"eval_samples_per_second": 3.349,
"eval_steps_per_second": 0.422,
"eval_wer": 0.20439139404656645,
"step": 18000
},
{
"epoch": 12.899340815962944,
"grad_norm": 23.298507690429688,
"learning_rate": 2.4692732645282858e-05,
"loss": 1.4439,
"step": 18100
},
{
"epoch": 12.970603955104222,
"grad_norm": 18.416128158569336,
"learning_rate": 2.4264485461008097e-05,
"loss": 1.3723,
"step": 18200
},
{
"epoch": 13.041332620701942,
"grad_norm": 27.486845016479492,
"learning_rate": 2.383623827673333e-05,
"loss": 1.3505,
"step": 18300
},
{
"epoch": 13.112595759843222,
"grad_norm": 13.97179889678955,
"learning_rate": 2.3407991092458568e-05,
"loss": 1.2488,
"step": 18400
},
{
"epoch": 13.112595759843222,
"eval_loss": 0.09486553817987442,
"eval_runtime": 147.7955,
"eval_samples_per_second": 3.383,
"eval_steps_per_second": 0.426,
"eval_wer": 0.19658119658119658,
"step": 18400
},
{
"epoch": 13.1838588989845,
"grad_norm": 22.142789840698242,
"learning_rate": 2.2979743908183804e-05,
"loss": 1.3619,
"step": 18500
},
{
"epoch": 13.25512203812578,
"grad_norm": 24.32270050048828,
"learning_rate": 2.255149672390904e-05,
"loss": 1.3079,
"step": 18600
},
{
"epoch": 13.326385177267058,
"grad_norm": 20.621620178222656,
"learning_rate": 2.2123249539634278e-05,
"loss": 1.3346,
"step": 18700
},
{
"epoch": 13.397648316408338,
"grad_norm": 19.052913665771484,
"learning_rate": 2.1695002355359514e-05,
"loss": 1.277,
"step": 18800
},
{
"epoch": 13.397648316408338,
"eval_loss": 0.09550650417804718,
"eval_runtime": 150.9097,
"eval_samples_per_second": 3.313,
"eval_steps_per_second": 0.417,
"eval_wer": 0.20837017388741527,
"step": 18800
},
{
"epoch": 13.468911455549616,
"grad_norm": 29.400739669799805,
"learning_rate": 2.1266755171084753e-05,
"loss": 1.4286,
"step": 18900
},
{
"epoch": 13.540174594690896,
"grad_norm": 21.397193908691406,
"learning_rate": 2.083850798680999e-05,
"loss": 1.3856,
"step": 19000
},
{
"epoch": 13.611437733832176,
"grad_norm": 15.47054386138916,
"learning_rate": 2.0410260802535224e-05,
"loss": 1.285,
"step": 19100
},
{
"epoch": 13.682700872973454,
"grad_norm": 17.92173957824707,
"learning_rate": 1.9982013618260463e-05,
"loss": 1.2443,
"step": 19200
},
{
"epoch": 13.682700872973454,
"eval_loss": 0.09599152952432632,
"eval_runtime": 147.2957,
"eval_samples_per_second": 3.395,
"eval_steps_per_second": 0.428,
"eval_wer": 0.19952844090775126,
"step": 19200
},
{
"epoch": 13.753964012114734,
"grad_norm": 19.764711380004883,
"learning_rate": 1.95537664339857e-05,
"loss": 1.3789,
"step": 19300
},
{
"epoch": 13.825227151256012,
"grad_norm": 14.71243667602539,
"learning_rate": 1.9125519249710934e-05,
"loss": 1.3413,
"step": 19400
},
{
"epoch": 13.896490290397292,
"grad_norm": 23.086503982543945,
"learning_rate": 1.869727206543617e-05,
"loss": 1.361,
"step": 19500
},
{
"epoch": 13.967753429538572,
"grad_norm": 21.994956970214844,
"learning_rate": 1.8269024881161405e-05,
"loss": 1.3972,
"step": 19600
},
{
"epoch": 13.967753429538572,
"eval_loss": 0.09553589671850204,
"eval_runtime": 148.0027,
"eval_samples_per_second": 3.378,
"eval_steps_per_second": 0.426,
"eval_wer": 0.2027704096669614,
"step": 19600
},
{
"epoch": 14.03848209513629,
"grad_norm": 29.658489227294922,
"learning_rate": 1.7840777696886644e-05,
"loss": 1.259,
"step": 19700
},
{
"epoch": 14.10974523427757,
"grad_norm": 22.740825653076172,
"learning_rate": 1.741253051261188e-05,
"loss": 1.2666,
"step": 19800
},
{
"epoch": 14.18100837341885,
"grad_norm": 15.764113426208496,
"learning_rate": 1.6984283328337116e-05,
"loss": 1.2416,
"step": 19900
},
{
"epoch": 14.252271512560128,
"grad_norm": 16.60357093811035,
"learning_rate": 1.6556036144062355e-05,
"loss": 1.2847,
"step": 20000
},
{
"epoch": 14.252271512560128,
"eval_loss": 0.09485521167516708,
"eval_runtime": 148.9312,
"eval_samples_per_second": 3.357,
"eval_steps_per_second": 0.423,
"eval_wer": 0.20335985853227231,
"step": 20000
},
{
"epoch": 14.323534651701408,
"grad_norm": 13.188959121704102,
"learning_rate": 1.612778895978759e-05,
"loss": 1.2445,
"step": 20100
},
{
"epoch": 14.394797790842686,
"grad_norm": 15.217430114746094,
"learning_rate": 1.5699541775512826e-05,
"loss": 1.2725,
"step": 20200
},
{
"epoch": 14.466060929983966,
"grad_norm": 19.81468963623047,
"learning_rate": 1.5271294591238065e-05,
"loss": 1.2663,
"step": 20300
},
{
"epoch": 14.537324069125244,
"grad_norm": 16.013675689697266,
"learning_rate": 1.48430474069633e-05,
"loss": 1.3107,
"step": 20400
},
{
"epoch": 14.537324069125244,
"eval_loss": 0.09505146741867065,
"eval_runtime": 147.9624,
"eval_samples_per_second": 3.379,
"eval_steps_per_second": 0.426,
"eval_wer": 0.20129678750368404,
"step": 20400
},
{
"epoch": 14.608587208266524,
"grad_norm": 26.155323028564453,
"learning_rate": 1.4414800222688538e-05,
"loss": 1.3006,
"step": 20500
},
{
"epoch": 14.679850347407804,
"grad_norm": 25.04928970336914,
"learning_rate": 1.3986553038413775e-05,
"loss": 1.2639,
"step": 20600
},
{
"epoch": 14.751113486549082,
"grad_norm": 17.566251754760742,
"learning_rate": 1.3558305854139009e-05,
"loss": 1.262,
"step": 20700
},
{
"epoch": 14.822376625690362,
"grad_norm": 21.8209228515625,
"learning_rate": 1.3130058669864246e-05,
"loss": 1.2232,
"step": 20800
},
{
"epoch": 14.822376625690362,
"eval_loss": 0.0947200208902359,
"eval_runtime": 149.3452,
"eval_samples_per_second": 3.348,
"eval_steps_per_second": 0.422,
"eval_wer": 0.2002652519893899,
"step": 20800
},
{
"epoch": 14.89363976483164,
"grad_norm": 19.568490982055664,
"learning_rate": 1.2701811485589482e-05,
"loss": 1.2632,
"step": 20900
},
{
"epoch": 14.96490290397292,
"grad_norm": 14.820449829101562,
"learning_rate": 1.2273564301314719e-05,
"loss": 1.3063,
"step": 21000
},
{
"epoch": 15.03563156957064,
"grad_norm": 17.102474212646484,
"learning_rate": 1.1845317117039956e-05,
"loss": 1.2273,
"step": 21100
},
{
"epoch": 15.106894708711918,
"grad_norm": 16.626834869384766,
"learning_rate": 1.1417069932765192e-05,
"loss": 1.2233,
"step": 21200
},
{
"epoch": 15.106894708711918,
"eval_loss": 0.09491468966007233,
"eval_runtime": 148.8446,
"eval_samples_per_second": 3.359,
"eval_steps_per_second": 0.423,
"eval_wer": 0.19849690539345713,
"step": 21200
},
{
"epoch": 15.178157847853198,
"grad_norm": 18.95102310180664,
"learning_rate": 1.098882274849043e-05,
"loss": 1.1737,
"step": 21300
},
{
"epoch": 15.249420986994478,
"grad_norm": 15.553340911865234,
"learning_rate": 1.0560575564215666e-05,
"loss": 1.213,
"step": 21400
},
{
"epoch": 15.320684126135756,
"grad_norm": 25.440275192260742,
"learning_rate": 1.0132328379940904e-05,
"loss": 1.2196,
"step": 21500
},
{
"epoch": 15.391947265277036,
"grad_norm": 20.26975440979004,
"learning_rate": 9.708363667508887e-06,
"loss": 1.1999,
"step": 21600
},
{
"epoch": 15.391947265277036,
"eval_loss": 0.09457062929868698,
"eval_runtime": 149.2533,
"eval_samples_per_second": 3.35,
"eval_steps_per_second": 0.422,
"eval_wer": 0.20247568523430592,
"step": 21600
},
{
"epoch": 15.463210404418314,
"grad_norm": 12.798954010009766,
"learning_rate": 9.280116483234123e-06,
"loss": 1.2562,
"step": 21700
},
{
"epoch": 15.534473543559594,
"grad_norm": 15.019392967224121,
"learning_rate": 8.85186929895936e-06,
"loss": 1.2189,
"step": 21800
},
{
"epoch": 15.605736682700872,
"grad_norm": 16.623435974121094,
"learning_rate": 8.423622114684597e-06,
"loss": 1.2374,
"step": 21900
},
{
"epoch": 15.676999821842152,
"grad_norm": 26.94820213317871,
"learning_rate": 7.995374930409833e-06,
"loss": 1.236,
"step": 22000
},
{
"epoch": 15.676999821842152,
"eval_loss": 0.09487726539373398,
"eval_runtime": 150.1052,
"eval_samples_per_second": 3.331,
"eval_steps_per_second": 0.42,
"eval_wer": 0.20291777188328913,
"step": 22000
},
{
"epoch": 15.748262960983432,
"grad_norm": 29.318561553955078,
"learning_rate": 7.567127746135069e-06,
"loss": 1.2446,
"step": 22100
},
{
"epoch": 15.81952610012471,
"grad_norm": 11.961990356445312,
"learning_rate": 7.1388805618603064e-06,
"loss": 1.2262,
"step": 22200
},
{
"epoch": 15.89078923926599,
"grad_norm": 19.715322494506836,
"learning_rate": 6.710633377585543e-06,
"loss": 1.1431,
"step": 22300
},
{
"epoch": 15.962052378407268,
"grad_norm": 16.766855239868164,
"learning_rate": 6.282386193310779e-06,
"loss": 1.2252,
"step": 22400
},
{
"epoch": 15.962052378407268,
"eval_loss": 0.09453491866588593,
"eval_runtime": 148.1933,
"eval_samples_per_second": 3.374,
"eval_steps_per_second": 0.425,
"eval_wer": 0.19938107869142352,
"step": 22400
},
{
"epoch": 16.032781044004988,
"grad_norm": 24.855350494384766,
"learning_rate": 5.854139009036016e-06,
"loss": 1.0984,
"step": 22500
},
{
"epoch": 16.104044183146268,
"grad_norm": 16.3688907623291,
"learning_rate": 5.425891824761253e-06,
"loss": 1.2072,
"step": 22600
},
{
"epoch": 16.175307322287548,
"grad_norm": 21.17823600769043,
"learning_rate": 4.997644640486489e-06,
"loss": 1.156,
"step": 22700
},
{
"epoch": 16.246570461428824,
"grad_norm": 13.77876091003418,
"learning_rate": 4.569397456211726e-06,
"loss": 1.2094,
"step": 22800
},
{
"epoch": 16.246570461428824,
"eval_loss": 0.094144806265831,
"eval_runtime": 148.3747,
"eval_samples_per_second": 3.37,
"eval_steps_per_second": 0.425,
"eval_wer": 0.2049808429118774,
"step": 22800
},
{
"epoch": 16.317833600570104,
"grad_norm": 15.153692245483398,
"learning_rate": 4.141150271936962e-06,
"loss": 1.184,
"step": 22900
},
{
"epoch": 16.389096739711384,
"grad_norm": 17.136568069458008,
"learning_rate": 3.7129030876621984e-06,
"loss": 1.18,
"step": 23000
},
{
"epoch": 16.460359878852664,
"grad_norm": 17.46211051940918,
"learning_rate": 3.2846559033874353e-06,
"loss": 1.173,
"step": 23100
},
{
"epoch": 16.531623017993944,
"grad_norm": 13.692911148071289,
"learning_rate": 2.8564087191126717e-06,
"loss": 1.2505,
"step": 23200
},
{
"epoch": 16.531623017993944,
"eval_loss": 0.0941072553396225,
"eval_runtime": 147.9642,
"eval_samples_per_second": 3.379,
"eval_steps_per_second": 0.426,
"eval_wer": 0.2002652519893899,
"step": 23200
},
{
"epoch": 16.60288615713522,
"grad_norm": 12.050436019897461,
"learning_rate": 2.4281615348379086e-06,
"loss": 1.1627,
"step": 23300
},
{
"epoch": 16.6741492962765,
"grad_norm": 13.23355484008789,
"learning_rate": 1.999914350563145e-06,
"loss": 1.1094,
"step": 23400
},
{
"epoch": 16.74541243541778,
"grad_norm": 11.773246765136719,
"learning_rate": 1.5716671662883815e-06,
"loss": 1.1767,
"step": 23500
},
{
"epoch": 16.81667557455906,
"grad_norm": 11.992323875427246,
"learning_rate": 1.1434199820136182e-06,
"loss": 1.1193,
"step": 23600
},
{
"epoch": 16.81667557455906,
"eval_loss": 0.09421209245920181,
"eval_runtime": 146.8439,
"eval_samples_per_second": 3.405,
"eval_steps_per_second": 0.429,
"eval_wer": 0.19908635425876806,
"step": 23600
},
{
"epoch": 16.88793871370034,
"grad_norm": 13.370680809020996,
"learning_rate": 7.151727977388549e-07,
"loss": 1.1688,
"step": 23700
},
{
"epoch": 16.959201852841616,
"grad_norm": 11.178988456726074,
"learning_rate": 2.869256134640915e-07,
"loss": 1.1531,
"step": 23800
},
{
"epoch": 17.034918938179228,
"grad_norm": 21.082555770874023,
"learning_rate": 3.236731742588576e-05,
"loss": 1.1705,
"step": 23900
},
{
"epoch": 17.106182077320504,
"grad_norm": 11.843573570251465,
"learning_rate": 3.208098336948662e-05,
"loss": 1.1992,
"step": 24000
},
{
"epoch": 17.106182077320504,
"eval_loss": 0.09458199143409729,
"eval_runtime": 165.5872,
"eval_samples_per_second": 3.02,
"eval_steps_per_second": 0.38,
"eval_wer": 0.20203359858532272,
"step": 24000
},
{
"epoch": 17.177445216461784,
"grad_norm": 23.818212509155273,
"learning_rate": 3.179175704989154e-05,
"loss": 1.2774,
"step": 24100
},
{
"epoch": 17.248708355603064,
"grad_norm": 13.351399421691895,
"learning_rate": 3.150253073029646e-05,
"loss": 1.2139,
"step": 24200
},
{
"epoch": 17.319971494744344,
"grad_norm": 34.073760986328125,
"learning_rate": 3.121330441070138e-05,
"loss": 1.2695,
"step": 24300
},
{
"epoch": 17.391234633885624,
"grad_norm": 28.087905883789062,
"learning_rate": 3.092407809110629e-05,
"loss": 1.2794,
"step": 24400
},
{
"epoch": 17.391234633885624,
"eval_loss": 0.09535403549671173,
"eval_runtime": 164.8794,
"eval_samples_per_second": 3.033,
"eval_steps_per_second": 0.382,
"eval_wer": 0.21175950486295314,
"step": 24400
},
{
"epoch": 17.4624977730269,
"grad_norm": 13.820633888244629,
"learning_rate": 3.063485177151121e-05,
"loss": 1.2422,
"step": 24500
},
{
"epoch": 17.53376091216818,
"grad_norm": 23.503931045532227,
"learning_rate": 3.0345625451916128e-05,
"loss": 1.2676,
"step": 24600
},
{
"epoch": 17.60502405130946,
"grad_norm": 25.35702133178711,
"learning_rate": 3.0056399132321044e-05,
"loss": 1.2898,
"step": 24700
},
{
"epoch": 17.67628719045074,
"grad_norm": 22.99468231201172,
"learning_rate": 2.9767172812725957e-05,
"loss": 1.2362,
"step": 24800
},
{
"epoch": 17.67628719045074,
"eval_loss": 0.0947684794664383,
"eval_runtime": 160.2628,
"eval_samples_per_second": 3.12,
"eval_steps_per_second": 0.393,
"eval_wer": 0.20247568523430592,
"step": 24800
},
{
"epoch": 17.74755032959202,
"grad_norm": 18.09113311767578,
"learning_rate": 2.9477946493130876e-05,
"loss": 1.2968,
"step": 24900
},
{
"epoch": 17.818813468733296,
"grad_norm": 19.54793930053711,
"learning_rate": 2.9188720173535792e-05,
"loss": 1.2264,
"step": 25000
},
{
"epoch": 17.890076607874576,
"grad_norm": 20.854331970214844,
"learning_rate": 2.8899493853940712e-05,
"loss": 1.2457,
"step": 25100
},
{
"epoch": 17.961339747015856,
"grad_norm": 19.649433135986328,
"learning_rate": 2.8610267534345625e-05,
"loss": 1.3528,
"step": 25200
},
{
"epoch": 17.961339747015856,
"eval_loss": 0.09558839350938797,
"eval_runtime": 162.932,
"eval_samples_per_second": 3.069,
"eval_steps_per_second": 0.387,
"eval_wer": 0.20704391394046567,
"step": 25200
},
{
"epoch": 18.032068412613576,
"grad_norm": 12.970821380615234,
"learning_rate": 2.8321041214750544e-05,
"loss": 1.3198,
"step": 25300
},
{
"epoch": 18.103331551754856,
"grad_norm": 17.95279884338379,
"learning_rate": 2.803181489515546e-05,
"loss": 1.2166,
"step": 25400
},
{
"epoch": 18.174594690896132,
"grad_norm": 15.008156776428223,
"learning_rate": 2.774258857556038e-05,
"loss": 1.1703,
"step": 25500
},
{
"epoch": 18.245857830037412,
"grad_norm": 28.749645233154297,
"learning_rate": 2.7453362255965293e-05,
"loss": 1.1863,
"step": 25600
},
{
"epoch": 18.245857830037412,
"eval_loss": 0.09352152794599533,
"eval_runtime": 161.799,
"eval_samples_per_second": 3.09,
"eval_steps_per_second": 0.389,
"eval_wer": 0.2036545829649278,
"step": 25600
},
{
"epoch": 18.317120969178692,
"grad_norm": 25.244779586791992,
"learning_rate": 2.7164135936370212e-05,
"loss": 1.1881,
"step": 25700
},
{
"epoch": 18.388384108319972,
"grad_norm": 28.115192413330078,
"learning_rate": 2.687490961677513e-05,
"loss": 1.1862,
"step": 25800
},
{
"epoch": 18.459647247461252,
"grad_norm": 14.245773315429688,
"learning_rate": 2.6585683297180048e-05,
"loss": 1.2374,
"step": 25900
},
{
"epoch": 18.530910386602528,
"grad_norm": 17.683853149414062,
"learning_rate": 2.629645697758496e-05,
"loss": 1.2936,
"step": 26000
},
{
"epoch": 18.530910386602528,
"eval_loss": 0.09395604580640793,
"eval_runtime": 159.852,
"eval_samples_per_second": 3.128,
"eval_steps_per_second": 0.394,
"eval_wer": 0.2032124963159446,
"step": 26000
},
{
"epoch": 18.602173525743808,
"grad_norm": 23.95792579650879,
"learning_rate": 2.6007230657989877e-05,
"loss": 1.2801,
"step": 26100
},
{
"epoch": 18.673436664885088,
"grad_norm": 33.534027099609375,
"learning_rate": 2.5718004338394796e-05,
"loss": 1.1837,
"step": 26200
},
{
"epoch": 18.744699804026368,
"grad_norm": 25.70989227294922,
"learning_rate": 2.542877801879971e-05,
"loss": 1.2212,
"step": 26300
},
{
"epoch": 18.815962943167648,
"grad_norm": 19.731300354003906,
"learning_rate": 2.513955169920463e-05,
"loss": 1.2434,
"step": 26400
},
{
"epoch": 18.815962943167648,
"eval_loss": 0.09383614361286163,
"eval_runtime": 162.1968,
"eval_samples_per_second": 3.083,
"eval_steps_per_second": 0.388,
"eval_wer": 0.20291777188328913,
"step": 26400
},
{
"epoch": 18.887226082308924,
"grad_norm": 21.168506622314453,
"learning_rate": 2.4850325379609545e-05,
"loss": 1.2603,
"step": 26500
},
{
"epoch": 18.958489221450204,
"grad_norm": 22.403703689575195,
"learning_rate": 2.456109906001446e-05,
"loss": 1.2183,
"step": 26600
},
{
"epoch": 19.029217887047924,
"grad_norm": 18.090112686157227,
"learning_rate": 2.427187274041938e-05,
"loss": 1.1921,
"step": 26700
},
{
"epoch": 19.100481026189204,
"grad_norm": 17.219758987426758,
"learning_rate": 2.3982646420824293e-05,
"loss": 1.1254,
"step": 26800
},
{
"epoch": 19.100481026189204,
"eval_loss": 0.09326955676078796,
"eval_runtime": 160.0801,
"eval_samples_per_second": 3.123,
"eval_steps_per_second": 0.394,
"eval_wer": 0.20262304745063367,
"step": 26800
},
{
"epoch": 19.171744165330484,
"grad_norm": 12.749887466430664,
"learning_rate": 2.3696312364425163e-05,
"loss": 1.1345,
"step": 26900
},
{
"epoch": 19.24300730447176,
"grad_norm": 31.437406539916992,
"learning_rate": 2.3407086044830083e-05,
"loss": 1.1849,
"step": 27000
},
{
"epoch": 19.31427044361304,
"grad_norm": 12.56886100769043,
"learning_rate": 2.3117859725234995e-05,
"loss": 1.2,
"step": 27100
},
{
"epoch": 19.38553358275432,
"grad_norm": 15.81522274017334,
"learning_rate": 2.2828633405639915e-05,
"loss": 1.2345,
"step": 27200
},
{
"epoch": 19.38553358275432,
"eval_loss": 0.0934244766831398,
"eval_runtime": 158.6427,
"eval_samples_per_second": 3.152,
"eval_steps_per_second": 0.397,
"eval_wer": 0.20085470085470086,
"step": 27200
},
{
"epoch": 19.4567967218956,
"grad_norm": 22.67652130126953,
"learning_rate": 2.253940708604483e-05,
"loss": 1.1491,
"step": 27300
},
{
"epoch": 19.52805986103688,
"grad_norm": 25.45644760131836,
"learning_rate": 2.225018076644975e-05,
"loss": 1.2354,
"step": 27400
},
{
"epoch": 19.599323000178156,
"grad_norm": 17.67236328125,
"learning_rate": 2.1960954446854663e-05,
"loss": 1.1475,
"step": 27500
},
{
"epoch": 19.670586139319436,
"grad_norm": 16.663982391357422,
"learning_rate": 2.1671728127259583e-05,
"loss": 1.2177,
"step": 27600
},
{
"epoch": 19.670586139319436,
"eval_loss": 0.09379982203245163,
"eval_runtime": 158.8505,
"eval_samples_per_second": 3.148,
"eval_steps_per_second": 0.397,
"eval_wer": 0.2036545829649278,
"step": 27600
},
{
"epoch": 19.741849278460716,
"grad_norm": 17.305145263671875,
"learning_rate": 2.13825018076645e-05,
"loss": 1.1699,
"step": 27700
},
{
"epoch": 19.813112417601996,
"grad_norm": 15.652444839477539,
"learning_rate": 2.1093275488069415e-05,
"loss": 1.1853,
"step": 27800
},
{
"epoch": 19.884375556743276,
"grad_norm": 12.307770729064941,
"learning_rate": 2.080404916847433e-05,
"loss": 1.1209,
"step": 27900
},
{
"epoch": 19.955638695884552,
"grad_norm": 16.08258056640625,
"learning_rate": 2.0514822848879248e-05,
"loss": 1.1479,
"step": 28000
},
{
"epoch": 19.955638695884552,
"eval_loss": 0.09380526840686798,
"eval_runtime": 158.3203,
"eval_samples_per_second": 3.158,
"eval_steps_per_second": 0.398,
"eval_wer": 0.20070733863837312,
"step": 28000
},
{
"epoch": 20.026367361482272,
"grad_norm": 13.121270179748535,
"learning_rate": 2.0225596529284167e-05,
"loss": 1.1111,
"step": 28100
},
{
"epoch": 20.097630500623552,
"grad_norm": 26.73927116394043,
"learning_rate": 1.993637020968908e-05,
"loss": 1.1596,
"step": 28200
},
{
"epoch": 20.168893639764832,
"grad_norm": 14.311580657958984,
"learning_rate": 1.9647143890094e-05,
"loss": 1.1404,
"step": 28300
},
{
"epoch": 20.240156778906112,
"grad_norm": 17.167621612548828,
"learning_rate": 1.9357917570498916e-05,
"loss": 1.1077,
"step": 28400
},
{
"epoch": 20.240156778906112,
"eval_loss": 0.09331627190113068,
"eval_runtime": 158.3839,
"eval_samples_per_second": 3.157,
"eval_steps_per_second": 0.398,
"eval_wer": 0.19952844090775126,
"step": 28400
},
{
"epoch": 20.31141991804739,
"grad_norm": 26.617481231689453,
"learning_rate": 1.9068691250903832e-05,
"loss": 1.1133,
"step": 28500
},
{
"epoch": 20.382683057188668,
"grad_norm": 15.19737434387207,
"learning_rate": 1.8779464931308748e-05,
"loss": 1.1067,
"step": 28600
},
{
"epoch": 20.453946196329948,
"grad_norm": 16.4631404876709,
"learning_rate": 1.8490238611713668e-05,
"loss": 1.0911,
"step": 28700
},
{
"epoch": 20.525209335471228,
"grad_norm": 16.38656997680664,
"learning_rate": 1.8201012292118584e-05,
"loss": 1.1615,
"step": 28800
},
{
"epoch": 20.525209335471228,
"eval_loss": 0.09307432919740677,
"eval_runtime": 160.2363,
"eval_samples_per_second": 3.12,
"eval_steps_per_second": 0.393,
"eval_wer": 0.20247568523430592,
"step": 28800
},
{
"epoch": 20.596472474612508,
"grad_norm": 12.079431533813477,
"learning_rate": 1.79117859725235e-05,
"loss": 1.1665,
"step": 28900
},
{
"epoch": 20.667735613753784,
"grad_norm": 20.58789825439453,
"learning_rate": 1.7622559652928416e-05,
"loss": 1.1225,
"step": 29000
},
{
"epoch": 20.738998752895064,
"grad_norm": 20.81654930114746,
"learning_rate": 1.7333333333333336e-05,
"loss": 1.0583,
"step": 29100
},
{
"epoch": 20.810261892036344,
"grad_norm": 19.925395965576172,
"learning_rate": 1.7044107013738252e-05,
"loss": 1.0642,
"step": 29200
},
{
"epoch": 20.810261892036344,
"eval_loss": 0.09396182000637054,
"eval_runtime": 160.8477,
"eval_samples_per_second": 3.109,
"eval_steps_per_second": 0.392,
"eval_wer": 0.2045387562628942,
"step": 29200
},
{
"epoch": 20.881525031177624,
"grad_norm": 15.387258529663086,
"learning_rate": 1.6754880694143168e-05,
"loss": 1.1447,
"step": 29300
},
{
"epoch": 20.952788170318904,
"grad_norm": 15.967385292053223,
"learning_rate": 1.6465654374548084e-05,
"loss": 1.1769,
"step": 29400
},
{
"epoch": 21.023516835916624,
"grad_norm": 14.936897277832031,
"learning_rate": 1.6176428054953004e-05,
"loss": 1.0766,
"step": 29500
},
{
"epoch": 21.0947799750579,
"grad_norm": 15.853533744812012,
"learning_rate": 1.5887201735357916e-05,
"loss": 1.0922,
"step": 29600
},
{
"epoch": 21.0947799750579,
"eval_loss": 0.09347377717494965,
"eval_runtime": 159.5968,
"eval_samples_per_second": 3.133,
"eval_steps_per_second": 0.395,
"eval_wer": 0.20114942528735633,
"step": 29600
},
{
"epoch": 21.16604311419918,
"grad_norm": 20.242115020751953,
"learning_rate": 1.5597975415762836e-05,
"loss": 1.1478,
"step": 29700
},
{
"epoch": 21.23730625334046,
"grad_norm": 19.621402740478516,
"learning_rate": 1.5308749096167752e-05,
"loss": 1.0792,
"step": 29800
},
{
"epoch": 21.30856939248174,
"grad_norm": 14.971497535705566,
"learning_rate": 1.501952277657267e-05,
"loss": 1.0581,
"step": 29900
},
{
"epoch": 21.379832531623016,
"grad_norm": 13.164278030395508,
"learning_rate": 1.4730296456977585e-05,
"loss": 1.0885,
"step": 30000
},
{
"epoch": 21.379832531623016,
"eval_loss": 0.0928725004196167,
"eval_runtime": 161.3647,
"eval_samples_per_second": 3.099,
"eval_steps_per_second": 0.39,
"eval_wer": 0.20100206307102858,
"step": 30000
},
{
"epoch": 21.451095670764296,
"grad_norm": 13.703166961669922,
"learning_rate": 1.4441070137382504e-05,
"loss": 1.0712,
"step": 30100
},
{
"epoch": 21.522358809905576,
"grad_norm": 11.881406784057617,
"learning_rate": 1.4151843817787419e-05,
"loss": 1.0534,
"step": 30200
},
{
"epoch": 21.593621949046856,
"grad_norm": 17.952503204345703,
"learning_rate": 1.3862617498192338e-05,
"loss": 1.0741,
"step": 30300
},
{
"epoch": 21.664885088188136,
"grad_norm": 14.229839324951172,
"learning_rate": 1.3573391178597253e-05,
"loss": 1.107,
"step": 30400
},
{
"epoch": 21.664885088188136,
"eval_loss": 0.09303069859743118,
"eval_runtime": 157.8883,
"eval_samples_per_second": 3.167,
"eval_steps_per_second": 0.399,
"eval_wer": 0.1987916298261126,
"step": 30400
},
{
"epoch": 21.736148227329412,
"grad_norm": 17.410789489746094,
"learning_rate": 1.328416485900217e-05,
"loss": 1.0532,
"step": 30500
},
{
"epoch": 21.807411366470692,
"grad_norm": 19.12940788269043,
"learning_rate": 1.2994938539407087e-05,
"loss": 1.0863,
"step": 30600
},
{
"epoch": 21.878674505611972,
"grad_norm": 10.864148139953613,
"learning_rate": 1.2705712219812004e-05,
"loss": 1.0953,
"step": 30700
},
{
"epoch": 21.949937644753252,
"grad_norm": 15.553997993469238,
"learning_rate": 1.241648590021692e-05,
"loss": 1.0449,
"step": 30800
},
{
"epoch": 21.949937644753252,
"eval_loss": 0.09309829026460648,
"eval_runtime": 158.4381,
"eval_samples_per_second": 3.156,
"eval_steps_per_second": 0.398,
"eval_wer": 0.2001178897730622,
"step": 30800
},
{
"epoch": 22.020666310350972,
"grad_norm": 16.839780807495117,
"learning_rate": 1.2130151843817789e-05,
"loss": 1.0535,
"step": 30900
},
{
"epoch": 22.091929449492252,
"grad_norm": 14.113849639892578,
"learning_rate": 1.1840925524222705e-05,
"loss": 1.0341,
"step": 31000
},
{
"epoch": 22.163192588633528,
"grad_norm": 16.825448989868164,
"learning_rate": 1.1551699204627623e-05,
"loss": 1.0091,
"step": 31100
},
{
"epoch": 22.234455727774808,
"grad_norm": 17.561140060424805,
"learning_rate": 1.1262472885032537e-05,
"loss": 1.033,
"step": 31200
},
{
"epoch": 22.234455727774808,
"eval_loss": 0.09307766705751419,
"eval_runtime": 159.9295,
"eval_samples_per_second": 3.126,
"eval_steps_per_second": 0.394,
"eval_wer": 0.20483348069554966,
"step": 31200
},
{
"epoch": 22.305718866916088,
"grad_norm": 11.815247535705566,
"learning_rate": 1.0973246565437455e-05,
"loss": 1.0777,
"step": 31300
},
{
"epoch": 22.376982006057368,
"grad_norm": 14.315788269042969,
"learning_rate": 1.0684020245842371e-05,
"loss": 1.0391,
"step": 31400
},
{
"epoch": 22.448245145198648,
"grad_norm": 20.650876998901367,
"learning_rate": 1.0394793926247289e-05,
"loss": 1.0338,
"step": 31500
},
{
"epoch": 22.519508284339924,
"grad_norm": 11.829896926879883,
"learning_rate": 1.0105567606652205e-05,
"loss": 1.057,
"step": 31600
},
{
"epoch": 22.519508284339924,
"eval_loss": 0.09323979169130325,
"eval_runtime": 157.3931,
"eval_samples_per_second": 3.177,
"eval_steps_per_second": 0.4,
"eval_wer": 0.1987916298261126,
"step": 31600
},
{
"epoch": 22.590771423481204,
"grad_norm": 16.371854782104492,
"learning_rate": 9.816341287057123e-06,
"loss": 1.0875,
"step": 31700
},
{
"epoch": 22.662034562622484,
"grad_norm": 12.373185157775879,
"learning_rate": 9.527114967462039e-06,
"loss": 1.0525,
"step": 31800
},
{
"epoch": 22.733297701763764,
"grad_norm": 16.901241302490234,
"learning_rate": 9.237888647866955e-06,
"loss": 0.9976,
"step": 31900
},
{
"epoch": 22.80456084090504,
"grad_norm": 11.734404563903809,
"learning_rate": 8.948662328271873e-06,
"loss": 1.0248,
"step": 32000
},
{
"epoch": 22.80456084090504,
"eval_loss": 0.09294673800468445,
"eval_runtime": 159.1061,
"eval_samples_per_second": 3.143,
"eval_steps_per_second": 0.396,
"eval_wer": 0.201886236368995,
"step": 32000
},
{
"epoch": 22.87582398004632,
"grad_norm": 13.594120025634766,
"learning_rate": 8.65943600867679e-06,
"loss": 1.0777,
"step": 32100
},
{
"epoch": 22.9470871191876,
"grad_norm": 15.224822044372559,
"learning_rate": 8.370209689081707e-06,
"loss": 1.0544,
"step": 32200
},
{
"epoch": 23.01781578478532,
"grad_norm": 11.079718589782715,
"learning_rate": 8.080983369486623e-06,
"loss": 1.0563,
"step": 32300
},
{
"epoch": 23.0890789239266,
"grad_norm": 12.417219161987305,
"learning_rate": 7.791757049891541e-06,
"loss": 0.9784,
"step": 32400
},
{
"epoch": 23.0890789239266,
"eval_loss": 0.0926588624715805,
"eval_runtime": 159.1992,
"eval_samples_per_second": 3.141,
"eval_steps_per_second": 0.396,
"eval_wer": 0.19510757441791923,
"step": 32400
},
{
"epoch": 23.16034206306788,
"grad_norm": 15.528765678405762,
"learning_rate": 7.502530730296457e-06,
"loss": 0.9855,
"step": 32500
},
{
"epoch": 23.231605202209156,
"grad_norm": 11.74361801147461,
"learning_rate": 7.213304410701374e-06,
"loss": 1.0193,
"step": 32600
},
{
"epoch": 23.302868341350436,
"grad_norm": 11.150557518005371,
"learning_rate": 6.924078091106291e-06,
"loss": 0.9878,
"step": 32700
},
{
"epoch": 23.374131480491716,
"grad_norm": 11.871788024902344,
"learning_rate": 6.634851771511208e-06,
"loss": 1.0443,
"step": 32800
},
{
"epoch": 23.374131480491716,
"eval_loss": 0.09271341562271118,
"eval_runtime": 157.7057,
"eval_samples_per_second": 3.17,
"eval_steps_per_second": 0.399,
"eval_wer": 0.19952844090775126,
"step": 32800
},
{
"epoch": 23.445394619632996,
"grad_norm": 17.588510513305664,
"learning_rate": 6.3456254519161245e-06,
"loss": 1.019,
"step": 32900
},
{
"epoch": 23.516657758774272,
"grad_norm": 13.247255325317383,
"learning_rate": 6.0563991323210416e-06,
"loss": 1.0711,
"step": 33000
},
{
"epoch": 23.587920897915552,
"grad_norm": 10.2926025390625,
"learning_rate": 5.7671728127259586e-06,
"loss": 1.0257,
"step": 33100
},
{
"epoch": 23.659184037056832,
"grad_norm": 13.390459060668945,
"learning_rate": 5.4779464931308756e-06,
"loss": 0.9972,
"step": 33200
},
{
"epoch": 23.659184037056832,
"eval_loss": 0.09225763380527496,
"eval_runtime": 158.7123,
"eval_samples_per_second": 3.15,
"eval_steps_per_second": 0.397,
"eval_wer": 0.19952844090775126,
"step": 33200
},
{
"epoch": 23.730447176198112,
"grad_norm": 18.851301193237305,
"learning_rate": 5.188720173535792e-06,
"loss": 1.0099,
"step": 33300
},
{
"epoch": 23.801710315339392,
"grad_norm": 12.088775634765625,
"learning_rate": 4.899493853940709e-06,
"loss": 0.9931,
"step": 33400
},
{
"epoch": 23.872973454480668,
"grad_norm": 12.231010437011719,
"learning_rate": 4.610267534345626e-06,
"loss": 1.0545,
"step": 33500
},
{
"epoch": 23.944236593621948,
"grad_norm": 12.138983726501465,
"learning_rate": 4.323933477946494e-06,
"loss": 1.0527,
"step": 33600
},
{
"epoch": 23.944236593621948,
"eval_loss": 0.09295401722192764,
"eval_runtime": 158.8474,
"eval_samples_per_second": 3.148,
"eval_steps_per_second": 0.397,
"eval_wer": 0.19643383436486886,
"step": 33600
},
{
"epoch": 24.014965259219668,
"grad_norm": 12.55706787109375,
"learning_rate": 4.03470715835141e-06,
"loss": 0.9904,
"step": 33700
},
{
"epoch": 24.086228398360948,
"grad_norm": 15.65674114227295,
"learning_rate": 3.745480838756327e-06,
"loss": 1.0292,
"step": 33800
},
{
"epoch": 24.157491537502228,
"grad_norm": 12.829319953918457,
"learning_rate": 3.4562545191612434e-06,
"loss": 0.9981,
"step": 33900
},
{
"epoch": 24.228754676643508,
"grad_norm": 10.557638168334961,
"learning_rate": 3.1670281995661605e-06,
"loss": 0.9927,
"step": 34000
},
{
"epoch": 24.228754676643508,
"eval_loss": 0.09272577613592148,
"eval_runtime": 160.295,
"eval_samples_per_second": 3.119,
"eval_steps_per_second": 0.393,
"eval_wer": 0.19790745652814618,
"step": 34000
},
{
"epoch": 24.300017815784784,
"grad_norm": 11.129683494567871,
"learning_rate": 2.8778018799710775e-06,
"loss": 0.9646,
"step": 34100
},
{
"epoch": 24.371280954926064,
"grad_norm": 11.022905349731445,
"learning_rate": 2.588575560375994e-06,
"loss": 0.9873,
"step": 34200
},
{
"epoch": 24.442544094067344,
"grad_norm": 9.839102745056152,
"learning_rate": 2.299349240780911e-06,
"loss": 0.9995,
"step": 34300
},
{
"epoch": 24.513807233208624,
"grad_norm": 10.704620361328125,
"learning_rate": 2.010122921185828e-06,
"loss": 0.9504,
"step": 34400
},
{
"epoch": 24.513807233208624,
"eval_loss": 0.09273621439933777,
"eval_runtime": 158.2643,
"eval_samples_per_second": 3.159,
"eval_steps_per_second": 0.398,
"eval_wer": 0.19599174771588565,
"step": 34400
},
{
"epoch": 24.585070372349904,
"grad_norm": 15.199273109436035,
"learning_rate": 1.7208966015907449e-06,
"loss": 0.99,
"step": 34500
},
{
"epoch": 24.65633351149118,
"grad_norm": 8.751911163330078,
"learning_rate": 1.4316702819956617e-06,
"loss": 0.9746,
"step": 34600
},
{
"epoch": 24.72759665063246,
"grad_norm": 11.815445899963379,
"learning_rate": 1.1424439624005785e-06,
"loss": 1.0184,
"step": 34700
},
{
"epoch": 24.79885978977374,
"grad_norm": 12.593855857849121,
"learning_rate": 8.532176428054954e-07,
"loss": 1.0567,
"step": 34800
},
{
"epoch": 24.79885978977374,
"eval_loss": 0.09248282015323639,
"eval_runtime": 161.1515,
"eval_samples_per_second": 3.103,
"eval_steps_per_second": 0.391,
"eval_wer": 0.19864426760978485,
"step": 34800
},
{
"epoch": 24.875111348654908,
"grad_norm": 10.561070442199707,
"learning_rate": 5.639913232104122e-07,
"loss": 0.9481,
"step": 34900
},
{
"epoch": 24.946374487796188,
"grad_norm": 11.20162296295166,
"learning_rate": 2.74765003615329e-07,
"loss": 0.9673,
"step": 35000
},
{
"epoch": 25.01781578478532,
"grad_norm": 15.058631896972656,
"learning_rate": 1.6855013224332772e-05,
"loss": 0.9582,
"step": 35100
},
{
"epoch": 25.0890789239266,
"grad_norm": 11.464241027832031,
"learning_rate": 1.6614570810290935e-05,
"loss": 1.0316,
"step": 35200
},
{
"epoch": 25.0890789239266,
"eval_loss": 0.09259089827537537,
"eval_runtime": 149.1152,
"eval_samples_per_second": 3.353,
"eval_steps_per_second": 0.422,
"eval_wer": 0.19834954317712938,
"step": 35200
},
{
"epoch": 25.16034206306788,
"grad_norm": 23.382781982421875,
"learning_rate": 1.63741283962491e-05,
"loss": 0.9861,
"step": 35300
},
{
"epoch": 25.231605202209156,
"grad_norm": 16.841032028198242,
"learning_rate": 1.6133685982207263e-05,
"loss": 1.0558,
"step": 35400
},
{
"epoch": 25.302868341350436,
"grad_norm": 15.60053825378418,
"learning_rate": 1.5893243568165425e-05,
"loss": 1.0135,
"step": 35500
},
{
"epoch": 25.374131480491716,
"grad_norm": 14.612176895141602,
"learning_rate": 1.5655205578264005e-05,
"loss": 0.9926,
"step": 35600
},
{
"epoch": 25.374131480491716,
"eval_loss": 0.09281676262617111,
"eval_runtime": 146.8753,
"eval_samples_per_second": 3.404,
"eval_steps_per_second": 0.429,
"eval_wer": 0.19820218096080164,
"step": 35600
},
{
"epoch": 25.445394619632996,
"grad_norm": 21.735061645507812,
"learning_rate": 1.541476316422217e-05,
"loss": 1.0212,
"step": 35700
},
{
"epoch": 25.516657758774272,
"grad_norm": 15.282707214355469,
"learning_rate": 1.5174320750180331e-05,
"loss": 1.0118,
"step": 35800
},
{
"epoch": 25.587920897915552,
"grad_norm": 25.81006622314453,
"learning_rate": 1.4933878336138495e-05,
"loss": 1.0535,
"step": 35900
},
{
"epoch": 25.659184037056832,
"grad_norm": 12.21265697479248,
"learning_rate": 1.4693435922096658e-05,
"loss": 1.0646,
"step": 36000
},
{
"epoch": 25.659184037056832,
"eval_loss": 0.0927177220582962,
"eval_runtime": 148.3711,
"eval_samples_per_second": 3.37,
"eval_steps_per_second": 0.425,
"eval_wer": 0.2005599764220454,
"step": 36000
},
{
"epoch": 25.730447176198112,
"grad_norm": 13.201019287109375,
"learning_rate": 1.4452993508054822e-05,
"loss": 1.0161,
"step": 36100
},
{
"epoch": 25.801710315339392,
"grad_norm": 9.418228149414062,
"learning_rate": 1.4212551094012986e-05,
"loss": 1.0435,
"step": 36200
},
{
"epoch": 25.872973454480668,
"grad_norm": 19.827404022216797,
"learning_rate": 1.3972108679971146e-05,
"loss": 1.0984,
"step": 36300
},
{
"epoch": 25.944236593621948,
"grad_norm": 18.476207733154297,
"learning_rate": 1.373166626592931e-05,
"loss": 1.0316,
"step": 36400
},
{
"epoch": 25.944236593621948,
"eval_loss": 0.09291358292102814,
"eval_runtime": 147.4091,
"eval_samples_per_second": 3.392,
"eval_steps_per_second": 0.427,
"eval_wer": 0.20380194518125552,
"step": 36400
},
{
"epoch": 26.014965259219668,
"grad_norm": 11.769281387329102,
"learning_rate": 1.3491223851887474e-05,
"loss": 1.0278,
"step": 36500
},
{
"epoch": 26.086228398360948,
"grad_norm": 11.849235534667969,
"learning_rate": 1.3250781437845635e-05,
"loss": 0.9957,
"step": 36600
},
{
"epoch": 26.157491537502228,
"grad_norm": 15.007696151733398,
"learning_rate": 1.3010339023803799e-05,
"loss": 1.0203,
"step": 36700
},
{
"epoch": 26.228754676643508,
"grad_norm": 15.471383094787598,
"learning_rate": 1.2769896609761963e-05,
"loss": 1.0315,
"step": 36800
},
{
"epoch": 26.228754676643508,
"eval_loss": 0.09275855869054794,
"eval_runtime": 147.5238,
"eval_samples_per_second": 3.389,
"eval_steps_per_second": 0.427,
"eval_wer": 0.20218096080165046,
"step": 36800
},
{
"epoch": 26.300017815784784,
"grad_norm": 17.358564376831055,
"learning_rate": 1.2529454195720125e-05,
"loss": 1.0202,
"step": 36900
},
{
"epoch": 26.371280954926064,
"grad_norm": 15.642189025878906,
"learning_rate": 1.228901178167829e-05,
"loss": 1.0203,
"step": 37000
},
{
"epoch": 26.442544094067344,
"grad_norm": 10.802166938781738,
"learning_rate": 1.2048569367636452e-05,
"loss": 1.0047,
"step": 37100
},
{
"epoch": 26.513807233208624,
"grad_norm": 15.042948722839355,
"learning_rate": 1.1808126953594614e-05,
"loss": 1.0131,
"step": 37200
},
{
"epoch": 26.513807233208624,
"eval_loss": 0.0926903486251831,
"eval_runtime": 147.4732,
"eval_samples_per_second": 3.39,
"eval_steps_per_second": 0.427,
"eval_wer": 0.20350722074860006,
"step": 37200
},
{
"epoch": 26.585070372349904,
"grad_norm": 17.35483169555664,
"learning_rate": 1.1567684539552778e-05,
"loss": 1.0161,
"step": 37300
},
{
"epoch": 26.65633351149118,
"grad_norm": 15.87969970703125,
"learning_rate": 1.132724212551094e-05,
"loss": 1.0339,
"step": 37400
},
{
"epoch": 26.72759665063246,
"grad_norm": 12.541560173034668,
"learning_rate": 1.1086799711469103e-05,
"loss": 0.9692,
"step": 37500
},
{
"epoch": 26.79885978977374,
"grad_norm": 13.724421501159668,
"learning_rate": 1.0848761721567685e-05,
"loss": 0.9659,
"step": 37600
},
{
"epoch": 26.79885978977374,
"eval_loss": 0.09249202907085419,
"eval_runtime": 146.0218,
"eval_samples_per_second": 3.424,
"eval_steps_per_second": 0.431,
"eval_wer": 0.19997052755673445,
"step": 37600
},
{
"epoch": 26.87012292891502,
"grad_norm": 9.979654312133789,
"learning_rate": 1.0608319307525847e-05,
"loss": 1.0082,
"step": 37700
},
{
"epoch": 26.941386068056296,
"grad_norm": 21.3048095703125,
"learning_rate": 1.0367876893484011e-05,
"loss": 0.9767,
"step": 37800
},
{
"epoch": 27.012114733654016,
"grad_norm": 13.462841987609863,
"learning_rate": 1.0127434479442175e-05,
"loss": 1.0252,
"step": 37900
},
{
"epoch": 27.083377872795296,
"grad_norm": 10.681748390197754,
"learning_rate": 9.886992065400337e-06,
"loss": 1.0056,
"step": 38000
},
{
"epoch": 27.083377872795296,
"eval_loss": 0.09220070391893387,
"eval_runtime": 148.5209,
"eval_samples_per_second": 3.367,
"eval_steps_per_second": 0.424,
"eval_wer": 0.19923371647509577,
"step": 38000
},
{
"epoch": 27.154641011936576,
"grad_norm": 13.021913528442383,
"learning_rate": 9.6465496513585e-06,
"loss": 0.9757,
"step": 38100
},
{
"epoch": 27.225904151077856,
"grad_norm": 16.946664810180664,
"learning_rate": 9.406107237316664e-06,
"loss": 1.0595,
"step": 38200
},
{
"epoch": 27.297167290219136,
"grad_norm": 13.715904235839844,
"learning_rate": 9.165664823274826e-06,
"loss": 0.9515,
"step": 38300
},
{
"epoch": 27.368430429360412,
"grad_norm": 12.851482391357422,
"learning_rate": 8.925222409232988e-06,
"loss": 1.007,
"step": 38400
},
{
"epoch": 27.368430429360412,
"eval_loss": 0.09224120527505875,
"eval_runtime": 146.7128,
"eval_samples_per_second": 3.408,
"eval_steps_per_second": 0.429,
"eval_wer": 0.19967580312407898,
"step": 38400
},
{
"epoch": 27.439693568501692,
"grad_norm": 14.637819290161133,
"learning_rate": 8.684779995191152e-06,
"loss": 0.9863,
"step": 38500
},
{
"epoch": 27.510956707642972,
"grad_norm": 11.744438171386719,
"learning_rate": 8.444337581149316e-06,
"loss": 0.9084,
"step": 38600
},
{
"epoch": 27.582219846784252,
"grad_norm": 12.19479751586914,
"learning_rate": 8.203895167107479e-06,
"loss": 1.0345,
"step": 38700
},
{
"epoch": 27.65348298592553,
"grad_norm": 18.433185577392578,
"learning_rate": 7.963452753065641e-06,
"loss": 0.9602,
"step": 38800
},
{
"epoch": 27.65348298592553,
"eval_loss": 0.09226806461811066,
"eval_runtime": 147.8247,
"eval_samples_per_second": 3.382,
"eval_steps_per_second": 0.426,
"eval_wer": 0.20173887415266725,
"step": 38800
},
{
"epoch": 27.724746125066808,
"grad_norm": 9.473343849182129,
"learning_rate": 7.723010339023805e-06,
"loss": 0.9717,
"step": 38900
},
{
"epoch": 27.796009264208088,
"grad_norm": 12.731324195861816,
"learning_rate": 7.482567924981967e-06,
"loss": 0.9844,
"step": 39000
},
{
"epoch": 27.867272403349368,
"grad_norm": 23.304418563842773,
"learning_rate": 7.24212551094013e-06,
"loss": 0.9744,
"step": 39100
},
{
"epoch": 27.938535542490648,
"grad_norm": 11.320829391479492,
"learning_rate": 7.001683096898293e-06,
"loss": 0.9353,
"step": 39200
},
{
"epoch": 27.938535542490648,
"eval_loss": 0.09226758033037186,
"eval_runtime": 146.6121,
"eval_samples_per_second": 3.41,
"eval_steps_per_second": 0.43,
"eval_wer": 0.1989389920424403,
"step": 39200
},
{
"epoch": 28.009264208088368,
"grad_norm": 12.030290603637695,
"learning_rate": 6.761240682856457e-06,
"loss": 1.0218,
"step": 39300
},
{
"epoch": 28.080527347229644,
"grad_norm": 12.404072761535645,
"learning_rate": 6.520798268814619e-06,
"loss": 0.9249,
"step": 39400
},
{
"epoch": 28.151790486370924,
"grad_norm": 9.461714744567871,
"learning_rate": 6.280355854772782e-06,
"loss": 0.9299,
"step": 39500
},
{
"epoch": 28.223053625512204,
"grad_norm": 14.217653274536133,
"learning_rate": 6.039913440730945e-06,
"loss": 0.951,
"step": 39600
},
{
"epoch": 28.223053625512204,
"eval_loss": 0.09204767644405365,
"eval_runtime": 147.0572,
"eval_samples_per_second": 3.4,
"eval_steps_per_second": 0.428,
"eval_wer": 0.19834954317712938,
"step": 39600
},
{
"epoch": 28.294316764653484,
"grad_norm": 12.441349029541016,
"learning_rate": 5.799471026689108e-06,
"loss": 0.9283,
"step": 39700
},
{
"epoch": 28.365579903794764,
"grad_norm": 16.366159439086914,
"learning_rate": 5.56143303678769e-06,
"loss": 1.0031,
"step": 39800
},
{
"epoch": 28.43684304293604,
"grad_norm": 11.885308265686035,
"learning_rate": 5.320990622745853e-06,
"loss": 0.964,
"step": 39900
},
{
"epoch": 28.50810618207732,
"grad_norm": 9.691365242004395,
"learning_rate": 5.080548208704015e-06,
"loss": 0.9675,
"step": 40000
},
{
"epoch": 28.50810618207732,
"eval_loss": 0.09217335283756256,
"eval_runtime": 145.8792,
"eval_samples_per_second": 3.427,
"eval_steps_per_second": 0.432,
"eval_wer": 0.19687592101385204,
"step": 40000
},
{
"epoch": 28.5793693212186,
"grad_norm": 11.783367156982422,
"learning_rate": 4.840105794662179e-06,
"loss": 0.9698,
"step": 40100
},
{
"epoch": 28.65063246035988,
"grad_norm": 12.037036895751953,
"learning_rate": 4.5996633806203414e-06,
"loss": 1.0061,
"step": 40200
},
{
"epoch": 28.72189559950116,
"grad_norm": 9.481605529785156,
"learning_rate": 4.359220966578505e-06,
"loss": 0.9748,
"step": 40300
},
{
"epoch": 28.793158738642436,
"grad_norm": 10.295605659484863,
"learning_rate": 4.118778552536667e-06,
"loss": 0.9398,
"step": 40400
},
{
"epoch": 28.793158738642436,
"eval_loss": 0.0921778753399849,
"eval_runtime": 145.4752,
"eval_samples_per_second": 3.437,
"eval_steps_per_second": 0.433,
"eval_wer": 0.19982316534040673,
"step": 40400
},
{
"epoch": 28.864421877783716,
"grad_norm": 16.17976188659668,
"learning_rate": 3.878336138494831e-06,
"loss": 0.9462,
"step": 40500
},
{
"epoch": 28.935685016924996,
"grad_norm": 12.729133605957031,
"learning_rate": 3.637893724452994e-06,
"loss": 0.9777,
"step": 40600
},
{
"epoch": 29.006413682522716,
"grad_norm": 15.731132507324219,
"learning_rate": 3.3974513104111564e-06,
"loss": 0.9736,
"step": 40700
},
{
"epoch": 29.077676821663996,
"grad_norm": 9.655588150024414,
"learning_rate": 3.15700889636932e-06,
"loss": 0.9533,
"step": 40800
},
{
"epoch": 29.077676821663996,
"eval_loss": 0.09240180999040604,
"eval_runtime": 145.3169,
"eval_samples_per_second": 3.441,
"eval_steps_per_second": 0.434,
"eval_wer": 0.1976127320954907,
"step": 40800
},
{
"epoch": 29.148939960805272,
"grad_norm": 12.47839641571045,
"learning_rate": 2.9165664823274828e-06,
"loss": 0.9257,
"step": 40900
},
{
"epoch": 29.220203099946552,
"grad_norm": 9.78573989868164,
"learning_rate": 2.676124068285646e-06,
"loss": 0.9376,
"step": 41000
},
{
"epoch": 29.291466239087832,
"grad_norm": 10.32386589050293,
"learning_rate": 2.4356816542438087e-06,
"loss": 0.9428,
"step": 41100
},
{
"epoch": 29.362729378229112,
"grad_norm": 12.5963773727417,
"learning_rate": 2.195239240201972e-06,
"loss": 0.9519,
"step": 41200
},
{
"epoch": 29.362729378229112,
"eval_loss": 0.09224098175764084,
"eval_runtime": 145.4228,
"eval_samples_per_second": 3.438,
"eval_steps_per_second": 0.433,
"eval_wer": 0.19687592101385204,
"step": 41200
},
{
"epoch": 29.433992517370392,
"grad_norm": 9.201010704040527,
"learning_rate": 1.954796826160135e-06,
"loss": 0.9134,
"step": 41300
},
{
"epoch": 29.505255656511668,
"grad_norm": 12.38512134552002,
"learning_rate": 1.7143544121182977e-06,
"loss": 0.9591,
"step": 41400
},
{
"epoch": 29.576518795652948,
"grad_norm": 12.809625625610352,
"learning_rate": 1.4739119980764607e-06,
"loss": 0.8946,
"step": 41500
},
{
"epoch": 29.647781934794228,
"grad_norm": 11.387064933776855,
"learning_rate": 1.2334695840346237e-06,
"loss": 0.9297,
"step": 41600
},
{
"epoch": 29.647781934794228,
"eval_loss": 0.09202717989683151,
"eval_runtime": 145.5856,
"eval_samples_per_second": 3.434,
"eval_steps_per_second": 0.433,
"eval_wer": 0.19820218096080164,
"step": 41600
},
{
"epoch": 29.719045073935508,
"grad_norm": 11.282709121704102,
"learning_rate": 9.930271699927866e-07,
"loss": 0.9995,
"step": 41700
},
{
"epoch": 29.790308213076784,
"grad_norm": 10.81017780303955,
"learning_rate": 7.525847559509498e-07,
"loss": 0.9876,
"step": 41800
},
{
"epoch": 29.861571352218064,
"grad_norm": 11.903881072998047,
"learning_rate": 5.121423419091128e-07,
"loss": 0.9524,
"step": 41900
},
{
"epoch": 29.932834491359344,
"grad_norm": 11.278658866882324,
"learning_rate": 2.7410435200769414e-07,
"loss": 0.9491,
"step": 42000
},
{
"epoch": 29.932834491359344,
"eval_loss": 0.09202782809734344,
"eval_runtime": 145.311,
"eval_samples_per_second": 3.441,
"eval_steps_per_second": 0.434,
"eval_wer": 0.19908635425876806,
"step": 42000
}
],
"logging_steps": 100,
"max_steps": 42090,
"num_input_tokens_seen": 0,
"num_train_epochs": 30,
"save_steps": 400,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 5.472222473355264e+19,
"train_batch_size": 8,
"trial_name": null,
"trial_params": null
}