medium-poleval-le3 / trainer_state.json
Aspik101's picture
Upload 8 files
ea875ba verified
{
"best_metric": 0.1262135922330097,
"best_model_checkpoint": "results3\\checkpoint-124000",
"epoch": 2.002992,
"eval_steps": 4000,
"global_step": 124000,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.0008,
"grad_norm": 12.33169937133789,
"learning_rate": 1.8800000000000002e-06,
"loss": 1.7644,
"step": 100
},
{
"epoch": 0.0016,
"grad_norm": 9.583354949951172,
"learning_rate": 3.88e-06,
"loss": 0.929,
"step": 200
},
{
"epoch": 0.0024,
"grad_norm": 12.06747817993164,
"learning_rate": 5.8800000000000005e-06,
"loss": 0.4652,
"step": 300
},
{
"epoch": 0.0032,
"grad_norm": 9.123275756835938,
"learning_rate": 7.88e-06,
"loss": 0.3351,
"step": 400
},
{
"epoch": 0.004,
"grad_norm": 15.981216430664062,
"learning_rate": 9.88e-06,
"loss": 0.4079,
"step": 500
},
{
"epoch": 0.0048,
"grad_norm": 20.674148559570312,
"learning_rate": 9.992449799196789e-06,
"loss": 0.3966,
"step": 600
},
{
"epoch": 0.0056,
"grad_norm": 11.142045974731445,
"learning_rate": 9.984417670682733e-06,
"loss": 0.3628,
"step": 700
},
{
"epoch": 0.0064,
"grad_norm": 6.716693878173828,
"learning_rate": 9.976385542168675e-06,
"loss": 0.3647,
"step": 800
},
{
"epoch": 0.0072,
"grad_norm": 15.127047538757324,
"learning_rate": 9.968353413654619e-06,
"loss": 0.4007,
"step": 900
},
{
"epoch": 0.008,
"grad_norm": 5.561180114746094,
"learning_rate": 9.960321285140563e-06,
"loss": 0.3748,
"step": 1000
},
{
"epoch": 0.0088,
"grad_norm": 11.087468147277832,
"learning_rate": 9.952289156626507e-06,
"loss": 0.3802,
"step": 1100
},
{
"epoch": 0.0096,
"grad_norm": 13.053077697753906,
"learning_rate": 9.94425702811245e-06,
"loss": 0.3555,
"step": 1200
},
{
"epoch": 0.0104,
"grad_norm": 8.72202205657959,
"learning_rate": 9.936224899598395e-06,
"loss": 0.3532,
"step": 1300
},
{
"epoch": 0.0112,
"grad_norm": 13.878268241882324,
"learning_rate": 9.928192771084338e-06,
"loss": 0.3571,
"step": 1400
},
{
"epoch": 0.012,
"grad_norm": 7.951385498046875,
"learning_rate": 9.920160642570282e-06,
"loss": 0.3683,
"step": 1500
},
{
"epoch": 0.0128,
"grad_norm": 7.835740566253662,
"learning_rate": 9.912128514056226e-06,
"loss": 0.3409,
"step": 1600
},
{
"epoch": 0.0136,
"grad_norm": 10.594538688659668,
"learning_rate": 9.904096385542169e-06,
"loss": 0.3455,
"step": 1700
},
{
"epoch": 0.0144,
"grad_norm": 8.736641883850098,
"learning_rate": 9.896064257028112e-06,
"loss": 0.3567,
"step": 1800
},
{
"epoch": 0.0152,
"grad_norm": 6.575601100921631,
"learning_rate": 9.888032128514056e-06,
"loss": 0.3606,
"step": 1900
},
{
"epoch": 0.016,
"grad_norm": 11.653617858886719,
"learning_rate": 9.88e-06,
"loss": 0.3683,
"step": 2000
},
{
"epoch": 0.0168,
"grad_norm": 12.467767715454102,
"learning_rate": 9.871967871485944e-06,
"loss": 0.3723,
"step": 2100
},
{
"epoch": 0.0176,
"grad_norm": 11.205801963806152,
"learning_rate": 9.863935742971888e-06,
"loss": 0.3561,
"step": 2200
},
{
"epoch": 0.0184,
"grad_norm": 10.407919883728027,
"learning_rate": 9.855903614457832e-06,
"loss": 0.3414,
"step": 2300
},
{
"epoch": 0.0192,
"grad_norm": 6.497184753417969,
"learning_rate": 9.847871485943776e-06,
"loss": 0.3439,
"step": 2400
},
{
"epoch": 0.02,
"grad_norm": 10.192763328552246,
"learning_rate": 9.83983935742972e-06,
"loss": 0.3334,
"step": 2500
},
{
"epoch": 0.0208,
"grad_norm": 10.233382225036621,
"learning_rate": 9.831807228915664e-06,
"loss": 0.3715,
"step": 2600
},
{
"epoch": 0.0216,
"grad_norm": 10.373730659484863,
"learning_rate": 9.823775100401608e-06,
"loss": 0.3557,
"step": 2700
},
{
"epoch": 0.0224,
"grad_norm": 9.90769100189209,
"learning_rate": 9.81574297188755e-06,
"loss": 0.345,
"step": 2800
},
{
"epoch": 0.0232,
"grad_norm": 12.89505386352539,
"learning_rate": 9.807710843373494e-06,
"loss": 0.3466,
"step": 2900
},
{
"epoch": 0.024,
"grad_norm": 11.755387306213379,
"learning_rate": 9.799678714859438e-06,
"loss": 0.3518,
"step": 3000
},
{
"epoch": 0.0248,
"grad_norm": 23.758481979370117,
"learning_rate": 9.791646586345382e-06,
"loss": 0.3367,
"step": 3100
},
{
"epoch": 0.0256,
"grad_norm": 8.041315078735352,
"learning_rate": 9.783614457831326e-06,
"loss": 0.3385,
"step": 3200
},
{
"epoch": 0.0264,
"grad_norm": 6.949316501617432,
"learning_rate": 9.77558232931727e-06,
"loss": 0.3557,
"step": 3300
},
{
"epoch": 0.0272,
"grad_norm": 20.79637908935547,
"learning_rate": 9.767550200803213e-06,
"loss": 0.3497,
"step": 3400
},
{
"epoch": 0.028,
"grad_norm": 10.800033569335938,
"learning_rate": 9.759518072289157e-06,
"loss": 0.351,
"step": 3500
},
{
"epoch": 0.0288,
"grad_norm": 7.772594451904297,
"learning_rate": 9.751485943775101e-06,
"loss": 0.3272,
"step": 3600
},
{
"epoch": 0.0296,
"grad_norm": 7.19946813583374,
"learning_rate": 9.743453815261045e-06,
"loss": 0.3216,
"step": 3700
},
{
"epoch": 0.0304,
"grad_norm": 11.165406227111816,
"learning_rate": 9.735421686746989e-06,
"loss": 0.3305,
"step": 3800
},
{
"epoch": 0.0312,
"grad_norm": 12.210046768188477,
"learning_rate": 9.727389558232933e-06,
"loss": 0.3524,
"step": 3900
},
{
"epoch": 0.032,
"grad_norm": 6.451345443725586,
"learning_rate": 9.719357429718877e-06,
"loss": 0.3394,
"step": 4000
},
{
"epoch": 0.032,
"eval_test1_cer": 0.09286074389424154,
"eval_test1_cer_norm": 0.07516507791677672,
"eval_test1_loss": 0.2566915452480316,
"eval_test1_runtime": 6436.6351,
"eval_test1_samples_per_second": 0.388,
"eval_test1_steps_per_second": 0.097,
"eval_test1_wer": 0.2368582174407417,
"eval_test1_wer_norm": 0.17320786697448787,
"step": 4000
},
{
"epoch": 0.032,
"eval_test2_cer": 0.2131304364057192,
"eval_test2_cer_norm": 0.18118124418964984,
"eval_test2_loss": 0.4192444682121277,
"eval_test2_runtime": 3419.6713,
"eval_test2_samples_per_second": 0.731,
"eval_test2_steps_per_second": 0.183,
"eval_test2_wer": 0.42229343099107347,
"eval_test2_wer_norm": 0.35425739170295667,
"step": 4000
},
{
"epoch": 0.0328,
"grad_norm": 43.34160232543945,
"learning_rate": 9.711325301204821e-06,
"loss": 0.3384,
"step": 4100
},
{
"epoch": 0.0336,
"grad_norm": 7.907329559326172,
"learning_rate": 9.703293172690765e-06,
"loss": 0.3674,
"step": 4200
},
{
"epoch": 0.0344,
"grad_norm": 6.71475076675415,
"learning_rate": 9.695261044176709e-06,
"loss": 0.3364,
"step": 4300
},
{
"epoch": 0.0352,
"grad_norm": 9.054850578308105,
"learning_rate": 9.687228915662651e-06,
"loss": 0.3417,
"step": 4400
},
{
"epoch": 0.036,
"grad_norm": 14.793953895568848,
"learning_rate": 9.679196787148595e-06,
"loss": 0.3286,
"step": 4500
},
{
"epoch": 0.0368,
"grad_norm": 5.203219413757324,
"learning_rate": 9.671164658634539e-06,
"loss": 0.3684,
"step": 4600
},
{
"epoch": 0.0376,
"grad_norm": 12.02910327911377,
"learning_rate": 9.663132530120483e-06,
"loss": 0.3393,
"step": 4700
},
{
"epoch": 0.0384,
"grad_norm": 10.032449722290039,
"learning_rate": 9.655100401606427e-06,
"loss": 0.3503,
"step": 4800
},
{
"epoch": 0.0392,
"grad_norm": 9.57917594909668,
"learning_rate": 9.64706827309237e-06,
"loss": 0.3239,
"step": 4900
},
{
"epoch": 0.04,
"grad_norm": 9.18687629699707,
"learning_rate": 9.639036144578314e-06,
"loss": 0.3407,
"step": 5000
},
{
"epoch": 0.0408,
"grad_norm": 9.08846378326416,
"learning_rate": 9.631004016064258e-06,
"loss": 0.3478,
"step": 5100
},
{
"epoch": 0.0416,
"grad_norm": 6.322196006774902,
"learning_rate": 9.622971887550202e-06,
"loss": 0.3294,
"step": 5200
},
{
"epoch": 0.0424,
"grad_norm": 11.071808815002441,
"learning_rate": 9.614939759036145e-06,
"loss": 0.3533,
"step": 5300
},
{
"epoch": 0.0432,
"grad_norm": 8.492830276489258,
"learning_rate": 9.606907630522088e-06,
"loss": 0.3297,
"step": 5400
},
{
"epoch": 0.044,
"grad_norm": 13.28835678100586,
"learning_rate": 9.598875502008032e-06,
"loss": 0.3228,
"step": 5500
},
{
"epoch": 0.0448,
"grad_norm": 9.030045509338379,
"learning_rate": 9.590843373493976e-06,
"loss": 0.3228,
"step": 5600
},
{
"epoch": 0.0456,
"grad_norm": 12.364481925964355,
"learning_rate": 9.58281124497992e-06,
"loss": 0.3406,
"step": 5700
},
{
"epoch": 0.0464,
"grad_norm": 6.908151149749756,
"learning_rate": 9.574779116465864e-06,
"loss": 0.3365,
"step": 5800
},
{
"epoch": 0.0472,
"grad_norm": 10.87695598602295,
"learning_rate": 9.566746987951808e-06,
"loss": 0.3332,
"step": 5900
},
{
"epoch": 0.048,
"grad_norm": 9.280797004699707,
"learning_rate": 9.558714859437752e-06,
"loss": 0.3414,
"step": 6000
},
{
"epoch": 0.0488,
"grad_norm": 13.637160301208496,
"learning_rate": 9.550682730923696e-06,
"loss": 0.3275,
"step": 6100
},
{
"epoch": 0.0496,
"grad_norm": 10.094959259033203,
"learning_rate": 9.542650602409638e-06,
"loss": 0.3152,
"step": 6200
},
{
"epoch": 0.0504,
"grad_norm": 6.261775016784668,
"learning_rate": 9.534618473895582e-06,
"loss": 0.3236,
"step": 6300
},
{
"epoch": 0.0512,
"grad_norm": 7.221704006195068,
"learning_rate": 9.526586345381526e-06,
"loss": 0.3461,
"step": 6400
},
{
"epoch": 0.052,
"grad_norm": 10.923733711242676,
"learning_rate": 9.51855421686747e-06,
"loss": 0.3249,
"step": 6500
},
{
"epoch": 0.0528,
"grad_norm": 5.432595729827881,
"learning_rate": 9.510522088353414e-06,
"loss": 0.3556,
"step": 6600
},
{
"epoch": 0.0536,
"grad_norm": 15.021200180053711,
"learning_rate": 9.502489959839358e-06,
"loss": 0.3208,
"step": 6700
},
{
"epoch": 0.0544,
"grad_norm": 6.641680717468262,
"learning_rate": 9.494457831325302e-06,
"loss": 0.3565,
"step": 6800
},
{
"epoch": 0.0552,
"grad_norm": 15.715243339538574,
"learning_rate": 9.486425702811246e-06,
"loss": 0.3123,
"step": 6900
},
{
"epoch": 0.056,
"grad_norm": 9.304129600524902,
"learning_rate": 9.47839357429719e-06,
"loss": 0.312,
"step": 7000
},
{
"epoch": 0.0568,
"grad_norm": 10.780037879943848,
"learning_rate": 9.470361445783133e-06,
"loss": 0.3445,
"step": 7100
},
{
"epoch": 0.0576,
"grad_norm": 11.77072811126709,
"learning_rate": 9.462329317269077e-06,
"loss": 0.3293,
"step": 7200
},
{
"epoch": 0.0584,
"grad_norm": 11.29255485534668,
"learning_rate": 9.454297188755021e-06,
"loss": 0.3045,
"step": 7300
},
{
"epoch": 0.0592,
"grad_norm": 9.00106143951416,
"learning_rate": 9.446265060240965e-06,
"loss": 0.3132,
"step": 7400
},
{
"epoch": 0.06,
"grad_norm": 8.445959091186523,
"learning_rate": 9.438232931726909e-06,
"loss": 0.3327,
"step": 7500
},
{
"epoch": 0.0608,
"grad_norm": 13.996682167053223,
"learning_rate": 9.430200803212853e-06,
"loss": 0.3496,
"step": 7600
},
{
"epoch": 0.0616,
"grad_norm": 6.321784019470215,
"learning_rate": 9.422168674698797e-06,
"loss": 0.2928,
"step": 7700
},
{
"epoch": 0.0624,
"grad_norm": 9.260029792785645,
"learning_rate": 9.414136546184741e-06,
"loss": 0.2978,
"step": 7800
},
{
"epoch": 0.0632,
"grad_norm": 14.37282657623291,
"learning_rate": 9.406104417670685e-06,
"loss": 0.3177,
"step": 7900
},
{
"epoch": 0.064,
"grad_norm": 13.508281707763672,
"learning_rate": 9.398072289156627e-06,
"loss": 0.3357,
"step": 8000
},
{
"epoch": 0.064,
"eval_test1_cer": 0.07946822018074538,
"eval_test1_cer_norm": 0.05569188657046126,
"eval_test1_loss": 0.2466106414794922,
"eval_test1_runtime": 3431.7176,
"eval_test1_samples_per_second": 0.728,
"eval_test1_steps_per_second": 0.182,
"eval_test1_wer": 0.21889851016064607,
"eval_test1_wer_norm": 0.15733948975715245,
"step": 8000
},
{
"epoch": 0.064,
"eval_test2_cer": 0.21952346287378205,
"eval_test2_cer_norm": 0.16622927641772545,
"eval_test2_loss": 0.410643607378006,
"eval_test2_runtime": 3658.9711,
"eval_test2_samples_per_second": 0.683,
"eval_test2_steps_per_second": 0.171,
"eval_test2_wer": 0.39210917830167086,
"eval_test2_wer_norm": 0.32348727939491173,
"step": 8000
},
{
"epoch": 0.0648,
"grad_norm": 7.349660873413086,
"learning_rate": 9.390040160642571e-06,
"loss": 0.3311,
"step": 8100
},
{
"epoch": 0.0656,
"grad_norm": 6.557274341583252,
"learning_rate": 9.382008032128515e-06,
"loss": 0.318,
"step": 8200
},
{
"epoch": 0.0664,
"grad_norm": 20.285615921020508,
"learning_rate": 9.373975903614459e-06,
"loss": 0.3116,
"step": 8300
},
{
"epoch": 0.0672,
"grad_norm": 7.684230327606201,
"learning_rate": 9.365943775100403e-06,
"loss": 0.3232,
"step": 8400
},
{
"epoch": 0.068,
"grad_norm": 8.374768257141113,
"learning_rate": 9.357911646586347e-06,
"loss": 0.3328,
"step": 8500
},
{
"epoch": 0.0688,
"grad_norm": 6.924304008483887,
"learning_rate": 9.34987951807229e-06,
"loss": 0.2923,
"step": 8600
},
{
"epoch": 0.0696,
"grad_norm": 5.943380832672119,
"learning_rate": 9.341847389558234e-06,
"loss": 0.3167,
"step": 8700
},
{
"epoch": 0.0704,
"grad_norm": 24.188907623291016,
"learning_rate": 9.333815261044178e-06,
"loss": 0.3521,
"step": 8800
},
{
"epoch": 0.0712,
"grad_norm": 13.702215194702148,
"learning_rate": 9.325783132530122e-06,
"loss": 0.2937,
"step": 8900
},
{
"epoch": 0.072,
"grad_norm": 10.463766098022461,
"learning_rate": 9.317751004016065e-06,
"loss": 0.3317,
"step": 9000
},
{
"epoch": 0.0728,
"grad_norm": 5.035882472991943,
"learning_rate": 9.309718875502008e-06,
"loss": 0.3048,
"step": 9100
},
{
"epoch": 0.0736,
"grad_norm": 11.715450286865234,
"learning_rate": 9.301686746987952e-06,
"loss": 0.3287,
"step": 9200
},
{
"epoch": 0.0744,
"grad_norm": 13.449250221252441,
"learning_rate": 9.293654618473896e-06,
"loss": 0.3345,
"step": 9300
},
{
"epoch": 0.0752,
"grad_norm": 8.16373348236084,
"learning_rate": 9.28562248995984e-06,
"loss": 0.3233,
"step": 9400
},
{
"epoch": 0.076,
"grad_norm": 6.236880302429199,
"learning_rate": 9.277590361445784e-06,
"loss": 0.3131,
"step": 9500
},
{
"epoch": 0.0768,
"grad_norm": 12.164971351623535,
"learning_rate": 9.269558232931728e-06,
"loss": 0.3281,
"step": 9600
},
{
"epoch": 0.0776,
"grad_norm": 15.774728775024414,
"learning_rate": 9.261526104417672e-06,
"loss": 0.315,
"step": 9700
},
{
"epoch": 0.0784,
"grad_norm": 16.81951332092285,
"learning_rate": 9.253493975903616e-06,
"loss": 0.3491,
"step": 9800
},
{
"epoch": 0.0792,
"grad_norm": 35.960697174072266,
"learning_rate": 9.245461847389558e-06,
"loss": 0.3258,
"step": 9900
},
{
"epoch": 0.08,
"grad_norm": 9.389575958251953,
"learning_rate": 9.237429718875502e-06,
"loss": 0.3082,
"step": 10000
},
{
"epoch": 0.0808,
"grad_norm": 5.77083683013916,
"learning_rate": 9.229477911646586e-06,
"loss": 0.316,
"step": 10100
},
{
"epoch": 0.0816,
"grad_norm": 7.730663776397705,
"learning_rate": 9.22144578313253e-06,
"loss": 0.3538,
"step": 10200
},
{
"epoch": 0.0824,
"grad_norm": 2.672037363052368,
"learning_rate": 9.213413654618474e-06,
"loss": 0.3089,
"step": 10300
},
{
"epoch": 0.0832,
"grad_norm": 10.987016677856445,
"learning_rate": 9.205381526104418e-06,
"loss": 0.3032,
"step": 10400
},
{
"epoch": 0.084,
"grad_norm": 5.871443271636963,
"learning_rate": 9.197349397590362e-06,
"loss": 0.3423,
"step": 10500
},
{
"epoch": 0.0848,
"grad_norm": 8.393383026123047,
"learning_rate": 9.189317269076306e-06,
"loss": 0.2872,
"step": 10600
},
{
"epoch": 0.0856,
"grad_norm": 8.698946952819824,
"learning_rate": 9.18128514056225e-06,
"loss": 0.3028,
"step": 10700
},
{
"epoch": 0.0864,
"grad_norm": 2.9550344944000244,
"learning_rate": 9.173253012048194e-06,
"loss": 0.3262,
"step": 10800
},
{
"epoch": 0.0872,
"grad_norm": 6.455613136291504,
"learning_rate": 9.165220883534138e-06,
"loss": 0.3273,
"step": 10900
},
{
"epoch": 0.088,
"grad_norm": 11.550117492675781,
"learning_rate": 9.157269076305222e-06,
"loss": 0.3307,
"step": 11000
},
{
"epoch": 0.0888,
"grad_norm": 9.657747268676758,
"learning_rate": 9.149236947791166e-06,
"loss": 0.3227,
"step": 11100
},
{
"epoch": 0.0896,
"grad_norm": 9.803390502929688,
"learning_rate": 9.14120481927711e-06,
"loss": 0.3102,
"step": 11200
},
{
"epoch": 0.0904,
"grad_norm": 12.699715614318848,
"learning_rate": 9.133172690763052e-06,
"loss": 0.3353,
"step": 11300
},
{
"epoch": 0.0912,
"grad_norm": 9.973093032836914,
"learning_rate": 9.125140562248996e-06,
"loss": 0.3053,
"step": 11400
},
{
"epoch": 0.092,
"grad_norm": 12.822247505187988,
"learning_rate": 9.11710843373494e-06,
"loss": 0.3088,
"step": 11500
},
{
"epoch": 0.0928,
"grad_norm": 13.785049438476562,
"learning_rate": 9.109076305220884e-06,
"loss": 0.3004,
"step": 11600
},
{
"epoch": 0.0936,
"grad_norm": 11.756546020507812,
"learning_rate": 9.101044176706828e-06,
"loss": 0.304,
"step": 11700
},
{
"epoch": 0.0944,
"grad_norm": 12.297760009765625,
"learning_rate": 9.093012048192772e-06,
"loss": 0.3029,
"step": 11800
},
{
"epoch": 0.0952,
"grad_norm": 7.838765621185303,
"learning_rate": 9.084979919678716e-06,
"loss": 0.3046,
"step": 11900
},
{
"epoch": 0.096,
"grad_norm": 6.533823013305664,
"learning_rate": 9.07694779116466e-06,
"loss": 0.3065,
"step": 12000
},
{
"epoch": 0.096,
"eval_test1_cer": 0.10149656434386436,
"eval_test1_cer_norm": 0.06880207457920139,
"eval_test1_loss": 0.23942527174949646,
"eval_test1_runtime": 3496.8354,
"eval_test1_samples_per_second": 0.715,
"eval_test1_steps_per_second": 0.179,
"eval_test1_wer": 0.23831598588880143,
"eval_test1_wer_norm": 0.1713375610041205,
"step": 12000
},
{
"epoch": 0.096,
"eval_test2_cer": 0.2408444394669056,
"eval_test2_cer_norm": 0.178324488689185,
"eval_test2_loss": 0.3968297839164734,
"eval_test2_runtime": 3795.2628,
"eval_test2_samples_per_second": 0.659,
"eval_test2_steps_per_second": 0.165,
"eval_test2_wer": 0.4381437399862669,
"eval_test2_wer_norm": 0.36356864542745815,
"step": 12000
},
{
"epoch": 0.0968,
"grad_norm": 6.5714826583862305,
"learning_rate": 9.068915662650603e-06,
"loss": 0.3177,
"step": 12100
},
{
"epoch": 0.0976,
"grad_norm": 10.972187995910645,
"learning_rate": 9.060883534136546e-06,
"loss": 0.3032,
"step": 12200
},
{
"epoch": 0.0984,
"grad_norm": 12.001550674438477,
"learning_rate": 9.05285140562249e-06,
"loss": 0.3009,
"step": 12300
},
{
"epoch": 0.0992,
"grad_norm": 10.76689338684082,
"learning_rate": 9.044819277108433e-06,
"loss": 0.304,
"step": 12400
},
{
"epoch": 0.1,
"grad_norm": 6.645873069763184,
"learning_rate": 9.036787148594377e-06,
"loss": 0.2801,
"step": 12500
},
{
"epoch": 0.1008,
"grad_norm": 7.068800926208496,
"learning_rate": 9.028755020080321e-06,
"loss": 0.3198,
"step": 12600
},
{
"epoch": 0.1016,
"grad_norm": 7.24527645111084,
"learning_rate": 9.020722891566265e-06,
"loss": 0.3236,
"step": 12700
},
{
"epoch": 0.1024,
"grad_norm": 8.540508270263672,
"learning_rate": 9.01269076305221e-06,
"loss": 0.3132,
"step": 12800
},
{
"epoch": 0.1032,
"grad_norm": 10.710851669311523,
"learning_rate": 9.004658634538153e-06,
"loss": 0.3034,
"step": 12900
},
{
"epoch": 0.104,
"grad_norm": 7.53853702545166,
"learning_rate": 8.996626506024097e-06,
"loss": 0.2957,
"step": 13000
},
{
"epoch": 0.1048,
"grad_norm": 8.962101936340332,
"learning_rate": 8.988594377510041e-06,
"loss": 0.312,
"step": 13100
},
{
"epoch": 0.1056,
"grad_norm": 7.6881561279296875,
"learning_rate": 8.980642570281125e-06,
"loss": 0.3304,
"step": 13200
},
{
"epoch": 0.1064,
"grad_norm": 3.7053704261779785,
"learning_rate": 8.97261044176707e-06,
"loss": 0.2913,
"step": 13300
},
{
"epoch": 0.1072,
"grad_norm": 5.796387195587158,
"learning_rate": 8.964578313253013e-06,
"loss": 0.3162,
"step": 13400
},
{
"epoch": 0.108,
"grad_norm": 14.237732887268066,
"learning_rate": 8.956546184738957e-06,
"loss": 0.3096,
"step": 13500
},
{
"epoch": 0.1088,
"grad_norm": 9.336588859558105,
"learning_rate": 8.948514056224901e-06,
"loss": 0.3018,
"step": 13600
},
{
"epoch": 0.1096,
"grad_norm": 16.10430335998535,
"learning_rate": 8.940481927710845e-06,
"loss": 0.3149,
"step": 13700
},
{
"epoch": 0.1104,
"grad_norm": 70.292236328125,
"learning_rate": 8.932449799196787e-06,
"loss": 0.2721,
"step": 13800
},
{
"epoch": 0.1112,
"grad_norm": 5.2695770263671875,
"learning_rate": 8.924417670682731e-06,
"loss": 0.294,
"step": 13900
},
{
"epoch": 0.112,
"grad_norm": 7.514453411102295,
"learning_rate": 8.916385542168675e-06,
"loss": 0.3363,
"step": 14000
},
{
"epoch": 0.1128,
"grad_norm": 6.603350639343262,
"learning_rate": 8.908353413654619e-06,
"loss": 0.3214,
"step": 14100
},
{
"epoch": 0.1136,
"grad_norm": 12.648730278015137,
"learning_rate": 8.900321285140563e-06,
"loss": 0.313,
"step": 14200
},
{
"epoch": 0.1144,
"grad_norm": 19.52114486694336,
"learning_rate": 8.892289156626507e-06,
"loss": 0.2778,
"step": 14300
},
{
"epoch": 0.1152,
"grad_norm": 9.065464973449707,
"learning_rate": 8.88425702811245e-06,
"loss": 0.3182,
"step": 14400
},
{
"epoch": 0.116,
"grad_norm": 6.192761421203613,
"learning_rate": 8.876224899598395e-06,
"loss": 0.3268,
"step": 14500
},
{
"epoch": 0.1168,
"grad_norm": 6.336472034454346,
"learning_rate": 8.868192771084338e-06,
"loss": 0.3125,
"step": 14600
},
{
"epoch": 0.1176,
"grad_norm": 7.2869343757629395,
"learning_rate": 8.86016064257028e-06,
"loss": 0.2938,
"step": 14700
},
{
"epoch": 0.1184,
"grad_norm": 8.914804458618164,
"learning_rate": 8.852128514056225e-06,
"loss": 0.2831,
"step": 14800
},
{
"epoch": 0.1192,
"grad_norm": 8.745502471923828,
"learning_rate": 8.844096385542168e-06,
"loss": 0.306,
"step": 14900
},
{
"epoch": 0.12,
"grad_norm": 8.596148490905762,
"learning_rate": 8.836064257028112e-06,
"loss": 0.2907,
"step": 15000
},
{
"epoch": 0.1208,
"grad_norm": 6.889795303344727,
"learning_rate": 8.828032128514056e-06,
"loss": 0.2806,
"step": 15100
},
{
"epoch": 0.1216,
"grad_norm": 9.58425521850586,
"learning_rate": 8.82e-06,
"loss": 0.2853,
"step": 15200
},
{
"epoch": 0.1224,
"grad_norm": 5.967700004577637,
"learning_rate": 8.811967871485944e-06,
"loss": 0.3109,
"step": 15300
},
{
"epoch": 0.1232,
"grad_norm": 10.997330665588379,
"learning_rate": 8.80401606425703e-06,
"loss": 0.3129,
"step": 15400
},
{
"epoch": 0.124,
"grad_norm": 8.48721981048584,
"learning_rate": 8.795983935742972e-06,
"loss": 0.3227,
"step": 15500
},
{
"epoch": 0.1248,
"grad_norm": 9.547022819519043,
"learning_rate": 8.787951807228916e-06,
"loss": 0.282,
"step": 15600
},
{
"epoch": 0.1256,
"grad_norm": 12.319640159606934,
"learning_rate": 8.77991967871486e-06,
"loss": 0.3019,
"step": 15700
},
{
"epoch": 0.1264,
"grad_norm": 9.99170970916748,
"learning_rate": 8.771887550200804e-06,
"loss": 0.2872,
"step": 15800
},
{
"epoch": 0.1272,
"grad_norm": 9.307220458984375,
"learning_rate": 8.763855421686748e-06,
"loss": 0.2955,
"step": 15900
},
{
"epoch": 0.128,
"grad_norm": 6.70652961730957,
"learning_rate": 8.755823293172692e-06,
"loss": 0.2813,
"step": 16000
},
{
"epoch": 0.128,
"eval_test1_cer": 0.11614478303084622,
"eval_test1_cer_norm": 0.08919730112613153,
"eval_test1_loss": 0.2319698929786682,
"eval_test1_runtime": 2494.875,
"eval_test1_samples_per_second": 1.002,
"eval_test1_steps_per_second": 0.251,
"eval_test1_wer": 0.2696580075220852,
"eval_test1_wer_norm": 0.2076331862415617,
"step": 16000
},
{
"epoch": 0.128,
"eval_test2_cer": 0.25404113935864414,
"eval_test2_cer_norm": 0.1936541292221878,
"eval_test2_loss": 0.38276800513267517,
"eval_test2_runtime": 2638.5507,
"eval_test2_samples_per_second": 0.947,
"eval_test2_steps_per_second": 0.237,
"eval_test2_wer": 0.46260585946440835,
"eval_test2_wer_norm": 0.3951409580563832,
"step": 16000
},
{
"epoch": 0.1288,
"grad_norm": 9.956653594970703,
"learning_rate": 8.747791164658636e-06,
"loss": 0.2785,
"step": 16100
},
{
"epoch": 0.1296,
"grad_norm": 10.152132987976074,
"learning_rate": 8.73975903614458e-06,
"loss": 0.2918,
"step": 16200
},
{
"epoch": 0.1304,
"grad_norm": 10.379541397094727,
"learning_rate": 8.731726907630524e-06,
"loss": 0.2879,
"step": 16300
},
{
"epoch": 0.1312,
"grad_norm": 5.655648708343506,
"learning_rate": 8.723694779116466e-06,
"loss": 0.2935,
"step": 16400
},
{
"epoch": 0.132,
"grad_norm": 9.877130508422852,
"learning_rate": 8.71566265060241e-06,
"loss": 0.3331,
"step": 16500
},
{
"epoch": 0.1328,
"grad_norm": 8.812068939208984,
"learning_rate": 8.707630522088354e-06,
"loss": 0.2902,
"step": 16600
},
{
"epoch": 0.1336,
"grad_norm": 8.972885131835938,
"learning_rate": 8.699598393574298e-06,
"loss": 0.3017,
"step": 16700
},
{
"epoch": 0.1344,
"grad_norm": 6.562330722808838,
"learning_rate": 8.691566265060242e-06,
"loss": 0.2866,
"step": 16800
},
{
"epoch": 0.1352,
"grad_norm": 10.704404830932617,
"learning_rate": 8.683534136546186e-06,
"loss": 0.3078,
"step": 16900
},
{
"epoch": 0.136,
"grad_norm": 9.468811988830566,
"learning_rate": 8.67550200803213e-06,
"loss": 0.2906,
"step": 17000
},
{
"epoch": 0.1368,
"grad_norm": 5.418923854827881,
"learning_rate": 8.667469879518073e-06,
"loss": 0.2918,
"step": 17100
},
{
"epoch": 0.1376,
"grad_norm": 10.071795463562012,
"learning_rate": 8.659437751004017e-06,
"loss": 0.3149,
"step": 17200
},
{
"epoch": 0.1384,
"grad_norm": 10.977407455444336,
"learning_rate": 8.65140562248996e-06,
"loss": 0.3091,
"step": 17300
},
{
"epoch": 0.1392,
"grad_norm": 7.745774745941162,
"learning_rate": 8.643373493975904e-06,
"loss": 0.2843,
"step": 17400
},
{
"epoch": 0.14,
"grad_norm": 5.645944595336914,
"learning_rate": 8.635341365461847e-06,
"loss": 0.2824,
"step": 17500
},
{
"epoch": 0.1408,
"grad_norm": 11.466683387756348,
"learning_rate": 8.627309236947791e-06,
"loss": 0.3228,
"step": 17600
},
{
"epoch": 0.1416,
"grad_norm": 7.580329418182373,
"learning_rate": 8.619277108433735e-06,
"loss": 0.2967,
"step": 17700
},
{
"epoch": 0.1424,
"grad_norm": 11.873896598815918,
"learning_rate": 8.61124497991968e-06,
"loss": 0.3073,
"step": 17800
},
{
"epoch": 0.1432,
"grad_norm": 6.72385835647583,
"learning_rate": 8.603212851405623e-06,
"loss": 0.3105,
"step": 17900
},
{
"epoch": 0.144,
"grad_norm": 11.72990608215332,
"learning_rate": 8.595180722891567e-06,
"loss": 0.3141,
"step": 18000
},
{
"epoch": 0.1448,
"grad_norm": 16.776378631591797,
"learning_rate": 8.587148594377511e-06,
"loss": 0.2964,
"step": 18100
},
{
"epoch": 0.1456,
"grad_norm": 9.208036422729492,
"learning_rate": 8.579116465863455e-06,
"loss": 0.2818,
"step": 18200
},
{
"epoch": 0.1464,
"grad_norm": 10.971381187438965,
"learning_rate": 8.571084337349397e-06,
"loss": 0.2694,
"step": 18300
},
{
"epoch": 0.1472,
"grad_norm": 10.638858795166016,
"learning_rate": 8.563052208835341e-06,
"loss": 0.2927,
"step": 18400
},
{
"epoch": 0.148,
"grad_norm": 6.251519203186035,
"learning_rate": 8.555020080321285e-06,
"loss": 0.2835,
"step": 18500
},
{
"epoch": 0.1488,
"grad_norm": 10.34622859954834,
"learning_rate": 8.546987951807229e-06,
"loss": 0.298,
"step": 18600
},
{
"epoch": 0.1496,
"grad_norm": 9.303108215332031,
"learning_rate": 8.538955823293173e-06,
"loss": 0.2856,
"step": 18700
},
{
"epoch": 0.1504,
"grad_norm": 8.3678617477417,
"learning_rate": 8.530923694779117e-06,
"loss": 0.2782,
"step": 18800
},
{
"epoch": 0.1512,
"grad_norm": 11.090498924255371,
"learning_rate": 8.52289156626506e-06,
"loss": 0.2863,
"step": 18900
},
{
"epoch": 0.152,
"grad_norm": 5.836663722991943,
"learning_rate": 8.514859437751005e-06,
"loss": 0.2693,
"step": 19000
},
{
"epoch": 0.1528,
"grad_norm": 6.694278717041016,
"learning_rate": 8.506827309236948e-06,
"loss": 0.2817,
"step": 19100
},
{
"epoch": 0.1536,
"grad_norm": 5.725604057312012,
"learning_rate": 8.498795180722892e-06,
"loss": 0.265,
"step": 19200
},
{
"epoch": 0.1544,
"grad_norm": 8.070332527160645,
"learning_rate": 8.490763052208836e-06,
"loss": 0.302,
"step": 19300
},
{
"epoch": 0.1552,
"grad_norm": 7.241143226623535,
"learning_rate": 8.48281124497992e-06,
"loss": 0.2898,
"step": 19400
},
{
"epoch": 0.156,
"grad_norm": 8.699810981750488,
"learning_rate": 8.474779116465865e-06,
"loss": 0.2868,
"step": 19500
},
{
"epoch": 0.1568,
"grad_norm": 4.879644870758057,
"learning_rate": 8.466746987951808e-06,
"loss": 0.2917,
"step": 19600
},
{
"epoch": 0.1576,
"grad_norm": 6.502651691436768,
"learning_rate": 8.458714859437752e-06,
"loss": 0.2888,
"step": 19700
},
{
"epoch": 0.1584,
"grad_norm": 5.7493791580200195,
"learning_rate": 8.450682730923695e-06,
"loss": 0.2936,
"step": 19800
},
{
"epoch": 0.1592,
"grad_norm": 10.4688720703125,
"learning_rate": 8.442650602409639e-06,
"loss": 0.2806,
"step": 19900
},
{
"epoch": 0.16,
"grad_norm": 7.386699676513672,
"learning_rate": 8.434618473895582e-06,
"loss": 0.3055,
"step": 20000
},
{
"epoch": 0.16,
"eval_test1_cer": 0.09210452610351781,
"eval_test1_cer_norm": 0.07215885898143924,
"eval_test1_loss": 0.22505834698677063,
"eval_test1_runtime": 2739.9634,
"eval_test1_samples_per_second": 0.912,
"eval_test1_steps_per_second": 0.228,
"eval_test1_wer": 0.23679990670281933,
"eval_test1_wer_norm": 0.17437680820596743,
"step": 20000
},
{
"epoch": 0.16,
"eval_test2_cer": 0.15537387538731473,
"eval_test2_cer_norm": 0.12383792996591261,
"eval_test2_loss": 0.3772575259208679,
"eval_test2_runtime": 3715.4173,
"eval_test2_samples_per_second": 0.673,
"eval_test2_steps_per_second": 0.168,
"eval_test2_wer": 0.31795033188372623,
"eval_test2_wer_norm": 0.24954159981663993,
"step": 20000
},
{
"epoch": 0.1608,
"grad_norm": 9.69093132019043,
"learning_rate": 8.426586345381526e-06,
"loss": 0.2777,
"step": 20100
},
{
"epoch": 0.1616,
"grad_norm": 8.120789527893066,
"learning_rate": 8.41855421686747e-06,
"loss": 0.2859,
"step": 20200
},
{
"epoch": 0.1624,
"grad_norm": 7.084224700927734,
"learning_rate": 8.410522088353414e-06,
"loss": 0.3035,
"step": 20300
},
{
"epoch": 0.1632,
"grad_norm": 10.893829345703125,
"learning_rate": 8.402489959839358e-06,
"loss": 0.2792,
"step": 20400
},
{
"epoch": 0.164,
"grad_norm": 36.892173767089844,
"learning_rate": 8.394457831325302e-06,
"loss": 0.2835,
"step": 20500
},
{
"epoch": 0.1648,
"grad_norm": 8.867621421813965,
"learning_rate": 8.386425702811246e-06,
"loss": 0.2752,
"step": 20600
},
{
"epoch": 0.1656,
"grad_norm": 5.385016918182373,
"learning_rate": 8.378393574297188e-06,
"loss": 0.2763,
"step": 20700
},
{
"epoch": 0.1664,
"grad_norm": 6.431478023529053,
"learning_rate": 8.370361445783132e-06,
"loss": 0.2874,
"step": 20800
},
{
"epoch": 0.1672,
"grad_norm": 7.890101432800293,
"learning_rate": 8.362329317269076e-06,
"loss": 0.2737,
"step": 20900
},
{
"epoch": 0.168,
"grad_norm": 5.053267955780029,
"learning_rate": 8.35429718875502e-06,
"loss": 0.2885,
"step": 21000
},
{
"epoch": 0.1688,
"grad_norm": 7.141018390655518,
"learning_rate": 8.346265060240964e-06,
"loss": 0.2735,
"step": 21100
},
{
"epoch": 0.1696,
"grad_norm": 12.165709495544434,
"learning_rate": 8.338232931726908e-06,
"loss": 0.3086,
"step": 21200
},
{
"epoch": 0.1704,
"grad_norm": 5.507659912109375,
"learning_rate": 8.330200803212852e-06,
"loss": 0.2773,
"step": 21300
},
{
"epoch": 0.1712,
"grad_norm": 6.527307987213135,
"learning_rate": 8.322248995983936e-06,
"loss": 0.2584,
"step": 21400
},
{
"epoch": 0.172,
"grad_norm": 8.494378089904785,
"learning_rate": 8.31421686746988e-06,
"loss": 0.3005,
"step": 21500
},
{
"epoch": 0.1728,
"grad_norm": 6.213225364685059,
"learning_rate": 8.306184738955824e-06,
"loss": 0.2708,
"step": 21600
},
{
"epoch": 0.1736,
"grad_norm": 10.444897651672363,
"learning_rate": 8.298152610441768e-06,
"loss": 0.2817,
"step": 21700
},
{
"epoch": 0.1744,
"grad_norm": 6.7094950675964355,
"learning_rate": 8.290120481927712e-06,
"loss": 0.2555,
"step": 21800
},
{
"epoch": 0.1752,
"grad_norm": 7.899540901184082,
"learning_rate": 8.282088353413656e-06,
"loss": 0.2902,
"step": 21900
},
{
"epoch": 0.176,
"grad_norm": 7.049533367156982,
"learning_rate": 8.2740562248996e-06,
"loss": 0.2648,
"step": 22000
},
{
"epoch": 0.1768,
"grad_norm": 7.662154674530029,
"learning_rate": 8.266024096385543e-06,
"loss": 0.2984,
"step": 22100
},
{
"epoch": 0.1776,
"grad_norm": 5.362339019775391,
"learning_rate": 8.257991967871487e-06,
"loss": 0.2533,
"step": 22200
},
{
"epoch": 0.1784,
"grad_norm": 11.844931602478027,
"learning_rate": 8.249959839357431e-06,
"loss": 0.2967,
"step": 22300
},
{
"epoch": 0.1792,
"grad_norm": 8.356042861938477,
"learning_rate": 8.241927710843374e-06,
"loss": 0.2817,
"step": 22400
},
{
"epoch": 0.18,
"grad_norm": 6.400118827819824,
"learning_rate": 8.233895582329317e-06,
"loss": 0.2975,
"step": 22500
},
{
"epoch": 0.1808,
"grad_norm": 7.8221611976623535,
"learning_rate": 8.225863453815261e-06,
"loss": 0.2928,
"step": 22600
},
{
"epoch": 0.1816,
"grad_norm": 9.997267723083496,
"learning_rate": 8.217831325301205e-06,
"loss": 0.2846,
"step": 22700
},
{
"epoch": 0.1824,
"grad_norm": 3.741245746612549,
"learning_rate": 8.20979919678715e-06,
"loss": 0.2712,
"step": 22800
},
{
"epoch": 0.1832,
"grad_norm": 6.534188270568848,
"learning_rate": 8.201767068273093e-06,
"loss": 0.2879,
"step": 22900
},
{
"epoch": 0.184,
"grad_norm": 5.730849266052246,
"learning_rate": 8.193734939759037e-06,
"loss": 0.2722,
"step": 23000
},
{
"epoch": 0.1848,
"grad_norm": 4.018589973449707,
"learning_rate": 8.185702811244981e-06,
"loss": 0.2979,
"step": 23100
},
{
"epoch": 0.1856,
"grad_norm": 7.68761682510376,
"learning_rate": 8.177670682730925e-06,
"loss": 0.2919,
"step": 23200
},
{
"epoch": 0.1864,
"grad_norm": 9.94321060180664,
"learning_rate": 8.169638554216867e-06,
"loss": 0.2731,
"step": 23300
},
{
"epoch": 0.1872,
"grad_norm": 2.9645307064056396,
"learning_rate": 8.161686746987953e-06,
"loss": 0.2698,
"step": 23400
},
{
"epoch": 0.188,
"grad_norm": 4.740218639373779,
"learning_rate": 8.153654618473897e-06,
"loss": 0.2731,
"step": 23500
},
{
"epoch": 0.1888,
"grad_norm": 5.427436828613281,
"learning_rate": 8.145622489959841e-06,
"loss": 0.269,
"step": 23600
},
{
"epoch": 0.1896,
"grad_norm": 13.593464851379395,
"learning_rate": 8.137590361445785e-06,
"loss": 0.263,
"step": 23700
},
{
"epoch": 0.1904,
"grad_norm": 7.22516393661499,
"learning_rate": 8.129558232931729e-06,
"loss": 0.3125,
"step": 23800
},
{
"epoch": 0.1912,
"grad_norm": 14.015786170959473,
"learning_rate": 8.121526104417673e-06,
"loss": 0.2832,
"step": 23900
},
{
"epoch": 0.192,
"grad_norm": 8.892059326171875,
"learning_rate": 8.113493975903615e-06,
"loss": 0.3009,
"step": 24000
},
{
"epoch": 0.192,
"eval_test1_cer": 0.07456214056314886,
"eval_test1_cer_norm": 0.05662352630441569,
"eval_test1_loss": 0.22622939944267273,
"eval_test1_runtime": 3568.0234,
"eval_test1_samples_per_second": 0.701,
"eval_test1_steps_per_second": 0.175,
"eval_test1_wer": 0.19636140995364296,
"eval_test1_wer_norm": 0.13466202986644846,
"step": 24000
},
{
"epoch": 0.192,
"eval_test2_cer": 0.17353100384514877,
"eval_test2_cer_norm": 0.14121571893399443,
"eval_test2_loss": 0.37227049469947815,
"eval_test2_runtime": 3757.9926,
"eval_test2_samples_per_second": 0.665,
"eval_test2_steps_per_second": 0.166,
"eval_test2_wer": 0.3350022888532845,
"eval_test2_wer_norm": 0.2660726564290626,
"step": 24000
},
{
"epoch": 0.1928,
"grad_norm": 6.917489051818848,
"learning_rate": 8.105461847389559e-06,
"loss": 0.2745,
"step": 24100
},
{
"epoch": 0.1936,
"grad_norm": 9.797795295715332,
"learning_rate": 8.097429718875503e-06,
"loss": 0.2792,
"step": 24200
},
{
"epoch": 0.1944,
"grad_norm": 7.5331315994262695,
"learning_rate": 8.089397590361447e-06,
"loss": 0.2631,
"step": 24300
},
{
"epoch": 0.1952,
"grad_norm": 28.057600021362305,
"learning_rate": 8.08136546184739e-06,
"loss": 0.2752,
"step": 24400
},
{
"epoch": 0.196,
"grad_norm": 6.589796543121338,
"learning_rate": 8.073413654618475e-06,
"loss": 0.2601,
"step": 24500
},
{
"epoch": 0.1968,
"grad_norm": 7.585195541381836,
"learning_rate": 8.065381526104419e-06,
"loss": 0.2845,
"step": 24600
},
{
"epoch": 0.1976,
"grad_norm": 12.819061279296875,
"learning_rate": 8.057349397590361e-06,
"loss": 0.2807,
"step": 24700
},
{
"epoch": 0.1984,
"grad_norm": 6.550510883331299,
"learning_rate": 8.049317269076305e-06,
"loss": 0.2641,
"step": 24800
},
{
"epoch": 0.1992,
"grad_norm": 4.5952653884887695,
"learning_rate": 8.041285140562249e-06,
"loss": 0.2773,
"step": 24900
},
{
"epoch": 0.2,
"grad_norm": 7.8463134765625,
"learning_rate": 8.033253012048193e-06,
"loss": 0.299,
"step": 25000
},
{
"epoch": 0.2008,
"grad_norm": 8.69621753692627,
"learning_rate": 8.025220883534137e-06,
"loss": 0.2568,
"step": 25100
},
{
"epoch": 0.2016,
"grad_norm": 19.935726165771484,
"learning_rate": 8.01718875502008e-06,
"loss": 0.2875,
"step": 25200
},
{
"epoch": 0.2024,
"grad_norm": 7.2960662841796875,
"learning_rate": 8.009156626506025e-06,
"loss": 0.2757,
"step": 25300
},
{
"epoch": 0.2032,
"grad_norm": 14.16852855682373,
"learning_rate": 8.001124497991969e-06,
"loss": 0.2733,
"step": 25400
},
{
"epoch": 0.204,
"grad_norm": 10.322851181030273,
"learning_rate": 7.993092369477912e-06,
"loss": 0.2732,
"step": 25500
},
{
"epoch": 0.2048,
"grad_norm": 6.632780075073242,
"learning_rate": 7.985060240963856e-06,
"loss": 0.2705,
"step": 25600
},
{
"epoch": 0.2056,
"grad_norm": 9.688406944274902,
"learning_rate": 7.9770281124498e-06,
"loss": 0.3111,
"step": 25700
},
{
"epoch": 0.2064,
"grad_norm": 11.941987037658691,
"learning_rate": 7.968995983935744e-06,
"loss": 0.2688,
"step": 25800
},
{
"epoch": 0.2072,
"grad_norm": 6.580658912658691,
"learning_rate": 7.960963855421688e-06,
"loss": 0.2842,
"step": 25900
},
{
"epoch": 0.208,
"grad_norm": 9.130199432373047,
"learning_rate": 7.95293172690763e-06,
"loss": 0.2789,
"step": 26000
},
{
"epoch": 0.2088,
"grad_norm": 7.189250469207764,
"learning_rate": 7.944899598393574e-06,
"loss": 0.2687,
"step": 26100
},
{
"epoch": 0.2096,
"grad_norm": 6.851436614990234,
"learning_rate": 7.936867469879518e-06,
"loss": 0.3035,
"step": 26200
},
{
"epoch": 0.2104,
"grad_norm": 5.916522979736328,
"learning_rate": 7.928835341365462e-06,
"loss": 0.2895,
"step": 26300
},
{
"epoch": 0.2112,
"grad_norm": 9.012747764587402,
"learning_rate": 7.920803212851406e-06,
"loss": 0.258,
"step": 26400
},
{
"epoch": 0.212,
"grad_norm": 4.732649803161621,
"learning_rate": 7.91277108433735e-06,
"loss": 0.2707,
"step": 26500
},
{
"epoch": 0.2128,
"grad_norm": 7.508853912353516,
"learning_rate": 7.904738955823294e-06,
"loss": 0.2985,
"step": 26600
},
{
"epoch": 0.2136,
"grad_norm": 7.832762241363525,
"learning_rate": 7.896706827309238e-06,
"loss": 0.2799,
"step": 26700
},
{
"epoch": 0.2144,
"grad_norm": 11.426980972290039,
"learning_rate": 7.888674698795182e-06,
"loss": 0.2877,
"step": 26800
},
{
"epoch": 0.2152,
"grad_norm": 9.61413288116455,
"learning_rate": 7.880642570281126e-06,
"loss": 0.2869,
"step": 26900
},
{
"epoch": 0.216,
"grad_norm": 11.701996803283691,
"learning_rate": 7.87261044176707e-06,
"loss": 0.2607,
"step": 27000
},
{
"epoch": 0.2168,
"grad_norm": 7.202447891235352,
"learning_rate": 7.864578313253013e-06,
"loss": 0.2763,
"step": 27100
},
{
"epoch": 0.2176,
"grad_norm": 9.715481758117676,
"learning_rate": 7.856546184738957e-06,
"loss": 0.2871,
"step": 27200
},
{
"epoch": 0.2184,
"grad_norm": 8.603879928588867,
"learning_rate": 7.848514056224901e-06,
"loss": 0.2876,
"step": 27300
},
{
"epoch": 0.2192,
"grad_norm": 7.729073524475098,
"learning_rate": 7.840481927710844e-06,
"loss": 0.25,
"step": 27400
},
{
"epoch": 0.22,
"grad_norm": 4.843418598175049,
"learning_rate": 7.832449799196787e-06,
"loss": 0.2913,
"step": 27500
},
{
"epoch": 0.2208,
"grad_norm": 13.062045097351074,
"learning_rate": 7.824417670682731e-06,
"loss": 0.2676,
"step": 27600
},
{
"epoch": 0.2216,
"grad_norm": 9.212933540344238,
"learning_rate": 7.816385542168675e-06,
"loss": 0.2814,
"step": 27700
},
{
"epoch": 0.2224,
"grad_norm": 8.371397972106934,
"learning_rate": 7.80835341365462e-06,
"loss": 0.2556,
"step": 27800
},
{
"epoch": 0.2232,
"grad_norm": 11.034435272216797,
"learning_rate": 7.800321285140563e-06,
"loss": 0.2656,
"step": 27900
},
{
"epoch": 0.224,
"grad_norm": 7.741024494171143,
"learning_rate": 7.792289156626507e-06,
"loss": 0.2481,
"step": 28000
},
{
"epoch": 0.224,
"eval_test1_cer": 0.08188158189558593,
"eval_test1_cer_norm": 0.06053257137368838,
"eval_test1_loss": 0.21281211078166962,
"eval_test1_runtime": 3606.8392,
"eval_test1_samples_per_second": 0.693,
"eval_test1_steps_per_second": 0.173,
"eval_test1_wer": 0.20149275489081314,
"eval_test1_wer_norm": 0.13837341827639615,
"step": 28000
},
{
"epoch": 0.224,
"eval_test2_cer": 0.22203867547691045,
"eval_test2_cer_norm": 0.17609234583204214,
"eval_test2_loss": 0.35916781425476074,
"eval_test2_runtime": 3907.2621,
"eval_test2_samples_per_second": 0.64,
"eval_test2_steps_per_second": 0.16,
"eval_test2_wer": 0.3924525062943465,
"eval_test2_wer_norm": 0.32526358010543205,
"step": 28000
},
{
"epoch": 0.2248,
"grad_norm": 8.185860633850098,
"learning_rate": 7.784257028112451e-06,
"loss": 0.2649,
"step": 28100
},
{
"epoch": 0.2256,
"grad_norm": 9.275001525878906,
"learning_rate": 7.776224899598395e-06,
"loss": 0.2835,
"step": 28200
},
{
"epoch": 0.2264,
"grad_norm": 11.974363327026367,
"learning_rate": 7.768192771084337e-06,
"loss": 0.279,
"step": 28300
},
{
"epoch": 0.2272,
"grad_norm": 8.101808547973633,
"learning_rate": 7.760160642570281e-06,
"loss": 0.2934,
"step": 28400
},
{
"epoch": 0.228,
"grad_norm": 6.571002006530762,
"learning_rate": 7.752128514056225e-06,
"loss": 0.2728,
"step": 28500
},
{
"epoch": 0.2288,
"grad_norm": 5.396939277648926,
"learning_rate": 7.744096385542169e-06,
"loss": 0.2606,
"step": 28600
},
{
"epoch": 0.2296,
"grad_norm": 11.369711875915527,
"learning_rate": 7.736144578313253e-06,
"loss": 0.2621,
"step": 28700
},
{
"epoch": 0.2304,
"grad_norm": 11.154069900512695,
"learning_rate": 7.728112449799197e-06,
"loss": 0.2668,
"step": 28800
},
{
"epoch": 0.2312,
"grad_norm": 9.739912033081055,
"learning_rate": 7.720080321285141e-06,
"loss": 0.2448,
"step": 28900
},
{
"epoch": 0.232,
"grad_norm": 7.0749125480651855,
"learning_rate": 7.712048192771085e-06,
"loss": 0.2571,
"step": 29000
},
{
"epoch": 0.2328,
"grad_norm": 11.714834213256836,
"learning_rate": 7.704016064257029e-06,
"loss": 0.2687,
"step": 29100
},
{
"epoch": 0.2336,
"grad_norm": 7.279887676239014,
"learning_rate": 7.695983935742973e-06,
"loss": 0.2606,
"step": 29200
},
{
"epoch": 0.2344,
"grad_norm": 13.882575035095215,
"learning_rate": 7.687951807228917e-06,
"loss": 0.2906,
"step": 29300
},
{
"epoch": 0.2352,
"grad_norm": 11.140130996704102,
"learning_rate": 7.67991967871486e-06,
"loss": 0.2795,
"step": 29400
},
{
"epoch": 0.236,
"grad_norm": 8.242719650268555,
"learning_rate": 7.671887550200805e-06,
"loss": 0.2531,
"step": 29500
},
{
"epoch": 0.2368,
"grad_norm": 3.9837567806243896,
"learning_rate": 7.663855421686748e-06,
"loss": 0.2684,
"step": 29600
},
{
"epoch": 0.2376,
"grad_norm": 4.7640275955200195,
"learning_rate": 7.655823293172692e-06,
"loss": 0.2812,
"step": 29700
},
{
"epoch": 0.2384,
"grad_norm": 5.323102951049805,
"learning_rate": 7.647791164658636e-06,
"loss": 0.2742,
"step": 29800
},
{
"epoch": 0.2392,
"grad_norm": 5.6695966720581055,
"learning_rate": 7.63975903614458e-06,
"loss": 0.2662,
"step": 29900
},
{
"epoch": 0.24,
"grad_norm": 11.021814346313477,
"learning_rate": 7.631726907630522e-06,
"loss": 0.2591,
"step": 30000
},
{
"epoch": 0.2408,
"grad_norm": 7.066900253295898,
"learning_rate": 7.623694779116466e-06,
"loss": 0.2799,
"step": 30100
},
{
"epoch": 0.2416,
"grad_norm": 7.006994247436523,
"learning_rate": 7.61566265060241e-06,
"loss": 0.2763,
"step": 30200
},
{
"epoch": 0.2424,
"grad_norm": 6.179232597351074,
"learning_rate": 7.607630522088354e-06,
"loss": 0.2607,
"step": 30300
},
{
"epoch": 0.2432,
"grad_norm": 11.01894760131836,
"learning_rate": 7.599598393574298e-06,
"loss": 0.2892,
"step": 30400
},
{
"epoch": 0.244,
"grad_norm": 8.360103607177734,
"learning_rate": 7.591566265060242e-06,
"loss": 0.2682,
"step": 30500
},
{
"epoch": 0.2448,
"grad_norm": 12.027982711791992,
"learning_rate": 7.583614457831326e-06,
"loss": 0.2795,
"step": 30600
},
{
"epoch": 0.2456,
"grad_norm": 9.705676078796387,
"learning_rate": 7.5755823293172694e-06,
"loss": 0.2604,
"step": 30700
},
{
"epoch": 0.2464,
"grad_norm": 78.23721313476562,
"learning_rate": 7.567550200803213e-06,
"loss": 0.2744,
"step": 30800
},
{
"epoch": 0.2472,
"grad_norm": 7.072327613830566,
"learning_rate": 7.559518072289157e-06,
"loss": 0.2518,
"step": 30900
},
{
"epoch": 0.248,
"grad_norm": 6.551907062530518,
"learning_rate": 7.551485943775101e-06,
"loss": 0.252,
"step": 31000
},
{
"epoch": 0.2488,
"grad_norm": 7.861015796661377,
"learning_rate": 7.543453815261045e-06,
"loss": 0.2616,
"step": 31100
},
{
"epoch": 0.2496,
"grad_norm": 13.113914489746094,
"learning_rate": 7.535421686746989e-06,
"loss": 0.2929,
"step": 31200
},
{
"epoch": 0.2504,
"grad_norm": 10.613142967224121,
"learning_rate": 7.527389558232933e-06,
"loss": 0.2603,
"step": 31300
},
{
"epoch": 0.2512,
"grad_norm": 9.165702819824219,
"learning_rate": 7.519357429718877e-06,
"loss": 0.2633,
"step": 31400
},
{
"epoch": 0.252,
"grad_norm": 5.82572078704834,
"learning_rate": 7.511325301204821e-06,
"loss": 0.2553,
"step": 31500
},
{
"epoch": 0.2528,
"grad_norm": 4.581362247467041,
"learning_rate": 7.503293172690763e-06,
"loss": 0.2735,
"step": 31600
},
{
"epoch": 0.2536,
"grad_norm": 6.59524393081665,
"learning_rate": 7.495261044176707e-06,
"loss": 0.2667,
"step": 31700
},
{
"epoch": 0.2544,
"grad_norm": 10.484724044799805,
"learning_rate": 7.487228915662651e-06,
"loss": 0.2679,
"step": 31800
},
{
"epoch": 0.2552,
"grad_norm": 8.551301002502441,
"learning_rate": 7.479196787148595e-06,
"loss": 0.2538,
"step": 31900
},
{
"epoch": 0.256,
"grad_norm": 6.701406002044678,
"learning_rate": 7.471164658634539e-06,
"loss": 0.2507,
"step": 32000
},
{
"epoch": 0.256,
"eval_test1_cer": 0.08450033609679587,
"eval_test1_cer_norm": 0.058093019905395345,
"eval_test1_loss": 0.20838095247745514,
"eval_test1_runtime": 3620.5796,
"eval_test1_samples_per_second": 0.69,
"eval_test1_steps_per_second": 0.173,
"eval_test1_wer": 0.20155106562873554,
"eval_test1_wer_norm": 0.13854875946111808,
"step": 32000
},
{
"epoch": 0.256,
"eval_test2_cer": 0.17676018217792214,
"eval_test2_cer_norm": 0.13646575766966224,
"eval_test2_loss": 0.35614562034606934,
"eval_test2_runtime": 3787.2574,
"eval_test2_samples_per_second": 0.66,
"eval_test2_steps_per_second": 0.165,
"eval_test2_wer": 0.3315690089265278,
"eval_test2_wer_norm": 0.26286385514554206,
"step": 32000
},
{
"epoch": 0.2568,
"grad_norm": 9.485516548156738,
"learning_rate": 7.463132530120483e-06,
"loss": 0.2773,
"step": 32100
},
{
"epoch": 0.2576,
"grad_norm": 7.148492336273193,
"learning_rate": 7.4551004016064265e-06,
"loss": 0.2609,
"step": 32200
},
{
"epoch": 0.2584,
"grad_norm": 7.948647499084473,
"learning_rate": 7.4470682730923705e-06,
"loss": 0.2901,
"step": 32300
},
{
"epoch": 0.2592,
"grad_norm": 11.446524620056152,
"learning_rate": 7.439036144578314e-06,
"loss": 0.2531,
"step": 32400
},
{
"epoch": 0.26,
"grad_norm": 7.954312324523926,
"learning_rate": 7.4310040160642574e-06,
"loss": 0.2474,
"step": 32500
},
{
"epoch": 0.2608,
"grad_norm": 6.527101039886475,
"learning_rate": 7.422971887550201e-06,
"loss": 0.2713,
"step": 32600
},
{
"epoch": 0.2616,
"grad_norm": 4.255068302154541,
"learning_rate": 7.414939759036145e-06,
"loss": 0.2774,
"step": 32700
},
{
"epoch": 0.2624,
"grad_norm": 12.29005241394043,
"learning_rate": 7.406907630522089e-06,
"loss": 0.2869,
"step": 32800
},
{
"epoch": 0.2632,
"grad_norm": 7.255998611450195,
"learning_rate": 7.398875502008033e-06,
"loss": 0.2776,
"step": 32900
},
{
"epoch": 0.264,
"grad_norm": 5.739976406097412,
"learning_rate": 7.390843373493977e-06,
"loss": 0.2658,
"step": 33000
},
{
"epoch": 0.2648,
"grad_norm": 9.824767112731934,
"learning_rate": 7.38281124497992e-06,
"loss": 0.2741,
"step": 33100
},
{
"epoch": 0.2656,
"grad_norm": 5.075085163116455,
"learning_rate": 7.374779116465864e-06,
"loss": 0.2626,
"step": 33200
},
{
"epoch": 0.2664,
"grad_norm": 11.58659839630127,
"learning_rate": 7.366746987951808e-06,
"loss": 0.2834,
"step": 33300
},
{
"epoch": 0.2672,
"grad_norm": 3.7526087760925293,
"learning_rate": 7.358795180722892e-06,
"loss": 0.2896,
"step": 33400
},
{
"epoch": 0.268,
"grad_norm": 5.333492755889893,
"learning_rate": 7.350763052208836e-06,
"loss": 0.2575,
"step": 33500
},
{
"epoch": 0.2688,
"grad_norm": 8.13162612915039,
"learning_rate": 7.34273092369478e-06,
"loss": 0.2513,
"step": 33600
},
{
"epoch": 0.2696,
"grad_norm": 8.09466552734375,
"learning_rate": 7.334698795180723e-06,
"loss": 0.3005,
"step": 33700
},
{
"epoch": 0.2704,
"grad_norm": 4.5985331535339355,
"learning_rate": 7.326666666666667e-06,
"loss": 0.3217,
"step": 33800
},
{
"epoch": 0.2712,
"grad_norm": 6.758331775665283,
"learning_rate": 7.318634538152611e-06,
"loss": 0.3054,
"step": 33900
},
{
"epoch": 0.272,
"grad_norm": 9.720072746276855,
"learning_rate": 7.310602409638555e-06,
"loss": 0.2554,
"step": 34000
},
{
"epoch": 0.2728,
"grad_norm": 10.393074989318848,
"learning_rate": 7.302570281124498e-06,
"loss": 0.2348,
"step": 34100
},
{
"epoch": 0.2736,
"grad_norm": 5.730389595031738,
"learning_rate": 7.294538152610442e-06,
"loss": 0.264,
"step": 34200
},
{
"epoch": 0.2744,
"grad_norm": 9.347960472106934,
"learning_rate": 7.286506024096386e-06,
"loss": 0.3121,
"step": 34300
},
{
"epoch": 0.2752,
"grad_norm": 5.9382171630859375,
"learning_rate": 7.27847389558233e-06,
"loss": 0.256,
"step": 34400
},
{
"epoch": 0.276,
"grad_norm": 13.730400085449219,
"learning_rate": 7.270441767068274e-06,
"loss": 0.2474,
"step": 34500
},
{
"epoch": 0.2768,
"grad_norm": 5.6664347648620605,
"learning_rate": 7.262409638554218e-06,
"loss": 0.2642,
"step": 34600
},
{
"epoch": 0.2776,
"grad_norm": 4.812320232391357,
"learning_rate": 7.2543775100401615e-06,
"loss": 0.2697,
"step": 34700
},
{
"epoch": 0.2784,
"grad_norm": 6.524253845214844,
"learning_rate": 7.2463453815261055e-06,
"loss": 0.2511,
"step": 34800
},
{
"epoch": 0.2792,
"grad_norm": 10.69779109954834,
"learning_rate": 7.238313253012049e-06,
"loss": 0.2455,
"step": 34900
},
{
"epoch": 0.28,
"grad_norm": 6.162754535675049,
"learning_rate": 7.230281124497992e-06,
"loss": 0.3028,
"step": 35000
},
{
"epoch": 0.2808,
"grad_norm": 10.023324966430664,
"learning_rate": 7.2222489959839355e-06,
"loss": 0.2618,
"step": 35100
},
{
"epoch": 0.2816,
"grad_norm": 7.126629829406738,
"learning_rate": 7.2142168674698794e-06,
"loss": 0.2482,
"step": 35200
},
{
"epoch": 0.2824,
"grad_norm": 5.309641361236572,
"learning_rate": 7.206184738955823e-06,
"loss": 0.2641,
"step": 35300
},
{
"epoch": 0.2832,
"grad_norm": 8.043412208557129,
"learning_rate": 7.198152610441767e-06,
"loss": 0.2779,
"step": 35400
},
{
"epoch": 0.284,
"grad_norm": 6.982008934020996,
"learning_rate": 7.190120481927711e-06,
"loss": 0.2519,
"step": 35500
},
{
"epoch": 0.2848,
"grad_norm": 14.100213050842285,
"learning_rate": 7.182088353413655e-06,
"loss": 0.2621,
"step": 35600
},
{
"epoch": 0.2856,
"grad_norm": 7.013679027557373,
"learning_rate": 7.174056224899599e-06,
"loss": 0.2511,
"step": 35700
},
{
"epoch": 0.2864,
"grad_norm": 12.395895004272461,
"learning_rate": 7.166024096385543e-06,
"loss": 0.2574,
"step": 35800
},
{
"epoch": 0.2872,
"grad_norm": 5.958752155303955,
"learning_rate": 7.157991967871486e-06,
"loss": 0.2376,
"step": 35900
},
{
"epoch": 0.288,
"grad_norm": 7.98759126663208,
"learning_rate": 7.14995983935743e-06,
"loss": 0.2558,
"step": 36000
},
{
"epoch": 0.288,
"eval_test1_cer": 0.09376167002763462,
"eval_test1_cer_norm": 0.06997382764664922,
"eval_test1_loss": 0.2042306363582611,
"eval_test1_runtime": 3524.6047,
"eval_test1_samples_per_second": 0.709,
"eval_test1_steps_per_second": 0.177,
"eval_test1_wer": 0.2142628064958162,
"eval_test1_wer_norm": 0.15251760717729917,
"step": 36000
},
{
"epoch": 0.288,
"eval_test2_cer": 0.18322787172882368,
"eval_test2_cer_norm": 0.1433607065385807,
"eval_test2_loss": 0.34622320532798767,
"eval_test2_runtime": 2531.2967,
"eval_test2_samples_per_second": 0.988,
"eval_test2_steps_per_second": 0.247,
"eval_test2_wer": 0.3355745021744106,
"eval_test2_wer_norm": 0.2676197570479028,
"step": 36000
},
{
"epoch": 0.2888,
"grad_norm": 7.333773612976074,
"learning_rate": 7.141927710843374e-06,
"loss": 0.2581,
"step": 36100
},
{
"epoch": 0.2896,
"grad_norm": 6.082370758056641,
"learning_rate": 7.133895582329318e-06,
"loss": 0.2444,
"step": 36200
},
{
"epoch": 0.2904,
"grad_norm": 7.414235591888428,
"learning_rate": 7.125863453815262e-06,
"loss": 0.2501,
"step": 36300
},
{
"epoch": 0.2912,
"grad_norm": 7.523865699768066,
"learning_rate": 7.117831325301206e-06,
"loss": 0.2732,
"step": 36400
},
{
"epoch": 0.292,
"grad_norm": 4.851339817047119,
"learning_rate": 7.1097991967871496e-06,
"loss": 0.2674,
"step": 36500
},
{
"epoch": 0.2928,
"grad_norm": 14.28159236907959,
"learning_rate": 7.1017670682730935e-06,
"loss": 0.2652,
"step": 36600
},
{
"epoch": 0.2936,
"grad_norm": 9.865965843200684,
"learning_rate": 7.093734939759037e-06,
"loss": 0.2586,
"step": 36700
},
{
"epoch": 0.2944,
"grad_norm": 8.199728965759277,
"learning_rate": 7.085702811244981e-06,
"loss": 0.2606,
"step": 36800
},
{
"epoch": 0.2952,
"grad_norm": 6.126343727111816,
"learning_rate": 7.0776706827309235e-06,
"loss": 0.2244,
"step": 36900
},
{
"epoch": 0.296,
"grad_norm": 12.142101287841797,
"learning_rate": 7.0696385542168675e-06,
"loss": 0.2737,
"step": 37000
},
{
"epoch": 0.2968,
"grad_norm": 18.582122802734375,
"learning_rate": 7.061606425702811e-06,
"loss": 0.2617,
"step": 37100
},
{
"epoch": 0.2976,
"grad_norm": 6.676758766174316,
"learning_rate": 7.053574297188755e-06,
"loss": 0.2679,
"step": 37200
},
{
"epoch": 0.2984,
"grad_norm": 11.17313003540039,
"learning_rate": 7.045542168674699e-06,
"loss": 0.2896,
"step": 37300
},
{
"epoch": 0.2992,
"grad_norm": 6.7832159996032715,
"learning_rate": 7.037510040160643e-06,
"loss": 0.2645,
"step": 37400
},
{
"epoch": 0.3,
"grad_norm": 7.81158971786499,
"learning_rate": 7.029558232931728e-06,
"loss": 0.2486,
"step": 37500
},
{
"epoch": 0.3008,
"grad_norm": 5.17408561706543,
"learning_rate": 7.0215261044176705e-06,
"loss": 0.274,
"step": 37600
},
{
"epoch": 0.3016,
"grad_norm": 4.426403045654297,
"learning_rate": 7.0134939759036144e-06,
"loss": 0.2574,
"step": 37700
},
{
"epoch": 0.3024,
"grad_norm": 4.935957908630371,
"learning_rate": 7.005461847389558e-06,
"loss": 0.239,
"step": 37800
},
{
"epoch": 0.3032,
"grad_norm": 5.82489013671875,
"learning_rate": 6.997429718875502e-06,
"loss": 0.2549,
"step": 37900
},
{
"epoch": 0.304,
"grad_norm": 5.054333209991455,
"learning_rate": 6.989397590361446e-06,
"loss": 0.2675,
"step": 38000
},
{
"epoch": 0.3048,
"grad_norm": 9.367691040039062,
"learning_rate": 6.98136546184739e-06,
"loss": 0.2699,
"step": 38100
},
{
"epoch": 0.3056,
"grad_norm": 11.023392677307129,
"learning_rate": 6.973333333333334e-06,
"loss": 0.2483,
"step": 38200
},
{
"epoch": 0.3064,
"grad_norm": 12.55837345123291,
"learning_rate": 6.965301204819278e-06,
"loss": 0.2557,
"step": 38300
},
{
"epoch": 0.3072,
"grad_norm": 5.479390621185303,
"learning_rate": 6.957269076305222e-06,
"loss": 0.2625,
"step": 38400
},
{
"epoch": 0.308,
"grad_norm": 11.671854972839355,
"learning_rate": 6.949236947791165e-06,
"loss": 0.2608,
"step": 38500
},
{
"epoch": 0.3088,
"grad_norm": 5.811091423034668,
"learning_rate": 6.941204819277109e-06,
"loss": 0.2671,
"step": 38600
},
{
"epoch": 0.3096,
"grad_norm": 5.157212734222412,
"learning_rate": 6.933172690763053e-06,
"loss": 0.2701,
"step": 38700
},
{
"epoch": 0.3104,
"grad_norm": 6.742706775665283,
"learning_rate": 6.925140562248997e-06,
"loss": 0.264,
"step": 38800
},
{
"epoch": 0.3112,
"grad_norm": 10.843968391418457,
"learning_rate": 6.917108433734941e-06,
"loss": 0.2476,
"step": 38900
},
{
"epoch": 0.312,
"grad_norm": 5.495096206665039,
"learning_rate": 6.9090763052208846e-06,
"loss": 0.2674,
"step": 39000
},
{
"epoch": 0.3128,
"grad_norm": 11.919147491455078,
"learning_rate": 6.901124497991969e-06,
"loss": 0.2397,
"step": 39100
},
{
"epoch": 0.3136,
"grad_norm": 9.006637573242188,
"learning_rate": 6.893092369477912e-06,
"loss": 0.2502,
"step": 39200
},
{
"epoch": 0.3144,
"grad_norm": 4.003146171569824,
"learning_rate": 6.885060240963856e-06,
"loss": 0.2697,
"step": 39300
},
{
"epoch": 0.3152,
"grad_norm": 10.066797256469727,
"learning_rate": 6.8770281124498e-06,
"loss": 0.2295,
"step": 39400
},
{
"epoch": 0.316,
"grad_norm": 8.48736572265625,
"learning_rate": 6.868995983935744e-06,
"loss": 0.2566,
"step": 39500
},
{
"epoch": 0.3168,
"grad_norm": 6.740108966827393,
"learning_rate": 6.860963855421688e-06,
"loss": 0.2529,
"step": 39600
},
{
"epoch": 0.3176,
"grad_norm": 7.446850299835205,
"learning_rate": 6.8529317269076315e-06,
"loss": 0.2571,
"step": 39700
},
{
"epoch": 0.3184,
"grad_norm": 4.642506122589111,
"learning_rate": 6.8448995983935755e-06,
"loss": 0.3017,
"step": 39800
},
{
"epoch": 0.3192,
"grad_norm": 5.234320163726807,
"learning_rate": 6.836867469879519e-06,
"loss": 0.2668,
"step": 39900
},
{
"epoch": 0.32,
"grad_norm": 7.142499923706055,
"learning_rate": 6.828835341365463e-06,
"loss": 0.2566,
"step": 40000
},
{
"epoch": 0.32,
"eval_test1_cer": 0.06914724774068265,
"eval_test1_cer_norm": 0.04478113669652076,
"eval_test1_loss": 0.19836583733558655,
"eval_test1_runtime": 2432.2319,
"eval_test1_samples_per_second": 1.028,
"eval_test1_steps_per_second": 0.257,
"eval_test1_wer": 0.17665238053587567,
"eval_test1_wer_norm": 0.11657266430930185,
"step": 40000
},
{
"epoch": 0.32,
"eval_test2_cer": 0.15002146563631613,
"eval_test2_cer_norm": 0.10912806011775643,
"eval_test2_loss": 0.33799564838409424,
"eval_test2_runtime": 2481.3184,
"eval_test2_samples_per_second": 1.008,
"eval_test2_steps_per_second": 0.252,
"eval_test2_wer": 0.28587777523460745,
"eval_test2_wer_norm": 0.2201466880586752,
"step": 40000
},
{
"epoch": 0.3208,
"grad_norm": 6.141145706176758,
"learning_rate": 6.8208032128514055e-06,
"loss": 0.2562,
"step": 40100
},
{
"epoch": 0.3216,
"grad_norm": 5.139159679412842,
"learning_rate": 6.8127710843373495e-06,
"loss": 0.2703,
"step": 40200
},
{
"epoch": 0.3224,
"grad_norm": 7.440242767333984,
"learning_rate": 6.804738955823293e-06,
"loss": 0.2821,
"step": 40300
},
{
"epoch": 0.3232,
"grad_norm": 6.795882225036621,
"learning_rate": 6.796706827309237e-06,
"loss": 0.2452,
"step": 40400
},
{
"epoch": 0.324,
"grad_norm": 7.417604446411133,
"learning_rate": 6.788674698795181e-06,
"loss": 0.2521,
"step": 40500
},
{
"epoch": 0.3248,
"grad_norm": 6.449469089508057,
"learning_rate": 6.780642570281125e-06,
"loss": 0.2813,
"step": 40600
},
{
"epoch": 0.3256,
"grad_norm": 10.74730110168457,
"learning_rate": 6.772610441767069e-06,
"loss": 0.2551,
"step": 40700
},
{
"epoch": 0.3264,
"grad_norm": 3.9926106929779053,
"learning_rate": 6.764578313253013e-06,
"loss": 0.242,
"step": 40800
},
{
"epoch": 0.3272,
"grad_norm": 3.6253511905670166,
"learning_rate": 6.756546184738957e-06,
"loss": 0.2359,
"step": 40900
},
{
"epoch": 0.328,
"grad_norm": 7.051224231719971,
"learning_rate": 6.7485140562249e-06,
"loss": 0.2618,
"step": 41000
},
{
"epoch": 0.3288,
"grad_norm": 9.776261329650879,
"learning_rate": 6.740481927710844e-06,
"loss": 0.2649,
"step": 41100
},
{
"epoch": 0.3296,
"grad_norm": 5.289026260375977,
"learning_rate": 6.732449799196788e-06,
"loss": 0.2446,
"step": 41200
},
{
"epoch": 0.3304,
"grad_norm": 9.757101058959961,
"learning_rate": 6.724417670682732e-06,
"loss": 0.2605,
"step": 41300
},
{
"epoch": 0.3312,
"grad_norm": 7.856915473937988,
"learning_rate": 6.716385542168675e-06,
"loss": 0.2566,
"step": 41400
},
{
"epoch": 0.332,
"grad_norm": 4.344404697418213,
"learning_rate": 6.708353413654619e-06,
"loss": 0.2482,
"step": 41500
},
{
"epoch": 0.3328,
"grad_norm": 7.893800735473633,
"learning_rate": 6.700321285140563e-06,
"loss": 0.2555,
"step": 41600
},
{
"epoch": 0.3336,
"grad_norm": 10.901208877563477,
"learning_rate": 6.6922891566265066e-06,
"loss": 0.2455,
"step": 41700
},
{
"epoch": 0.3344,
"grad_norm": 8.60644817352295,
"learning_rate": 6.6842570281124505e-06,
"loss": 0.242,
"step": 41800
},
{
"epoch": 0.3352,
"grad_norm": 10.857059478759766,
"learning_rate": 6.6762248995983936e-06,
"loss": 0.2611,
"step": 41900
},
{
"epoch": 0.336,
"grad_norm": 4.682589054107666,
"learning_rate": 6.6681927710843375e-06,
"loss": 0.2365,
"step": 42000
},
{
"epoch": 0.3368,
"grad_norm": 10.836357116699219,
"learning_rate": 6.660160642570281e-06,
"loss": 0.2477,
"step": 42100
},
{
"epoch": 0.3376,
"grad_norm": 4.692202568054199,
"learning_rate": 6.652128514056225e-06,
"loss": 0.2494,
"step": 42200
},
{
"epoch": 0.3384,
"grad_norm": 10.428234100341797,
"learning_rate": 6.644096385542169e-06,
"loss": 0.2605,
"step": 42300
},
{
"epoch": 0.3392,
"grad_norm": 4.078035831451416,
"learning_rate": 6.636064257028113e-06,
"loss": 0.2358,
"step": 42400
},
{
"epoch": 0.34,
"grad_norm": 7.381161689758301,
"learning_rate": 6.628032128514057e-06,
"loss": 0.2371,
"step": 42500
},
{
"epoch": 0.3408,
"grad_norm": 2.6584036350250244,
"learning_rate": 6.620000000000001e-06,
"loss": 0.2795,
"step": 42600
},
{
"epoch": 0.3416,
"grad_norm": 10.325010299682617,
"learning_rate": 6.611967871485945e-06,
"loss": 0.2476,
"step": 42700
},
{
"epoch": 0.3424,
"grad_norm": 6.986746788024902,
"learning_rate": 6.603935742971887e-06,
"loss": 0.2354,
"step": 42800
},
{
"epoch": 0.3432,
"grad_norm": 6.040923118591309,
"learning_rate": 6.595903614457831e-06,
"loss": 0.2748,
"step": 42900
},
{
"epoch": 0.344,
"grad_norm": 4.043920993804932,
"learning_rate": 6.587871485943775e-06,
"loss": 0.2579,
"step": 43000
},
{
"epoch": 0.3448,
"grad_norm": 7.066967964172363,
"learning_rate": 6.57991967871486e-06,
"loss": 0.2558,
"step": 43100
},
{
"epoch": 0.3456,
"grad_norm": 7.57224702835083,
"learning_rate": 6.571887550200804e-06,
"loss": 0.2542,
"step": 43200
},
{
"epoch": 0.3464,
"grad_norm": 11.219292640686035,
"learning_rate": 6.563855421686748e-06,
"loss": 0.2688,
"step": 43300
},
{
"epoch": 0.3472,
"grad_norm": 9.798463821411133,
"learning_rate": 6.555823293172692e-06,
"loss": 0.2555,
"step": 43400
},
{
"epoch": 0.348,
"grad_norm": 6.011776924133301,
"learning_rate": 6.547791164658636e-06,
"loss": 0.2517,
"step": 43500
},
{
"epoch": 0.3488,
"grad_norm": 10.116564750671387,
"learning_rate": 6.539759036144578e-06,
"loss": 0.2585,
"step": 43600
},
{
"epoch": 0.3496,
"grad_norm": 8.014334678649902,
"learning_rate": 6.531807228915663e-06,
"loss": 0.2386,
"step": 43700
},
{
"epoch": 0.3504,
"grad_norm": 6.700802803039551,
"learning_rate": 6.523775100401607e-06,
"loss": 0.2412,
"step": 43800
},
{
"epoch": 0.3512,
"grad_norm": 5.972689151763916,
"learning_rate": 6.515742971887551e-06,
"loss": 0.2461,
"step": 43900
},
{
"epoch": 0.352,
"grad_norm": 6.2880167961120605,
"learning_rate": 6.507710843373495e-06,
"loss": 0.2535,
"step": 44000
},
{
"epoch": 0.352,
"eval_test1_cer": 0.07431940398834865,
"eval_test1_cer_norm": 0.051710807501140536,
"eval_test1_loss": 0.19776684045791626,
"eval_test1_runtime": 2427.0653,
"eval_test1_samples_per_second": 1.03,
"eval_test1_steps_per_second": 0.258,
"eval_test1_wer": 0.17638998221522492,
"eval_test1_wer_norm": 0.11625120547064496,
"step": 44000
},
{
"epoch": 0.352,
"eval_test2_cer": 0.1173377011236794,
"eval_test2_cer_norm": 0.09236519987604587,
"eval_test2_loss": 0.3366641104221344,
"eval_test2_runtime": 2441.4105,
"eval_test2_samples_per_second": 1.024,
"eval_test2_steps_per_second": 0.256,
"eval_test2_wer": 0.2464522774090181,
"eval_test2_wer_norm": 0.1820135228054091,
"step": 44000
},
{
"epoch": 0.3528,
"grad_norm": 13.718138694763184,
"learning_rate": 6.499678714859439e-06,
"loss": 0.2509,
"step": 44100
},
{
"epoch": 0.3536,
"grad_norm": 8.281696319580078,
"learning_rate": 6.491646586345383e-06,
"loss": 0.2489,
"step": 44200
},
{
"epoch": 0.3544,
"grad_norm": 4.6960225105285645,
"learning_rate": 6.483614457831325e-06,
"loss": 0.2693,
"step": 44300
},
{
"epoch": 0.3552,
"grad_norm": 6.965161323547363,
"learning_rate": 6.475582329317269e-06,
"loss": 0.2242,
"step": 44400
},
{
"epoch": 0.356,
"grad_norm": 4.1582722663879395,
"learning_rate": 6.467550200803213e-06,
"loss": 0.2551,
"step": 44500
},
{
"epoch": 0.3568,
"grad_norm": 6.540956020355225,
"learning_rate": 6.459518072289157e-06,
"loss": 0.2504,
"step": 44600
},
{
"epoch": 0.3576,
"grad_norm": 8.521894454956055,
"learning_rate": 6.451485943775101e-06,
"loss": 0.2578,
"step": 44700
},
{
"epoch": 0.3584,
"grad_norm": 9.609855651855469,
"learning_rate": 6.443453815261045e-06,
"loss": 0.2727,
"step": 44800
},
{
"epoch": 0.3592,
"grad_norm": 7.131938457489014,
"learning_rate": 6.4354216867469885e-06,
"loss": 0.2519,
"step": 44900
},
{
"epoch": 0.36,
"grad_norm": 3.9930307865142822,
"learning_rate": 6.4273895582329325e-06,
"loss": 0.246,
"step": 45000
},
{
"epoch": 0.3608,
"grad_norm": 12.197285652160645,
"learning_rate": 6.419357429718876e-06,
"loss": 0.2592,
"step": 45100
},
{
"epoch": 0.3616,
"grad_norm": 11.38996410369873,
"learning_rate": 6.4113253012048195e-06,
"loss": 0.2645,
"step": 45200
},
{
"epoch": 0.3624,
"grad_norm": 9.129228591918945,
"learning_rate": 6.403293172690763e-06,
"loss": 0.254,
"step": 45300
},
{
"epoch": 0.3632,
"grad_norm": 4.596703052520752,
"learning_rate": 6.395261044176707e-06,
"loss": 0.2555,
"step": 45400
},
{
"epoch": 0.364,
"grad_norm": 4.702436923980713,
"learning_rate": 6.387228915662651e-06,
"loss": 0.2617,
"step": 45500
},
{
"epoch": 0.3648,
"grad_norm": 6.283945083618164,
"learning_rate": 6.379196787148595e-06,
"loss": 0.2546,
"step": 45600
},
{
"epoch": 0.3656,
"grad_norm": 8.981857299804688,
"learning_rate": 6.371164658634539e-06,
"loss": 0.2353,
"step": 45700
},
{
"epoch": 0.3664,
"grad_norm": 7.928671360015869,
"learning_rate": 6.363132530120483e-06,
"loss": 0.2657,
"step": 45800
},
{
"epoch": 0.3672,
"grad_norm": 7.3669114112854,
"learning_rate": 6.355100401606427e-06,
"loss": 0.2486,
"step": 45900
},
{
"epoch": 0.368,
"grad_norm": 6.917601108551025,
"learning_rate": 6.347068273092371e-06,
"loss": 0.2524,
"step": 46000
},
{
"epoch": 0.3688,
"grad_norm": 4.770781517028809,
"learning_rate": 6.339036144578313e-06,
"loss": 0.2434,
"step": 46100
},
{
"epoch": 0.3696,
"grad_norm": 12.580122947692871,
"learning_rate": 6.331004016064257e-06,
"loss": 0.2463,
"step": 46200
},
{
"epoch": 0.3704,
"grad_norm": 6.896516799926758,
"learning_rate": 6.322971887550201e-06,
"loss": 0.2365,
"step": 46300
},
{
"epoch": 0.3712,
"grad_norm": 7.098085880279541,
"learning_rate": 6.314939759036145e-06,
"loss": 0.2412,
"step": 46400
},
{
"epoch": 0.372,
"grad_norm": 5.216635704040527,
"learning_rate": 6.306907630522089e-06,
"loss": 0.2209,
"step": 46500
},
{
"epoch": 0.3728,
"grad_norm": 7.694733619689941,
"learning_rate": 6.298875502008033e-06,
"loss": 0.2256,
"step": 46600
},
{
"epoch": 0.3736,
"grad_norm": 37.76133346557617,
"learning_rate": 6.290923694779118e-06,
"loss": 0.24,
"step": 46700
},
{
"epoch": 0.3744,
"grad_norm": 4.706324100494385,
"learning_rate": 6.28289156626506e-06,
"loss": 0.2224,
"step": 46800
},
{
"epoch": 0.3752,
"grad_norm": 9.157281875610352,
"learning_rate": 6.274859437751004e-06,
"loss": 0.2609,
"step": 46900
},
{
"epoch": 0.376,
"grad_norm": 6.885075092315674,
"learning_rate": 6.266827309236948e-06,
"loss": 0.2711,
"step": 47000
},
{
"epoch": 0.3768,
"grad_norm": 5.564688205718994,
"learning_rate": 6.258795180722892e-06,
"loss": 0.2634,
"step": 47100
},
{
"epoch": 0.3776,
"grad_norm": 2.6855292320251465,
"learning_rate": 6.250763052208836e-06,
"loss": 0.2509,
"step": 47200
},
{
"epoch": 0.3784,
"grad_norm": 9.584918975830078,
"learning_rate": 6.24273092369478e-06,
"loss": 0.2586,
"step": 47300
},
{
"epoch": 0.3792,
"grad_norm": 9.060691833496094,
"learning_rate": 6.2346987951807236e-06,
"loss": 0.2658,
"step": 47400
},
{
"epoch": 0.38,
"grad_norm": 3.5146710872650146,
"learning_rate": 6.2266666666666675e-06,
"loss": 0.2441,
"step": 47500
},
{
"epoch": 0.3808,
"grad_norm": 6.001715183258057,
"learning_rate": 6.218634538152611e-06,
"loss": 0.2403,
"step": 47600
},
{
"epoch": 0.3816,
"grad_norm": 3.3255221843719482,
"learning_rate": 6.2106024096385545e-06,
"loss": 0.237,
"step": 47700
},
{
"epoch": 0.3824,
"grad_norm": 9.07324504852295,
"learning_rate": 6.202570281124498e-06,
"loss": 0.2522,
"step": 47800
},
{
"epoch": 0.3832,
"grad_norm": 10.312811851501465,
"learning_rate": 6.194538152610442e-06,
"loss": 0.2356,
"step": 47900
},
{
"epoch": 0.384,
"grad_norm": 8.543242454528809,
"learning_rate": 6.186506024096386e-06,
"loss": 0.25,
"step": 48000
},
{
"epoch": 0.384,
"eval_test1_cer": 0.07607457614459631,
"eval_test1_cer_norm": 0.056354599370903063,
"eval_test1_loss": 0.19436757266521454,
"eval_test1_runtime": 2446.1615,
"eval_test1_samples_per_second": 1.022,
"eval_test1_steps_per_second": 0.256,
"eval_test1_wer": 0.1832123385521444,
"eval_test1_wer_norm": 0.12332329992109646,
"step": 48000
},
{
"epoch": 0.384,
"eval_test2_cer": 0.12698323813790272,
"eval_test2_cer_norm": 0.10085799504183453,
"eval_test2_loss": 0.3315908908843994,
"eval_test2_runtime": 2495.6856,
"eval_test2_samples_per_second": 1.002,
"eval_test2_steps_per_second": 0.25,
"eval_test2_wer": 0.25469214923323413,
"eval_test2_wer_norm": 0.1900641760256704,
"step": 48000
},
{
"epoch": 0.3848,
"grad_norm": 7.8291215896606445,
"learning_rate": 6.17847389558233e-06,
"loss": 0.2452,
"step": 48100
},
{
"epoch": 0.3856,
"grad_norm": 4.892631530761719,
"learning_rate": 6.170441767068274e-06,
"loss": 0.2568,
"step": 48200
},
{
"epoch": 0.3864,
"grad_norm": 9.617656707763672,
"learning_rate": 6.162409638554218e-06,
"loss": 0.2197,
"step": 48300
},
{
"epoch": 0.3872,
"grad_norm": 9.659303665161133,
"learning_rate": 6.154377510040162e-06,
"loss": 0.2392,
"step": 48400
},
{
"epoch": 0.388,
"grad_norm": 6.675602912902832,
"learning_rate": 6.146345381526105e-06,
"loss": 0.2209,
"step": 48500
},
{
"epoch": 0.3888,
"grad_norm": 5.2548298835754395,
"learning_rate": 6.138313253012048e-06,
"loss": 0.2642,
"step": 48600
},
{
"epoch": 0.3896,
"grad_norm": 5.428624629974365,
"learning_rate": 6.130281124497992e-06,
"loss": 0.2139,
"step": 48700
},
{
"epoch": 0.3904,
"grad_norm": 10.75515079498291,
"learning_rate": 6.122248995983936e-06,
"loss": 0.2238,
"step": 48800
},
{
"epoch": 0.3912,
"grad_norm": 9.192462921142578,
"learning_rate": 6.11421686746988e-06,
"loss": 0.2422,
"step": 48900
},
{
"epoch": 0.392,
"grad_norm": 5.55458927154541,
"learning_rate": 6.106184738955824e-06,
"loss": 0.2559,
"step": 49000
},
{
"epoch": 0.3928,
"grad_norm": 10.406610488891602,
"learning_rate": 6.098152610441768e-06,
"loss": 0.2474,
"step": 49100
},
{
"epoch": 0.3936,
"grad_norm": 6.274507999420166,
"learning_rate": 6.0901204819277116e-06,
"loss": 0.2361,
"step": 49200
},
{
"epoch": 0.3944,
"grad_norm": 10.830498695373535,
"learning_rate": 6.0820883534136555e-06,
"loss": 0.2409,
"step": 49300
},
{
"epoch": 0.3952,
"grad_norm": 11.912911415100098,
"learning_rate": 6.074056224899599e-06,
"loss": 0.2467,
"step": 49400
},
{
"epoch": 0.396,
"grad_norm": 8.484502792358398,
"learning_rate": 6.0660240963855425e-06,
"loss": 0.2411,
"step": 49500
},
{
"epoch": 0.3968,
"grad_norm": 9.443007469177246,
"learning_rate": 6.057991967871486e-06,
"loss": 0.2449,
"step": 49600
},
{
"epoch": 0.3976,
"grad_norm": 10.93896770477295,
"learning_rate": 6.04995983935743e-06,
"loss": 0.239,
"step": 49700
},
{
"epoch": 0.3984,
"grad_norm": 7.4916768074035645,
"learning_rate": 6.041927710843373e-06,
"loss": 0.2349,
"step": 49800
},
{
"epoch": 0.3992,
"grad_norm": 7.075978755950928,
"learning_rate": 6.033895582329317e-06,
"loss": 0.2418,
"step": 49900
},
{
"epoch": 0.4,
"grad_norm": 5.759922981262207,
"learning_rate": 6.025863453815261e-06,
"loss": 0.2536,
"step": 50000
},
{
"epoch": 0.4008,
"grad_norm": 10.531188011169434,
"learning_rate": 6.017831325301205e-06,
"loss": 0.2162,
"step": 50100
},
{
"epoch": 0.4016,
"grad_norm": 7.113009452819824,
"learning_rate": 6.009799196787149e-06,
"loss": 0.2373,
"step": 50200
},
{
"epoch": 0.4024,
"grad_norm": 8.859579086303711,
"learning_rate": 6.001767068273093e-06,
"loss": 0.2327,
"step": 50300
},
{
"epoch": 0.4032,
"grad_norm": 5.7207512855529785,
"learning_rate": 5.993734939759036e-06,
"loss": 0.2605,
"step": 50400
},
{
"epoch": 0.404,
"grad_norm": 5.358285427093506,
"learning_rate": 5.98570281124498e-06,
"loss": 0.2442,
"step": 50500
},
{
"epoch": 0.4048,
"grad_norm": 9.501456260681152,
"learning_rate": 5.977670682730924e-06,
"loss": 0.2333,
"step": 50600
},
{
"epoch": 0.4056,
"grad_norm": 9.682687759399414,
"learning_rate": 5.969638554216868e-06,
"loss": 0.2404,
"step": 50700
},
{
"epoch": 0.4064,
"grad_norm": 6.1207380294799805,
"learning_rate": 5.961686746987952e-06,
"loss": 0.2401,
"step": 50800
},
{
"epoch": 0.4072,
"grad_norm": 3.812893867492676,
"learning_rate": 5.953654618473896e-06,
"loss": 0.2411,
"step": 50900
},
{
"epoch": 0.408,
"grad_norm": 8.557660102844238,
"learning_rate": 5.94562248995984e-06,
"loss": 0.2486,
"step": 51000
},
{
"epoch": 0.4088,
"grad_norm": 8.239738464355469,
"learning_rate": 5.937590361445784e-06,
"loss": 0.219,
"step": 51100
},
{
"epoch": 0.4096,
"grad_norm": 8.825906753540039,
"learning_rate": 5.929558232931727e-06,
"loss": 0.2418,
"step": 51200
},
{
"epoch": 0.4104,
"grad_norm": 9.038442611694336,
"learning_rate": 5.921526104417671e-06,
"loss": 0.2495,
"step": 51300
},
{
"epoch": 0.4112,
"grad_norm": 10.421882629394531,
"learning_rate": 5.913493975903615e-06,
"loss": 0.2294,
"step": 51400
},
{
"epoch": 0.412,
"grad_norm": 5.004858493804932,
"learning_rate": 5.905461847389559e-06,
"loss": 0.2553,
"step": 51500
},
{
"epoch": 0.4128,
"grad_norm": 5.850473880767822,
"learning_rate": 5.897429718875503e-06,
"loss": 0.2443,
"step": 51600
},
{
"epoch": 0.4136,
"grad_norm": 5.483931064605713,
"learning_rate": 5.8893975903614466e-06,
"loss": 0.2849,
"step": 51700
},
{
"epoch": 0.4144,
"grad_norm": 9.20142650604248,
"learning_rate": 5.881445783132531e-06,
"loss": 0.2617,
"step": 51800
},
{
"epoch": 0.4152,
"grad_norm": 5.675454139709473,
"learning_rate": 5.873413654618474e-06,
"loss": 0.2666,
"step": 51900
},
{
"epoch": 0.416,
"grad_norm": 7.959702968597412,
"learning_rate": 5.865381526104418e-06,
"loss": 0.2286,
"step": 52000
},
{
"epoch": 0.416,
"eval_test1_cer": 0.06314885353648518,
"eval_test1_cer_norm": 0.04443057122962038,
"eval_test1_loss": 0.1907189041376114,
"eval_test1_runtime": 2490.0845,
"eval_test1_samples_per_second": 1.004,
"eval_test1_steps_per_second": 0.251,
"eval_test1_wer": 0.18128808420070555,
"eval_test1_wer_norm": 0.12083929980420234,
"step": 52000
},
{
"epoch": 0.416,
"eval_test2_cer": 0.1296011124799343,
"eval_test2_cer_norm": 0.09485396653238302,
"eval_test2_loss": 0.3308376371860504,
"eval_test2_runtime": 3341.4359,
"eval_test2_samples_per_second": 0.748,
"eval_test2_steps_per_second": 0.187,
"eval_test2_wer": 0.2722304875257496,
"eval_test2_wer_norm": 0.2069390327756131,
"step": 52000
},
{
"epoch": 0.4168,
"grad_norm": 13.770788192749023,
"learning_rate": 5.857349397590362e-06,
"loss": 0.2467,
"step": 52100
},
{
"epoch": 0.4176,
"grad_norm": 5.342738151550293,
"learning_rate": 5.849317269076306e-06,
"loss": 0.233,
"step": 52200
},
{
"epoch": 0.4184,
"grad_norm": 6.316587448120117,
"learning_rate": 5.84128514056225e-06,
"loss": 0.2513,
"step": 52300
},
{
"epoch": 0.4192,
"grad_norm": 6.252966403961182,
"learning_rate": 5.8332530120481936e-06,
"loss": 0.2495,
"step": 52400
},
{
"epoch": 0.42,
"grad_norm": 8.304986953735352,
"learning_rate": 5.8252208835341375e-06,
"loss": 0.2573,
"step": 52500
},
{
"epoch": 0.4208,
"grad_norm": 8.028290748596191,
"learning_rate": 5.817188755020081e-06,
"loss": 0.2393,
"step": 52600
},
{
"epoch": 0.4216,
"grad_norm": 10.630465507507324,
"learning_rate": 5.809156626506025e-06,
"loss": 0.2453,
"step": 52700
},
{
"epoch": 0.4224,
"grad_norm": 9.712656021118164,
"learning_rate": 5.8011244979919675e-06,
"loss": 0.2553,
"step": 52800
},
{
"epoch": 0.4232,
"grad_norm": 9.605881690979004,
"learning_rate": 5.7930923694779115e-06,
"loss": 0.2393,
"step": 52900
},
{
"epoch": 0.424,
"grad_norm": 10.953594207763672,
"learning_rate": 5.785060240963855e-06,
"loss": 0.2484,
"step": 53000
},
{
"epoch": 0.4248,
"grad_norm": 4.043101787567139,
"learning_rate": 5.777028112449799e-06,
"loss": 0.2146,
"step": 53100
},
{
"epoch": 0.4256,
"grad_norm": 8.587929725646973,
"learning_rate": 5.768995983935743e-06,
"loss": 0.2555,
"step": 53200
},
{
"epoch": 0.4264,
"grad_norm": 6.47918701171875,
"learning_rate": 5.760963855421687e-06,
"loss": 0.2277,
"step": 53300
},
{
"epoch": 0.4272,
"grad_norm": 6.032763957977295,
"learning_rate": 5.752931726907631e-06,
"loss": 0.24,
"step": 53400
},
{
"epoch": 0.428,
"grad_norm": 4.708799362182617,
"learning_rate": 5.744899598393575e-06,
"loss": 0.2262,
"step": 53500
},
{
"epoch": 0.4288,
"grad_norm": 7.626307964324951,
"learning_rate": 5.736867469879519e-06,
"loss": 0.2344,
"step": 53600
},
{
"epoch": 0.4296,
"grad_norm": 5.346840858459473,
"learning_rate": 5.728835341365462e-06,
"loss": 0.2071,
"step": 53700
},
{
"epoch": 0.4304,
"grad_norm": 6.508141040802002,
"learning_rate": 5.720803212851406e-06,
"loss": 0.2275,
"step": 53800
},
{
"epoch": 0.4312,
"grad_norm": 7.617433071136475,
"learning_rate": 5.71277108433735e-06,
"loss": 0.2359,
"step": 53900
},
{
"epoch": 0.432,
"grad_norm": 5.544378757476807,
"learning_rate": 5.704738955823294e-06,
"loss": 0.2491,
"step": 54000
},
{
"epoch": 0.4328,
"grad_norm": 5.247679233551025,
"learning_rate": 5.696706827309238e-06,
"loss": 0.2353,
"step": 54100
},
{
"epoch": 0.4336,
"grad_norm": 9.137139320373535,
"learning_rate": 5.688674698795182e-06,
"loss": 0.2397,
"step": 54200
},
{
"epoch": 0.4344,
"grad_norm": 6.74301815032959,
"learning_rate": 5.6806425702811255e-06,
"loss": 0.2602,
"step": 54300
},
{
"epoch": 0.4352,
"grad_norm": 6.35469913482666,
"learning_rate": 5.672610441767069e-06,
"loss": 0.2515,
"step": 54400
},
{
"epoch": 0.436,
"grad_norm": 10.135574340820312,
"learning_rate": 5.664578313253013e-06,
"loss": 0.2219,
"step": 54500
},
{
"epoch": 0.4368,
"grad_norm": 5.893723011016846,
"learning_rate": 5.6565461847389556e-06,
"loss": 0.2569,
"step": 54600
},
{
"epoch": 0.4376,
"grad_norm": 14.050215721130371,
"learning_rate": 5.6485140562248995e-06,
"loss": 0.2906,
"step": 54700
},
{
"epoch": 0.4384,
"grad_norm": 6.908878326416016,
"learning_rate": 5.640481927710843e-06,
"loss": 0.2444,
"step": 54800
},
{
"epoch": 0.4392,
"grad_norm": 7.57582950592041,
"learning_rate": 5.632449799196787e-06,
"loss": 0.2467,
"step": 54900
},
{
"epoch": 0.44,
"grad_norm": 11.190139770507812,
"learning_rate": 5.624417670682731e-06,
"loss": 0.2593,
"step": 55000
},
{
"epoch": 0.4408,
"grad_norm": 12.42068099975586,
"learning_rate": 5.616385542168675e-06,
"loss": 0.2453,
"step": 55100
},
{
"epoch": 0.4416,
"grad_norm": 10.345560073852539,
"learning_rate": 5.608353413654619e-06,
"loss": 0.2411,
"step": 55200
},
{
"epoch": 0.4424,
"grad_norm": 7.91195011138916,
"learning_rate": 5.600321285140563e-06,
"loss": 0.2414,
"step": 55300
},
{
"epoch": 0.4432,
"grad_norm": 6.820638179779053,
"learning_rate": 5.592289156626507e-06,
"loss": 0.2363,
"step": 55400
},
{
"epoch": 0.444,
"grad_norm": 4.268233299255371,
"learning_rate": 5.58425702811245e-06,
"loss": 0.2325,
"step": 55500
},
{
"epoch": 0.4448,
"grad_norm": 8.063896179199219,
"learning_rate": 5.576224899598394e-06,
"loss": 0.2607,
"step": 55600
},
{
"epoch": 0.4456,
"grad_norm": 3.5322296619415283,
"learning_rate": 5.568192771084338e-06,
"loss": 0.2346,
"step": 55700
},
{
"epoch": 0.4464,
"grad_norm": 9.244215965270996,
"learning_rate": 5.560160642570282e-06,
"loss": 0.2155,
"step": 55800
},
{
"epoch": 0.4472,
"grad_norm": 10.747200965881348,
"learning_rate": 5.552208835341366e-06,
"loss": 0.2351,
"step": 55900
},
{
"epoch": 0.448,
"grad_norm": 14.072000503540039,
"learning_rate": 5.54417670682731e-06,
"loss": 0.2413,
"step": 56000
},
{
"epoch": 0.448,
"eval_test1_cer": 0.0663977892299649,
"eval_test1_cer_norm": 0.0478834009652556,
"eval_test1_loss": 0.1882741004228592,
"eval_test1_runtime": 3418.3969,
"eval_test1_samples_per_second": 0.731,
"eval_test1_steps_per_second": 0.183,
"eval_test1_wer": 0.16968424735415027,
"eval_test1_wer_norm": 0.1093544522049154,
"step": 56000
},
{
"epoch": 0.448,
"eval_test2_cer": 0.13036174263635344,
"eval_test2_cer_norm": 0.09973950263402541,
"eval_test2_loss": 0.3249567449092865,
"eval_test2_runtime": 3504.8089,
"eval_test2_samples_per_second": 0.713,
"eval_test2_steps_per_second": 0.178,
"eval_test2_wer": 0.25629434653238725,
"eval_test2_wer_norm": 0.18966307586523035,
"step": 56000
},
{
"epoch": 0.4488,
"grad_norm": 8.576178550720215,
"learning_rate": 5.536144578313254e-06,
"loss": 0.2404,
"step": 56100
},
{
"epoch": 0.4496,
"grad_norm": 6.107906818389893,
"learning_rate": 5.528112449799197e-06,
"loss": 0.2217,
"step": 56200
},
{
"epoch": 0.4504,
"grad_norm": 5.073305606842041,
"learning_rate": 5.520080321285141e-06,
"loss": 0.229,
"step": 56300
},
{
"epoch": 0.4512,
"grad_norm": 5.073732376098633,
"learning_rate": 5.512048192771085e-06,
"loss": 0.2404,
"step": 56400
},
{
"epoch": 0.452,
"grad_norm": 6.29484748840332,
"learning_rate": 5.504016064257029e-06,
"loss": 0.2283,
"step": 56500
},
{
"epoch": 0.4528,
"grad_norm": 13.772027015686035,
"learning_rate": 5.495983935742973e-06,
"loss": 0.2433,
"step": 56600
},
{
"epoch": 0.4536,
"grad_norm": 6.613615989685059,
"learning_rate": 5.487951807228917e-06,
"loss": 0.2634,
"step": 56700
},
{
"epoch": 0.4544,
"grad_norm": 5.8526082038879395,
"learning_rate": 5.480000000000001e-06,
"loss": 0.2277,
"step": 56800
},
{
"epoch": 0.4552,
"grad_norm": 6.082682132720947,
"learning_rate": 5.471967871485945e-06,
"loss": 0.2372,
"step": 56900
},
{
"epoch": 0.456,
"grad_norm": 4.50082540512085,
"learning_rate": 5.463935742971888e-06,
"loss": 0.2578,
"step": 57000
},
{
"epoch": 0.4568,
"grad_norm": 5.06675910949707,
"learning_rate": 5.455903614457832e-06,
"loss": 0.2033,
"step": 57100
},
{
"epoch": 0.4576,
"grad_norm": 3.3439102172851562,
"learning_rate": 5.447871485943776e-06,
"loss": 0.2395,
"step": 57200
},
{
"epoch": 0.4584,
"grad_norm": 12.235926628112793,
"learning_rate": 5.43983935742972e-06,
"loss": 0.2335,
"step": 57300
},
{
"epoch": 0.4592,
"grad_norm": 6.169180393218994,
"learning_rate": 5.431807228915663e-06,
"loss": 0.2432,
"step": 57400
},
{
"epoch": 0.46,
"grad_norm": 3.7379937171936035,
"learning_rate": 5.423775100401607e-06,
"loss": 0.2377,
"step": 57500
},
{
"epoch": 0.4608,
"grad_norm": 6.2007060050964355,
"learning_rate": 5.4157429718875506e-06,
"loss": 0.2475,
"step": 57600
},
{
"epoch": 0.4616,
"grad_norm": 11.921826362609863,
"learning_rate": 5.4077108433734945e-06,
"loss": 0.2359,
"step": 57700
},
{
"epoch": 0.4624,
"grad_norm": 4.8144612312316895,
"learning_rate": 5.399678714859438e-06,
"loss": 0.2222,
"step": 57800
},
{
"epoch": 0.4632,
"grad_norm": 7.309458255767822,
"learning_rate": 5.3916465863453815e-06,
"loss": 0.2315,
"step": 57900
},
{
"epoch": 0.464,
"grad_norm": 6.312740325927734,
"learning_rate": 5.383614457831325e-06,
"loss": 0.2557,
"step": 58000
},
{
"epoch": 0.4648,
"grad_norm": 7.315298557281494,
"learning_rate": 5.375582329317269e-06,
"loss": 0.2673,
"step": 58100
},
{
"epoch": 0.4656,
"grad_norm": 9.65597152709961,
"learning_rate": 5.367550200803213e-06,
"loss": 0.2417,
"step": 58200
},
{
"epoch": 0.4664,
"grad_norm": 5.868946552276611,
"learning_rate": 5.359518072289157e-06,
"loss": 0.2289,
"step": 58300
},
{
"epoch": 0.4672,
"grad_norm": 7.116505146026611,
"learning_rate": 5.351485943775101e-06,
"loss": 0.2238,
"step": 58400
},
{
"epoch": 0.468,
"grad_norm": 4.75971794128418,
"learning_rate": 5.343453815261045e-06,
"loss": 0.2466,
"step": 58500
},
{
"epoch": 0.4688,
"grad_norm": 5.3648762702941895,
"learning_rate": 5.335421686746989e-06,
"loss": 0.2539,
"step": 58600
},
{
"epoch": 0.4696,
"grad_norm": 2.8886609077453613,
"learning_rate": 5.327389558232933e-06,
"loss": 0.2386,
"step": 58700
},
{
"epoch": 0.4704,
"grad_norm": 19.657976150512695,
"learning_rate": 5.319357429718875e-06,
"loss": 0.2257,
"step": 58800
},
{
"epoch": 0.4712,
"grad_norm": 6.010655879974365,
"learning_rate": 5.311325301204819e-06,
"loss": 0.2402,
"step": 58900
},
{
"epoch": 0.472,
"grad_norm": 11.614595413208008,
"learning_rate": 5.303293172690763e-06,
"loss": 0.2422,
"step": 59000
},
{
"epoch": 0.4728,
"grad_norm": 9.773826599121094,
"learning_rate": 5.295261044176707e-06,
"loss": 0.2224,
"step": 59100
},
{
"epoch": 0.4736,
"grad_norm": 3.5956804752349854,
"learning_rate": 5.287228915662651e-06,
"loss": 0.2287,
"step": 59200
},
{
"epoch": 0.4744,
"grad_norm": 5.884477615356445,
"learning_rate": 5.279196787148595e-06,
"loss": 0.2284,
"step": 59300
},
{
"epoch": 0.4752,
"grad_norm": 6.785783290863037,
"learning_rate": 5.271164658634539e-06,
"loss": 0.2486,
"step": 59400
},
{
"epoch": 0.476,
"grad_norm": 5.1648054122924805,
"learning_rate": 5.2631325301204825e-06,
"loss": 0.2352,
"step": 59500
},
{
"epoch": 0.4768,
"grad_norm": 7.0386643409729,
"learning_rate": 5.255180722891566e-06,
"loss": 0.2489,
"step": 59600
},
{
"epoch": 0.4776,
"grad_norm": 8.528487205505371,
"learning_rate": 5.24714859437751e-06,
"loss": 0.2434,
"step": 59700
},
{
"epoch": 0.4784,
"grad_norm": 10.057316780090332,
"learning_rate": 5.239116465863454e-06,
"loss": 0.232,
"step": 59800
},
{
"epoch": 0.4792,
"grad_norm": 5.452738285064697,
"learning_rate": 5.231084337349398e-06,
"loss": 0.2138,
"step": 59900
},
{
"epoch": 0.48,
"grad_norm": 12.630691528320312,
"learning_rate": 5.223052208835342e-06,
"loss": 0.2406,
"step": 60000
},
{
"epoch": 0.48,
"eval_test1_cer": 0.047455000373440887,
"eval_test1_cer_norm": 0.03161812375441208,
"eval_test1_loss": 0.18266192078590393,
"eval_test1_runtime": 3355.1591,
"eval_test1_samples_per_second": 0.745,
"eval_test1_steps_per_second": 0.186,
"eval_test1_wer": 0.1444065424647949,
"eval_test1_wer_norm": 0.08430988632046524,
"step": 60000
},
{
"epoch": 0.48,
"eval_test2_cer": 0.11545245828200246,
"eval_test2_cer_norm": 0.0919003718624109,
"eval_test2_loss": 0.32027342915534973,
"eval_test2_runtime": 3441.9039,
"eval_test2_samples_per_second": 0.726,
"eval_test2_steps_per_second": 0.182,
"eval_test2_wer": 0.24284733348592355,
"eval_test2_wer_norm": 0.17731492092596837,
"step": 60000
},
{
"epoch": 0.4808,
"grad_norm": 1.9925609827041626,
"learning_rate": 5.2150200803212856e-06,
"loss": 0.2353,
"step": 60100
},
{
"epoch": 0.4816,
"grad_norm": 5.7619476318359375,
"learning_rate": 5.2069879518072295e-06,
"loss": 0.2124,
"step": 60200
},
{
"epoch": 0.4824,
"grad_norm": 6.040704727172852,
"learning_rate": 5.198955823293173e-06,
"loss": 0.2304,
"step": 60300
},
{
"epoch": 0.4832,
"grad_norm": 14.028429985046387,
"learning_rate": 5.1909236947791165e-06,
"loss": 0.2549,
"step": 60400
},
{
"epoch": 0.484,
"grad_norm": 14.17192268371582,
"learning_rate": 5.18289156626506e-06,
"loss": 0.2215,
"step": 60500
},
{
"epoch": 0.4848,
"grad_norm": 8.68266773223877,
"learning_rate": 5.174859437751004e-06,
"loss": 0.2144,
"step": 60600
},
{
"epoch": 0.4856,
"grad_norm": 7.074737548828125,
"learning_rate": 5.166827309236948e-06,
"loss": 0.2262,
"step": 60700
},
{
"epoch": 0.4864,
"grad_norm": 7.095781326293945,
"learning_rate": 5.158795180722892e-06,
"loss": 0.2509,
"step": 60800
},
{
"epoch": 0.4872,
"grad_norm": 14.873151779174805,
"learning_rate": 5.150763052208836e-06,
"loss": 0.2417,
"step": 60900
},
{
"epoch": 0.488,
"grad_norm": 7.113455295562744,
"learning_rate": 5.14273092369478e-06,
"loss": 0.2232,
"step": 61000
},
{
"epoch": 0.4888,
"grad_norm": 4.443511009216309,
"learning_rate": 5.134698795180724e-06,
"loss": 0.2333,
"step": 61100
},
{
"epoch": 0.4896,
"grad_norm": 11.819351196289062,
"learning_rate": 5.126666666666668e-06,
"loss": 0.2252,
"step": 61200
},
{
"epoch": 0.4904,
"grad_norm": 9.132615089416504,
"learning_rate": 5.11863453815261e-06,
"loss": 0.2469,
"step": 61300
},
{
"epoch": 0.4912,
"grad_norm": 5.749680042266846,
"learning_rate": 5.110602409638554e-06,
"loss": 0.2322,
"step": 61400
},
{
"epoch": 0.492,
"grad_norm": 5.154579162597656,
"learning_rate": 5.102570281124498e-06,
"loss": 0.2293,
"step": 61500
},
{
"epoch": 0.4928,
"grad_norm": 4.018552780151367,
"learning_rate": 5.094538152610442e-06,
"loss": 0.2087,
"step": 61600
},
{
"epoch": 0.4936,
"grad_norm": 12.570889472961426,
"learning_rate": 5.086506024096386e-06,
"loss": 0.2374,
"step": 61700
},
{
"epoch": 0.4944,
"grad_norm": 4.278166770935059,
"learning_rate": 5.07847389558233e-06,
"loss": 0.2298,
"step": 61800
},
{
"epoch": 1.000696,
"grad_norm": 4.555546760559082,
"learning_rate": 5.070441767068274e-06,
"loss": 0.2077,
"step": 61900
},
{
"epoch": 1.001496,
"grad_norm": 15.487360954284668,
"learning_rate": 5.0624096385542175e-06,
"loss": 0.2044,
"step": 62000
},
{
"epoch": 1.002296,
"grad_norm": 4.297917366027832,
"learning_rate": 5.0543775100401614e-06,
"loss": 0.1977,
"step": 62100
},
{
"epoch": 1.003096,
"grad_norm": 8.327834129333496,
"learning_rate": 5.0463453815261045e-06,
"loss": 0.1749,
"step": 62200
},
{
"epoch": 1.003896,
"grad_norm": 8.811240196228027,
"learning_rate": 5.038313253012048e-06,
"loss": 0.1868,
"step": 62300
},
{
"epoch": 1.004696,
"grad_norm": 5.721954822540283,
"learning_rate": 5.030281124497992e-06,
"loss": 0.195,
"step": 62400
},
{
"epoch": 1.005496,
"grad_norm": 7.047698020935059,
"learning_rate": 5.022248995983936e-06,
"loss": 0.1605,
"step": 62500
},
{
"epoch": 1.006296,
"grad_norm": 6.009364128112793,
"learning_rate": 5.0142971887550206e-06,
"loss": 0.176,
"step": 62600
},
{
"epoch": 1.007096,
"grad_norm": 5.495716571807861,
"learning_rate": 5.0062650602409645e-06,
"loss": 0.1991,
"step": 62700
},
{
"epoch": 1.007896,
"grad_norm": 2.433659076690674,
"learning_rate": 4.9982329317269076e-06,
"loss": 0.1832,
"step": 62800
},
{
"epoch": 1.008696,
"grad_norm": 6.076290607452393,
"learning_rate": 4.9902008032128515e-06,
"loss": 0.1807,
"step": 62900
},
{
"epoch": 1.009496,
"grad_norm": 5.725612640380859,
"learning_rate": 4.982168674698795e-06,
"loss": 0.1627,
"step": 63000
},
{
"epoch": 1.010296,
"grad_norm": 4.852512836456299,
"learning_rate": 4.974136546184739e-06,
"loss": 0.1692,
"step": 63100
},
{
"epoch": 1.011096,
"grad_norm": 6.572232246398926,
"learning_rate": 4.966104417670683e-06,
"loss": 0.1616,
"step": 63200
},
{
"epoch": 1.011896,
"grad_norm": 9.408519744873047,
"learning_rate": 4.958072289156627e-06,
"loss": 0.1566,
"step": 63300
},
{
"epoch": 1.012696,
"grad_norm": 4.744565963745117,
"learning_rate": 4.950040160642571e-06,
"loss": 0.1519,
"step": 63400
},
{
"epoch": 1.013496,
"grad_norm": 13.262666702270508,
"learning_rate": 4.942008032128515e-06,
"loss": 0.1804,
"step": 63500
},
{
"epoch": 1.014296,
"grad_norm": 3.7456908226013184,
"learning_rate": 4.933975903614458e-06,
"loss": 0.1703,
"step": 63600
},
{
"epoch": 1.015096,
"grad_norm": 6.4614152908325195,
"learning_rate": 4.925943775100402e-06,
"loss": 0.1796,
"step": 63700
},
{
"epoch": 1.015896,
"grad_norm": 4.488025665283203,
"learning_rate": 4.917911646586346e-06,
"loss": 0.1678,
"step": 63800
},
{
"epoch": 1.016696,
"grad_norm": 5.777276992797852,
"learning_rate": 4.90987951807229e-06,
"loss": 0.1839,
"step": 63900
},
{
"epoch": 1.017496,
"grad_norm": 4.160243034362793,
"learning_rate": 4.901847389558234e-06,
"loss": 0.1627,
"step": 64000
},
{
"epoch": 1.017496,
"eval_test1_cer": 0.07008084995145268,
"eval_test1_cer_norm": 0.050106850433404564,
"eval_test1_loss": 0.18447566032409668,
"eval_test1_runtime": 3416.4242,
"eval_test1_samples_per_second": 0.732,
"eval_test1_steps_per_second": 0.183,
"eval_test1_wer": 0.17367853290183388,
"eval_test1_wer_norm": 0.11408866419240772,
"step": 64000
},
{
"epoch": 1.017496,
"eval_test2_cer": 0.14694627991189757,
"eval_test2_cer_norm": 0.11825515184381778,
"eval_test2_loss": 0.3210515081882477,
"eval_test2_runtime": 3532.7138,
"eval_test2_samples_per_second": 0.708,
"eval_test2_steps_per_second": 0.177,
"eval_test2_wer": 0.2882810711833371,
"eval_test2_wer_norm": 0.22352738941095576,
"step": 64000
},
{
"epoch": 1.018296,
"grad_norm": 5.366184234619141,
"learning_rate": 4.893815261044177e-06,
"loss": 0.1663,
"step": 64100
},
{
"epoch": 1.019096,
"grad_norm": 3.319096326828003,
"learning_rate": 4.885783132530121e-06,
"loss": 0.1597,
"step": 64200
},
{
"epoch": 1.019896,
"grad_norm": 2.7266533374786377,
"learning_rate": 4.877751004016065e-06,
"loss": 0.1636,
"step": 64300
},
{
"epoch": 1.020696,
"grad_norm": 5.649815082550049,
"learning_rate": 4.869718875502009e-06,
"loss": 0.1719,
"step": 64400
},
{
"epoch": 1.021496,
"grad_norm": 9.476170539855957,
"learning_rate": 4.861686746987952e-06,
"loss": 0.1612,
"step": 64500
},
{
"epoch": 1.022296,
"grad_norm": 6.443653106689453,
"learning_rate": 4.853654618473896e-06,
"loss": 0.1589,
"step": 64600
},
{
"epoch": 1.023096,
"grad_norm": 4.278164863586426,
"learning_rate": 4.8456224899598395e-06,
"loss": 0.1575,
"step": 64700
},
{
"epoch": 1.023896,
"grad_norm": 7.513009071350098,
"learning_rate": 4.837590361445783e-06,
"loss": 0.1678,
"step": 64800
},
{
"epoch": 1.024696,
"grad_norm": 7.260671615600586,
"learning_rate": 4.829558232931727e-06,
"loss": 0.1605,
"step": 64900
},
{
"epoch": 1.025496,
"grad_norm": 5.455900192260742,
"learning_rate": 4.821526104417671e-06,
"loss": 0.1555,
"step": 65000
},
{
"epoch": 1.026296,
"grad_norm": 11.086517333984375,
"learning_rate": 4.813493975903615e-06,
"loss": 0.165,
"step": 65100
},
{
"epoch": 1.027096,
"grad_norm": 4.682340145111084,
"learning_rate": 4.805461847389558e-06,
"loss": 0.1791,
"step": 65200
},
{
"epoch": 1.027896,
"grad_norm": 8.538371086120605,
"learning_rate": 4.797429718875502e-06,
"loss": 0.1656,
"step": 65300
},
{
"epoch": 1.028696,
"grad_norm": 6.110259532928467,
"learning_rate": 4.789397590361446e-06,
"loss": 0.1564,
"step": 65400
},
{
"epoch": 1.029496,
"grad_norm": 6.267505168914795,
"learning_rate": 4.78136546184739e-06,
"loss": 0.1624,
"step": 65500
},
{
"epoch": 1.030296,
"grad_norm": 7.841426372528076,
"learning_rate": 4.773333333333334e-06,
"loss": 0.1497,
"step": 65600
},
{
"epoch": 1.031096,
"grad_norm": 3.335782051086426,
"learning_rate": 4.765301204819278e-06,
"loss": 0.1684,
"step": 65700
},
{
"epoch": 1.031896,
"grad_norm": 6.293437480926514,
"learning_rate": 4.757269076305222e-06,
"loss": 0.1687,
"step": 65800
},
{
"epoch": 1.032696,
"grad_norm": 9.889056205749512,
"learning_rate": 4.749236947791165e-06,
"loss": 0.1605,
"step": 65900
},
{
"epoch": 1.033496,
"grad_norm": 7.943349838256836,
"learning_rate": 4.741204819277109e-06,
"loss": 0.1781,
"step": 66000
},
{
"epoch": 1.034296,
"grad_norm": 9.100912094116211,
"learning_rate": 4.733172690763053e-06,
"loss": 0.1789,
"step": 66100
},
{
"epoch": 1.035096,
"grad_norm": 5.020230293273926,
"learning_rate": 4.725140562248997e-06,
"loss": 0.1679,
"step": 66200
},
{
"epoch": 1.035896,
"grad_norm": 3.7011711597442627,
"learning_rate": 4.71710843373494e-06,
"loss": 0.1617,
"step": 66300
},
{
"epoch": 1.036696,
"grad_norm": 13.44151782989502,
"learning_rate": 4.709076305220884e-06,
"loss": 0.1717,
"step": 66400
},
{
"epoch": 1.037496,
"grad_norm": 5.3160319328308105,
"learning_rate": 4.7010441767068275e-06,
"loss": 0.18,
"step": 66500
},
{
"epoch": 1.0382959999999999,
"grad_norm": 9.991889953613281,
"learning_rate": 4.693092369477912e-06,
"loss": 0.1698,
"step": 66600
},
{
"epoch": 1.039096,
"grad_norm": 5.639313697814941,
"learning_rate": 4.685060240963856e-06,
"loss": 0.1595,
"step": 66700
},
{
"epoch": 1.039896,
"grad_norm": 4.916014194488525,
"learning_rate": 4.6770281124498e-06,
"loss": 0.1703,
"step": 66800
},
{
"epoch": 1.040696,
"grad_norm": 5.738508701324463,
"learning_rate": 4.668995983935744e-06,
"loss": 0.1864,
"step": 66900
},
{
"epoch": 1.041496,
"grad_norm": 8.484038352966309,
"learning_rate": 4.6609638554216875e-06,
"loss": 0.1587,
"step": 67000
},
{
"epoch": 1.042296,
"grad_norm": 6.696768283843994,
"learning_rate": 4.652931726907631e-06,
"loss": 0.1601,
"step": 67100
},
{
"epoch": 1.043096,
"grad_norm": 3.6157071590423584,
"learning_rate": 4.6448995983935745e-06,
"loss": 0.1657,
"step": 67200
},
{
"epoch": 1.043896,
"grad_norm": 6.275000095367432,
"learning_rate": 4.6368674698795184e-06,
"loss": 0.1559,
"step": 67300
},
{
"epoch": 1.044696,
"grad_norm": 11.618382453918457,
"learning_rate": 4.628835341365462e-06,
"loss": 0.1636,
"step": 67400
},
{
"epoch": 1.045496,
"grad_norm": 3.8046157360076904,
"learning_rate": 4.620803212851405e-06,
"loss": 0.165,
"step": 67500
},
{
"epoch": 1.046296,
"grad_norm": 7.296670913696289,
"learning_rate": 4.612771084337349e-06,
"loss": 0.1692,
"step": 67600
},
{
"epoch": 1.047096,
"grad_norm": 5.143111228942871,
"learning_rate": 4.604738955823293e-06,
"loss": 0.1567,
"step": 67700
},
{
"epoch": 1.047896,
"grad_norm": 6.818264961242676,
"learning_rate": 4.596706827309237e-06,
"loss": 0.1569,
"step": 67800
},
{
"epoch": 1.048696,
"grad_norm": 3.8615856170654297,
"learning_rate": 4.588674698795181e-06,
"loss": 0.1646,
"step": 67900
},
{
"epoch": 1.049496,
"grad_norm": 4.350166320800781,
"learning_rate": 4.580642570281125e-06,
"loss": 0.1562,
"step": 68000
},
{
"epoch": 1.049496,
"eval_test1_cer": 0.04518167899021585,
"eval_test1_cer_norm": 0.02964919441976613,
"eval_test1_loss": 0.18461571633815765,
"eval_test1_runtime": 3355.9439,
"eval_test1_samples_per_second": 0.745,
"eval_test1_steps_per_second": 0.186,
"eval_test1_wer": 0.1442899209889501,
"eval_test1_wer_norm": 0.08173821561121014,
"step": 68000
},
{
"epoch": 1.049496,
"eval_test2_cer": 0.08435995072236532,
"eval_test2_cer_norm": 0.06550201425472575,
"eval_test2_loss": 0.3253004252910614,
"eval_test2_runtime": 3365.7371,
"eval_test2_samples_per_second": 0.743,
"eval_test2_steps_per_second": 0.186,
"eval_test2_wer": 0.19904440375371937,
"eval_test2_wer_norm": 0.13302200320880128,
"step": 68000
},
{
"epoch": 1.050296,
"grad_norm": 7.118165016174316,
"learning_rate": 4.572610441767069e-06,
"loss": 0.1473,
"step": 68100
},
{
"epoch": 1.051096,
"grad_norm": 7.220821857452393,
"learning_rate": 4.564578313253013e-06,
"loss": 0.1617,
"step": 68200
},
{
"epoch": 1.051896,
"grad_norm": 7.148895740509033,
"learning_rate": 4.556546184738957e-06,
"loss": 0.1629,
"step": 68300
},
{
"epoch": 1.052696,
"grad_norm": 3.8178248405456543,
"learning_rate": 4.5485140562249e-06,
"loss": 0.181,
"step": 68400
},
{
"epoch": 1.053496,
"grad_norm": 6.640100479125977,
"learning_rate": 4.540481927710844e-06,
"loss": 0.1545,
"step": 68500
},
{
"epoch": 1.054296,
"grad_norm": 6.102357387542725,
"learning_rate": 4.532530120481928e-06,
"loss": 0.1716,
"step": 68600
},
{
"epoch": 1.055096,
"grad_norm": 5.749415397644043,
"learning_rate": 4.524497991967872e-06,
"loss": 0.1419,
"step": 68700
},
{
"epoch": 1.055896,
"grad_norm": 5.71680212020874,
"learning_rate": 4.516465863453816e-06,
"loss": 0.1487,
"step": 68800
},
{
"epoch": 1.056696,
"grad_norm": 5.209336757659912,
"learning_rate": 4.50843373493976e-06,
"loss": 0.1722,
"step": 68900
},
{
"epoch": 1.057496,
"grad_norm": 6.00562047958374,
"learning_rate": 4.500401606425703e-06,
"loss": 0.1659,
"step": 69000
},
{
"epoch": 1.058296,
"grad_norm": 4.806443214416504,
"learning_rate": 4.492369477911647e-06,
"loss": 0.1499,
"step": 69100
},
{
"epoch": 1.059096,
"grad_norm": 7.675013065338135,
"learning_rate": 4.484337349397591e-06,
"loss": 0.1624,
"step": 69200
},
{
"epoch": 1.059896,
"grad_norm": 5.567787170410156,
"learning_rate": 4.476305220883535e-06,
"loss": 0.1572,
"step": 69300
},
{
"epoch": 1.060696,
"grad_norm": 20.12483787536621,
"learning_rate": 4.468273092369479e-06,
"loss": 0.169,
"step": 69400
},
{
"epoch": 1.061496,
"grad_norm": 11.066375732421875,
"learning_rate": 4.4602409638554225e-06,
"loss": 0.1618,
"step": 69500
},
{
"epoch": 1.062296,
"grad_norm": 6.3260579109191895,
"learning_rate": 4.452208835341366e-06,
"loss": 0.1465,
"step": 69600
},
{
"epoch": 1.063096,
"grad_norm": 5.906326770782471,
"learning_rate": 4.4441767068273095e-06,
"loss": 0.1532,
"step": 69700
},
{
"epoch": 1.063896,
"grad_norm": 4.177192211151123,
"learning_rate": 4.4361445783132534e-06,
"loss": 0.1624,
"step": 69800
},
{
"epoch": 1.064696,
"grad_norm": 4.535795211791992,
"learning_rate": 4.428112449799197e-06,
"loss": 0.1637,
"step": 69900
},
{
"epoch": 1.065496,
"grad_norm": 8.556774139404297,
"learning_rate": 4.420080321285141e-06,
"loss": 0.1512,
"step": 70000
},
{
"epoch": 1.066296,
"grad_norm": 5.175086975097656,
"learning_rate": 4.412048192771084e-06,
"loss": 0.1487,
"step": 70100
},
{
"epoch": 1.067096,
"grad_norm": 5.598880767822266,
"learning_rate": 4.404016064257028e-06,
"loss": 0.1604,
"step": 70200
},
{
"epoch": 1.067896,
"grad_norm": 4.312199115753174,
"learning_rate": 4.395983935742972e-06,
"loss": 0.1752,
"step": 70300
},
{
"epoch": 1.068696,
"grad_norm": 4.8355255126953125,
"learning_rate": 4.387951807228916e-06,
"loss": 0.1375,
"step": 70400
},
{
"epoch": 1.069496,
"grad_norm": 0.6319503784179688,
"learning_rate": 4.379919678714859e-06,
"loss": 0.1518,
"step": 70500
},
{
"epoch": 1.070296,
"grad_norm": 7.824057579040527,
"learning_rate": 4.371887550200803e-06,
"loss": 0.171,
"step": 70600
},
{
"epoch": 1.071096,
"grad_norm": 6.540349960327148,
"learning_rate": 4.363855421686747e-06,
"loss": 0.146,
"step": 70700
},
{
"epoch": 1.071896,
"grad_norm": 11.684809684753418,
"learning_rate": 4.355823293172691e-06,
"loss": 0.1504,
"step": 70800
},
{
"epoch": 1.072696,
"grad_norm": 9.533754348754883,
"learning_rate": 4.347791164658635e-06,
"loss": 0.1548,
"step": 70900
},
{
"epoch": 1.073496,
"grad_norm": 6.979349136352539,
"learning_rate": 4.339759036144579e-06,
"loss": 0.1608,
"step": 71000
},
{
"epoch": 1.074296,
"grad_norm": 6.119946002960205,
"learning_rate": 4.331807228915663e-06,
"loss": 0.1642,
"step": 71100
},
{
"epoch": 1.075096,
"grad_norm": 5.234997272491455,
"learning_rate": 4.323775100401606e-06,
"loss": 0.159,
"step": 71200
},
{
"epoch": 1.075896,
"grad_norm": 11.375402450561523,
"learning_rate": 4.31574297188755e-06,
"loss": 0.1598,
"step": 71300
},
{
"epoch": 1.076696,
"grad_norm": 6.117208003997803,
"learning_rate": 4.307710843373494e-06,
"loss": 0.1581,
"step": 71400
},
{
"epoch": 1.077496,
"grad_norm": 5.760785102844238,
"learning_rate": 4.299678714859438e-06,
"loss": 0.1581,
"step": 71500
},
{
"epoch": 1.078296,
"grad_norm": 12.158953666687012,
"learning_rate": 4.291646586345382e-06,
"loss": 0.1645,
"step": 71600
},
{
"epoch": 1.079096,
"grad_norm": 3.9681832790374756,
"learning_rate": 4.283614457831326e-06,
"loss": 0.1664,
"step": 71700
},
{
"epoch": 1.079896,
"grad_norm": 4.026464939117432,
"learning_rate": 4.27558232931727e-06,
"loss": 0.1478,
"step": 71800
},
{
"epoch": 1.080696,
"grad_norm": 9.842916488647461,
"learning_rate": 4.267550200803214e-06,
"loss": 0.1516,
"step": 71900
},
{
"epoch": 1.081496,
"grad_norm": 6.5888776779174805,
"learning_rate": 4.2595180722891575e-06,
"loss": 0.1743,
"step": 72000
},
{
"epoch": 1.081496,
"eval_test1_cer": 0.056235529165733066,
"eval_test1_cer_norm": 0.03523423055682282,
"eval_test1_loss": 0.18351316452026367,
"eval_test1_runtime": 3389.512,
"eval_test1_samples_per_second": 0.738,
"eval_test1_steps_per_second": 0.184,
"eval_test1_wer": 0.15283244409458002,
"eval_test1_wer_norm": 0.09255092200239633,
"step": 72000
},
{
"epoch": 1.081496,
"eval_test2_cer": 0.10279239929816703,
"eval_test2_cer_norm": 0.0767353579175705,
"eval_test2_loss": 0.32321926951408386,
"eval_test2_runtime": 3400.9588,
"eval_test2_samples_per_second": 0.735,
"eval_test2_steps_per_second": 0.184,
"eval_test2_wer": 0.21498054474708173,
"eval_test2_wer_norm": 0.14843570937428374,
"step": 72000
},
{
"epoch": 1.082296,
"grad_norm": 11.681933403015137,
"learning_rate": 4.251485943775101e-06,
"loss": 0.1564,
"step": 72100
},
{
"epoch": 1.083096,
"grad_norm": 5.162091255187988,
"learning_rate": 4.2434538152610445e-06,
"loss": 0.1469,
"step": 72200
},
{
"epoch": 1.083896,
"grad_norm": 7.147501468658447,
"learning_rate": 4.2354216867469884e-06,
"loss": 0.1663,
"step": 72300
},
{
"epoch": 1.084696,
"grad_norm": 7.40344762802124,
"learning_rate": 4.227389558232932e-06,
"loss": 0.1284,
"step": 72400
},
{
"epoch": 1.085496,
"grad_norm": 3.5983896255493164,
"learning_rate": 4.219357429718876e-06,
"loss": 0.1558,
"step": 72500
},
{
"epoch": 1.086296,
"grad_norm": 4.397582054138184,
"learning_rate": 4.211325301204819e-06,
"loss": 0.1599,
"step": 72600
},
{
"epoch": 1.087096,
"grad_norm": 2.2897703647613525,
"learning_rate": 4.203293172690763e-06,
"loss": 0.1611,
"step": 72700
},
{
"epoch": 1.087896,
"grad_norm": 4.20655632019043,
"learning_rate": 4.195261044176707e-06,
"loss": 0.1597,
"step": 72800
},
{
"epoch": 1.088696,
"grad_norm": 4.474186420440674,
"learning_rate": 4.187228915662651e-06,
"loss": 0.145,
"step": 72900
},
{
"epoch": 1.089496,
"grad_norm": 4.5594401359558105,
"learning_rate": 4.179196787148594e-06,
"loss": 0.1545,
"step": 73000
},
{
"epoch": 1.090296,
"grad_norm": 3.502408266067505,
"learning_rate": 4.171164658634538e-06,
"loss": 0.1648,
"step": 73100
},
{
"epoch": 1.091096,
"grad_norm": 2.6792593002319336,
"learning_rate": 4.163132530120482e-06,
"loss": 0.1491,
"step": 73200
},
{
"epoch": 1.091896,
"grad_norm": 6.028800010681152,
"learning_rate": 4.155100401606426e-06,
"loss": 0.1544,
"step": 73300
},
{
"epoch": 1.0926960000000001,
"grad_norm": 5.256026268005371,
"learning_rate": 4.14706827309237e-06,
"loss": 0.1526,
"step": 73400
},
{
"epoch": 1.093496,
"grad_norm": 5.323427677154541,
"learning_rate": 4.139036144578314e-06,
"loss": 0.1412,
"step": 73500
},
{
"epoch": 1.094296,
"grad_norm": 3.4750475883483887,
"learning_rate": 4.131004016064257e-06,
"loss": 0.1372,
"step": 73600
},
{
"epoch": 1.095096,
"grad_norm": 2.8323254585266113,
"learning_rate": 4.122971887550201e-06,
"loss": 0.156,
"step": 73700
},
{
"epoch": 1.095896,
"grad_norm": 11.332598686218262,
"learning_rate": 4.114939759036145e-06,
"loss": 0.1443,
"step": 73800
},
{
"epoch": 1.0966960000000001,
"grad_norm": 3.0766139030456543,
"learning_rate": 4.106907630522089e-06,
"loss": 0.1513,
"step": 73900
},
{
"epoch": 1.097496,
"grad_norm": 8.13890266418457,
"learning_rate": 4.0988755020080325e-06,
"loss": 0.1441,
"step": 74000
},
{
"epoch": 1.098296,
"grad_norm": 6.174729824066162,
"learning_rate": 4.0908433734939765e-06,
"loss": 0.1465,
"step": 74100
},
{
"epoch": 1.099096,
"grad_norm": 7.782393455505371,
"learning_rate": 4.08281124497992e-06,
"loss": 0.1525,
"step": 74200
},
{
"epoch": 1.099896,
"grad_norm": 4.48836088180542,
"learning_rate": 4.074779116465864e-06,
"loss": 0.1419,
"step": 74300
},
{
"epoch": 1.100696,
"grad_norm": 9.527304649353027,
"learning_rate": 4.066746987951807e-06,
"loss": 0.1471,
"step": 74400
},
{
"epoch": 1.101496,
"grad_norm": 5.195642471313477,
"learning_rate": 4.058714859437751e-06,
"loss": 0.1454,
"step": 74500
},
{
"epoch": 1.102296,
"grad_norm": 4.529500961303711,
"learning_rate": 4.050682730923695e-06,
"loss": 0.1582,
"step": 74600
},
{
"epoch": 1.103096,
"grad_norm": 4.841000080108643,
"learning_rate": 4.042650602409639e-06,
"loss": 0.1521,
"step": 74700
},
{
"epoch": 1.103896,
"grad_norm": 4.943587303161621,
"learning_rate": 4.034618473895583e-06,
"loss": 0.148,
"step": 74800
},
{
"epoch": 1.104696,
"grad_norm": 7.331528186798096,
"learning_rate": 4.026586345381526e-06,
"loss": 0.1469,
"step": 74900
},
{
"epoch": 1.105496,
"grad_norm": 8.253586769104004,
"learning_rate": 4.01855421686747e-06,
"loss": 0.1736,
"step": 75000
},
{
"epoch": 1.106296,
"grad_norm": 6.348284721374512,
"learning_rate": 4.010602409638554e-06,
"loss": 0.1398,
"step": 75100
},
{
"epoch": 1.107096,
"grad_norm": 2.285827398300171,
"learning_rate": 4.002570281124498e-06,
"loss": 0.146,
"step": 75200
},
{
"epoch": 1.107896,
"grad_norm": 7.164559364318848,
"learning_rate": 3.994538152610442e-06,
"loss": 0.1564,
"step": 75300
},
{
"epoch": 1.108696,
"grad_norm": 2.311056137084961,
"learning_rate": 3.986506024096386e-06,
"loss": 0.1509,
"step": 75400
},
{
"epoch": 1.109496,
"grad_norm": 3.1986663341522217,
"learning_rate": 3.97847389558233e-06,
"loss": 0.1521,
"step": 75500
},
{
"epoch": 1.110296,
"grad_norm": 10.752972602844238,
"learning_rate": 3.970441767068273e-06,
"loss": 0.1321,
"step": 75600
},
{
"epoch": 1.111096,
"grad_norm": 5.435462474822998,
"learning_rate": 3.962409638554217e-06,
"loss": 0.1413,
"step": 75700
},
{
"epoch": 1.111896,
"grad_norm": 3.7016944885253906,
"learning_rate": 3.954457831325301e-06,
"loss": 0.1661,
"step": 75800
},
{
"epoch": 1.112696,
"grad_norm": 9.687053680419922,
"learning_rate": 3.946425702811245e-06,
"loss": 0.1516,
"step": 75900
},
{
"epoch": 1.113496,
"grad_norm": 2.699572801589966,
"learning_rate": 3.938393574297189e-06,
"loss": 0.153,
"step": 76000
},
{
"epoch": 1.113496,
"eval_test1_cer": 0.07078571962058407,
"eval_test1_cer_norm": 0.04619780536413187,
"eval_test1_loss": 0.18133017420768738,
"eval_test1_runtime": 3430.9909,
"eval_test1_samples_per_second": 0.729,
"eval_test1_steps_per_second": 0.182,
"eval_test1_wer": 0.16875127554739205,
"eval_test1_wer_norm": 0.11002659341301615,
"step": 76000
},
{
"epoch": 1.113496,
"eval_test2_cer": 0.13182233919438532,
"eval_test2_cer_norm": 0.09897931515339324,
"eval_test2_loss": 0.32256972789764404,
"eval_test2_runtime": 3529.0223,
"eval_test2_samples_per_second": 0.708,
"eval_test2_steps_per_second": 0.177,
"eval_test2_wer": 0.26012817578393227,
"eval_test2_wer_norm": 0.19547902819161128,
"step": 76000
},
{
"epoch": 1.114296,
"grad_norm": 4.699991703033447,
"learning_rate": 3.930361445783133e-06,
"loss": 0.1306,
"step": 76100
},
{
"epoch": 1.115096,
"grad_norm": 5.847876071929932,
"learning_rate": 3.922329317269077e-06,
"loss": 0.159,
"step": 76200
},
{
"epoch": 1.115896,
"grad_norm": 14.771410942077637,
"learning_rate": 3.91429718875502e-06,
"loss": 0.1593,
"step": 76300
},
{
"epoch": 1.116696,
"grad_norm": 3.190328598022461,
"learning_rate": 3.906265060240964e-06,
"loss": 0.1513,
"step": 76400
},
{
"epoch": 1.117496,
"grad_norm": 3.2370731830596924,
"learning_rate": 3.898232931726908e-06,
"loss": 0.1524,
"step": 76500
},
{
"epoch": 1.118296,
"grad_norm": 3.932393789291382,
"learning_rate": 3.890200803212852e-06,
"loss": 0.129,
"step": 76600
},
{
"epoch": 1.119096,
"grad_norm": 3.998230457305908,
"learning_rate": 3.882168674698796e-06,
"loss": 0.1534,
"step": 76700
},
{
"epoch": 1.119896,
"grad_norm": 5.405791282653809,
"learning_rate": 3.874136546184739e-06,
"loss": 0.1455,
"step": 76800
},
{
"epoch": 1.120696,
"grad_norm": 5.66414737701416,
"learning_rate": 3.866104417670683e-06,
"loss": 0.1321,
"step": 76900
},
{
"epoch": 1.121496,
"grad_norm": 2.428980827331543,
"learning_rate": 3.858072289156627e-06,
"loss": 0.1251,
"step": 77000
},
{
"epoch": 1.122296,
"grad_norm": 8.633134841918945,
"learning_rate": 3.850040160642571e-06,
"loss": 0.1581,
"step": 77100
},
{
"epoch": 1.123096,
"grad_norm": 6.84693717956543,
"learning_rate": 3.8420080321285145e-06,
"loss": 0.1516,
"step": 77200
},
{
"epoch": 1.123896,
"grad_norm": 1.603908896446228,
"learning_rate": 3.8339759036144584e-06,
"loss": 0.1614,
"step": 77300
},
{
"epoch": 1.124696,
"grad_norm": 12.588580131530762,
"learning_rate": 3.8259437751004015e-06,
"loss": 0.1423,
"step": 77400
},
{
"epoch": 1.125496,
"grad_norm": 8.18181324005127,
"learning_rate": 3.8179116465863454e-06,
"loss": 0.1409,
"step": 77500
},
{
"epoch": 1.126296,
"grad_norm": 4.0206217765808105,
"learning_rate": 3.8098795180722898e-06,
"loss": 0.1407,
"step": 77600
},
{
"epoch": 1.1270959999999999,
"grad_norm": 3.093519449234009,
"learning_rate": 3.8018473895582333e-06,
"loss": 0.128,
"step": 77700
},
{
"epoch": 1.127896,
"grad_norm": 5.9990925788879395,
"learning_rate": 3.7938152610441768e-06,
"loss": 0.138,
"step": 77800
},
{
"epoch": 1.128696,
"grad_norm": 5.512988567352295,
"learning_rate": 3.7857831325301207e-06,
"loss": 0.1366,
"step": 77900
},
{
"epoch": 1.129496,
"grad_norm": 2.798251152038574,
"learning_rate": 3.7777510040160646e-06,
"loss": 0.141,
"step": 78000
},
{
"epoch": 1.130296,
"grad_norm": 4.432553768157959,
"learning_rate": 3.769718875502008e-06,
"loss": 0.1415,
"step": 78100
},
{
"epoch": 1.1310959999999999,
"grad_norm": 4.363057613372803,
"learning_rate": 3.761686746987952e-06,
"loss": 0.1516,
"step": 78200
},
{
"epoch": 1.131896,
"grad_norm": 8.875186920166016,
"learning_rate": 3.753654618473896e-06,
"loss": 0.1578,
"step": 78300
},
{
"epoch": 1.132696,
"grad_norm": 5.948533535003662,
"learning_rate": 3.74562248995984e-06,
"loss": 0.1436,
"step": 78400
},
{
"epoch": 1.133496,
"grad_norm": 3.6585094928741455,
"learning_rate": 3.737590361445784e-06,
"loss": 0.1452,
"step": 78500
},
{
"epoch": 1.134296,
"grad_norm": 4.998322486877441,
"learning_rate": 3.729558232931727e-06,
"loss": 0.1395,
"step": 78600
},
{
"epoch": 1.1350959999999999,
"grad_norm": 2.1381640434265137,
"learning_rate": 3.7215261044176708e-06,
"loss": 0.1386,
"step": 78700
},
{
"epoch": 1.135896,
"grad_norm": 3.784778594970703,
"learning_rate": 3.7134939759036147e-06,
"loss": 0.1494,
"step": 78800
},
{
"epoch": 1.136696,
"grad_norm": 12.97223949432373,
"learning_rate": 3.7054618473895586e-06,
"loss": 0.1389,
"step": 78900
},
{
"epoch": 1.137496,
"grad_norm": 2.718700647354126,
"learning_rate": 3.697429718875502e-06,
"loss": 0.1488,
"step": 79000
},
{
"epoch": 1.138296,
"grad_norm": 5.164200782775879,
"learning_rate": 3.689397590361446e-06,
"loss": 0.1554,
"step": 79100
},
{
"epoch": 1.1390959999999999,
"grad_norm": 5.673985004425049,
"learning_rate": 3.68136546184739e-06,
"loss": 0.136,
"step": 79200
},
{
"epoch": 1.139896,
"grad_norm": 2.9805924892425537,
"learning_rate": 3.673333333333334e-06,
"loss": 0.1394,
"step": 79300
},
{
"epoch": 1.140696,
"grad_norm": 7.971452713012695,
"learning_rate": 3.665301204819278e-06,
"loss": 0.1515,
"step": 79400
},
{
"epoch": 1.141496,
"grad_norm": 3.0903499126434326,
"learning_rate": 3.657269076305221e-06,
"loss": 0.1378,
"step": 79500
},
{
"epoch": 1.142296,
"grad_norm": 4.741939067840576,
"learning_rate": 3.649236947791165e-06,
"loss": 0.1472,
"step": 79600
},
{
"epoch": 1.143096,
"grad_norm": 5.633648872375488,
"learning_rate": 3.6412048192771087e-06,
"loss": 0.1573,
"step": 79700
},
{
"epoch": 1.143896,
"grad_norm": 3.337674379348755,
"learning_rate": 3.633253012048193e-06,
"loss": 0.1531,
"step": 79800
},
{
"epoch": 1.144696,
"grad_norm": 15.155500411987305,
"learning_rate": 3.625220883534137e-06,
"loss": 0.1438,
"step": 79900
},
{
"epoch": 1.145496,
"grad_norm": 7.186131477355957,
"learning_rate": 3.617188755020081e-06,
"loss": 0.1315,
"step": 80000
},
{
"epoch": 1.145496,
"eval_test1_cer": 0.050181118828889384,
"eval_test1_cer_norm": 0.03300117655533412,
"eval_test1_loss": 0.17834880948066711,
"eval_test1_runtime": 3408.2771,
"eval_test1_samples_per_second": 0.734,
"eval_test1_steps_per_second": 0.183,
"eval_test1_wer": 0.1458643109128546,
"eval_test1_wer_norm": 0.08547882755194483,
"step": 80000
},
{
"epoch": 1.145496,
"eval_test2_cer": 0.10706219434800462,
"eval_test2_cer_norm": 0.08262802138208862,
"eval_test2_loss": 0.3213089108467102,
"eval_test2_runtime": 3477.8704,
"eval_test2_samples_per_second": 0.719,
"eval_test2_steps_per_second": 0.18,
"eval_test2_wer": 0.23108834973678188,
"eval_test2_wer_norm": 0.16551111620444647,
"step": 80000
},
{
"epoch": 1.146296,
"grad_norm": 5.269469738006592,
"learning_rate": 3.6091566265060248e-06,
"loss": 0.1244,
"step": 80100
},
{
"epoch": 1.147096,
"grad_norm": 5.666225433349609,
"learning_rate": 3.6012048192771087e-06,
"loss": 0.13,
"step": 80200
},
{
"epoch": 1.147896,
"grad_norm": 5.9655375480651855,
"learning_rate": 3.5931726907630526e-06,
"loss": 0.1464,
"step": 80300
},
{
"epoch": 1.148696,
"grad_norm": 2.792882204055786,
"learning_rate": 3.5851405622489965e-06,
"loss": 0.1482,
"step": 80400
},
{
"epoch": 1.149496,
"grad_norm": 2.3720922470092773,
"learning_rate": 3.57710843373494e-06,
"loss": 0.1375,
"step": 80500
},
{
"epoch": 1.150296,
"grad_norm": 3.8564910888671875,
"learning_rate": 3.569076305220884e-06,
"loss": 0.1291,
"step": 80600
},
{
"epoch": 1.151096,
"grad_norm": 5.648325443267822,
"learning_rate": 3.561044176706828e-06,
"loss": 0.1341,
"step": 80700
},
{
"epoch": 1.151896,
"grad_norm": 8.060567855834961,
"learning_rate": 3.5530120481927718e-06,
"loss": 0.1285,
"step": 80800
},
{
"epoch": 1.152696,
"grad_norm": 2.7208454608917236,
"learning_rate": 3.544979919678715e-06,
"loss": 0.1413,
"step": 80900
},
{
"epoch": 1.153496,
"grad_norm": 5.355711460113525,
"learning_rate": 3.5369477911646588e-06,
"loss": 0.1259,
"step": 81000
},
{
"epoch": 1.154296,
"grad_norm": 7.103275775909424,
"learning_rate": 3.5289156626506027e-06,
"loss": 0.149,
"step": 81100
},
{
"epoch": 1.155096,
"grad_norm": 3.506634473800659,
"learning_rate": 3.5208835341365466e-06,
"loss": 0.1414,
"step": 81200
},
{
"epoch": 1.155896,
"grad_norm": 5.049018383026123,
"learning_rate": 3.5128514056224905e-06,
"loss": 0.1404,
"step": 81300
},
{
"epoch": 1.156696,
"grad_norm": 4.586015701293945,
"learning_rate": 3.504819277108434e-06,
"loss": 0.1446,
"step": 81400
},
{
"epoch": 1.157496,
"grad_norm": 4.436155796051025,
"learning_rate": 3.496787148594378e-06,
"loss": 0.1406,
"step": 81500
},
{
"epoch": 1.158296,
"grad_norm": 2.5560381412506104,
"learning_rate": 3.4887550200803214e-06,
"loss": 0.1384,
"step": 81600
},
{
"epoch": 1.159096,
"grad_norm": 5.3852057456970215,
"learning_rate": 3.4807228915662654e-06,
"loss": 0.1412,
"step": 81700
},
{
"epoch": 1.159896,
"grad_norm": 5.565433502197266,
"learning_rate": 3.472690763052209e-06,
"loss": 0.1441,
"step": 81800
},
{
"epoch": 1.160696,
"grad_norm": 10.139933586120605,
"learning_rate": 3.4646586345381528e-06,
"loss": 0.1387,
"step": 81900
},
{
"epoch": 1.161496,
"grad_norm": 10.769366264343262,
"learning_rate": 3.4566265060240967e-06,
"loss": 0.1424,
"step": 82000
},
{
"epoch": 1.162296,
"grad_norm": 7.583619594573975,
"learning_rate": 3.4485943775100406e-06,
"loss": 0.1476,
"step": 82100
},
{
"epoch": 1.163096,
"grad_norm": 3.2200429439544678,
"learning_rate": 3.4405622489959845e-06,
"loss": 0.1391,
"step": 82200
},
{
"epoch": 1.163896,
"grad_norm": 4.72890567779541,
"learning_rate": 3.4325301204819276e-06,
"loss": 0.1371,
"step": 82300
},
{
"epoch": 1.164696,
"grad_norm": 3.9485466480255127,
"learning_rate": 3.4244979919678715e-06,
"loss": 0.1349,
"step": 82400
},
{
"epoch": 1.165496,
"grad_norm": 7.35235071182251,
"learning_rate": 3.4164658634538154e-06,
"loss": 0.1419,
"step": 82500
},
{
"epoch": 1.166296,
"grad_norm": 7.841335773468018,
"learning_rate": 3.4084337349397594e-06,
"loss": 0.1459,
"step": 82600
},
{
"epoch": 1.167096,
"grad_norm": 3.344604015350342,
"learning_rate": 3.400401606425703e-06,
"loss": 0.1198,
"step": 82700
},
{
"epoch": 1.167896,
"grad_norm": 4.77374792098999,
"learning_rate": 3.3923694779116468e-06,
"loss": 0.1531,
"step": 82800
},
{
"epoch": 1.168696,
"grad_norm": 6.152327537536621,
"learning_rate": 3.3843373493975907e-06,
"loss": 0.1303,
"step": 82900
},
{
"epoch": 1.169496,
"grad_norm": 5.056685924530029,
"learning_rate": 3.3763052208835346e-06,
"loss": 0.1479,
"step": 83000
},
{
"epoch": 1.170296,
"grad_norm": 7.288972854614258,
"learning_rate": 3.3682730923694785e-06,
"loss": 0.1502,
"step": 83100
},
{
"epoch": 1.171096,
"grad_norm": 6.630878448486328,
"learning_rate": 3.3602409638554216e-06,
"loss": 0.1435,
"step": 83200
},
{
"epoch": 1.171896,
"grad_norm": 6.784937381744385,
"learning_rate": 3.3522088353413655e-06,
"loss": 0.1445,
"step": 83300
},
{
"epoch": 1.172696,
"grad_norm": 5.813895225524902,
"learning_rate": 3.3441767068273095e-06,
"loss": 0.1362,
"step": 83400
},
{
"epoch": 1.173496,
"grad_norm": 6.4298415184021,
"learning_rate": 3.3361445783132534e-06,
"loss": 0.138,
"step": 83500
},
{
"epoch": 1.174296,
"grad_norm": 3.369779586791992,
"learning_rate": 3.328112449799197e-06,
"loss": 0.1251,
"step": 83600
},
{
"epoch": 1.175096,
"grad_norm": 2.4179258346557617,
"learning_rate": 3.320080321285141e-06,
"loss": 0.1491,
"step": 83700
},
{
"epoch": 1.175896,
"grad_norm": 6.803851127624512,
"learning_rate": 3.3120481927710847e-06,
"loss": 0.1209,
"step": 83800
},
{
"epoch": 1.176696,
"grad_norm": 5.81317138671875,
"learning_rate": 3.3040160642570286e-06,
"loss": 0.1463,
"step": 83900
},
{
"epoch": 1.177496,
"grad_norm": 3.342420816421509,
"learning_rate": 3.2959839357429726e-06,
"loss": 0.1108,
"step": 84000
},
{
"epoch": 1.177496,
"eval_test1_cer": 0.04321177832549108,
"eval_test1_cer_norm": 0.028405407352270273,
"eval_test1_loss": 0.17878110706806183,
"eval_test1_runtime": 3382.137,
"eval_test1_samples_per_second": 0.739,
"eval_test1_steps_per_second": 0.185,
"eval_test1_wer": 0.13557246566955306,
"eval_test1_wer_norm": 0.07484146234548059,
"step": 84000
},
{
"epoch": 1.177496,
"eval_test2_cer": 0.10907343114197185,
"eval_test2_cer_norm": 0.08453091106290672,
"eval_test2_loss": 0.32212311029434204,
"eval_test2_runtime": 3480.5629,
"eval_test2_samples_per_second": 0.718,
"eval_test2_steps_per_second": 0.18,
"eval_test2_wer": 0.2316891737239643,
"eval_test2_wer_norm": 0.16588356635342655,
"step": 84000
},
{
"epoch": 1.178296,
"grad_norm": 5.4935503005981445,
"learning_rate": 3.2879518072289156e-06,
"loss": 0.1458,
"step": 84100
},
{
"epoch": 1.179096,
"grad_norm": 7.605450630187988,
"learning_rate": 3.2799196787148595e-06,
"loss": 0.1377,
"step": 84200
},
{
"epoch": 1.179896,
"grad_norm": 6.349878311157227,
"learning_rate": 3.2719678714859443e-06,
"loss": 0.1365,
"step": 84300
},
{
"epoch": 1.180696,
"grad_norm": 6.173891544342041,
"learning_rate": 3.2639357429718878e-06,
"loss": 0.1394,
"step": 84400
},
{
"epoch": 1.181496,
"grad_norm": 1.916458249092102,
"learning_rate": 3.2559036144578317e-06,
"loss": 0.1343,
"step": 84500
},
{
"epoch": 1.182296,
"grad_norm": 4.605401039123535,
"learning_rate": 3.2478714859437756e-06,
"loss": 0.1438,
"step": 84600
},
{
"epoch": 1.183096,
"grad_norm": 10.40539264678955,
"learning_rate": 3.2398393574297195e-06,
"loss": 0.1426,
"step": 84700
},
{
"epoch": 1.183896,
"grad_norm": 5.316925525665283,
"learning_rate": 3.2318072289156626e-06,
"loss": 0.1312,
"step": 84800
},
{
"epoch": 1.184696,
"grad_norm": 3.3205056190490723,
"learning_rate": 3.2237751004016065e-06,
"loss": 0.1339,
"step": 84900
},
{
"epoch": 1.185496,
"grad_norm": 4.223670959472656,
"learning_rate": 3.2157429718875504e-06,
"loss": 0.1477,
"step": 85000
},
{
"epoch": 1.186296,
"grad_norm": 2.418142318725586,
"learning_rate": 3.2077108433734944e-06,
"loss": 0.13,
"step": 85100
},
{
"epoch": 1.187096,
"grad_norm": 5.908360958099365,
"learning_rate": 3.1996787148594383e-06,
"loss": 0.1351,
"step": 85200
},
{
"epoch": 1.187896,
"grad_norm": 3.79148268699646,
"learning_rate": 3.1916465863453818e-06,
"loss": 0.1193,
"step": 85300
},
{
"epoch": 1.188696,
"grad_norm": 4.751871109008789,
"learning_rate": 3.1836144578313257e-06,
"loss": 0.1357,
"step": 85400
},
{
"epoch": 1.189496,
"grad_norm": 5.901039123535156,
"learning_rate": 3.175582329317269e-06,
"loss": 0.1214,
"step": 85500
},
{
"epoch": 1.190296,
"grad_norm": 4.6331305503845215,
"learning_rate": 3.167550200803213e-06,
"loss": 0.1593,
"step": 85600
},
{
"epoch": 1.191096,
"grad_norm": 5.742949962615967,
"learning_rate": 3.1595180722891566e-06,
"loss": 0.1321,
"step": 85700
},
{
"epoch": 1.191896,
"grad_norm": 3.329071283340454,
"learning_rate": 3.1514859437751005e-06,
"loss": 0.1514,
"step": 85800
},
{
"epoch": 1.192696,
"grad_norm": 6.989814758300781,
"learning_rate": 3.1434538152610445e-06,
"loss": 0.135,
"step": 85900
},
{
"epoch": 1.1934960000000001,
"grad_norm": 4.263644218444824,
"learning_rate": 3.1354216867469884e-06,
"loss": 0.1364,
"step": 86000
},
{
"epoch": 1.194296,
"grad_norm": 3.34621524810791,
"learning_rate": 3.1273895582329323e-06,
"loss": 0.1215,
"step": 86100
},
{
"epoch": 1.195096,
"grad_norm": 5.672510623931885,
"learning_rate": 3.1193574297188754e-06,
"loss": 0.1323,
"step": 86200
},
{
"epoch": 1.195896,
"grad_norm": 7.123562812805176,
"learning_rate": 3.1113253012048193e-06,
"loss": 0.128,
"step": 86300
},
{
"epoch": 1.196696,
"grad_norm": 5.052427768707275,
"learning_rate": 3.1033734939759036e-06,
"loss": 0.1343,
"step": 86400
},
{
"epoch": 1.1974960000000001,
"grad_norm": 9.78446102142334,
"learning_rate": 3.0953413654618475e-06,
"loss": 0.1354,
"step": 86500
},
{
"epoch": 1.198296,
"grad_norm": 3.7782175540924072,
"learning_rate": 3.0873092369477914e-06,
"loss": 0.1369,
"step": 86600
},
{
"epoch": 1.199096,
"grad_norm": 3.037858247756958,
"learning_rate": 3.0792771084337354e-06,
"loss": 0.1368,
"step": 86700
},
{
"epoch": 1.199896,
"grad_norm": 6.603598117828369,
"learning_rate": 3.0712449799196793e-06,
"loss": 0.1313,
"step": 86800
},
{
"epoch": 1.200696,
"grad_norm": 5.845118522644043,
"learning_rate": 3.0632128514056224e-06,
"loss": 0.1381,
"step": 86900
},
{
"epoch": 1.2014960000000001,
"grad_norm": 5.331429958343506,
"learning_rate": 3.0551807228915663e-06,
"loss": 0.1394,
"step": 87000
},
{
"epoch": 1.202296,
"grad_norm": 2.769038677215576,
"learning_rate": 3.04714859437751e-06,
"loss": 0.138,
"step": 87100
},
{
"epoch": 1.203096,
"grad_norm": 4.4109320640563965,
"learning_rate": 3.039116465863454e-06,
"loss": 0.1355,
"step": 87200
},
{
"epoch": 1.203896,
"grad_norm": 5.107141494750977,
"learning_rate": 3.031084337349398e-06,
"loss": 0.1242,
"step": 87300
},
{
"epoch": 1.204696,
"grad_norm": 3.513662815093994,
"learning_rate": 3.0230522088353415e-06,
"loss": 0.1341,
"step": 87400
},
{
"epoch": 1.2054960000000001,
"grad_norm": 5.542943000793457,
"learning_rate": 3.0150200803212855e-06,
"loss": 0.1582,
"step": 87500
},
{
"epoch": 1.206296,
"grad_norm": 6.642317771911621,
"learning_rate": 3.0069879518072294e-06,
"loss": 0.1328,
"step": 87600
},
{
"epoch": 1.207096,
"grad_norm": 4.840209007263184,
"learning_rate": 2.9989558232931733e-06,
"loss": 0.1442,
"step": 87700
},
{
"epoch": 1.207896,
"grad_norm": 8.443116188049316,
"learning_rate": 2.9909236947791164e-06,
"loss": 0.1301,
"step": 87800
},
{
"epoch": 1.208696,
"grad_norm": 5.006475925445557,
"learning_rate": 2.9828915662650603e-06,
"loss": 0.137,
"step": 87900
},
{
"epoch": 1.209496,
"grad_norm": 16.58333969116211,
"learning_rate": 2.974859437751004e-06,
"loss": 0.141,
"step": 88000
},
{
"epoch": 1.209496,
"eval_test1_cer": 0.05272518485323773,
"eval_test1_cer_norm": 0.035959372823972914,
"eval_test1_loss": 0.1771300584077835,
"eval_test1_runtime": 3405.3955,
"eval_test1_samples_per_second": 0.734,
"eval_test1_steps_per_second": 0.184,
"eval_test1_wer": 0.15084987900521882,
"eval_test1_wer_norm": 0.0903883807241591,
"step": 88000
},
{
"epoch": 1.209496,
"eval_test2_cer": 0.10232108858774779,
"eval_test2_cer_norm": 0.08082681282925318,
"eval_test2_loss": 0.31750166416168213,
"eval_test2_runtime": 3462.1713,
"eval_test2_samples_per_second": 0.722,
"eval_test2_steps_per_second": 0.181,
"eval_test2_wer": 0.2193579766536965,
"eval_test2_wer_norm": 0.15279051111620445,
"step": 88000
},
{
"epoch": 1.210296,
"grad_norm": 7.229742527008057,
"learning_rate": 2.966827309236948e-06,
"loss": 0.1565,
"step": 88100
},
{
"epoch": 1.211096,
"grad_norm": 4.775784015655518,
"learning_rate": 2.958795180722892e-06,
"loss": 0.1191,
"step": 88200
},
{
"epoch": 1.211896,
"grad_norm": 5.914107799530029,
"learning_rate": 2.9507630522088355e-06,
"loss": 0.1272,
"step": 88300
},
{
"epoch": 1.212696,
"grad_norm": 3.42290997505188,
"learning_rate": 2.9427309236947795e-06,
"loss": 0.1533,
"step": 88400
},
{
"epoch": 1.213496,
"grad_norm": 5.825013637542725,
"learning_rate": 2.9347791164658634e-06,
"loss": 0.1337,
"step": 88500
},
{
"epoch": 1.214296,
"grad_norm": 2.929975748062134,
"learning_rate": 2.9267469879518073e-06,
"loss": 0.1363,
"step": 88600
},
{
"epoch": 1.215096,
"grad_norm": 4.544336318969727,
"learning_rate": 2.918714859437751e-06,
"loss": 0.1416,
"step": 88700
},
{
"epoch": 1.215896,
"grad_norm": 6.057349681854248,
"learning_rate": 2.910682730923695e-06,
"loss": 0.1298,
"step": 88800
},
{
"epoch": 1.216696,
"grad_norm": 4.51986026763916,
"learning_rate": 2.902650602409639e-06,
"loss": 0.1333,
"step": 88900
},
{
"epoch": 1.217496,
"grad_norm": 4.948803424835205,
"learning_rate": 2.8946184738955825e-06,
"loss": 0.134,
"step": 89000
},
{
"epoch": 1.218296,
"grad_norm": 4.863702774047852,
"learning_rate": 2.8865863453815264e-06,
"loss": 0.1378,
"step": 89100
},
{
"epoch": 1.219096,
"grad_norm": 4.049374580383301,
"learning_rate": 2.8785542168674704e-06,
"loss": 0.125,
"step": 89200
},
{
"epoch": 1.219896,
"grad_norm": 3.5720324516296387,
"learning_rate": 2.870522088353414e-06,
"loss": 0.1395,
"step": 89300
},
{
"epoch": 1.220696,
"grad_norm": 1.931492567062378,
"learning_rate": 2.8624899598393574e-06,
"loss": 0.131,
"step": 89400
},
{
"epoch": 1.221496,
"grad_norm": 6.574014663696289,
"learning_rate": 2.8544578313253013e-06,
"loss": 0.132,
"step": 89500
},
{
"epoch": 1.222296,
"grad_norm": 7.461375713348389,
"learning_rate": 2.846425702811245e-06,
"loss": 0.1289,
"step": 89600
},
{
"epoch": 1.223096,
"grad_norm": 5.367936611175537,
"learning_rate": 2.838393574297189e-06,
"loss": 0.1261,
"step": 89700
},
{
"epoch": 1.223896,
"grad_norm": 5.853065013885498,
"learning_rate": 2.830361445783133e-06,
"loss": 0.1161,
"step": 89800
},
{
"epoch": 1.224696,
"grad_norm": 4.45815896987915,
"learning_rate": 2.8223293172690765e-06,
"loss": 0.1285,
"step": 89900
},
{
"epoch": 1.225496,
"grad_norm": 5.519855499267578,
"learning_rate": 2.81429718875502e-06,
"loss": 0.1464,
"step": 90000
},
{
"epoch": 1.226296,
"grad_norm": 3.830179214477539,
"learning_rate": 2.806265060240964e-06,
"loss": 0.1278,
"step": 90100
},
{
"epoch": 1.227096,
"grad_norm": 3.485785722732544,
"learning_rate": 2.798232931726908e-06,
"loss": 0.1599,
"step": 90200
},
{
"epoch": 1.2278959999999999,
"grad_norm": 5.015017509460449,
"learning_rate": 2.7902008032128514e-06,
"loss": 0.1213,
"step": 90300
},
{
"epoch": 1.228696,
"grad_norm": 5.003856658935547,
"learning_rate": 2.7821686746987953e-06,
"loss": 0.1268,
"step": 90400
},
{
"epoch": 1.229496,
"grad_norm": 8.828622817993164,
"learning_rate": 2.77421686746988e-06,
"loss": 0.128,
"step": 90500
},
{
"epoch": 1.230296,
"grad_norm": 5.4841790199279785,
"learning_rate": 2.766184738955823e-06,
"loss": 0.1267,
"step": 90600
},
{
"epoch": 1.231096,
"grad_norm": 1.1707990169525146,
"learning_rate": 2.758152610441767e-06,
"loss": 0.1173,
"step": 90700
},
{
"epoch": 1.2318959999999999,
"grad_norm": 4.860354423522949,
"learning_rate": 2.750120481927711e-06,
"loss": 0.1248,
"step": 90800
},
{
"epoch": 1.232696,
"grad_norm": 4.709814071655273,
"learning_rate": 2.742088353413655e-06,
"loss": 0.1288,
"step": 90900
},
{
"epoch": 1.233496,
"grad_norm": 5.921780109405518,
"learning_rate": 2.7340562248995988e-06,
"loss": 0.1165,
"step": 91000
},
{
"epoch": 1.234296,
"grad_norm": 31.39832305908203,
"learning_rate": 2.7260240963855423e-06,
"loss": 0.1362,
"step": 91100
},
{
"epoch": 1.235096,
"grad_norm": 9.929832458496094,
"learning_rate": 2.717991967871486e-06,
"loss": 0.1356,
"step": 91200
},
{
"epoch": 1.2358959999999999,
"grad_norm": 6.635560035705566,
"learning_rate": 2.70995983935743e-06,
"loss": 0.1242,
"step": 91300
},
{
"epoch": 1.236696,
"grad_norm": 3.8171298503875732,
"learning_rate": 2.701927710843374e-06,
"loss": 0.1322,
"step": 91400
},
{
"epoch": 1.237496,
"grad_norm": 5.8909430503845215,
"learning_rate": 2.693895582329317e-06,
"loss": 0.1404,
"step": 91500
},
{
"epoch": 1.238296,
"grad_norm": 12.169204711914062,
"learning_rate": 2.685863453815261e-06,
"loss": 0.1312,
"step": 91600
},
{
"epoch": 1.239096,
"grad_norm": 2.663769483566284,
"learning_rate": 2.677831325301205e-06,
"loss": 0.1266,
"step": 91700
},
{
"epoch": 1.2398959999999999,
"grad_norm": 10.018877983093262,
"learning_rate": 2.669799196787149e-06,
"loss": 0.1279,
"step": 91800
},
{
"epoch": 1.240696,
"grad_norm": 3.6418004035949707,
"learning_rate": 2.661767068273093e-06,
"loss": 0.1233,
"step": 91900
},
{
"epoch": 1.241496,
"grad_norm": 6.836270332336426,
"learning_rate": 2.6537349397590363e-06,
"loss": 0.137,
"step": 92000
},
{
"epoch": 1.241496,
"eval_test1_cer": 0.04675013070430951,
"eval_test1_cer_norm": 0.031027444954018296,
"eval_test1_loss": 0.17643441259860992,
"eval_test1_runtime": 3383.8187,
"eval_test1_samples_per_second": 0.739,
"eval_test1_steps_per_second": 0.185,
"eval_test1_wer": 0.1424531327443949,
"eval_test1_wer_norm": 0.08281948625032877,
"step": 92000
},
{
"epoch": 1.241496,
"eval_test2_cer": 0.10391234554074738,
"eval_test2_cer_norm": 0.07987294700960644,
"eval_test2_loss": 0.31574827432632446,
"eval_test2_runtime": 3451.2496,
"eval_test2_samples_per_second": 0.724,
"eval_test2_steps_per_second": 0.181,
"eval_test2_wer": 0.22485122453650722,
"eval_test2_wer_norm": 0.1593513637405455,
"step": 92000
},
{
"epoch": 1.242296,
"grad_norm": 6.066310405731201,
"learning_rate": 2.64570281124498e-06,
"loss": 0.1255,
"step": 92100
},
{
"epoch": 1.243096,
"grad_norm": 4.294830799102783,
"learning_rate": 2.637670682730924e-06,
"loss": 0.1399,
"step": 92200
},
{
"epoch": 1.243896,
"grad_norm": 6.00723934173584,
"learning_rate": 2.629638554216868e-06,
"loss": 0.1446,
"step": 92300
},
{
"epoch": 1.244696,
"grad_norm": 10.248518943786621,
"learning_rate": 2.621606425702811e-06,
"loss": 0.1317,
"step": 92400
},
{
"epoch": 1.245496,
"grad_norm": 2.9732255935668945,
"learning_rate": 2.613654618473896e-06,
"loss": 0.1329,
"step": 92500
},
{
"epoch": 1.246296,
"grad_norm": 5.735791206359863,
"learning_rate": 2.6056224899598398e-06,
"loss": 0.1246,
"step": 92600
},
{
"epoch": 1.247096,
"grad_norm": 4.58491849899292,
"learning_rate": 2.5975903614457833e-06,
"loss": 0.118,
"step": 92700
},
{
"epoch": 1.247896,
"grad_norm": 6.321929931640625,
"learning_rate": 2.589558232931727e-06,
"loss": 0.1317,
"step": 92800
},
{
"epoch": 1.248696,
"grad_norm": 2.8514037132263184,
"learning_rate": 2.581526104417671e-06,
"loss": 0.124,
"step": 92900
},
{
"epoch": 1.249496,
"grad_norm": 6.127628803253174,
"learning_rate": 2.573493975903615e-06,
"loss": 0.1429,
"step": 93000
},
{
"epoch": 1.250296,
"grad_norm": 5.147157192230225,
"learning_rate": 2.565461847389558e-06,
"loss": 0.1294,
"step": 93100
},
{
"epoch": 1.251096,
"grad_norm": 1.7546809911727905,
"learning_rate": 2.557429718875502e-06,
"loss": 0.1228,
"step": 93200
},
{
"epoch": 1.251896,
"grad_norm": 5.040311813354492,
"learning_rate": 2.549397590361446e-06,
"loss": 0.1272,
"step": 93300
},
{
"epoch": 1.252696,
"grad_norm": 6.703260898590088,
"learning_rate": 2.54136546184739e-06,
"loss": 0.1312,
"step": 93400
},
{
"epoch": 1.253496,
"grad_norm": 4.4106974601745605,
"learning_rate": 2.5333333333333338e-06,
"loss": 0.1209,
"step": 93500
},
{
"epoch": 1.254296,
"grad_norm": 4.892759799957275,
"learning_rate": 2.5253012048192773e-06,
"loss": 0.1273,
"step": 93600
},
{
"epoch": 1.255096,
"grad_norm": 3.4537353515625,
"learning_rate": 2.517269076305221e-06,
"loss": 0.1292,
"step": 93700
},
{
"epoch": 1.255896,
"grad_norm": 6.59584379196167,
"learning_rate": 2.5092369477911647e-06,
"loss": 0.1194,
"step": 93800
},
{
"epoch": 1.256696,
"grad_norm": 7.888189792633057,
"learning_rate": 2.5012048192771086e-06,
"loss": 0.1292,
"step": 93900
},
{
"epoch": 1.257496,
"grad_norm": 3.4533464908599854,
"learning_rate": 2.4931726907630525e-06,
"loss": 0.1271,
"step": 94000
},
{
"epoch": 1.258296,
"grad_norm": 1.2868270874023438,
"learning_rate": 2.4851405622489965e-06,
"loss": 0.1416,
"step": 94100
},
{
"epoch": 1.259096,
"grad_norm": 4.686305999755859,
"learning_rate": 2.47710843373494e-06,
"loss": 0.135,
"step": 94200
},
{
"epoch": 1.259896,
"grad_norm": 2.333994150161743,
"learning_rate": 2.469076305220884e-06,
"loss": 0.1143,
"step": 94300
},
{
"epoch": 1.260696,
"grad_norm": 24.702146530151367,
"learning_rate": 2.4610441767068274e-06,
"loss": 0.122,
"step": 94400
},
{
"epoch": 1.261496,
"grad_norm": 2.577688694000244,
"learning_rate": 2.4530120481927713e-06,
"loss": 0.1323,
"step": 94500
},
{
"epoch": 1.262296,
"grad_norm": 4.011758804321289,
"learning_rate": 2.4449799196787148e-06,
"loss": 0.1375,
"step": 94600
},
{
"epoch": 1.263096,
"grad_norm": 3.8181052207946777,
"learning_rate": 2.4369477911646587e-06,
"loss": 0.1364,
"step": 94700
},
{
"epoch": 1.263896,
"grad_norm": 5.570916652679443,
"learning_rate": 2.4289959839357434e-06,
"loss": 0.1334,
"step": 94800
},
{
"epoch": 1.264696,
"grad_norm": 5.255520820617676,
"learning_rate": 2.420963855421687e-06,
"loss": 0.1346,
"step": 94900
},
{
"epoch": 1.265496,
"grad_norm": 6.02402400970459,
"learning_rate": 2.412931726907631e-06,
"loss": 0.1299,
"step": 95000
},
{
"epoch": 1.266296,
"grad_norm": 5.61665153503418,
"learning_rate": 2.4048995983935744e-06,
"loss": 0.1332,
"step": 95100
},
{
"epoch": 1.267096,
"grad_norm": 3.135876417160034,
"learning_rate": 2.3968674698795183e-06,
"loss": 0.1466,
"step": 95200
},
{
"epoch": 1.267896,
"grad_norm": 3.9527804851531982,
"learning_rate": 2.3888353413654618e-06,
"loss": 0.118,
"step": 95300
},
{
"epoch": 1.268696,
"grad_norm": 8.5547513961792,
"learning_rate": 2.3808032128514057e-06,
"loss": 0.1289,
"step": 95400
},
{
"epoch": 1.269496,
"grad_norm": 4.350017070770264,
"learning_rate": 2.3727710843373496e-06,
"loss": 0.1485,
"step": 95500
},
{
"epoch": 1.270296,
"grad_norm": 7.225109100341797,
"learning_rate": 2.3647389558232935e-06,
"loss": 0.1724,
"step": 95600
},
{
"epoch": 1.271096,
"grad_norm": 3.152327060699463,
"learning_rate": 2.356706827309237e-06,
"loss": 0.151,
"step": 95700
},
{
"epoch": 1.271896,
"grad_norm": 9.343049049377441,
"learning_rate": 2.348674698795181e-06,
"loss": 0.1219,
"step": 95800
},
{
"epoch": 1.272696,
"grad_norm": 3.7528295516967773,
"learning_rate": 2.340642570281125e-06,
"loss": 0.1011,
"step": 95900
},
{
"epoch": 1.273496,
"grad_norm": 1.7038061618804932,
"learning_rate": 2.3326104417670684e-06,
"loss": 0.1287,
"step": 96000
},
{
"epoch": 1.273496,
"eval_test1_cer": 0.04843528269474942,
"eval_test1_cer_norm": 0.03259298388839532,
"eval_test1_loss": 0.17615145444869995,
"eval_test1_runtime": 3390.194,
"eval_test1_samples_per_second": 0.737,
"eval_test1_steps_per_second": 0.184,
"eval_test1_wer": 0.14195749147205458,
"eval_test1_wer_norm": 0.08346240392764254,
"step": 96000
},
{
"epoch": 1.273496,
"eval_test2_cer": 0.10040784709000634,
"eval_test2_cer_norm": 0.07324914781530834,
"eval_test2_loss": 0.31477001309394836,
"eval_test2_runtime": 3429.6171,
"eval_test2_samples_per_second": 0.729,
"eval_test2_steps_per_second": 0.182,
"eval_test2_wer": 0.21283474479285877,
"eval_test2_wer_norm": 0.14789135915654367,
"step": 96000
},
{
"epoch": 1.274296,
"grad_norm": 4.975541591644287,
"learning_rate": 2.3245783132530123e-06,
"loss": 0.1545,
"step": 96100
},
{
"epoch": 1.275096,
"grad_norm": 2.7354822158813477,
"learning_rate": 2.316546184738956e-06,
"loss": 0.1379,
"step": 96200
},
{
"epoch": 1.275896,
"grad_norm": 4.287178993225098,
"learning_rate": 2.3085140562248997e-06,
"loss": 0.1109,
"step": 96300
},
{
"epoch": 1.276696,
"grad_norm": 7.494585990905762,
"learning_rate": 2.3004819277108436e-06,
"loss": 0.1361,
"step": 96400
},
{
"epoch": 1.277496,
"grad_norm": 5.967809677124023,
"learning_rate": 2.292449799196787e-06,
"loss": 0.1282,
"step": 96500
},
{
"epoch": 1.278296,
"grad_norm": 6.256405830383301,
"learning_rate": 2.284417670682731e-06,
"loss": 0.1254,
"step": 96600
},
{
"epoch": 1.279096,
"grad_norm": 4.7275214195251465,
"learning_rate": 2.276385542168675e-06,
"loss": 0.1125,
"step": 96700
},
{
"epoch": 1.279896,
"grad_norm": 6.460155010223389,
"learning_rate": 2.2685140562248997e-06,
"loss": 0.1531,
"step": 96800
},
{
"epoch": 1.280696,
"grad_norm": 7.814698696136475,
"learning_rate": 2.2604819277108436e-06,
"loss": 0.1353,
"step": 96900
},
{
"epoch": 1.281496,
"grad_norm": 2.7123336791992188,
"learning_rate": 2.252449799196787e-06,
"loss": 0.1223,
"step": 97000
},
{
"epoch": 1.282296,
"grad_norm": 4.504587650299072,
"learning_rate": 2.244417670682731e-06,
"loss": 0.1358,
"step": 97100
},
{
"epoch": 1.283096,
"grad_norm": 7.585097789764404,
"learning_rate": 2.236385542168675e-06,
"loss": 0.1378,
"step": 97200
},
{
"epoch": 1.283896,
"grad_norm": 3.129997491836548,
"learning_rate": 2.228353413654619e-06,
"loss": 0.1234,
"step": 97300
},
{
"epoch": 1.284696,
"grad_norm": 6.103450775146484,
"learning_rate": 2.2203212851405628e-06,
"loss": 0.1188,
"step": 97400
},
{
"epoch": 1.285496,
"grad_norm": 5.75684928894043,
"learning_rate": 2.2122891566265063e-06,
"loss": 0.118,
"step": 97500
},
{
"epoch": 1.286296,
"grad_norm": 4.268872261047363,
"learning_rate": 2.20425702811245e-06,
"loss": 0.1242,
"step": 97600
},
{
"epoch": 1.287096,
"grad_norm": 10.057229042053223,
"learning_rate": 2.1962248995983937e-06,
"loss": 0.107,
"step": 97700
},
{
"epoch": 1.287896,
"grad_norm": 6.790497303009033,
"learning_rate": 2.1881927710843376e-06,
"loss": 0.1169,
"step": 97800
},
{
"epoch": 1.288696,
"grad_norm": 7.399913311004639,
"learning_rate": 2.180160642570281e-06,
"loss": 0.1389,
"step": 97900
},
{
"epoch": 1.289496,
"grad_norm": 8.95304012298584,
"learning_rate": 2.172128514056225e-06,
"loss": 0.1119,
"step": 98000
},
{
"epoch": 1.290296,
"grad_norm": 3.782831907272339,
"learning_rate": 2.164096385542169e-06,
"loss": 0.1158,
"step": 98100
},
{
"epoch": 1.291096,
"grad_norm": 9.19373893737793,
"learning_rate": 2.1560642570281124e-06,
"loss": 0.1373,
"step": 98200
},
{
"epoch": 1.291896,
"grad_norm": 8.121448516845703,
"learning_rate": 2.1480321285140563e-06,
"loss": 0.1379,
"step": 98300
},
{
"epoch": 1.292696,
"grad_norm": 5.722080707550049,
"learning_rate": 2.1400000000000003e-06,
"loss": 0.1171,
"step": 98400
},
{
"epoch": 1.293496,
"grad_norm": 4.1129255294799805,
"learning_rate": 2.131967871485944e-06,
"loss": 0.1302,
"step": 98500
},
{
"epoch": 1.2942960000000001,
"grad_norm": 3.102550983428955,
"learning_rate": 2.1239357429718877e-06,
"loss": 0.1256,
"step": 98600
},
{
"epoch": 1.295096,
"grad_norm": 5.705405235290527,
"learning_rate": 2.1159036144578316e-06,
"loss": 0.1103,
"step": 98700
},
{
"epoch": 1.295896,
"grad_norm": 5.043685436248779,
"learning_rate": 2.107871485943775e-06,
"loss": 0.1207,
"step": 98800
},
{
"epoch": 1.296696,
"grad_norm": 5.753846645355225,
"learning_rate": 2.099839357429719e-06,
"loss": 0.1195,
"step": 98900
},
{
"epoch": 1.297496,
"grad_norm": 5.277960300445557,
"learning_rate": 2.091807228915663e-06,
"loss": 0.1326,
"step": 99000
},
{
"epoch": 1.2982960000000001,
"grad_norm": 4.641193866729736,
"learning_rate": 2.0837751004016064e-06,
"loss": 0.1409,
"step": 99100
},
{
"epoch": 1.299096,
"grad_norm": 6.307822227478027,
"learning_rate": 2.0757429718875504e-06,
"loss": 0.1271,
"step": 99200
},
{
"epoch": 1.299896,
"grad_norm": 4.50252628326416,
"learning_rate": 2.0677108433734943e-06,
"loss": 0.1242,
"step": 99300
},
{
"epoch": 1.300696,
"grad_norm": 6.755404949188232,
"learning_rate": 2.059678714859438e-06,
"loss": 0.1349,
"step": 99400
},
{
"epoch": 1.301496,
"grad_norm": 4.805443286895752,
"learning_rate": 2.0516465863453817e-06,
"loss": 0.1227,
"step": 99500
},
{
"epoch": 1.3022960000000001,
"grad_norm": 3.4150850772857666,
"learning_rate": 2.0436144578313256e-06,
"loss": 0.1147,
"step": 99600
},
{
"epoch": 1.303096,
"grad_norm": 6.213440895080566,
"learning_rate": 2.035582329317269e-06,
"loss": 0.1236,
"step": 99700
},
{
"epoch": 1.303896,
"grad_norm": 6.957757472991943,
"learning_rate": 2.027550200803213e-06,
"loss": 0.1323,
"step": 99800
},
{
"epoch": 1.304696,
"grad_norm": 7.19875955581665,
"learning_rate": 2.019518072289157e-06,
"loss": 0.1283,
"step": 99900
},
{
"epoch": 1.305496,
"grad_norm": 2.5074002742767334,
"learning_rate": 2.0114859437751004e-06,
"loss": 0.1128,
"step": 100000
},
{
"epoch": 1.305496,
"eval_test1_cer": 0.0477444170587796,
"eval_test1_cer_norm": 0.03308281508872188,
"eval_test1_loss": 0.17668992280960083,
"eval_test1_runtime": 3382.0305,
"eval_test1_samples_per_second": 0.739,
"eval_test1_steps_per_second": 0.185,
"eval_test1_wer": 0.1387795562552844,
"eval_test1_wer_norm": 0.07951722727139893,
"step": 100000
},
{
"epoch": 1.305496,
"eval_test2_cer": 0.10429499384029567,
"eval_test2_cer_norm": 0.0824295010845987,
"eval_test2_loss": 0.3178161084651947,
"eval_test2_runtime": 3460.99,
"eval_test2_samples_per_second": 0.722,
"eval_test2_steps_per_second": 0.181,
"eval_test2_wer": 0.2208743419546807,
"eval_test2_wer_norm": 0.15585606234242494,
"step": 100000
},
{
"epoch": 1.3062960000000001,
"grad_norm": 4.352635383605957,
"learning_rate": 2.0034538152610444e-06,
"loss": 0.1269,
"step": 100100
},
{
"epoch": 1.307096,
"grad_norm": 6.455359935760498,
"learning_rate": 1.995421686746988e-06,
"loss": 0.1228,
"step": 100200
},
{
"epoch": 1.307896,
"grad_norm": 4.56587028503418,
"learning_rate": 1.9873895582329318e-06,
"loss": 0.1302,
"step": 100300
},
{
"epoch": 1.308696,
"grad_norm": 6.275040149688721,
"learning_rate": 1.9793574297188757e-06,
"loss": 0.1316,
"step": 100400
},
{
"epoch": 1.309496,
"grad_norm": 2.4755728244781494,
"learning_rate": 1.9713253012048196e-06,
"loss": 0.1324,
"step": 100500
},
{
"epoch": 1.3102960000000001,
"grad_norm": 5.687671184539795,
"learning_rate": 1.963293172690763e-06,
"loss": 0.1296,
"step": 100600
},
{
"epoch": 1.311096,
"grad_norm": 3.915098190307617,
"learning_rate": 1.955261044176707e-06,
"loss": 0.1198,
"step": 100700
},
{
"epoch": 1.311896,
"grad_norm": 5.507267475128174,
"learning_rate": 1.947228915662651e-06,
"loss": 0.1308,
"step": 100800
},
{
"epoch": 1.3126959999999999,
"grad_norm": 6.799343109130859,
"learning_rate": 1.939277108433735e-06,
"loss": 0.1106,
"step": 100900
},
{
"epoch": 1.313496,
"grad_norm": 2.4747636318206787,
"learning_rate": 1.9312449799196788e-06,
"loss": 0.1198,
"step": 101000
},
{
"epoch": 1.3142960000000001,
"grad_norm": 1.4364855289459229,
"learning_rate": 1.9232128514056227e-06,
"loss": 0.1251,
"step": 101100
},
{
"epoch": 1.315096,
"grad_norm": 3.4942729473114014,
"learning_rate": 1.9151807228915666e-06,
"loss": 0.105,
"step": 101200
},
{
"epoch": 1.315896,
"grad_norm": 7.960775852203369,
"learning_rate": 1.9071485943775103e-06,
"loss": 0.121,
"step": 101300
},
{
"epoch": 1.3166959999999999,
"grad_norm": 4.901790618896484,
"learning_rate": 1.899116465863454e-06,
"loss": 0.1248,
"step": 101400
},
{
"epoch": 1.317496,
"grad_norm": 3.4517650604248047,
"learning_rate": 1.891084337349398e-06,
"loss": 0.1101,
"step": 101500
},
{
"epoch": 1.3182960000000001,
"grad_norm": 2.9589340686798096,
"learning_rate": 1.8830522088353414e-06,
"loss": 0.1569,
"step": 101600
},
{
"epoch": 1.319096,
"grad_norm": 4.821226119995117,
"learning_rate": 1.8750200803212854e-06,
"loss": 0.1248,
"step": 101700
},
{
"epoch": 1.319896,
"grad_norm": 5.956510543823242,
"learning_rate": 1.866987951807229e-06,
"loss": 0.129,
"step": 101800
},
{
"epoch": 1.3206959999999999,
"grad_norm": 8.236087799072266,
"learning_rate": 1.8589558232931728e-06,
"loss": 0.1288,
"step": 101900
},
{
"epoch": 1.321496,
"grad_norm": 7.885124206542969,
"learning_rate": 1.8509236947791165e-06,
"loss": 0.133,
"step": 102000
},
{
"epoch": 1.3222960000000001,
"grad_norm": 4.618575096130371,
"learning_rate": 1.8428915662650604e-06,
"loss": 0.131,
"step": 102100
},
{
"epoch": 1.323096,
"grad_norm": 3.3860082626342773,
"learning_rate": 1.8348594377510043e-06,
"loss": 0.1275,
"step": 102200
},
{
"epoch": 1.323896,
"grad_norm": 0.5851134061813354,
"learning_rate": 1.8268273092369478e-06,
"loss": 0.1163,
"step": 102300
},
{
"epoch": 1.3246959999999999,
"grad_norm": 8.403120040893555,
"learning_rate": 1.8187951807228917e-06,
"loss": 0.142,
"step": 102400
},
{
"epoch": 1.325496,
"grad_norm": 4.019391059875488,
"learning_rate": 1.8107630522088354e-06,
"loss": 0.1234,
"step": 102500
},
{
"epoch": 1.326296,
"grad_norm": 7.64513635635376,
"learning_rate": 1.8028112449799198e-06,
"loss": 0.1202,
"step": 102600
},
{
"epoch": 1.327096,
"grad_norm": 4.8131489753723145,
"learning_rate": 1.7947791164658637e-06,
"loss": 0.1169,
"step": 102700
},
{
"epoch": 1.327896,
"grad_norm": 7.043356895446777,
"learning_rate": 1.7867469879518074e-06,
"loss": 0.1209,
"step": 102800
},
{
"epoch": 1.3286959999999999,
"grad_norm": 4.337855815887451,
"learning_rate": 1.7787148594377513e-06,
"loss": 0.1365,
"step": 102900
},
{
"epoch": 1.329496,
"grad_norm": 6.657166481018066,
"learning_rate": 1.7706827309236948e-06,
"loss": 0.1256,
"step": 103000
},
{
"epoch": 1.330296,
"grad_norm": 4.905508041381836,
"learning_rate": 1.7626506024096387e-06,
"loss": 0.1118,
"step": 103100
},
{
"epoch": 1.331096,
"grad_norm": 5.122402667999268,
"learning_rate": 1.7546184738955824e-06,
"loss": 0.1276,
"step": 103200
},
{
"epoch": 1.331896,
"grad_norm": 4.78582239151001,
"learning_rate": 1.7465863453815264e-06,
"loss": 0.1194,
"step": 103300
},
{
"epoch": 1.3326959999999999,
"grad_norm": 5.943844318389893,
"learning_rate": 1.7385542168674698e-06,
"loss": 0.1206,
"step": 103400
},
{
"epoch": 1.333496,
"grad_norm": 8.206818580627441,
"learning_rate": 1.7305220883534138e-06,
"loss": 0.1151,
"step": 103500
},
{
"epoch": 1.334296,
"grad_norm": 5.556063175201416,
"learning_rate": 1.7224899598393577e-06,
"loss": 0.1114,
"step": 103600
},
{
"epoch": 1.335096,
"grad_norm": 7.587924957275391,
"learning_rate": 1.7144578313253014e-06,
"loss": 0.1207,
"step": 103700
},
{
"epoch": 1.335896,
"grad_norm": 7.417220592498779,
"learning_rate": 1.706425702811245e-06,
"loss": 0.1163,
"step": 103800
},
{
"epoch": 1.3366959999999999,
"grad_norm": 7.316643238067627,
"learning_rate": 1.6983935742971888e-06,
"loss": 0.1157,
"step": 103900
},
{
"epoch": 1.337496,
"grad_norm": 1.7829804420471191,
"learning_rate": 1.6903614457831327e-06,
"loss": 0.1155,
"step": 104000
},
{
"epoch": 1.337496,
"eval_test1_cer": 0.045685824184031665,
"eval_test1_cer_norm": 0.03030710495353807,
"eval_test1_loss": 0.17226466536521912,
"eval_test1_runtime": 3385.0143,
"eval_test1_samples_per_second": 0.739,
"eval_test1_steps_per_second": 0.185,
"eval_test1_wer": 0.13615557304877693,
"eval_test1_wer_norm": 0.07755925070867062,
"step": 104000
},
{
"epoch": 1.337496,
"eval_test2_cer": 0.10362769253742488,
"eval_test2_cer_norm": 0.08312674310505114,
"eval_test2_loss": 0.31396326422691345,
"eval_test2_runtime": 3448.444,
"eval_test2_samples_per_second": 0.725,
"eval_test2_steps_per_second": 0.181,
"eval_test2_wer": 0.21904325932707713,
"eval_test2_wer_norm": 0.15513981205592484,
"step": 104000
},
{
"epoch": 1.338296,
"grad_norm": 3.8256428241729736,
"learning_rate": 1.6823293172690762e-06,
"loss": 0.1251,
"step": 104100
},
{
"epoch": 1.339096,
"grad_norm": 4.764732837677002,
"learning_rate": 1.6742971887550201e-06,
"loss": 0.1183,
"step": 104200
},
{
"epoch": 1.339896,
"grad_norm": 3.5565683841705322,
"learning_rate": 1.666265060240964e-06,
"loss": 0.1152,
"step": 104300
},
{
"epoch": 1.3406959999999999,
"grad_norm": 5.896035671234131,
"learning_rate": 1.6582329317269078e-06,
"loss": 0.136,
"step": 104400
},
{
"epoch": 1.341496,
"grad_norm": 6.5424628257751465,
"learning_rate": 1.6502008032128517e-06,
"loss": 0.128,
"step": 104500
},
{
"epoch": 1.342296,
"grad_norm": 5.04475736618042,
"learning_rate": 1.6421686746987952e-06,
"loss": 0.1149,
"step": 104600
},
{
"epoch": 1.343096,
"grad_norm": 5.732762813568115,
"learning_rate": 1.6341365461847391e-06,
"loss": 0.131,
"step": 104700
},
{
"epoch": 1.343896,
"grad_norm": 5.360403060913086,
"learning_rate": 1.6261044176706828e-06,
"loss": 0.1283,
"step": 104800
},
{
"epoch": 1.344696,
"grad_norm": 6.674147129058838,
"learning_rate": 1.6180722891566267e-06,
"loss": 0.1357,
"step": 104900
},
{
"epoch": 1.345496,
"grad_norm": 3.1230456829071045,
"learning_rate": 1.6100401606425702e-06,
"loss": 0.1237,
"step": 105000
},
{
"epoch": 1.346296,
"grad_norm": 6.3545756340026855,
"learning_rate": 1.6020080321285142e-06,
"loss": 0.1248,
"step": 105100
},
{
"epoch": 1.347096,
"grad_norm": 9.923174858093262,
"learning_rate": 1.593975903614458e-06,
"loss": 0.1382,
"step": 105200
},
{
"epoch": 1.347896,
"grad_norm": 3.6945223808288574,
"learning_rate": 1.5859437751004018e-06,
"loss": 0.1244,
"step": 105300
},
{
"epoch": 1.348696,
"grad_norm": 8.398360252380371,
"learning_rate": 1.5779116465863457e-06,
"loss": 0.123,
"step": 105400
},
{
"epoch": 1.349496,
"grad_norm": 1.4128965139389038,
"learning_rate": 1.5698795180722892e-06,
"loss": 0.1175,
"step": 105500
},
{
"epoch": 1.350296,
"grad_norm": 11.629409790039062,
"learning_rate": 1.5618473895582331e-06,
"loss": 0.1117,
"step": 105600
},
{
"epoch": 1.351096,
"grad_norm": 4.094301700592041,
"learning_rate": 1.5538152610441768e-06,
"loss": 0.1193,
"step": 105700
},
{
"epoch": 1.351896,
"grad_norm": 4.269906044006348,
"learning_rate": 1.5457831325301205e-06,
"loss": 0.1256,
"step": 105800
},
{
"epoch": 1.352696,
"grad_norm": 5.27727746963501,
"learning_rate": 1.5377510040160642e-06,
"loss": 0.1224,
"step": 105900
},
{
"epoch": 1.353496,
"grad_norm": 4.353626728057861,
"learning_rate": 1.5297188755020082e-06,
"loss": 0.1196,
"step": 106000
},
{
"epoch": 1.354296,
"grad_norm": 3.226297616958618,
"learning_rate": 1.521686746987952e-06,
"loss": 0.1364,
"step": 106100
},
{
"epoch": 1.355096,
"grad_norm": 0.7694222331047058,
"learning_rate": 1.5136546184738956e-06,
"loss": 0.103,
"step": 106200
},
{
"epoch": 1.355896,
"grad_norm": 3.6190733909606934,
"learning_rate": 1.5056224899598395e-06,
"loss": 0.1235,
"step": 106300
},
{
"epoch": 1.356696,
"grad_norm": 3.509673833847046,
"learning_rate": 1.4975903614457832e-06,
"loss": 0.121,
"step": 106400
},
{
"epoch": 1.357496,
"grad_norm": 8.607564926147461,
"learning_rate": 1.4895582329317271e-06,
"loss": 0.1318,
"step": 106500
},
{
"epoch": 1.358296,
"grad_norm": 3.327017307281494,
"learning_rate": 1.4815261044176706e-06,
"loss": 0.1283,
"step": 106600
},
{
"epoch": 1.359096,
"grad_norm": 8.573492050170898,
"learning_rate": 1.4735742971887552e-06,
"loss": 0.1222,
"step": 106700
},
{
"epoch": 1.359896,
"grad_norm": 4.089103698730469,
"learning_rate": 1.465542168674699e-06,
"loss": 0.1257,
"step": 106800
},
{
"epoch": 1.360696,
"grad_norm": 3.942911386489868,
"learning_rate": 1.4575100401606426e-06,
"loss": 0.1205,
"step": 106900
},
{
"epoch": 1.361496,
"grad_norm": 6.010968208312988,
"learning_rate": 1.4494779116465865e-06,
"loss": 0.1295,
"step": 107000
},
{
"epoch": 1.362296,
"grad_norm": 6.7945756912231445,
"learning_rate": 1.4414457831325302e-06,
"loss": 0.1246,
"step": 107100
},
{
"epoch": 1.363096,
"grad_norm": 2.255643129348755,
"learning_rate": 1.4334136546184741e-06,
"loss": 0.1254,
"step": 107200
},
{
"epoch": 1.363896,
"grad_norm": 4.047929763793945,
"learning_rate": 1.4253815261044176e-06,
"loss": 0.1197,
"step": 107300
},
{
"epoch": 1.364696,
"grad_norm": 4.005350112915039,
"learning_rate": 1.4173493975903615e-06,
"loss": 0.1261,
"step": 107400
},
{
"epoch": 1.365496,
"grad_norm": 5.783274173736572,
"learning_rate": 1.4093172690763055e-06,
"loss": 0.1157,
"step": 107500
},
{
"epoch": 1.366296,
"grad_norm": 5.45566463470459,
"learning_rate": 1.4012851405622492e-06,
"loss": 0.125,
"step": 107600
},
{
"epoch": 1.367096,
"grad_norm": 4.851735591888428,
"learning_rate": 1.3932530120481929e-06,
"loss": 0.1242,
"step": 107700
},
{
"epoch": 1.367896,
"grad_norm": 5.266098976135254,
"learning_rate": 1.3852208835341366e-06,
"loss": 0.1267,
"step": 107800
},
{
"epoch": 1.368696,
"grad_norm": 4.1477742195129395,
"learning_rate": 1.3771887550200805e-06,
"loss": 0.1158,
"step": 107900
},
{
"epoch": 1.369496,
"grad_norm": 12.537534713745117,
"learning_rate": 1.369156626506024e-06,
"loss": 0.1114,
"step": 108000
},
{
"epoch": 1.369496,
"eval_test1_cer": 0.041395922025543354,
"eval_test1_cer_norm": 0.027075179484716785,
"eval_test1_loss": 0.1729966700077057,
"eval_test1_runtime": 3379.841,
"eval_test1_samples_per_second": 0.74,
"eval_test1_steps_per_second": 0.185,
"eval_test1_wer": 0.12840024490509927,
"eval_test1_wer_norm": 0.06993190917326632,
"step": 108000
},
{
"epoch": 1.369496,
"eval_test2_cer": 0.1033523724194572,
"eval_test2_cer_norm": 0.077180818097304,
"eval_test2_loss": 0.3105059862136841,
"eval_test2_runtime": 3449.5535,
"eval_test2_samples_per_second": 0.725,
"eval_test2_steps_per_second": 0.181,
"eval_test2_wer": 0.2153238727397574,
"eval_test2_wer_norm": 0.15038391015356406,
"step": 108000
},
{
"epoch": 1.370296,
"grad_norm": 2.571197986602783,
"learning_rate": 1.361124497991968e-06,
"loss": 0.1133,
"step": 108100
},
{
"epoch": 1.371096,
"grad_norm": 2.839953660964966,
"learning_rate": 1.3531726907630524e-06,
"loss": 0.1225,
"step": 108200
},
{
"epoch": 1.371896,
"grad_norm": 2.108365535736084,
"learning_rate": 1.345140562248996e-06,
"loss": 0.1111,
"step": 108300
},
{
"epoch": 1.372696,
"grad_norm": 4.538883686065674,
"learning_rate": 1.3371084337349399e-06,
"loss": 0.0988,
"step": 108400
},
{
"epoch": 1.373496,
"grad_norm": 5.8930277824401855,
"learning_rate": 1.3290763052208836e-06,
"loss": 0.1124,
"step": 108500
},
{
"epoch": 1.374296,
"grad_norm": 3.4291086196899414,
"learning_rate": 1.3210441767068275e-06,
"loss": 0.0998,
"step": 108600
},
{
"epoch": 1.375096,
"grad_norm": 3.634300947189331,
"learning_rate": 1.313012048192771e-06,
"loss": 0.123,
"step": 108700
},
{
"epoch": 1.375896,
"grad_norm": 10.641935348510742,
"learning_rate": 1.304979919678715e-06,
"loss": 0.1331,
"step": 108800
},
{
"epoch": 1.376696,
"grad_norm": 3.8561360836029053,
"learning_rate": 1.2969477911646588e-06,
"loss": 0.1386,
"step": 108900
},
{
"epoch": 1.377496,
"grad_norm": 5.578998565673828,
"learning_rate": 1.2889156626506025e-06,
"loss": 0.1252,
"step": 109000
},
{
"epoch": 1.378296,
"grad_norm": 6.097825050354004,
"learning_rate": 1.2808835341365464e-06,
"loss": 0.1226,
"step": 109100
},
{
"epoch": 1.379096,
"grad_norm": 7.930078983306885,
"learning_rate": 1.27285140562249e-06,
"loss": 0.1324,
"step": 109200
},
{
"epoch": 1.379896,
"grad_norm": 4.191389083862305,
"learning_rate": 1.2648192771084339e-06,
"loss": 0.1265,
"step": 109300
},
{
"epoch": 1.380696,
"grad_norm": 4.194836139678955,
"learning_rate": 1.2567871485943776e-06,
"loss": 0.1131,
"step": 109400
},
{
"epoch": 1.381496,
"grad_norm": 0.9715979695320129,
"learning_rate": 1.2487550200803215e-06,
"loss": 0.1149,
"step": 109500
},
{
"epoch": 1.382296,
"grad_norm": 2.0711517333984375,
"learning_rate": 1.2407228915662652e-06,
"loss": 0.1196,
"step": 109600
},
{
"epoch": 1.383096,
"grad_norm": 3.178311824798584,
"learning_rate": 1.232690763052209e-06,
"loss": 0.1116,
"step": 109700
},
{
"epoch": 1.383896,
"grad_norm": 4.997854232788086,
"learning_rate": 1.2246586345381526e-06,
"loss": 0.1215,
"step": 109800
},
{
"epoch": 1.384696,
"grad_norm": 4.431860446929932,
"learning_rate": 1.2166265060240963e-06,
"loss": 0.1129,
"step": 109900
},
{
"epoch": 1.385496,
"grad_norm": 4.320796966552734,
"learning_rate": 1.2085943775100402e-06,
"loss": 0.1375,
"step": 110000
},
{
"epoch": 1.386296,
"grad_norm": 7.0986552238464355,
"learning_rate": 1.2005622489959842e-06,
"loss": 0.1051,
"step": 110100
},
{
"epoch": 1.387096,
"grad_norm": 2.893299102783203,
"learning_rate": 1.1925301204819279e-06,
"loss": 0.1133,
"step": 110200
},
{
"epoch": 1.387896,
"grad_norm": 1.8434147834777832,
"learning_rate": 1.1844979919678716e-06,
"loss": 0.1099,
"step": 110300
},
{
"epoch": 1.388696,
"grad_norm": 3.1360116004943848,
"learning_rate": 1.1764658634538153e-06,
"loss": 0.1276,
"step": 110400
},
{
"epoch": 1.389496,
"grad_norm": 2.8671581745147705,
"learning_rate": 1.1684337349397592e-06,
"loss": 0.1113,
"step": 110500
},
{
"epoch": 1.390296,
"grad_norm": 8.354147911071777,
"learning_rate": 1.160401606425703e-06,
"loss": 0.1085,
"step": 110600
},
{
"epoch": 1.391096,
"grad_norm": 14.839973449707031,
"learning_rate": 1.1523694779116466e-06,
"loss": 0.1096,
"step": 110700
},
{
"epoch": 1.391896,
"grad_norm": 3.917928695678711,
"learning_rate": 1.1443373493975903e-06,
"loss": 0.1219,
"step": 110800
},
{
"epoch": 1.392696,
"grad_norm": 6.117668151855469,
"learning_rate": 1.1363052208835343e-06,
"loss": 0.1115,
"step": 110900
},
{
"epoch": 1.393496,
"grad_norm": 1.9348158836364746,
"learning_rate": 1.128273092369478e-06,
"loss": 0.1219,
"step": 111000
},
{
"epoch": 1.394296,
"grad_norm": 5.231213569641113,
"learning_rate": 1.1202409638554219e-06,
"loss": 0.1178,
"step": 111100
},
{
"epoch": 1.3950960000000001,
"grad_norm": 5.860185623168945,
"learning_rate": 1.1122088353413656e-06,
"loss": 0.1136,
"step": 111200
},
{
"epoch": 1.395896,
"grad_norm": 4.386104106903076,
"learning_rate": 1.10425702811245e-06,
"loss": 0.1159,
"step": 111300
},
{
"epoch": 1.396696,
"grad_norm": 1.6194918155670166,
"learning_rate": 1.0962248995983938e-06,
"loss": 0.1187,
"step": 111400
},
{
"epoch": 1.397496,
"grad_norm": 5.940235137939453,
"learning_rate": 1.0881927710843375e-06,
"loss": 0.1157,
"step": 111500
},
{
"epoch": 1.398296,
"grad_norm": 4.942705154418945,
"learning_rate": 1.0801606425702812e-06,
"loss": 0.1005,
"step": 111600
},
{
"epoch": 1.3990960000000001,
"grad_norm": 5.09475564956665,
"learning_rate": 1.072128514056225e-06,
"loss": 0.1204,
"step": 111700
},
{
"epoch": 1.399896,
"grad_norm": 5.27365255355835,
"learning_rate": 1.0640963855421687e-06,
"loss": 0.127,
"step": 111800
},
{
"epoch": 1.400696,
"grad_norm": 6.454215049743652,
"learning_rate": 1.0560642570281126e-06,
"loss": 0.1025,
"step": 111900
},
{
"epoch": 1.401496,
"grad_norm": 5.676671028137207,
"learning_rate": 1.0480321285140563e-06,
"loss": 0.1141,
"step": 112000
},
{
"epoch": 1.401496,
"eval_test1_cer": 0.04145660616924341,
"eval_test1_cer_norm": 0.02695992508463995,
"eval_test1_loss": 0.17103232443332672,
"eval_test1_runtime": 2490.4053,
"eval_test1_samples_per_second": 1.004,
"eval_test1_steps_per_second": 0.251,
"eval_test1_wer": 0.12912912912912913,
"eval_test1_wer_norm": 0.07016569741956223,
"step": 112000
},
{
"epoch": 1.401496,
"eval_test2_cer": 0.08579254862433269,
"eval_test2_cer_norm": 0.06814572358227455,
"eval_test2_loss": 0.3084600865840912,
"eval_test2_runtime": 2471.9908,
"eval_test2_samples_per_second": 1.011,
"eval_test2_steps_per_second": 0.253,
"eval_test2_wer": 0.19641222247653925,
"eval_test2_wer_norm": 0.13070135228054092,
"step": 112000
},
{
"epoch": 1.402296,
"grad_norm": 2.858332633972168,
"learning_rate": 1.04e-06,
"loss": 0.1044,
"step": 112100
},
{
"epoch": 1.4030960000000001,
"grad_norm": 5.488587379455566,
"learning_rate": 1.0319678714859437e-06,
"loss": 0.1366,
"step": 112200
},
{
"epoch": 1.403896,
"grad_norm": 13.426318168640137,
"learning_rate": 1.0239357429718876e-06,
"loss": 0.1186,
"step": 112300
},
{
"epoch": 1.404696,
"grad_norm": 4.668258190155029,
"learning_rate": 1.0159036144578315e-06,
"loss": 0.1056,
"step": 112400
},
{
"epoch": 1.405496,
"grad_norm": 3.497905969619751,
"learning_rate": 1.0078714859437753e-06,
"loss": 0.1168,
"step": 112500
},
{
"epoch": 1.406296,
"grad_norm": 5.060064315795898,
"learning_rate": 9.99839357429719e-07,
"loss": 0.1167,
"step": 112600
},
{
"epoch": 1.4070960000000001,
"grad_norm": 1.985062837600708,
"learning_rate": 9.918072289156627e-07,
"loss": 0.1234,
"step": 112700
},
{
"epoch": 1.407896,
"grad_norm": 3.2145867347717285,
"learning_rate": 9.837751004016064e-07,
"loss": 0.1231,
"step": 112800
},
{
"epoch": 1.408696,
"grad_norm": 6.194746971130371,
"learning_rate": 9.757429718875503e-07,
"loss": 0.1038,
"step": 112900
},
{
"epoch": 1.409496,
"grad_norm": 4.408815860748291,
"learning_rate": 9.67710843373494e-07,
"loss": 0.1157,
"step": 113000
},
{
"epoch": 1.410296,
"grad_norm": 4.587121486663818,
"learning_rate": 9.59678714859438e-07,
"loss": 0.1163,
"step": 113100
},
{
"epoch": 1.4110960000000001,
"grad_norm": 4.01561975479126,
"learning_rate": 9.516465863453816e-07,
"loss": 0.1077,
"step": 113200
},
{
"epoch": 1.411896,
"grad_norm": 4.253355026245117,
"learning_rate": 9.436144578313254e-07,
"loss": 0.1209,
"step": 113300
},
{
"epoch": 1.412696,
"grad_norm": 1.5479423999786377,
"learning_rate": 9.355823293172692e-07,
"loss": 0.1378,
"step": 113400
},
{
"epoch": 1.4134959999999999,
"grad_norm": 4.244668960571289,
"learning_rate": 9.27550200803213e-07,
"loss": 0.1381,
"step": 113500
},
{
"epoch": 1.414296,
"grad_norm": 4.061355113983154,
"learning_rate": 9.195180722891567e-07,
"loss": 0.1283,
"step": 113600
},
{
"epoch": 1.4150960000000001,
"grad_norm": 8.999344825744629,
"learning_rate": 9.114859437751005e-07,
"loss": 0.1291,
"step": 113700
},
{
"epoch": 1.415896,
"grad_norm": 3.263817310333252,
"learning_rate": 9.034538152610442e-07,
"loss": 0.119,
"step": 113800
},
{
"epoch": 1.416696,
"grad_norm": 5.394161224365234,
"learning_rate": 8.954216867469879e-07,
"loss": 0.1218,
"step": 113900
},
{
"epoch": 1.4174959999999999,
"grad_norm": 11.564488410949707,
"learning_rate": 8.873895582329318e-07,
"loss": 0.1222,
"step": 114000
},
{
"epoch": 1.418296,
"grad_norm": 3.848464250564575,
"learning_rate": 8.793574297188756e-07,
"loss": 0.1175,
"step": 114100
},
{
"epoch": 1.4190960000000001,
"grad_norm": 3.4740781784057617,
"learning_rate": 8.713253012048194e-07,
"loss": 0.1267,
"step": 114200
},
{
"epoch": 1.419896,
"grad_norm": 3.503969192504883,
"learning_rate": 8.632931726907632e-07,
"loss": 0.1333,
"step": 114300
},
{
"epoch": 1.420696,
"grad_norm": 3.658344268798828,
"learning_rate": 8.552610441767069e-07,
"loss": 0.1137,
"step": 114400
},
{
"epoch": 1.4214959999999999,
"grad_norm": 5.1479973793029785,
"learning_rate": 8.472289156626507e-07,
"loss": 0.1254,
"step": 114500
},
{
"epoch": 1.422296,
"grad_norm": 5.20347261428833,
"learning_rate": 8.391967871485944e-07,
"loss": 0.1167,
"step": 114600
},
{
"epoch": 1.4230960000000001,
"grad_norm": 5.197832107543945,
"learning_rate": 8.311646586345381e-07,
"loss": 0.1297,
"step": 114700
},
{
"epoch": 1.423896,
"grad_norm": 2.3857178688049316,
"learning_rate": 8.23132530120482e-07,
"loss": 0.1173,
"step": 114800
},
{
"epoch": 1.424696,
"grad_norm": 5.162271976470947,
"learning_rate": 8.151004016064258e-07,
"loss": 0.1089,
"step": 114900
},
{
"epoch": 1.4254959999999999,
"grad_norm": 4.035723686218262,
"learning_rate": 8.070682730923695e-07,
"loss": 0.1186,
"step": 115000
},
{
"epoch": 1.426296,
"grad_norm": 7.1610283851623535,
"learning_rate": 7.991164658634539e-07,
"loss": 0.1196,
"step": 115100
},
{
"epoch": 1.427096,
"grad_norm": 6.167757987976074,
"learning_rate": 7.910843373493976e-07,
"loss": 0.112,
"step": 115200
},
{
"epoch": 1.427896,
"grad_norm": 8.109100341796875,
"learning_rate": 7.830522088353415e-07,
"loss": 0.1092,
"step": 115300
},
{
"epoch": 1.428696,
"grad_norm": 2.2968735694885254,
"learning_rate": 7.750200803212853e-07,
"loss": 0.1058,
"step": 115400
},
{
"epoch": 1.4294959999999999,
"grad_norm": 4.020580768585205,
"learning_rate": 7.66987951807229e-07,
"loss": 0.1089,
"step": 115500
},
{
"epoch": 1.430296,
"grad_norm": 1.7853455543518066,
"learning_rate": 7.589558232931728e-07,
"loss": 0.1077,
"step": 115600
},
{
"epoch": 1.431096,
"grad_norm": 2.6213812828063965,
"learning_rate": 7.509236947791165e-07,
"loss": 0.114,
"step": 115700
},
{
"epoch": 1.431896,
"grad_norm": 7.169987678527832,
"learning_rate": 7.428915662650602e-07,
"loss": 0.1247,
"step": 115800
},
{
"epoch": 1.432696,
"grad_norm": 5.280975818634033,
"learning_rate": 7.348594377510041e-07,
"loss": 0.1117,
"step": 115900
},
{
"epoch": 1.4334959999999999,
"grad_norm": 7.396920204162598,
"learning_rate": 7.268273092369478e-07,
"loss": 0.1218,
"step": 116000
},
{
"epoch": 1.4334959999999999,
"eval_test1_cer": 0.04053700799163493,
"eval_test1_cer_norm": 0.02625399188416933,
"eval_test1_loss": 0.16978052258491516,
"eval_test1_runtime": 2459.5813,
"eval_test1_samples_per_second": 1.016,
"eval_test1_steps_per_second": 0.254,
"eval_test1_wer": 0.12650514592262166,
"eval_test1_wer_norm": 0.06905520324965662,
"step": 116000
},
{
"epoch": 1.4334959999999999,
"eval_test2_cer": 0.0906269832381379,
"eval_test2_cer_norm": 0.07077006507592191,
"eval_test2_loss": 0.30657845735549927,
"eval_test2_runtime": 2479.2829,
"eval_test2_samples_per_second": 1.008,
"eval_test2_steps_per_second": 0.252,
"eval_test2_wer": 0.20224879835202564,
"eval_test2_wer_norm": 0.13654595461838184,
"step": 116000
},
{
"epoch": 1.434296,
"grad_norm": 7.435978412628174,
"learning_rate": 7.187951807228916e-07,
"loss": 0.1312,
"step": 116100
},
{
"epoch": 1.435096,
"grad_norm": 6.736993789672852,
"learning_rate": 7.107630522088355e-07,
"loss": 0.1245,
"step": 116200
},
{
"epoch": 1.435896,
"grad_norm": 3.6509387493133545,
"learning_rate": 7.027309236947792e-07,
"loss": 0.1041,
"step": 116300
},
{
"epoch": 1.436696,
"grad_norm": 5.3854851722717285,
"learning_rate": 6.94698795180723e-07,
"loss": 0.126,
"step": 116400
},
{
"epoch": 1.4374959999999999,
"grad_norm": 5.283120155334473,
"learning_rate": 6.866666666666667e-07,
"loss": 0.1615,
"step": 116500
},
{
"epoch": 1.438296,
"grad_norm": 3.9473323822021484,
"learning_rate": 6.786345381526105e-07,
"loss": 0.1218,
"step": 116600
},
{
"epoch": 1.439096,
"grad_norm": 11.495280265808105,
"learning_rate": 6.706024096385542e-07,
"loss": 0.1165,
"step": 116700
},
{
"epoch": 1.439896,
"grad_norm": 6.106124401092529,
"learning_rate": 6.62570281124498e-07,
"loss": 0.1291,
"step": 116800
},
{
"epoch": 1.440696,
"grad_norm": 3.252533435821533,
"learning_rate": 6.545381526104418e-07,
"loss": 0.1302,
"step": 116900
},
{
"epoch": 1.4414959999999999,
"grad_norm": 4.417702674865723,
"learning_rate": 6.465060240963857e-07,
"loss": 0.1185,
"step": 117000
},
{
"epoch": 1.442296,
"grad_norm": 7.446925163269043,
"learning_rate": 6.384738955823294e-07,
"loss": 0.1181,
"step": 117100
},
{
"epoch": 1.443096,
"grad_norm": 4.945169448852539,
"learning_rate": 6.304417670682732e-07,
"loss": 0.1146,
"step": 117200
},
{
"epoch": 1.443896,
"grad_norm": 4.255674362182617,
"learning_rate": 6.224096385542169e-07,
"loss": 0.1149,
"step": 117300
},
{
"epoch": 1.444696,
"grad_norm": 7.243592739105225,
"learning_rate": 6.143775100401607e-07,
"loss": 0.1266,
"step": 117400
},
{
"epoch": 1.445496,
"grad_norm": 4.7151265144348145,
"learning_rate": 6.063453815261044e-07,
"loss": 0.1228,
"step": 117500
},
{
"epoch": 1.446296,
"grad_norm": 3.492438793182373,
"learning_rate": 5.983132530120483e-07,
"loss": 0.107,
"step": 117600
},
{
"epoch": 1.447096,
"grad_norm": 5.758606910705566,
"learning_rate": 5.902811244979921e-07,
"loss": 0.1104,
"step": 117700
},
{
"epoch": 1.447896,
"grad_norm": 5.847075462341309,
"learning_rate": 5.822489959839358e-07,
"loss": 0.1234,
"step": 117800
},
{
"epoch": 1.448696,
"grad_norm": 4.088864803314209,
"learning_rate": 5.742168674698796e-07,
"loss": 0.1247,
"step": 117900
},
{
"epoch": 1.449496,
"grad_norm": 2.7524282932281494,
"learning_rate": 5.661847389558233e-07,
"loss": 0.0971,
"step": 118000
},
{
"epoch": 1.450296,
"grad_norm": 6.989465236663818,
"learning_rate": 5.581526104417671e-07,
"loss": 0.126,
"step": 118100
},
{
"epoch": 1.451096,
"grad_norm": 9.567557334899902,
"learning_rate": 5.501204819277109e-07,
"loss": 0.1191,
"step": 118200
},
{
"epoch": 1.451896,
"grad_norm": 10.12435531616211,
"learning_rate": 5.420883534136546e-07,
"loss": 0.1096,
"step": 118300
},
{
"epoch": 1.452696,
"grad_norm": 1.7563780546188354,
"learning_rate": 5.340562248995985e-07,
"loss": 0.1197,
"step": 118400
},
{
"epoch": 1.453496,
"grad_norm": 5.218570232391357,
"learning_rate": 5.260240963855423e-07,
"loss": 0.1207,
"step": 118500
},
{
"epoch": 1.454296,
"grad_norm": 4.665302753448486,
"learning_rate": 5.17991967871486e-07,
"loss": 0.115,
"step": 118600
},
{
"epoch": 1.455096,
"grad_norm": 5.446451663970947,
"learning_rate": 5.099598393574298e-07,
"loss": 0.119,
"step": 118700
},
{
"epoch": 1.455896,
"grad_norm": 4.688202381134033,
"learning_rate": 5.020080321285141e-07,
"loss": 0.1369,
"step": 118800
},
{
"epoch": 1.456696,
"grad_norm": 3.5095245838165283,
"learning_rate": 4.939759036144578e-07,
"loss": 0.1094,
"step": 118900
},
{
"epoch": 1.457496,
"grad_norm": 3.919813394546509,
"learning_rate": 4.859437751004017e-07,
"loss": 0.1113,
"step": 119000
},
{
"epoch": 1.458296,
"grad_norm": 3.1427431106567383,
"learning_rate": 4.779116465863454e-07,
"loss": 0.1127,
"step": 119100
},
{
"epoch": 1.459096,
"grad_norm": 5.548600673675537,
"learning_rate": 4.698795180722892e-07,
"loss": 0.1175,
"step": 119200
},
{
"epoch": 1.459896,
"grad_norm": 3.4401326179504395,
"learning_rate": 4.6184738955823296e-07,
"loss": 0.1118,
"step": 119300
},
{
"epoch": 1.460696,
"grad_norm": 0.7513042688369751,
"learning_rate": 4.538152610441767e-07,
"loss": 0.1227,
"step": 119400
},
{
"epoch": 1.461496,
"grad_norm": 4.836828708648682,
"learning_rate": 4.4578313253012054e-07,
"loss": 0.1208,
"step": 119500
},
{
"epoch": 1.462296,
"grad_norm": 9.92773723602295,
"learning_rate": 4.377510040160643e-07,
"loss": 0.1021,
"step": 119600
},
{
"epoch": 1.463096,
"grad_norm": 4.6646575927734375,
"learning_rate": 4.2971887550200806e-07,
"loss": 0.1142,
"step": 119700
},
{
"epoch": 1.463896,
"grad_norm": 0.4946214258670807,
"learning_rate": 4.216867469879518e-07,
"loss": 0.1334,
"step": 119800
},
{
"epoch": 1.464696,
"grad_norm": 7.3409247398376465,
"learning_rate": 4.1365461847389564e-07,
"loss": 0.1387,
"step": 119900
},
{
"epoch": 1.465496,
"grad_norm": 2.3244385719299316,
"learning_rate": 4.056224899598394e-07,
"loss": 0.1216,
"step": 120000
},
{
"epoch": 1.465496,
"eval_test1_cer": 0.04472888191799238,
"eval_test1_cer_norm": 0.029865296419910196,
"eval_test1_loss": 0.16931886970996857,
"eval_test1_runtime": 2461.9854,
"eval_test1_samples_per_second": 1.015,
"eval_test1_steps_per_second": 0.254,
"eval_test1_wer": 0.13493104755240679,
"eval_test1_wer_norm": 0.076477980069552,
"step": 120000
},
{
"epoch": 1.465496,
"eval_test2_cer": 0.09805129353791019,
"eval_test2_cer_norm": 0.07766985590331578,
"eval_test2_loss": 0.3047462999820709,
"eval_test2_runtime": 2488.9087,
"eval_test2_samples_per_second": 1.004,
"eval_test2_steps_per_second": 0.251,
"eval_test2_wer": 0.21137560082398718,
"eval_test2_wer_norm": 0.1462296584918634,
"step": 120000
},
{
"epoch": 1.466296,
"grad_norm": 5.934924602508545,
"learning_rate": 3.9759036144578316e-07,
"loss": 0.1166,
"step": 120100
},
{
"epoch": 1.467096,
"grad_norm": 6.291118621826172,
"learning_rate": 3.895582329317269e-07,
"loss": 0.1144,
"step": 120200
},
{
"epoch": 1.467896,
"grad_norm": 7.283012866973877,
"learning_rate": 3.8152610441767073e-07,
"loss": 0.1258,
"step": 120300
},
{
"epoch": 1.468696,
"grad_norm": 2.082486391067505,
"learning_rate": 3.734939759036145e-07,
"loss": 0.1209,
"step": 120400
},
{
"epoch": 1.469496,
"grad_norm": 7.8560967445373535,
"learning_rate": 3.6546184738955826e-07,
"loss": 0.115,
"step": 120500
},
{
"epoch": 1.470296,
"grad_norm": 6.016510963439941,
"learning_rate": 3.57429718875502e-07,
"loss": 0.1082,
"step": 120600
},
{
"epoch": 1.471096,
"grad_norm": 2.1153485774993896,
"learning_rate": 3.4939759036144583e-07,
"loss": 0.1286,
"step": 120700
},
{
"epoch": 1.471896,
"grad_norm": 5.02449369430542,
"learning_rate": 3.413654618473896e-07,
"loss": 0.111,
"step": 120800
},
{
"epoch": 1.472696,
"grad_norm": 2.6848514080047607,
"learning_rate": 3.3333333333333335e-07,
"loss": 0.1161,
"step": 120900
},
{
"epoch": 1.473496,
"grad_norm": 3.306321620941162,
"learning_rate": 3.253012048192771e-07,
"loss": 0.1164,
"step": 121000
},
{
"epoch": 1.474296,
"grad_norm": 5.202853202819824,
"learning_rate": 3.1726907630522093e-07,
"loss": 0.1148,
"step": 121100
},
{
"epoch": 1.475096,
"grad_norm": 6.273738861083984,
"learning_rate": 3.092369477911647e-07,
"loss": 0.1257,
"step": 121200
},
{
"epoch": 1.475896,
"grad_norm": 1.6317389011383057,
"learning_rate": 3.01285140562249e-07,
"loss": 0.1169,
"step": 121300
},
{
"epoch": 1.476696,
"grad_norm": 3.1588313579559326,
"learning_rate": 2.932530120481928e-07,
"loss": 0.1243,
"step": 121400
},
{
"epoch": 1.477496,
"grad_norm": 3.1574482917785645,
"learning_rate": 2.852208835341366e-07,
"loss": 0.1254,
"step": 121500
},
{
"epoch": 1.478296,
"grad_norm": 3.313225507736206,
"learning_rate": 2.7718875502008034e-07,
"loss": 0.1196,
"step": 121600
},
{
"epoch": 1.479096,
"grad_norm": 3.645125389099121,
"learning_rate": 2.691566265060241e-07,
"loss": 0.0956,
"step": 121700
},
{
"epoch": 1.479896,
"grad_norm": 0.8086249232292175,
"learning_rate": 2.611244979919679e-07,
"loss": 0.1171,
"step": 121800
},
{
"epoch": 1.480696,
"grad_norm": 4.813057899475098,
"learning_rate": 2.530923694779117e-07,
"loss": 0.1254,
"step": 121900
},
{
"epoch": 1.481496,
"grad_norm": 3.0106194019317627,
"learning_rate": 2.4506024096385544e-07,
"loss": 0.109,
"step": 122000
},
{
"epoch": 1.482296,
"grad_norm": 3.1917545795440674,
"learning_rate": 2.370281124497992e-07,
"loss": 0.1082,
"step": 122100
},
{
"epoch": 1.483096,
"grad_norm": 8.249537467956543,
"learning_rate": 2.28995983935743e-07,
"loss": 0.1334,
"step": 122200
},
{
"epoch": 1.483896,
"grad_norm": 2.7427356243133545,
"learning_rate": 2.2096385542168677e-07,
"loss": 0.1209,
"step": 122300
},
{
"epoch": 1.484696,
"grad_norm": 2.4920859336853027,
"learning_rate": 2.1293172690763056e-07,
"loss": 0.1041,
"step": 122400
},
{
"epoch": 1.485496,
"grad_norm": 5.013920307159424,
"learning_rate": 2.0489959839357432e-07,
"loss": 0.1114,
"step": 122500
},
{
"epoch": 1.486296,
"grad_norm": 0.5811383128166199,
"learning_rate": 1.968674698795181e-07,
"loss": 0.1228,
"step": 122600
},
{
"epoch": 1.487096,
"grad_norm": 5.773433685302734,
"learning_rate": 1.8883534136546187e-07,
"loss": 0.1284,
"step": 122700
},
{
"epoch": 1.487896,
"grad_norm": 3.8643128871917725,
"learning_rate": 1.8080321285140566e-07,
"loss": 0.1159,
"step": 122800
},
{
"epoch": 1.488696,
"grad_norm": 5.357062816619873,
"learning_rate": 1.7277108433734942e-07,
"loss": 0.1225,
"step": 122900
},
{
"epoch": 1.489496,
"grad_norm": 4.911788463592529,
"learning_rate": 1.647389558232932e-07,
"loss": 0.1116,
"step": 123000
},
{
"epoch": 1.490296,
"grad_norm": 4.041905403137207,
"learning_rate": 1.5670682730923697e-07,
"loss": 0.1236,
"step": 123100
},
{
"epoch": 1.491096,
"grad_norm": 5.671024799346924,
"learning_rate": 1.4867469879518073e-07,
"loss": 0.1168,
"step": 123200
},
{
"epoch": 1.491896,
"grad_norm": 8.380229949951172,
"learning_rate": 1.4064257028112452e-07,
"loss": 0.1192,
"step": 123300
},
{
"epoch": 1.492696,
"grad_norm": 8.371071815490723,
"learning_rate": 1.3261044176706828e-07,
"loss": 0.102,
"step": 123400
},
{
"epoch": 1.493496,
"grad_norm": 4.045628547668457,
"learning_rate": 1.2457831325301207e-07,
"loss": 0.1177,
"step": 123500
},
{
"epoch": 1.494296,
"grad_norm": 7.542957782745361,
"learning_rate": 1.1654618473895584e-07,
"loss": 0.108,
"step": 123600
},
{
"epoch": 2.000592,
"grad_norm": 5.552263259887695,
"learning_rate": 1.0851405622489961e-07,
"loss": 0.1093,
"step": 123700
},
{
"epoch": 2.001392,
"grad_norm": 4.822888374328613,
"learning_rate": 1.0048192771084339e-07,
"loss": 0.1042,
"step": 123800
},
{
"epoch": 2.002192,
"grad_norm": 0.32907435297966003,
"learning_rate": 9.244979919678716e-08,
"loss": 0.1014,
"step": 123900
},
{
"epoch": 2.002992,
"grad_norm": 5.526859760284424,
"learning_rate": 8.441767068273094e-08,
"loss": 0.0819,
"step": 124000
},
{
"epoch": 2.002992,
"eval_test1_cer": 0.04037829561580402,
"eval_test1_cer_norm": 0.026181957884121306,
"eval_test1_loss": 0.1684691458940506,
"eval_test1_runtime": 2457.6805,
"eval_test1_samples_per_second": 1.017,
"eval_test1_steps_per_second": 0.254,
"eval_test1_wer": 0.1262135922330097,
"eval_test1_wer_norm": 0.06817849732604693,
"step": 124000
},
{
"epoch": 2.002992,
"eval_test2_cer": 0.09239089857020197,
"eval_test2_cer_norm": 0.07192245119305857,
"eval_test2_loss": 0.30393916368484497,
"eval_test2_runtime": 2485.4178,
"eval_test2_samples_per_second": 1.006,
"eval_test2_steps_per_second": 0.251,
"eval_test2_wer": 0.20482375829709315,
"eval_test2_wer_norm": 0.13981205592482238,
"step": 124000
}
],
"logging_steps": 100,
"max_steps": 125000,
"num_input_tokens_seen": 0,
"num_train_epochs": 9223372036854775807,
"save_steps": 4000,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": false
},
"attributes": {}
}
},
"total_flos": 5.062139105004749e+20,
"train_batch_size": 4,
"trial_name": null,
"trial_params": null
}