{ "best_metric": 22.283451007404857, "best_model_checkpoint": "./whisper-large-v2/second2/checkpoint-500", "epoch": 0.2133902373966391, "eval_steps": 500, "global_step": 1000, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0010669511869831954, "grad_norm": 19.418699264526367, "learning_rate": 3.3333333333333334e-08, "loss": 2.4187, "step": 5 }, { "epoch": 0.002133902373966391, "grad_norm": 21.75494956970215, "learning_rate": 6.666666666666667e-08, "loss": 2.3362, "step": 10 }, { "epoch": 0.0032008535609495867, "grad_norm": 21.226938247680664, "learning_rate": 1e-07, "loss": 2.4288, "step": 15 }, { "epoch": 0.004267804747932782, "grad_norm": 27.21125602722168, "learning_rate": 1.3333333333333334e-07, "loss": 2.4004, "step": 20 }, { "epoch": 0.005334755934915978, "grad_norm": 13.559316635131836, "learning_rate": 1.6666666666666665e-07, "loss": 2.1607, "step": 25 }, { "epoch": 0.0064017071218991735, "grad_norm": 13.044219970703125, "learning_rate": 1.9333333333333332e-07, "loss": 2.1591, "step": 30 }, { "epoch": 0.007468658308882369, "grad_norm": 13.89301872253418, "learning_rate": 2.2666666666666663e-07, "loss": 2.0855, "step": 35 }, { "epoch": 0.008535609495865563, "grad_norm": 14.154369354248047, "learning_rate": 2.6e-07, "loss": 1.9557, "step": 40 }, { "epoch": 0.00960256068284876, "grad_norm": 9.126311302185059, "learning_rate": 2.933333333333333e-07, "loss": 1.8661, "step": 45 }, { "epoch": 0.010669511869831956, "grad_norm": 13.619270324707031, "learning_rate": 3.2666666666666663e-07, "loss": 1.8424, "step": 50 }, { "epoch": 0.01173646305681515, "grad_norm": 9.875323295593262, "learning_rate": 3.6e-07, "loss": 1.6208, "step": 55 }, { "epoch": 0.012803414243798347, "grad_norm": 11.143688201904297, "learning_rate": 3.933333333333333e-07, "loss": 1.5396, "step": 60 }, { "epoch": 0.013870365430781541, "grad_norm": 10.829161643981934, "learning_rate": 4.266666666666667e-07, "loss": 1.4174, "step": 65 }, { "epoch": 0.014937316617764738, "grad_norm": 7.393435001373291, "learning_rate": 4.6e-07, "loss": 1.1693, "step": 70 }, { "epoch": 0.016004267804747934, "grad_norm": 4.182456016540527, "learning_rate": 4.933333333333333e-07, "loss": 1.1779, "step": 75 }, { "epoch": 0.017071218991731127, "grad_norm": 8.4662504196167, "learning_rate": 5.266666666666666e-07, "loss": 1.1281, "step": 80 }, { "epoch": 0.018138170178714323, "grad_norm": 4.043624401092529, "learning_rate": 5.6e-07, "loss": 1.0614, "step": 85 }, { "epoch": 0.01920512136569752, "grad_norm": 3.8517227172851562, "learning_rate": 5.933333333333334e-07, "loss": 1.0066, "step": 90 }, { "epoch": 0.020272072552680716, "grad_norm": 4.19964075088501, "learning_rate": 6.266666666666667e-07, "loss": 1.0659, "step": 95 }, { "epoch": 0.021339023739663912, "grad_norm": 4.7087178230285645, "learning_rate": 6.6e-07, "loss": 0.9793, "step": 100 }, { "epoch": 0.022405974926647105, "grad_norm": 4.864803314208984, "learning_rate": 6.933333333333333e-07, "loss": 0.9675, "step": 105 }, { "epoch": 0.0234729261136303, "grad_norm": 3.8939764499664307, "learning_rate": 7.266666666666667e-07, "loss": 0.9513, "step": 110 }, { "epoch": 0.024539877300613498, "grad_norm": 3.557497978210449, "learning_rate": 7.599999999999999e-07, "loss": 0.9164, "step": 115 }, { "epoch": 0.025606828487596694, "grad_norm": 4.033596038818359, "learning_rate": 7.933333333333333e-07, "loss": 0.9069, "step": 120 }, { "epoch": 0.026673779674579887, "grad_norm": 4.726081371307373, "learning_rate": 8.266666666666667e-07, "loss": 0.8976, "step": 125 }, { "epoch": 0.027740730861563083, "grad_norm": 18.711572647094727, "learning_rate": 8.599999999999999e-07, "loss": 0.9157, "step": 130 }, { "epoch": 0.02880768204854628, "grad_norm": 4.13236141204834, "learning_rate": 8.933333333333333e-07, "loss": 0.8808, "step": 135 }, { "epoch": 0.029874633235529476, "grad_norm": 3.7335939407348633, "learning_rate": 9.266666666666665e-07, "loss": 0.7707, "step": 140 }, { "epoch": 0.03094158442251267, "grad_norm": 4.5814313888549805, "learning_rate": 9.6e-07, "loss": 0.8349, "step": 145 }, { "epoch": 0.03200853560949587, "grad_norm": 4.993039608001709, "learning_rate": 9.933333333333333e-07, "loss": 0.7002, "step": 150 }, { "epoch": 0.033075486796479064, "grad_norm": 5.326510906219482, "learning_rate": 9.996791443850267e-07, "loss": 0.6418, "step": 155 }, { "epoch": 0.034142437983462254, "grad_norm": 4.4001994132995605, "learning_rate": 9.99144385026738e-07, "loss": 0.5632, "step": 160 }, { "epoch": 0.03520938917044545, "grad_norm": 71.39772033691406, "learning_rate": 9.98716577540107e-07, "loss": 0.6374, "step": 165 }, { "epoch": 0.036276340357428646, "grad_norm": 9.561308860778809, "learning_rate": 9.98288770053476e-07, "loss": 0.5493, "step": 170 }, { "epoch": 0.03734329154441184, "grad_norm": 2.866039991378784, "learning_rate": 9.97754010695187e-07, "loss": 0.473, "step": 175 }, { "epoch": 0.03841024273139504, "grad_norm": 3.0788004398345947, "learning_rate": 9.972192513368983e-07, "loss": 0.4903, "step": 180 }, { "epoch": 0.039477193918378235, "grad_norm": 3.4059643745422363, "learning_rate": 9.966844919786097e-07, "loss": 0.5017, "step": 185 }, { "epoch": 0.04054414510536143, "grad_norm": 2.9747097492218018, "learning_rate": 9.961497326203208e-07, "loss": 0.4269, "step": 190 }, { "epoch": 0.04161109629234463, "grad_norm": 3.476583242416382, "learning_rate": 9.95614973262032e-07, "loss": 0.4214, "step": 195 }, { "epoch": 0.042678047479327824, "grad_norm": 3.0810964107513428, "learning_rate": 9.950802139037432e-07, "loss": 0.4291, "step": 200 }, { "epoch": 0.043744998666311014, "grad_norm": 87.03646850585938, "learning_rate": 9.945454545454544e-07, "loss": 0.4035, "step": 205 }, { "epoch": 0.04481194985329421, "grad_norm": 2.6879165172576904, "learning_rate": 9.940106951871658e-07, "loss": 0.428, "step": 210 }, { "epoch": 0.045878901040277406, "grad_norm": 3.602217197418213, "learning_rate": 9.93475935828877e-07, "loss": 0.4534, "step": 215 }, { "epoch": 0.0469458522272606, "grad_norm": 2.9568989276885986, "learning_rate": 9.929411764705881e-07, "loss": 0.42, "step": 220 }, { "epoch": 0.0480128034142438, "grad_norm": 2.7602076530456543, "learning_rate": 9.924064171122995e-07, "loss": 0.448, "step": 225 }, { "epoch": 0.049079754601226995, "grad_norm": 2.7191200256347656, "learning_rate": 9.918716577540107e-07, "loss": 0.4046, "step": 230 }, { "epoch": 0.05014670578821019, "grad_norm": 2.7233424186706543, "learning_rate": 9.913368983957219e-07, "loss": 0.4028, "step": 235 }, { "epoch": 0.05121365697519339, "grad_norm": 2.730400323867798, "learning_rate": 9.90802139037433e-07, "loss": 0.4361, "step": 240 }, { "epoch": 0.05228060816217658, "grad_norm": 4.667561054229736, "learning_rate": 9.902673796791442e-07, "loss": 0.4183, "step": 245 }, { "epoch": 0.05334755934915977, "grad_norm": 2.981497049331665, "learning_rate": 9.897326203208556e-07, "loss": 0.4148, "step": 250 }, { "epoch": 0.05441451053614297, "grad_norm": 9.336112976074219, "learning_rate": 9.891978609625668e-07, "loss": 0.4386, "step": 255 }, { "epoch": 0.055481461723126166, "grad_norm": 4.434041500091553, "learning_rate": 9.88663101604278e-07, "loss": 0.4235, "step": 260 }, { "epoch": 0.05654841291010936, "grad_norm": 2.695866346359253, "learning_rate": 9.881283422459892e-07, "loss": 0.4424, "step": 265 }, { "epoch": 0.05761536409709256, "grad_norm": 2.8247873783111572, "learning_rate": 9.875935828877004e-07, "loss": 0.387, "step": 270 }, { "epoch": 0.058682315284075755, "grad_norm": 3.4680376052856445, "learning_rate": 9.870588235294118e-07, "loss": 0.4294, "step": 275 }, { "epoch": 0.05974926647105895, "grad_norm": 2.6994473934173584, "learning_rate": 9.86524064171123e-07, "loss": 0.4043, "step": 280 }, { "epoch": 0.06081621765804215, "grad_norm": 2.6048827171325684, "learning_rate": 9.859893048128341e-07, "loss": 0.4064, "step": 285 }, { "epoch": 0.06188316884502534, "grad_norm": 3.028503179550171, "learning_rate": 9.854545454545455e-07, "loss": 0.4269, "step": 290 }, { "epoch": 0.06295012003200853, "grad_norm": 3.3406739234924316, "learning_rate": 9.849197860962567e-07, "loss": 0.4339, "step": 295 }, { "epoch": 0.06401707121899174, "grad_norm": 3.1800878047943115, "learning_rate": 9.843850267379679e-07, "loss": 0.3972, "step": 300 }, { "epoch": 0.06508402240597493, "grad_norm": 2.818852424621582, "learning_rate": 9.83850267379679e-07, "loss": 0.3898, "step": 305 }, { "epoch": 0.06615097359295813, "grad_norm": 3.256840705871582, "learning_rate": 9.833155080213902e-07, "loss": 0.4377, "step": 310 }, { "epoch": 0.06721792477994132, "grad_norm": 3.388472318649292, "learning_rate": 9.827807486631016e-07, "loss": 0.4187, "step": 315 }, { "epoch": 0.06828487596692451, "grad_norm": 4.434413909912109, "learning_rate": 9.822459893048128e-07, "loss": 0.4009, "step": 320 }, { "epoch": 0.06935182715390771, "grad_norm": 3.08677077293396, "learning_rate": 9.81711229946524e-07, "loss": 0.3837, "step": 325 }, { "epoch": 0.0704187783408909, "grad_norm": 3.7852623462677, "learning_rate": 9.811764705882352e-07, "loss": 0.3999, "step": 330 }, { "epoch": 0.0714857295278741, "grad_norm": 2.947720766067505, "learning_rate": 9.806417112299463e-07, "loss": 0.3677, "step": 335 }, { "epoch": 0.07255268071485729, "grad_norm": 3.1685659885406494, "learning_rate": 9.801069518716577e-07, "loss": 0.4035, "step": 340 }, { "epoch": 0.0736196319018405, "grad_norm": 2.661599636077881, "learning_rate": 9.79572192513369e-07, "loss": 0.3658, "step": 345 }, { "epoch": 0.07468658308882369, "grad_norm": 2.6443653106689453, "learning_rate": 9.7903743315508e-07, "loss": 0.3805, "step": 350 }, { "epoch": 0.07575353427580689, "grad_norm": 3.653778314590454, "learning_rate": 9.785026737967915e-07, "loss": 0.407, "step": 355 }, { "epoch": 0.07682048546279008, "grad_norm": 2.991504430770874, "learning_rate": 9.779679144385027e-07, "loss": 0.4135, "step": 360 }, { "epoch": 0.07788743664977327, "grad_norm": 2.7770955562591553, "learning_rate": 9.774331550802139e-07, "loss": 0.3934, "step": 365 }, { "epoch": 0.07895438783675647, "grad_norm": 2.686368703842163, "learning_rate": 9.76898395721925e-07, "loss": 0.409, "step": 370 }, { "epoch": 0.08002133902373966, "grad_norm": 3.307638168334961, "learning_rate": 9.763636363636362e-07, "loss": 0.3872, "step": 375 }, { "epoch": 0.08108829021072286, "grad_norm": 8.76164722442627, "learning_rate": 9.758288770053476e-07, "loss": 0.3882, "step": 380 }, { "epoch": 0.08215524139770605, "grad_norm": 2.868077039718628, "learning_rate": 9.752941176470588e-07, "loss": 0.4123, "step": 385 }, { "epoch": 0.08322219258468926, "grad_norm": 4.677827835083008, "learning_rate": 9.7475935828877e-07, "loss": 0.362, "step": 390 }, { "epoch": 0.08428914377167245, "grad_norm": 3.836914539337158, "learning_rate": 9.742245989304812e-07, "loss": 0.4023, "step": 395 }, { "epoch": 0.08535609495865565, "grad_norm": 2.7120203971862793, "learning_rate": 9.736898395721923e-07, "loss": 0.419, "step": 400 }, { "epoch": 0.08642304614563884, "grad_norm": 3.1341233253479004, "learning_rate": 9.731550802139037e-07, "loss": 0.4275, "step": 405 }, { "epoch": 0.08748999733262203, "grad_norm": 3.0901923179626465, "learning_rate": 9.72620320855615e-07, "loss": 0.3799, "step": 410 }, { "epoch": 0.08855694851960523, "grad_norm": 2.7124781608581543, "learning_rate": 9.72085561497326e-07, "loss": 0.3788, "step": 415 }, { "epoch": 0.08962389970658842, "grad_norm": 2.8177292346954346, "learning_rate": 9.715508021390375e-07, "loss": 0.3436, "step": 420 }, { "epoch": 0.09069085089357162, "grad_norm": 3.063669204711914, "learning_rate": 9.710160427807487e-07, "loss": 0.4312, "step": 425 }, { "epoch": 0.09175780208055481, "grad_norm": 2.78320574760437, "learning_rate": 9.704812834224598e-07, "loss": 0.4122, "step": 430 }, { "epoch": 0.09282475326753802, "grad_norm": 3.080400228500366, "learning_rate": 9.69946524064171e-07, "loss": 0.3917, "step": 435 }, { "epoch": 0.0938917044545212, "grad_norm": 3.02703857421875, "learning_rate": 9.694117647058822e-07, "loss": 0.3546, "step": 440 }, { "epoch": 0.09495865564150441, "grad_norm": 2.6559834480285645, "learning_rate": 9.688770053475936e-07, "loss": 0.3536, "step": 445 }, { "epoch": 0.0960256068284876, "grad_norm": 3.939589738845825, "learning_rate": 9.683422459893048e-07, "loss": 0.3729, "step": 450 }, { "epoch": 0.09709255801547079, "grad_norm": 2.419799327850342, "learning_rate": 9.67807486631016e-07, "loss": 0.3474, "step": 455 }, { "epoch": 0.09815950920245399, "grad_norm": 2.237165927886963, "learning_rate": 9.672727272727271e-07, "loss": 0.3387, "step": 460 }, { "epoch": 0.09922646038943718, "grad_norm": 2.6400506496429443, "learning_rate": 9.667379679144385e-07, "loss": 0.3462, "step": 465 }, { "epoch": 0.10029341157642038, "grad_norm": 3.294222354888916, "learning_rate": 9.662032085561497e-07, "loss": 0.3724, "step": 470 }, { "epoch": 0.10136036276340357, "grad_norm": 2.5825111865997314, "learning_rate": 9.656684491978609e-07, "loss": 0.3652, "step": 475 }, { "epoch": 0.10242731395038678, "grad_norm": 2.6807165145874023, "learning_rate": 9.65133689839572e-07, "loss": 0.3824, "step": 480 }, { "epoch": 0.10349426513736996, "grad_norm": 2.541398048400879, "learning_rate": 9.645989304812835e-07, "loss": 0.3509, "step": 485 }, { "epoch": 0.10456121632435315, "grad_norm": 3.071420192718506, "learning_rate": 9.640641711229946e-07, "loss": 0.3511, "step": 490 }, { "epoch": 0.10562816751133636, "grad_norm": 3.117623805999756, "learning_rate": 9.635294117647058e-07, "loss": 0.3823, "step": 495 }, { "epoch": 0.10669511869831955, "grad_norm": 2.9759390354156494, "learning_rate": 9.62994652406417e-07, "loss": 0.3595, "step": 500 }, { "epoch": 0.10669511869831955, "eval_bleu": 0.5858525890811711, "eval_cer": 5.81035140115651, "eval_loss": 0.35532379150390625, "eval_runtime": 1038.2775, "eval_samples_per_second": 0.884, "eval_steps_per_second": 0.222, "eval_wer": 22.283451007404857, "step": 500 }, { "epoch": 0.10776206988530275, "grad_norm": 2.9231984615325928, "learning_rate": 9.475679090334807e-07, "loss": 0.3783, "step": 505 }, { "epoch": 0.10882902107228594, "grad_norm": 2.2739880084991455, "learning_rate": 9.470414824173509e-07, "loss": 0.3238, "step": 510 }, { "epoch": 0.10989597225926914, "grad_norm": 2.603306531906128, "learning_rate": 9.465150558012212e-07, "loss": 0.3538, "step": 515 }, { "epoch": 0.11096292344625233, "grad_norm": 2.871011972427368, "learning_rate": 9.459886291850916e-07, "loss": 0.3955, "step": 520 }, { "epoch": 0.11202987463323553, "grad_norm": 3.134646415710449, "learning_rate": 9.454622025689618e-07, "loss": 0.3508, "step": 525 }, { "epoch": 0.11309682582021872, "grad_norm": 3.014796733856201, "learning_rate": 9.449357759528321e-07, "loss": 0.3664, "step": 530 }, { "epoch": 0.11416377700720191, "grad_norm": 3.0675926208496094, "learning_rate": 9.444093493367024e-07, "loss": 0.3497, "step": 535 }, { "epoch": 0.11523072819418512, "grad_norm": 4.508389472961426, "learning_rate": 9.438829227205727e-07, "loss": 0.3172, "step": 540 }, { "epoch": 0.1162976793811683, "grad_norm": 2.528317451477051, "learning_rate": 9.43356496104443e-07, "loss": 0.3495, "step": 545 }, { "epoch": 0.11736463056815151, "grad_norm": 2.759575843811035, "learning_rate": 9.428300694883133e-07, "loss": 0.3326, "step": 550 }, { "epoch": 0.1184315817551347, "grad_norm": 6.699812889099121, "learning_rate": 9.423036428721835e-07, "loss": 0.4124, "step": 555 }, { "epoch": 0.1194985329421179, "grad_norm": 2.358922243118286, "learning_rate": 9.417772162560539e-07, "loss": 0.3292, "step": 560 }, { "epoch": 0.12056548412910109, "grad_norm": 2.3635692596435547, "learning_rate": 9.412507896399241e-07, "loss": 0.3715, "step": 565 }, { "epoch": 0.1216324353160843, "grad_norm": 2.8696706295013428, "learning_rate": 9.407243630237944e-07, "loss": 0.3348, "step": 570 }, { "epoch": 0.12269938650306748, "grad_norm": 3.717510223388672, "learning_rate": 9.401979364076647e-07, "loss": 0.355, "step": 575 }, { "epoch": 0.12376633769005067, "grad_norm": 3.9285202026367188, "learning_rate": 9.39671509791535e-07, "loss": 0.3752, "step": 580 }, { "epoch": 0.12483328887703388, "grad_norm": 3.560582399368286, "learning_rate": 9.391450831754053e-07, "loss": 0.3422, "step": 585 }, { "epoch": 0.12590024006401707, "grad_norm": 6.333406925201416, "learning_rate": 9.386186565592757e-07, "loss": 0.3436, "step": 590 }, { "epoch": 0.12696719125100026, "grad_norm": 2.94331431388855, "learning_rate": 9.380922299431458e-07, "loss": 0.3346, "step": 595 }, { "epoch": 0.12803414243798347, "grad_norm": 2.933142900466919, "learning_rate": 9.375658033270162e-07, "loss": 0.3414, "step": 600 }, { "epoch": 0.12910109362496666, "grad_norm": 3.2017970085144043, "learning_rate": 9.370393767108865e-07, "loss": 0.3731, "step": 605 }, { "epoch": 0.13016804481194985, "grad_norm": 3.2127702236175537, "learning_rate": 9.365129500947567e-07, "loss": 0.3399, "step": 610 }, { "epoch": 0.13123499599893304, "grad_norm": 3.1433818340301514, "learning_rate": 9.35986523478627e-07, "loss": 0.342, "step": 615 }, { "epoch": 0.13230194718591626, "grad_norm": 3.0913007259368896, "learning_rate": 9.354600968624973e-07, "loss": 0.3543, "step": 620 }, { "epoch": 0.13336889837289945, "grad_norm": 3.459428310394287, "learning_rate": 9.349336702463676e-07, "loss": 0.292, "step": 625 }, { "epoch": 0.13443584955988264, "grad_norm": 2.965162754058838, "learning_rate": 9.34407243630238e-07, "loss": 0.3591, "step": 630 }, { "epoch": 0.13550280074686583, "grad_norm": 2.476099967956543, "learning_rate": 9.338808170141081e-07, "loss": 0.3351, "step": 635 }, { "epoch": 0.13656975193384902, "grad_norm": 2.853848695755005, "learning_rate": 9.333543903979785e-07, "loss": 0.3828, "step": 640 }, { "epoch": 0.13763670312083223, "grad_norm": 2.560877561569214, "learning_rate": 9.328279637818488e-07, "loss": 0.3183, "step": 645 }, { "epoch": 0.13870365430781542, "grad_norm": 2.7191262245178223, "learning_rate": 9.32301537165719e-07, "loss": 0.3572, "step": 650 }, { "epoch": 0.1397706054947986, "grad_norm": 3.388456106185913, "learning_rate": 9.317751105495893e-07, "loss": 0.3207, "step": 655 }, { "epoch": 0.1408375566817818, "grad_norm": 2.827470302581787, "learning_rate": 9.312486839334597e-07, "loss": 0.3576, "step": 660 }, { "epoch": 0.14190450786876502, "grad_norm": 2.630094528198242, "learning_rate": 9.307222573173299e-07, "loss": 0.3616, "step": 665 }, { "epoch": 0.1429714590557482, "grad_norm": 2.7217891216278076, "learning_rate": 9.301958307012003e-07, "loss": 0.3434, "step": 670 }, { "epoch": 0.1440384102427314, "grad_norm": 2.558335065841675, "learning_rate": 9.296694040850705e-07, "loss": 0.3516, "step": 675 }, { "epoch": 0.14510536142971459, "grad_norm": 3.1991679668426514, "learning_rate": 9.291429774689408e-07, "loss": 0.338, "step": 680 }, { "epoch": 0.14617231261669777, "grad_norm": 4.686666011810303, "learning_rate": 9.286165508528111e-07, "loss": 0.3539, "step": 685 }, { "epoch": 0.147239263803681, "grad_norm": 3.0826447010040283, "learning_rate": 9.280901242366814e-07, "loss": 0.3313, "step": 690 }, { "epoch": 0.14830621499066418, "grad_norm": 2.656141757965088, "learning_rate": 9.275636976205516e-07, "loss": 0.3454, "step": 695 }, { "epoch": 0.14937316617764737, "grad_norm": 4.859818935394287, "learning_rate": 9.27037271004422e-07, "loss": 0.3331, "step": 700 }, { "epoch": 0.15044011736463056, "grad_norm": 5.376903057098389, "learning_rate": 9.265108443882922e-07, "loss": 0.3377, "step": 705 }, { "epoch": 0.15150706855161378, "grad_norm": 2.7976577281951904, "learning_rate": 9.259844177721626e-07, "loss": 0.3768, "step": 710 }, { "epoch": 0.15257401973859697, "grad_norm": 2.993427038192749, "learning_rate": 9.254579911560328e-07, "loss": 0.3509, "step": 715 }, { "epoch": 0.15364097092558016, "grad_norm": 2.8581597805023193, "learning_rate": 9.249315645399031e-07, "loss": 0.3338, "step": 720 }, { "epoch": 0.15470792211256335, "grad_norm": 2.8063721656799316, "learning_rate": 9.244051379237734e-07, "loss": 0.3597, "step": 725 }, { "epoch": 0.15577487329954653, "grad_norm": 2.7333686351776123, "learning_rate": 9.238787113076438e-07, "loss": 0.3563, "step": 730 }, { "epoch": 0.15684182448652975, "grad_norm": 3.4143598079681396, "learning_rate": 9.233522846915139e-07, "loss": 0.336, "step": 735 }, { "epoch": 0.15790877567351294, "grad_norm": 2.84298038482666, "learning_rate": 9.228258580753843e-07, "loss": 0.3651, "step": 740 }, { "epoch": 0.15897572686049613, "grad_norm": 5.912104606628418, "learning_rate": 9.222994314592545e-07, "loss": 0.3531, "step": 745 }, { "epoch": 0.16004267804747932, "grad_norm": 3.0039525032043457, "learning_rate": 9.217730048431249e-07, "loss": 0.3212, "step": 750 }, { "epoch": 0.16110962923446254, "grad_norm": 3.2109341621398926, "learning_rate": 9.212465782269951e-07, "loss": 0.3206, "step": 755 }, { "epoch": 0.16217658042144573, "grad_norm": 2.7238008975982666, "learning_rate": 9.207201516108654e-07, "loss": 0.3696, "step": 760 }, { "epoch": 0.16324353160842892, "grad_norm": 2.6420886516571045, "learning_rate": 9.201937249947357e-07, "loss": 0.3339, "step": 765 }, { "epoch": 0.1643104827954121, "grad_norm": 2.8822038173675537, "learning_rate": 9.196672983786061e-07, "loss": 0.3611, "step": 770 }, { "epoch": 0.1653774339823953, "grad_norm": 2.558979034423828, "learning_rate": 9.191408717624762e-07, "loss": 0.3247, "step": 775 }, { "epoch": 0.1664443851693785, "grad_norm": 2.649867057800293, "learning_rate": 9.186144451463465e-07, "loss": 0.3146, "step": 780 }, { "epoch": 0.1675113363563617, "grad_norm": 2.6174542903900146, "learning_rate": 9.180880185302169e-07, "loss": 0.3319, "step": 785 }, { "epoch": 0.1685782875433449, "grad_norm": 2.65977144241333, "learning_rate": 9.175615919140871e-07, "loss": 0.3643, "step": 790 }, { "epoch": 0.16964523873032808, "grad_norm": 3.2722222805023193, "learning_rate": 9.170351652979574e-07, "loss": 0.3349, "step": 795 }, { "epoch": 0.1707121899173113, "grad_norm": 2.346200466156006, "learning_rate": 9.165087386818277e-07, "loss": 0.3028, "step": 800 }, { "epoch": 0.17177914110429449, "grad_norm": 2.675050735473633, "learning_rate": 9.15982312065698e-07, "loss": 0.3256, "step": 805 }, { "epoch": 0.17284609229127768, "grad_norm": 2.4576005935668945, "learning_rate": 9.154558854495683e-07, "loss": 0.3127, "step": 810 }, { "epoch": 0.17391304347826086, "grad_norm": 2.73311710357666, "learning_rate": 9.150347441566645e-07, "loss": 0.3508, "step": 815 }, { "epoch": 0.17497999466524405, "grad_norm": 2.6573374271392822, "learning_rate": 9.145083175405348e-07, "loss": 0.3643, "step": 820 }, { "epoch": 0.17604694585222727, "grad_norm": 2.7585701942443848, "learning_rate": 9.139818909244052e-07, "loss": 0.3321, "step": 825 }, { "epoch": 0.17711389703921046, "grad_norm": 2.8242616653442383, "learning_rate": 9.134554643082753e-07, "loss": 0.3175, "step": 830 }, { "epoch": 0.17818084822619365, "grad_norm": 3.1244609355926514, "learning_rate": 9.129290376921457e-07, "loss": 0.3641, "step": 835 }, { "epoch": 0.17924779941317684, "grad_norm": 2.4807212352752686, "learning_rate": 9.12402611076016e-07, "loss": 0.3464, "step": 840 }, { "epoch": 0.18031475060016006, "grad_norm": 4.587203502655029, "learning_rate": 9.118761844598863e-07, "loss": 0.3199, "step": 845 }, { "epoch": 0.18138170178714325, "grad_norm": 3.1150426864624023, "learning_rate": 9.113497578437566e-07, "loss": 0.3129, "step": 850 }, { "epoch": 0.18244865297412644, "grad_norm": 2.80098557472229, "learning_rate": 9.108233312276268e-07, "loss": 0.3296, "step": 855 }, { "epoch": 0.18351560416110962, "grad_norm": 2.6048800945281982, "learning_rate": 9.102969046114971e-07, "loss": 0.3426, "step": 860 }, { "epoch": 0.1845825553480928, "grad_norm": 2.4776954650878906, "learning_rate": 9.097704779953675e-07, "loss": 0.3389, "step": 865 }, { "epoch": 0.18564950653507603, "grad_norm": 3.662856101989746, "learning_rate": 9.092440513792377e-07, "loss": 0.3191, "step": 870 }, { "epoch": 0.18671645772205922, "grad_norm": 5.855990886688232, "learning_rate": 9.08717624763108e-07, "loss": 0.3491, "step": 875 }, { "epoch": 0.1877834089090424, "grad_norm": 2.6081950664520264, "learning_rate": 9.081911981469783e-07, "loss": 0.3256, "step": 880 }, { "epoch": 0.1888503600960256, "grad_norm": 3.206470251083374, "learning_rate": 9.076647715308485e-07, "loss": 0.3419, "step": 885 }, { "epoch": 0.18991731128300882, "grad_norm": 3.7217066287994385, "learning_rate": 9.071383449147189e-07, "loss": 0.3369, "step": 890 }, { "epoch": 0.190984262469992, "grad_norm": 4.7247633934021, "learning_rate": 9.066119182985891e-07, "loss": 0.3664, "step": 895 }, { "epoch": 0.1920512136569752, "grad_norm": 3.455446481704712, "learning_rate": 9.060854916824594e-07, "loss": 0.3171, "step": 900 }, { "epoch": 0.19311816484395838, "grad_norm": 2.6066224575042725, "learning_rate": 9.055590650663297e-07, "loss": 0.3266, "step": 905 }, { "epoch": 0.19418511603094157, "grad_norm": 3.2545228004455566, "learning_rate": 9.050326384502e-07, "loss": 0.3201, "step": 910 }, { "epoch": 0.1952520672179248, "grad_norm": 3.8174829483032227, "learning_rate": 9.045062118340702e-07, "loss": 0.3537, "step": 915 }, { "epoch": 0.19631901840490798, "grad_norm": 3.725991725921631, "learning_rate": 9.039797852179406e-07, "loss": 0.3373, "step": 920 }, { "epoch": 0.19738596959189117, "grad_norm": 2.5402047634124756, "learning_rate": 9.034533586018108e-07, "loss": 0.3107, "step": 925 }, { "epoch": 0.19845292077887436, "grad_norm": 2.3887853622436523, "learning_rate": 9.029269319856812e-07, "loss": 0.3255, "step": 930 }, { "epoch": 0.19951987196585758, "grad_norm": 2.83178973197937, "learning_rate": 9.024005053695514e-07, "loss": 0.3192, "step": 935 }, { "epoch": 0.20058682315284077, "grad_norm": 2.540933132171631, "learning_rate": 9.018740787534217e-07, "loss": 0.3414, "step": 940 }, { "epoch": 0.20165377433982395, "grad_norm": 2.799577474594116, "learning_rate": 9.01347652137292e-07, "loss": 0.3383, "step": 945 }, { "epoch": 0.20272072552680714, "grad_norm": 2.5945661067962646, "learning_rate": 9.008212255211624e-07, "loss": 0.3414, "step": 950 }, { "epoch": 0.20378767671379033, "grad_norm": 3.102372407913208, "learning_rate": 9.002947989050325e-07, "loss": 0.33, "step": 955 }, { "epoch": 0.20485462790077355, "grad_norm": 3.1992921829223633, "learning_rate": 8.997683722889029e-07, "loss": 0.3719, "step": 960 }, { "epoch": 0.20592157908775674, "grad_norm": 3.082578420639038, "learning_rate": 8.992419456727732e-07, "loss": 0.3412, "step": 965 }, { "epoch": 0.20698853027473993, "grad_norm": 2.406475067138672, "learning_rate": 8.987155190566435e-07, "loss": 0.3069, "step": 970 }, { "epoch": 0.20805548146172312, "grad_norm": 3.033055067062378, "learning_rate": 8.981890924405137e-07, "loss": 0.3683, "step": 975 }, { "epoch": 0.2091224326487063, "grad_norm": 2.807874917984009, "learning_rate": 8.97662665824384e-07, "loss": 0.3024, "step": 980 }, { "epoch": 0.21018938383568952, "grad_norm": 2.6849870681762695, "learning_rate": 8.971362392082543e-07, "loss": 0.3248, "step": 985 }, { "epoch": 0.21125633502267271, "grad_norm": 3.144150495529175, "learning_rate": 8.966098125921247e-07, "loss": 0.3158, "step": 990 }, { "epoch": 0.2123232862096559, "grad_norm": 2.495368003845215, "learning_rate": 8.960833859759948e-07, "loss": 0.312, "step": 995 }, { "epoch": 0.2133902373966391, "grad_norm": 2.2597124576568604, "learning_rate": 8.955569593598652e-07, "loss": 0.3053, "step": 1000 }, { "epoch": 0.2133902373966391, "eval_bleu": 0.6173395426659279, "eval_cer": 5.420346953040605, "eval_loss": 0.3327235281467438, "eval_runtime": 1022.0953, "eval_samples_per_second": 0.898, "eval_steps_per_second": 0.225, "eval_wer": 20.46380804775845, "step": 1000 } ], "logging_steps": 5, "max_steps": 9500, "num_input_tokens_seen": 0, "num_train_epochs": 3, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 1.358829453312e+20, "train_batch_size": 8, "trial_name": null, "trial_params": null }