| { |
| "best_global_step": null, |
| "best_metric": null, |
| "best_model_checkpoint": null, |
| "epoch": 1.0, |
| "eval_steps": 578, |
| "global_step": 2309, |
| "is_hyper_param_search": false, |
| "is_local_process_zero": true, |
| "is_world_process_zero": true, |
| "log_history": [ |
| { |
| "epoch": 0.0043308791684712, |
| "grad_norm": 0.9375, |
| "learning_rate": 4.5e-06, |
| "loss": 0.44086565971374514, |
| "step": 10 |
| }, |
| { |
| "epoch": 0.0086617583369424, |
| "grad_norm": 1.7265625, |
| "learning_rate": 9.5e-06, |
| "loss": 0.38978335857391355, |
| "step": 20 |
| }, |
| { |
| "epoch": 0.0129926375054136, |
| "grad_norm": 1.1171875, |
| "learning_rate": 1.45e-05, |
| "loss": 0.3998324632644653, |
| "step": 30 |
| }, |
| { |
| "epoch": 0.0173235166738848, |
| "grad_norm": 0.8359375, |
| "learning_rate": 1.9500000000000003e-05, |
| "loss": 0.36765055656433104, |
| "step": 40 |
| }, |
| { |
| "epoch": 0.021654395842355997, |
| "grad_norm": 0.9140625, |
| "learning_rate": 2.45e-05, |
| "loss": 0.38308374881744384, |
| "step": 50 |
| }, |
| { |
| "epoch": 0.0259852750108272, |
| "grad_norm": 1.3125, |
| "learning_rate": 2.95e-05, |
| "loss": 0.37063548564910886, |
| "step": 60 |
| }, |
| { |
| "epoch": 0.030316154179298397, |
| "grad_norm": 0.921875, |
| "learning_rate": 3.45e-05, |
| "loss": 0.3595526456832886, |
| "step": 70 |
| }, |
| { |
| "epoch": 0.0346470333477696, |
| "grad_norm": 1.25, |
| "learning_rate": 3.9500000000000005e-05, |
| "loss": 0.3661433935165405, |
| "step": 80 |
| }, |
| { |
| "epoch": 0.0389779125162408, |
| "grad_norm": 0.8671875, |
| "learning_rate": 4.4500000000000004e-05, |
| "loss": 0.37132949829101564, |
| "step": 90 |
| }, |
| { |
| "epoch": 0.043308791684711995, |
| "grad_norm": 1.1015625, |
| "learning_rate": 4.9500000000000004e-05, |
| "loss": 0.3292886257171631, |
| "step": 100 |
| }, |
| { |
| "epoch": 0.04763967085318319, |
| "grad_norm": 1.03125, |
| "learning_rate": 4.999795215342401e-05, |
| "loss": 0.3398285865783691, |
| "step": 110 |
| }, |
| { |
| "epoch": 0.0519705500216544, |
| "grad_norm": 0.94921875, |
| "learning_rate": 4.999087360832838e-05, |
| "loss": 0.3274256229400635, |
| "step": 120 |
| }, |
| { |
| "epoch": 0.056301429190125596, |
| "grad_norm": 1.0078125, |
| "learning_rate": 4.9978740513992555e-05, |
| "loss": 0.32653069496154785, |
| "step": 130 |
| }, |
| { |
| "epoch": 0.060632308358596794, |
| "grad_norm": 0.92578125, |
| "learning_rate": 4.996155532440529e-05, |
| "loss": 0.34501988887786866, |
| "step": 140 |
| }, |
| { |
| "epoch": 0.06496318752706799, |
| "grad_norm": 0.58984375, |
| "learning_rate": 4.9939321515370904e-05, |
| "loss": 0.33666629791259767, |
| "step": 150 |
| }, |
| { |
| "epoch": 0.0692940666955392, |
| "grad_norm": 0.875, |
| "learning_rate": 4.991204358380634e-05, |
| "loss": 0.34445130825042725, |
| "step": 160 |
| }, |
| { |
| "epoch": 0.07362494586401039, |
| "grad_norm": 0.7109375, |
| "learning_rate": 4.9879727046831594e-05, |
| "loss": 0.31984710693359375, |
| "step": 170 |
| }, |
| { |
| "epoch": 0.0779558250324816, |
| "grad_norm": 0.65625, |
| "learning_rate": 4.984237844065383e-05, |
| "loss": 0.2977360486984253, |
| "step": 180 |
| }, |
| { |
| "epoch": 0.0822867042009528, |
| "grad_norm": 1.0234375, |
| "learning_rate": 4.980000531924542e-05, |
| "loss": 0.33499305248260497, |
| "step": 190 |
| }, |
| { |
| "epoch": 0.08661758336942399, |
| "grad_norm": 1.1953125, |
| "learning_rate": 4.975261625281614e-05, |
| "loss": 0.3221716403961182, |
| "step": 200 |
| }, |
| { |
| "epoch": 0.0909484625378952, |
| "grad_norm": 0.8828125, |
| "learning_rate": 4.97002208260797e-05, |
| "loss": 0.3402083396911621, |
| "step": 210 |
| }, |
| { |
| "epoch": 0.09527934170636639, |
| "grad_norm": 0.875, |
| "learning_rate": 4.964282963631529e-05, |
| "loss": 0.32864036560058596, |
| "step": 220 |
| }, |
| { |
| "epoch": 0.09961022087483759, |
| "grad_norm": 0.8046875, |
| "learning_rate": 4.9580454291224156e-05, |
| "loss": 0.319922137260437, |
| "step": 230 |
| }, |
| { |
| "epoch": 0.1039411000433088, |
| "grad_norm": 0.88671875, |
| "learning_rate": 4.951310740658187e-05, |
| "loss": 0.34775371551513673, |
| "step": 240 |
| }, |
| { |
| "epoch": 0.10827197921177999, |
| "grad_norm": 0.703125, |
| "learning_rate": 4.944080260368675e-05, |
| "loss": 0.3269499778747559, |
| "step": 250 |
| }, |
| { |
| "epoch": 0.11260285838025119, |
| "grad_norm": 1.0078125, |
| "learning_rate": 4.936355450660487e-05, |
| "loss": 0.3431516170501709, |
| "step": 260 |
| }, |
| { |
| "epoch": 0.1169337375487224, |
| "grad_norm": 0.86328125, |
| "learning_rate": 4.9281378739212225e-05, |
| "loss": 0.33036587238311765, |
| "step": 270 |
| }, |
| { |
| "epoch": 0.12126461671719359, |
| "grad_norm": 0.87109375, |
| "learning_rate": 4.919429192203473e-05, |
| "loss": 0.31238555908203125, |
| "step": 280 |
| }, |
| { |
| "epoch": 0.1255954958856648, |
| "grad_norm": 0.828125, |
| "learning_rate": 4.9102311668886634e-05, |
| "loss": 0.34010791778564453, |
| "step": 290 |
| }, |
| { |
| "epoch": 0.12992637505413598, |
| "grad_norm": 1.0390625, |
| "learning_rate": 4.9005456583308016e-05, |
| "loss": 0.37534093856811523, |
| "step": 300 |
| }, |
| { |
| "epoch": 0.1342572542226072, |
| "grad_norm": 0.90625, |
| "learning_rate": 4.8903746254802096e-05, |
| "loss": 0.33760197162628175, |
| "step": 310 |
| }, |
| { |
| "epoch": 0.1385881333910784, |
| "grad_norm": 1.1328125, |
| "learning_rate": 4.879720125487317e-05, |
| "loss": 0.3542941093444824, |
| "step": 320 |
| }, |
| { |
| "epoch": 0.14291901255954959, |
| "grad_norm": 0.875, |
| "learning_rate": 4.868584313286589e-05, |
| "loss": 0.3184266805648804, |
| "step": 330 |
| }, |
| { |
| "epoch": 0.14724989172802078, |
| "grad_norm": 1.0625, |
| "learning_rate": 4.8569694411606784e-05, |
| "loss": 0.3295663595199585, |
| "step": 340 |
| }, |
| { |
| "epoch": 0.151580770896492, |
| "grad_norm": 0.9609375, |
| "learning_rate": 4.844877858284886e-05, |
| "loss": 0.3455744504928589, |
| "step": 350 |
| }, |
| { |
| "epoch": 0.1559116500649632, |
| "grad_norm": 0.8203125, |
| "learning_rate": 4.8323120102520334e-05, |
| "loss": 0.32991776466369627, |
| "step": 360 |
| }, |
| { |
| "epoch": 0.16024252923343438, |
| "grad_norm": 0.6953125, |
| "learning_rate": 4.8192744385778185e-05, |
| "loss": 0.33624818325042727, |
| "step": 370 |
| }, |
| { |
| "epoch": 0.1645734084019056, |
| "grad_norm": 0.75390625, |
| "learning_rate": 4.805767780186786e-05, |
| "loss": 0.29813156127929685, |
| "step": 380 |
| }, |
| { |
| "epoch": 0.1689042875703768, |
| "grad_norm": 39.5, |
| "learning_rate": 4.7917947668789926e-05, |
| "loss": 0.3126053810119629, |
| "step": 390 |
| }, |
| { |
| "epoch": 0.17323516673884798, |
| "grad_norm": 0.94921875, |
| "learning_rate": 4.7773582247774806e-05, |
| "loss": 0.30437936782836916, |
| "step": 400 |
| }, |
| { |
| "epoch": 0.1775660459073192, |
| "grad_norm": 0.609375, |
| "learning_rate": 4.7624610737566846e-05, |
| "loss": 0.313838791847229, |
| "step": 410 |
| }, |
| { |
| "epoch": 0.1818969250757904, |
| "grad_norm": 1.015625, |
| "learning_rate": 4.747106326851864e-05, |
| "loss": 0.3429024457931519, |
| "step": 420 |
| }, |
| { |
| "epoch": 0.18622780424426158, |
| "grad_norm": 0.76953125, |
| "learning_rate": 4.731297089649703e-05, |
| "loss": 0.3326719284057617, |
| "step": 430 |
| }, |
| { |
| "epoch": 0.19055868341273277, |
| "grad_norm": 0.65234375, |
| "learning_rate": 4.7150365596601876e-05, |
| "loss": 0.3007481336593628, |
| "step": 440 |
| }, |
| { |
| "epoch": 0.194889562581204, |
| "grad_norm": 0.86328125, |
| "learning_rate": 4.6983280256698864e-05, |
| "loss": 0.3022452354431152, |
| "step": 450 |
| }, |
| { |
| "epoch": 0.19922044174967518, |
| "grad_norm": 0.8046875, |
| "learning_rate": 4.68117486707678e-05, |
| "loss": 0.31449906826019286, |
| "step": 460 |
| }, |
| { |
| "epoch": 0.20355132091814637, |
| "grad_norm": 1.015625, |
| "learning_rate": 4.663580553206755e-05, |
| "loss": 0.316056227684021, |
| "step": 470 |
| }, |
| { |
| "epoch": 0.2078822000866176, |
| "grad_norm": 0.8671875, |
| "learning_rate": 4.645548642611913e-05, |
| "loss": 0.31620476245880125, |
| "step": 480 |
| }, |
| { |
| "epoch": 0.21221307925508878, |
| "grad_norm": 0.765625, |
| "learning_rate": 4.627082782350833e-05, |
| "loss": 0.3331968069076538, |
| "step": 490 |
| }, |
| { |
| "epoch": 0.21654395842355997, |
| "grad_norm": 0.67578125, |
| "learning_rate": 4.6081867072509334e-05, |
| "loss": 0.32191483974456786, |
| "step": 500 |
| }, |
| { |
| "epoch": 0.2208748375920312, |
| "grad_norm": 0.91015625, |
| "learning_rate": 4.58886423915308e-05, |
| "loss": 0.3320103406906128, |
| "step": 510 |
| }, |
| { |
| "epoch": 0.22520571676050238, |
| "grad_norm": 0.91015625, |
| "learning_rate": 4.569119286138598e-05, |
| "loss": 0.3349184036254883, |
| "step": 520 |
| }, |
| { |
| "epoch": 0.22953659592897357, |
| "grad_norm": 1.0703125, |
| "learning_rate": 4.548955841738839e-05, |
| "loss": 0.3133854389190674, |
| "step": 530 |
| }, |
| { |
| "epoch": 0.2338674750974448, |
| "grad_norm": 1.1328125, |
| "learning_rate": 4.528377984127466e-05, |
| "loss": 0.31101830005645753, |
| "step": 540 |
| }, |
| { |
| "epoch": 0.23819835426591598, |
| "grad_norm": 0.87890625, |
| "learning_rate": 4.50738987529562e-05, |
| "loss": 0.32517337799072266, |
| "step": 550 |
| }, |
| { |
| "epoch": 0.24252923343438718, |
| "grad_norm": 0.6328125, |
| "learning_rate": 4.485995760210132e-05, |
| "loss": 0.31145153045654295, |
| "step": 560 |
| }, |
| { |
| "epoch": 0.24686011260285837, |
| "grad_norm": 0.9609375, |
| "learning_rate": 4.464199965954954e-05, |
| "loss": 0.33439595699310304, |
| "step": 570 |
| }, |
| { |
| "epoch": 0.25032481593763534, |
| "eval_text_loss": 1.6730986833572388, |
| "eval_text_model_preparation_time": 0.0199, |
| "eval_text_runtime": 23.6907, |
| "eval_text_samples_per_second": 13.93, |
| "eval_text_steps_per_second": 13.93, |
| "step": 578 |
| }, |
| { |
| "epoch": 0.25032481593763534, |
| "eval_audio_loss": 0.8745406866073608, |
| "eval_audio_model_preparation_time": 0.0199, |
| "eval_audio_runtime": 45.6232, |
| "eval_audio_samples_per_second": 5.611, |
| "eval_audio_steps_per_second": 5.611, |
| "step": 578 |
| }, |
| { |
| "epoch": 0.2511909917713296, |
| "grad_norm": 0.6015625, |
| "learning_rate": 4.442006900855983e-05, |
| "loss": 0.3008127689361572, |
| "step": 580 |
| }, |
| { |
| "epoch": 0.2555218709398008, |
| "grad_norm": 0.984375, |
| "learning_rate": 4.4194210535894475e-05, |
| "loss": 0.3019113063812256, |
| "step": 590 |
| }, |
| { |
| "epoch": 0.25985275010827197, |
| "grad_norm": 0.890625, |
| "learning_rate": 4.3964469922740526e-05, |
| "loss": 0.33276543617248533, |
| "step": 600 |
| }, |
| { |
| "epoch": 0.26418362927674316, |
| "grad_norm": 0.83984375, |
| "learning_rate": 4.3730893635470456e-05, |
| "loss": 0.32658073902130125, |
| "step": 610 |
| }, |
| { |
| "epoch": 0.2685145084452144, |
| "grad_norm": 0.71875, |
| "learning_rate": 4.3493528916244094e-05, |
| "loss": 0.2932913303375244, |
| "step": 620 |
| }, |
| { |
| "epoch": 0.2728453876136856, |
| "grad_norm": 1.0390625, |
| "learning_rate": 4.3252423773453623e-05, |
| "loss": 0.28568110466003416, |
| "step": 630 |
| }, |
| { |
| "epoch": 0.2771762667821568, |
| "grad_norm": 0.765625, |
| "learning_rate": 4.3007626972013596e-05, |
| "loss": 0.31102354526519777, |
| "step": 640 |
| }, |
| { |
| "epoch": 0.281507145950628, |
| "grad_norm": 0.82421875, |
| "learning_rate": 4.2759188023497984e-05, |
| "loss": 0.3104224681854248, |
| "step": 650 |
| }, |
| { |
| "epoch": 0.28583802511909917, |
| "grad_norm": 0.8671875, |
| "learning_rate": 4.250715717612611e-05, |
| "loss": 0.30080304145812986, |
| "step": 660 |
| }, |
| { |
| "epoch": 0.29016890428757036, |
| "grad_norm": 0.78515625, |
| "learning_rate": 4.22515854045997e-05, |
| "loss": 0.319173264503479, |
| "step": 670 |
| }, |
| { |
| "epoch": 0.29449978345604155, |
| "grad_norm": 0.96875, |
| "learning_rate": 4.1992524399792945e-05, |
| "loss": 0.33753151893615724, |
| "step": 680 |
| }, |
| { |
| "epoch": 0.2988306626245128, |
| "grad_norm": 0.75, |
| "learning_rate": 4.173002655829771e-05, |
| "loss": 0.30184378623962405, |
| "step": 690 |
| }, |
| { |
| "epoch": 0.303161541792984, |
| "grad_norm": 0.921875, |
| "learning_rate": 4.1464144971826056e-05, |
| "loss": 0.2979575157165527, |
| "step": 700 |
| }, |
| { |
| "epoch": 0.3074924209614552, |
| "grad_norm": 0.625, |
| "learning_rate": 4.119493341647208e-05, |
| "loss": 0.3160278558731079, |
| "step": 710 |
| }, |
| { |
| "epoch": 0.3118233001299264, |
| "grad_norm": 0.92578125, |
| "learning_rate": 4.0922446341835405e-05, |
| "loss": 0.3350363254547119, |
| "step": 720 |
| }, |
| { |
| "epoch": 0.31615417929839756, |
| "grad_norm": 0.734375, |
| "learning_rate": 4.064673886000847e-05, |
| "loss": 0.2890045642852783, |
| "step": 730 |
| }, |
| { |
| "epoch": 0.32048505846686876, |
| "grad_norm": 0.83203125, |
| "learning_rate": 4.036786673442974e-05, |
| "loss": 0.3226451396942139, |
| "step": 740 |
| }, |
| { |
| "epoch": 0.32481593763533995, |
| "grad_norm": 0.67578125, |
| "learning_rate": 4.0085886368605256e-05, |
| "loss": 0.3058380842208862, |
| "step": 750 |
| }, |
| { |
| "epoch": 0.3291468168038112, |
| "grad_norm": 1.109375, |
| "learning_rate": 3.9800854794700685e-05, |
| "loss": 0.3117943286895752, |
| "step": 760 |
| }, |
| { |
| "epoch": 0.3334776959722824, |
| "grad_norm": 1.3125, |
| "learning_rate": 3.951282966200624e-05, |
| "loss": 0.3291532039642334, |
| "step": 770 |
| }, |
| { |
| "epoch": 0.3378085751407536, |
| "grad_norm": 0.8671875, |
| "learning_rate": 3.922186922527677e-05, |
| "loss": 0.32024178504943845, |
| "step": 780 |
| }, |
| { |
| "epoch": 0.34213945430922477, |
| "grad_norm": 0.78515625, |
| "learning_rate": 3.892803233294942e-05, |
| "loss": 0.32744786739349363, |
| "step": 790 |
| }, |
| { |
| "epoch": 0.34647033347769596, |
| "grad_norm": 0.796875, |
| "learning_rate": 3.8631378415241135e-05, |
| "loss": 0.26966466903686526, |
| "step": 800 |
| }, |
| { |
| "epoch": 0.35080121264616715, |
| "grad_norm": 0.89453125, |
| "learning_rate": 3.833196747212865e-05, |
| "loss": 0.33243422508239745, |
| "step": 810 |
| }, |
| { |
| "epoch": 0.3551320918146384, |
| "grad_norm": 0.95703125, |
| "learning_rate": 3.802986006121304e-05, |
| "loss": 0.3162668228149414, |
| "step": 820 |
| }, |
| { |
| "epoch": 0.3594629709831096, |
| "grad_norm": 0.498046875, |
| "learning_rate": 3.772511728547168e-05, |
| "loss": 0.2786050319671631, |
| "step": 830 |
| }, |
| { |
| "epoch": 0.3637938501515808, |
| "grad_norm": 0.703125, |
| "learning_rate": 3.741780078089975e-05, |
| "loss": 0.3166247844696045, |
| "step": 840 |
| }, |
| { |
| "epoch": 0.36812472932005197, |
| "grad_norm": 0.84765625, |
| "learning_rate": 3.710797270404405e-05, |
| "loss": 0.3263724327087402, |
| "step": 850 |
| }, |
| { |
| "epoch": 0.37245560848852316, |
| "grad_norm": 0.498046875, |
| "learning_rate": 3.6795695719431436e-05, |
| "loss": 0.3051194667816162, |
| "step": 860 |
| }, |
| { |
| "epoch": 0.37678648765699435, |
| "grad_norm": 0.9375, |
| "learning_rate": 3.6481032986894566e-05, |
| "loss": 0.3087512254714966, |
| "step": 870 |
| }, |
| { |
| "epoch": 0.38111736682546554, |
| "grad_norm": 1.0859375, |
| "learning_rate": 3.616404814879748e-05, |
| "loss": 0.31677100658416746, |
| "step": 880 |
| }, |
| { |
| "epoch": 0.3854482459939368, |
| "grad_norm": 0.59375, |
| "learning_rate": 3.5844805317163525e-05, |
| "loss": 0.3010247707366943, |
| "step": 890 |
| }, |
| { |
| "epoch": 0.389779125162408, |
| "grad_norm": 0.86328125, |
| "learning_rate": 3.552336906070838e-05, |
| "loss": 0.2823305606842041, |
| "step": 900 |
| }, |
| { |
| "epoch": 0.39411000433087917, |
| "grad_norm": 1.046875, |
| "learning_rate": 3.5199804391780594e-05, |
| "loss": 0.3044945240020752, |
| "step": 910 |
| }, |
| { |
| "epoch": 0.39844088349935036, |
| "grad_norm": 1.265625, |
| "learning_rate": 3.48741767532125e-05, |
| "loss": 0.31587924957275393, |
| "step": 920 |
| }, |
| { |
| "epoch": 0.40277176266782155, |
| "grad_norm": 1.203125, |
| "learning_rate": 3.454655200508402e-05, |
| "loss": 0.32800912857055664, |
| "step": 930 |
| }, |
| { |
| "epoch": 0.40710264183629274, |
| "grad_norm": 0.8125, |
| "learning_rate": 3.4216996411402077e-05, |
| "loss": 0.30738139152526855, |
| "step": 940 |
| }, |
| { |
| "epoch": 0.411433521004764, |
| "grad_norm": 0.63671875, |
| "learning_rate": 3.388557662669831e-05, |
| "loss": 0.2966631889343262, |
| "step": 950 |
| }, |
| { |
| "epoch": 0.4157644001732352, |
| "grad_norm": 0.83984375, |
| "learning_rate": 3.355235968254782e-05, |
| "loss": 0.30721249580383303, |
| "step": 960 |
| }, |
| { |
| "epoch": 0.4200952793417064, |
| "grad_norm": 0.90234375, |
| "learning_rate": 3.321741297401162e-05, |
| "loss": 0.29646241664886475, |
| "step": 970 |
| }, |
| { |
| "epoch": 0.42442615851017756, |
| "grad_norm": 1.1875, |
| "learning_rate": 3.288080424600563e-05, |
| "loss": 0.30686221122741697, |
| "step": 980 |
| }, |
| { |
| "epoch": 0.42875703767864876, |
| "grad_norm": 0.76953125, |
| "learning_rate": 3.254260157959884e-05, |
| "loss": 0.2886993408203125, |
| "step": 990 |
| }, |
| { |
| "epoch": 0.43308791684711995, |
| "grad_norm": 0.6328125, |
| "learning_rate": 3.220287337824355e-05, |
| "loss": 0.2889398097991943, |
| "step": 1000 |
| }, |
| { |
| "epoch": 0.43741879601559114, |
| "grad_norm": 0.80859375, |
| "learning_rate": 3.186168835394032e-05, |
| "loss": 0.28464765548706056, |
| "step": 1010 |
| }, |
| { |
| "epoch": 0.4417496751840624, |
| "grad_norm": 0.87890625, |
| "learning_rate": 3.151911551334066e-05, |
| "loss": 0.2970520734786987, |
| "step": 1020 |
| }, |
| { |
| "epoch": 0.4460805543525336, |
| "grad_norm": 1.046875, |
| "learning_rate": 3.1175224143789946e-05, |
| "loss": 0.27363173961639403, |
| "step": 1030 |
| }, |
| { |
| "epoch": 0.45041143352100477, |
| "grad_norm": 1.0546875, |
| "learning_rate": 3.083008379931369e-05, |
| "loss": 0.3226848840713501, |
| "step": 1040 |
| }, |
| { |
| "epoch": 0.45474231268947596, |
| "grad_norm": 0.9921875, |
| "learning_rate": 3.0483764286549843e-05, |
| "loss": 0.3007674217224121, |
| "step": 1050 |
| }, |
| { |
| "epoch": 0.45907319185794715, |
| "grad_norm": 1.0859375, |
| "learning_rate": 3.013633565062999e-05, |
| "loss": 0.2832650661468506, |
| "step": 1060 |
| }, |
| { |
| "epoch": 0.46340407102641834, |
| "grad_norm": 0.60546875, |
| "learning_rate": 2.978786816101229e-05, |
| "loss": 0.30285255908966063, |
| "step": 1070 |
| }, |
| { |
| "epoch": 0.4677349501948896, |
| "grad_norm": 0.98828125, |
| "learning_rate": 2.9438432297269113e-05, |
| "loss": 0.3115823268890381, |
| "step": 1080 |
| }, |
| { |
| "epoch": 0.4720658293633608, |
| "grad_norm": 0.5625, |
| "learning_rate": 2.9088098734832105e-05, |
| "loss": 0.31290392875671386, |
| "step": 1090 |
| }, |
| { |
| "epoch": 0.47639670853183197, |
| "grad_norm": 0.95703125, |
| "learning_rate": 2.873693833069769e-05, |
| "loss": 0.32068135738372805, |
| "step": 1100 |
| }, |
| { |
| "epoch": 0.48072758770030316, |
| "grad_norm": 1.1640625, |
| "learning_rate": 2.8385022109095828e-05, |
| "loss": 0.29875409603118896, |
| "step": 1110 |
| }, |
| { |
| "epoch": 0.48505846686877435, |
| "grad_norm": 0.86328125, |
| "learning_rate": 2.803242124712493e-05, |
| "loss": 0.30289008617401125, |
| "step": 1120 |
| }, |
| { |
| "epoch": 0.48938934603724554, |
| "grad_norm": 0.8828125, |
| "learning_rate": 2.7679207060355912e-05, |
| "loss": 0.3129941463470459, |
| "step": 1130 |
| }, |
| { |
| "epoch": 0.49372022520571673, |
| "grad_norm": 0.8203125, |
| "learning_rate": 2.7325450988408185e-05, |
| "loss": 0.28847951889038087, |
| "step": 1140 |
| }, |
| { |
| "epoch": 0.498051104374188, |
| "grad_norm": 0.64453125, |
| "learning_rate": 2.6971224580500592e-05, |
| "loss": 0.3122821092605591, |
| "step": 1150 |
| }, |
| { |
| "epoch": 0.5006496318752707, |
| "eval_text_loss": 1.580957055091858, |
| "eval_text_model_preparation_time": 0.0199, |
| "eval_text_runtime": 23.7145, |
| "eval_text_samples_per_second": 13.916, |
| "eval_text_steps_per_second": 13.916, |
| "step": 1156 |
| }, |
| { |
| "epoch": 0.5006496318752707, |
| "eval_audio_loss": 0.8451715111732483, |
| "eval_audio_model_preparation_time": 0.0199, |
| "eval_audio_runtime": 46.4407, |
| "eval_audio_samples_per_second": 5.512, |
| "eval_audio_steps_per_second": 5.512, |
| "step": 1156 |
| }, |
| { |
| "epoch": 0.5023819835426592, |
| "grad_norm": 0.66796875, |
| "learning_rate": 2.6616599480980143e-05, |
| "loss": 0.32015998363494874, |
| "step": 1160 |
| }, |
| { |
| "epoch": 0.5067128627111304, |
| "grad_norm": 0.8828125, |
| "learning_rate": 2.626164741483154e-05, |
| "loss": 0.30613036155700685, |
| "step": 1170 |
| }, |
| { |
| "epoch": 0.5110437418796016, |
| "grad_norm": 0.84375, |
| "learning_rate": 2.5906440173170386e-05, |
| "loss": 0.30779805183410647, |
| "step": 1180 |
| }, |
| { |
| "epoch": 0.5153746210480727, |
| "grad_norm": 0.890625, |
| "learning_rate": 2.5551049598723027e-05, |
| "loss": 0.29211466312408446, |
| "step": 1190 |
| }, |
| { |
| "epoch": 0.5197055002165439, |
| "grad_norm": 1.0703125, |
| "learning_rate": 2.5195547571295898e-05, |
| "loss": 0.29266557693481443, |
| "step": 1200 |
| }, |
| { |
| "epoch": 0.5240363793850151, |
| "grad_norm": 0.5703125, |
| "learning_rate": 2.484000599323747e-05, |
| "loss": 0.2896392583847046, |
| "step": 1210 |
| }, |
| { |
| "epoch": 0.5283672585534863, |
| "grad_norm": 0.81640625, |
| "learning_rate": 2.448449677489555e-05, |
| "loss": 0.298609185218811, |
| "step": 1220 |
| }, |
| { |
| "epoch": 0.5326981377219575, |
| "grad_norm": 0.54296875, |
| "learning_rate": 2.4129091820073015e-05, |
| "loss": 0.29924740791320803, |
| "step": 1230 |
| }, |
| { |
| "epoch": 0.5370290168904288, |
| "grad_norm": 1.0078125, |
| "learning_rate": 2.377386301148482e-05, |
| "loss": 0.2744253635406494, |
| "step": 1240 |
| }, |
| { |
| "epoch": 0.5413598960589, |
| "grad_norm": 0.6796875, |
| "learning_rate": 2.341888219621934e-05, |
| "loss": 0.278640079498291, |
| "step": 1250 |
| }, |
| { |
| "epoch": 0.5456907752273712, |
| "grad_norm": 0.859375, |
| "learning_rate": 2.3064221171206856e-05, |
| "loss": 0.28020195960998534, |
| "step": 1260 |
| }, |
| { |
| "epoch": 0.5500216543958424, |
| "grad_norm": 0.625, |
| "learning_rate": 2.270995166869821e-05, |
| "loss": 0.30299355983734133, |
| "step": 1270 |
| }, |
| { |
| "epoch": 0.5543525335643136, |
| "grad_norm": 0.84375, |
| "learning_rate": 2.2356145341756548e-05, |
| "loss": 0.3237317562103271, |
| "step": 1280 |
| }, |
| { |
| "epoch": 0.5586834127327848, |
| "grad_norm": 0.76171875, |
| "learning_rate": 2.2002873749765076e-05, |
| "loss": 0.3258683681488037, |
| "step": 1290 |
| }, |
| { |
| "epoch": 0.563014291901256, |
| "grad_norm": 0.90625, |
| "learning_rate": 2.1650208343953747e-05, |
| "loss": 0.32569427490234376, |
| "step": 1300 |
| }, |
| { |
| "epoch": 0.5673451710697271, |
| "grad_norm": 0.76171875, |
| "learning_rate": 2.1298220452947826e-05, |
| "loss": 0.28685925006866453, |
| "step": 1310 |
| }, |
| { |
| "epoch": 0.5716760502381983, |
| "grad_norm": 0.765625, |
| "learning_rate": 2.0946981268341297e-05, |
| "loss": 0.27852678298950195, |
| "step": 1320 |
| }, |
| { |
| "epoch": 0.5760069294066695, |
| "grad_norm": 0.6640625, |
| "learning_rate": 2.059656183029792e-05, |
| "loss": 0.30152413845062254, |
| "step": 1330 |
| }, |
| { |
| "epoch": 0.5803378085751407, |
| "grad_norm": 0.984375, |
| "learning_rate": 2.0247033013182955e-05, |
| "loss": 0.30128428936004636, |
| "step": 1340 |
| }, |
| { |
| "epoch": 0.5846686877436119, |
| "grad_norm": 1.15625, |
| "learning_rate": 1.9898465511228416e-05, |
| "loss": 0.31016106605529786, |
| "step": 1350 |
| }, |
| { |
| "epoch": 0.5889995669120831, |
| "grad_norm": 0.7109375, |
| "learning_rate": 1.9550929824234736e-05, |
| "loss": 0.2842914342880249, |
| "step": 1360 |
| }, |
| { |
| "epoch": 0.5933304460805544, |
| "grad_norm": 0.8046875, |
| "learning_rate": 1.920449624331179e-05, |
| "loss": 0.2784018278121948, |
| "step": 1370 |
| }, |
| { |
| "epoch": 0.5976613252490256, |
| "grad_norm": 0.8359375, |
| "learning_rate": 1.8859234836662117e-05, |
| "loss": 0.28922812938690184, |
| "step": 1380 |
| }, |
| { |
| "epoch": 0.6019922044174968, |
| "grad_norm": 0.80859375, |
| "learning_rate": 1.851521543540916e-05, |
| "loss": 0.31638059616088865, |
| "step": 1390 |
| }, |
| { |
| "epoch": 0.606323083585968, |
| "grad_norm": 1.09375, |
| "learning_rate": 1.8172507619473614e-05, |
| "loss": 0.30915663242340086, |
| "step": 1400 |
| }, |
| { |
| "epoch": 0.6106539627544392, |
| "grad_norm": 0.68359375, |
| "learning_rate": 1.78311807035004e-05, |
| "loss": 0.31648037433624265, |
| "step": 1410 |
| }, |
| { |
| "epoch": 0.6149848419229104, |
| "grad_norm": 0.56640625, |
| "learning_rate": 1.749130372283942e-05, |
| "loss": 0.3143639326095581, |
| "step": 1420 |
| }, |
| { |
| "epoch": 0.6193157210913816, |
| "grad_norm": 0.9375, |
| "learning_rate": 1.715294541958274e-05, |
| "loss": 0.2913862943649292, |
| "step": 1430 |
| }, |
| { |
| "epoch": 0.6236466002598527, |
| "grad_norm": 0.703125, |
| "learning_rate": 1.6816174228661097e-05, |
| "loss": 0.28313374519348145, |
| "step": 1440 |
| }, |
| { |
| "epoch": 0.6279774794283239, |
| "grad_norm": 0.90234375, |
| "learning_rate": 1.648105826400256e-05, |
| "loss": 0.28623785972595217, |
| "step": 1450 |
| }, |
| { |
| "epoch": 0.6323083585967951, |
| "grad_norm": 0.9765625, |
| "learning_rate": 1.6147665304756084e-05, |
| "loss": 0.2705575227737427, |
| "step": 1460 |
| }, |
| { |
| "epoch": 0.6366392377652663, |
| "grad_norm": 0.78515625, |
| "learning_rate": 1.581606278158274e-05, |
| "loss": 0.30232059955596924, |
| "step": 1470 |
| }, |
| { |
| "epoch": 0.6409701169337375, |
| "grad_norm": 0.734375, |
| "learning_rate": 1.548631776301756e-05, |
| "loss": 0.29305553436279297, |
| "step": 1480 |
| }, |
| { |
| "epoch": 0.6453009961022087, |
| "grad_norm": 0.8515625, |
| "learning_rate": 1.5158496941904462e-05, |
| "loss": 0.2765927314758301, |
| "step": 1490 |
| }, |
| { |
| "epoch": 0.6496318752706799, |
| "grad_norm": 0.91015625, |
| "learning_rate": 1.4832666621907265e-05, |
| "loss": 0.2911043882369995, |
| "step": 1500 |
| }, |
| { |
| "epoch": 0.6539627544391512, |
| "grad_norm": 0.87109375, |
| "learning_rate": 1.4508892704099392e-05, |
| "loss": 0.2826990604400635, |
| "step": 1510 |
| }, |
| { |
| "epoch": 0.6582936336076224, |
| "grad_norm": 0.91796875, |
| "learning_rate": 1.4187240673634964e-05, |
| "loss": 0.2976381778717041, |
| "step": 1520 |
| }, |
| { |
| "epoch": 0.6626245127760936, |
| "grad_norm": 0.8359375, |
| "learning_rate": 1.3867775586504094e-05, |
| "loss": 0.28958122730255126, |
| "step": 1530 |
| }, |
| { |
| "epoch": 0.6669553919445648, |
| "grad_norm": 1.0625, |
| "learning_rate": 1.3550562056374908e-05, |
| "loss": 0.34233570098876953, |
| "step": 1540 |
| }, |
| { |
| "epoch": 0.671286271113036, |
| "grad_norm": 0.80859375, |
| "learning_rate": 1.3235664241525052e-05, |
| "loss": 0.30748577117919923, |
| "step": 1550 |
| }, |
| { |
| "epoch": 0.6756171502815072, |
| "grad_norm": 0.7265625, |
| "learning_rate": 1.2923145831865325e-05, |
| "loss": 0.279698920249939, |
| "step": 1560 |
| }, |
| { |
| "epoch": 0.6799480294499783, |
| "grad_norm": 0.78515625, |
| "learning_rate": 1.2613070036058005e-05, |
| "loss": 0.2836109399795532, |
| "step": 1570 |
| }, |
| { |
| "epoch": 0.6842789086184495, |
| "grad_norm": 0.80078125, |
| "learning_rate": 1.2305499568732554e-05, |
| "loss": 0.30521063804626464, |
| "step": 1580 |
| }, |
| { |
| "epoch": 0.6886097877869207, |
| "grad_norm": 0.7265625, |
| "learning_rate": 1.2000496637801195e-05, |
| "loss": 0.2962735891342163, |
| "step": 1590 |
| }, |
| { |
| "epoch": 0.6929406669553919, |
| "grad_norm": 0.9140625, |
| "learning_rate": 1.1698122931877018e-05, |
| "loss": 0.2767336845397949, |
| "step": 1600 |
| }, |
| { |
| "epoch": 0.6972715461238631, |
| "grad_norm": 0.78515625, |
| "learning_rate": 1.1398439607797098e-05, |
| "loss": 0.2979475975036621, |
| "step": 1610 |
| }, |
| { |
| "epoch": 0.7016024252923343, |
| "grad_norm": 0.8828125, |
| "learning_rate": 1.1101507278253167e-05, |
| "loss": 0.26526603698730467, |
| "step": 1620 |
| }, |
| { |
| "epoch": 0.7059333044608055, |
| "grad_norm": 0.9609375, |
| "learning_rate": 1.080738599953233e-05, |
| "loss": 0.28164918422698976, |
| "step": 1630 |
| }, |
| { |
| "epoch": 0.7102641836292768, |
| "grad_norm": 0.72265625, |
| "learning_rate": 1.0516135259370355e-05, |
| "loss": 0.3067033767700195, |
| "step": 1640 |
| }, |
| { |
| "epoch": 0.714595062797748, |
| "grad_norm": 0.91015625, |
| "learning_rate": 1.0227813964919938e-05, |
| "loss": 0.299686861038208, |
| "step": 1650 |
| }, |
| { |
| "epoch": 0.7189259419662192, |
| "grad_norm": 0.74609375, |
| "learning_rate": 9.94248043083636e-06, |
| "loss": 0.27935218811035156, |
| "step": 1660 |
| }, |
| { |
| "epoch": 0.7232568211346904, |
| "grad_norm": 1.1015625, |
| "learning_rate": 9.660192367483038e-06, |
| "loss": 0.2881230115890503, |
| "step": 1670 |
| }, |
| { |
| "epoch": 0.7275877003031616, |
| "grad_norm": 0.765625, |
| "learning_rate": 9.381006869259243e-06, |
| "loss": 0.29773364067077634, |
| "step": 1680 |
| }, |
| { |
| "epoch": 0.7319185794716327, |
| "grad_norm": 1.0390625, |
| "learning_rate": 9.104980403052458e-06, |
| "loss": 0.31337528228759765, |
| "step": 1690 |
| }, |
| { |
| "epoch": 0.7362494586401039, |
| "grad_norm": 0.58984375, |
| "learning_rate": 8.8321687968176e-06, |
| "loss": 0.29222004413604735, |
| "step": 1700 |
| }, |
| { |
| "epoch": 0.7405803378085751, |
| "grad_norm": 0.94140625, |
| "learning_rate": 8.562627228285478e-06, |
| "loss": 0.2752734661102295, |
| "step": 1710 |
| }, |
| { |
| "epoch": 0.7449112169770463, |
| "grad_norm": 0.76171875, |
| "learning_rate": 8.296410213802813e-06, |
| "loss": 0.2992286443710327, |
| "step": 1720 |
| }, |
| { |
| "epoch": 0.7492420961455175, |
| "grad_norm": 1.09375, |
| "learning_rate": 8.033571597305977e-06, |
| "loss": 0.3042688608169556, |
| "step": 1730 |
| }, |
| { |
| "epoch": 0.750974447812906, |
| "eval_text_loss": 1.555040955543518, |
| "eval_text_model_preparation_time": 0.0199, |
| "eval_text_runtime": 24.438, |
| "eval_text_samples_per_second": 13.504, |
| "eval_text_steps_per_second": 13.504, |
| "step": 1734 |
| }, |
| { |
| "epoch": 0.750974447812906, |
| "eval_audio_loss": 0.8402940034866333, |
| "eval_audio_model_preparation_time": 0.0199, |
| "eval_audio_runtime": 45.5183, |
| "eval_audio_samples_per_second": 5.624, |
| "eval_audio_steps_per_second": 5.624, |
| "step": 1734 |
| }, |
| { |
| "epoch": 0.7535729753139887, |
| "grad_norm": 0.8984375, |
| "learning_rate": 7.774164539430734e-06, |
| "loss": 0.300627326965332, |
| "step": 1740 |
| }, |
| { |
| "epoch": 0.7579038544824599, |
| "grad_norm": 0.87109375, |
| "learning_rate": 7.518241506760196e-06, |
| "loss": 0.2855875253677368, |
| "step": 1750 |
| }, |
| { |
| "epoch": 0.7622347336509311, |
| "grad_norm": 1.5, |
| "learning_rate": 7.265854261213156e-06, |
| "loss": 0.29245338439941404, |
| "step": 1760 |
| }, |
| { |
| "epoch": 0.7665656128194024, |
| "grad_norm": 0.56640625, |
| "learning_rate": 7.017053849574945e-06, |
| "loss": 0.2890751361846924, |
| "step": 1770 |
| }, |
| { |
| "epoch": 0.7708964919878736, |
| "grad_norm": 0.70703125, |
| "learning_rate": 6.771890593172911e-06, |
| "loss": 0.2851340055465698, |
| "step": 1780 |
| }, |
| { |
| "epoch": 0.7752273711563448, |
| "grad_norm": 1.0078125, |
| "learning_rate": 6.530414077698649e-06, |
| "loss": 0.2948427677154541, |
| "step": 1790 |
| }, |
| { |
| "epoch": 0.779558250324816, |
| "grad_norm": 0.8359375, |
| "learning_rate": 6.2926731431789954e-06, |
| "loss": 0.3046332597732544, |
| "step": 1800 |
| }, |
| { |
| "epoch": 0.7838891294932872, |
| "grad_norm": 0.7734375, |
| "learning_rate": 6.058715874097864e-06, |
| "loss": 0.31103029251098635, |
| "step": 1810 |
| }, |
| { |
| "epoch": 0.7882200086617583, |
| "grad_norm": 0.625, |
| "learning_rate": 5.828589589670871e-06, |
| "loss": 0.29731192588806155, |
| "step": 1820 |
| }, |
| { |
| "epoch": 0.7925508878302295, |
| "grad_norm": 0.87109375, |
| "learning_rate": 5.60234083427475e-06, |
| "loss": 0.3095412254333496, |
| "step": 1830 |
| }, |
| { |
| "epoch": 0.7968817669987007, |
| "grad_norm": 0.8125, |
| "learning_rate": 5.380015368033476e-06, |
| "loss": 0.29828534126281736, |
| "step": 1840 |
| }, |
| { |
| "epoch": 0.8012126461671719, |
| "grad_norm": 0.578125, |
| "learning_rate": 5.161658157563026e-06, |
| "loss": 0.2998827934265137, |
| "step": 1850 |
| }, |
| { |
| "epoch": 0.8055435253356431, |
| "grad_norm": 0.546875, |
| "learning_rate": 4.947313366876619e-06, |
| "loss": 0.3060743808746338, |
| "step": 1860 |
| }, |
| { |
| "epoch": 0.8098744045041143, |
| "grad_norm": 0.76953125, |
| "learning_rate": 4.737024348452282e-06, |
| "loss": 0.28541877269744875, |
| "step": 1870 |
| }, |
| { |
| "epoch": 0.8142052836725855, |
| "grad_norm": 0.77734375, |
| "learning_rate": 4.530833634464548e-06, |
| "loss": 0.3088233947753906, |
| "step": 1880 |
| }, |
| { |
| "epoch": 0.8185361628410567, |
| "grad_norm": 1.25, |
| "learning_rate": 4.328782928182104e-06, |
| "loss": 0.3078837156295776, |
| "step": 1890 |
| }, |
| { |
| "epoch": 0.822867042009528, |
| "grad_norm": 0.83984375, |
| "learning_rate": 4.130913095533046e-06, |
| "loss": 0.2788443088531494, |
| "step": 1900 |
| }, |
| { |
| "epoch": 0.8271979211779992, |
| "grad_norm": 1.1796875, |
| "learning_rate": 3.9372641568395125e-06, |
| "loss": 0.2981949090957642, |
| "step": 1910 |
| }, |
| { |
| "epoch": 0.8315288003464704, |
| "grad_norm": 0.8984375, |
| "learning_rate": 3.747875278723334e-06, |
| "loss": 0.29841878414154055, |
| "step": 1920 |
| }, |
| { |
| "epoch": 0.8358596795149416, |
| "grad_norm": 0.87890625, |
| "learning_rate": 3.562784766184371e-06, |
| "loss": 0.3009059190750122, |
| "step": 1930 |
| }, |
| { |
| "epoch": 0.8401905586834127, |
| "grad_norm": 0.7578125, |
| "learning_rate": 3.3820300548530943e-06, |
| "loss": 0.29599244594573976, |
| "step": 1940 |
| }, |
| { |
| "epoch": 0.8445214378518839, |
| "grad_norm": 0.8203125, |
| "learning_rate": 3.205647703419015e-06, |
| "loss": 0.2934562683105469, |
| "step": 1950 |
| }, |
| { |
| "epoch": 0.8488523170203551, |
| "grad_norm": 0.80859375, |
| "learning_rate": 3.0336733862364684e-06, |
| "loss": 0.29344346523284914, |
| "step": 1960 |
| }, |
| { |
| "epoch": 0.8531831961888263, |
| "grad_norm": 0.96875, |
| "learning_rate": 2.866141886109286e-06, |
| "loss": 0.2922006845474243, |
| "step": 1970 |
| }, |
| { |
| "epoch": 0.8575140753572975, |
| "grad_norm": 0.62109375, |
| "learning_rate": 2.7030870872557593e-06, |
| "loss": 0.2850653648376465, |
| "step": 1980 |
| }, |
| { |
| "epoch": 0.8618449545257687, |
| "grad_norm": 0.90625, |
| "learning_rate": 2.544541968455372e-06, |
| "loss": 0.2891335725784302, |
| "step": 1990 |
| }, |
| { |
| "epoch": 0.8661758336942399, |
| "grad_norm": 0.7109375, |
| "learning_rate": 2.39053859637863e-06, |
| "loss": 0.28892719745635986, |
| "step": 2000 |
| }, |
| { |
| "epoch": 0.8705067128627111, |
| "grad_norm": 0.72265625, |
| "learning_rate": 2.2411081191014093e-06, |
| "loss": 0.29591898918151854, |
| "step": 2010 |
| }, |
| { |
| "epoch": 0.8748375920311823, |
| "grad_norm": 1.046875, |
| "learning_rate": 2.096280759805069e-06, |
| "loss": 0.2645926237106323, |
| "step": 2020 |
| }, |
| { |
| "epoch": 0.8791684711996536, |
| "grad_norm": 1.1640625, |
| "learning_rate": 1.9560858106636408e-06, |
| "loss": 0.2921916007995605, |
| "step": 2030 |
| }, |
| { |
| "epoch": 0.8834993503681248, |
| "grad_norm": 0.9375, |
| "learning_rate": 1.8205516269193063e-06, |
| "loss": 0.2916860580444336, |
| "step": 2040 |
| }, |
| { |
| "epoch": 0.887830229536596, |
| "grad_norm": 0.75390625, |
| "learning_rate": 1.6897056211474054e-06, |
| "loss": 0.28798465728759765, |
| "step": 2050 |
| }, |
| { |
| "epoch": 0.8921611087050672, |
| "grad_norm": 0.84765625, |
| "learning_rate": 1.5635742577120638e-06, |
| "loss": 0.28080263137817385, |
| "step": 2060 |
| }, |
| { |
| "epoch": 0.8964919878735383, |
| "grad_norm": 0.59765625, |
| "learning_rate": 1.4421830474136339e-06, |
| "loss": 0.2958329677581787, |
| "step": 2070 |
| }, |
| { |
| "epoch": 0.9008228670420095, |
| "grad_norm": 0.859375, |
| "learning_rate": 1.3255565423289818e-06, |
| "loss": 0.29818201065063477, |
| "step": 2080 |
| }, |
| { |
| "epoch": 0.9051537462104807, |
| "grad_norm": 0.9453125, |
| "learning_rate": 1.2137183308456867e-06, |
| "loss": 0.30882911682128905, |
| "step": 2090 |
| }, |
| { |
| "epoch": 0.9094846253789519, |
| "grad_norm": 0.796875, |
| "learning_rate": 1.1066910328911579e-06, |
| "loss": 0.29765470027923585, |
| "step": 2100 |
| }, |
| { |
| "epoch": 0.9138155045474231, |
| "grad_norm": 0.8125, |
| "learning_rate": 1.0044962953576238e-06, |
| "loss": 0.28506147861480713, |
| "step": 2110 |
| }, |
| { |
| "epoch": 0.9181463837158943, |
| "grad_norm": 0.77734375, |
| "learning_rate": 9.071547877239017e-07, |
| "loss": 0.2811078310012817, |
| "step": 2120 |
| }, |
| { |
| "epoch": 0.9224772628843655, |
| "grad_norm": 0.73046875, |
| "learning_rate": 8.146861978749115e-07, |
| "loss": 0.2883314847946167, |
| "step": 2130 |
| }, |
| { |
| "epoch": 0.9268081420528367, |
| "grad_norm": 0.5234375, |
| "learning_rate": 7.271092281196573e-07, |
| "loss": 0.27544965744018557, |
| "step": 2140 |
| }, |
| { |
| "epoch": 0.9311390212213079, |
| "grad_norm": 0.97265625, |
| "learning_rate": 6.4444159140859e-07, |
| "loss": 0.3128631353378296, |
| "step": 2150 |
| }, |
| { |
| "epoch": 0.9354699003897792, |
| "grad_norm": 0.92578125, |
| "learning_rate": 5.667000077510559e-07, |
| "loss": 0.2900621652603149, |
| "step": 2160 |
| }, |
| { |
| "epoch": 0.9398007795582504, |
| "grad_norm": 0.63671875, |
| "learning_rate": 4.939002008335802e-07, |
| "loss": 0.3027785062789917, |
| "step": 2170 |
| }, |
| { |
| "epoch": 0.9441316587267216, |
| "grad_norm": 1.140625, |
| "learning_rate": 4.2605689483966037e-07, |
| "loss": 0.3042114734649658, |
| "step": 2180 |
| }, |
| { |
| "epoch": 0.9484625378951927, |
| "grad_norm": 0.91796875, |
| "learning_rate": 3.6318381147171234e-07, |
| "loss": 0.2658639669418335, |
| "step": 2190 |
| }, |
| { |
| "epoch": 0.9527934170636639, |
| "grad_norm": 0.75390625, |
| "learning_rate": 3.052936671757739e-07, |
| "loss": 0.28366703987121583, |
| "step": 2200 |
| }, |
| { |
| "epoch": 0.9571242962321351, |
| "grad_norm": 1.1796875, |
| "learning_rate": 2.523981705695427e-07, |
| "loss": 0.31629841327667235, |
| "step": 2210 |
| }, |
| { |
| "epoch": 0.9614551754006063, |
| "grad_norm": 1.171875, |
| "learning_rate": 2.0450802007422053e-07, |
| "loss": 0.3071019172668457, |
| "step": 2220 |
| }, |
| { |
| "epoch": 0.9657860545690775, |
| "grad_norm": 1.0078125, |
| "learning_rate": 1.6163290175071343e-07, |
| "loss": 0.30133790969848634, |
| "step": 2230 |
| }, |
| { |
| "epoch": 0.9701169337375487, |
| "grad_norm": 0.61328125, |
| "learning_rate": 1.2378148734056017e-07, |
| "loss": 0.2814929962158203, |
| "step": 2240 |
| }, |
| { |
| "epoch": 0.9744478129060199, |
| "grad_norm": 0.796875, |
| "learning_rate": 9.096143251202115e-08, |
| "loss": 0.28168976306915283, |
| "step": 2250 |
| }, |
| { |
| "epoch": 0.9787786920744911, |
| "grad_norm": 0.86328125, |
| "learning_rate": 6.317937531168106e-08, |
| "loss": 0.31104719638824463, |
| "step": 2260 |
| }, |
| { |
| "epoch": 0.9831095712429623, |
| "grad_norm": 0.7890625, |
| "learning_rate": 4.0440934821864286e-08, |
| "loss": 0.2767423868179321, |
| "step": 2270 |
| }, |
| { |
| "epoch": 0.9874404504114335, |
| "grad_norm": 0.87890625, |
| "learning_rate": 2.2750710024138646e-08, |
| "loss": 0.29538898468017577, |
| "step": 2280 |
| }, |
| { |
| "epoch": 0.9917713295799048, |
| "grad_norm": 0.71484375, |
| "learning_rate": 1.0112278869145453e-08, |
| "loss": 0.275164270401001, |
| "step": 2290 |
| }, |
| { |
| "epoch": 0.996102208748376, |
| "grad_norm": 1.1484375, |
| "learning_rate": 2.528197552942313e-09, |
| "loss": 0.30729990005493163, |
| "step": 2300 |
| }, |
| { |
| "epoch": 1.0, |
| "eval_text_loss": 1.5541865825653076, |
| "eval_text_model_preparation_time": 0.0199, |
| "eval_text_runtime": 24.8156, |
| "eval_text_samples_per_second": 13.298, |
| "eval_text_steps_per_second": 13.298, |
| "step": 2309 |
| }, |
| { |
| "epoch": 1.0, |
| "eval_audio_loss": 0.8393442034721375, |
| "eval_audio_model_preparation_time": 0.0199, |
| "eval_audio_runtime": 44.203, |
| "eval_audio_samples_per_second": 5.791, |
| "eval_audio_steps_per_second": 5.791, |
| "step": 2309 |
| } |
| ], |
| "logging_steps": 10, |
| "max_steps": 2309, |
| "num_input_tokens_seen": 0, |
| "num_train_epochs": 1, |
| "save_steps": 47, |
| "stateful_callbacks": { |
| "TrainerControl": { |
| "args": { |
| "should_epoch_stop": false, |
| "should_evaluate": false, |
| "should_log": false, |
| "should_save": true, |
| "should_training_stop": true |
| }, |
| "attributes": {} |
| } |
| }, |
| "total_flos": 2.2190042187360576e+17, |
| "train_batch_size": 4, |
| "trial_name": null, |
| "trial_params": null |
| } |
|
|