{ "best_global_step": 1200, "best_metric": 0.4533848760350299, "best_model_checkpoint": "./SALAMA_NEW99/checkpoint-1200", "epoch": 1.9575856443719413, "eval_steps": 600, "global_step": 1200, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.01631321370309951, "grad_norm": 0.8075992465019226, "learning_rate": 1.8e-07, "loss": 0.0109, "step": 10 }, { "epoch": 0.03262642740619902, "grad_norm": 1.4262839555740356, "learning_rate": 3.8e-07, "loss": 0.0094, "step": 20 }, { "epoch": 0.048939641109298535, "grad_norm": 2.3042991161346436, "learning_rate": 5.800000000000001e-07, "loss": 0.0093, "step": 30 }, { "epoch": 0.06525285481239804, "grad_norm": 0.546521782875061, "learning_rate": 7.8e-07, "loss": 0.0218, "step": 40 }, { "epoch": 0.08156606851549755, "grad_norm": 2.1409292221069336, "learning_rate": 9.800000000000001e-07, "loss": 0.0077, "step": 50 }, { "epoch": 0.09787928221859707, "grad_norm": 1.9533841609954834, "learning_rate": 1.1800000000000001e-06, "loss": 0.0061, "step": 60 }, { "epoch": 0.11419249592169657, "grad_norm": 1.0335050821304321, "learning_rate": 1.3800000000000001e-06, "loss": 0.0245, "step": 70 }, { "epoch": 0.13050570962479607, "grad_norm": 1.158021330833435, "learning_rate": 1.5800000000000001e-06, "loss": 0.009, "step": 80 }, { "epoch": 0.1468189233278956, "grad_norm": 2.7232327461242676, "learning_rate": 1.7800000000000001e-06, "loss": 0.0197, "step": 90 }, { "epoch": 0.1631321370309951, "grad_norm": 1.2299067974090576, "learning_rate": 1.98e-06, "loss": 0.007, "step": 100 }, { "epoch": 0.17944535073409462, "grad_norm": 2.173428535461426, "learning_rate": 2.1800000000000003e-06, "loss": 0.0098, "step": 110 }, { "epoch": 0.19575856443719414, "grad_norm": 1.4795992374420166, "learning_rate": 2.38e-06, "loss": 0.0098, "step": 120 }, { "epoch": 0.21207177814029363, "grad_norm": 0.48007962107658386, "learning_rate": 2.5800000000000003e-06, "loss": 0.0258, "step": 130 }, { "epoch": 0.22838499184339314, "grad_norm": 0.5283637642860413, "learning_rate": 2.7800000000000005e-06, "loss": 0.0078, "step": 140 }, { "epoch": 0.24469820554649266, "grad_norm": 1.748868465423584, "learning_rate": 2.9800000000000003e-06, "loss": 0.0083, "step": 150 }, { "epoch": 0.26101141924959215, "grad_norm": 2.2862608432769775, "learning_rate": 3.1800000000000005e-06, "loss": 0.012, "step": 160 }, { "epoch": 0.27732463295269166, "grad_norm": 1.9614684581756592, "learning_rate": 3.3800000000000007e-06, "loss": 0.0093, "step": 170 }, { "epoch": 0.2936378466557912, "grad_norm": 0.2269970178604126, "learning_rate": 3.58e-06, "loss": 0.0079, "step": 180 }, { "epoch": 0.3099510603588907, "grad_norm": 3.722224712371826, "learning_rate": 3.7800000000000002e-06, "loss": 0.0123, "step": 190 }, { "epoch": 0.3262642740619902, "grad_norm": 1.0197768211364746, "learning_rate": 3.980000000000001e-06, "loss": 0.0126, "step": 200 }, { "epoch": 0.3425774877650897, "grad_norm": 2.238046884536743, "learning_rate": 4.18e-06, "loss": 0.0155, "step": 210 }, { "epoch": 0.35889070146818924, "grad_norm": 2.2540082931518555, "learning_rate": 4.38e-06, "loss": 0.0064, "step": 220 }, { "epoch": 0.37520391517128876, "grad_norm": 4.272034168243408, "learning_rate": 4.58e-06, "loss": 0.0144, "step": 230 }, { "epoch": 0.3915171288743883, "grad_norm": 1.297784447669983, "learning_rate": 4.78e-06, "loss": 0.0183, "step": 240 }, { "epoch": 0.4078303425774878, "grad_norm": 1.1175105571746826, "learning_rate": 4.980000000000001e-06, "loss": 0.01, "step": 250 }, { "epoch": 0.42414355628058725, "grad_norm": 0.989632248878479, "learning_rate": 5.18e-06, "loss": 0.0185, "step": 260 }, { "epoch": 0.44045676998368677, "grad_norm": 2.2630860805511475, "learning_rate": 5.380000000000001e-06, "loss": 0.004, "step": 270 }, { "epoch": 0.4567699836867863, "grad_norm": 3.4459376335144043, "learning_rate": 5.580000000000001e-06, "loss": 0.0338, "step": 280 }, { "epoch": 0.4730831973898858, "grad_norm": 2.0962679386138916, "learning_rate": 5.78e-06, "loss": 0.0078, "step": 290 }, { "epoch": 0.4893964110929853, "grad_norm": 1.109755277633667, "learning_rate": 5.98e-06, "loss": 0.0133, "step": 300 }, { "epoch": 0.5057096247960848, "grad_norm": 1.8200187683105469, "learning_rate": 6.18e-06, "loss": 0.0128, "step": 310 }, { "epoch": 0.5220228384991843, "grad_norm": 0.8147674798965454, "learning_rate": 6.380000000000001e-06, "loss": 0.0139, "step": 320 }, { "epoch": 0.5383360522022839, "grad_norm": 2.6240575313568115, "learning_rate": 6.5800000000000005e-06, "loss": 0.0153, "step": 330 }, { "epoch": 0.5546492659053833, "grad_norm": 0.49103260040283203, "learning_rate": 6.780000000000001e-06, "loss": 0.0119, "step": 340 }, { "epoch": 0.5709624796084829, "grad_norm": 1.4433389902114868, "learning_rate": 6.98e-06, "loss": 0.0169, "step": 350 }, { "epoch": 0.5872756933115824, "grad_norm": 1.6671804189682007, "learning_rate": 7.180000000000001e-06, "loss": 0.0103, "step": 360 }, { "epoch": 0.6035889070146819, "grad_norm": 1.7344307899475098, "learning_rate": 7.3800000000000005e-06, "loss": 0.0234, "step": 370 }, { "epoch": 0.6199021207177814, "grad_norm": 1.4996588230133057, "learning_rate": 7.58e-06, "loss": 0.0138, "step": 380 }, { "epoch": 0.636215334420881, "grad_norm": 1.5783976316452026, "learning_rate": 7.78e-06, "loss": 0.0219, "step": 390 }, { "epoch": 0.6525285481239804, "grad_norm": 0.5455562472343445, "learning_rate": 7.980000000000002e-06, "loss": 0.0113, "step": 400 }, { "epoch": 0.6688417618270799, "grad_norm": 1.6277507543563843, "learning_rate": 8.18e-06, "loss": 0.018, "step": 410 }, { "epoch": 0.6851549755301795, "grad_norm": 2.473850965499878, "learning_rate": 8.380000000000001e-06, "loss": 0.0224, "step": 420 }, { "epoch": 0.7014681892332789, "grad_norm": 1.4439480304718018, "learning_rate": 8.580000000000001e-06, "loss": 0.0156, "step": 430 }, { "epoch": 0.7177814029363785, "grad_norm": 0.7241881489753723, "learning_rate": 8.78e-06, "loss": 0.0131, "step": 440 }, { "epoch": 0.734094616639478, "grad_norm": 2.1885428428649902, "learning_rate": 8.98e-06, "loss": 0.019, "step": 450 }, { "epoch": 0.7504078303425775, "grad_norm": 2.702345371246338, "learning_rate": 9.180000000000002e-06, "loss": 0.0253, "step": 460 }, { "epoch": 0.766721044045677, "grad_norm": 3.975011110305786, "learning_rate": 9.38e-06, "loss": 0.0107, "step": 470 }, { "epoch": 0.7830342577487766, "grad_norm": 2.3299198150634766, "learning_rate": 9.58e-06, "loss": 0.0212, "step": 480 }, { "epoch": 0.799347471451876, "grad_norm": 1.317580223083496, "learning_rate": 9.780000000000001e-06, "loss": 0.0227, "step": 490 }, { "epoch": 0.8156606851549756, "grad_norm": 2.4023892879486084, "learning_rate": 9.980000000000001e-06, "loss": 0.0365, "step": 500 }, { "epoch": 0.831973898858075, "grad_norm": 2.641399383544922, "learning_rate": 9.932785660941001e-06, "loss": 0.0138, "step": 510 }, { "epoch": 0.8482871125611745, "grad_norm": 2.7177205085754395, "learning_rate": 9.858103061986558e-06, "loss": 0.0145, "step": 520 }, { "epoch": 0.8646003262642741, "grad_norm": 1.9215168952941895, "learning_rate": 9.783420463032114e-06, "loss": 0.0305, "step": 530 }, { "epoch": 0.8809135399673735, "grad_norm": 2.8683998584747314, "learning_rate": 9.708737864077671e-06, "loss": 0.0183, "step": 540 }, { "epoch": 0.8972267536704731, "grad_norm": 1.9078762531280518, "learning_rate": 9.634055265123227e-06, "loss": 0.0235, "step": 550 }, { "epoch": 0.9135399673735726, "grad_norm": 3.7367427349090576, "learning_rate": 9.559372666168784e-06, "loss": 0.0175, "step": 560 }, { "epoch": 0.9298531810766721, "grad_norm": 1.7846318483352661, "learning_rate": 9.48469006721434e-06, "loss": 0.0138, "step": 570 }, { "epoch": 0.9461663947797716, "grad_norm": 3.428025245666504, "learning_rate": 9.410007468259897e-06, "loss": 0.0168, "step": 580 }, { "epoch": 0.9624796084828712, "grad_norm": 2.601677417755127, "learning_rate": 9.335324869305452e-06, "loss": 0.0192, "step": 590 }, { "epoch": 0.9787928221859706, "grad_norm": 3.509207248687744, "learning_rate": 9.26064227035101e-06, "loss": 0.0177, "step": 600 }, { "epoch": 0.9787928221859706, "eval_loss": 0.011165497824549675, "eval_runtime": 1766.1438, "eval_samples_per_second": 2.775, "eval_steps_per_second": 0.347, "eval_wer": 1.0189228529839884, "step": 600 }, { "epoch": 0.9951060358890701, "grad_norm": 2.4465436935424805, "learning_rate": 9.185959671396565e-06, "loss": 0.0168, "step": 610 }, { "epoch": 1.0114192495921697, "grad_norm": 3.492405414581299, "learning_rate": 9.111277072442123e-06, "loss": 0.0096, "step": 620 }, { "epoch": 1.0277324632952691, "grad_norm": 0.9142910242080688, "learning_rate": 9.036594473487678e-06, "loss": 0.0106, "step": 630 }, { "epoch": 1.0440456769983686, "grad_norm": 0.7352218627929688, "learning_rate": 8.961911874533234e-06, "loss": 0.0085, "step": 640 }, { "epoch": 1.0603588907014683, "grad_norm": 1.1586534976959229, "learning_rate": 8.887229275578791e-06, "loss": 0.0068, "step": 650 }, { "epoch": 1.0766721044045677, "grad_norm": 2.758241891860962, "learning_rate": 8.812546676624347e-06, "loss": 0.0084, "step": 660 }, { "epoch": 1.0929853181076672, "grad_norm": 0.44431474804878235, "learning_rate": 8.737864077669904e-06, "loss": 0.0087, "step": 670 }, { "epoch": 1.1092985318107667, "grad_norm": 1.2769174575805664, "learning_rate": 8.66318147871546e-06, "loss": 0.0096, "step": 680 }, { "epoch": 1.1256117455138663, "grad_norm": 2.656780242919922, "learning_rate": 8.588498879761017e-06, "loss": 0.0127, "step": 690 }, { "epoch": 1.1419249592169658, "grad_norm": 1.0895354747772217, "learning_rate": 8.513816280806572e-06, "loss": 0.0112, "step": 700 }, { "epoch": 1.1582381729200653, "grad_norm": 1.0903260707855225, "learning_rate": 8.43913368185213e-06, "loss": 0.0091, "step": 710 }, { "epoch": 1.1745513866231647, "grad_norm": 2.6359965801239014, "learning_rate": 8.364451082897685e-06, "loss": 0.0193, "step": 720 }, { "epoch": 1.1908646003262642, "grad_norm": 0.9313237071037292, "learning_rate": 8.289768483943242e-06, "loss": 0.0073, "step": 730 }, { "epoch": 1.2071778140293639, "grad_norm": 0.7475559115409851, "learning_rate": 8.215085884988798e-06, "loss": 0.0075, "step": 740 }, { "epoch": 1.2234910277324633, "grad_norm": 0.24283719062805176, "learning_rate": 8.140403286034355e-06, "loss": 0.0071, "step": 750 }, { "epoch": 1.2398042414355628, "grad_norm": 0.6300885677337646, "learning_rate": 8.06572068707991e-06, "loss": 0.006, "step": 760 }, { "epoch": 1.2561174551386622, "grad_norm": 1.7239911556243896, "learning_rate": 7.991038088125468e-06, "loss": 0.0137, "step": 770 }, { "epoch": 1.272430668841762, "grad_norm": 1.5679051876068115, "learning_rate": 7.916355489171023e-06, "loss": 0.0047, "step": 780 }, { "epoch": 1.2887438825448614, "grad_norm": 1.0264155864715576, "learning_rate": 7.84167289021658e-06, "loss": 0.0074, "step": 790 }, { "epoch": 1.3050570962479608, "grad_norm": 1.0627104043960571, "learning_rate": 7.766990291262136e-06, "loss": 0.0106, "step": 800 }, { "epoch": 1.3213703099510603, "grad_norm": 1.262370228767395, "learning_rate": 7.692307692307694e-06, "loss": 0.0072, "step": 810 }, { "epoch": 1.3376835236541598, "grad_norm": 1.5836150646209717, "learning_rate": 7.617625093353249e-06, "loss": 0.0092, "step": 820 }, { "epoch": 1.3539967373572595, "grad_norm": 1.4232890605926514, "learning_rate": 7.542942494398806e-06, "loss": 0.0089, "step": 830 }, { "epoch": 1.370309951060359, "grad_norm": 0.49621379375457764, "learning_rate": 7.468259895444362e-06, "loss": 0.0066, "step": 840 }, { "epoch": 1.3866231647634584, "grad_norm": 1.060057520866394, "learning_rate": 7.393577296489919e-06, "loss": 0.0065, "step": 850 }, { "epoch": 1.4029363784665578, "grad_norm": 1.5507824420928955, "learning_rate": 7.318894697535475e-06, "loss": 0.0086, "step": 860 }, { "epoch": 1.4192495921696575, "grad_norm": 0.6949241757392883, "learning_rate": 7.244212098581031e-06, "loss": 0.0066, "step": 870 }, { "epoch": 1.435562805872757, "grad_norm": 3.595614194869995, "learning_rate": 7.169529499626588e-06, "loss": 0.0084, "step": 880 }, { "epoch": 1.4518760195758564, "grad_norm": 0.38814422488212585, "learning_rate": 7.094846900672144e-06, "loss": 0.0086, "step": 890 }, { "epoch": 1.468189233278956, "grad_norm": 0.612360417842865, "learning_rate": 7.0201643017177005e-06, "loss": 0.0087, "step": 900 }, { "epoch": 1.4845024469820554, "grad_norm": 1.359798789024353, "learning_rate": 6.945481702763257e-06, "loss": 0.0052, "step": 910 }, { "epoch": 1.5008156606851548, "grad_norm": 2.678591251373291, "learning_rate": 6.870799103808813e-06, "loss": 0.0079, "step": 920 }, { "epoch": 1.5171288743882545, "grad_norm": 1.1047519445419312, "learning_rate": 6.79611650485437e-06, "loss": 0.0047, "step": 930 }, { "epoch": 1.533442088091354, "grad_norm": 1.798194169998169, "learning_rate": 6.721433905899925e-06, "loss": 0.0125, "step": 940 }, { "epoch": 1.5497553017944536, "grad_norm": 0.44330868124961853, "learning_rate": 6.6467513069454825e-06, "loss": 0.0074, "step": 950 }, { "epoch": 1.566068515497553, "grad_norm": 1.6538296937942505, "learning_rate": 6.572068707991038e-06, "loss": 0.0118, "step": 960 }, { "epoch": 1.5823817292006526, "grad_norm": 1.2385334968566895, "learning_rate": 6.497386109036595e-06, "loss": 0.0091, "step": 970 }, { "epoch": 1.598694942903752, "grad_norm": 1.1827936172485352, "learning_rate": 6.422703510082151e-06, "loss": 0.008, "step": 980 }, { "epoch": 1.6150081566068515, "grad_norm": 1.1670842170715332, "learning_rate": 6.348020911127708e-06, "loss": 0.004, "step": 990 }, { "epoch": 1.631321370309951, "grad_norm": 1.9073982238769531, "learning_rate": 6.273338312173264e-06, "loss": 0.0105, "step": 1000 }, { "epoch": 1.6476345840130504, "grad_norm": 0.6385743021965027, "learning_rate": 6.198655713218821e-06, "loss": 0.0047, "step": 1010 }, { "epoch": 1.66394779771615, "grad_norm": 0.717432975769043, "learning_rate": 6.123973114264377e-06, "loss": 0.0104, "step": 1020 }, { "epoch": 1.6802610114192496, "grad_norm": 1.6229373216629028, "learning_rate": 6.049290515309934e-06, "loss": 0.014, "step": 1030 }, { "epoch": 1.6965742251223492, "grad_norm": 1.5651880502700806, "learning_rate": 5.974607916355489e-06, "loss": 0.0081, "step": 1040 }, { "epoch": 1.7128874388254487, "grad_norm": 1.444212794303894, "learning_rate": 5.899925317401046e-06, "loss": 0.0099, "step": 1050 }, { "epoch": 1.7292006525285482, "grad_norm": 1.1551159620285034, "learning_rate": 5.825242718446602e-06, "loss": 0.0087, "step": 1060 }, { "epoch": 1.7455138662316476, "grad_norm": 0.7630689740180969, "learning_rate": 5.750560119492159e-06, "loss": 0.0056, "step": 1070 }, { "epoch": 1.761827079934747, "grad_norm": 0.6756051778793335, "learning_rate": 5.675877520537715e-06, "loss": 0.0057, "step": 1080 }, { "epoch": 1.7781402936378465, "grad_norm": 0.47443750500679016, "learning_rate": 5.6011949215832715e-06, "loss": 0.0068, "step": 1090 }, { "epoch": 1.7944535073409462, "grad_norm": 0.8193678855895996, "learning_rate": 5.526512322628828e-06, "loss": 0.0079, "step": 1100 }, { "epoch": 1.8107667210440457, "grad_norm": 0.7489306330680847, "learning_rate": 5.451829723674384e-06, "loss": 0.0072, "step": 1110 }, { "epoch": 1.8270799347471451, "grad_norm": 0.9458585977554321, "learning_rate": 5.37714712471994e-06, "loss": 0.0047, "step": 1120 }, { "epoch": 1.8433931484502448, "grad_norm": 0.8624395132064819, "learning_rate": 5.302464525765497e-06, "loss": 0.0053, "step": 1130 }, { "epoch": 1.8597063621533443, "grad_norm": 0.5177104473114014, "learning_rate": 5.227781926811053e-06, "loss": 0.0074, "step": 1140 }, { "epoch": 1.8760195758564437, "grad_norm": 1.2944625616073608, "learning_rate": 5.15309932785661e-06, "loss": 0.0151, "step": 1150 }, { "epoch": 1.8923327895595432, "grad_norm": 1.9204165935516357, "learning_rate": 5.0784167289021655e-06, "loss": 0.0082, "step": 1160 }, { "epoch": 1.9086460032626427, "grad_norm": 2.2888307571411133, "learning_rate": 5.003734129947723e-06, "loss": 0.0057, "step": 1170 }, { "epoch": 1.9249592169657421, "grad_norm": 1.0358986854553223, "learning_rate": 4.929051530993279e-06, "loss": 0.0087, "step": 1180 }, { "epoch": 1.9412724306688418, "grad_norm": 0.8882740139961243, "learning_rate": 4.854368932038836e-06, "loss": 0.0111, "step": 1190 }, { "epoch": 1.9575856443719413, "grad_norm": 1.2315634489059448, "learning_rate": 4.779686333084392e-06, "loss": 0.0055, "step": 1200 }, { "epoch": 1.9575856443719413, "eval_loss": 0.005837247706949711, "eval_runtime": 1757.0337, "eval_samples_per_second": 2.789, "eval_steps_per_second": 0.349, "eval_wer": 0.4533848760350299, "step": 1200 } ], "logging_steps": 10, "max_steps": 1839, "num_input_tokens_seen": 0, "num_train_epochs": 3, "save_steps": 600, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 1.107792776208384e+19, "train_batch_size": 16, "trial_name": null, "trial_params": null }