{ "best_metric": 35.9517533349309, "best_model_checkpoint": "./whisper-lora-15k-adapters/checkpoint-1896", "epoch": 2.2201405152224822, "eval_steps": 237, "global_step": 1896, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.02927400468384075, "grad_norm": 0.578233540058136, "learning_rate": 0.0005, "loss": 0.99, "step": 25 }, { "epoch": 0.0585480093676815, "grad_norm": 0.3136115074157715, "learning_rate": 0.001, "loss": 0.7869, "step": 50 }, { "epoch": 0.08782201405152225, "grad_norm": 0.3556165397167206, "learning_rate": 0.0009940758293838863, "loss": 0.6559, "step": 75 }, { "epoch": 0.117096018735363, "grad_norm": 0.43047836422920227, "learning_rate": 0.0009881516587677726, "loss": 0.6799, "step": 100 }, { "epoch": 0.14637002341920374, "grad_norm": 0.41161859035491943, "learning_rate": 0.0009822274881516586, "loss": 0.6179, "step": 125 }, { "epoch": 0.1756440281030445, "grad_norm": 0.3486640453338623, "learning_rate": 0.000976303317535545, "loss": 0.6218, "step": 150 }, { "epoch": 0.20491803278688525, "grad_norm": 0.33961209654808044, "learning_rate": 0.0009703791469194313, "loss": 0.5623, "step": 175 }, { "epoch": 0.234192037470726, "grad_norm": 0.4211539328098297, "learning_rate": 0.0009644549763033176, "loss": 0.6293, "step": 200 }, { "epoch": 0.26346604215456676, "grad_norm": 0.4401342272758484, "learning_rate": 0.0009585308056872039, "loss": 0.65, "step": 225 }, { "epoch": 0.2775175644028103, "eval_loss": 0.6182317733764648, "eval_runtime": 12305.8898, "eval_samples_per_second": 0.123, "eval_steps_per_second": 0.008, "eval_wer": 42.11536326582399, "step": 237 }, { "epoch": 0.2927400468384075, "grad_norm": 0.5222854614257812, "learning_rate": 0.0009526066350710901, "loss": 0.6494, "step": 250 }, { "epoch": 0.32201405152224827, "grad_norm": 0.5739536285400391, "learning_rate": 0.0009466824644549763, "loss": 0.5656, "step": 275 }, { "epoch": 0.351288056206089, "grad_norm": 0.4213266968727112, "learning_rate": 0.0009407582938388626, "loss": 0.5781, "step": 300 }, { "epoch": 0.3805620608899297, "grad_norm": 0.5185717344284058, "learning_rate": 0.0009348341232227489, "loss": 0.5993, "step": 325 }, { "epoch": 0.4098360655737705, "grad_norm": 0.41156110167503357, "learning_rate": 0.0009289099526066352, "loss": 0.5504, "step": 350 }, { "epoch": 0.43911007025761123, "grad_norm": 0.44983068108558655, "learning_rate": 0.0009229857819905212, "loss": 0.642, "step": 375 }, { "epoch": 0.468384074941452, "grad_norm": 0.7018289566040039, "learning_rate": 0.0009170616113744075, "loss": 0.6313, "step": 400 }, { "epoch": 0.49765807962529274, "grad_norm": 0.41570019721984863, "learning_rate": 0.0009111374407582938, "loss": 0.642, "step": 425 }, { "epoch": 0.5269320843091335, "grad_norm": 0.2906375229358673, "learning_rate": 0.0009052132701421801, "loss": 0.5501, "step": 450 }, { "epoch": 0.5550351288056206, "eval_loss": 0.5893104076385498, "eval_runtime": 12217.9829, "eval_samples_per_second": 0.124, "eval_steps_per_second": 0.008, "eval_wer": 39.71420868158533, "step": 474 }, { "epoch": 0.5562060889929742, "grad_norm": 0.42602404952049255, "learning_rate": 0.0008992890995260664, "loss": 0.6419, "step": 475 }, { "epoch": 0.585480093676815, "grad_norm": 0.45508912205696106, "learning_rate": 0.0008933649289099525, "loss": 0.5816, "step": 500 }, { "epoch": 0.6147540983606558, "grad_norm": 0.5000929236412048, "learning_rate": 0.0008874407582938388, "loss": 0.6941, "step": 525 }, { "epoch": 0.6440281030444965, "grad_norm": 0.4415169656276703, "learning_rate": 0.0008815165876777251, "loss": 0.5615, "step": 550 }, { "epoch": 0.6733021077283372, "grad_norm": 0.5120753049850464, "learning_rate": 0.0008755924170616114, "loss": 0.559, "step": 575 }, { "epoch": 0.702576112412178, "grad_norm": 0.3653784990310669, "learning_rate": 0.0008696682464454977, "loss": 0.5836, "step": 600 }, { "epoch": 0.7318501170960188, "grad_norm": 0.5504665374755859, "learning_rate": 0.0008637440758293838, "loss": 0.6163, "step": 625 }, { "epoch": 0.7611241217798594, "grad_norm": 0.49855440855026245, "learning_rate": 0.0008578199052132701, "loss": 0.5482, "step": 650 }, { "epoch": 0.7903981264637002, "grad_norm": 0.3784034848213196, "learning_rate": 0.0008518957345971564, "loss": 0.5572, "step": 675 }, { "epoch": 0.819672131147541, "grad_norm": 0.5111596584320068, "learning_rate": 0.0008459715639810427, "loss": 0.565, "step": 700 }, { "epoch": 0.832552693208431, "eval_loss": 0.5823442339897156, "eval_runtime": 12225.5714, "eval_samples_per_second": 0.124, "eval_steps_per_second": 0.008, "eval_wer": 37.68099852505035, "step": 711 }, { "epoch": 0.8489461358313818, "grad_norm": 0.5943437218666077, "learning_rate": 0.000840047393364929, "loss": 0.5146, "step": 725 }, { "epoch": 0.8782201405152225, "grad_norm": 0.5228826403617859, "learning_rate": 0.0008341232227488151, "loss": 0.5338, "step": 750 }, { "epoch": 0.9074941451990632, "grad_norm": 0.44550982117652893, "learning_rate": 0.0008281990521327014, "loss": 0.5631, "step": 775 }, { "epoch": 0.936768149882904, "grad_norm": 0.5326892733573914, "learning_rate": 0.0008222748815165877, "loss": 0.5489, "step": 800 }, { "epoch": 0.9660421545667447, "grad_norm": 0.5083812475204468, "learning_rate": 0.000816350710900474, "loss": 0.5336, "step": 825 }, { "epoch": 0.9953161592505855, "grad_norm": 0.4346718192100525, "learning_rate": 0.0008104265402843603, "loss": 0.6155, "step": 850 }, { "epoch": 1.0245901639344261, "grad_norm": 0.4419436454772949, "learning_rate": 0.0008045023696682464, "loss": 0.5506, "step": 875 }, { "epoch": 1.053864168618267, "grad_norm": 0.5935924649238586, "learning_rate": 0.0007985781990521327, "loss": 0.5407, "step": 900 }, { "epoch": 1.0831381733021077, "grad_norm": 0.4228830635547638, "learning_rate": 0.000792654028436019, "loss": 0.5527, "step": 925 }, { "epoch": 1.1100702576112411, "eval_loss": 0.5060898065567017, "eval_runtime": 12332.7416, "eval_samples_per_second": 0.123, "eval_steps_per_second": 0.008, "eval_wer": 37.76065419091052, "step": 948 }, { "epoch": 1.1124121779859484, "grad_norm": 0.37129494547843933, "learning_rate": 0.0007867298578199053, "loss": 0.5191, "step": 950 }, { "epoch": 1.1416861826697893, "grad_norm": 0.7254778146743774, "learning_rate": 0.0007808056872037916, "loss": 0.5537, "step": 975 }, { "epoch": 1.17096018735363, "grad_norm": 0.4878183603286743, "learning_rate": 0.0007748815165876777, "loss": 0.5281, "step": 1000 }, { "epoch": 1.2002341920374708, "grad_norm": 0.35084593296051025, "learning_rate": 0.000768957345971564, "loss": 0.5166, "step": 1025 }, { "epoch": 1.2295081967213115, "grad_norm": 0.5030648708343506, "learning_rate": 0.0007630331753554502, "loss": 0.5284, "step": 1050 }, { "epoch": 1.2587822014051522, "grad_norm": 0.5004339218139648, "learning_rate": 0.0007571090047393365, "loss": 0.5695, "step": 1075 }, { "epoch": 1.288056206088993, "grad_norm": 0.5789551734924316, "learning_rate": 0.0007511848341232228, "loss": 0.5511, "step": 1100 }, { "epoch": 1.3173302107728337, "grad_norm": 0.389371782541275, "learning_rate": 0.0007452606635071089, "loss": 0.5661, "step": 1125 }, { "epoch": 1.3466042154566744, "grad_norm": 0.38161447644233704, "learning_rate": 0.0007393364928909952, "loss": 0.5087, "step": 1150 }, { "epoch": 1.3758782201405153, "grad_norm": 0.40263721346855164, "learning_rate": 0.0007334123222748815, "loss": 0.5147, "step": 1175 }, { "epoch": 1.3875878220140514, "eval_loss": 0.5079419016838074, "eval_runtime": 12322.9428, "eval_samples_per_second": 0.123, "eval_steps_per_second": 0.008, "eval_wer": 38.37238559521937, "step": 1185 }, { "epoch": 1.405152224824356, "grad_norm": 0.5195249319076538, "learning_rate": 0.0007274881516587678, "loss": 0.5409, "step": 1200 }, { "epoch": 1.4344262295081966, "grad_norm": 0.40098896622657776, "learning_rate": 0.0007215639810426541, "loss": 0.5121, "step": 1225 }, { "epoch": 1.4637002341920375, "grad_norm": 0.42950162291526794, "learning_rate": 0.0007156398104265402, "loss": 0.5155, "step": 1250 }, { "epoch": 1.4929742388758782, "grad_norm": 0.38044580817222595, "learning_rate": 0.0007097156398104265, "loss": 0.4727, "step": 1275 }, { "epoch": 1.5222482435597189, "grad_norm": 0.38700923323631287, "learning_rate": 0.0007037914691943128, "loss": 0.498, "step": 1300 }, { "epoch": 1.5515222482435598, "grad_norm": 0.4633864760398865, "learning_rate": 0.0006978672985781991, "loss": 0.5297, "step": 1325 }, { "epoch": 1.5807962529274004, "grad_norm": 0.48980265855789185, "learning_rate": 0.0006919431279620854, "loss": 0.4732, "step": 1350 }, { "epoch": 1.6100702576112411, "grad_norm": 0.3389205038547516, "learning_rate": 0.0006860189573459715, "loss": 0.5461, "step": 1375 }, { "epoch": 1.639344262295082, "grad_norm": 0.3686542510986328, "learning_rate": 0.0006800947867298578, "loss": 0.5033, "step": 1400 }, { "epoch": 1.6651053864168617, "eval_loss": 0.5103082060813904, "eval_runtime": 12252.9597, "eval_samples_per_second": 0.124, "eval_steps_per_second": 0.008, "eval_wer": 39.455228053358894, "step": 1422 }, { "epoch": 1.6686182669789227, "grad_norm": 0.46823108196258545, "learning_rate": 0.0006741706161137441, "loss": 0.5392, "step": 1425 }, { "epoch": 1.6978922716627634, "grad_norm": 0.5638931393623352, "learning_rate": 0.0006682464454976304, "loss": 0.5253, "step": 1450 }, { "epoch": 1.7271662763466042, "grad_norm": 0.5234322547912598, "learning_rate": 0.0006623222748815167, "loss": 0.51, "step": 1475 }, { "epoch": 1.756440281030445, "grad_norm": 0.5467631816864014, "learning_rate": 0.0006563981042654028, "loss": 0.5436, "step": 1500 }, { "epoch": 1.7857142857142856, "grad_norm": 0.3867318034172058, "learning_rate": 0.0006504739336492891, "loss": 0.5142, "step": 1525 }, { "epoch": 1.8149882903981265, "grad_norm": 0.4091216027736664, "learning_rate": 0.0006445497630331754, "loss": 0.5345, "step": 1550 }, { "epoch": 1.8442622950819674, "grad_norm": 0.44898247718811035, "learning_rate": 0.0006386255924170617, "loss": 0.4937, "step": 1575 }, { "epoch": 1.8735362997658078, "grad_norm": 0.3484508991241455, "learning_rate": 0.000632701421800948, "loss": 0.489, "step": 1600 }, { "epoch": 1.9028103044496487, "grad_norm": 0.5735388398170471, "learning_rate": 0.0006267772511848341, "loss": 0.4742, "step": 1625 }, { "epoch": 1.9320843091334896, "grad_norm": 0.7618733048439026, "learning_rate": 0.0006208530805687204, "loss": 0.5559, "step": 1650 }, { "epoch": 1.9426229508196722, "eval_loss": 0.5032439827919006, "eval_runtime": 12233.7713, "eval_samples_per_second": 0.124, "eval_steps_per_second": 0.008, "eval_wer": 39.20600686955827, "step": 1659 }, { "epoch": 1.96135831381733, "grad_norm": 0.3670201003551483, "learning_rate": 0.0006149289099526067, "loss": 0.4868, "step": 1675 }, { "epoch": 1.990632318501171, "grad_norm": 0.4840170443058014, "learning_rate": 0.000609004739336493, "loss": 0.5458, "step": 1700 }, { "epoch": 2.019906323185012, "grad_norm": 0.30357852578163147, "learning_rate": 0.0006030805687203791, "loss": 0.4845, "step": 1725 }, { "epoch": 2.0491803278688523, "grad_norm": 0.43158742785453796, "learning_rate": 0.0005971563981042653, "loss": 0.5007, "step": 1750 }, { "epoch": 2.078454332552693, "grad_norm": 0.46644917130470276, "learning_rate": 0.0005912322274881516, "loss": 0.4558, "step": 1775 }, { "epoch": 2.107728337236534, "grad_norm": 0.42779088020324707, "learning_rate": 0.0005853080568720379, "loss": 0.4736, "step": 1800 }, { "epoch": 2.1370023419203745, "grad_norm": 0.4596354067325592, "learning_rate": 0.0005793838862559242, "loss": 0.4338, "step": 1825 }, { "epoch": 2.1662763466042154, "grad_norm": 0.5213513970375061, "learning_rate": 0.0005734597156398104, "loss": 0.4657, "step": 1850 }, { "epoch": 2.1955503512880563, "grad_norm": 0.30604368448257446, "learning_rate": 0.0005675355450236966, "loss": 0.4697, "step": 1875 }, { "epoch": 2.2201405152224822, "eval_loss": 0.49964743852615356, "eval_runtime": 12221.7958, "eval_samples_per_second": 0.124, "eval_steps_per_second": 0.008, "eval_wer": 35.9517533349309, "step": 1896 } ], "logging_steps": 25, "max_steps": 4270, "num_input_tokens_seen": 0, "num_train_epochs": 5, "save_steps": 237, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 3.137332291043328e+19, "train_batch_size": 16, "trial_name": null, "trial_params": null }