| { | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 2.0, | |
| "eval_steps": 500, | |
| "global_step": 776, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.041237113402061855, | |
| "grad_norm": 0.09626785665750504, | |
| "learning_rate": 0.0006666666666666666, | |
| "loss": 4.9272, | |
| "step": 16 | |
| }, | |
| { | |
| "epoch": 0.08247422680412371, | |
| "grad_norm": 0.09776347130537033, | |
| "learning_rate": 0.0009997207818651273, | |
| "loss": 4.8403, | |
| "step": 32 | |
| }, | |
| { | |
| "epoch": 0.12371134020618557, | |
| "grad_norm": 0.18136683106422424, | |
| "learning_rate": 0.0009974889075442521, | |
| "loss": 4.9176, | |
| "step": 48 | |
| }, | |
| { | |
| "epoch": 0.16494845360824742, | |
| "grad_norm": 0.10246019065380096, | |
| "learning_rate": 0.0009930351269950143, | |
| "loss": 4.8586, | |
| "step": 64 | |
| }, | |
| { | |
| "epoch": 0.20618556701030927, | |
| "grad_norm": 0.26106250286102295, | |
| "learning_rate": 0.0009863793318825185, | |
| "loss": 4.8768, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.24742268041237114, | |
| "grad_norm": 0.13016994297504425, | |
| "learning_rate": 0.0009775512486034563, | |
| "loss": 4.9537, | |
| "step": 96 | |
| }, | |
| { | |
| "epoch": 0.28865979381443296, | |
| "grad_norm": 0.14532029628753662, | |
| "learning_rate": 0.0009665903055208013, | |
| "loss": 4.9051, | |
| "step": 112 | |
| }, | |
| { | |
| "epoch": 0.32989690721649484, | |
| "grad_norm": 0.12766750156879425, | |
| "learning_rate": 0.0009535454568671704, | |
| "loss": 4.9145, | |
| "step": 128 | |
| }, | |
| { | |
| "epoch": 0.3711340206185567, | |
| "grad_norm": 0.16211313009262085, | |
| "learning_rate": 0.0009384749641033358, | |
| "loss": 4.9382, | |
| "step": 144 | |
| }, | |
| { | |
| "epoch": 0.41237113402061853, | |
| "grad_norm": 0.18606019020080566, | |
| "learning_rate": 0.0009214461357083985, | |
| "loss": 4.8932, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 0.4536082474226804, | |
| "grad_norm": 0.10332559049129486, | |
| "learning_rate": 0.0009025350265637815, | |
| "loss": 4.8723, | |
| "step": 176 | |
| }, | |
| { | |
| "epoch": 0.4948453608247423, | |
| "grad_norm": 0.21772044897079468, | |
| "learning_rate": 0.0008818260982736661, | |
| "loss": 4.875, | |
| "step": 192 | |
| }, | |
| { | |
| "epoch": 0.5360824742268041, | |
| "grad_norm": 0.1752215325832367, | |
| "learning_rate": 0.0008594118419389648, | |
| "loss": 4.9184, | |
| "step": 208 | |
| }, | |
| { | |
| "epoch": 0.5773195876288659, | |
| "grad_norm": 0.15603622794151306, | |
| "learning_rate": 0.0008353923650696118, | |
| "loss": 4.8573, | |
| "step": 224 | |
| }, | |
| { | |
| "epoch": 0.6185567010309279, | |
| "grad_norm": 0.19524474442005157, | |
| "learning_rate": 0.0008098749444801224, | |
| "loss": 4.9145, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 0.6597938144329897, | |
| "grad_norm": 0.14055490493774414, | |
| "learning_rate": 0.0007829735471652976, | |
| "loss": 4.9301, | |
| "step": 256 | |
| }, | |
| { | |
| "epoch": 0.7010309278350515, | |
| "grad_norm": 0.1492496281862259, | |
| "learning_rate": 0.0007548083212959587, | |
| "loss": 4.8914, | |
| "step": 272 | |
| }, | |
| { | |
| "epoch": 0.7422680412371134, | |
| "grad_norm": 0.20217546820640564, | |
| "learning_rate": 0.0007255050596080509, | |
| "loss": 4.9268, | |
| "step": 288 | |
| }, | |
| { | |
| "epoch": 0.7835051546391752, | |
| "grad_norm": 0.12439503520727158, | |
| "learning_rate": 0.0006951946375817474, | |
| "loss": 4.8838, | |
| "step": 304 | |
| }, | |
| { | |
| "epoch": 0.8247422680412371, | |
| "grad_norm": 0.15163038671016693, | |
| "learning_rate": 0.0006640124289197845, | |
| "loss": 4.9025, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 0.865979381443299, | |
| "grad_norm": 0.15424683690071106, | |
| "learning_rate": 0.0006320977009356432, | |
| "loss": 4.867, | |
| "step": 336 | |
| }, | |
| { | |
| "epoch": 0.9072164948453608, | |
| "grad_norm": 0.12109530717134476, | |
| "learning_rate": 0.000599592992551918, | |
| "loss": 4.9068, | |
| "step": 352 | |
| }, | |
| { | |
| "epoch": 0.9484536082474226, | |
| "grad_norm": 0.15034383535385132, | |
| "learning_rate": 0.0005666434776868895, | |
| "loss": 4.8993, | |
| "step": 368 | |
| }, | |
| { | |
| "epoch": 0.9896907216494846, | |
| "grad_norm": 0.1772756278514862, | |
| "learning_rate": 0.0005333963168725609, | |
| "loss": 4.8522, | |
| "step": 384 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "eval_bleu": 0.08197595877772341, | |
| "eval_cap_loss": 1.7311131333567433, | |
| "eval_con_loss": 2.0061442756775727, | |
| "eval_loss": 3.737257410570518, | |
| "step": 388 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "eval_bleu": 0.08197595877772341, | |
| "eval_cap_loss": 1.7311131333567433, | |
| "eval_con_loss": 2.0061442756775727, | |
| "eval_loss": 3.737257410570518, | |
| "eval_runtime": 142.0128, | |
| "eval_samples_per_second": 21.836, | |
| "eval_steps_per_second": 2.732, | |
| "step": 388 | |
| }, | |
| { | |
| "epoch": 1.0309278350515463, | |
| "grad_norm": 0.1598144769668579, | |
| "learning_rate": 0.0005, | |
| "loss": 4.8689, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 1.0721649484536082, | |
| "grad_norm": 0.15655860304832458, | |
| "learning_rate": 0.0004666036831274392, | |
| "loss": 4.8865, | |
| "step": 416 | |
| }, | |
| { | |
| "epoch": 1.1134020618556701, | |
| "grad_norm": 0.15307271480560303, | |
| "learning_rate": 0.00043335652231311076, | |
| "loss": 4.8548, | |
| "step": 432 | |
| }, | |
| { | |
| "epoch": 1.1546391752577319, | |
| "grad_norm": 0.07996341586112976, | |
| "learning_rate": 0.00040040700744808206, | |
| "loss": 4.8571, | |
| "step": 448 | |
| }, | |
| { | |
| "epoch": 1.1958762886597938, | |
| "grad_norm": 0.14137883484363556, | |
| "learning_rate": 0.000367902299064357, | |
| "loss": 4.9429, | |
| "step": 464 | |
| }, | |
| { | |
| "epoch": 1.2371134020618557, | |
| "grad_norm": 0.14231054484844208, | |
| "learning_rate": 0.00033598757108021545, | |
| "loss": 4.8607, | |
| "step": 480 | |
| }, | |
| { | |
| "epoch": 1.2783505154639174, | |
| "grad_norm": 0.1808304786682129, | |
| "learning_rate": 0.00030480536241825264, | |
| "loss": 4.8901, | |
| "step": 496 | |
| }, | |
| { | |
| "epoch": 1.3195876288659794, | |
| "grad_norm": 0.14347229897975922, | |
| "learning_rate": 0.000274494940391949, | |
| "loss": 4.839, | |
| "step": 512 | |
| }, | |
| { | |
| "epoch": 1.3608247422680413, | |
| "grad_norm": 0.12875933945178986, | |
| "learning_rate": 0.00024519167870404123, | |
| "loss": 4.8809, | |
| "step": 528 | |
| }, | |
| { | |
| "epoch": 1.402061855670103, | |
| "grad_norm": 0.17960529029369354, | |
| "learning_rate": 0.00021702645283470234, | |
| "loss": 4.848, | |
| "step": 544 | |
| }, | |
| { | |
| "epoch": 1.443298969072165, | |
| "grad_norm": 0.12456343322992325, | |
| "learning_rate": 0.00019012505551987762, | |
| "loss": 4.8828, | |
| "step": 560 | |
| }, | |
| { | |
| "epoch": 1.4845360824742269, | |
| "grad_norm": 0.1851247102022171, | |
| "learning_rate": 0.0001646076349303884, | |
| "loss": 4.8708, | |
| "step": 576 | |
| }, | |
| { | |
| "epoch": 1.5257731958762886, | |
| "grad_norm": 0.1590530276298523, | |
| "learning_rate": 0.00014058815806103543, | |
| "loss": 4.8611, | |
| "step": 592 | |
| }, | |
| { | |
| "epoch": 1.5670103092783505, | |
| "grad_norm": 0.13291551172733307, | |
| "learning_rate": 0.00011817390172633402, | |
| "loss": 4.9032, | |
| "step": 608 | |
| }, | |
| { | |
| "epoch": 1.6082474226804124, | |
| "grad_norm": 0.20336174964904785, | |
| "learning_rate": 9.746497343621857e-05, | |
| "loss": 4.8976, | |
| "step": 624 | |
| }, | |
| { | |
| "epoch": 1.6494845360824741, | |
| "grad_norm": 0.13562831282615662, | |
| "learning_rate": 7.85538642916015e-05, | |
| "loss": 4.8911, | |
| "step": 640 | |
| }, | |
| { | |
| "epoch": 1.690721649484536, | |
| "grad_norm": 0.13604836165905, | |
| "learning_rate": 6.152503589666425e-05, | |
| "loss": 4.8916, | |
| "step": 656 | |
| }, | |
| { | |
| "epoch": 1.731958762886598, | |
| "grad_norm": 0.15518011152744293, | |
| "learning_rate": 4.645454313282965e-05, | |
| "loss": 4.8618, | |
| "step": 672 | |
| }, | |
| { | |
| "epoch": 1.7731958762886597, | |
| "grad_norm": 0.11967703700065613, | |
| "learning_rate": 3.340969447919873e-05, | |
| "loss": 4.9006, | |
| "step": 688 | |
| }, | |
| { | |
| "epoch": 1.8144329896907216, | |
| "grad_norm": 0.13278520107269287, | |
| "learning_rate": 2.2448751396543787e-05, | |
| "loss": 4.9169, | |
| "step": 704 | |
| }, | |
| { | |
| "epoch": 1.8556701030927836, | |
| "grad_norm": 0.15354037284851074, | |
| "learning_rate": 1.3620668117481471e-05, | |
| "loss": 4.8556, | |
| "step": 720 | |
| }, | |
| { | |
| "epoch": 1.8969072164948453, | |
| "grad_norm": 0.14049023389816284, | |
| "learning_rate": 6.964873004985717e-06, | |
| "loss": 4.8451, | |
| "step": 736 | |
| }, | |
| { | |
| "epoch": 1.9381443298969072, | |
| "grad_norm": 0.13459183275699615, | |
| "learning_rate": 2.511092455747932e-06, | |
| "loss": 4.9148, | |
| "step": 752 | |
| }, | |
| { | |
| "epoch": 1.9793814432989691, | |
| "grad_norm": 0.1388990879058838, | |
| "learning_rate": 2.792181348726941e-07, | |
| "loss": 4.8364, | |
| "step": 768 | |
| }, | |
| { | |
| "epoch": 2.0, | |
| "eval_bleu": 0.083526303042587, | |
| "eval_cap_loss": 1.723467393327005, | |
| "eval_con_loss": 1.9934953146988583, | |
| "eval_loss": 3.7169627135561916, | |
| "step": 776 | |
| }, | |
| { | |
| "epoch": 2.0, | |
| "eval_bleu": 0.083526303042587, | |
| "eval_cap_loss": 1.723467393327005, | |
| "eval_con_loss": 1.9934953146988583, | |
| "eval_loss": 3.7169627135561916, | |
| "eval_runtime": 142.3121, | |
| "eval_samples_per_second": 21.79, | |
| "eval_steps_per_second": 2.726, | |
| "step": 776 | |
| } | |
| ], | |
| "logging_steps": 16, | |
| "max_steps": 776, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 2, | |
| "save_steps": 500, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": true | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 0.0, | |
| "train_batch_size": 32, | |
| "trial_name": null, | |
| "trial_params": null, | |
| "tau_value": 2.5431 | |
| } | |