| { | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 3.0, | |
| "eval_steps": 500, | |
| "global_step": 549, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.0546448087431694, | |
| "grad_norm": 9.216311123646195, | |
| "learning_rate": 5e-06, | |
| "loss": 1.0197, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.1092896174863388, | |
| "grad_norm": 2.84822970052699, | |
| "learning_rate": 5e-06, | |
| "loss": 0.9201, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.16393442622950818, | |
| "grad_norm": 1.1976026595489564, | |
| "learning_rate": 5e-06, | |
| "loss": 0.8893, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.2185792349726776, | |
| "grad_norm": 0.7151750657967473, | |
| "learning_rate": 5e-06, | |
| "loss": 0.8686, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.273224043715847, | |
| "grad_norm": 0.7178785706016054, | |
| "learning_rate": 5e-06, | |
| "loss": 0.8566, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.32786885245901637, | |
| "grad_norm": 1.3738089444275707, | |
| "learning_rate": 5e-06, | |
| "loss": 0.8464, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.3825136612021858, | |
| "grad_norm": 0.7307703509770203, | |
| "learning_rate": 5e-06, | |
| "loss": 0.837, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 0.4371584699453552, | |
| "grad_norm": 0.9793900236646151, | |
| "learning_rate": 5e-06, | |
| "loss": 0.834, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.4918032786885246, | |
| "grad_norm": 0.6074469784310406, | |
| "learning_rate": 5e-06, | |
| "loss": 0.8239, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 0.546448087431694, | |
| "grad_norm": 0.7581415259721198, | |
| "learning_rate": 5e-06, | |
| "loss": 0.8203, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.6010928961748634, | |
| "grad_norm": 0.6207562995736402, | |
| "learning_rate": 5e-06, | |
| "loss": 0.8191, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 0.6557377049180327, | |
| "grad_norm": 0.712118501956403, | |
| "learning_rate": 5e-06, | |
| "loss": 0.8148, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 0.7103825136612022, | |
| "grad_norm": 0.5406181591483699, | |
| "learning_rate": 5e-06, | |
| "loss": 0.816, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 0.7650273224043715, | |
| "grad_norm": 0.801157400991226, | |
| "learning_rate": 5e-06, | |
| "loss": 0.8084, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 0.819672131147541, | |
| "grad_norm": 1.3907122538264838, | |
| "learning_rate": 5e-06, | |
| "loss": 0.8101, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 0.8743169398907104, | |
| "grad_norm": 0.9374210060236701, | |
| "learning_rate": 5e-06, | |
| "loss": 0.8095, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 0.9289617486338798, | |
| "grad_norm": 0.7357026277429077, | |
| "learning_rate": 5e-06, | |
| "loss": 0.8056, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 0.9836065573770492, | |
| "grad_norm": 0.6439903150890136, | |
| "learning_rate": 5e-06, | |
| "loss": 0.8067, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "eval_loss": 0.8027574419975281, | |
| "eval_runtime": 18.4703, | |
| "eval_samples_per_second": 266.536, | |
| "eval_steps_per_second": 1.083, | |
| "step": 183 | |
| }, | |
| { | |
| "epoch": 1.0382513661202186, | |
| "grad_norm": 0.9901943553002889, | |
| "learning_rate": 5e-06, | |
| "loss": 0.7796, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 1.092896174863388, | |
| "grad_norm": 0.6647757287607619, | |
| "learning_rate": 5e-06, | |
| "loss": 0.7669, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 1.1475409836065573, | |
| "grad_norm": 0.9580968284481025, | |
| "learning_rate": 5e-06, | |
| "loss": 0.7664, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 1.2021857923497268, | |
| "grad_norm": 0.8889371423217748, | |
| "learning_rate": 5e-06, | |
| "loss": 0.7675, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 1.2568306010928962, | |
| "grad_norm": 0.9247757328115808, | |
| "learning_rate": 5e-06, | |
| "loss": 0.7625, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 1.3114754098360657, | |
| "grad_norm": 0.7030804123659421, | |
| "learning_rate": 5e-06, | |
| "loss": 0.7669, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 1.366120218579235, | |
| "grad_norm": 0.6820492241392998, | |
| "learning_rate": 5e-06, | |
| "loss": 0.7679, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 1.4207650273224044, | |
| "grad_norm": 0.5636718084741676, | |
| "learning_rate": 5e-06, | |
| "loss": 0.7606, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 1.4754098360655736, | |
| "grad_norm": 0.639581027268557, | |
| "learning_rate": 5e-06, | |
| "loss": 0.7593, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 1.530054644808743, | |
| "grad_norm": 0.6987258742060128, | |
| "learning_rate": 5e-06, | |
| "loss": 0.7612, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 1.5846994535519126, | |
| "grad_norm": 0.6936443486720386, | |
| "learning_rate": 5e-06, | |
| "loss": 0.7582, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 1.639344262295082, | |
| "grad_norm": 0.5505691855153351, | |
| "learning_rate": 5e-06, | |
| "loss": 0.7625, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 1.6939890710382515, | |
| "grad_norm": 0.5640162883535096, | |
| "learning_rate": 5e-06, | |
| "loss": 0.7626, | |
| "step": 310 | |
| }, | |
| { | |
| "epoch": 1.748633879781421, | |
| "grad_norm": 0.6321872003984678, | |
| "learning_rate": 5e-06, | |
| "loss": 0.764, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 1.8032786885245902, | |
| "grad_norm": 0.6338738498810821, | |
| "learning_rate": 5e-06, | |
| "loss": 0.762, | |
| "step": 330 | |
| }, | |
| { | |
| "epoch": 1.8579234972677594, | |
| "grad_norm": 0.5229579892525594, | |
| "learning_rate": 5e-06, | |
| "loss": 0.7672, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 1.9125683060109289, | |
| "grad_norm": 0.7287452929104943, | |
| "learning_rate": 5e-06, | |
| "loss": 0.7615, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 1.9672131147540983, | |
| "grad_norm": 0.7034416943673282, | |
| "learning_rate": 5e-06, | |
| "loss": 0.7596, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 2.0, | |
| "eval_loss": 0.7888949513435364, | |
| "eval_runtime": 18.3314, | |
| "eval_samples_per_second": 268.556, | |
| "eval_steps_per_second": 1.091, | |
| "step": 366 | |
| }, | |
| { | |
| "epoch": 2.021857923497268, | |
| "grad_norm": 1.0144238515169914, | |
| "learning_rate": 5e-06, | |
| "loss": 0.7482, | |
| "step": 370 | |
| }, | |
| { | |
| "epoch": 2.0765027322404372, | |
| "grad_norm": 0.8603434912719397, | |
| "learning_rate": 5e-06, | |
| "loss": 0.7216, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 2.1311475409836067, | |
| "grad_norm": 0.7563179754120521, | |
| "learning_rate": 5e-06, | |
| "loss": 0.7175, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 2.185792349726776, | |
| "grad_norm": 0.9559125673842068, | |
| "learning_rate": 5e-06, | |
| "loss": 0.7222, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 2.240437158469945, | |
| "grad_norm": 0.769800549346291, | |
| "learning_rate": 5e-06, | |
| "loss": 0.7184, | |
| "step": 410 | |
| }, | |
| { | |
| "epoch": 2.2950819672131146, | |
| "grad_norm": 0.6419920496502396, | |
| "learning_rate": 5e-06, | |
| "loss": 0.7222, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 2.349726775956284, | |
| "grad_norm": 0.7532650987823766, | |
| "learning_rate": 5e-06, | |
| "loss": 0.7246, | |
| "step": 430 | |
| }, | |
| { | |
| "epoch": 2.4043715846994536, | |
| "grad_norm": 0.5920199594002061, | |
| "learning_rate": 5e-06, | |
| "loss": 0.7228, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 2.459016393442623, | |
| "grad_norm": 0.6161056536206126, | |
| "learning_rate": 5e-06, | |
| "loss": 0.722, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 2.5136612021857925, | |
| "grad_norm": 0.7438641211945972, | |
| "learning_rate": 5e-06, | |
| "loss": 0.7215, | |
| "step": 460 | |
| }, | |
| { | |
| "epoch": 2.5683060109289615, | |
| "grad_norm": 0.6857540384247945, | |
| "learning_rate": 5e-06, | |
| "loss": 0.7255, | |
| "step": 470 | |
| }, | |
| { | |
| "epoch": 2.6229508196721314, | |
| "grad_norm": 0.5697372445993161, | |
| "learning_rate": 5e-06, | |
| "loss": 0.7231, | |
| "step": 480 | |
| }, | |
| { | |
| "epoch": 2.6775956284153004, | |
| "grad_norm": 0.868972289179613, | |
| "learning_rate": 5e-06, | |
| "loss": 0.7238, | |
| "step": 490 | |
| }, | |
| { | |
| "epoch": 2.73224043715847, | |
| "grad_norm": 0.6159867401214144, | |
| "learning_rate": 5e-06, | |
| "loss": 0.7208, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 2.7868852459016393, | |
| "grad_norm": 0.584105759494615, | |
| "learning_rate": 5e-06, | |
| "loss": 0.7263, | |
| "step": 510 | |
| }, | |
| { | |
| "epoch": 2.841530054644809, | |
| "grad_norm": 0.5906676150264403, | |
| "learning_rate": 5e-06, | |
| "loss": 0.7221, | |
| "step": 520 | |
| }, | |
| { | |
| "epoch": 2.8961748633879782, | |
| "grad_norm": 0.5994810510441251, | |
| "learning_rate": 5e-06, | |
| "loss": 0.7266, | |
| "step": 530 | |
| }, | |
| { | |
| "epoch": 2.9508196721311473, | |
| "grad_norm": 0.7788656644195687, | |
| "learning_rate": 5e-06, | |
| "loss": 0.7194, | |
| "step": 540 | |
| }, | |
| { | |
| "epoch": 3.0, | |
| "eval_loss": 0.7868959903717041, | |
| "eval_runtime": 18.0047, | |
| "eval_samples_per_second": 273.428, | |
| "eval_steps_per_second": 1.111, | |
| "step": 549 | |
| }, | |
| { | |
| "epoch": 3.0, | |
| "step": 549, | |
| "total_flos": 919595447746560.0, | |
| "train_loss": 0.7765402672286459, | |
| "train_runtime": 3626.8952, | |
| "train_samples_per_second": 77.367, | |
| "train_steps_per_second": 0.151 | |
| } | |
| ], | |
| "logging_steps": 10, | |
| "max_steps": 549, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 3, | |
| "save_steps": 500, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": true | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 919595447746560.0, | |
| "train_batch_size": 16, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |