{ "best_metric": 0.0019312179647386074, "best_model_checkpoint": "./vision_transformer_model_progress/checkpoint-200", "epoch": 9.0, "eval_steps": 100, "global_step": 297, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.30303030303030304, "grad_norm": 0.12155108898878098, "learning_rate": 2.0000000000000003e-06, "loss": 0.0158, "step": 10 }, { "epoch": 0.6060606060606061, "grad_norm": 1.0541080236434937, "learning_rate": 4.000000000000001e-06, "loss": 0.0479, "step": 20 }, { "epoch": 0.9090909090909091, "grad_norm": 0.03966180980205536, "learning_rate": 6e-06, "loss": 0.0047, "step": 30 }, { "epoch": 1.2121212121212122, "grad_norm": 0.06989104300737381, "learning_rate": 8.000000000000001e-06, "loss": 0.0045, "step": 40 }, { "epoch": 1.5151515151515151, "grad_norm": 0.0959181860089302, "learning_rate": 1e-05, "loss": 0.0027, "step": 50 }, { "epoch": 1.8181818181818183, "grad_norm": 0.05803044140338898, "learning_rate": 1.2e-05, "loss": 0.0281, "step": 60 }, { "epoch": 2.121212121212121, "grad_norm": 0.24542172253131866, "learning_rate": 1.4000000000000001e-05, "loss": 0.0049, "step": 70 }, { "epoch": 2.4242424242424243, "grad_norm": 0.0824781134724617, "learning_rate": 1.6000000000000003e-05, "loss": 0.0018, "step": 80 }, { "epoch": 2.7272727272727275, "grad_norm": 0.05576428771018982, "learning_rate": 1.8e-05, "loss": 0.0015, "step": 90 }, { "epoch": 3.0303030303030303, "grad_norm": 0.02234538644552231, "learning_rate": 2e-05, "loss": 0.0018, "step": 100 }, { "epoch": 3.0303030303030303, "eval_loss": 0.0030471435748040676, "eval_runtime": 12.9149, "eval_samples_per_second": 10.066, "eval_steps_per_second": 0.697, "step": 100 }, { "epoch": 3.3333333333333335, "grad_norm": 0.014021256007254124, "learning_rate": 2.2000000000000003e-05, "loss": 0.001, "step": 110 }, { "epoch": 3.6363636363636362, "grad_norm": 0.03260282799601555, "learning_rate": 2.4e-05, "loss": 0.0011, "step": 120 }, { "epoch": 3.9393939393939394, "grad_norm": 0.020618194714188576, "learning_rate": 2.6000000000000002e-05, "loss": 0.0012, "step": 130 }, { "epoch": 4.242424242424242, "grad_norm": 0.021722646430134773, "learning_rate": 2.8000000000000003e-05, "loss": 0.0009, "step": 140 }, { "epoch": 4.545454545454545, "grad_norm": 0.015248863026499748, "learning_rate": 3e-05, "loss": 0.0007, "step": 150 }, { "epoch": 4.848484848484849, "grad_norm": 0.00668317498639226, "learning_rate": 3.2000000000000005e-05, "loss": 0.0007, "step": 160 }, { "epoch": 5.151515151515151, "grad_norm": 0.009767434559762478, "learning_rate": 3.4000000000000007e-05, "loss": 0.0006, "step": 170 }, { "epoch": 5.454545454545454, "grad_norm": 0.0070835500955581665, "learning_rate": 3.6e-05, "loss": 0.0005, "step": 180 }, { "epoch": 5.757575757575758, "grad_norm": 0.005514961667358875, "learning_rate": 3.8e-05, "loss": 0.0005, "step": 190 }, { "epoch": 6.0606060606060606, "grad_norm": 0.007456360850483179, "learning_rate": 4e-05, "loss": 0.0005, "step": 200 }, { "epoch": 6.0606060606060606, "eval_loss": 0.0019312179647386074, "eval_runtime": 13.4333, "eval_samples_per_second": 9.677, "eval_steps_per_second": 0.67, "step": 200 }, { "epoch": 6.363636363636363, "grad_norm": 0.0049119917675852776, "learning_rate": 4.2e-05, "loss": 0.0004, "step": 210 }, { "epoch": 6.666666666666667, "grad_norm": 0.0046443757601082325, "learning_rate": 4.4000000000000006e-05, "loss": 0.0004, "step": 220 }, { "epoch": 6.96969696969697, "grad_norm": 0.006061806809157133, "learning_rate": 4.600000000000001e-05, "loss": 0.0004, "step": 230 }, { "epoch": 7.2727272727272725, "grad_norm": 0.004031210206449032, "learning_rate": 4.8e-05, "loss": 0.0004, "step": 240 }, { "epoch": 7.575757575757576, "grad_norm": 0.004312009084969759, "learning_rate": 5e-05, "loss": 0.0003, "step": 250 }, { "epoch": 7.878787878787879, "grad_norm": 0.004320390522480011, "learning_rate": 5.2000000000000004e-05, "loss": 0.0003, "step": 260 }, { "epoch": 8.181818181818182, "grad_norm": 0.0036691236309707165, "learning_rate": 5.4000000000000005e-05, "loss": 0.0003, "step": 270 }, { "epoch": 8.484848484848484, "grad_norm": 0.0040830825455486774, "learning_rate": 5.6000000000000006e-05, "loss": 0.0003, "step": 280 }, { "epoch": 8.787878787878787, "grad_norm": 0.003878690768033266, "learning_rate": 5.8e-05, "loss": 0.0002, "step": 290 } ], "logging_steps": 10, "max_steps": 297, "num_input_tokens_seen": 0, "num_train_epochs": 9, "save_steps": 100, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 3.62740521701376e+17, "train_batch_size": 16, "trial_name": null, "trial_params": null }