| { |
| "best_metric": 0.43980079889297485, |
| "best_model_checkpoint": "./drive/Shareddrives/CS198-Drones/training_output/beit-base-patch16-224_rice-leaf-disease-augmented_tl_020125/checkpoint-3250", |
| "epoch": 13.0, |
| "eval_steps": 500, |
| "global_step": 3250, |
| "is_hyper_param_search": false, |
| "is_local_process_zero": true, |
| "is_world_process_zero": true, |
| "log_history": [ |
| { |
| "epoch": 1.0, |
| "grad_norm": 5.276273250579834, |
| "learning_rate": 6.666666666666667e-05, |
| "loss": 1.9619, |
| "step": 250 |
| }, |
| { |
| "epoch": 1.0, |
| "eval_accuracy": 0.496, |
| "eval_loss": 1.4686135053634644, |
| "eval_runtime": 90.1192, |
| "eval_samples_per_second": 22.193, |
| "eval_steps_per_second": 0.355, |
| "step": 250 |
| }, |
| { |
| "epoch": 2.0, |
| "grad_norm": 5.132676124572754, |
| "learning_rate": 9.62962962962963e-05, |
| "loss": 1.1487, |
| "step": 500 |
| }, |
| { |
| "epoch": 2.0, |
| "eval_accuracy": 0.716, |
| "eval_loss": 0.8857663869857788, |
| "eval_runtime": 86.7331, |
| "eval_samples_per_second": 23.059, |
| "eval_steps_per_second": 0.369, |
| "step": 500 |
| }, |
| { |
| "epoch": 3.0, |
| "grad_norm": 2.637998342514038, |
| "learning_rate": 8.888888888888889e-05, |
| "loss": 0.8166, |
| "step": 750 |
| }, |
| { |
| "epoch": 3.0, |
| "eval_accuracy": 0.7685, |
| "eval_loss": 0.7041126489639282, |
| "eval_runtime": 86.3454, |
| "eval_samples_per_second": 23.163, |
| "eval_steps_per_second": 0.371, |
| "step": 750 |
| }, |
| { |
| "epoch": 4.0, |
| "grad_norm": 3.51906418800354, |
| "learning_rate": 8.148148148148148e-05, |
| "loss": 0.691, |
| "step": 1000 |
| }, |
| { |
| "epoch": 4.0, |
| "eval_accuracy": 0.797, |
| "eval_loss": 0.6215672492980957, |
| "eval_runtime": 87.5308, |
| "eval_samples_per_second": 22.849, |
| "eval_steps_per_second": 0.366, |
| "step": 1000 |
| }, |
| { |
| "epoch": 5.0, |
| "grad_norm": 2.1984190940856934, |
| "learning_rate": 7.407407407407407e-05, |
| "loss": 0.6259, |
| "step": 1250 |
| }, |
| { |
| "epoch": 5.0, |
| "eval_accuracy": 0.824, |
| "eval_loss": 0.5643388032913208, |
| "eval_runtime": 87.0773, |
| "eval_samples_per_second": 22.968, |
| "eval_steps_per_second": 0.367, |
| "step": 1250 |
| }, |
| { |
| "epoch": 6.0, |
| "grad_norm": 3.236772060394287, |
| "learning_rate": 6.666666666666667e-05, |
| "loss": 0.582, |
| "step": 1500 |
| }, |
| { |
| "epoch": 6.0, |
| "eval_accuracy": 0.8305, |
| "eval_loss": 0.5285928845405579, |
| "eval_runtime": 84.9775, |
| "eval_samples_per_second": 23.536, |
| "eval_steps_per_second": 0.377, |
| "step": 1500 |
| }, |
| { |
| "epoch": 7.0, |
| "grad_norm": 2.1035590171813965, |
| "learning_rate": 5.925925925925926e-05, |
| "loss": 0.547, |
| "step": 1750 |
| }, |
| { |
| "epoch": 7.0, |
| "eval_accuracy": 0.836, |
| "eval_loss": 0.5041353702545166, |
| "eval_runtime": 87.6919, |
| "eval_samples_per_second": 22.807, |
| "eval_steps_per_second": 0.365, |
| "step": 1750 |
| }, |
| { |
| "epoch": 8.0, |
| "grad_norm": 2.690420150756836, |
| "learning_rate": 5.185185185185185e-05, |
| "loss": 0.5244, |
| "step": 2000 |
| }, |
| { |
| "epoch": 8.0, |
| "eval_accuracy": 0.844, |
| "eval_loss": 0.4841882288455963, |
| "eval_runtime": 89.4014, |
| "eval_samples_per_second": 22.371, |
| "eval_steps_per_second": 0.358, |
| "step": 2000 |
| }, |
| { |
| "epoch": 9.0, |
| "grad_norm": 3.189504623413086, |
| "learning_rate": 4.4444444444444447e-05, |
| "loss": 0.5081, |
| "step": 2250 |
| }, |
| { |
| "epoch": 9.0, |
| "eval_accuracy": 0.843, |
| "eval_loss": 0.4707326591014862, |
| "eval_runtime": 89.3049, |
| "eval_samples_per_second": 22.395, |
| "eval_steps_per_second": 0.358, |
| "step": 2250 |
| }, |
| { |
| "epoch": 10.0, |
| "grad_norm": 2.4485363960266113, |
| "learning_rate": 3.7037037037037037e-05, |
| "loss": 0.493, |
| "step": 2500 |
| }, |
| { |
| "epoch": 10.0, |
| "eval_accuracy": 0.8475, |
| "eval_loss": 0.4588555097579956, |
| "eval_runtime": 87.1428, |
| "eval_samples_per_second": 22.951, |
| "eval_steps_per_second": 0.367, |
| "step": 2500 |
| }, |
| { |
| "epoch": 11.0, |
| "grad_norm": 2.1555464267730713, |
| "learning_rate": 2.962962962962963e-05, |
| "loss": 0.4824, |
| "step": 2750 |
| }, |
| { |
| "epoch": 11.0, |
| "eval_accuracy": 0.85, |
| "eval_loss": 0.45093804597854614, |
| "eval_runtime": 86.5567, |
| "eval_samples_per_second": 23.106, |
| "eval_steps_per_second": 0.37, |
| "step": 2750 |
| }, |
| { |
| "epoch": 12.0, |
| "grad_norm": 2.3292956352233887, |
| "learning_rate": 2.2222222222222223e-05, |
| "loss": 0.4759, |
| "step": 3000 |
| }, |
| { |
| "epoch": 12.0, |
| "eval_accuracy": 0.853, |
| "eval_loss": 0.4436955153942108, |
| "eval_runtime": 86.9601, |
| "eval_samples_per_second": 22.999, |
| "eval_steps_per_second": 0.368, |
| "step": 3000 |
| }, |
| { |
| "epoch": 13.0, |
| "grad_norm": 2.5893406867980957, |
| "learning_rate": 1.4814814814814815e-05, |
| "loss": 0.4677, |
| "step": 3250 |
| }, |
| { |
| "epoch": 13.0, |
| "eval_accuracy": 0.855, |
| "eval_loss": 0.43980079889297485, |
| "eval_runtime": 87.1496, |
| "eval_samples_per_second": 22.949, |
| "eval_steps_per_second": 0.367, |
| "step": 3250 |
| } |
| ], |
| "logging_steps": 500, |
| "max_steps": 3750, |
| "num_input_tokens_seen": 0, |
| "num_train_epochs": 15, |
| "save_steps": 500, |
| "stateful_callbacks": { |
| "TrainerControl": { |
| "args": { |
| "should_epoch_stop": false, |
| "should_evaluate": false, |
| "should_log": false, |
| "should_save": true, |
| "should_training_stop": false |
| }, |
| "attributes": {} |
| } |
| }, |
| "total_flos": 1.6112311457808384e+19, |
| "train_batch_size": 64, |
| "trial_name": null, |
| "trial_params": null |
| } |
|
|