| { |
| "best_metric": 0.5048543689320387, |
| "best_model_checkpoint": "models/jorgeortizfuentes-patana-bert_lr-3.027890490221129e-05_lr_scheduler_type-linear_weight_decay-0_optim-adafactor_aug-2/checkpoint-430", |
| "epoch": 4.0, |
| "global_step": 1720, |
| "is_hyper_param_search": false, |
| "is_local_process_zero": true, |
| "is_world_process_zero": true, |
| "log_history": [ |
| { |
| "epoch": 1.0, |
| "eval_f1": 0.5048543689320387, |
| "eval_loss": 0.10201934725046158, |
| "eval_runtime": 6.9759, |
| "eval_samples_per_second": 229.361, |
| "eval_steps_per_second": 3.584, |
| "step": 430 |
| }, |
| { |
| "epoch": 1.16, |
| "learning_rate": 2.6366901684871328e-05, |
| "loss": 0.3341, |
| "step": 500 |
| }, |
| { |
| "epoch": 2.0, |
| "eval_f1": 0.44047619047619047, |
| "eval_loss": 0.14018036425113678, |
| "eval_runtime": 6.9723, |
| "eval_samples_per_second": 229.479, |
| "eval_steps_per_second": 3.586, |
| "step": 860 |
| }, |
| { |
| "epoch": 2.33, |
| "learning_rate": 2.245489846753137e-05, |
| "loss": 0.1899, |
| "step": 1000 |
| }, |
| { |
| "epoch": 3.0, |
| "eval_f1": 0.326241134751773, |
| "eval_loss": 0.31355568766593933, |
| "eval_runtime": 6.9727, |
| "eval_samples_per_second": 229.468, |
| "eval_steps_per_second": 3.585, |
| "step": 1290 |
| }, |
| { |
| "epoch": 3.49, |
| "learning_rate": 1.854289525019141e-05, |
| "loss": 0.1032, |
| "step": 1500 |
| }, |
| { |
| "epoch": 4.0, |
| "eval_f1": 0.39647577092511016, |
| "eval_loss": 0.2821248769760132, |
| "eval_runtime": 6.9682, |
| "eval_samples_per_second": 229.613, |
| "eval_steps_per_second": 3.588, |
| "step": 1720 |
| }, |
| { |
| "epoch": 4.0, |
| "step": 1720, |
| "total_flos": 2.890222320918528e+16, |
| "train_loss": 0.1912146679190702, |
| "train_runtime": 1340.9092, |
| "train_samples_per_second": 184.321, |
| "train_steps_per_second": 2.886 |
| }, |
| { |
| "epoch": 4.0, |
| "eval_f1": 0.5048543689320387, |
| "eval_loss": 0.10201934725046158, |
| "eval_runtime": 6.9602, |
| "eval_samples_per_second": 229.88, |
| "eval_steps_per_second": 3.592, |
| "step": 1720 |
| } |
| ], |
| "max_steps": 3870, |
| "num_train_epochs": 9, |
| "total_flos": 2.890222320918528e+16, |
| "trial_name": null, |
| "trial_params": null |
| } |
|
|