| { |
| "best_global_step": 250, |
| "best_metric": 0.8147702755950403, |
| "best_model_checkpoint": "./results/run-2/checkpoint-250", |
| "epoch": 1.0, |
| "eval_steps": 500, |
| "global_step": 250, |
| "is_hyper_param_search": true, |
| "is_local_process_zero": true, |
| "is_world_process_zero": true, |
| "log_history": [ |
| { |
| "epoch": 0.4, |
| "grad_norm": 1.9651439189910889, |
| "learning_rate": 1.386e-05, |
| "loss": 1.62, |
| "step": 100 |
| }, |
| { |
| "epoch": 0.8, |
| "grad_norm": 1.7124375104904175, |
| "learning_rate": 2.7859999999999998e-05, |
| "loss": 0.6715, |
| "step": 200 |
| }, |
| { |
| "epoch": 1.0, |
| "eval_clientelism_accuracy": 0.906, |
| "eval_clientelism_f1": 0.897770957127279, |
| "eval_discipline_among_poor_accuracy": 0.968, |
| "eval_discipline_among_poor_f1": 0.9615640823822642, |
| "eval_economic_policy_accuracy": 0.875, |
| "eval_economic_policy_f1": 0.860741831908876, |
| "eval_loss": 0.47158536314964294, |
| "eval_marcos_duterte_alliance_accuracy": 0.846, |
| "eval_marcos_duterte_alliance_f1": 0.839269429776376, |
| "eval_overall_accuracy": 0.8419375, |
| "eval_overall_f1": 0.8147702755950403, |
| "eval_populism_accuracy": 0.645, |
| "eval_populism_f1": 0.5639441230465639, |
| "eval_regionalism_accuracy": 0.963, |
| "eval_regionalism_f1": 0.9482946255564284, |
| "eval_runtime": 2.5786, |
| "eval_samples_per_second": 775.6, |
| "eval_security_accuracy": 0.8275, |
| "eval_security_f1": 0.8006718398296887, |
| "eval_steps_per_second": 48.475, |
| "eval_uniteam_positive_campaign_accuracy": 0.705, |
| "eval_uniteam_positive_campaign_f1": 0.645905315132846, |
| "step": 250 |
| } |
| ], |
| "logging_steps": 100, |
| "max_steps": 1250, |
| "num_input_tokens_seen": 0, |
| "num_train_epochs": 5, |
| "save_steps": 500, |
| "stateful_callbacks": { |
| "EarlyStoppingCallback": { |
| "args": { |
| "early_stopping_patience": 2, |
| "early_stopping_threshold": 0.0 |
| }, |
| "attributes": { |
| "early_stopping_patience_counter": 0 |
| } |
| }, |
| "TrainerControl": { |
| "args": { |
| "should_epoch_stop": false, |
| "should_evaluate": false, |
| "should_log": false, |
| "should_save": true, |
| "should_training_stop": false |
| }, |
| "attributes": {} |
| } |
| }, |
| "total_flos": 1684364328960000.0, |
| "train_batch_size": 16, |
| "trial_name": null, |
| "trial_params": { |
| "gradient_accumulation_steps": 2, |
| "learning_rate": 7e-05, |
| "num_train_epochs": 5 |
| } |
| } |
|
|