PhilippinesPoliBERT / run-2 /checkpoint-250 /trainer_state.json
YagiASAFAS's picture
Training in progress, epoch 1
de7723e verified
{
"best_global_step": 250,
"best_metric": 0.8147702755950403,
"best_model_checkpoint": "./results/run-2/checkpoint-250",
"epoch": 1.0,
"eval_steps": 500,
"global_step": 250,
"is_hyper_param_search": true,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.4,
"grad_norm": 1.9651439189910889,
"learning_rate": 1.386e-05,
"loss": 1.62,
"step": 100
},
{
"epoch": 0.8,
"grad_norm": 1.7124375104904175,
"learning_rate": 2.7859999999999998e-05,
"loss": 0.6715,
"step": 200
},
{
"epoch": 1.0,
"eval_clientelism_accuracy": 0.906,
"eval_clientelism_f1": 0.897770957127279,
"eval_discipline_among_poor_accuracy": 0.968,
"eval_discipline_among_poor_f1": 0.9615640823822642,
"eval_economic_policy_accuracy": 0.875,
"eval_economic_policy_f1": 0.860741831908876,
"eval_loss": 0.47158536314964294,
"eval_marcos_duterte_alliance_accuracy": 0.846,
"eval_marcos_duterte_alliance_f1": 0.839269429776376,
"eval_overall_accuracy": 0.8419375,
"eval_overall_f1": 0.8147702755950403,
"eval_populism_accuracy": 0.645,
"eval_populism_f1": 0.5639441230465639,
"eval_regionalism_accuracy": 0.963,
"eval_regionalism_f1": 0.9482946255564284,
"eval_runtime": 2.5786,
"eval_samples_per_second": 775.6,
"eval_security_accuracy": 0.8275,
"eval_security_f1": 0.8006718398296887,
"eval_steps_per_second": 48.475,
"eval_uniteam_positive_campaign_accuracy": 0.705,
"eval_uniteam_positive_campaign_f1": 0.645905315132846,
"step": 250
}
],
"logging_steps": 100,
"max_steps": 1250,
"num_input_tokens_seen": 0,
"num_train_epochs": 5,
"save_steps": 500,
"stateful_callbacks": {
"EarlyStoppingCallback": {
"args": {
"early_stopping_patience": 2,
"early_stopping_threshold": 0.0
},
"attributes": {
"early_stopping_patience_counter": 0
}
},
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": false
},
"attributes": {}
}
},
"total_flos": 1684364328960000.0,
"train_batch_size": 16,
"trial_name": null,
"trial_params": {
"gradient_accumulation_steps": 2,
"learning_rate": 7e-05,
"num_train_epochs": 5
}
}