MyPoliBERT-ver03 / checkpoint-4718 /trainer_state.json
YagiASAFAS's picture
Add tokenizer files
e04e294
{
"best_metric": 0.9203949718676169,
"best_model_checkpoint": "./results/checkpoint-3370",
"epoch": 7.0,
"eval_steps": 500,
"global_step": 4718,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.7423904974016332,
"grad_norm": 1.0851198434829712,
"learning_rate": 2.8606983655274892e-05,
"loss": 0.448,
"step": 500
},
{
"epoch": 1.0,
"eval_administration_accuracy": 0.8700092850510678,
"eval_administration_f1": 0.830681542112301,
"eval_corruption_accuracy": 0.9301764159702879,
"eval_corruption_f1": 0.9210203978746075,
"eval_democracy_accuracy": 0.9201485608170845,
"eval_democracy_f1": 0.8972549864645475,
"eval_development_accuracy": 0.8672237697307336,
"eval_development_f1": 0.8462048329953127,
"eval_economy_accuracy": 0.9062209842154132,
"eval_economy_f1": 0.8952254345075943,
"eval_education_accuracy": 0.9467038068709378,
"eval_education_f1": 0.9344408657756663,
"eval_environment_accuracy": 0.9619312906220984,
"eval_environment_f1": 0.9564784032426319,
"eval_instability_accuracy": 0.9084493964716805,
"eval_instability_f1": 0.8872576883139862,
"eval_leadership_accuracy": 0.833983286908078,
"eval_leadership_f1": 0.8199437016930928,
"eval_loss": 0.2781185507774353,
"eval_overall_accuracy": 0.9090374497059734,
"eval_overall_f1": 0.8943279031006514,
"eval_race_accuracy": 0.9385329619312907,
"eval_race_f1": 0.9345823913548884,
"eval_religion_accuracy": 0.9303621169916435,
"eval_religion_f1": 0.9219388519537324,
"eval_runtime": 6.3896,
"eval_safety_accuracy": 0.8947075208913648,
"eval_safety_f1": 0.8869057409194557,
"eval_samples_per_second": 842.781,
"eval_steps_per_second": 52.742,
"step": 674
},
{
"epoch": 1.4840386043058649,
"grad_norm": 1.3921394348144531,
"learning_rate": 2.7213967310549777e-05,
"loss": 0.2646,
"step": 1000
},
{
"epoch": 2.0,
"eval_administration_accuracy": 0.886908077994429,
"eval_administration_f1": 0.8723732992505114,
"eval_corruption_accuracy": 0.9424326833797586,
"eval_corruption_f1": 0.9384898864061544,
"eval_democracy_accuracy": 0.933519034354689,
"eval_democracy_f1": 0.9231869597067329,
"eval_development_accuracy": 0.8739090064995357,
"eval_development_f1": 0.8668624356641937,
"eval_economy_accuracy": 0.9143918291550603,
"eval_economy_f1": 0.9111034488927022,
"eval_education_accuracy": 0.9580315691736304,
"eval_education_f1": 0.9543147194312244,
"eval_environment_accuracy": 0.9734447539461467,
"eval_environment_f1": 0.9731508098039792,
"eval_instability_accuracy": 0.9277623026926648,
"eval_instability_f1": 0.9222378035229967,
"eval_leadership_accuracy": 0.840297121634169,
"eval_leadership_f1": 0.8406461991345243,
"eval_loss": 0.23724055290222168,
"eval_overall_accuracy": 0.9205354379449089,
"eval_overall_f1": 0.9156647148374422,
"eval_race_accuracy": 0.9467038068709378,
"eval_race_f1": 0.943797657082642,
"eval_religion_accuracy": 0.9409470752089136,
"eval_religion_f1": 0.9380295054445577,
"eval_runtime": 6.3736,
"eval_safety_accuracy": 0.9080779944289693,
"eval_safety_f1": 0.9037838537090879,
"eval_samples_per_second": 844.886,
"eval_steps_per_second": 52.874,
"step": 1348
},
{
"epoch": 2.2256867112100966,
"grad_norm": 0.9178161025047302,
"learning_rate": 2.582095096582467e-05,
"loss": 0.2085,
"step": 1500
},
{
"epoch": 2.9680772086117297,
"grad_norm": 1.1747676134109497,
"learning_rate": 2.4427934621099553e-05,
"loss": 0.1696,
"step": 2000
},
{
"epoch": 3.0,
"eval_administration_accuracy": 0.8919220055710306,
"eval_administration_f1": 0.8805907698259323,
"eval_corruption_accuracy": 0.9454038997214484,
"eval_corruption_f1": 0.9422755927443824,
"eval_democracy_accuracy": 0.9333333333333333,
"eval_democracy_f1": 0.9277380787973176,
"eval_development_accuracy": 0.8846796657381616,
"eval_development_f1": 0.8767801428288444,
"eval_economy_accuracy": 0.9177344475394614,
"eval_economy_f1": 0.913190124488594,
"eval_education_accuracy": 0.959702878365831,
"eval_education_f1": 0.9592018840910198,
"eval_environment_accuracy": 0.9766016713091922,
"eval_environment_f1": 0.9753052667455842,
"eval_instability_accuracy": 0.9255338904363974,
"eval_instability_f1": 0.9219441146038561,
"eval_leadership_accuracy": 0.8503249767873723,
"eval_leadership_f1": 0.8464660955301658,
"eval_loss": 0.22912514209747314,
"eval_overall_accuracy": 0.923785205818632,
"eval_overall_f1": 0.9198927823234619,
"eval_race_accuracy": 0.9468895078922934,
"eval_race_f1": 0.9441406664359246,
"eval_religion_accuracy": 0.9418755803156917,
"eval_religion_f1": 0.9406664102976392,
"eval_runtime": 6.3983,
"eval_safety_accuracy": 0.9114206128133705,
"eval_safety_f1": 0.9104142414922826,
"eval_samples_per_second": 841.634,
"eval_steps_per_second": 52.67,
"step": 2022
},
{
"epoch": 3.7097253155159615,
"grad_norm": 1.7257879972457886,
"learning_rate": 2.3034918276374445e-05,
"loss": 0.1309,
"step": 2500
},
{
"epoch": 4.0,
"eval_administration_accuracy": 0.875766016713092,
"eval_administration_f1": 0.8761692862915822,
"eval_corruption_accuracy": 0.9465181058495822,
"eval_corruption_f1": 0.9432932911595113,
"eval_democracy_accuracy": 0.9344475394614671,
"eval_democracy_f1": 0.9289539940724922,
"eval_development_accuracy": 0.8804085422469824,
"eval_development_f1": 0.8733152085361343,
"eval_economy_accuracy": 0.9175487465181058,
"eval_economy_f1": 0.9168308054942083,
"eval_education_accuracy": 0.959702878365831,
"eval_education_f1": 0.9577025286068839,
"eval_environment_accuracy": 0.9740018570102136,
"eval_environment_f1": 0.9740155750677182,
"eval_instability_accuracy": 0.92330547818013,
"eval_instability_f1": 0.9214961270553598,
"eval_leadership_accuracy": 0.8469823584029712,
"eval_leadership_f1": 0.8454208641557382,
"eval_loss": 0.23740428686141968,
"eval_overall_accuracy": 0.9211853915196534,
"eval_overall_f1": 0.9191929747362485,
"eval_race_accuracy": 0.9452181987000928,
"eval_race_f1": 0.9441252497587079,
"eval_religion_accuracy": 0.940761374187558,
"eval_religion_f1": 0.9389083250935241,
"eval_runtime": 6.3816,
"eval_safety_accuracy": 0.9095636025998143,
"eval_safety_f1": 0.9100844415431236,
"eval_samples_per_second": 843.83,
"eval_steps_per_second": 52.808,
"step": 2696
},
{
"epoch": 4.451373422420193,
"grad_norm": 1.5328147411346436,
"learning_rate": 2.1641901931649333e-05,
"loss": 0.1085,
"step": 3000
},
{
"epoch": 5.0,
"eval_administration_accuracy": 0.8921077065923863,
"eval_administration_f1": 0.8861580465319954,
"eval_corruption_accuracy": 0.9463324048282266,
"eval_corruption_f1": 0.943458379396768,
"eval_democracy_accuracy": 0.9346332404828227,
"eval_democracy_f1": 0.9313693520601084,
"eval_development_accuracy": 0.8807799442896936,
"eval_development_f1": 0.8746742414193872,
"eval_economy_accuracy": 0.9175487465181058,
"eval_economy_f1": 0.9165794145446547,
"eval_education_accuracy": 0.9587743732590529,
"eval_education_f1": 0.9574397967431261,
"eval_environment_accuracy": 0.9736304549675023,
"eval_environment_f1": 0.9731615305259281,
"eval_instability_accuracy": 0.9257195914577531,
"eval_instability_f1": 0.9217833078621397,
"eval_leadership_accuracy": 0.8458681522748375,
"eval_leadership_f1": 0.849217017756863,
"eval_loss": 0.24142640829086304,
"eval_overall_accuracy": 0.9226245744351593,
"eval_overall_f1": 0.9203949718676169,
"eval_race_accuracy": 0.9452181987000928,
"eval_race_f1": 0.9418839421000289,
"eval_religion_accuracy": 0.9426183844011142,
"eval_religion_f1": 0.9420269885121756,
"eval_runtime": 6.3895,
"eval_safety_accuracy": 0.9082636954503249,
"eval_safety_f1": 0.9069876449582265,
"eval_samples_per_second": 842.785,
"eval_steps_per_second": 52.743,
"step": 3370
},
{
"epoch": 5.193021529324425,
"grad_norm": 1.2458630800247192,
"learning_rate": 2.024888558692422e-05,
"loss": 0.0925,
"step": 3500
},
{
"epoch": 5.935412026726058,
"grad_norm": 1.964983582496643,
"learning_rate": 1.885586924219911e-05,
"loss": 0.0759,
"step": 4000
},
{
"epoch": 6.0,
"eval_administration_accuracy": 0.8885793871866295,
"eval_administration_f1": 0.8796540295327503,
"eval_corruption_accuracy": 0.944661095636026,
"eval_corruption_f1": 0.9430977714288632,
"eval_democracy_accuracy": 0.9312906220984215,
"eval_democracy_f1": 0.9311108081348881,
"eval_development_accuracy": 0.8809656453110493,
"eval_development_f1": 0.8743050043114865,
"eval_economy_accuracy": 0.9162488393686166,
"eval_economy_f1": 0.9153006920861309,
"eval_education_accuracy": 0.9600742804085423,
"eval_education_f1": 0.9587642973646463,
"eval_environment_accuracy": 0.9756731662024141,
"eval_environment_f1": 0.9753440362005743,
"eval_instability_accuracy": 0.9205199628597958,
"eval_instability_f1": 0.9185046153727873,
"eval_leadership_accuracy": 0.8510677808727948,
"eval_leadership_f1": 0.8491756854928866,
"eval_loss": 0.25557437539100647,
"eval_overall_accuracy": 0.9218198700092851,
"eval_overall_f1": 0.9198843837682896,
"eval_race_accuracy": 0.9472609099350047,
"eval_race_f1": 0.9465463984676691,
"eval_religion_accuracy": 0.9420612813370474,
"eval_religion_f1": 0.941883774169552,
"eval_runtime": 6.406,
"eval_safety_accuracy": 0.903435468895079,
"eval_safety_f1": 0.9049254926572401,
"eval_samples_per_second": 840.618,
"eval_steps_per_second": 52.607,
"step": 4044
},
{
"epoch": 6.67706013363029,
"grad_norm": 1.1733307838439941,
"learning_rate": 1.7462852897474e-05,
"loss": 0.0618,
"step": 4500
},
{
"epoch": 7.0,
"eval_administration_accuracy": 0.8891364902506964,
"eval_administration_f1": 0.8830637895995853,
"eval_corruption_accuracy": 0.9441039925719591,
"eval_corruption_f1": 0.9419800967289127,
"eval_democracy_accuracy": 0.9318477251624884,
"eval_democracy_f1": 0.9312436052041511,
"eval_development_accuracy": 0.8748375116063138,
"eval_development_f1": 0.8709535653934616,
"eval_economy_accuracy": 0.9151346332404828,
"eval_economy_f1": 0.9143035893385101,
"eval_education_accuracy": 0.9567316620241412,
"eval_education_f1": 0.956532871707756,
"eval_environment_accuracy": 0.9745589600742804,
"eval_environment_f1": 0.9745453763641572,
"eval_instability_accuracy": 0.9197771587743733,
"eval_instability_f1": 0.9164183053432433,
"eval_leadership_accuracy": 0.8493964716805943,
"eval_leadership_f1": 0.848764162637075,
"eval_loss": 0.2655349373817444,
"eval_overall_accuracy": 0.9205663881151346,
"eval_overall_f1": 0.9191255338624242,
"eval_race_accuracy": 0.9455896007428041,
"eval_race_f1": 0.9449102601982086,
"eval_religion_accuracy": 0.9424326833797586,
"eval_religion_f1": 0.9425971967753438,
"eval_runtime": 6.378,
"eval_safety_accuracy": 0.9032497678737234,
"eval_safety_f1": 0.9041935870586874,
"eval_samples_per_second": 844.304,
"eval_steps_per_second": 52.838,
"step": 4718
}
],
"logging_steps": 500,
"max_steps": 10768,
"num_input_tokens_seen": 0,
"num_train_epochs": 16,
"save_steps": 500,
"stateful_callbacks": {
"EarlyStoppingCallback": {
"args": {
"early_stopping_patience": 2,
"early_stopping_threshold": 0.0
},
"attributes": {
"early_stopping_patience_counter": 2
}
},
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 3.968458495200461e+16,
"train_batch_size": 16,
"trial_name": null,
"trial_params": null
}