| { |
| "best_metric": 0.9203949718676169, |
| "best_model_checkpoint": "./results/checkpoint-3370", |
| "epoch": 7.0, |
| "eval_steps": 500, |
| "global_step": 4718, |
| "is_hyper_param_search": false, |
| "is_local_process_zero": true, |
| "is_world_process_zero": true, |
| "log_history": [ |
| { |
| "epoch": 0.7423904974016332, |
| "grad_norm": 1.0851198434829712, |
| "learning_rate": 2.8606983655274892e-05, |
| "loss": 0.448, |
| "step": 500 |
| }, |
| { |
| "epoch": 1.0, |
| "eval_administration_accuracy": 0.8700092850510678, |
| "eval_administration_f1": 0.830681542112301, |
| "eval_corruption_accuracy": 0.9301764159702879, |
| "eval_corruption_f1": 0.9210203978746075, |
| "eval_democracy_accuracy": 0.9201485608170845, |
| "eval_democracy_f1": 0.8972549864645475, |
| "eval_development_accuracy": 0.8672237697307336, |
| "eval_development_f1": 0.8462048329953127, |
| "eval_economy_accuracy": 0.9062209842154132, |
| "eval_economy_f1": 0.8952254345075943, |
| "eval_education_accuracy": 0.9467038068709378, |
| "eval_education_f1": 0.9344408657756663, |
| "eval_environment_accuracy": 0.9619312906220984, |
| "eval_environment_f1": 0.9564784032426319, |
| "eval_instability_accuracy": 0.9084493964716805, |
| "eval_instability_f1": 0.8872576883139862, |
| "eval_leadership_accuracy": 0.833983286908078, |
| "eval_leadership_f1": 0.8199437016930928, |
| "eval_loss": 0.2781185507774353, |
| "eval_overall_accuracy": 0.9090374497059734, |
| "eval_overall_f1": 0.8943279031006514, |
| "eval_race_accuracy": 0.9385329619312907, |
| "eval_race_f1": 0.9345823913548884, |
| "eval_religion_accuracy": 0.9303621169916435, |
| "eval_religion_f1": 0.9219388519537324, |
| "eval_runtime": 6.3896, |
| "eval_safety_accuracy": 0.8947075208913648, |
| "eval_safety_f1": 0.8869057409194557, |
| "eval_samples_per_second": 842.781, |
| "eval_steps_per_second": 52.742, |
| "step": 674 |
| }, |
| { |
| "epoch": 1.4840386043058649, |
| "grad_norm": 1.3921394348144531, |
| "learning_rate": 2.7213967310549777e-05, |
| "loss": 0.2646, |
| "step": 1000 |
| }, |
| { |
| "epoch": 2.0, |
| "eval_administration_accuracy": 0.886908077994429, |
| "eval_administration_f1": 0.8723732992505114, |
| "eval_corruption_accuracy": 0.9424326833797586, |
| "eval_corruption_f1": 0.9384898864061544, |
| "eval_democracy_accuracy": 0.933519034354689, |
| "eval_democracy_f1": 0.9231869597067329, |
| "eval_development_accuracy": 0.8739090064995357, |
| "eval_development_f1": 0.8668624356641937, |
| "eval_economy_accuracy": 0.9143918291550603, |
| "eval_economy_f1": 0.9111034488927022, |
| "eval_education_accuracy": 0.9580315691736304, |
| "eval_education_f1": 0.9543147194312244, |
| "eval_environment_accuracy": 0.9734447539461467, |
| "eval_environment_f1": 0.9731508098039792, |
| "eval_instability_accuracy": 0.9277623026926648, |
| "eval_instability_f1": 0.9222378035229967, |
| "eval_leadership_accuracy": 0.840297121634169, |
| "eval_leadership_f1": 0.8406461991345243, |
| "eval_loss": 0.23724055290222168, |
| "eval_overall_accuracy": 0.9205354379449089, |
| "eval_overall_f1": 0.9156647148374422, |
| "eval_race_accuracy": 0.9467038068709378, |
| "eval_race_f1": 0.943797657082642, |
| "eval_religion_accuracy": 0.9409470752089136, |
| "eval_religion_f1": 0.9380295054445577, |
| "eval_runtime": 6.3736, |
| "eval_safety_accuracy": 0.9080779944289693, |
| "eval_safety_f1": 0.9037838537090879, |
| "eval_samples_per_second": 844.886, |
| "eval_steps_per_second": 52.874, |
| "step": 1348 |
| }, |
| { |
| "epoch": 2.2256867112100966, |
| "grad_norm": 0.9178161025047302, |
| "learning_rate": 2.582095096582467e-05, |
| "loss": 0.2085, |
| "step": 1500 |
| }, |
| { |
| "epoch": 2.9680772086117297, |
| "grad_norm": 1.1747676134109497, |
| "learning_rate": 2.4427934621099553e-05, |
| "loss": 0.1696, |
| "step": 2000 |
| }, |
| { |
| "epoch": 3.0, |
| "eval_administration_accuracy": 0.8919220055710306, |
| "eval_administration_f1": 0.8805907698259323, |
| "eval_corruption_accuracy": 0.9454038997214484, |
| "eval_corruption_f1": 0.9422755927443824, |
| "eval_democracy_accuracy": 0.9333333333333333, |
| "eval_democracy_f1": 0.9277380787973176, |
| "eval_development_accuracy": 0.8846796657381616, |
| "eval_development_f1": 0.8767801428288444, |
| "eval_economy_accuracy": 0.9177344475394614, |
| "eval_economy_f1": 0.913190124488594, |
| "eval_education_accuracy": 0.959702878365831, |
| "eval_education_f1": 0.9592018840910198, |
| "eval_environment_accuracy": 0.9766016713091922, |
| "eval_environment_f1": 0.9753052667455842, |
| "eval_instability_accuracy": 0.9255338904363974, |
| "eval_instability_f1": 0.9219441146038561, |
| "eval_leadership_accuracy": 0.8503249767873723, |
| "eval_leadership_f1": 0.8464660955301658, |
| "eval_loss": 0.22912514209747314, |
| "eval_overall_accuracy": 0.923785205818632, |
| "eval_overall_f1": 0.9198927823234619, |
| "eval_race_accuracy": 0.9468895078922934, |
| "eval_race_f1": 0.9441406664359246, |
| "eval_religion_accuracy": 0.9418755803156917, |
| "eval_religion_f1": 0.9406664102976392, |
| "eval_runtime": 6.3983, |
| "eval_safety_accuracy": 0.9114206128133705, |
| "eval_safety_f1": 0.9104142414922826, |
| "eval_samples_per_second": 841.634, |
| "eval_steps_per_second": 52.67, |
| "step": 2022 |
| }, |
| { |
| "epoch": 3.7097253155159615, |
| "grad_norm": 1.7257879972457886, |
| "learning_rate": 2.3034918276374445e-05, |
| "loss": 0.1309, |
| "step": 2500 |
| }, |
| { |
| "epoch": 4.0, |
| "eval_administration_accuracy": 0.875766016713092, |
| "eval_administration_f1": 0.8761692862915822, |
| "eval_corruption_accuracy": 0.9465181058495822, |
| "eval_corruption_f1": 0.9432932911595113, |
| "eval_democracy_accuracy": 0.9344475394614671, |
| "eval_democracy_f1": 0.9289539940724922, |
| "eval_development_accuracy": 0.8804085422469824, |
| "eval_development_f1": 0.8733152085361343, |
| "eval_economy_accuracy": 0.9175487465181058, |
| "eval_economy_f1": 0.9168308054942083, |
| "eval_education_accuracy": 0.959702878365831, |
| "eval_education_f1": 0.9577025286068839, |
| "eval_environment_accuracy": 0.9740018570102136, |
| "eval_environment_f1": 0.9740155750677182, |
| "eval_instability_accuracy": 0.92330547818013, |
| "eval_instability_f1": 0.9214961270553598, |
| "eval_leadership_accuracy": 0.8469823584029712, |
| "eval_leadership_f1": 0.8454208641557382, |
| "eval_loss": 0.23740428686141968, |
| "eval_overall_accuracy": 0.9211853915196534, |
| "eval_overall_f1": 0.9191929747362485, |
| "eval_race_accuracy": 0.9452181987000928, |
| "eval_race_f1": 0.9441252497587079, |
| "eval_religion_accuracy": 0.940761374187558, |
| "eval_religion_f1": 0.9389083250935241, |
| "eval_runtime": 6.3816, |
| "eval_safety_accuracy": 0.9095636025998143, |
| "eval_safety_f1": 0.9100844415431236, |
| "eval_samples_per_second": 843.83, |
| "eval_steps_per_second": 52.808, |
| "step": 2696 |
| }, |
| { |
| "epoch": 4.451373422420193, |
| "grad_norm": 1.5328147411346436, |
| "learning_rate": 2.1641901931649333e-05, |
| "loss": 0.1085, |
| "step": 3000 |
| }, |
| { |
| "epoch": 5.0, |
| "eval_administration_accuracy": 0.8921077065923863, |
| "eval_administration_f1": 0.8861580465319954, |
| "eval_corruption_accuracy": 0.9463324048282266, |
| "eval_corruption_f1": 0.943458379396768, |
| "eval_democracy_accuracy": 0.9346332404828227, |
| "eval_democracy_f1": 0.9313693520601084, |
| "eval_development_accuracy": 0.8807799442896936, |
| "eval_development_f1": 0.8746742414193872, |
| "eval_economy_accuracy": 0.9175487465181058, |
| "eval_economy_f1": 0.9165794145446547, |
| "eval_education_accuracy": 0.9587743732590529, |
| "eval_education_f1": 0.9574397967431261, |
| "eval_environment_accuracy": 0.9736304549675023, |
| "eval_environment_f1": 0.9731615305259281, |
| "eval_instability_accuracy": 0.9257195914577531, |
| "eval_instability_f1": 0.9217833078621397, |
| "eval_leadership_accuracy": 0.8458681522748375, |
| "eval_leadership_f1": 0.849217017756863, |
| "eval_loss": 0.24142640829086304, |
| "eval_overall_accuracy": 0.9226245744351593, |
| "eval_overall_f1": 0.9203949718676169, |
| "eval_race_accuracy": 0.9452181987000928, |
| "eval_race_f1": 0.9418839421000289, |
| "eval_religion_accuracy": 0.9426183844011142, |
| "eval_religion_f1": 0.9420269885121756, |
| "eval_runtime": 6.3895, |
| "eval_safety_accuracy": 0.9082636954503249, |
| "eval_safety_f1": 0.9069876449582265, |
| "eval_samples_per_second": 842.785, |
| "eval_steps_per_second": 52.743, |
| "step": 3370 |
| }, |
| { |
| "epoch": 5.193021529324425, |
| "grad_norm": 1.2458630800247192, |
| "learning_rate": 2.024888558692422e-05, |
| "loss": 0.0925, |
| "step": 3500 |
| }, |
| { |
| "epoch": 5.935412026726058, |
| "grad_norm": 1.964983582496643, |
| "learning_rate": 1.885586924219911e-05, |
| "loss": 0.0759, |
| "step": 4000 |
| }, |
| { |
| "epoch": 6.0, |
| "eval_administration_accuracy": 0.8885793871866295, |
| "eval_administration_f1": 0.8796540295327503, |
| "eval_corruption_accuracy": 0.944661095636026, |
| "eval_corruption_f1": 0.9430977714288632, |
| "eval_democracy_accuracy": 0.9312906220984215, |
| "eval_democracy_f1": 0.9311108081348881, |
| "eval_development_accuracy": 0.8809656453110493, |
| "eval_development_f1": 0.8743050043114865, |
| "eval_economy_accuracy": 0.9162488393686166, |
| "eval_economy_f1": 0.9153006920861309, |
| "eval_education_accuracy": 0.9600742804085423, |
| "eval_education_f1": 0.9587642973646463, |
| "eval_environment_accuracy": 0.9756731662024141, |
| "eval_environment_f1": 0.9753440362005743, |
| "eval_instability_accuracy": 0.9205199628597958, |
| "eval_instability_f1": 0.9185046153727873, |
| "eval_leadership_accuracy": 0.8510677808727948, |
| "eval_leadership_f1": 0.8491756854928866, |
| "eval_loss": 0.25557437539100647, |
| "eval_overall_accuracy": 0.9218198700092851, |
| "eval_overall_f1": 0.9198843837682896, |
| "eval_race_accuracy": 0.9472609099350047, |
| "eval_race_f1": 0.9465463984676691, |
| "eval_religion_accuracy": 0.9420612813370474, |
| "eval_religion_f1": 0.941883774169552, |
| "eval_runtime": 6.406, |
| "eval_safety_accuracy": 0.903435468895079, |
| "eval_safety_f1": 0.9049254926572401, |
| "eval_samples_per_second": 840.618, |
| "eval_steps_per_second": 52.607, |
| "step": 4044 |
| }, |
| { |
| "epoch": 6.67706013363029, |
| "grad_norm": 1.1733307838439941, |
| "learning_rate": 1.7462852897474e-05, |
| "loss": 0.0618, |
| "step": 4500 |
| }, |
| { |
| "epoch": 7.0, |
| "eval_administration_accuracy": 0.8891364902506964, |
| "eval_administration_f1": 0.8830637895995853, |
| "eval_corruption_accuracy": 0.9441039925719591, |
| "eval_corruption_f1": 0.9419800967289127, |
| "eval_democracy_accuracy": 0.9318477251624884, |
| "eval_democracy_f1": 0.9312436052041511, |
| "eval_development_accuracy": 0.8748375116063138, |
| "eval_development_f1": 0.8709535653934616, |
| "eval_economy_accuracy": 0.9151346332404828, |
| "eval_economy_f1": 0.9143035893385101, |
| "eval_education_accuracy": 0.9567316620241412, |
| "eval_education_f1": 0.956532871707756, |
| "eval_environment_accuracy": 0.9745589600742804, |
| "eval_environment_f1": 0.9745453763641572, |
| "eval_instability_accuracy": 0.9197771587743733, |
| "eval_instability_f1": 0.9164183053432433, |
| "eval_leadership_accuracy": 0.8493964716805943, |
| "eval_leadership_f1": 0.848764162637075, |
| "eval_loss": 0.2655349373817444, |
| "eval_overall_accuracy": 0.9205663881151346, |
| "eval_overall_f1": 0.9191255338624242, |
| "eval_race_accuracy": 0.9455896007428041, |
| "eval_race_f1": 0.9449102601982086, |
| "eval_religion_accuracy": 0.9424326833797586, |
| "eval_religion_f1": 0.9425971967753438, |
| "eval_runtime": 6.378, |
| "eval_safety_accuracy": 0.9032497678737234, |
| "eval_safety_f1": 0.9041935870586874, |
| "eval_samples_per_second": 844.304, |
| "eval_steps_per_second": 52.838, |
| "step": 4718 |
| } |
| ], |
| "logging_steps": 500, |
| "max_steps": 10768, |
| "num_input_tokens_seen": 0, |
| "num_train_epochs": 16, |
| "save_steps": 500, |
| "stateful_callbacks": { |
| "EarlyStoppingCallback": { |
| "args": { |
| "early_stopping_patience": 2, |
| "early_stopping_threshold": 0.0 |
| }, |
| "attributes": { |
| "early_stopping_patience_counter": 2 |
| } |
| }, |
| "TrainerControl": { |
| "args": { |
| "should_epoch_stop": false, |
| "should_evaluate": false, |
| "should_log": false, |
| "should_save": true, |
| "should_training_stop": true |
| }, |
| "attributes": {} |
| } |
| }, |
| "total_flos": 3.968458495200461e+16, |
| "train_batch_size": 16, |
| "trial_name": null, |
| "trial_params": null |
| } |
|
|