{ "best_epoch": 3, "best_score": 0.8008182905163272, "best_selection_score": 0.8008182905163272, "best_val_metrics_at_selection": { "acc": 0.8749348619072433, "f1": 0.8198198198198198, "precision": 0.7203166226912929, "recall": 0.9512195121951219 }, "model_selection": { "metric": "recall-floor-constrained-fbeta", "beta": 0.5, "recall_floor": 0.9, "recall_floor_penalty": 2.0, "acc_weight": 0.05 }, "data_provenance": { "train_csv": { "path": "data_for_deberta/Risk-only-pick/task1_risk_training_20260225.csv", "sha256_16": "149b65cfd4785835", "size_bytes": 61496472 }, "val_csv_source": { "path": "data_for_deberta/dedup_split/risk_val.csv", "sha256_16": "61c445d1558d1b34", "size_bytes": 6044780 }, "final_eval_csv": { "path": "data_for_deberta/dedup_split/risk_test.csv", "sha256_16": "7bb9f6e80cd447ae", "size_bytes": 9345127 } }, "train_rows": 17206, "val_rows": 1919, "final_eval_rows": 2879, "val_metrics": { "acc": 0.8749348619072433, "f1": 0.8198198198198198, "precision": 0.7203166226912929, "recall": 0.9512195121951219 }, "final_eval_metrics": { "acc": 0.8749565821465787, "f1": 0.8123044838373306, "precision": 0.7049773755656109, "recall": 0.958179581795818 }, "final_eval_baseline": { "always_pick_acc": 0.28238971865230983, "always_decline_acc": 0.7176102813476901 } }