{ "best_epoch": 2, "best_score": 0.8753029434576742, "best_selection_score": 0.8753029434576742, "best_val_metrics_at_selection": { "acc": 0.90563086548488, "f1": 0.8571428571428571, "precision": 0.812874251497006, "recall": 0.9065108514190318 }, "model_selection": { "metric": "recall-floor-constrained-fbeta", "beta": 0.5, "recall_floor": 0.9, "recall_floor_penalty": 2.0, "acc_weight": 0.05 }, "data_provenance": { "train_csv": { "path": "data_for_deberta/Risk-only-pick/task1_risk_training_20260225.csv", "sha256_16": "e8a3504a9b457cce", "size_bytes": 61519951 }, "val_csv_source": { "path": "data_for_deberta/dedup_split/risk_val.csv", "sha256_16": "5a878fcd1d1713be", "size_bytes": 6146696 }, "final_eval_csv": { "path": "data_for_deberta/dedup_split/risk_test.csv", "sha256_16": "1aa054c802b71ee8", "size_bytes": 9140917 } }, "train_rows": 17214, "val_rows": 1918, "final_eval_rows": 2879, "val_metrics": { "acc": 0.90563086548488, "f1": 0.8571428571428571, "precision": 0.812874251497006, "recall": 0.9065108514190318 }, "final_eval_metrics": { "acc": 0.8982285515804098, "f1": 0.8418780356179169, "precision": 0.8024691358024691, "recall": 0.8853575482406356 }, "final_eval_baseline": { "always_pick_acc": 0.3060090309135116, "always_decline_acc": 0.6939909690864884 } }