{ "best_epoch": 6, "best_score": 0.8533304437182528, "best_selection_score": 0.8533304437182528, "best_val_metrics_at_selection": { "acc": 0.8983845752996352, "f1": 0.8418491484184915, "precision": 0.787556904400607, "recall": 0.9041811846689896 }, "model_selection": { "metric": "recall-floor-constrained-fbeta", "beta": 0.5, "recall_floor": 0.9, "recall_floor_penalty": 2.0, "acc_weight": 0.05 }, "data_provenance": { "train_csv": { "path": "data_for_deberta/Risk-only-pick/task1_risk_training_20260225.csv", "sha256_16": "149b65cfd4785835", "size_bytes": 61496472 }, "val_csv_source": { "path": "data_for_deberta/dedup_split/risk_val.csv", "sha256_16": "61c445d1558d1b34", "size_bytes": 6044780 }, "final_eval_csv": { "path": "data_for_deberta/dedup_split/risk_test.csv", "sha256_16": "7bb9f6e80cd447ae", "size_bytes": 9345127 } }, "train_rows": 17206, "val_rows": 1919, "final_eval_rows": 2879, "val_metrics": { "acc": 0.8983845752996352, "f1": 0.8418491484184915, "precision": 0.787556904400607, "recall": 0.9041811846689896 }, "final_eval_metrics": { "acc": 0.9051754081278222, "f1": 0.8455008488964346, "precision": 0.7830188679245284, "recall": 0.9188191881918819 }, "final_eval_baseline": { "always_pick_acc": 0.28238971865230983, "always_decline_acc": 0.7176102813476901 } }