| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| |
|
| | import argparse, json, os, re, time, glob |
| | from pathlib import Path |
| | from collections import Counter |
| |
|
| | import numpy as np |
| | import pandas as pd |
| |
|
| | from sklearn.model_selection import StratifiedGroupKFold |
| | from sklearn.metrics import ( |
| | f1_score, confusion_matrix, accuracy_score, precision_score, recall_score, |
| | fbeta_score, roc_auc_score, average_precision_score, precision_recall_curve |
| | ) |
| | from sklearn.preprocessing import StandardScaler |
| | from sklearn.pipeline import Pipeline |
| | from sklearn.compose import ColumnTransformer |
| | from sklearn.ensemble import RandomForestClassifier, ExtraTreesClassifier |
| | from sklearn.impute import SimpleImputer |
| | from sklearn.calibration import CalibratedClassifierCV |
| | from sklearn.base import BaseEstimator, ClassifierMixin, clone |
| |
|
| | import joblib |
| |
|
| | try: |
| | from xgboost import XGBClassifier |
| | _HAS_XGB = True |
| | except Exception: |
| | _HAS_XGB = False |
| |
|
| | import optuna |
| | from optuna.pruners import MedianPruner |
| |
|
| | ANTICODONS64 = [a+b+c for a in "ACGT" for b in "ACGT" for c in "ACGT"] |
| | TETRA_KEYS = [a+b+c+d for a in "ACGT" for b in "ACGT" for c in "ACGT" for d in "ACGT"] |
| |
|
| | AA_SET = set(list("ACDEFGHIKLMNPQRSTVWY")) |
| | STAR = "*" |
| | QMARK = "?" |
| |
|
| | |
| | |
| | |
| | def fit_pipeline(model, X, y, sample_weight=None): |
| | """ |
| | Fit helper that correctly routes sample_weight into Pipeline |
| | (to the 'clf' step) and also supports plain estimators. |
| | """ |
| | if sample_weight is None: |
| | return model.fit(X, y) |
| |
|
| | sw = np.asarray(sample_weight) |
| |
|
| | |
| | if isinstance(model, Pipeline): |
| | return model.fit(X, y, clf__sample_weight=sw) |
| |
|
| | |
| | try: |
| | return model.fit(X, y, sample_weight=sw) |
| | except TypeError: |
| | return model.fit(X, y) |
| |
|
| | |
| | |
| | |
| | class PUBaggingClassifier(BaseEstimator, ClassifierMixin): |
| | """ |
| | PU-learning via bagging: |
| | - y==1: positives (P) [assemblies from Supp2] |
| | - y==0: unlabeled (U) [everything else] |
| | For each bag: |
| | - train on all P + sampled subset of U as pseudo-negative |
| | predict_proba: average over bags. |
| | |
| | sklearn calibration fix: |
| | - _estimator_type='classifier' |
| | - get tags / is_classifier compatibility |
| | """ |
| | _estimator_type = "classifier" |
| |
|
| | def __init__(self, base_estimator, n_bags=15, u_ratio=3.0, random_state=42): |
| | self.base_estimator = base_estimator |
| | self.n_bags = int(n_bags) |
| | self.u_ratio = float(u_ratio) |
| | self.random_state = int(random_state) |
| | self.models_ = None |
| | self.classes_ = np.array([0, 1], dtype=int) |
| |
|
| | def _more_tags(self): |
| | |
| | return {"requires_y": True, "binary_only": True} |
| |
|
| | def fit(self, X, y, sample_weight=None): |
| | y = np.asarray(y).astype(int) |
| | self.classes_ = np.array([0, 1], dtype=int) |
| |
|
| | pos_idx = np.where(y == 1)[0] |
| | unl_idx = np.where(y == 0)[0] |
| |
|
| | if pos_idx.size == 0: |
| | raise ValueError("PU training requires at least one positive sample (y==1).") |
| |
|
| | rng = np.random.RandomState(self.random_state) |
| | self.models_ = [] |
| |
|
| | |
| | if unl_idx.size == 0: |
| | m = clone(self.base_estimator) |
| | fit_pipeline(m, X, y, sample_weight) |
| | self.models_.append(m) |
| | return self |
| |
|
| | k_u = int(min(unl_idx.size, max(1, round(self.u_ratio * pos_idx.size)))) |
| |
|
| | for _ in range(self.n_bags): |
| | if k_u <= unl_idx.size: |
| | u_b = rng.choice(unl_idx, size=k_u, replace=False) |
| | else: |
| | u_b = rng.choice(unl_idx, size=k_u, replace=True) |
| |
|
| | idx_b = np.concatenate([pos_idx, u_b]) |
| |
|
| | X_b = X.iloc[idx_b] if hasattr(X, "iloc") else X[idx_b] |
| | y_b = y[idx_b] |
| |
|
| | sw_b = None |
| | if sample_weight is not None: |
| | sw_b = np.asarray(sample_weight)[idx_b] |
| |
|
| | m = clone(self.base_estimator) |
| | fit_pipeline(m, X_b, y_b, sw_b) |
| | self.models_.append(m) |
| |
|
| | return self |
| |
|
| | def predict_proba(self, X): |
| | if not self.models_: |
| | raise RuntimeError("PUBaggingClassifier not fitted") |
| | probs = [m.predict_proba(X) for m in self.models_] |
| | return np.mean(np.stack(probs, axis=0), axis=0) |
| |
|
| | def predict(self, X): |
| | return (self.predict_proba(X)[:, 1] >= 0.5).astype(int) |
| |
|
| | |
| | |
| | |
| | def extract_acc_base(acc: str) -> str: |
| | m = re.match(r'^(G[CF]A_\d+)', str(acc)) |
| | return m.group(1) if m else str(acc).split('.')[0] |
| |
|
| | def load_alt_list_from_supp2(supp2_xlsx: str) -> set: |
| | df = pd.read_excel(supp2_xlsx, header=None, usecols=[3]) |
| | alt = df.iloc[:, 0].dropna().astype(str).unique().tolist() |
| | return set(extract_acc_base(x) for x in alt) |
| |
|
| | def label_from_supp2(meta: pd.DataFrame, supp2_xlsx: str) -> pd.Series: |
| | alt = load_alt_list_from_supp2(supp2_xlsx) |
| | return meta["acc_base"].map(lambda a: 1 if a in alt else 0).astype(int) |
| |
|
| | def parse_anticodon_from_label(label_tail: str) -> str: |
| | if "_" in label_tail: |
| | return label_tail.split("_", 1)[1].upper().replace("U","T") |
| | return label_tail.upper().replace("U","T") |
| |
|
| | def build_vec64_from_record(rec: dict) -> np.ndarray: |
| | counter = Counter() |
| | for t in rec.get("trna_type_per_acc", []): |
| | c = int(t.get("count", 0) or 0) |
| | tail = str(t.get("label", "")).split("_genome_")[-1] |
| | ac = parse_anticodon_from_label(tail) |
| | if len(ac) == 3 and set(ac) <= {"A","C","G","T"}: |
| | counter[ac] += c |
| | return np.array([counter.get(k, 0) for k in ANTICODONS64], dtype=float) |
| |
|
| | def build_feature_matrices(ndjson_path: str, include_gc_len_tetra: bool = True): |
| | rows = [] |
| | with open(ndjson_path, "r") as fh: |
| | for line in fh: |
| | line = line.strip() |
| | if not line: |
| | continue |
| | try: |
| | rows.append(json.loads(line)) |
| | except Exception: |
| | continue |
| |
|
| | feat_rows, meta_rows = [], [] |
| | for rec in rows: |
| | acc = rec.get("acc") |
| | if not acc: |
| | continue |
| | acc_base = extract_acc_base(acc) |
| | gc = rec.get("gc", {}) |
| | tetra = rec.get("tetra_norm", {}) |
| |
|
| | v64 = build_vec64_from_record(rec) |
| | vals = np.log1p(v64) |
| | cols = [f"ac_{k}" for k in ANTICODONS64] |
| | trna_dict = {c: float(v) for c, v in zip(cols, vals)} |
| |
|
| | extra = {} |
| | if include_gc_len_tetra: |
| | extra["gc_percent"] = float(gc.get("percent", np.nan)) |
| | extra["genome_length"] = int(gc.get("length", 0) or 0) |
| | for k in TETRA_KEYS: |
| | extra[f"tetra_{k}"] = float(tetra.get(k, np.nan)) |
| |
|
| | feat_rows.append({**trna_dict, **extra}) |
| | meta_rows.append({"acc": acc, "acc_base": acc_base}) |
| |
|
| | X = pd.DataFrame(feat_rows) |
| | meta = pd.DataFrame(meta_rows) |
| | return X, meta |
| |
|
| | def _count_lines(path: Path) -> int: |
| | try: |
| | return sum(1 for _ in open(path, 'r')) |
| | except Exception: |
| | return -1 |
| |
|
| | def _load_train_test(path_str: str, supp2_path: str): |
| | p = Path(path_str) |
| | if not p.is_dir(): |
| | raise FileNotFoundError(f"--ndjson must be a directory containing train.jsonl and test.jsonl: {p}") |
| | tr = p / "train.jsonl" |
| | te = p / "test.jsonl" |
| | if not tr.exists() or not te.exists(): |
| | raise FileNotFoundError(f"{p} must contain train.jsonl and test.jsonl") |
| |
|
| | print(f"[LOAD] train: {tr} (lines≈{_count_lines(tr)})") |
| | Xtr, mtr = build_feature_matrices(str(tr), True) |
| | print(f"[LOAD] test: {te} (lines≈{_count_lines(te)})") |
| | Xte, mte = build_feature_matrices(str(te), True) |
| |
|
| | ytr = label_from_supp2(mtr, supp2_path) |
| | yte = label_from_supp2(mte, supp2_path) |
| |
|
| | gtr = mtr["acc_base"].tolist() |
| | gte = mte["acc_base"].tolist() |
| | return Xtr, ytr, gtr, mtr, Xte, yte, gte, mte |
| |
|
| | def make_preprocess_pipeline(X: pd.DataFrame) -> ColumnTransformer: |
| | num_cols = list(X.columns) |
| | pre_num = Pipeline([ |
| | ("imputer", SimpleImputer(strategy="median")), |
| | ("scaler", StandardScaler(with_mean=True, with_std=True)) |
| | ]) |
| | return ColumnTransformer( |
| | [("num", pre_num, num_cols)], |
| | remainder="drop", |
| | verbose_feature_names_out=False |
| | ) |
| |
|
| | |
| | |
| | |
| | def _norm_code_str(x) -> str: |
| | if pd.isna(x): |
| | return "" |
| | return str(x).replace(",", "").replace(" ", "").strip().upper() |
| |
|
| | def _analyze_expected_inferred(expected: str, inferred: str): |
| | """ |
| | Return counts: |
| | aa_aa: AA<->AA different |
| | aa_q : AA<->? |
| | stop_aa: *<->AA |
| | stop_q : *<->? |
| | total_q: positions with '?' in either string |
| | valid: both length 64 |
| | """ |
| | e = _norm_code_str(expected) |
| | i = _norm_code_str(inferred) |
| | if len(e) != 64 or len(i) != 64: |
| | return dict(valid=False, aa_aa=0, aa_q=0, stop_aa=0, stop_q=0, total_q=0) |
| |
|
| | aa_aa = aa_q = stop_aa = stop_q = total_q = 0 |
| |
|
| | for a, b in zip(e, i): |
| | if a == QMARK or b == QMARK: |
| | total_q += 1 |
| |
|
| | if a == b: |
| | continue |
| |
|
| | a_is_aa = a in AA_SET |
| | b_is_aa = b in AA_SET |
| | a_is_q = a == QMARK |
| | b_is_q = b == QMARK |
| | a_is_s = a == STAR |
| | b_is_s = b == STAR |
| |
|
| | if a_is_aa and b_is_aa: |
| | aa_aa += 1 |
| | elif (a_is_aa and b_is_q) or (a_is_q and b_is_aa): |
| | aa_q += 1 |
| | elif (a_is_s and b_is_aa) or (a_is_aa and b_is_s): |
| | stop_aa += 1 |
| | elif (a_is_s and b_is_q) or (a_is_q and b_is_s): |
| | stop_q += 1 |
| | else: |
| | pass |
| |
|
| | return dict(valid=True, aa_aa=aa_aa, aa_q=aa_q, stop_aa=stop_aa, stop_q=stop_q, total_q=total_q) |
| |
|
| | def load_supp1_code_map(supp1_csv: str): |
| | """ |
| | Returns dict base_assembly -> (expected_str, inferred_str) |
| | Uses columns: |
| | - assembly |
| | - expected genetic code |
| | - Codetta inferred genetic code |
| | """ |
| | df = pd.read_csv(supp1_csv) |
| | cols = {c.lower(): c for c in df.columns} |
| |
|
| | def pick(key_sub): |
| | for k in cols: |
| | if k == key_sub: |
| | return cols[k] |
| | for k in cols: |
| | if key_sub in k: |
| | return cols[k] |
| | return None |
| |
|
| | asm_col = pick("assembly") |
| | exp_col = pick("expected genetic code") |
| | inf_col = pick("codetta inferred genetic code") |
| |
|
| | if not (asm_col and exp_col and inf_col): |
| | raise ValueError(f"[ERROR] Supp1.csv missing required columns. Found: {list(df.columns)}") |
| |
|
| | out = {} |
| | for _, row in df.iterrows(): |
| | asm = str(row[asm_col]) |
| | base = extract_acc_base(asm) |
| | out[base] = (row[exp_col], row[inf_col]) |
| | return out |
| |
|
| | def gentle_gc_penalty(gc_val, gc_median, gc_iqr): |
| | """ |
| | Very gentle penalty: near 1.0 for most of the range. |
| | Uses robust z = |gc - median| / IQR. |
| | Returns in [0.90, 1.00] (delicate gating). |
| | """ |
| | if not np.isfinite(gc_val) or gc_iqr <= 1e-9 or not np.isfinite(gc_median): |
| | return 1.0 |
| | z = abs(float(gc_val) - float(gc_median)) / float(gc_iqr) |
| | pen = float(np.exp(-0.08 * z)) |
| | return float(min(1.0, max(0.90, pen))) |
| |
|
| | def weight_for_supp2_genome(base, gc_percent, supp1_map, gc_median, gc_iqr): |
| | """ |
| | Weight only for genomes in Supp2: |
| | - STOP<->AA => very confident ALT signal (1.0) |
| | - AA<->AA => confident, but apply gentle GC penalty (close to 1.0) |
| | - STOP<->? treated like AA<->AA (your requirement) |
| | - AA<->? => less confident (downweights with total_q) |
| | If Supp1 missing or invalid: return 0.85 (still candidate but less trusted) |
| | """ |
| | if base not in supp1_map: |
| | return 0.85 |
| |
|
| | exp, inf = supp1_map[base] |
| | a = _analyze_expected_inferred(exp, inf) |
| | if not a["valid"]: |
| | return 0.85 |
| |
|
| | aa_aa = a["aa_aa"] |
| | aa_q = a["aa_q"] |
| | stop_aa = a["stop_aa"] |
| | stop_q = a["stop_q"] |
| | total_q = a["total_q"] |
| |
|
| | gc_pen = gentle_gc_penalty(gc_percent, gc_median, gc_iqr) |
| |
|
| | if stop_aa > 0: |
| | w = 1.00 * gc_pen |
| | return float(max(0.90, min(1.00, w))) |
| |
|
| | |
| | aa_like = aa_aa + stop_q |
| |
|
| | if aa_like > 0 and aa_q == 0: |
| | w = 0.98 * gc_pen |
| | return float(max(0.90, min(1.00, w))) |
| |
|
| | if aa_like > 0 and aa_q > 0: |
| | w = 0.95 * (0.97 ** max(total_q, 0)) * gc_pen |
| | return float(max(0.75, min(0.98, w))) |
| |
|
| | if aa_like == 0 and aa_q > 0: |
| | w = 0.70 * (0.95 ** max(total_q, 0)) * gc_pen |
| | return float(max(0.35, min(0.75, w))) |
| |
|
| | w = 0.80 * gc_pen |
| | return float(max(0.50, min(0.90, w))) |
| |
|
| | def build_sample_weights_for_dataset(meta_df, X_df, y_series, supp2_set, supp1_map): |
| | """ |
| | Returns np.array weights (len == n_samples). |
| | Only genomes in Supp2 get graded weights; others -> 1.0. |
| | """ |
| | gc_vals = X_df["gc_percent"].astype(float).values if "gc_percent" in X_df.columns else np.full(len(X_df), np.nan) |
| |
|
| | gc_clean = gc_vals[np.isfinite(gc_vals)] |
| | if gc_clean.size >= 10: |
| | gc_median = float(np.median(gc_clean)) |
| | q1 = float(np.quantile(gc_clean, 0.25)) |
| | q3 = float(np.quantile(gc_clean, 0.75)) |
| | gc_iqr = float(max(1e-6, q3 - q1)) |
| | else: |
| | gc_median, gc_iqr = float("nan"), 1.0 |
| |
|
| | weights = np.ones(len(X_df), dtype=float) |
| | bases = meta_df["acc_base"].astype(str).tolist() |
| |
|
| | for i, base in enumerate(bases): |
| | if base in supp2_set: |
| | gc = gc_vals[i] |
| | weights[i] = weight_for_supp2_genome(base, gc, supp1_map, gc_median, gc_iqr) |
| | else: |
| | weights[i] = 1.0 |
| | return weights |
| |
|
| | |
| | |
| | |
| | def metrics_from_pred(y_true, y_pred, y_proba=None): |
| | y_true = np.asarray(y_true); y_pred = np.asarray(y_pred) |
| | cm = confusion_matrix(y_true, y_pred, labels=[0,1]) |
| | tn, fp, fn, tp = int(cm[0,0]), int(cm[0,1]), int(cm[1,0]), int(cm[1,1]) |
| | row = dict(tn=tn, fp=fp, fn=fn, tp=tp) |
| | row["n"] = int(len(y_true)); row["positives"] = int(y_true.sum()) |
| | row["accuracy"] = float(accuracy_score(y_true, y_pred)) |
| | row["precision"] = float(precision_score(y_true, y_pred, pos_label=1, zero_division=0)) |
| | row["recall"] = float(recall_score(y_true, y_pred, pos_label=1, zero_division=0)) |
| | row["specificity"] = float(tn / (tn + fp)) if (tn + fp) > 0 else 0.0 |
| | row["f1"] = float(f1_score(y_true, y_pred, pos_label=1, zero_division=0)) |
| | if y_proba is not None and not np.isnan(y_proba).all(): |
| | try: row["roc_auc"] = float(roc_auc_score(y_true, y_proba)) |
| | except Exception: row["roc_auc"] = float("nan") |
| | try: row["pr_auc"] = float(average_precision_score(y_true, y_proba)) |
| | except Exception: row["pr_auc"] = float("nan") |
| | else: |
| | row["roc_auc"] = float("nan"); row["pr_auc"] = float("nan") |
| | return row |
| |
|
| | def best_threshold_f1(y_true, y_score): |
| | p, r, t = precision_recall_curve(y_true, y_score) |
| | f1 = (2*p*r) / np.maximum(p+r, 1e-12) |
| | idx = int(np.nanargmax(f1)) |
| | return float(t[max(0, idx-1)]) if idx == len(t) else float(t[idx]) |
| |
|
| | def best_threshold_fbeta(y_true, y_score, beta=2.0): |
| | p, r, t = precision_recall_curve(y_true, y_score) |
| | if t.size == 0: |
| | return float(np.median(y_score)) |
| | p = p[1:].astype(float) |
| | r = r[1:].astype(float) |
| | t = t.astype(float) |
| | valid = np.isfinite(p) & np.isfinite(r) & np.isfinite(t) |
| | p, r, t = p[valid], r[valid], t[valid] |
| | if t.size == 0: |
| | return float(np.median(y_score)) |
| | fbeta_vals = (1.0 + beta**2) * p * r / np.maximum(beta**2 * p + r, 1e-12) |
| | j = int(np.nanargmax(fbeta_vals)) |
| | return float(t[j]) |
| |
|
| | |
| | |
| | |
| | def make_objective(model_type: str, |
| | train_mode: str, |
| | X: pd.DataFrame, y: pd.Series, groups, |
| | metric_obj: str, |
| | n_splits: int, seed: int, threads: int, |
| | pu_bags: int, pu_u_ratio: float, |
| | fb_beta: float, |
| | sample_weight: np.ndarray | None): |
| | sgkf = StratifiedGroupKFold(n_splits=n_splits, shuffle=True, random_state=seed) |
| | pre = make_preprocess_pipeline(X) |
| |
|
| | def build_base_estimator(trial): |
| | if model_type == "RF": |
| | params = dict( |
| | n_estimators = trial.suggest_int("n_estimators", 400, 1400), |
| | max_depth = trial.suggest_int("max_depth", 6, 16), |
| | min_samples_split = trial.suggest_int("min_samples_split", 2, 20), |
| | min_samples_leaf = trial.suggest_int("min_samples_leaf", 2, 12), |
| | max_features = trial.suggest_categorical("max_features", ["sqrt"]), |
| | class_weight = trial.suggest_categorical("class_weight", [None, "balanced", "balanced_subsample"]), |
| | n_jobs = threads, |
| | random_state = seed, |
| | ) |
| | return RandomForestClassifier(**params) |
| |
|
| | if model_type == "ET": |
| | params = dict( |
| | n_estimators = trial.suggest_int("n_estimators", 600, 2000), |
| | max_depth = trial.suggest_int("max_depth", 6, 18), |
| | min_samples_split = trial.suggest_int("min_samples_split", 2, 20), |
| | min_samples_leaf = trial.suggest_int("min_samples_leaf", 2, 12), |
| | max_features = trial.suggest_categorical("max_features", ["sqrt"]), |
| | class_weight = trial.suggest_categorical("class_weight", [None, "balanced"]), |
| | n_jobs = threads, |
| | random_state = seed, |
| | ) |
| | return ExtraTreesClassifier(**params) |
| |
|
| | if not _HAS_XGB: |
| | raise RuntimeError("xgboost not installed") |
| | params = dict( |
| | n_estimators = trial.suggest_int("n_estimators", 400, 2000), |
| | max_depth = trial.suggest_int("max_depth", 3, 9), |
| | learning_rate= trial.suggest_float("learning_rate", 1e-3, 0.15, log=True), |
| | subsample = trial.suggest_float("subsample", 0.6, 1.0), |
| | colsample_bytree = trial.suggest_float("colsample_bytree", 0.5, 1.0), |
| | reg_alpha = trial.suggest_float("reg_alpha", 1e-4, 10.0, log=True), |
| | reg_lambda = trial.suggest_float("reg_lambda", 1e-3, 10.0, log=True), |
| | gamma = trial.suggest_float("gamma", 0.0, 5.0), |
| | min_child_weight = trial.suggest_float("min_child_weight", 1e-3, 10.0, log=True), |
| | n_jobs = threads, |
| | random_state = seed, |
| | tree_method = trial.suggest_categorical("tree_method", ["hist", "approx"]), |
| | objective = "binary:logistic", |
| | eval_metric = "logloss", |
| | ) |
| | return XGBClassifier(**params) |
| |
|
| | def objective(trial): |
| | clf = build_base_estimator(trial) |
| | base_pipe = Pipeline([("pre", pre), ("clf", clf)]) |
| |
|
| | if train_mode == "pu": |
| | model = PUBaggingClassifier( |
| | base_estimator=base_pipe, |
| | n_bags=pu_bags, |
| | u_ratio=pu_u_ratio, |
| | random_state=seed |
| | ) |
| | else: |
| | model = base_pipe |
| |
|
| | y_true_all = [] |
| | proba_all = [] |
| | f_scores = [] |
| |
|
| | for tr_idx, va_idx in sgkf.split(X, y, groups): |
| | Xtr, Xva = X.iloc[tr_idx], X.iloc[va_idx] |
| | ytr, yva = y.iloc[tr_idx], y.iloc[va_idx] |
| |
|
| | sw_tr = None |
| | if sample_weight is not None: |
| | sw_tr = np.asarray(sample_weight)[tr_idx] |
| |
|
| | model_fold = clone(model) |
| | fit_pipeline(model_fold, Xtr, ytr, sw_tr) |
| |
|
| | proba = model_fold.predict_proba(Xva)[:, 1] |
| | y_true_all.append(yva.values) |
| | proba_all.append(proba) |
| |
|
| | if metric_obj == "F1": |
| | thr = best_threshold_f1(yva.values, proba) |
| | yhat = (proba >= thr).astype(int) |
| | f_scores.append(f1_score(yva.values, yhat, zero_division=0)) |
| |
|
| | y_true_all = np.concatenate(y_true_all) |
| | proba_all = np.concatenate(proba_all) |
| |
|
| | if metric_obj == "F1": |
| | return float(np.mean(f_scores)) |
| | if metric_obj == "PR_AUC": |
| | return float(average_precision_score(y_true_all, proba_all)) |
| | if metric_obj == "FB": |
| | thr = best_threshold_fbeta(y_true_all, proba_all, beta=fb_beta) |
| | yhat = (proba_all >= thr).astype(int) |
| | return float(fbeta_score(y_true_all, yhat, beta=fb_beta, zero_division=0)) |
| |
|
| | raise ValueError("metric_obj must be one of: F1, PR_AUC, FB") |
| |
|
| | return objective, pre |
| |
|
| | |
| | |
| | |
| | def fit_best_model(model_type: str, |
| | train_mode: str, |
| | weight_mode: str, |
| | metric_obj: str, |
| | X: pd.DataFrame, y: pd.Series, groups, |
| | seed: int, timeout: int, n_trials: int, outdir: Path, tag: str, |
| | X_test: pd.DataFrame = None, y_test: pd.Series = None, |
| | threads: int = 1, |
| | pu_bags: int = 15, pu_u_ratio: float = 3.0, |
| | fb_beta: float = 2.0, |
| | calibrate: bool = True, |
| | sample_weight: np.ndarray | None = None, |
| | cv_folds: int = 5): |
| |
|
| | t0 = time.time() |
| |
|
| | def _make_eta_cb(t0, timeout, n_trials): |
| | def _cb(study, trial): |
| | try: |
| | elapsed = time.time() - t0 |
| | done = trial.number + 1 |
| | avg = elapsed / max(1, done) |
| | rem_trials = max(0, (n_trials or 0) - done) |
| | eta_trials = rem_trials * avg if n_trials else float('inf') |
| | eta_timeout = max(0.0, (timeout or 0) - elapsed) if timeout else float('inf') |
| | eta = min(eta_trials, eta_timeout) |
| | print(f"[TRIAL] #{trial.number:03d} value={trial.value:.5f} | best={study.best_value:.5f} | elapsed={elapsed/60:.1f}m | ETA~{eta/60:.1f}m") |
| | except Exception: |
| | print(f"[TRIAL] #{trial.number:03d} done") |
| | return _cb |
| |
|
| | study = optuna.create_study( |
| | direction="maximize", |
| | pruner=MedianPruner(n_startup_trials=8, n_warmup_steps=2), |
| | study_name=f"{model_type}_{train_mode}_{weight_mode}_{metric_obj}_{tag}" |
| | ) |
| |
|
| | objective, pre = make_objective( |
| | model_type=model_type, |
| | train_mode=train_mode, |
| | X=X, y=y, groups=groups, |
| | metric_obj=metric_obj, |
| | n_splits=cv_folds, |
| | seed=seed, threads=threads, |
| | pu_bags=pu_bags, pu_u_ratio=pu_u_ratio, |
| | fb_beta=fb_beta, |
| | sample_weight=sample_weight |
| | ) |
| |
|
| | study.optimize( |
| | objective, |
| | timeout=timeout, |
| | n_trials=n_trials, |
| | gc_after_trial=True, |
| | callbacks=[_make_eta_cb(t0, timeout, n_trials)] |
| | ) |
| |
|
| | best_params = dict(study.best_params) |
| |
|
| | |
| | if model_type == "RF": |
| | clf = RandomForestClassifier(**{**best_params, "n_jobs": threads, "random_state": seed}) |
| | elif model_type == "ET": |
| | clf = ExtraTreesClassifier(**{**best_params, "n_jobs": threads, "random_state": seed}) |
| | else: |
| | if not _HAS_XGB: |
| | raise RuntimeError("xgboost not installed") |
| | clf = XGBClassifier(**{ |
| | **best_params, |
| | "n_jobs": threads, |
| | "random_state": seed, |
| | "objective": "binary:logistic", |
| | "eval_metric": "logloss", |
| | }) |
| |
|
| | base_pipe = Pipeline([("pre", pre), ("clf", clf)]) |
| |
|
| | if train_mode == "pu": |
| | model = PUBaggingClassifier( |
| | base_estimator=base_pipe, |
| | n_bags=pu_bags, |
| | u_ratio=pu_u_ratio, |
| | random_state=seed |
| | ) |
| | else: |
| | model = base_pipe |
| |
|
| | |
| | fit_pipeline(model, X, y, sample_weight) |
| |
|
| | |
| | final_model = model |
| | if calibrate: |
| | |
| | try: |
| | calib = CalibratedClassifierCV(estimator=model, method="isotonic", cv=cv_folds) |
| | except TypeError: |
| | calib = CalibratedClassifierCV(base_estimator=model, method="isotonic", cv=cv_folds) |
| |
|
| | |
| | calib.fit(X, y) |
| | final_model = calib |
| |
|
| | |
| | proba_tr = final_model.predict_proba(X)[:, 1] |
| | if metric_obj == "F1": |
| | tau = best_threshold_f1(y.values, proba_tr) |
| | elif metric_obj == "FB": |
| | tau = best_threshold_fbeta(y.values, proba_tr, beta=fb_beta) |
| | else: |
| | tau = best_threshold_f1(y.values, proba_tr) |
| |
|
| | yhat_tr = (proba_tr >= tau).astype(int) |
| | train_row = metrics_from_pred(y.values, yhat_tr, proba_tr) |
| | train_row = {f"train_{k}": v for k, v in train_row.items()} |
| |
|
| | metrics = dict( |
| | threshold_used=float(tau), |
| | study_best=float(study.best_value), |
| | model_type=model_type, |
| | train_mode=train_mode, |
| | weight_mode=weight_mode, |
| | metric_obj=metric_obj, |
| | fb_beta=float(fb_beta), |
| | pu_bags=int(pu_bags), |
| | pu_u_ratio=float(pu_u_ratio), |
| | calibrated=bool(calibrate), |
| | **train_row |
| | ) |
| |
|
| | |
| | if X_test is not None and y_test is not None: |
| | X_te = X_test.reindex(columns=list(X.columns)) |
| | proba_te = final_model.predict_proba(X_te)[:, 1] |
| | yhat_te = (proba_te >= tau).astype(int) |
| | test_row = metrics_from_pred(y_test.values, yhat_te, proba_te) |
| | metrics.update({f"test_{k}": v for k, v in test_row.items()}) |
| |
|
| | return final_model, best_params, metrics, study |
| |
|
| | |
| | |
| | |
| | def model_run_id(model_type, train_mode, weight_mode, metric_obj, tag): |
| | return f"{model_type}_{train_mode}_{weight_mode}_{metric_obj}_{tag}" |
| |
|
| | def artifact_paths(results_dir: Path, run_id: str): |
| | return { |
| | "model": results_dir / f"model_{run_id}.joblib", |
| | "params": results_dir / f"best_params_{run_id}.json", |
| | "metrics": results_dir / f"metrics_{run_id}.json", |
| | } |
| |
|
| | def is_run_complete(results_dir: Path, run_id: str) -> bool: |
| | p = artifact_paths(results_dir, run_id) |
| | return p["model"].exists() and p["metrics"].exists() and p["params"].exists() |
| |
|
| | def load_existing_metrics(results_dir: Path) -> list: |
| | rows = [] |
| | for mf in sorted(results_dir.glob("metrics_*.json")): |
| | try: |
| | d = json.loads(mf.read_text()) |
| | rows.append(d) |
| | except Exception: |
| | continue |
| | return rows |
| |
|
| | def append_metrics_summary(results_dir: Path, rows: list): |
| | if not rows: |
| | return |
| | df = pd.DataFrame(rows) |
| | df.to_csv(results_dir / "metrics_summary.csv", index=False) |
| |
|
| | |
| | |
| | |
| | def main(): |
| | ap = argparse.ArgumentParser(description="Train RF/XGB/ET with grouped CV. supervised + PU. normal + weighted.") |
| | ap.add_argument("--ndjson", required=True, help="Folder with train.jsonl and test.jsonl") |
| | ap.add_argument("--supp2", required=True, help="Supp2.xlsx (positives list for PU; label=1)") |
| | ap.add_argument("--supp1", required=False, default=None, help="Supp1.csv (for weighted grading of Supp2 genomes)") |
| | ap.add_argument("--outdir", required=True, help="Output directory root (will create results_models/)") |
| |
|
| | ap.add_argument("--train_mode", choices=["supervised","pu","both"], default="both", |
| | help="Training mode: classic supervised vs PU-bagging vs both") |
| |
|
| | ap.add_argument("--weight_mode", choices=["normal","weighted","both"], default="both", |
| | help="normal: all weights=1; weighted: grade only Supp2 genomes via Supp1 expected/inferred; both: run both") |
| |
|
| | ap.add_argument("--model", choices=["RF","XGB","ET","all"], default="all", |
| | help="Model(s): RF, XGB, ET (ExtraTrees baseline), or all") |
| |
|
| | ap.add_argument("--metric", choices=["F1","PR_AUC","FB","all"], default="all", |
| | help="Optuna objective metric") |
| |
|
| | ap.add_argument("--fb_beta", type=float, default=2.0, help="Beta for F-beta metric (FB objective)") |
| |
|
| | ap.add_argument("--timeout", type=int, default=3600, help="Optuna time budget per run (seconds)") |
| | ap.add_argument("--n_trials", type=int, default=60, help="Upper limit of trials if timeout not reached") |
| | ap.add_argument("--threads", type=int, default=0, help="Threads (0=auto cpu_count()-4)") |
| | ap.add_argument("--seed", type=int, default=42, help="Random seed") |
| |
|
| | ap.add_argument("--pu_bags", type=int, default=15, help="PU bagging: number of bags") |
| | ap.add_argument("--pu_u_ratio", type=float, default=3.0, help="PU bagging: U sampled per bag = ratio * |P|") |
| |
|
| | ap.add_argument("--no_calibration", action="store_true", |
| | help="Disable isotonic calibration (use raw predict_proba from base model / PU ensemble).") |
| |
|
| | ap.add_argument("--force_calibration_pu", action="store_true", |
| | help="Force calibration even for PU (can be extremely slow: cv_folds * pu_bags fits).") |
| |
|
| | ap.add_argument("--cv", type=int, default=5, help="Grouped CV folds for objective & calibration") |
| | ap.add_argument("--overwrite", action="store_true", |
| | help="Re-train even if artifacts already exist for a run_id (otherwise skipped).") |
| |
|
| | args = ap.parse_args() |
| |
|
| | cpu_total = os.cpu_count() or 8 |
| | eff_threads = max(1, cpu_total - 4) if args.threads in (None, 0, -1) else max(1, args.threads) |
| | print(f"[CPU ] total={cpu_total} using={eff_threads} (flag={args.threads})") |
| |
|
| | os.environ['OMP_NUM_THREADS'] = str(eff_threads) |
| | os.environ['OPENBLAS_NUM_THREADS'] = str(eff_threads) |
| | os.environ['MKL_NUM_THREADS'] = str(eff_threads) |
| | os.environ['VECLIB_MAXIMUM_THREADS'] = str(eff_threads) |
| | os.environ['NUMEXPR_NUM_THREADS'] = str(eff_threads) |
| |
|
| | out_root = Path(args.outdir) |
| | out_root.mkdir(parents=True, exist_ok=True) |
| |
|
| | results_dir = out_root / "results_models" |
| | results_dir.mkdir(parents=True, exist_ok=True) |
| |
|
| | |
| | Xtr, ytr, gtr, mtr, Xte, yte, gte, mte = _load_train_test(args.ndjson, args.supp2) |
| | tag = "64log_gc_tetra" |
| |
|
| | supp2_set = load_alt_list_from_supp2(args.supp2) |
| |
|
| | |
| | supp1_map = {} |
| | if args.supp1: |
| | print("[LOAD] Supp1 mapping for weights:", args.supp1) |
| | supp1_map = load_supp1_code_map(args.supp1) |
| |
|
| | weights_train_weighted = None |
| | weights_test_weighted = None |
| | if args.weight_mode in ("weighted", "both"): |
| | if not args.supp1: |
| | raise ValueError("--weight_mode weighted/both requires --supp1 Supp1.csv") |
| | weights_train_weighted = build_sample_weights_for_dataset(mtr, Xtr, ytr, supp2_set, supp1_map) |
| | weights_test_weighted = build_sample_weights_for_dataset(mte, Xte, yte, supp2_set, supp1_map) |
| |
|
| | |
| | snap_tr = pd.DataFrame({ |
| | "acc_base": mtr["acc_base"].astype(str).values, |
| | "y": ytr.astype(int).values, |
| | "gc_percent": Xtr["gc_percent"].astype(float).values, |
| | "weight": weights_train_weighted |
| | }) |
| | snap_te = pd.DataFrame({ |
| | "acc_base": mte["acc_base"].astype(str).values, |
| | "y": yte.astype(int).values, |
| | "gc_percent": Xte["gc_percent"].astype(float).values, |
| | "weight": weights_test_weighted |
| | }) |
| | snap_tr.to_csv(results_dir / "weights_snapshot_train.tsv", sep="\t", index=False) |
| | snap_te.to_csv(results_dir / "weights_snapshot_test.tsv", sep="\t", index=False) |
| | print("[WRITE] weights_snapshot_train.tsv / weights_snapshot_test.tsv") |
| |
|
| | wP = weights_train_weighted[ytr.values == 1] |
| | if wP.size: |
| | print(f"[WEIGHTS] Supp2(P) weights: n={wP.size} mean={wP.mean():.3f} sd={wP.std():.3f} min={wP.min():.3f} max={wP.max():.3f}") |
| |
|
| | print("\n" + "="*72) |
| | print(f"[DATA] train={Xtr.shape} P(from Supp2)={int(ytr.sum())} ({100.0*float(ytr.mean()):.3f}%)") |
| | print(f"[DATA] test={Xte.shape} P(from Supp2)={int(yte.sum())} ({100.0*float(yte.mean()):.3f}%)") |
| | print("="*72) |
| |
|
| | train_modes = ["supervised","pu"] if args.train_mode == "both" else [args.train_mode] |
| | weight_modes = ["normal","weighted"] if args.weight_mode == "both" else [args.weight_mode] |
| | models = ["RF","XGB","ET"] if args.model == "all" else [args.model] |
| | metrics = ["F1","PR_AUC","FB"] if args.metric == "all" else [args.metric] |
| |
|
| | |
| | global_rows = load_existing_metrics(results_dir) |
| | if global_rows: |
| | append_metrics_summary(results_dir, global_rows) |
| | print(f"[INFO] Found existing metrics: {len(global_rows)} runs. metrics_summary.csv refreshed.") |
| |
|
| | for tm in train_modes: |
| | for wm in weight_modes: |
| | for met in metrics: |
| | for mdl in models: |
| | run_id = model_run_id(mdl, tm, wm, met, tag) |
| |
|
| | |
| | calibrate = (not args.no_calibration) |
| | if tm == "pu" and calibrate and not args.force_calibration_pu: |
| | print(f"[INFO] Auto-disabling calibration for PU run {run_id} (use --force_calibration_pu to override).") |
| | calibrate = False |
| |
|
| | |
| | if (not args.overwrite) and is_run_complete(results_dir, run_id): |
| | print(f"[SKIP] already exists: {run_id}") |
| | continue |
| |
|
| | print("\n" + "-"*72) |
| | print(f"[RUN ] mode={tm} | weights={wm} | metric={met} | model={mdl} | timeout={args.timeout}s | trials={args.n_trials}") |
| | print(f"[RUN ] run_id={run_id} | calibrate={calibrate}") |
| | print("-"*72) |
| |
|
| | |
| | if wm == "weighted": |
| | sw_tr = weights_train_weighted |
| | sw_te = weights_test_weighted |
| | else: |
| | sw_tr = None |
| | sw_te = None |
| |
|
| | t0 = time.time() |
| | final_model, best_params, met_dict, study = fit_best_model( |
| | model_type=mdl, |
| | train_mode=tm, |
| | weight_mode=wm, |
| | metric_obj=met, |
| | X=Xtr, y=ytr, groups=gtr, |
| | seed=args.seed, |
| | timeout=args.timeout, |
| | n_trials=args.n_trials, |
| | outdir=results_dir, |
| | tag=tag, |
| | X_test=Xte, |
| | y_test=yte, |
| | threads=eff_threads, |
| | pu_bags=args.pu_bags, |
| | pu_u_ratio=args.pu_u_ratio, |
| | fb_beta=args.fb_beta, |
| | calibrate=calibrate, |
| | sample_weight=sw_tr, |
| | cv_folds=args.cv |
| | ) |
| |
|
| | dt = time.time() - t0 |
| | print(f"[DONE] {mdl} | {tm} | {wm} | {met} in {dt/60:.1f} min best={study.best_value:.5f}") |
| |
|
| | |
| | model_path = results_dir / f"model_{run_id}.joblib" |
| | params_path = results_dir / f"best_params_{run_id}.json" |
| | metrics_path = results_dir / f"metrics_{run_id}.json" |
| | cols_path = results_dir / f"feature_columns_{tag}.json" |
| |
|
| | joblib.dump(final_model, model_path) |
| | params_path.write_text(json.dumps(best_params, indent=2)) |
| | metrics_path.write_text(json.dumps(met_dict, indent=2)) |
| | if not cols_path.exists(): |
| | cols_path.write_text(json.dumps(list(Xtr.columns), indent=2)) |
| |
|
| | |
| | |
| | global_rows = [r for r in global_rows if not ( |
| | r.get("model_type")==mdl and r.get("train_mode")==tm and r.get("weight_mode")==wm |
| | and r.get("metric_obj")==met |
| | )] |
| | met_row = dict(met_dict) |
| | met_row["elapsed_min"] = float(dt/60.0) |
| | global_rows.append(met_row) |
| |
|
| | append_metrics_summary(results_dir, global_rows) |
| | print("[WRITE] metrics_summary.csv updated") |
| |
|
| | print("[DONE] Saved artifacts to", results_dir.resolve()) |
| |
|
| | if __name__ == "__main__": |
| | main() |
| |
|
| |
|