FireProtDB2 / src /07_gen_subsets.py
drake463's picture
final pipeline and updated subsets
ae74a72
#!/usr/bin/env python3
from __future__ import annotations
import os
import argparse
import pandas as pd
OUT_SPLITS = ['train', 'validation', 'test']
def ensure_dir(p: str):
os.makedirs(p, exist_ok=True)
def write_subset(out_dir: str, name: str, df: pd.DataFrame):
outp = os.path.join(out_dir, name)
ensure_dir(outp)
for split in OUT_SPLITS:
fp = os.path.join(outp, f"{split}.parquet")
d_split = df[df["split"] == split].copy()
d_split.to_parquet(fp, index=False)
print(f"{name}/{split}: {len(d_split):,} -> {fp}")
def main():
ap = argparse.ArgumentParser()
ap.add_argument("--input", default="../data/fireprotdb_with_cluster_splits.parquet",
help="Parquet with MMseqs2 cluster_id and split columns already assigned")
ap.add_argument("--out_dir", default="../data/subsets",
help="Output dir for HF subset parquets")
ap.add_argument("--drop_columns", default="",
help="Comma-separated columns to drop in outputs (optional)")
ap.add_argument("--require_sequence", action="store_true",
help="If set, drop rows without a UniProt sequence before creating subsets")
ap.add_argument("--strict_split_check", action="store_true",
help="If set, error if any protein/cluster appears in multiple splits")
args = ap.parse_args()
df = pd.read_parquet(args.input)
# Basic required columns
if "split" not in df.columns:
raise ValueError("Input parquet must contain a 'split' column (train/validation/test).")
if "mutation" not in df.columns:
raise ValueError("Input parquet must contain a 'mutation' column.")
# Normalize split values
df["split"] = df["split"].astype("string").str.strip().str.lower()
bad = sorted(set(df["split"].dropna().unique()) - set(OUT_SPLITS))
if bad:
raise ValueError(f"Unexpected split values: {bad}. Expected {OUT_SPLITS}.")
if args.require_sequence and "sequence" in df.columns:
before = len(df)
df = df[df["sequence"].notna()].copy()
print(f"Dropped rows without sequence: {before - len(df):,}")
# Optional leakage check: ensure each cluster/protein lives in exactly one split
# Prefer cluster_id if available, else fall back to uniprotkb/sequence_id
if args.strict_split_check:
if "cluster_id" in df.columns:
key = "cluster_id"
elif "uniprotkb" in df.columns:
key = "uniprotkb"
else:
key = "sequence_id"
counts = df.groupby(key, dropna=False)["split"].nunique()
leaked = counts[counts > 1]
if len(leaked):
example = leaked.head(10)
raise ValueError(
f"Leakage detected: {len(leaked):,} {key}s appear in multiple splits.\n"
f"Examples:\n{example}"
)
print(f"Split leakage check passed using key={key}.")
# Base filter: parsed mutation only (matches your fireprotdb.py intent)
has_mut = df["mutation"].notna()
df_mut = df[has_mut].copy()
# Build subsets (you said you only want these)
dg = df_mut[df_mut["dg"].notna()].copy()
ddg = df_mut[df_mut["ddg"].notna()].copy()
tm = df_mut[df_mut["tm"].notna()].copy()
dtm = df_mut[df_mut["dtm"].notna()].copy()
fitness = df_mut[df_mut["fitness"].notna()].copy()
binary = df_mut[df_mut["stabilizing"].notna()].copy()
# Optional: drop columns to slim outputs
if args.drop_columns.strip():
drop_cols = [c.strip() for c in args.drop_columns.split(",") if c.strip()]
keep_drop = [c for c in drop_cols if c in df.columns]
if keep_drop:
dg = dg.drop(columns=keep_drop, errors="ignore")
ddg = ddg.drop(columns=keep_drop, errors="ignore")
tm = tm.drop(columns=keep_drop, errors="ignore")
dtm = dtm.drop(columns=keep_drop, errors="ignore")
fitness = fitness.drop(columns=keep_drop, errors="ignore")
binary = binary.drop(columns=keep_drop, errors="ignore")
write_subset(args.out_dir, "mutation_dg", dg)
write_subset(args.out_dir, "mutation_ddg", ddg)
write_subset(args.out_dir, "mutation_tm", tm)
write_subset(args.out_dir, "mutation_dtm", dtm)
write_subset(args.out_dir, "mutation_fitness", fitness)
write_subset(args.out_dir, "mutation_binary", binary)
if __name__ == "__main__":
main()