NegBioDB / scripts_ct /train_ct_baseline.py
jang1563's picture
NegBioDB final: 4 domains, fully audited
6d1bbc7
#!/usr/bin/env python3
"""Unified training harness for CT baseline models.
Supports XGBoost, MLP, GNN+Tab on CT-M1 (binary) and CT-M2 (7/8-way classification).
Usage:
python scripts_ct/train_ct_baseline.py \\
--model xgboost \\
--task m1 \\
--split random \\
--dataset balanced \\
--negative negbiodb \\
--epochs 100 --patience 15 \\
--batch_size 256 --lr 0.001 --seed 42 \\
--data_dir exports/ct/ \\
--output_dir results/ct_baselines/
Outputs:
results/ct_baselines/{model}_{task}_{dataset}_{split}_{negative}_seed{seed}/
best.pt|best.json — checkpoint (pt for MLP/GNN, json for XGBoost)
results.json — test-set metrics
training_log.csv — per-epoch log (MLP/GNN only)
"""
from __future__ import annotations
import argparse
import csv
import json
import logging
import random
import sys
from pathlib import Path
import numpy as np
import pandas as pd
logging.basicConfig(
level=logging.INFO,
format="%(asctime)s %(levelname)s %(message)s",
datefmt="%H:%M:%S",
stream=sys.stderr,
)
logger = logging.getLogger(__name__)
ROOT = Path(__file__).parent.parent
# ---------------------------------------------------------------------------
# Dataset maps
# ---------------------------------------------------------------------------
# M1 parquets by (dataset, negative)
_M1_DATASET_MAP: dict[tuple[str, str], str] = {
("balanced", "negbiodb"): "negbiodb_ct_m1_balanced.parquet",
("realistic", "negbiodb"): "negbiodb_ct_m1_realistic.parquet",
("smiles_only", "negbiodb"): "negbiodb_ct_m1_smiles_only.parquet",
("balanced", "uniform_random"): "negbiodb_ct_m1_uniform_random.parquet",
("balanced", "degree_matched"): "negbiodb_ct_m1_degree_matched.parquet",
}
# M2 has a single parquet
_M2_PARQUET = "negbiodb_ct_m2.parquet"
# Pre-computed split columns in M1 parquets
_M1_PRECOMPUTED_SPLITS = {"random", "cold_drug", "cold_condition"}
# Failure category mapping (must match ct_export.CATEGORY_TO_INT)
CATEGORY_TO_INT: dict[str, int] = {
"efficacy": 0,
"enrollment": 1,
"other": 2,
"strategic": 3,
"safety": 4,
"design": 5,
"regulatory": 6,
"pharmacokinetic": 7,
}
# ---------------------------------------------------------------------------
# Utilities
# ---------------------------------------------------------------------------
def set_seed(seed: int) -> None:
random.seed(seed)
np.random.seed(seed)
try:
import torch
torch.manual_seed(seed)
if torch.cuda.is_available():
torch.cuda.manual_seed_all(seed)
except ImportError:
pass
def _json_safe(value):
"""Convert NaN/Inf to JSON-safe nulls."""
if isinstance(value, dict):
return {k: _json_safe(v) for k, v in value.items()}
if isinstance(value, (list, tuple)):
return [_json_safe(v) for v in value]
if isinstance(value, (np.floating, float)):
value = float(value)
return value if np.isfinite(value) else None
if isinstance(value, np.integer):
return int(value)
if isinstance(value, np.ndarray):
return _json_safe(value.tolist())
return value
def write_results_json(path: Path, payload: dict) -> None:
with open(path, "w") as f:
json.dump(_json_safe(payload), f, indent=2, allow_nan=False)
# ---------------------------------------------------------------------------
# Data Loading + Split Resolution
# ---------------------------------------------------------------------------
def _resolve_m1_parquet(dataset: str, negative: str) -> str | None:
return _M1_DATASET_MAP.get((dataset, negative))
def _get_split_column(task: str, split: str) -> str:
"""Return the split column name in the parquet."""
return f"split_{split}"
def _compute_runtime_split(df: pd.DataFrame, split: str, seed: int) -> pd.Series:
"""Compute temporal/scaffold/degree_balanced splits at runtime for M1.
M1 parquets may have NaN pair_id for success rows (from CTO).
We assign synthetic integer IDs for the split functions.
"""
from negbiodb_ct.ct_export import (
generate_ct_degree_balanced_split,
generate_ct_scaffold_split,
generate_ct_temporal_split,
)
# Use synthetic IDs: M1 pair_id can have NaN for success rows
df = df.copy()
synthetic_id_col = "_split_id"
df[synthetic_id_col] = range(len(df))
# Temporarily set as pair_id/result_id for split functions
orig_id_col = "pair_id" if "pair_id" in df.columns else "result_id"
orig_ids = df[orig_id_col].copy()
df[orig_id_col] = df[synthetic_id_col]
if split == "temporal":
fold_map = generate_ct_temporal_split(df)
elif split == "scaffold":
fold_map = generate_ct_scaffold_split(df, seed=seed)
elif split == "degree_balanced":
fold_map = generate_ct_degree_balanced_split(df, seed=seed)
else:
raise ValueError(f"Cannot compute runtime split for: {split}")
# Map back using synthetic IDs
return pd.Series(df[synthetic_id_col].map(fold_map).values, index=df.index)
def load_data(args) -> tuple[pd.DataFrame, pd.DataFrame, pd.DataFrame]:
"""Load and split data for the given task/split/dataset/negative combo.
Returns (train_df, val_df, test_df).
"""
if args.task == "m1":
filename = _resolve_m1_parquet(args.dataset, args.negative)
if filename is None:
raise ValueError(
f"M1 dataset combo not found: dataset={args.dataset}, negative={args.negative}"
)
parquet_path = args.data_dir / filename
if not parquet_path.exists():
raise FileNotFoundError(f"M1 parquet not found: {parquet_path}")
df = pd.read_parquet(parquet_path)
logger.info("Loaded M1 data: %d rows from %s", len(df), parquet_path.name)
# Get split column
split_col = _get_split_column(args.task, args.split)
if split_col in df.columns:
logger.info("Using pre-computed split: %s", split_col)
elif args.split in _M1_PRECOMPUTED_SPLITS:
raise ValueError(f"Split column {split_col} not found in {parquet_path.name}")
else:
# Runtime split for temporal/scaffold/degree_balanced
if args.split == "temporal" and args.negative != "negbiodb":
raise ValueError(
f"Temporal split requires negbiodb negatives "
f"(earliest_completion_year column), got --negative={args.negative}"
)
logger.info("Computing runtime split: %s", args.split)
df[split_col] = _compute_runtime_split(df, args.split, args.seed)
# Drop rows with None fold (e.g. scaffold split for non-SMILES)
before = len(df)
df = df.dropna(subset=[split_col]).reset_index(drop=True)
if len(df) < before:
logger.info("Dropped %d rows with NULL fold", before - len(df))
else: # m2
parquet_path = args.data_dir / _M2_PARQUET
if not parquet_path.exists():
raise FileNotFoundError(f"M2 parquet not found: {parquet_path}")
df = pd.read_parquet(parquet_path)
logger.info("Loaded M2 data: %d rows from %s", len(df), parquet_path.name)
split_col = _get_split_column(args.task, args.split)
if split_col not in df.columns:
raise ValueError(f"Split column {split_col} not in {parquet_path.name}")
# Split into folds
train_df = df[df[split_col] == "train"].reset_index(drop=True)
val_df = df[df[split_col] == "val"].reset_index(drop=True)
test_df = df[df[split_col] == "test"].reset_index(drop=True)
logger.info(
"Split sizes — train: %d, val: %d, test: %d",
len(train_df), len(val_df), len(test_df),
)
return train_df, val_df, test_df
# ---------------------------------------------------------------------------
# Metrics
# ---------------------------------------------------------------------------
def compute_m1_metrics(y_true: np.ndarray, y_score: np.ndarray) -> dict[str, float]:
"""Compute M1 binary metrics (9 total: 7 DTI + accuracy + F1)."""
from negbiodb.metrics import compute_all_metrics
from sklearn.metrics import accuracy_score, f1_score
if len(np.unique(y_true)) < 2:
logger.warning("Test set has only one class — metrics will be NaN.")
keys = ["auroc", "auprc", "mcc", "log_auc", "bedroc", "ef_1pct", "ef_5pct", "accuracy", "f1"]
return {k: float("nan") for k in keys}
base = compute_all_metrics(y_true, y_score)
y_pred = (y_score >= 0.5).astype(int)
base["accuracy"] = float(accuracy_score(y_true, y_pred))
base["f1"] = float(f1_score(y_true, y_pred, zero_division=0))
return base
def compute_m2_metrics(y_true: np.ndarray, y_pred: np.ndarray) -> dict[str, float]:
"""Compute M2 multiclass metrics."""
from sklearn.metrics import (
accuracy_score,
confusion_matrix,
f1_score,
matthews_corrcoef,
)
n_classes = len(CATEGORY_TO_INT) # always 8
all_labels = list(range(n_classes))
per_class_acc = {}
for c in all_labels:
mask = y_true == c
cat_name = [k for k, v in CATEGORY_TO_INT.items() if v == c]
cat_name = cat_name[0] if cat_name else str(c)
if mask.sum() > 0:
per_class_acc[cat_name] = float((y_pred[mask] == c).mean())
else:
per_class_acc[cat_name] = float("nan")
return {
"macro_f1": float(f1_score(y_true, y_pred, average="macro", zero_division=0, labels=all_labels)),
"weighted_f1": float(f1_score(y_true, y_pred, average="weighted", zero_division=0, labels=all_labels)),
"mcc": float(matthews_corrcoef(y_true, y_pred)),
"accuracy": float(accuracy_score(y_true, y_pred)),
"per_class_accuracy": per_class_acc,
"confusion_matrix": confusion_matrix(y_true, y_pred, labels=all_labels).tolist(),
}
# ---------------------------------------------------------------------------
# XGBoost Training
# ---------------------------------------------------------------------------
def train_xgboost(
train_df: pd.DataFrame,
val_df: pd.DataFrame,
test_df: pd.DataFrame,
task: str,
output_dir: Path,
seed: int,
) -> dict:
"""Train and evaluate XGBoost."""
import xgboost as xgb
from negbiodb_ct.ct_features import build_xgboost_features
train_X = build_xgboost_features(train_df, task)
val_X = build_xgboost_features(val_df, task)
test_X = build_xgboost_features(test_df, task)
if task == "m1":
train_y = train_df["Y"].values.astype(np.float32)
val_y = val_df["Y"].values.astype(np.float32)
test_y = test_df["Y"].values.astype(np.float32)
else:
train_y = train_df["failure_category_int"].values.astype(np.int64)
val_y = val_df["failure_category_int"].values.astype(np.int64)
test_y = test_df["failure_category_int"].values.astype(np.int64)
params: dict = {
"max_depth": 6,
"learning_rate": 0.1,
"n_estimators": 500,
"early_stopping_rounds": 20,
"tree_method": "hist",
"random_state": seed,
"verbosity": 0,
}
if task == "m1":
params["objective"] = "binary:logistic"
params["eval_metric"] = "logloss"
n_neg = (train_y == 0).sum()
n_pos = (train_y == 1).sum()
if n_pos > 0:
params["scale_pos_weight"] = float(n_neg / n_pos)
else:
params["objective"] = "multi:softprob"
params["eval_metric"] = "mlogloss"
params["num_class"] = 8
logger.info("XGBoost params: %s", {k: v for k, v in params.items() if k != "verbosity"})
model = xgb.XGBClassifier(**params)
model.fit(
train_X, train_y,
eval_set=[(val_X, val_y)],
verbose=False,
)
# Save model
model_path = output_dir / "best.json"
model.save_model(str(model_path))
logger.info("Saved XGBoost model → %s", model_path)
# Evaluate
if task == "m1":
y_score = model.predict_proba(test_X)[:, 1]
test_metrics = compute_m1_metrics(test_y, y_score)
else:
y_proba = model.predict_proba(test_X)
y_pred = np.argmax(y_proba, axis=1)
test_metrics = compute_m2_metrics(test_y, y_pred)
return test_metrics
# ---------------------------------------------------------------------------
# MLP/GNN PyTorch Datasets + Collate
# ---------------------------------------------------------------------------
class CTTabularDataset:
"""Tabular dataset for MLP."""
def __init__(self, df: pd.DataFrame, task: str) -> None:
from negbiodb_ct.ct_features import build_mlp_features
self.X = build_mlp_features(df, task).astype(np.float32)
if task == "m1":
self.y = df["Y"].values.astype(np.float32)
else:
self.y = df["failure_category_int"].values.astype(np.int64)
self.task = task
def __len__(self) -> int:
return len(self.y)
def __getitem__(self, idx: int):
return self.X[idx], self.y[idx]
class CTGraphDataset:
"""Graph + tabular dataset for GNN+Tab. SMILES-only rows."""
def __init__(
self,
df: pd.DataFrame,
task: str,
graph_cache: dict,
) -> None:
from negbiodb_ct.ct_features import build_gnn_tab_features
# Filter to SMILES-only
has_smiles = df["smiles"].notna()
self.df = df[has_smiles].reset_index(drop=True)
if len(self.df) < len(df):
logger.info(
"GNN dataset: dropped %d rows without SMILES (%d → %d)",
len(df) - len(self.df), len(df), len(self.df),
)
self.smiles = self.df["smiles"].tolist()
self.tab = build_gnn_tab_features(self.df, task).astype(np.float32)
if task == "m1":
self.y = self.df["Y"].values.astype(np.float32)
else:
self.y = self.df["failure_category_int"].values.astype(np.int64)
self.task = task
self.graph_cache = graph_cache
def __len__(self) -> int:
return len(self.y)
def __getitem__(self, idx: int):
graph = self.graph_cache.get(self.smiles[idx])
return graph, self.tab[idx], self.y[idx]
def _collate_mlp(batch, device):
import torch
X_list, y_list = zip(*batch)
X = torch.tensor(np.array(X_list), dtype=torch.float32).to(device)
y = torch.tensor(np.array(y_list)).to(device)
return X, y
def _collate_gnn(batch, device):
import torch
from torch_geometric.data import Batch, Data
from negbiodb.models.graphdta import NODE_FEATURE_DIM
graphs, tabs, labels = zip(*batch)
# Placeholder for None graphs
placeholder = Data(
x=torch.zeros(1, NODE_FEATURE_DIM),
edge_index=torch.zeros(2, 0, dtype=torch.long),
)
graphs_clean = [g if g is not None else placeholder for g in graphs]
drug_batch = Batch.from_data_list(graphs_clean).to(device)
tab_tensor = torch.tensor(np.array(tabs), dtype=torch.float32).to(device)
label_tensor = torch.tensor(np.array(labels)).to(device)
return drug_batch, tab_tensor, label_tensor
def _prepare_ct_graph_cache(parquet_path: Path, cache_path: Path) -> dict:
"""Build/load graph cache for CT GNN training."""
import torch
from negbiodb.models.graphdta import smiles_to_graph
smiles_series = pd.read_parquet(parquet_path, columns=["smiles"])["smiles"].dropna()
smiles_list = smiles_series.unique().tolist()
cache: dict = {}
if cache_path.exists():
logger.info("Loading graph cache: %s", cache_path)
cache = torch.load(cache_path, weights_only=False)
logger.info("Graph cache loaded: %d entries", len(cache))
missing = [smi for smi in smiles_list if smi not in cache]
if missing:
logger.info("Building graphs for %d missing SMILES...", len(missing))
failed = 0
for smi in missing:
g = smiles_to_graph(smi)
cache[smi] = g
if g is None:
failed += 1
cache_path.parent.mkdir(parents=True, exist_ok=True)
torch.save(cache, cache_path)
logger.info("Saved graph cache → %s (%d total, %d failed)", cache_path, len(cache), failed)
return cache
# ---------------------------------------------------------------------------
# MLP/GNN Training Loop
# ---------------------------------------------------------------------------
def _run_epoch_mlp(model, loader, criterion, optimizer, device, is_train: bool, task: str):
"""Run one epoch for MLP."""
import torch
model.train(is_train)
total_loss = 0.0
all_labels = []
all_preds = []
n_batches = 0
with torch.set_grad_enabled(is_train):
for X, y in loader:
logits = model(X)
if task == "m1":
loss = criterion(logits, y.float())
else:
loss = criterion(logits, y.long())
if is_train:
optimizer.zero_grad()
loss.backward()
optimizer.step()
total_loss += loss.item()
all_labels.extend(y.cpu().tolist())
if task == "m1":
all_preds.extend(torch.sigmoid(logits).detach().cpu().tolist())
else:
all_preds.extend(torch.argmax(logits, dim=1).detach().cpu().tolist())
n_batches += 1
avg_loss = total_loss / max(n_batches, 1)
return avg_loss, np.array(all_labels), np.array(all_preds)
def _run_epoch_gnn(model, loader, criterion, optimizer, device, is_train: bool, task: str):
"""Run one epoch for GNN+Tab."""
import torch
model.train(is_train)
total_loss = 0.0
all_labels = []
all_preds = []
n_batches = 0
with torch.set_grad_enabled(is_train):
for drug_batch, tab, y in loader:
logits = model(drug_batch, tab)
if task == "m1":
loss = criterion(logits, y.float())
else:
loss = criterion(logits, y.long())
if is_train:
optimizer.zero_grad()
loss.backward()
optimizer.step()
total_loss += loss.item()
all_labels.extend(y.cpu().tolist())
if task == "m1":
all_preds.extend(torch.sigmoid(logits).detach().cpu().tolist())
else:
all_preds.extend(torch.argmax(logits, dim=1).detach().cpu().tolist())
n_batches += 1
avg_loss = total_loss / max(n_batches, 1)
return avg_loss, np.array(all_labels), np.array(all_preds)
def _compute_val_metric(y_true, y_pred, task: str) -> float:
"""Compute primary validation metric."""
if task == "m1":
from negbiodb.metrics import auroc as compute_auroc
if len(np.unique(y_true)) < 2:
return float("nan")
return compute_auroc(y_true, y_pred)
else:
from sklearn.metrics import f1_score
return float(f1_score(y_true.astype(int), y_pred.astype(int), average="macro", zero_division=0))
def train_neural(
model_name: str,
train_df: pd.DataFrame,
val_df: pd.DataFrame,
test_df: pd.DataFrame,
task: str,
output_dir: Path,
args,
) -> dict:
"""Train and evaluate MLP or GNN+Tab."""
import torch
import torch.nn as nn
from negbiodb_ct.ct_models import build_ct_model
# Device
if torch.cuda.is_available():
device = torch.device("cuda")
elif hasattr(torch.backends, "mps") and torch.backends.mps.is_available():
device = torch.device("mps")
else:
device = torch.device("cpu")
logger.info("Device: %s", device)
# Build datasets
if model_name == "mlp":
train_ds = CTTabularDataset(train_df, task)
val_ds = CTTabularDataset(val_df, task)
test_ds = CTTabularDataset(test_df, task)
from torch.utils.data import DataLoader
collate_fn = lambda b: _collate_mlp(b, device) # noqa: E731
train_loader = DataLoader(train_ds, batch_size=args.batch_size, shuffle=True, num_workers=0, collate_fn=collate_fn)
val_loader = DataLoader(val_ds, batch_size=args.batch_size, shuffle=False, num_workers=0, collate_fn=collate_fn)
test_loader = DataLoader(test_ds, batch_size=args.batch_size, shuffle=False, num_workers=0, collate_fn=collate_fn)
run_epoch = _run_epoch_mlp
else:
# GNN: build graph cache
parquet_name = _resolve_m1_parquet(args.dataset, args.negative) if task == "m1" else _M2_PARQUET
parquet_path = args.data_dir / parquet_name
cache_path = args.data_dir / "ct_graph_cache.pt"
graph_cache = _prepare_ct_graph_cache(parquet_path, cache_path)
train_ds = CTGraphDataset(train_df, task, graph_cache)
val_ds = CTGraphDataset(val_df, task, graph_cache)
test_ds = CTGraphDataset(test_df, task, graph_cache)
from torch.utils.data import DataLoader
collate_fn = lambda b: _collate_gnn(b, device) # noqa: E731
train_loader = DataLoader(train_ds, batch_size=args.batch_size, shuffle=True, num_workers=0, collate_fn=collate_fn)
val_loader = DataLoader(val_ds, batch_size=args.batch_size, shuffle=False, num_workers=0, collate_fn=collate_fn)
test_loader = DataLoader(test_ds, batch_size=args.batch_size, shuffle=False, num_workers=0, collate_fn=collate_fn)
run_epoch = _run_epoch_gnn
if len(train_ds) == 0 or len(val_ds) == 0 or len(test_ds) == 0:
raise ValueError("Empty split detected (likely SMILES filtering for GNN)")
# Build model
model = build_ct_model(model_name, task).to(device)
n_params = sum(p.numel() for p in model.parameters() if p.requires_grad)
logger.info("Model: %s | task: %s | params: %d", model_name, task, n_params)
# Loss function
if task == "m1":
criterion = nn.BCEWithLogitsLoss()
else:
# Compute class weights from actual training data
# For GNN, use SMILES-only subset (matches CTGraphDataset filtering)
if model_name == "gnn":
train_labels = train_ds.y
else:
train_labels = train_df["failure_category_int"].values
class_counts = np.bincount(train_labels, minlength=8).astype(np.float64)
# pharmacokinetic (idx 7) may have 0 records → weight 0.0
weights = np.zeros(8, dtype=np.float64)
for c in range(8):
if class_counts[c] > 0:
weights[c] = class_counts.sum() / (8 * class_counts[c])
else:
weights[c] = 0.0
weight_tensor = torch.tensor(weights, dtype=torch.float32).to(device)
criterion = nn.CrossEntropyLoss(weight=weight_tensor)
logger.info("M2 class weights: %s", {k: f"{weights[v]:.2f}" for k, v in CATEGORY_TO_INT.items()})
optimizer = torch.optim.Adam(model.parameters(), lr=args.lr)
# Training loop
best_val_metric = float("-inf")
patience_counter = 0
training_log = []
for epoch in range(1, args.epochs + 1):
train_loss, _, _ = run_epoch(model, train_loader, criterion, optimizer, device, True, task)
val_loss, val_y, val_pred = run_epoch(model, val_loader, criterion, None, device, False, task)
val_metric = _compute_val_metric(val_y, val_pred, task)
metric_name = "val_auroc" if task == "m1" else "val_macro_f1"
row = {"epoch": epoch, "train_loss": train_loss, "val_loss": val_loss, metric_name: val_metric}
training_log.append(row)
logger.info(
"Epoch %3d | train_loss=%.4f | val_loss=%.4f | %s=%.4f",
epoch, train_loss, val_loss, metric_name, val_metric,
)
if not np.isnan(val_metric) and val_metric > best_val_metric:
best_val_metric = val_metric
patience_counter = 0
torch.save(model.state_dict(), output_dir / "best.pt")
logger.info(" Saved best checkpoint (%s=%.4f)", metric_name, best_val_metric)
else:
patience_counter += 1
if patience_counter >= args.patience:
logger.info("Early stopping at epoch %d (patience=%d)", epoch, args.patience)
break
# Save training log
log_path = output_dir / "training_log.csv"
with open(log_path, "w", newline="") as f:
writer = csv.DictWriter(f, fieldnames=training_log[0].keys())
writer.writeheader()
writer.writerows(training_log)
logger.info("Training log saved → %s", log_path)
# Evaluate on test set
checkpoint = output_dir / "best.pt"
if not checkpoint.exists():
logger.error("No checkpoint saved — all epochs produced NaN val metric.")
return {}
state = torch.load(checkpoint, map_location=device, weights_only=True)
model.load_state_dict(state)
_, test_y, test_pred = run_epoch(model, test_loader, criterion, None, device, False, task)
if task == "m1":
test_metrics = compute_m1_metrics(test_y, test_pred)
else:
test_metrics = compute_m2_metrics(test_y.astype(int), test_pred.astype(int))
# Return actual dataset sizes (may differ from df sizes for GNN SMILES filtering)
test_metrics["_n_train"] = len(train_ds)
test_metrics["_n_val"] = len(val_ds)
test_metrics["_n_test"] = len(test_ds)
return test_metrics
# ---------------------------------------------------------------------------
# Main
# ---------------------------------------------------------------------------
def main(argv: list[str] | None = None) -> int:
parser = argparse.ArgumentParser(description="Train CT baseline model.")
parser.add_argument("--model", required=True, choices=["xgboost", "mlp", "gnn"])
parser.add_argument("--task", required=True, choices=["m1", "m2"])
parser.add_argument("--split", required=True,
choices=["random", "cold_drug", "cold_condition",
"temporal", "scaffold", "degree_balanced"])
parser.add_argument("--dataset", default="balanced",
choices=["balanced", "realistic", "smiles_only"])
parser.add_argument("--negative", default="negbiodb",
choices=["negbiodb", "uniform_random", "degree_matched"])
parser.add_argument("--epochs", type=int, default=100)
parser.add_argument("--patience", type=int, default=15)
parser.add_argument("--batch_size", type=int, default=256)
parser.add_argument("--lr", type=float, default=1e-3)
parser.add_argument("--seed", type=int, default=42)
parser.add_argument("--data_dir", type=Path, default=ROOT / "exports" / "ct")
parser.add_argument("--output_dir", type=Path, default=ROOT / "results" / "ct_baselines")
args = parser.parse_args(argv)
set_seed(args.seed)
# Build run name
run_name = f"{args.model}_{args.task}_{args.dataset}_{args.split}_{args.negative}_seed{args.seed}"
out_dir = args.output_dir / run_name
out_dir.mkdir(parents=True, exist_ok=True)
logger.info("Run: %s → %s", run_name, out_dir)
# Load data
try:
train_df, val_df, test_df = load_data(args)
except (ValueError, FileNotFoundError) as e:
logger.error("%s", e)
return 1
if len(train_df) == 0 or len(val_df) == 0 or len(test_df) == 0:
logger.error("Empty split detected.")
return 1
# Train
if args.model == "xgboost":
test_metrics = train_xgboost(train_df, val_df, test_df, args.task, out_dir, args.seed)
else:
test_metrics = train_neural(args.model, train_df, val_df, test_df, args.task, out_dir, args)
if not test_metrics:
logger.warning("No test metrics produced (single-class val/test set?).")
null_metrics = {k: None for k in (
["auroc", "auprc", "mcc", "log_auc", "bedroc", "ef_1pct", "ef_5pct", "accuracy", "f1"]
if args.task == "m1" else
["macro_f1", "weighted_f1", "mcc", "accuracy"]
)}
results = {
"run_name": run_name,
"model": args.model,
"task": args.task,
"split": args.split,
"negative": args.negative,
"dataset": args.dataset,
"seed": args.seed,
"test_metrics": null_metrics,
"n_train": len(train_df),
"n_val": len(val_df),
"n_test": len(test_df),
}
results_path = out_dir / "results.json"
write_results_json(results_path, results)
logger.info("Null results saved → %s", results_path)
return 0
# Save results — use actual dataset sizes from neural models (GNN filters SMILES)
n_train = test_metrics.pop("_n_train", len(train_df))
n_val = test_metrics.pop("_n_val", len(val_df))
n_test = test_metrics.pop("_n_test", len(test_df))
results = {
"run_name": run_name,
"model": args.model,
"task": args.task,
"split": args.split,
"negative": args.negative,
"dataset": args.dataset,
"seed": args.seed,
"test_metrics": test_metrics,
"n_train": n_train,
"n_val": n_val,
"n_test": n_test,
}
results_path = out_dir / "results.json"
write_results_json(results_path, results)
logger.info("Test metrics:")
for k, v in test_metrics.items():
if isinstance(v, (int, float)):
logger.info(" %-20s = %.4f", k, v if np.isfinite(v) else float("nan"))
else:
logger.info(" %-20s = %s", k, str(v)[:80])
logger.info("Results saved → %s", results_path)
return 0
if __name__ == "__main__":
sys.exit(main())