NegBioDB / src /negbiodb_ct /ct_models.py
jang1563's picture
NegBioDB final: 4 domains, fully audited
6d1bbc7
"""CT domain ML model definitions.
Models:
CT_MLP: MLP for tabular features (CT-M1 binary, CT-M2 multiclass)
CT_GNN_Tab: GNN for drug graph + tabular condition/trial features
Reuses _GCNDrugEncoder and smiles_to_graph from negbiodb.models.graphdta.
"""
from __future__ import annotations
import warnings
import torch
import torch.nn as nn
import torch.nn.functional as F
from negbiodb_ct.ct_features import (
CONDITION_DIM,
DRUG_TAB_DIM,
M2_TRIAL_DIM,
TOTAL_M1_DIM,
TOTAL_M2_DIM,
)
try:
from torch_geometric.data import Batch, Data
from torch_geometric.nn import GCNConv, global_max_pool
HAS_TORCH_GEOMETRIC = True
except ImportError:
HAS_TORCH_GEOMETRIC = False
warnings.warn(
"torch_geometric not found. CT_GNN_Tab requires `pip install negbiodb[ml]`.",
stacklevel=1,
)
# Import GCN drug encoder from DTI domain
from negbiodb.models.graphdta import NODE_FEATURE_DIM, _GCNDrugEncoder
# ---------------------------------------------------------------------------
# CT_MLP
# ---------------------------------------------------------------------------
class CT_MLP(nn.Module):
"""MLP for CT-M1 (binary) or CT-M2 (multiclass classification).
Args:
input_dim: Input feature dimension (1044 for M1, 1066 for M2).
num_classes: 1 for binary (M1), 8 for multiclass (M2).
hidden_dims: Tuple of hidden layer sizes.
dropout: Dropout rate.
"""
def __init__(
self,
input_dim: int,
num_classes: int = 1,
hidden_dims: tuple[int, ...] = (512, 256),
dropout: float = 0.3,
) -> None:
super().__init__()
self.num_classes = num_classes
layers: list[nn.Module] = []
in_dim = input_dim
for h_dim in hidden_dims:
layers.extend([nn.Linear(in_dim, h_dim), nn.ReLU(), nn.Dropout(dropout)])
in_dim = h_dim
layers.append(nn.Linear(in_dim, num_classes))
self.fc = nn.Sequential(*layers)
def forward(self, x: torch.Tensor) -> torch.Tensor:
"""
Args:
x: (B, input_dim) feature tensor.
Returns:
(B,) raw logits for binary, (B, num_classes) for multiclass.
"""
out = self.fc(x)
if self.num_classes == 1:
return out.squeeze(-1) # (B,)
return out # (B, num_classes)
# ---------------------------------------------------------------------------
# CT_GNN_Tab
# ---------------------------------------------------------------------------
class CT_GNN_Tab(nn.Module):
"""GNN encoder for drug molecular graph + tabular condition/trial features.
Drug: _GCNDrugEncoder (3-layer GCN, 128-dim) → 128-dim graph embedding
Tab: FC(tab_dim → 64) → ReLU
Concat: (128 + 64) = 192 → FC(256) → ReLU → Dropout → FC(128) → ReLU → Dropout → FC(out)
Args:
tab_dim: Tabular feature dimension (14 for M1, 36 for M2).
num_classes: 1 for binary, 8 for multiclass.
gnn_hidden: GCN hidden dimension (128, smaller than DTI's 256).
fc_dims: FC layer sizes after concat.
dropout: Dropout rate.
"""
def __init__(
self,
tab_dim: int,
num_classes: int = 1,
gnn_hidden: int = 128,
fc_dims: tuple[int, ...] = (256, 128),
dropout: float = 0.3,
) -> None:
super().__init__()
if not HAS_TORCH_GEOMETRIC:
raise RuntimeError("torch_geometric required for CT_GNN_Tab.")
self.num_classes = num_classes
# Drug graph encoder (reuse DTI GCN, smaller hidden)
self.drug_encoder = _GCNDrugEncoder(NODE_FEATURE_DIM, gnn_hidden)
# Tabular encoder
self.tab_encoder = nn.Sequential(
nn.Linear(tab_dim, 64),
nn.ReLU(),
)
# Classification head
concat_dim = gnn_hidden + 64
layers: list[nn.Module] = []
in_dim = concat_dim
for f_dim in fc_dims:
layers.extend([nn.Linear(in_dim, f_dim), nn.ReLU(), nn.Dropout(dropout)])
in_dim = f_dim
layers.append(nn.Linear(in_dim, num_classes))
self.fc = nn.Sequential(*layers)
def forward(
self,
drug_graph: "Batch",
tab_features: torch.Tensor,
) -> torch.Tensor:
"""
Args:
drug_graph: PyG Batch of molecular graphs (x, edge_index, batch).
tab_features: (B, tab_dim) tabular features.
Returns:
(B,) raw logits for binary, (B, num_classes) for multiclass.
"""
d = self.drug_encoder(drug_graph.x, drug_graph.edge_index, drug_graph.batch)
t = self.tab_encoder(tab_features)
h = torch.cat([d, t], dim=1)
out = self.fc(h)
if self.num_classes == 1:
return out.squeeze(-1)
return out
# ---------------------------------------------------------------------------
# Factory
# ---------------------------------------------------------------------------
# Tab dims for GNN+Tab
GNN_TAB_DIM_M1 = DRUG_TAB_DIM + CONDITION_DIM # 13 + 1 = 14
GNN_TAB_DIM_M2 = GNN_TAB_DIM_M1 + M2_TRIAL_DIM # 14 + 22 = 36
def build_ct_model(
model_name: str,
task: str = "m1",
num_classes: int | None = None,
**kwargs,
) -> nn.Module:
"""Factory function to create CT models.
Args:
model_name: "mlp" or "gnn"
task: "m1" or "m2"
num_classes: override (default: 1 for M1, 8 for M2)
**kwargs: passed to model constructor
Returns:
nn.Module instance
"""
if num_classes is None:
num_classes = 1 if task == "m1" else 8
if model_name == "mlp":
input_dim = TOTAL_M1_DIM if task == "m1" else TOTAL_M2_DIM
return CT_MLP(input_dim=input_dim, num_classes=num_classes, **kwargs)
elif model_name == "gnn":
tab_dim = GNN_TAB_DIM_M1 if task == "m1" else GNN_TAB_DIM_M2
return CT_GNN_Tab(tab_dim=tab_dim, num_classes=num_classes, **kwargs)
else:
raise ValueError(f"Unknown model: {model_name!r}. Choose 'mlp' or 'gnn'.")