File size: 6,234 Bytes
6d1bbc7
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
"""GraphDTA: Drug-Target Affinity Prediction via Graph Neural Networks.

Reference: Nguyen et al., 2020 (arXiv:2003.06751).
Architecture: GCN for drug molecular graph + 1D-CNN for protein sequence.

Requires: torch-geometric (install with `pip install negbiodb[ml]`).
"""

from __future__ import annotations

import warnings

import torch
import torch.nn as nn
import torch.nn.functional as F

from negbiodb.models.deepdta import (
    AA_VOCAB_SIZE,
    MAX_SEQ_LEN,
    _CNNEncoder,
    seq_to_tensor,
)

try:
    from torch_geometric.data import Batch, Data
    from torch_geometric.nn import GCNConv, global_max_pool

    HAS_TORCH_GEOMETRIC = True
except ImportError:
    HAS_TORCH_GEOMETRIC = False
    warnings.warn(
        "torch_geometric not found. GraphDTA and DrugBAN require `pip install negbiodb[ml]`.",
        stacklevel=1,
    )

# --- Atom feature constants --------------------------------------------------
# One-hot feature sizes (must match smiles_to_graph)
ATOM_TYPES = [
    "C", "N", "O", "S", "F", "Si", "P", "Cl", "Br", "Mg", "Na", "Ca", "Fe",
    "As", "Al", "I", "B", "V", "K", "Tl", "Yb", "Sb", "Sn", "Ag", "Pd",
    "Co", "Se", "Ti", "Zn", "H", "Li", "Ge", "Cu", "Au", "Ni", "Cd", "In",
    "Mn", "Zr", "Cr", "Pt", "Hg", "Pb", "Unknown",
]  # 44 atom types
ATOM_DEGREES = list(range(11))  # 0-10 → 11 values
ATOM_H_COUNTS = list(range(11))  # 0-10 → 11 values
ATOM_VALENCES = list(range(11))  # 0-10 → 11 values
# _one_hot returns len(choices)+1 (includes unknown bucket).
# Total: (44+1) + (11+1) + (11+1) + (11+1) + 1 (aromaticity) = 82
NODE_FEATURE_DIM = 82


def _one_hot(value: int | str, choices: list) -> list[int]:
    """Return one-hot vector; last element is 1 if value not in choices."""
    vec = [0] * (len(choices) + 1)
    idx = choices.index(value) if value in choices else len(choices)
    vec[idx] = 1
    return vec


def smiles_to_graph(smiles: str) -> "Data | None":
    """Convert a SMILES string to a PyG Data object.

    Returns None if the SMILES is invalid.

    Node features (82-dim):
        (44+1) atom type + (11+1) degree + (11+1) H-count + (11+1) valence + 1 aromaticity
        Each _one_hot adds +1 unknown bucket, giving 82 total.
    """
    if not HAS_TORCH_GEOMETRIC:
        raise RuntimeError("torch_geometric required for GraphDTA.")
    try:
        from rdkit import Chem
    except ImportError as e:
        raise RuntimeError("rdkit required for graph construction.") from e

    mol = Chem.MolFromSmiles(smiles)
    if mol is None:
        return None

    node_features = []
    for atom in mol.GetAtoms():
        feat = (
            _one_hot(atom.GetSymbol(), ATOM_TYPES)
            + _one_hot(atom.GetDegree(), ATOM_DEGREES)
            + _one_hot(atom.GetTotalNumHs(), ATOM_H_COUNTS)
            + _one_hot(atom.GetImplicitValence(), ATOM_VALENCES)
            + [int(atom.GetIsAromatic())]
        )
        node_features.append(feat)

    if len(node_features) == 0:
        return None  # Mol with no atoms — treat as invalid

    x = torch.tensor(node_features, dtype=torch.float)  # (N_atoms, 82)

    edges = []
    for bond in mol.GetBonds():
        i, j = bond.GetBeginAtomIdx(), bond.GetEndAtomIdx()
        edges += [[i, j], [j, i]]  # undirected

    if edges:
        edge_index = torch.tensor(edges, dtype=torch.long).t().contiguous()
    else:
        edge_index = torch.zeros((2, 0), dtype=torch.long)

    return Data(x=x, edge_index=edge_index)


class _GCNDrugEncoder(nn.Module):
    """3-layer GCN encoder for molecular graphs."""

    def __init__(self, in_dim: int = NODE_FEATURE_DIM, hidden: int = 256) -> None:
        super().__init__()
        if not HAS_TORCH_GEOMETRIC:
            raise RuntimeError("torch_geometric required.")
        self.conv1 = GCNConv(in_dim, hidden)
        self.conv2 = GCNConv(hidden, hidden)
        self.conv3 = GCNConv(hidden, hidden)

    def forward(self, x: torch.Tensor, edge_index: torch.Tensor, batch: torch.Tensor) -> torch.Tensor:
        h = F.relu(self.conv1(x, edge_index))
        h = F.relu(self.conv2(h, edge_index))
        h = F.relu(self.conv3(h, edge_index))
        return global_max_pool(h, batch)  # (B, hidden)


class GraphDTA(nn.Module):
    """GraphDTA model for binary DTI prediction (GCN variant).

    Args:
        gnn_hidden: Hidden dimension for GCN layers.
        target_embed_dim: Embedding dimension for amino acid characters.
        target_filters: Filter counts for target 1D-CNN.
        target_kernels: Kernel sizes for target 1D-CNN.
        fc_dims: Fully-connected layer sizes after concat.
        dropout: Dropout rate.
    """

    def __init__(
        self,
        gnn_hidden: int = 256,
        target_embed_dim: int = 128,
        target_filters: tuple[int, int, int] = (32, 64, 96),
        target_kernels: tuple[int, int, int] = (4, 8, 12),
        fc_dims: tuple[int, int] = (1024, 512),
        dropout: float = 0.2,
    ) -> None:
        super().__init__()
        if not HAS_TORCH_GEOMETRIC:
            raise RuntimeError("torch_geometric required for GraphDTA.")
        self.drug_encoder = _GCNDrugEncoder(NODE_FEATURE_DIM, gnn_hidden)
        self.target_encoder = _CNNEncoder(
            AA_VOCAB_SIZE, target_embed_dim, target_filters, target_kernels
        )

        target_out_dim = target_filters[-1]
        concat_dim = gnn_hidden + target_out_dim

        fc_layers: list[nn.Module] = []
        in_dim = concat_dim
        for out_dim in fc_dims:
            fc_layers += [nn.Linear(in_dim, out_dim), nn.ReLU(), nn.Dropout(dropout)]
            in_dim = out_dim
        fc_layers.append(nn.Linear(in_dim, 1))
        self.fc = nn.Sequential(*fc_layers)

    def forward(
        self,
        drug_graph: "Batch",
        target_tokens: torch.Tensor,
    ) -> torch.Tensor:
        """
        Args:
            drug_graph: PyG Batch of molecular graphs (x, edge_index, batch).
            target_tokens: (B, MAX_SEQ_LEN) integer token ids.

        Returns:
            (B,) raw logits.
        """
        d = self.drug_encoder(drug_graph.x, drug_graph.edge_index, drug_graph.batch)
        t = self.target_encoder(target_tokens)
        h = torch.cat([d, t], dim=1)
        return self.fc(h).squeeze(-1)