File size: 13,802 Bytes
521547b
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
import torch
import csv
import pickle
import os
import numpy as np
from rdkit import Chem
import pandas as pd
from torch_geometric.data import Data
from torch_geometric.nn import GCNConv
import torch.nn.functional as F
from transformers import AutoTokenizer, AutoModel
from matplotlib.colors import LinearSegmentedColormap
import matplotlib.pyplot as plt
from sklearn.preprocessing import MinMaxScaler

def load_csv_data(filename):
    mylist = []
    with open(filename) as data:
        csv_data = csv.reader(data, delimiter=',')
        next(csv_data)  # Skip header row
        for row in csv_data:
            if any(row):  # Check if any element in the row is non-empty
                mylist.append(row)
    return mylist

# Define the color (255, 102, 153) in RGB, and add transparency
color1 = (255/255, 102/255, 153/255)  # Normalize RGB values to [0, 1]
custom_cmap_pink = LinearSegmentedColormap.from_list("transparent_to_color",[(0, (*color1, 0)), (1, (*color1, 1))], N=256)

color2 = (102/255, 153/255, 255/255)  # Normalize RGB values to [0, 1]
custom_cmap_blue = LinearSegmentedColormap.from_list("transparent_to_color",[(0, (*color2, 0)), (1, (*color2, 1))], N=256)

color3 = (250/255, 147/255, 62/255)  # Normalize RGB values to [0, 1]
custom_cmap_orange = LinearSegmentedColormap.from_list("transparent_to_color",[(0, (*color3, 0)), (1, (*color3, 1))], N=256)

color4 = (20/255, 165/255, 248/255)  # Normalize RGB values to [0, 1]
custom_cmap_button_blue = LinearSegmentedColormap.from_list("transparent_to_color",[(0, (*color4, 0)), (1, (*color4, 1))], N=256)

color5 = (8/255, 253/255, 237/255)  # Normalize RGB values to [0, 1]
custom_cmap_turquoise_blue = LinearSegmentedColormap.from_list("transparent_to_color",[(0, (*color5, 0)), (1, (*color5, 1))], N=256)

color6 = (224/255, 36/255, 159/255)  # Normalize RGB values to [0, 1]
custom_cmap_barbie_pink = LinearSegmentedColormap.from_list("transparent_to_color",[(0, (*color6, 0)), (1, (*color6, 1))], N=256)

color7 = (80/255, 200/255, 120/255)  # Normalize RGB values to [0, 1]
custom_cmap_emerald_green = LinearSegmentedColormap.from_list("transparent_to_color",[(0, (*color7, 0)), (1, (*color7, 1))], N=256)


def check_dir(folder_path):
    if not os.path.exists(folder_path):
        os.makedirs(folder_path)
        print(f"Folder '{folder_path}' created.")
    else:
        print(f"Folder '{folder_path}' already exists.")

def extract_list_from_pickle(file_path):
    with open(file_path, 'rb') as file:
        data_list = pickle.load(file)
    return data_list

def extract_data_from_sdf(file_path):
    with open(file_path, 'r') as file:
        lines = file.readlines()
    
    mapped_data = []
    canonical_smiles_set = set()  # Set to track canonical Kekulé SMILES and avoid duplicates
    current_smiles = None
    current_generic_name = None
    index = 0  # Initialize a new index for the extracted data
    for i, line in enumerate(lines):
        if line.strip() == '> <SMILES>':
            current_smiles = lines[i + 1].strip()
        elif line.strip() == '> <GENERIC_NAME>':
            current_generic_name = lines[i + 1].strip()
        elif line.strip() == '$$$$':
            if current_smiles and current_generic_name:
                m = Chem.MolFromSmiles(current_smiles)
                if m is not None:
                    # Get the canonical SMILES with Kekulé form
                    canonical_kekule_smiles = Chem.MolToSmiles(m, canonical=True, kekuleSmiles=True)
                    
                    # Check if the canonical Kekulé SMILES is already in the set
                    if canonical_kekule_smiles not in canonical_smiles_set:
                        canonical_smiles_set.add(canonical_kekule_smiles)  # Add to the set if it's new
                        rdkit_smiles = Chem.MolToSmiles(m, canonical=True, kekuleSmiles=True)
                        
                        # Append the index along with the other data
                        mapped_data.append([index, current_smiles, rdkit_smiles, current_generic_name])
                        print([index, current_smiles, rdkit_smiles, current_generic_name])
                        index += 1  # Increment the index for each valid data entry
            
            # Reset for the next entry
            current_smiles = None
            current_generic_name = None
    
    return mapped_data

def extract_smiles_from_txt(file_path):
    smiles_list = []
    try:
        with open(file_path, 'r') as file:
            for line in file:
                smiles = line.strip()  # Remove any leading/trailing whitespace
                if smiles:  # Check if the line is not empty
                    smiles_list.append(smiles)
                else:
                    print(f"Empty line found in {file_path}")
    except FileNotFoundError:
        print(f"File not found: {file_path}")
    except IOError:
        print(f"Error reading file: {file_path}")
    
    return smiles_list

def one_hot_encoding(value, choices):
    encoding = [0] * len(choices)
    encoding[choices.index(value)] = 1
    return encoding

def smiles_to_graph(smiles):
    mol = Chem.MolFromSmiles(smiles)
    if mol is None:
        return None

    Chem.Kekulize(mol)
    adj = Chem.GetAdjacencyMatrix(mol)
    atom_features = []
    for atom in mol.GetAtoms():
        atom_features.append(one_hot_encoding(atom.GetAtomicNum(), [1, 3, 4, 5, 6, 7, 8, 9, 11, 12, 13, 14, 15, 16, 17, 19, 20, 22, 23, 24, 
        25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 37, 38, 39, 40, 42, 46, 47, 48, 50, 51, 52, 53, 55, 56, 57, 58, 60, 65, 73, 74, 75, 79, 80, 81, 82, 83, 92]))
    
    x = torch.tensor(atom_features, dtype=torch.float)
    edge_index = torch.tensor(np.where(adj), dtype=torch.long)
    return Data(x=x, edge_index=edge_index)

class GNN(torch.nn.Module):
    def __init__(self):
        super(GNN, self).__init__()
        self.conv1 = GCNConv(58, 64)
        self.conv2 = GCNConv(64, 128)
    # as atoms present in the mcdb is 51, updated the model from 48 to 64.
    # checks the claude_model2, where the model takes in leng of atoms list. 
    def forward(self, data):
        x, edge_index = data.x, data.edge_index
        x = self.conv1(x, edge_index)
        x = F.relu(x)
        x = self.conv2(x, edge_index)
        return x

def atom_molecule_agg(gnn_output):
    return torch.mean(gnn_output, dim=0)

def smiles_ChemBERTa_embedding(smiles_list, model_name='seyonec/PubChem10M_SMILES_BPE_450k'):
    # Load the tokenizer and model
    tokenizer = AutoTokenizer.from_pretrained(model_name)
    model = AutoModel.from_pretrained(model_name)

    # Tokenize the SMILES strings
    inputs = tokenizer(smiles_list, return_tensors='pt', padding=True, truncation=False)

    # Get the model outputs
    with torch.no_grad():
        outputs = model(**inputs)
    last_hidden_state = outputs.last_hidden_state

    # Get the attention mask to identify the padding tokens
    attention_mask = inputs['attention_mask']

    # Calculate the mean embedding for each sequence, ignoring the padding tokens
    mean_embeddings = []
    for i in range(last_hidden_state.size(0)):
        mask = attention_mask[i].unsqueeze(-1).expand_as(last_hidden_state[i])
        masked_embedding = last_hidden_state[i] * mask
        sum_embedding = masked_embedding.sum(dim=0)
        count_non_pad_tokens = attention_mask[i].sum()
        mean_embedding = sum_embedding / count_non_pad_tokens
        mean_embeddings.append(mean_embedding)

    mean_embeddings = torch.stack(mean_embeddings)

    # Calculate the final mean embedding across all sequences
    final_mean_embedding = mean_embeddings.mean(dim=0)
    return final_mean_embedding

def process_data(data_list):
    try:
        # Initialize lists
        generic_name, smiles, confidence, permeability_class, chemberta_embeddings, permeability_values = [], [], [], [], [], []

        # Process each item in the data list
        for i, item in enumerate(data_list):
            if len(item) < 6:
                raise IndexError(f"Item at index {i} does not have enough elements: {item}")
            
            #index.append(item[0])
            generic_name.append(item[0])
            smiles.append(item[1])
            confidence.append(item[2])
            permeability_class.append(item[3])
            chemberta_embeddings.append(item[4])
            permeability_values.append(item[5])

        # Stack the embeddings into a tensor
        stacked_tensor = torch.stack(chemberta_embeddings, dim=0)

        # Return processed data
        return generic_name, smiles, confidence, permeability_class, stacked_tensor, permeability_values

    except IndexError as e:
        print(f"IndexError: {e}. Check the structure and content of the data list.")
        return None, None, None
    except Exception as e:
        print(f"An unexpected error occurred: {e}")
        return None, None, None

def take_unique_tensors(list_of_tensors):
    unique_tensors = []
    for i, tensor in enumerate(list_of_tensors):  # Use enumerate to get index
        if not any(torch.equal(tensor, unique_tensor) for unique_tensor in unique_tensors):
            unique_tensors.append(tensor)
        else:
            print(f"Duplicate found at index: {i}")
    
    # Print the number of unique tensors
    print(f"Number of unique tensors: {len(unique_tensors)}")
    return unique_tensors

def process_data_old(data_list):
    try:
        # Initialize lists
        original_smiles, rdkit_smiles, chemberta_embeddings = [], [], []

        # Process each item in the data list
        for i, item in enumerate(data_list):
            if len(item) < 3:
                raise IndexError(f"Item at index {i} does not have enough elements: {item}")
            
            original_smiles.append(item[1])
            rdkit_smiles.append(item[2])
            chemberta_embeddings.append(item[3])

        # Stack the embeddings into a tensor
        stacked_tensor = torch.stack(chemberta_embeddings, dim=0)

        # Return processed data
        return original_smiles, rdkit_smiles, stacked_tensor

    except IndexError as e:
        print(f"IndexError: {e}. Check the structure and content of the data list.")
        return None, None, None
    except Exception as e:
        print(f"An unexpected error occurred: {e}")
        return None, None, None
    
def process_data1(data_list):
    try:
        # Initialize lists
        index, generic_name, smiles, confidence, permeability_class, chemberta_embeddings, permeability_values = [], [], [], [], [], [], []
        # Process each item in the data list
        for i, item in enumerate(data_list):
            if len(item) < 6:
                raise IndexError(f"Item at index {i} does not have enough elements: {item}")
            index.append(item[0])
            generic_name.append(item[1])
            smiles.append(item[2])
            confidence.append(item[3])
            permeability_class.append(item[4])
            chemberta_embeddings.append(item[5])
            permeability_values.append(item[6])
        # Stack the embeddings into a tensor
        stacked_tensor = torch.stack(chemberta_embeddings, dim=0)
        # Return processed data
        return index, generic_name, smiles, confidence, permeability_class, stacked_tensor, permeability_values
    except IndexError as e:
        print(f"IndexError: {e}. Check the structure and content of the data list.")
        return None, None, None
    except Exception as e:
        print(f"An unexpected error occurred: {e}")
        return None, None, None
    
def sort_by_distance_with_names_and_permeability(ref_points, data_points, sorted_names, normalized_permeability):
    ref_points = np.expand_dims(ref_points, axis=1)  
    distances = np.linalg.norm(data_points - ref_points, axis=2)  
    min_distances = np.min(distances, axis=0)
    sorted_indices = np.argsort(min_distances)
    # Sort data points, distances, and names using the sorted indices
    sorted_points = data_points[sorted_indices]
    sorted_distances = min_distances[sorted_indices]
    new_sorted_names = [sorted_names[i] for i in sorted_indices]  # Apply sorting to names
    sorted_permeability = [normalized_permeability[i] for i in sorted_indices]  # Apply sorting to permeability
    return sorted_points, sorted_distances, new_sorted_names, sorted_permeability

def calculate_distances(ref_points, data_points, cutoff):
    ref_points = np.expand_dims(ref_points, axis=1)
    distances = np.linalg.norm(data_points - ref_points, axis=2)
    mask = distances < cutoff
    return data_points[np.any(mask, axis=0)]

def sort_by_distance_with_names_and_permeability(ref_points, data_points, all_sorted_names, normalized_permeability, all_sorted_smiles):
    ref_points = np.expand_dims(ref_points, axis=1)  
    distances = np.linalg.norm(data_points - ref_points, axis=2)  
    min_distances = np.min(distances, axis=0)
    sorted_indices = np.argsort(min_distances)
    # Sort data points, distances, and names using the sorted indices
    sorted_points = data_points[sorted_indices]
    sorted_distances = min_distances[sorted_indices]
    new_sorted_names = [all_sorted_names[i] for i in sorted_indices]  # Apply sorting to names
    sorted_permeability = [normalized_permeability[i] for i in sorted_indices]  # Apply sorting to permeability
    sorted_smiles_string = [all_sorted_smiles[i] for i in sorted_indices]
    return sorted_points, sorted_distances, new_sorted_names, sorted_permeability, sorted_smiles_string

def calculate_distances(ref_points, data_points, cutoff):
    ref_points = np.expand_dims(ref_points, axis=1)
    distances = np.linalg.norm(data_points - ref_points, axis=2)
    mask = distances < cutoff
    return data_points[np.any(mask, axis=0)]