|
|
import torch |
|
|
import csv |
|
|
import pickle |
|
|
import os |
|
|
import numpy as np |
|
|
from rdkit import Chem |
|
|
import pandas as pd |
|
|
from torch_geometric.data import Data |
|
|
from torch_geometric.nn import GCNConv |
|
|
import torch.nn.functional as F |
|
|
from transformers import AutoTokenizer, AutoModel |
|
|
from matplotlib.colors import LinearSegmentedColormap |
|
|
import matplotlib.pyplot as plt |
|
|
from sklearn.preprocessing import MinMaxScaler |
|
|
|
|
|
def load_csv_data(filename): |
|
|
mylist = [] |
|
|
with open(filename) as data: |
|
|
csv_data = csv.reader(data, delimiter=',') |
|
|
next(csv_data) |
|
|
for row in csv_data: |
|
|
if any(row): |
|
|
mylist.append(row) |
|
|
return mylist |
|
|
|
|
|
|
|
|
color1 = (255/255, 102/255, 153/255) |
|
|
custom_cmap_pink = LinearSegmentedColormap.from_list("transparent_to_color",[(0, (*color1, 0)), (1, (*color1, 1))], N=256) |
|
|
|
|
|
color2 = (102/255, 153/255, 255/255) |
|
|
custom_cmap_blue = LinearSegmentedColormap.from_list("transparent_to_color",[(0, (*color2, 0)), (1, (*color2, 1))], N=256) |
|
|
|
|
|
color3 = (250/255, 147/255, 62/255) |
|
|
custom_cmap_orange = LinearSegmentedColormap.from_list("transparent_to_color",[(0, (*color3, 0)), (1, (*color3, 1))], N=256) |
|
|
|
|
|
color4 = (20/255, 165/255, 248/255) |
|
|
custom_cmap_button_blue = LinearSegmentedColormap.from_list("transparent_to_color",[(0, (*color4, 0)), (1, (*color4, 1))], N=256) |
|
|
|
|
|
color5 = (8/255, 253/255, 237/255) |
|
|
custom_cmap_turquoise_blue = LinearSegmentedColormap.from_list("transparent_to_color",[(0, (*color5, 0)), (1, (*color5, 1))], N=256) |
|
|
|
|
|
color6 = (224/255, 36/255, 159/255) |
|
|
custom_cmap_barbie_pink = LinearSegmentedColormap.from_list("transparent_to_color",[(0, (*color6, 0)), (1, (*color6, 1))], N=256) |
|
|
|
|
|
color7 = (80/255, 200/255, 120/255) |
|
|
custom_cmap_emerald_green = LinearSegmentedColormap.from_list("transparent_to_color",[(0, (*color7, 0)), (1, (*color7, 1))], N=256) |
|
|
|
|
|
|
|
|
def check_dir(folder_path): |
|
|
if not os.path.exists(folder_path): |
|
|
os.makedirs(folder_path) |
|
|
print(f"Folder '{folder_path}' created.") |
|
|
else: |
|
|
print(f"Folder '{folder_path}' already exists.") |
|
|
|
|
|
def extract_list_from_pickle(file_path): |
|
|
with open(file_path, 'rb') as file: |
|
|
data_list = pickle.load(file) |
|
|
return data_list |
|
|
|
|
|
def extract_data_from_sdf(file_path): |
|
|
with open(file_path, 'r') as file: |
|
|
lines = file.readlines() |
|
|
|
|
|
mapped_data = [] |
|
|
canonical_smiles_set = set() |
|
|
current_smiles = None |
|
|
current_generic_name = None |
|
|
index = 0 |
|
|
for i, line in enumerate(lines): |
|
|
if line.strip() == '> <SMILES>': |
|
|
current_smiles = lines[i + 1].strip() |
|
|
elif line.strip() == '> <GENERIC_NAME>': |
|
|
current_generic_name = lines[i + 1].strip() |
|
|
elif line.strip() == '$$$$': |
|
|
if current_smiles and current_generic_name: |
|
|
m = Chem.MolFromSmiles(current_smiles) |
|
|
if m is not None: |
|
|
|
|
|
canonical_kekule_smiles = Chem.MolToSmiles(m, canonical=True, kekuleSmiles=True) |
|
|
|
|
|
|
|
|
if canonical_kekule_smiles not in canonical_smiles_set: |
|
|
canonical_smiles_set.add(canonical_kekule_smiles) |
|
|
rdkit_smiles = Chem.MolToSmiles(m, canonical=True, kekuleSmiles=True) |
|
|
|
|
|
|
|
|
mapped_data.append([index, current_smiles, rdkit_smiles, current_generic_name]) |
|
|
print([index, current_smiles, rdkit_smiles, current_generic_name]) |
|
|
index += 1 |
|
|
|
|
|
|
|
|
current_smiles = None |
|
|
current_generic_name = None |
|
|
|
|
|
return mapped_data |
|
|
|
|
|
def extract_smiles_from_txt(file_path): |
|
|
smiles_list = [] |
|
|
try: |
|
|
with open(file_path, 'r') as file: |
|
|
for line in file: |
|
|
smiles = line.strip() |
|
|
if smiles: |
|
|
smiles_list.append(smiles) |
|
|
else: |
|
|
print(f"Empty line found in {file_path}") |
|
|
except FileNotFoundError: |
|
|
print(f"File not found: {file_path}") |
|
|
except IOError: |
|
|
print(f"Error reading file: {file_path}") |
|
|
|
|
|
return smiles_list |
|
|
|
|
|
def one_hot_encoding(value, choices): |
|
|
encoding = [0] * len(choices) |
|
|
encoding[choices.index(value)] = 1 |
|
|
return encoding |
|
|
|
|
|
def smiles_to_graph(smiles): |
|
|
mol = Chem.MolFromSmiles(smiles) |
|
|
if mol is None: |
|
|
return None |
|
|
|
|
|
Chem.Kekulize(mol) |
|
|
adj = Chem.GetAdjacencyMatrix(mol) |
|
|
atom_features = [] |
|
|
for atom in mol.GetAtoms(): |
|
|
atom_features.append(one_hot_encoding(atom.GetAtomicNum(), [1, 3, 4, 5, 6, 7, 8, 9, 11, 12, 13, 14, 15, 16, 17, 19, 20, 22, 23, 24, |
|
|
25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 37, 38, 39, 40, 42, 46, 47, 48, 50, 51, 52, 53, 55, 56, 57, 58, 60, 65, 73, 74, 75, 79, 80, 81, 82, 83, 92])) |
|
|
|
|
|
x = torch.tensor(atom_features, dtype=torch.float) |
|
|
edge_index = torch.tensor(np.where(adj), dtype=torch.long) |
|
|
return Data(x=x, edge_index=edge_index) |
|
|
|
|
|
class GNN(torch.nn.Module): |
|
|
def __init__(self): |
|
|
super(GNN, self).__init__() |
|
|
self.conv1 = GCNConv(58, 64) |
|
|
self.conv2 = GCNConv(64, 128) |
|
|
|
|
|
|
|
|
def forward(self, data): |
|
|
x, edge_index = data.x, data.edge_index |
|
|
x = self.conv1(x, edge_index) |
|
|
x = F.relu(x) |
|
|
x = self.conv2(x, edge_index) |
|
|
return x |
|
|
|
|
|
def atom_molecule_agg(gnn_output): |
|
|
return torch.mean(gnn_output, dim=0) |
|
|
|
|
|
def smiles_ChemBERTa_embedding(smiles_list, model_name='seyonec/PubChem10M_SMILES_BPE_450k'): |
|
|
|
|
|
tokenizer = AutoTokenizer.from_pretrained(model_name) |
|
|
model = AutoModel.from_pretrained(model_name) |
|
|
|
|
|
|
|
|
inputs = tokenizer(smiles_list, return_tensors='pt', padding=True, truncation=False) |
|
|
|
|
|
|
|
|
with torch.no_grad(): |
|
|
outputs = model(**inputs) |
|
|
last_hidden_state = outputs.last_hidden_state |
|
|
|
|
|
|
|
|
attention_mask = inputs['attention_mask'] |
|
|
|
|
|
|
|
|
mean_embeddings = [] |
|
|
for i in range(last_hidden_state.size(0)): |
|
|
mask = attention_mask[i].unsqueeze(-1).expand_as(last_hidden_state[i]) |
|
|
masked_embedding = last_hidden_state[i] * mask |
|
|
sum_embedding = masked_embedding.sum(dim=0) |
|
|
count_non_pad_tokens = attention_mask[i].sum() |
|
|
mean_embedding = sum_embedding / count_non_pad_tokens |
|
|
mean_embeddings.append(mean_embedding) |
|
|
|
|
|
mean_embeddings = torch.stack(mean_embeddings) |
|
|
|
|
|
|
|
|
final_mean_embedding = mean_embeddings.mean(dim=0) |
|
|
return final_mean_embedding |
|
|
|
|
|
def process_data(data_list): |
|
|
try: |
|
|
|
|
|
generic_name, smiles, confidence, permeability_class, chemberta_embeddings, permeability_values = [], [], [], [], [], [] |
|
|
|
|
|
|
|
|
for i, item in enumerate(data_list): |
|
|
if len(item) < 6: |
|
|
raise IndexError(f"Item at index {i} does not have enough elements: {item}") |
|
|
|
|
|
|
|
|
generic_name.append(item[0]) |
|
|
smiles.append(item[1]) |
|
|
confidence.append(item[2]) |
|
|
permeability_class.append(item[3]) |
|
|
chemberta_embeddings.append(item[4]) |
|
|
permeability_values.append(item[5]) |
|
|
|
|
|
|
|
|
stacked_tensor = torch.stack(chemberta_embeddings, dim=0) |
|
|
|
|
|
|
|
|
return generic_name, smiles, confidence, permeability_class, stacked_tensor, permeability_values |
|
|
|
|
|
except IndexError as e: |
|
|
print(f"IndexError: {e}. Check the structure and content of the data list.") |
|
|
return None, None, None |
|
|
except Exception as e: |
|
|
print(f"An unexpected error occurred: {e}") |
|
|
return None, None, None |
|
|
|
|
|
def take_unique_tensors(list_of_tensors): |
|
|
unique_tensors = [] |
|
|
for i, tensor in enumerate(list_of_tensors): |
|
|
if not any(torch.equal(tensor, unique_tensor) for unique_tensor in unique_tensors): |
|
|
unique_tensors.append(tensor) |
|
|
else: |
|
|
print(f"Duplicate found at index: {i}") |
|
|
|
|
|
|
|
|
print(f"Number of unique tensors: {len(unique_tensors)}") |
|
|
return unique_tensors |
|
|
|
|
|
def process_data_old(data_list): |
|
|
try: |
|
|
|
|
|
original_smiles, rdkit_smiles, chemberta_embeddings = [], [], [] |
|
|
|
|
|
|
|
|
for i, item in enumerate(data_list): |
|
|
if len(item) < 3: |
|
|
raise IndexError(f"Item at index {i} does not have enough elements: {item}") |
|
|
|
|
|
original_smiles.append(item[1]) |
|
|
rdkit_smiles.append(item[2]) |
|
|
chemberta_embeddings.append(item[3]) |
|
|
|
|
|
|
|
|
stacked_tensor = torch.stack(chemberta_embeddings, dim=0) |
|
|
|
|
|
|
|
|
return original_smiles, rdkit_smiles, stacked_tensor |
|
|
|
|
|
except IndexError as e: |
|
|
print(f"IndexError: {e}. Check the structure and content of the data list.") |
|
|
return None, None, None |
|
|
except Exception as e: |
|
|
print(f"An unexpected error occurred: {e}") |
|
|
return None, None, None |
|
|
|
|
|
def process_data1(data_list): |
|
|
try: |
|
|
|
|
|
index, generic_name, smiles, confidence, permeability_class, chemberta_embeddings, permeability_values = [], [], [], [], [], [], [] |
|
|
|
|
|
for i, item in enumerate(data_list): |
|
|
if len(item) < 6: |
|
|
raise IndexError(f"Item at index {i} does not have enough elements: {item}") |
|
|
index.append(item[0]) |
|
|
generic_name.append(item[1]) |
|
|
smiles.append(item[2]) |
|
|
confidence.append(item[3]) |
|
|
permeability_class.append(item[4]) |
|
|
chemberta_embeddings.append(item[5]) |
|
|
permeability_values.append(item[6]) |
|
|
|
|
|
stacked_tensor = torch.stack(chemberta_embeddings, dim=0) |
|
|
|
|
|
return index, generic_name, smiles, confidence, permeability_class, stacked_tensor, permeability_values |
|
|
except IndexError as e: |
|
|
print(f"IndexError: {e}. Check the structure and content of the data list.") |
|
|
return None, None, None |
|
|
except Exception as e: |
|
|
print(f"An unexpected error occurred: {e}") |
|
|
return None, None, None |
|
|
|
|
|
def sort_by_distance_with_names_and_permeability(ref_points, data_points, sorted_names, normalized_permeability): |
|
|
ref_points = np.expand_dims(ref_points, axis=1) |
|
|
distances = np.linalg.norm(data_points - ref_points, axis=2) |
|
|
min_distances = np.min(distances, axis=0) |
|
|
sorted_indices = np.argsort(min_distances) |
|
|
|
|
|
sorted_points = data_points[sorted_indices] |
|
|
sorted_distances = min_distances[sorted_indices] |
|
|
new_sorted_names = [sorted_names[i] for i in sorted_indices] |
|
|
sorted_permeability = [normalized_permeability[i] for i in sorted_indices] |
|
|
return sorted_points, sorted_distances, new_sorted_names, sorted_permeability |
|
|
|
|
|
def calculate_distances(ref_points, data_points, cutoff): |
|
|
ref_points = np.expand_dims(ref_points, axis=1) |
|
|
distances = np.linalg.norm(data_points - ref_points, axis=2) |
|
|
mask = distances < cutoff |
|
|
return data_points[np.any(mask, axis=0)] |
|
|
|
|
|
def sort_by_distance_with_names_and_permeability(ref_points, data_points, all_sorted_names, normalized_permeability, all_sorted_smiles): |
|
|
ref_points = np.expand_dims(ref_points, axis=1) |
|
|
distances = np.linalg.norm(data_points - ref_points, axis=2) |
|
|
min_distances = np.min(distances, axis=0) |
|
|
sorted_indices = np.argsort(min_distances) |
|
|
|
|
|
sorted_points = data_points[sorted_indices] |
|
|
sorted_distances = min_distances[sorted_indices] |
|
|
new_sorted_names = [all_sorted_names[i] for i in sorted_indices] |
|
|
sorted_permeability = [normalized_permeability[i] for i in sorted_indices] |
|
|
sorted_smiles_string = [all_sorted_smiles[i] for i in sorted_indices] |
|
|
return sorted_points, sorted_distances, new_sorted_names, sorted_permeability, sorted_smiles_string |
|
|
|
|
|
def calculate_distances(ref_points, data_points, cutoff): |
|
|
ref_points = np.expand_dims(ref_points, axis=1) |
|
|
distances = np.linalg.norm(data_points - ref_points, axis=2) |
|
|
mask = distances < cutoff |
|
|
return data_points[np.any(mask, axis=0)] |
|
|
|
|
|
|