TeLLAgent / tool /comget /generator.py
jinysun's picture
Update tool/comget/generator.py
f37d307 verified
# -*- coding: utf-8 -*-
import pandas as pd
import math
from tqdm import tqdm
import argparse
from .model import GPT, GPTConfig
import torch
import numpy as np
import re
import json
from rdkit.Chem import RDConfig
from torch.nn import functional as F
import selfies as sf
import os
import sys
sys.path.append(os.path.join(RDConfig.RDContribDir, 'SA_Score'))
from rdkit import Chem
import os
import os
import torch
torch.classes.__path__ = [os.path.join(torch.__path__[0], torch.classes.__file__)]
def get_mol(smiles_or_mol):
'''
Loads SMILES/molecule into RDKit's object
'''
if isinstance(smiles_or_mol, str):
if len(smiles_or_mol) == 0:
return None
mol = Chem.MolFromSmiles(smiles_or_mol)
if mol is None:
return None
try:
Chem.SanitizeMol(mol)
except ValueError:
return None
return mol
return smiles_or_mol
def top_k_logits(logits, k):
v, ix = torch.topk(logits, k)
out = logits.clone()
out[out < v[:, [-1]]] = -float('Inf')
return out
def sample(model, x, steps, temperature=1.0, sample=False, top_k=None, prop = None, scaffold = None):
"""
take a conditioning sequence of indices in x (of shape (b,t)) and predict the next token in
the sequence, feeding the predictions back into the model each time. Clearly the sampling
has quadratic complexity unlike an RNN that is only linear, and has a finite context window
of block_size, unlike an RNN that has an infinite context window.
"""
block_size = model.get_block_size()
model.eval()
for k in range(steps):
x_cond = x if x.size(1) <= block_size else x[:, -block_size:] # crop context if needed
logits, _, _ = model(x_cond, prop = prop, scaffold = scaffold) # for liggpt
# logits, _, _ = model(x_cond) # for char_rnn
# pluck the logits at the final step and scale by temperature
logits = logits[:, -1, :] / temperature
# optionally crop probabilities to only the top k options
if top_k is not None:
logits = top_k_logits(logits, top_k)
# apply softmax to convert to probabilities
probs = F.softmax(logits, dim=-1)
# sample from the distribution or take the most likely
if sample:
ix = torch.multinomial(probs, num_samples=1)
else:
_, ix = torch.topk(probs, k=1, dim=-1)
# append to the sequence and continue
x = torch.cat((x, ix), dim=1)
return x
def get_selfie_and_smiles_encodings_for_dataset(smiles):
"""
Returns encoding, alphabet and length of largest molecule in SMILES and
SELFIES, given a file containing SMILES molecules.
input:
csv file with molecules. Column's name must be 'smiles'.
output:
- selfies encoding
- selfies alphabet
- longest selfies string
- smiles encoding (equivalent to file content)
- smiles alphabet (character based)
- longest smiles string
"""
smiles_list = np.asanyarray(smiles)
smiles_alphabet = list(set("".join(smiles_list)))
smiles_alphabet.append(" ") # for padding
largest_smiles_len = len(max(smiles_list, key=len))
print("--> Translating SMILES to SELFIES...")
selfies_list = list(map(sf.encoder, smiles_list))
all_selfies_symbols = sf.get_alphabet_from_selfies(selfies_list)
all_selfies_symbols.add("[nop]")
selfies_alphabet = list(all_selfies_symbols)
largest_selfies_len = max(sf.len_selfies(s) for s in selfies_list)
print("Finished translating SMILES to SELFIES.")
return selfies_list, selfies_alphabet, largest_selfies_len, \
smiles_list, smiles_alphabet, largest_smiles_len
def generation(value):
parser = argparse.ArgumentParser()
#parser.add_argument('--model_weight', type=str, help="path of model weights", required=True)
parser.add_argument('--scaffold', action='store_true', default=False, help='condition on scaffold')
parser.add_argument('--lstm', action='store_true', default=False, help='use lstm for transforming scaffold')
#parser.add_argument('--csv_name', type=str, help="name to save the generated mols in csv format", required=True)
parser.add_argument('--data_name', type=str, default = 'moses2', help="name of the dataset to train on", required=False)
parser.add_argument('--batch_size', type=int, default = 512, help="batch size", required=False)
parser.add_argument('--gen_size', type=int, default = 10000, help="number of times to generate from a batch", required=False)
parser.add_argument('--vocab_size', type=int, default = 26, help="number of layers", required=False) # previously 28 .... 26 for moses. 94 for guacamol
parser.add_argument('--block_size', type=int, default = 54, help="number of layers", required=False) # previously 57... 54 for moses. 100 for guacamol.
# parser.add_argument('--num_props', type=int, default = 0, help="number of properties to use for condition", required=False)
parser.add_argument('--props', nargs="+", default = [], help="properties to be used for condition", required=False)
parser.add_argument('--n_layer', type=int, default = 8, help="number of layers", required=False)
parser.add_argument('--n_head', type=int, default = 8, help="number of heads", required=False)
parser.add_argument('--n_embd', type=int, default = 256, help="embedding dimension", required=False)
parser.add_argument('--lstm_layers', type=int, default = 2, help="number of layers in lstm", required=False)
args = parser.parse_args()
args.data_name = 'ppcenos'
args.vocab_size = 29 #
args.block_size = 196 #max_len
args.gen_size = 10
args.batch_size = 5
args.csv_name = 'ppcenos'
args.props = ['pce']
context = "[C]"
args.scaffold = False
pattern = "(\[[^\]]+]|<|Br?|Cl?|N|O|S|P|F|I|b|c|n|o|s|p|\(|\)|\.|=|#|-|\+|\\\\|\/|:|~|@|\?|>|\*|\$|\%[0-9]{2}|[0-9])"
regex = re.compile(pattern)
if ('moses' in args.data_name) and args.scaffold:
scaffold_max_len=48
elif ('guacamol' in args.data_name):
scaffold_max_len = 107
else:
scaffold_max_len = 181
stoi = json.load(open('tool/comget/' + f'{args.data_name}.json', 'r'))
# itos = { i:ch for i,ch in enumerate(chars) }
itos = { i:ch for ch,i in stoi.items() }
print(len(itos))
num_props = len(args.props)
mconf = GPTConfig(args.vocab_size, args.block_size, num_props = num_props,
n_layer=args.n_layer, n_head=args.n_head, n_embd=args.n_embd, scaffold = args.scaffold, scaffold_maxlen = scaffold_max_len,
lstm = args.lstm, lstm_layers = args.lstm_layers)
model = GPT(mconf)
args.model_weight = f'{args.csv_name}.pt'
model.load_state_dict(torch.load('tool/comget/' + args.model_weight, map_location=torch.device('cpu')))
model.to('cpu')
print('Model loaded')
gen_iter = math.ceil(args.gen_size / args.batch_size)
# gen_iter = 2
if 'guacamol1' in args.data_name:
prop2value = {'qed': [0.3, 0.5, 0.7], 'sas': [2.0, 3.0, 4.0], 'logp': [2.0, 4.0, 6.0], 'tpsa': [40.0, 80.0, 120.0],
'tpsa_logp': [[40.0, 2.0], [80.0, 2.0], [120.0, 2.0], [40.0, 4.0], [80.0, 4.0], [120.0, 4.0], [40.0, 6.0], [80.0, 6.0], [120.0, 6.0]],
'sas_logp': [[2.0, 2.0], [2.0, 4.0], [2.0, 6.0], [3.0, 2.0], [3.0, 4.0], [3.0, 6.0], [4.0, 2.0], [4.0, 4.0], [4.0, 6.0]],
'tpsa_sas': [[40.0, 2.0], [80.0, 2.0], [120.0, 2.0], [40.0, 3.0], [80.0, 3.0], [120.0, 3.0], [40.0, 4.0], [80.0, 4.0], [120.0, 4.0]],
'tpsa_logp_sas': [[40.0, 2.0, 2.0], [40.0, 2.0, 4.0], [40.0, 6.0, 4.0], [40.0, 6.0, 2.0], [80.0, 6.0, 4.0], [80.0, 2.0, 4.0], [80.0, 2.0, 2.0], [80.0, 6.0, 2.0]]}
else:
prop2value = { 'pce': [float(value)]}
prop_condition = None
if len(args.props) > 0:
prop_condition = prop2value['_'.join(args.props)]
scaf_condition = None
all_dfs = []
all_metrics = []
count = 0
if prop_condition is not None and scaf_condition is None :
for c in prop_condition:
molecules = []
selfies = []
count += 1
for i in tqdm(range(gen_iter)):
x = torch.tensor([stoi[s] for s in regex.findall(context)], dtype=torch.long)[None,...].repeat(args.batch_size, 1).to('cpu')
p = None
if len(args.props) == 1:
p = torch.tensor([c]).repeat(args.batch_size, 1).to('cpu') # for single condition
else:
p = torch.tensor([c]).repeat(args.batch_size, 1).unsqueeze(1).to('cpu') # for multiple conditions
sca = None
y = sample(model, x, 300, temperature= 1.0, sample=True, top_k = 10, prop = p, scaffold = sca)
for gen_mol in y:
completion = ''.join([itos[int(i)] for i in gen_mol])
completion = completion.replace('<', '')
selfies.append(completion)
file = pd.DataFrame(selfies)
for ind, i in enumerate( file[0]):
smi = (sf.decoder(eval(repr(i))))
mol = get_mol(smi)
# gen_smiles.append(completion)
if mol:
molecules.append(mol)
else:
print(ind)
print(i)
"Valid molecules % = {}".format(len(molecules))
mol_dict = []
for i in molecules:
mol_dict.append({'molecule' : i, 'smiles': Chem.MolToSmiles(i)})
# for i in gen_smiles:
# mol_dict.append({'temperature' : temp, 'smiles': i})
results = pd.DataFrame(mol_dict)
all_dfs.append(results)
results = pd.concat(all_dfs)
return results