|
|
import pandas as pd |
|
|
from rdkit import Chem |
|
|
from rdkit.Chem.Scaffolds import MurckoScaffold |
|
|
from collections import defaultdict |
|
|
from concurrent.futures import ProcessPoolExecutor |
|
|
from tqdm import tqdm |
|
|
import argparse |
|
|
import os |
|
|
import pyarrow as pa |
|
|
import pyarrow.parquet as pq |
|
|
|
|
|
|
|
|
def get_murcko(smiles): |
|
|
try: |
|
|
mol = Chem.MolFromSmiles(smiles) |
|
|
scaffold = MurckoScaffold.GetScaffoldForMol(mol) |
|
|
return Chem.MolToSmiles(scaffold) |
|
|
except: |
|
|
return None |
|
|
|
|
|
def parallel_get_murcko(smiles_list, n_jobs=10): |
|
|
with ProcessPoolExecutor(max_workers=n_jobs) as executor: |
|
|
scaffolds = list(tqdm(executor.map(get_murcko, smiles_list), total=len(smiles_list))) |
|
|
return scaffolds |
|
|
|
|
|
def scaffold_split(df, frac_train=0.8, frac_val=0.1, seed=42): |
|
|
scaffold_to_indices = defaultdict(list) |
|
|
for idx, scaffold in enumerate(df['scaffold']): |
|
|
scaffold_to_indices[scaffold].append(idx) |
|
|
|
|
|
scaffolds = list(scaffold_to_indices.keys()) |
|
|
rng = pd.Series(scaffolds).sample(frac=1, random_state=seed).tolist() |
|
|
|
|
|
train_idx, val_idx, test_idx = [], [], [] |
|
|
n_total = len(df) |
|
|
n_train, n_val = int(frac_train * n_total), int(frac_val * n_total) |
|
|
|
|
|
for scaffold in rng: |
|
|
indices = scaffold_to_indices[scaffold] |
|
|
if len(train_idx) + len(indices) <= n_train: |
|
|
train_idx.extend(indices) |
|
|
elif len(val_idx) + len(indices) <= n_val: |
|
|
val_idx.extend(indices) |
|
|
else: |
|
|
test_idx.extend(indices) |
|
|
|
|
|
return df.iloc[train_idx], df.iloc[val_idx], df.iloc[test_idx] |
|
|
|
|
|
|
|
|
def main(args): |
|
|
os.makedirs(args.output_dir, exist_ok=True) |
|
|
|
|
|
print(f"Loading data from {args.input_parquet} ...") |
|
|
df = pd.read_parquet(args.input_parquet) |
|
|
print(f"Loaded {len(df)} molecules") |
|
|
|
|
|
print(f"Extracting Murcko scaffolds using {args.num_cores} cores ...") |
|
|
df['scaffold'] = parallel_get_murcko(df[args.smiles_column], n_jobs=args.num_cores) |
|
|
|
|
|
print("Performing scaffold split ...") |
|
|
train_df, val_df, test_df = scaffold_split(df, frac_train=0.8, frac_val=0.1, seed=args.seed) |
|
|
|
|
|
print(f"Train: {len(train_df)} | Val: {len(val_df)} | Test: {len(test_df)}") |
|
|
|
|
|
print(f"Saving splits to {args.output_dir} ...") |
|
|
train_df.to_parquet(os.path.join(args.output_dir, "train.parquet"), index=False) |
|
|
val_df.to_parquet(os.path.join(args.output_dir, "val.parquet"), index=False) |
|
|
test_df.to_parquet(os.path.join(args.output_dir, "test.parquet"), index=False) |
|
|
|
|
|
print("Done!") |
|
|
|
|
|
|
|
|
if __name__ == "__main__": |
|
|
parser = argparse.ArgumentParser(description="Scaffold split virtual screening data and save as parquet.") |
|
|
parser.add_argument("--input_parquet", type=str, required=True, help="Path to input .parquet file") |
|
|
parser.add_argument("--output_dir", type=str, required=True, help="Output directory to save train/val/test splits") |
|
|
parser.add_argument("--smiles_column", type=str, default="smiles", help="Column name for SMILES") |
|
|
parser.add_argument("--num_cores", type=int, default=10, help="Number of CPU cores to use") |
|
|
parser.add_argument("--seed", type=int, default=42, help="Random seed for reproducibility") |
|
|
|
|
|
args = parser.parse_args() |
|
|
main(args) |
|
|
|