Upload 5 files
Browse files- CHAFF_processing_scripts/st1_download_pubchem.py +99 -0
- CHAFF_processing_scripts/st2_run_download_pubchem.py +20 -0
- CHAFF_processing_scripts/st3_extract_active_compounds.py +25 -0
- CHAFF_processing_scripts/st4_smiles_curation.py +91 -0
- CHAFF_processing_scripts/st5_detergent_smiles_curation.py +121 -0
CHAFF_processing_scripts/st1_download_pubchem.py
ADDED
|
@@ -0,0 +1,99 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
#!/usr/bin/env python3
|
| 2 |
+
"""
|
| 3 |
+
Download canonical SMILES and assay outcome for all compounds in a PubChem BioAssay (AID).
|
| 4 |
+
Uses ListKey interface to handle large assays (>10,000 compounds).
|
| 5 |
+
|
| 6 |
+
Usage:
|
| 7 |
+
python download_pubchem.py 584 --output smiles_with_outcome.csv
|
| 8 |
+
"""
|
| 9 |
+
|
| 10 |
+
import argparse
|
| 11 |
+
import csv
|
| 12 |
+
import sys
|
| 13 |
+
import time
|
| 14 |
+
import requests
|
| 15 |
+
from typing import Dict, Tuple
|
| 16 |
+
from io import StringIO
|
| 17 |
+
|
| 18 |
+
PUG_BASE = "https://pubchem.ncbi.nlm.nih.gov/rest/pug"
|
| 19 |
+
|
| 20 |
+
def get_listkey_for_aid(aid: str) -> Tuple[str, int]:
|
| 21 |
+
"""Get ListKey and size for a given AID (used to retrieve large CID sets)."""
|
| 22 |
+
url = f"{PUG_BASE}/assay/aid/{aid}/cids/JSON?list_return=listkey"
|
| 23 |
+
try:
|
| 24 |
+
resp = requests.get(url, timeout=20)
|
| 25 |
+
resp.raise_for_status()
|
| 26 |
+
data = resp.json()["IdentifierList"]
|
| 27 |
+
listkey = data["ListKey"]
|
| 28 |
+
size = data["Size"]
|
| 29 |
+
print(f"Retrieved ListKey: {listkey} with {size} CIDs.")
|
| 30 |
+
return listkey, size
|
| 31 |
+
except Exception as e:
|
| 32 |
+
print(f"Error retrieving ListKey for AID {aid}: {e}", file=sys.stderr)
|
| 33 |
+
sys.exit(1)
|
| 34 |
+
|
| 35 |
+
def fetch_smiles_from_listkey(listkey: str) -> Dict[int, str]:
|
| 36 |
+
"""Fetch SMILES in batches using ListKey."""
|
| 37 |
+
smiles_dict = {}
|
| 38 |
+
url = f"{PUG_BASE}/compound/listkey/{listkey}/property/SMILES/CSV"
|
| 39 |
+
try:
|
| 40 |
+
resp = requests.get(url, timeout=30)
|
| 41 |
+
resp.raise_for_status()
|
| 42 |
+
reader = csv.DictReader(StringIO(resp.text))
|
| 43 |
+
count = 0
|
| 44 |
+
for row in reader:
|
| 45 |
+
cid = int(row["CID"])
|
| 46 |
+
smiles = row["SMILES"]
|
| 47 |
+
smiles_dict[cid] = smiles
|
| 48 |
+
count += 1
|
| 49 |
+
print(f"Fetched {count} SMILES")
|
| 50 |
+
except Exception as e:
|
| 51 |
+
print(f"Error fetching SMILES: {e}", file=sys.stderr)
|
| 52 |
+
return smiles_dict
|
| 53 |
+
|
| 54 |
+
def fetch_outcomes_from_listkey(aid: int, listkey: str, size: int, batch_size: int, delay: float) -> Dict[int, str]:
|
| 55 |
+
"""Fetch outcomes in batches using ListKey."""
|
| 56 |
+
outcomes = {}
|
| 57 |
+
for start in range(0, size, batch_size):
|
| 58 |
+
url = f"{PUG_BASE}/assay/aid/{aid}/CSV?cid=listkey&listkey={listkey}&listkey_start={start}&listkey_count={batch_size}"
|
| 59 |
+
try:
|
| 60 |
+
resp = requests.get(url, timeout=30)
|
| 61 |
+
resp.raise_for_status()
|
| 62 |
+
reader = csv.DictReader(StringIO(resp.text))
|
| 63 |
+
for row in reader:
|
| 64 |
+
if "PUBCHEM_CID" in row and row["PUBCHEM_CID"].isdigit():
|
| 65 |
+
cid = int(row["PUBCHEM_CID"])
|
| 66 |
+
outcome = row.get("PUBCHEM_ACTIVITY_OUTCOME", "Unavailable")
|
| 67 |
+
outcomes[cid] = outcome
|
| 68 |
+
print(f"Fetched outcomes for {len(outcomes)} compounds.")
|
| 69 |
+
time.sleep(delay)
|
| 70 |
+
except Exception as e:
|
| 71 |
+
print(f"Error fetching assay outcomes: {e}", file=sys.stderr)
|
| 72 |
+
sys.exit(1)
|
| 73 |
+
return outcomes
|
| 74 |
+
|
| 75 |
+
def write_to_csv(smiles_dict: Dict[int, str], outcomes: Dict[int, str], output_file: str):
|
| 76 |
+
"""Write combined CID, SMILES, and outcome to CSV."""
|
| 77 |
+
with open(output_file, "w", newline='') as f:
|
| 78 |
+
writer = csv.writer(f)
|
| 79 |
+
writer.writerow(["CID", "CanonicalSMILES", "AssayOutcome"])
|
| 80 |
+
for cid, smiles in smiles_dict.items():
|
| 81 |
+
outcome = outcomes.get(cid, "Unavailable")
|
| 82 |
+
writer.writerow([cid, smiles, outcome])
|
| 83 |
+
print(f"Saved {len(smiles_dict)} entries to {output_file}")
|
| 84 |
+
|
| 85 |
+
def main():
|
| 86 |
+
parser = argparse.ArgumentParser(description="Download SMILES and outcome using ListKey for a PubChem assay")
|
| 87 |
+
parser.add_argument("aid", help="PubChem Assay ID (AID)")
|
| 88 |
+
parser.add_argument("--output", "-o", default="smiles_with_outcome.csv", help="Output CSV file")
|
| 89 |
+
parser.add_argument("--batch-size", type=int, default=10000, help="Batch size for ListKey download")
|
| 90 |
+
parser.add_argument("--delay", type=float, default=0.5, help="Delay between batches (in seconds)")
|
| 91 |
+
args = parser.parse_args()
|
| 92 |
+
|
| 93 |
+
listkey, size = get_listkey_for_aid(args.aid)
|
| 94 |
+
smiles_dict = fetch_smiles_from_listkey(listkey)
|
| 95 |
+
outcomes = fetch_outcomes_from_listkey(args.aid, listkey, size, batch_size=args.batch_size, delay=args.delay)
|
| 96 |
+
write_to_csv(smiles_dict, outcomes, args.output)
|
| 97 |
+
|
| 98 |
+
if __name__ == "__main__":
|
| 99 |
+
main()
|
CHAFF_processing_scripts/st2_run_download_pubchem.py
ADDED
|
@@ -0,0 +1,20 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
import os
|
| 2 |
+
import subprocess
|
| 3 |
+
|
| 4 |
+
aid_list = [
|
| 5 |
+
632, 1641, 1730, 1857, 1926, 435026, 504689, 720541, 1159604,
|
| 6 |
+
587, 588, 589, 590, 591, 592, 593, 594, 709, 923, 1480, 1483, 1696, 1775, 1776, 2124, 2757,
|
| 7 |
+
588517, 588620, 624483, 720675, 720678, 720680, 720681, 720682, 720686, 720687,
|
| 8 |
+
584, 585, 1476, 1478, 485294, 485341, # Detergents
|
| 9 |
+
411, 1006, 1269, 1379, 1891, 2515, 2530, 366887, 366889, 366891, 488838, 493175,
|
| 10 |
+
588342, 588498, 602357, 602358, 602364, 602474, 602475, 602476, 602477,
|
| 11 |
+
624030, 652016, 720522, 720835, 1224835, 1347047,
|
| 12 |
+
672, 682, 936,
|
| 13 |
+
878, 888, 929, 1234
|
| 14 |
+
]
|
| 15 |
+
|
| 16 |
+
for aid in aid_list:
|
| 17 |
+
output_file = f"./raw/pubchem_aid_{aid}.csv"
|
| 18 |
+
command = ["python", "st1_download_pubchem.py", str(aid), "--output", output_file]
|
| 19 |
+
print(f"Running: {' '.join(command)}")
|
| 20 |
+
subprocess.run(command)
|
CHAFF_processing_scripts/st3_extract_active_compounds.py
ADDED
|
@@ -0,0 +1,25 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
import os
|
| 2 |
+
import pandas as pd
|
| 3 |
+
import glob
|
| 4 |
+
|
| 5 |
+
file_path = "./raw/*.csv"
|
| 6 |
+
|
| 7 |
+
# Extract active compounds from the datasets
|
| 8 |
+
for file in glob.glob(file_path):
|
| 9 |
+
name = os.path.splitext(os.path.basename(file))[0]
|
| 10 |
+
output_path = f"./active/{name}_active.csv"
|
| 11 |
+
|
| 12 |
+
if os.path.exists(output_path):
|
| 13 |
+
print(f"Skipping {name}, already exists.")
|
| 14 |
+
continue
|
| 15 |
+
|
| 16 |
+
raw_df = pd.read_csv(file)
|
| 17 |
+
active_df = raw_df[raw_df['AssayOutcome'] == 'Active']
|
| 18 |
+
|
| 19 |
+
# Save new CSV file
|
| 20 |
+
active_df.to_csv(output_path, index=False)
|
| 21 |
+
print(f"Saved {len(active_df)} active compounds out of {len(raw_df)} to {output_path}")
|
| 22 |
+
|
| 23 |
+
|
| 24 |
+
|
| 25 |
+
|
CHAFF_processing_scripts/st4_smiles_curation.py
ADDED
|
@@ -0,0 +1,91 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
import os
|
| 2 |
+
import pandas as pd
|
| 3 |
+
import rdkit
|
| 4 |
+
import molvs
|
| 5 |
+
import tqdm
|
| 6 |
+
import glob
|
| 7 |
+
from rdkit import Chem
|
| 8 |
+
|
| 9 |
+
standardizer = molvs.Standardizer()
|
| 10 |
+
fragment_remover = molvs.fragment.FragmentRemover()
|
| 11 |
+
|
| 12 |
+
|
| 13 |
+
excluded_aids = {"584", "585", "1478", "1476", "485294", "485341"} # with/without detergents
|
| 14 |
+
|
| 15 |
+
file_path = "./active/*.csv"
|
| 16 |
+
|
| 17 |
+
for file in glob.glob(file_path):
|
| 18 |
+
file_name = os.path.splitext(os.path.basename(file))[0]
|
| 19 |
+
aid = file_name.replace("pubchem_aid_", "").replace("_active", "")
|
| 20 |
+
|
| 21 |
+
if aid in excluded_aids:
|
| 22 |
+
print(f"Skipping {aid} (excluded).")
|
| 23 |
+
continue
|
| 24 |
+
|
| 25 |
+
output_path = f"./curated/{file_name}_curated.csv"
|
| 26 |
+
if os.path.exists(output_path):
|
| 27 |
+
print(f"Skipping {file_name}, already exists.")
|
| 28 |
+
continue
|
| 29 |
+
|
| 30 |
+
active_df = pd.read_csv(file)
|
| 31 |
+
smiles_series = active_df["CanonicalSMILES"]
|
| 32 |
+
active_df["curated_SMILES"] = None
|
| 33 |
+
cid = active_df["CID"]
|
| 34 |
+
|
| 35 |
+
# --------- SMILES sanitization ---------
|
| 36 |
+
valid_indices = []
|
| 37 |
+
invalid_smiles = []
|
| 38 |
+
warning_smiles = []
|
| 39 |
+
|
| 40 |
+
for idx, smiles in smiles_series.items():
|
| 41 |
+
mol = Chem.MolFromSmiles(smiles)
|
| 42 |
+
compound_cid = cid.iloc[idx]
|
| 43 |
+
|
| 44 |
+
if mol is None:
|
| 45 |
+
invalid_smiles.append({
|
| 46 |
+
'CID': compound_cid,
|
| 47 |
+
'SMILES': smiles,
|
| 48 |
+
'Reason': "MolFromSmiles returned None"
|
| 49 |
+
})
|
| 50 |
+
continue
|
| 51 |
+
|
| 52 |
+
results = molvs.validate_smiles(smiles)
|
| 53 |
+
|
| 54 |
+
if len(results) > 0:
|
| 55 |
+
warning_smiles.append({
|
| 56 |
+
'CID': compound_cid,
|
| 57 |
+
'SMILES': smiles,
|
| 58 |
+
'Reason': results
|
| 59 |
+
})
|
| 60 |
+
continue
|
| 61 |
+
|
| 62 |
+
mol = standardizer.standardize(mol)
|
| 63 |
+
mol = fragment_remover.remove(mol)
|
| 64 |
+
standardized = Chem.MolToSmiles(mol)
|
| 65 |
+
|
| 66 |
+
active_df.at[idx, "curated_SMILES"] = standardized
|
| 67 |
+
valid_indices.append(idx)
|
| 68 |
+
|
| 69 |
+
# Save valid entries
|
| 70 |
+
valid_df = active_df.loc[valid_indices].reset_index(drop=True)
|
| 71 |
+
valid_df = valid_df.drop(columns=["CanonicalSMILES"])
|
| 72 |
+
valid_df = valid_df.rename(columns={"curated_SMILES": "SMILES"})
|
| 73 |
+
|
| 74 |
+
# Create DataFrames for invalid and warning
|
| 75 |
+
invalid_df = pd.DataFrame(invalid_smiles)
|
| 76 |
+
warning_df = pd.DataFrame(warning_smiles)
|
| 77 |
+
|
| 78 |
+
# Add AID column to dataframes
|
| 79 |
+
valid_df.insert(0, "AID", aid)
|
| 80 |
+
invalid_df.insert(0, "AID", aid)
|
| 81 |
+
warning_df.insert(0, "AID", aid)
|
| 82 |
+
|
| 83 |
+
# Save csv files
|
| 84 |
+
valid_df.to_csv(f'./curated/{file_name}_curated.csv', index=False)
|
| 85 |
+
invalid_df.to_csv(f'./curated/{file_name}_invalid_smiles.csv', index=False)
|
| 86 |
+
warning_df.to_csv(f'./curated/{file_name}_molvs_validation.csv', index=False)
|
| 87 |
+
|
| 88 |
+
print(f"Number of compounds in {file_name}:", len(active_df))
|
| 89 |
+
print(f"Number of invalid smiles in {file_name}: {len(invalid_df)}")
|
| 90 |
+
print(f"Number of warning smiles in {file_name}: {len(warning_df)}\n")
|
| 91 |
+
|
CHAFF_processing_scripts/st5_detergent_smiles_curation.py
ADDED
|
@@ -0,0 +1,121 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
import os
|
| 2 |
+
import pandas as pd
|
| 3 |
+
import rdkit
|
| 4 |
+
import molvs
|
| 5 |
+
import tqdm
|
| 6 |
+
import glob
|
| 7 |
+
from rdkit import Chem
|
| 8 |
+
|
| 9 |
+
standardizer = molvs.Standardizer()
|
| 10 |
+
fragment_remover = molvs.fragment.FragmentRemover()
|
| 11 |
+
|
| 12 |
+
|
| 13 |
+
# AID mapping {without : with detergent}
|
| 14 |
+
filter_map = {
|
| 15 |
+
"585": "584",
|
| 16 |
+
"1476": "1478",
|
| 17 |
+
"485341": "485294"
|
| 18 |
+
}
|
| 19 |
+
|
| 20 |
+
# AID -> CID set maaping
|
| 21 |
+
cid_sets = {}
|
| 22 |
+
|
| 23 |
+
# Gather CIDs
|
| 24 |
+
for target_aid in filter_map.values():
|
| 25 |
+
file_name = f"pubchem_aid_{target_aid}_active.csv"
|
| 26 |
+
file_full_path = os.path.join("./active", file_name)
|
| 27 |
+
|
| 28 |
+
if os.path.exists(file_full_path):
|
| 29 |
+
df = pd.read_csv(file_full_path)
|
| 30 |
+
cid_sets[target_aid] = set(df["CID"].tolist())
|
| 31 |
+
else:
|
| 32 |
+
print(f"Warning: file for AID {target_aid} not found!")
|
| 33 |
+
|
| 34 |
+
# Fliter
|
| 35 |
+
for target_aid, filter_aid in filter_map.items():
|
| 36 |
+
file_name = f"pubchem_aid_{target_aid}_active.csv"
|
| 37 |
+
file_full_path = os.path.join("./active", file_name)
|
| 38 |
+
|
| 39 |
+
if not os.path.exists(file_full_path):
|
| 40 |
+
print(f"Skipping {target_aid}, file not found.")
|
| 41 |
+
continue
|
| 42 |
+
|
| 43 |
+
df = pd.read_csv(file_full_path)
|
| 44 |
+
|
| 45 |
+
before = len(df)
|
| 46 |
+
df = df[~df["CID"].isin(cid_sets[filter_aid])]
|
| 47 |
+
after = len(df)
|
| 48 |
+
|
| 49 |
+
output_path = f"./active/pubchem_aid_{target_aid}_active_filtered.csv"
|
| 50 |
+
df.to_csv(output_path, index=False)
|
| 51 |
+
|
| 52 |
+
print(f"{target_aid}: removed {before - after} compounds from {before}, saved to {output_path}")
|
| 53 |
+
|
| 54 |
+
|
| 55 |
+
filtered_files = [
|
| 56 |
+
"pubchem_aid_585_active_filtered.csv",
|
| 57 |
+
"pubchem_aid_1476_active_filtered.csv",
|
| 58 |
+
"pubchem_aid_485341_active_filtered.csv"
|
| 59 |
+
]
|
| 60 |
+
|
| 61 |
+
|
| 62 |
+
for file in filtered_files:
|
| 63 |
+
file_path = os.path.join("./active", file)
|
| 64 |
+
file_name = os.path.splitext(file)[0]
|
| 65 |
+
|
| 66 |
+
# Read file
|
| 67 |
+
active_df = pd.read_csv(file_path)
|
| 68 |
+
smiles_series = active_df["CanonicalSMILES"]
|
| 69 |
+
active_df["curated_SMILES"] = None
|
| 70 |
+
cid = active_df["CID"]
|
| 71 |
+
|
| 72 |
+
valid_indices = []
|
| 73 |
+
invalid_smiles = []
|
| 74 |
+
warning_smiles = []
|
| 75 |
+
|
| 76 |
+
for idx, smiles in smiles_series.items():
|
| 77 |
+
mol = Chem.MolFromSmiles(smiles)
|
| 78 |
+
compound_cid = cid.iloc[idx]
|
| 79 |
+
|
| 80 |
+
if mol is None:
|
| 81 |
+
invalid_smiles.append({
|
| 82 |
+
'CID': compound_cid,
|
| 83 |
+
'SMILES': smiles,
|
| 84 |
+
'Reason': "MolFromSmiles returned None"
|
| 85 |
+
})
|
| 86 |
+
continue
|
| 87 |
+
|
| 88 |
+
results = molvs.validate_smiles(smiles)
|
| 89 |
+
if len(results) > 0:
|
| 90 |
+
warning_smiles.append({
|
| 91 |
+
'CID': compound_cid,
|
| 92 |
+
'SMILES': smiles,
|
| 93 |
+
'Reason': results
|
| 94 |
+
})
|
| 95 |
+
continue
|
| 96 |
+
|
| 97 |
+
mol = standardizer.standardize(mol)
|
| 98 |
+
mol = fragment_remover.remove(mol)
|
| 99 |
+
standardized = Chem.MolToSmiles(mol)
|
| 100 |
+
|
| 101 |
+
active_df.at[idx, "curated_SMILES"] = standardized
|
| 102 |
+
valid_indices.append(idx)
|
| 103 |
+
|
| 104 |
+
# Save outputs
|
| 105 |
+
valid_df = active_df.loc[valid_indices].reset_index(drop=True)
|
| 106 |
+
valid_df = valid_df.drop(columns=["CanonicalSMILES"])
|
| 107 |
+
valid_df = valid_df.rename(columns={"curated_SMILES": "SMILES"})
|
| 108 |
+
aid = file_name.replace("pubchem_aid_", "").replace("_active_filtered", "")
|
| 109 |
+
valid_df.insert(0, "AID", aid)
|
| 110 |
+
|
| 111 |
+
invalid_df = pd.DataFrame(invalid_smiles)
|
| 112 |
+
warning_df = pd.DataFrame(warning_smiles)
|
| 113 |
+
invalid_df.insert(0, "AID", aid)
|
| 114 |
+
warning_df.insert(0, "AID", aid)
|
| 115 |
+
|
| 116 |
+
valid_df.to_csv(f'./curated/{file_name}_curated.csv', index=False)
|
| 117 |
+
invalid_df.to_csv(f'./curated/{file_name}_invalid_smiles.csv', index=False)
|
| 118 |
+
warning_df.to_csv(f'./curated/{file_name}_molvs_validation.csv', index=False)
|
| 119 |
+
|
| 120 |
+
print(f"Finished curation for {file_name}")
|
| 121 |
+
print(f" Valid: {len(valid_df)}, Invalid: {len(invalid_df)}, Warnings: {len(warning_df)}\n")
|