Datasets:
Formats:
parquet
Languages:
English
Size:
10M - 100M
Tags:
biology
chemistry
drug-discovery
clinical-trials
protein-protein-interaction
gene-essentiality
License:
| """Full-scan audit for strict human-only PubChem ETL policy. | |
| This script performs an exhaustive scan over PubChem bioactivities and reports: | |
| - Legacy vs strict human-only filtering counts | |
| - Drop reasons under strict filtering | |
| - UniProt resolvability under each policy | |
| - Optional audit of already-loaded PubChem rows in the SQLite DB | |
| Outputs JSON and Markdown files under exports/. | |
| """ | |
| from __future__ import annotations | |
| import argparse | |
| import json | |
| import sqlite3 | |
| from datetime import datetime, timezone | |
| from pathlib import Path | |
| import pandas as pd | |
| from negbiodb.db import _PROJECT_ROOT | |
| from negbiodb.download import load_config | |
| from negbiodb.etl_pubchem import ( | |
| _normalize_accession, | |
| _resolve_pubchem_chunk, | |
| load_aid_to_uniprot_map, | |
| load_confirmatory_aids, | |
| load_confirmatory_human_aids, | |
| ) | |
| _HUMAN_TAXID = 9606 | |
| def _pct(n: int, d: int) -> float | None: | |
| if d == 0: | |
| return None | |
| return round((100.0 * n) / d, 6) | |
| def _file_meta(path: Path) -> dict: | |
| st = path.stat() | |
| return { | |
| "path": str(path), | |
| "bytes": int(st.st_size), | |
| "size_mb": round(st.st_size / (1024 * 1024), 2), | |
| "mtime_utc": datetime.fromtimestamp(st.st_mtime, tz=timezone.utc).isoformat(), | |
| } | |
| def _maybe_add_example(bucket: list[dict], sample_limit: int, row: pd.Series, reason: str, aid_is_human: bool) -> None: | |
| if len(bucket) >= sample_limit: | |
| return | |
| bucket.append( | |
| { | |
| "reason": reason, | |
| "aid": int(row["aid"]), | |
| "sid": int(row["sid"]), | |
| "target_taxid": None if pd.isna(row["target_taxid"]) else int(row["target_taxid"]), | |
| "protein_accession": None | |
| if pd.isna(row["protein_accession"]) | |
| else str(row["protein_accession"]), | |
| "aid_in_confirmatory_human_set": bool(aid_is_human), | |
| } | |
| ) | |
| def scan_full_bioactivities( | |
| bioactivities_path: Path, | |
| confirmatory_aids: set[int], | |
| confirmatory_human_aids: set[int], | |
| aid_to_uniprot: dict[int, str], | |
| chunksize: int, | |
| sample_limit: int, | |
| progress_rows: int, | |
| ) -> dict: | |
| rows_seen = 0 | |
| chunks_seen = 0 | |
| next_progress = progress_rows | |
| stats = { | |
| "rows_seen": 0, | |
| "rows_inactive_confirmatory_base": 0, | |
| "legacy_rows_kept": 0, | |
| "strict_rows_kept": 0, | |
| "strict_rows_kept_known_human_taxid": 0, | |
| "strict_rows_kept_missing_taxid_human_aid": 0, | |
| "strict_rows_dropped_known_nonhuman_taxid": 0, | |
| "strict_rows_dropped_missing_taxid_nonhuman_aid": 0, | |
| "legacy_rows_with_uniprot": 0, | |
| "strict_rows_with_uniprot": 0, | |
| "strict_rows_without_uniprot": 0, | |
| } | |
| unique = { | |
| "base_aids": set(), | |
| "legacy_aids": set(), | |
| "strict_aids": set(), | |
| "strict_drop_nonhuman_aids": set(), | |
| "strict_drop_missing_nonhuman_aids": set(), | |
| } | |
| samples = { | |
| "kept_missing_taxid_human_aid": [], | |
| "dropped_known_nonhuman_taxid": [], | |
| "dropped_missing_taxid_nonhuman_aid": [], | |
| } | |
| reader = pd.read_csv( | |
| bioactivities_path, | |
| sep="\t", | |
| compression="gzip" if bioactivities_path.suffix == ".gz" else None, | |
| chunksize=chunksize, | |
| low_memory=False, | |
| ) | |
| for chunk in reader: | |
| chunks_seen += 1 | |
| rows_seen += len(chunk) | |
| stats["rows_seen"] = int(rows_seen) | |
| base = _resolve_pubchem_chunk(chunk, confirmatory_aids) | |
| if base.empty: | |
| if rows_seen >= next_progress: | |
| print(f"[audit] rows_seen={rows_seen} base=0 strict={stats['strict_rows_kept']}", flush=True) | |
| next_progress += progress_rows | |
| continue | |
| stats["rows_inactive_confirmatory_base"] += len(base) | |
| unique["base_aids"].update(int(v) for v in base["aid"].unique().tolist()) | |
| known_taxid = base["target_taxid"].notna() | |
| human_taxid = base["target_taxid"] == _HUMAN_TAXID | |
| missing_taxid = ~known_taxid | |
| missing_taxid_human_aid = missing_taxid & base["aid"].isin(confirmatory_human_aids) | |
| strict_mask = human_taxid | missing_taxid_human_aid | |
| # Legacy behavior was chunk-level: only filter if chunk has any known taxid. | |
| legacy = base | |
| if known_taxid.any(): | |
| legacy = base[human_taxid] | |
| stats["legacy_rows_kept"] += len(legacy) | |
| unique["legacy_aids"].update(int(v) for v in legacy["aid"].unique().tolist()) | |
| strict = base[strict_mask] | |
| stats["strict_rows_kept"] += len(strict) | |
| unique["strict_aids"].update(int(v) for v in strict["aid"].unique().tolist()) | |
| kept_known_human = int(human_taxid.sum()) | |
| kept_missing_human = int(missing_taxid_human_aid.sum()) | |
| dropped_known_nonhuman = int((known_taxid & (~human_taxid)).sum()) | |
| dropped_missing_nonhuman = int((missing_taxid & (~base["aid"].isin(confirmatory_human_aids))).sum()) | |
| stats["strict_rows_kept_known_human_taxid"] += kept_known_human | |
| stats["strict_rows_kept_missing_taxid_human_aid"] += kept_missing_human | |
| stats["strict_rows_dropped_known_nonhuman_taxid"] += dropped_known_nonhuman | |
| stats["strict_rows_dropped_missing_taxid_nonhuman_aid"] += dropped_missing_nonhuman | |
| if dropped_known_nonhuman > 0: | |
| drop_rows = base[known_taxid & (~human_taxid)] | |
| unique["strict_drop_nonhuman_aids"].update(int(v) for v in drop_rows["aid"].unique().tolist()) | |
| for _, row in drop_rows.head(sample_limit).iterrows(): | |
| _maybe_add_example( | |
| samples["dropped_known_nonhuman_taxid"], | |
| sample_limit, | |
| row, | |
| "known_nonhuman_taxid", | |
| bool(int(row["aid"]) in confirmatory_human_aids), | |
| ) | |
| if dropped_missing_nonhuman > 0: | |
| drop_rows = base[missing_taxid & (~base["aid"].isin(confirmatory_human_aids))] | |
| unique["strict_drop_missing_nonhuman_aids"].update( | |
| int(v) for v in drop_rows["aid"].unique().tolist() | |
| ) | |
| for _, row in drop_rows.head(sample_limit).iterrows(): | |
| _maybe_add_example( | |
| samples["dropped_missing_taxid_nonhuman_aid"], | |
| sample_limit, | |
| row, | |
| "missing_taxid_nonhuman_aid", | |
| False, | |
| ) | |
| if kept_missing_human > 0: | |
| keep_rows = base[missing_taxid_human_aid] | |
| for _, row in keep_rows.head(sample_limit).iterrows(): | |
| _maybe_add_example( | |
| samples["kept_missing_taxid_human_aid"], | |
| sample_limit, | |
| row, | |
| "missing_taxid_human_aid", | |
| True, | |
| ) | |
| aid_mapped = base["aid"].map(aid_to_uniprot).map(_normalize_accession) | |
| resolved_uniprot = base["protein_accession"].where(base["protein_accession"].notna(), aid_mapped) | |
| legacy_uniprot = resolved_uniprot.loc[legacy.index] | |
| strict_uniprot = resolved_uniprot.loc[strict.index] | |
| stats["legacy_rows_with_uniprot"] += int(legacy_uniprot.notna().sum()) | |
| stats["strict_rows_with_uniprot"] += int(strict_uniprot.notna().sum()) | |
| stats["strict_rows_without_uniprot"] += int(strict_uniprot.isna().sum()) | |
| if rows_seen >= next_progress: | |
| print( | |
| f"[audit] rows_seen={rows_seen} base={stats['rows_inactive_confirmatory_base']} " | |
| f"legacy={stats['legacy_rows_kept']} strict={stats['strict_rows_kept']}", | |
| flush=True, | |
| ) | |
| next_progress += progress_rows | |
| stats["chunks_seen"] = chunks_seen | |
| stats["rows_seen"] = rows_seen | |
| stats["unique_base_aids"] = len(unique["base_aids"]) | |
| stats["unique_legacy_aids"] = len(unique["legacy_aids"]) | |
| stats["unique_strict_aids"] = len(unique["strict_aids"]) | |
| stats["unique_strict_drop_nonhuman_aids"] = len(unique["strict_drop_nonhuman_aids"]) | |
| stats["unique_strict_drop_missing_nonhuman_aids"] = len(unique["strict_drop_missing_nonhuman_aids"]) | |
| stats["strict_vs_legacy_row_retention_pct"] = _pct( | |
| stats["strict_rows_kept"], | |
| stats["legacy_rows_kept"], | |
| ) | |
| stats["strict_vs_base_row_retention_pct"] = _pct( | |
| stats["strict_rows_kept"], | |
| stats["rows_inactive_confirmatory_base"], | |
| ) | |
| stats["strict_with_uniprot_pct"] = _pct( | |
| stats["strict_rows_with_uniprot"], | |
| stats["strict_rows_kept"], | |
| ) | |
| return { | |
| "stats": stats, | |
| "samples": samples, | |
| } | |
| def audit_existing_db(db_path: Path, confirmatory_human_aids: set[int]) -> dict: | |
| out = { | |
| "db_path": str(db_path), | |
| "pubchem_rows_total": 0, | |
| "pubchem_assays_total": 0, | |
| "pubchem_assays_in_human_aid_set": 0, | |
| "pubchem_assays_not_in_human_aid_set": 0, | |
| "pubchem_rows_from_nonhuman_assays": 0, | |
| "species_top_counts": [], | |
| } | |
| if not db_path.exists(): | |
| out["db_exists"] = False | |
| return out | |
| out["db_exists"] = True | |
| conn = sqlite3.connect(str(db_path)) | |
| try: | |
| out["pubchem_rows_total"] = int( | |
| conn.execute( | |
| "SELECT COUNT(*) FROM negative_results WHERE source_db='pubchem'" | |
| ).fetchone()[0] | |
| ) | |
| out["pubchem_assays_total"] = int( | |
| conn.execute("SELECT COUNT(*) FROM assays WHERE source_db='pubchem'").fetchone()[0] | |
| ) | |
| species_rows = conn.execute( | |
| """ | |
| SELECT species_tested, COUNT(*) | |
| FROM negative_results | |
| WHERE source_db='pubchem' | |
| GROUP BY species_tested | |
| ORDER BY COUNT(*) DESC | |
| LIMIT 20 | |
| """ | |
| ).fetchall() | |
| out["species_top_counts"] = [ | |
| {"species_tested": row[0], "count": int(row[1])} for row in species_rows | |
| ] | |
| assay_ids = conn.execute( | |
| "SELECT assay_id, source_assay_id FROM assays WHERE source_db='pubchem'" | |
| ).fetchall() | |
| if not assay_ids: | |
| return out | |
| nonhuman_assay_ids: list[int] = [] | |
| human_assay_count = 0 | |
| for assay_id, source_assay_id in assay_ids: | |
| try: | |
| aid = int(source_assay_id) | |
| except (TypeError, ValueError): | |
| continue | |
| if aid in confirmatory_human_aids: | |
| human_assay_count += 1 | |
| else: | |
| nonhuman_assay_ids.append(int(assay_id)) | |
| out["pubchem_assays_in_human_aid_set"] = int(human_assay_count) | |
| out["pubchem_assays_not_in_human_aid_set"] = int(len(nonhuman_assay_ids)) | |
| if nonhuman_assay_ids: | |
| rows = 0 | |
| batch_size = 500 | |
| for i in range(0, len(nonhuman_assay_ids), batch_size): | |
| batch = nonhuman_assay_ids[i : i + batch_size] | |
| placeholders = ",".join("?" * len(batch)) | |
| rows += int( | |
| conn.execute( | |
| f""" | |
| SELECT COUNT(*) | |
| FROM negative_results | |
| WHERE source_db='pubchem' | |
| AND assay_id IN ({placeholders}) | |
| """, | |
| batch, | |
| ).fetchone()[0] | |
| ) | |
| out["pubchem_rows_from_nonhuman_assays"] = int(rows) | |
| finally: | |
| conn.close() | |
| return out | |
| def render_markdown(report: dict) -> str: | |
| fs = report["full_scan"]["stats"] | |
| db = report["existing_db_audit"] | |
| lines = [ | |
| "# PubChem Strict Human-Only Audit", | |
| "", | |
| f"- generated_at_utc: {report['generated_at_utc']}", | |
| "", | |
| "## File Metadata", | |
| "", | |
| "| file | size_mb | mtime_utc |", | |
| "|---|---:|---|", | |
| ] | |
| for key in ["bioactivities", "bioassays", "aid_uniprot"]: | |
| m = report["files"][key] | |
| lines.append(f"| {key} | {m['size_mb']} | {m['mtime_utc']} |") | |
| lines += [ | |
| "", | |
| "## Key Sets", | |
| "", | |
| f"- confirmatory_aids: {report['key_sets']['confirmatory_aids']}", | |
| f"- confirmatory_human_aids: {report['key_sets']['confirmatory_human_aids']}", | |
| f"- aid_to_uniprot: {report['key_sets']['aid_to_uniprot']}", | |
| "", | |
| "## Full Scan", | |
| "", | |
| f"- rows_seen: {fs['rows_seen']}", | |
| f"- rows_inactive_confirmatory_base: {fs['rows_inactive_confirmatory_base']}", | |
| f"- legacy_rows_kept: {fs['legacy_rows_kept']}", | |
| f"- strict_rows_kept: {fs['strict_rows_kept']}", | |
| f"- strict_vs_legacy_row_retention_pct: {fs['strict_vs_legacy_row_retention_pct']}", | |
| f"- strict_rows_dropped_known_nonhuman_taxid: {fs['strict_rows_dropped_known_nonhuman_taxid']}", | |
| f"- strict_rows_dropped_missing_taxid_nonhuman_aid: {fs['strict_rows_dropped_missing_taxid_nonhuman_aid']}", | |
| f"- strict_rows_kept_missing_taxid_human_aid: {fs['strict_rows_kept_missing_taxid_human_aid']}", | |
| f"- strict_rows_with_uniprot: {fs['strict_rows_with_uniprot']}", | |
| f"- strict_rows_without_uniprot: {fs['strict_rows_without_uniprot']}", | |
| "", | |
| "## Existing DB Audit", | |
| "", | |
| f"- db_exists: {db.get('db_exists')}", | |
| f"- pubchem_rows_total: {db.get('pubchem_rows_total')}", | |
| f"- pubchem_assays_total: {db.get('pubchem_assays_total')}", | |
| f"- pubchem_assays_in_human_aid_set: {db.get('pubchem_assays_in_human_aid_set')}", | |
| f"- pubchem_assays_not_in_human_aid_set: {db.get('pubchem_assays_not_in_human_aid_set')}", | |
| f"- pubchem_rows_from_nonhuman_assays: {db.get('pubchem_rows_from_nonhuman_assays')}", | |
| "", | |
| ] | |
| return "\n".join(lines) + "\n" | |
| def main() -> None: | |
| parser = argparse.ArgumentParser(description="Full audit for strict human-only PubChem ETL policy") | |
| parser.add_argument("--chunksize", type=int, default=200_000) | |
| parser.add_argument( | |
| "--sample-limit", | |
| type=int, | |
| default=20, | |
| help="Per-category sample rows in the report", | |
| ) | |
| parser.add_argument( | |
| "--progress-rows", | |
| type=int, | |
| default=5_000_000, | |
| help="Print progress every N scanned rows", | |
| ) | |
| parser.add_argument( | |
| "--output-prefix", | |
| type=str, | |
| default="pubchem_strict_audit", | |
| ) | |
| args = parser.parse_args() | |
| cfg = load_config() | |
| pubchem_cfg = cfg["downloads"]["pubchem"] | |
| bioactivities_path = _PROJECT_ROOT / pubchem_cfg["dest"] | |
| bioassays_path = _PROJECT_ROOT / pubchem_cfg["bioassays_dest"] | |
| aid_uniprot_path = _PROJECT_ROOT / pubchem_cfg["aid_uniprot_dest"] | |
| db_path = _PROJECT_ROOT / cfg["paths"]["database"] | |
| for p in [bioactivities_path, bioassays_path, aid_uniprot_path]: | |
| if not p.exists(): | |
| raise FileNotFoundError(f"Required file missing: {p}") | |
| confirmatory_aids = load_confirmatory_aids(bioassays_path, chunksize=args.chunksize) | |
| confirmatory_human_aids = load_confirmatory_human_aids(bioassays_path, chunksize=args.chunksize) | |
| aid_to_uniprot = load_aid_to_uniprot_map(aid_uniprot_path, chunksize=args.chunksize) | |
| print( | |
| f"[audit] loaded sets: confirm={len(confirmatory_aids)} " | |
| f"confirm_human={len(confirmatory_human_aids)} aid_map={len(aid_to_uniprot)}", | |
| flush=True, | |
| ) | |
| full_scan = scan_full_bioactivities( | |
| bioactivities_path=bioactivities_path, | |
| confirmatory_aids=confirmatory_aids, | |
| confirmatory_human_aids=confirmatory_human_aids, | |
| aid_to_uniprot=aid_to_uniprot, | |
| chunksize=args.chunksize, | |
| sample_limit=args.sample_limit, | |
| progress_rows=args.progress_rows, | |
| ) | |
| existing_db = audit_existing_db(db_path, confirmatory_human_aids) | |
| report = { | |
| "generated_at_utc": datetime.now(timezone.utc).isoformat(), | |
| "files": { | |
| "bioactivities": _file_meta(bioactivities_path), | |
| "bioassays": _file_meta(bioassays_path), | |
| "aid_uniprot": _file_meta(aid_uniprot_path), | |
| }, | |
| "key_sets": { | |
| "confirmatory_aids": len(confirmatory_aids), | |
| "confirmatory_human_aids": len(confirmatory_human_aids), | |
| "aid_to_uniprot": len(aid_to_uniprot), | |
| }, | |
| "full_scan": full_scan, | |
| "existing_db_audit": existing_db, | |
| } | |
| exports_dir = _PROJECT_ROOT / cfg["paths"]["exports_dir"] | |
| exports_dir.mkdir(parents=True, exist_ok=True) | |
| ts = datetime.now(timezone.utc).strftime("%Y%m%dT%H%M%SZ") | |
| json_path = exports_dir / f"{args.output_prefix}_{ts}.json" | |
| md_path = exports_dir / f"{args.output_prefix}_{ts}.md" | |
| json_path.write_text(json.dumps(report, indent=2), encoding="utf-8") | |
| md_path.write_text(render_markdown(report), encoding="utf-8") | |
| fs = report["full_scan"]["stats"] | |
| print(f"Audit JSON: {json_path}") | |
| print(f"Audit Markdown: {md_path}") | |
| print( | |
| "[audit] strict rows kept: " | |
| f"{fs['strict_rows_kept']} / legacy {fs['legacy_rows_kept']} " | |
| f"({fs['strict_vs_legacy_row_retention_pct']}%)" | |
| ) | |
| if __name__ == "__main__": | |
| main() | |