Datasets:
Formats:
parquet
Languages:
English
Size:
10M - 100M
Tags:
biology
chemistry
drug-discovery
clinical-trials
protein-protein-interaction
gene-essentiality
License:
| #!/usr/bin/env python3 | |
| """Build CT-L3 reasoning dataset for LLM benchmark. | |
| Generates 200 records from gold tier (with silver fallback cascade), | |
| Phase II-III, safety+efficacy. Requires ChEMBL resolution. | |
| Supports biologics (no SMILES, molecular context). | |
| Output: exports/ct_llm/ct_l3_dataset.jsonl | |
| """ | |
| from __future__ import annotations | |
| import argparse | |
| import logging | |
| import sys | |
| from pathlib import Path | |
| import numpy as np | |
| import pandas as pd | |
| logging.basicConfig(level=logging.INFO, format="%(asctime)s %(levelname)s %(message)s") | |
| logger = logging.getLogger(__name__) | |
| PROJECT_ROOT = Path(__file__).resolve().parent.parent | |
| OUTPUT_DIR = PROJECT_ROOT / "exports" / "ct_llm" | |
| N_TOTAL = 200 | |
| MIN_CHEMBL_COVERAGE_PCT = 25 | |
| MAX_ONCOLOGY_FRAC = 0.40 | |
| BIOLOGIC_TYPES = {"monoclonal_antibody", "antibody_drug_conjugate", "peptide", "other_biologic"} | |
| def load_intervention_targets(db_path: Path) -> dict[int, list[dict]]: | |
| """Load intervention_targets: intervention_id → list of {uniprot, gene_symbol}.""" | |
| from negbiodb_ct.ct_db import get_connection | |
| conn = get_connection(db_path) | |
| try: | |
| rows = conn.execute( | |
| "SELECT intervention_id, uniprot_accession, gene_symbol " | |
| "FROM intervention_targets" | |
| ).fetchall() | |
| finally: | |
| conn.close() | |
| targets: dict[int, list[dict]] = {} | |
| for iid, uniprot, gene in rows: | |
| targets.setdefault(iid, []).append({ | |
| "uniprot": uniprot, | |
| "gene_symbol": gene, | |
| }) | |
| return targets | |
| def format_l3_context(row: pd.Series, targets: list[dict] | None) -> str: | |
| """Generate CT-L3 reasoning context (gold tier, full quantitative).""" | |
| from negbiodb_ct.llm_dataset import infer_therapeutic_area | |
| lines = [f"Drug: {row.get('intervention_name', 'Unknown')}"] | |
| mol_type = row.get("molecular_type") | |
| if mol_type: | |
| lines.append(f"Drug type: {mol_type}") | |
| smiles = row.get("canonical_smiles") | |
| if smiles and pd.notna(smiles) and mol_type == "small_molecule": | |
| lines.append(f"SMILES: {smiles}") | |
| if targets: | |
| gene_list = [t["gene_symbol"] for t in targets if t.get("gene_symbol")] | |
| if gene_list: | |
| lines.append(f"Known targets: {', '.join(gene_list[:5])}") | |
| condition = row.get("condition_name", "Unknown") | |
| lines.append(f"Condition: {condition}") | |
| ta = infer_therapeutic_area(condition) | |
| lines.append(f"Therapeutic area: {ta}") | |
| phase = row.get("trial_phase") or row.get("highest_phase_reached") | |
| if phase: | |
| lines.append(f"Phase: {phase}") | |
| design = row.get("control_type") | |
| if design: | |
| lines.append(f"Design: {design}") | |
| blinding = row.get("blinding") | |
| if blinding and pd.notna(blinding): | |
| lines.append(f"Blinding: {blinding}") | |
| enrollment = row.get("enrollment_actual") | |
| if enrollment and pd.notna(enrollment): | |
| lines.append(f"Enrollment: {int(enrollment)}") | |
| endpoint_met = row.get("primary_endpoint_met") | |
| if endpoint_met and pd.notna(endpoint_met): | |
| lines.append(f"Primary endpoint met: {endpoint_met}") | |
| p_val = row.get("p_value_primary") | |
| if p_val and pd.notna(p_val): | |
| lines.append(f"p-value: {p_val}") | |
| effect = row.get("effect_size") | |
| if effect and pd.notna(effect): | |
| etype = row.get("effect_size_type", "") | |
| lines.append(f"Effect size ({etype}): {effect}" if etype else f"Effect size: {effect}") | |
| ci_lo, ci_hi = row.get("ci_lower"), row.get("ci_upper") | |
| if ci_lo and pd.notna(ci_lo) and ci_hi and pd.notna(ci_hi): | |
| lines.append(f"95% CI: [{ci_lo}, {ci_hi}]") | |
| saes = row.get("serious_adverse_events") | |
| if saes and pd.notna(saes): | |
| lines.append(f"Serious adverse events: {saes}") | |
| interp = row.get("result_interpretation") | |
| if interp and pd.notna(interp): | |
| lines.append(f"Interpretation: {interp}") | |
| return "\n".join(lines) | |
| def _build_gold_reasoning(row: pd.Series, targets: list[dict]) -> str: | |
| """Build a brief evidence-based reasoning summary for few-shot examples. | |
| This is a structured factual summary, not expert-level reasoning. | |
| It demonstrates the expected response format for few-shot prompting. | |
| """ | |
| parts = [] | |
| drug = row.get("intervention_name", "The drug") | |
| condition = row.get("condition_name", "the condition") | |
| category = row.get("failure_category", "unknown") | |
| # Opening | |
| parts.append( | |
| f"{drug} failed in a clinical trial for {condition}, " | |
| f"classified as a {category} failure." | |
| ) | |
| # Evidence | |
| p_val = row.get("p_value_primary") | |
| effect = row.get("effect_size") | |
| endpoint_met = row.get("primary_endpoint_met") | |
| if p_val and pd.notna(p_val): | |
| parts.append(f"The primary endpoint p-value was {p_val}.") | |
| if endpoint_met and pd.notna(endpoint_met): | |
| parts.append(f"Primary endpoint met: {endpoint_met}.") | |
| if effect and pd.notna(effect): | |
| etype = row.get("effect_size_type", "") | |
| if etype: | |
| parts.append(f"The effect size ({etype}) was {effect}.") | |
| else: | |
| parts.append(f"The effect size was {effect}.") | |
| # Targets | |
| if targets: | |
| gene_list = [t["gene_symbol"] for t in targets if t.get("gene_symbol")] | |
| if gene_list: | |
| parts.append( | |
| f"The drug targets {', '.join(gene_list[:3])}, " | |
| f"which may have been insufficient for this indication." | |
| ) | |
| # Safety | |
| saes = row.get("serious_adverse_events") | |
| if saes and pd.notna(saes) and category == "safety": | |
| parts.append(f"Serious adverse events were reported: {saes}.") | |
| # Interpretation | |
| interp = row.get("result_interpretation") | |
| if interp and pd.notna(interp): | |
| parts.append(f"The trial interpretation noted: {interp}.") | |
| return " ".join(parts) | |
| def build_l3_dataset(db_path: Path, seed: int) -> list[dict]: | |
| """Build CT-L3 dataset from DB with relaxation cascade.""" | |
| from negbiodb_ct.llm_dataset import ( | |
| apply_max_per_drug, | |
| infer_therapeutic_area, | |
| load_candidate_pool, | |
| ) | |
| rng = np.random.RandomState(seed) | |
| int_targets = load_intervention_targets(db_path) | |
| # Base filter: gold tier, safety+efficacy, Phase II-III, ChEMBL resolved | |
| extra_where = ( | |
| "tfr.failure_category IN ('safety', 'efficacy') " | |
| "AND ct.trial_phase IN ('phase_2', 'phase_2_3', 'phase_3') " | |
| "AND ct.has_results = 1 " | |
| "AND i.chembl_id IS NOT NULL " | |
| "AND (i.canonical_smiles IS NOT NULL OR i.molecular_type != 'small_molecule')" | |
| ) | |
| pool = load_candidate_pool(db_path, tier_filter="= 'gold'", extra_where=extra_where) | |
| logger.info("L3 gold pool (strict): %d records", len(pool)) | |
| # Relaxation cascade if pool too small | |
| relaxation_log = [] | |
| if len(pool) < N_TOTAL: | |
| # Relax 1: allow silver tier (keep has_results requirement) | |
| extra_silver = ( | |
| "tfr.failure_category IN ('safety', 'efficacy') " | |
| "AND ct.trial_phase IN ('phase_2', 'phase_2_3', 'phase_3') " | |
| "AND ct.has_results = 1 " | |
| "AND i.chembl_id IS NOT NULL " | |
| "AND (i.canonical_smiles IS NOT NULL OR i.molecular_type != 'small_molecule')" | |
| ) | |
| pool_r1 = load_candidate_pool( | |
| db_path, tier_filter="IN ('gold', 'silver')", extra_where=extra_silver | |
| ) | |
| relaxation_log.append(f"R1: allow silver → {len(pool_r1)} records") | |
| pool = pool_r1 | |
| if len(pool) < N_TOTAL: | |
| # Relax 2: drop has_results requirement | |
| extra_r2 = ( | |
| "tfr.failure_category IN ('safety', 'efficacy') " | |
| "AND ct.trial_phase IN ('phase_2', 'phase_2_3', 'phase_3') " | |
| "AND i.chembl_id IS NOT NULL " | |
| "AND (i.canonical_smiles IS NOT NULL OR i.molecular_type != 'small_molecule')" | |
| ) | |
| pool_r2 = load_candidate_pool( | |
| db_path, tier_filter="IN ('gold', 'silver')", extra_where=extra_r2 | |
| ) | |
| relaxation_log.append(f"R2: drop has_results → {len(pool_r2)} records") | |
| pool = pool_r2 | |
| if len(pool) < N_TOTAL: | |
| # Relax 3: drop SMILES requirement for biologics | |
| extra_r3 = ( | |
| "tfr.failure_category IN ('safety', 'efficacy') " | |
| "AND ct.trial_phase IN ('phase_2', 'phase_2_3', 'phase_3') " | |
| "AND i.chembl_id IS NOT NULL" | |
| ) | |
| pool_r3 = load_candidate_pool( | |
| db_path, tier_filter="IN ('gold', 'silver')", extra_where=extra_r3 | |
| ) | |
| relaxation_log.append(f"R3: drop SMILES req → {len(pool_r3)} records") | |
| pool = pool_r3 | |
| if relaxation_log: | |
| logger.info("Relaxation cascade applied:\n %s", "\n ".join(relaxation_log)) | |
| if len(pool) == 0: | |
| logger.error("No L3 candidates found even after relaxation!") | |
| return [] | |
| # Apply max-per-drug | |
| pool = apply_max_per_drug(pool, rng=rng) | |
| # Diversity constraints | |
| pool["therapeutic_area"] = pool["condition_name"].apply( | |
| lambda x: infer_therapeutic_area(x) if pd.notna(x) else "other" | |
| ) | |
| pool["is_biologic"] = pool["molecular_type"].isin(BIOLOGIC_TYPES) | |
| # Oncology cap | |
| oncology = pool[pool["therapeutic_area"] == "oncology"] | |
| non_oncology = pool[pool["therapeutic_area"] != "oncology"] | |
| max_oncology = int(N_TOTAL * MAX_ONCOLOGY_FRAC) | |
| if len(oncology) > max_oncology: | |
| oncology = oncology.sample(max_oncology, random_state=rng.randint(0, 2**31)) | |
| pool = pd.concat([oncology, non_oncology], ignore_index=True) | |
| # Sample | |
| n_target = min(N_TOTAL, len(pool)) | |
| # Balance safety/efficacy ~50/50 | |
| safety_pool = pool[pool["failure_category"] == "safety"] | |
| efficacy_pool = pool[pool["failure_category"] == "efficacy"] | |
| n_safety = min(n_target // 2, len(safety_pool)) | |
| n_efficacy = min(n_target - n_safety, len(efficacy_pool)) | |
| if n_efficacy < n_target - n_safety: | |
| n_safety = min(n_target - n_efficacy, len(safety_pool)) | |
| sampled_safety = safety_pool.sample(n_safety, random_state=rng.randint(0, 2**31)) | |
| sampled_efficacy = efficacy_pool.sample(n_efficacy, random_state=rng.randint(0, 2**31)) | |
| sampled = pd.concat([sampled_safety, sampled_efficacy], ignore_index=True) | |
| # Check biologic fraction | |
| n_biologic = sampled["is_biologic"].sum() | |
| logger.info( | |
| "Sampled %d (safety=%d, efficacy=%d, biologic=%d/%.0f%%)", | |
| len(sampled), n_safety, n_efficacy, | |
| n_biologic, 100 * n_biologic / max(len(sampled), 1), | |
| ) | |
| # Build records | |
| records = [] | |
| for _, row in sampled.iterrows(): | |
| iid = row.get("intervention_id") | |
| targets = int_targets.get(int(iid), []) if pd.notna(iid) else [] | |
| context_text = format_l3_context(row, targets) | |
| gold_reasoning = _build_gold_reasoning(row, targets) | |
| records.append({ | |
| "question_id": None, | |
| "task": "CT-L3", | |
| "gold_answer": row["failure_category"], | |
| "gold_reasoning": gold_reasoning, | |
| "gold_category": row["failure_category"], | |
| "difficulty": None, # L3 uses judge scoring, no difficulty | |
| "context_text": context_text, | |
| "metadata": { | |
| "result_id": int(row["result_id"]), | |
| "source_trial_id": row.get("source_trial_id"), | |
| "intervention_name": row.get("intervention_name"), | |
| "condition_name": row.get("condition_name"), | |
| "confidence_tier": row.get("confidence_tier"), | |
| "therapeutic_area": row.get("therapeutic_area"), | |
| "molecular_type": row.get("molecular_type"), | |
| "chembl_id": row.get("chembl_id"), | |
| "n_targets": len(targets), | |
| }, | |
| }) | |
| return records | |
| def main(argv: list[str] | None = None) -> int: | |
| from negbiodb_ct.llm_dataset import ( | |
| assign_splits, | |
| write_dataset_metadata, | |
| write_jsonl, | |
| ) | |
| parser = argparse.ArgumentParser(description="Build CT-L3 reasoning dataset") | |
| parser.add_argument("--db-path", type=Path, default=PROJECT_ROOT / "data" / "negbiodb_ct.db") | |
| parser.add_argument("--output-dir", type=Path, default=OUTPUT_DIR) | |
| parser.add_argument("--seed", type=int, default=42) | |
| args = parser.parse_args(argv) | |
| if not args.db_path.exists(): | |
| logger.error("CT database not found: %s", args.db_path) | |
| return 1 | |
| records = build_l3_dataset(args.db_path, args.seed) | |
| if not records: | |
| logger.error("No records generated!") | |
| return 1 | |
| df = pd.DataFrame(records) | |
| df = assign_splits(df, fewshot_size=20, val_size=20, test_size=160, seed=args.seed) | |
| output_records = [] | |
| for i, (_, row) in enumerate(df.iterrows()): | |
| rec = row.to_dict() | |
| rec["question_id"] = f"CTL3-{i:04d}" | |
| output_records.append(rec) | |
| output_path = args.output_dir / "ct_l3_dataset.jsonl" | |
| write_jsonl(output_records, output_path) | |
| from collections import Counter | |
| splits = Counter(r["split"] for r in output_records) | |
| cats = Counter(r["gold_category"] for r in output_records) | |
| logger.info("=== CT-L3 Dataset Summary ===") | |
| logger.info("Total: %d", len(output_records)) | |
| logger.info("Categories: %s", dict(cats)) | |
| logger.info("Splits: %s", dict(splits)) | |
| write_dataset_metadata(args.output_dir, "ct-l3", { | |
| "total": len(output_records), | |
| "categories": dict(cats), | |
| "splits": dict(splits), | |
| }) | |
| return 0 | |
| if __name__ == "__main__": | |
| sys.exit(main()) | |