Datasets:
NegBioDB — Execution Roadmap
Last updated: 2026-03-30 (v19 — DTI ✅ CT ✅ PPI ✅ GE near-complete: ML seed 42 done, LLM 4/5 models done)
Critical Findings (Updated March 2026)
- HCDT 2.0 License: CC BY-NC-ND 4.0 — Cannot redistribute derivatives. Must independently recreate from underlying sources (BindingDB, ChEMBL, GtoPdb, PubChem, TTD). Use 10 uM primary threshold (not 100 uM) to differentiate.
- InertDB License: CC BY-NC — Cannot include in commercial track. Provide optional download script only.
- Submission requirements: downloadable data, Croissant metadata, code available, Datasheet for Datasets.
- LIT-PCBA compromised (2025 audit found data leakage) — Creates urgency for NegBioDB as replacement gold-standard.
- Recommended NegBioDB License: CC BY-SA 4.0 — Compatible with ChEMBL (CC BY-SA 3.0) via one-way upgrade.
- No direct competitor exists as of March 2026.
- No LLM benchmark tests negative DTI tasks — ChemBench, Mol-Instructions, MedQA, SciBench all lack negative result evaluation. NegBioBench LLM track is first-of-kind.
- LLM evaluation also free — Gemini Flash free tier as LLM-as-Judge + ollama local models as baselines. Flagship models (GPT-4, Claude) added post-stabilization only.
- Data volume is NOT the bottleneck — ChEMBL alone has ~527K quality inactive records (pchembl < 5, validated). PubChem has ~61M target-annotated confirmatory inactives. Estimated 200K+ unique compound-target pairs available. Minimum target raised to 10K curated entries (from 5K).
- PubChem FTP bulk is far superior to API —
bioactivities.tsv.gz(3 GB) contains all 301M bioactivity rows. Processing: < 1 day. API approach would take weeks. - LLM-as-Judge rate limit (250 RPD) — Must-have tasks (L1, L2, L4) all use automated evaluation. Judge needed only for should-have L3 (1,530 calls = 6 days). All judge tasks with 3 models = 20 days. With 6 models = 39 days (NOT feasible for sprint).
- Paper narrative must be problem-first — "Existing benchmarks are broken" (Exp 1 + Exp 4), not "Here's a database." Database is the solution, not the contribution.
- Positive data protocol required — NegBioDB is negative-only. For ML benchmarking (M1), positive data must be sourced from ChEMBL (pChEMBL ≥ 6). Report two class ratios: balanced (1:1) and realistic (1:10). See §Positive Data Protocol below.
- Random negative baseline must be precisely defined — Exp 1 compares NegBioDB negatives against random negatives. Random = uniform sampling from untested compound-target pairs (TDC standard). See §Random Negative Control Design.
- Paper format: 9 pages + unlimited appendix. Croissant is mandatory (desk rejection if missing/invalid).
- GPU strategy: Kaggle free tier (30 hrs/week) is sufficient for 18 ML baseline runs (~36-72 GPU-hours over 4 weeks). Fallback: Colab Pro ($10/month).
- ChEMBL v36 (Sep 2025, 24.3M activities) should be used, not v35.
chembl_downloaderfetches latest by default. - Nature MI 2025 — Biologically driven negative subsampling paper independently shows "assumed negatives" distort DTI models. Related: EviDTI (Nature Comms 2025), DDB paper (BMC Biology 2025), LIT-PCBA audit (2025).
Positive Data Protocol (P0 — Expert Panel Finding)
NegBioDB is a negative-only database. For ML benchmarking (Task M1: binary DTI prediction), positive (active) data is required. This section defines the protocol.
Positive Data Source
-- Extract active DTIs from ChEMBL v36 SQLite
-- Threshold: pChEMBL ≥ 6 (IC50/Ki/Kd/EC50 ≤ 1 uM)
SELECT
a.molregno, a.pchembl_value, a.standard_type,
cs.canonical_smiles, cs.standard_inchi_key,
cp.accession AS uniprot_id
FROM activities a
JOIN compound_structures cs ON a.molregno = cs.molregno
JOIN assays ass ON a.assay_id = ass.assay_id
JOIN target_dictionary td ON ass.tid = td.tid
LEFT JOIN target_components tc ON td.tid = tc.tid
LEFT JOIN component_sequences cp ON tc.component_id = cp.component_id
WHERE a.pchembl_value >= 6
AND a.standard_type IN ('IC50', 'Ki', 'Kd', 'EC50')
AND a.data_validity_comment IS NULL
AND td.target_type = 'SINGLE PROTEIN'
AND cp.accession IS NOT NULL
Positive-Negative Pairing
| Setting | Ratio | Purpose | Primary Use |
|---|---|---|---|
| Balanced | 1:1 (active:inactive) | Fair model comparison | Exp 1, Exp 4, baselines |
| Realistic | 1:10 (active:inactive) | Real-world HTS simulation | Supplementary evaluation |
- Positives restricted to shared targets between ChEMBL actives and NegBioDB inactives (same target pool)
- Same compound standardization pipeline (RDKit) applied to positives
- DAVIS matrix known actives (pKd ≥ 7, Kd ≤ 100 nM) used as Gold-standard validation set
Overlap Prevention
- Active and inactive compound-target pairs must not overlap (same pair cannot be both active and inactive)
- Borderline zone (pChEMBL 4.5–5.5) excluded from both positive and negative sets for clean separation
- Overlap analysis: report % of NegBioDB negatives where the same compound appears as active against a different target
Random Negative Control Design (P0 — Expert Panel Finding)
Experiment 1 compares NegBioDB's experimentally confirmed negatives against random negatives. The random negative generation must be precisely defined.
Control Conditions for Exp 1
| Control | Method | What it Tests |
|---|---|---|
| Uniform random | Sample untested compound-target pairs uniformly at random from the full cross-product space | Standard TDC approach; tests baseline inflation |
| Degree-matched random | Sample untested pairs matching the degree distribution of NegBioDB pairs | Isolates the effect of experimental confirmation vs. degree bias |
All Exp 1 runs:
- 3 ML models (DeepDTA, GraphDTA, DrugBAN)
- Random split only (for controlled comparison)
- Same positive data, same split seed
- Only the negative set changes: NegBioDB confirmed vs. uniform random vs. degree-matched random
- Total: 3 models × 3 negative conditions = 9 runs (was 3 runs; updated)
- Note: The 3 NegBioDB-negative random-split runs are shared with the baseline count (9 baselines include random split). Thus Exp 1 adds only 6 new runs (uniform random + degree-matched random). Similarly, Exp 4 shares the random-split baseline and adds only 3 new DDB runs. Overall: 9 baseline + 6 Exp 1 + 3 Exp 4 = 18 total.
- Exp 4 definition: The DDB comparison uses a full-task degree-balanced split on the merged M1 balanced benchmark. Positives and negatives are reassigned together under the same split policy.
Reporting
- Table: [Model × Negative Source × Metric] for LogAUC, AUPRC, MCC
- Expected: NegBioDB > degree-matched > uniform random for precision-oriented metrics
- If NegBioDB ≈ uniform random → narrative shifts to Exp 4 (DDB bias) as primary result
Phase 1: Implementation Sprint (Weeks 0-11)
Week 1: Scaffolding + Download + Schema ✅ COMPLETE
- Project scaffolding: Create
src/negbiodb/,scripts/,tests/,migrations/,config.yaml,Makefile,pyproject.toml - Dependency management:
pyproject.tomlwith Python 3.11+, rdkit, pandas, pyarrow, mlcroissant, tqdm, scikit-learn - Makefile skeleton: Define target structure (full pipeline encoding in Week 2)
- Finalize database schema (SQLite for MVP) — apply
migrations/001_initial_schema.sql - Download all source data (see below — < 1 day total)
- Verify ChEMBL v36 (Sep 2025) downloaded, not v35
- [B7] Verify PubChem bioactivities.tsv.gz column names after download
- [B4] Hardware decision: Test local RAM/GPU. If < 32GB RAM → use Llama 3.1 8B + Mistral 7B (not 70B). If ≥ 32GB → quantized Llama 3.3 70B (Q4). Document choice.
- [B2] Verify citations: Search for Nature MI 2025 negative subsampling paper + Science 2025 editorial. If not found → substitute with EviDTI, DDB paper, LIT-PCBA audit
- [B3] Monitor submission deadlines
Week 2: Standardization + Extraction Start ✅ COMPLETE
- Implement compound standardization pipeline (RDKit: salt removal, normalization, InChIKey)
- Implement target standardization pipeline (UniProt accession as canonical ID)
- Set up cross-DB deduplication (InChIKey[0:14] connectivity layer)
- Makefile pipeline: Encode full data pipeline dependency graph as executable Makefile targets
- [B5] Check shared target pool size: Count intersection of NegBioDB targets ∩ ChEMBL pChEMBL ≥ 6 targets. If < 200 targets → expand NegBioDB target extraction
- [B6] Check borderline exclusion impact: Run pChEMBL distribution query on ChEMBL. Estimate data loss from excluding pChEMBL 4.5–5.5 zone
Week 2-4: Data Extraction ✅ COMPLETE
Result: 30.5M negative_results (>minimum target of 10K — far exceeded)
Data Sources (License-Safe Only):
| Source | Available Volume | Method | License |
|---|---|---|---|
| PubChem BioAssay (confirmatory inactive) | ~61M (target-annotated) | FTP bulk: bioactivities.tsv.gz (3 GB) + bioassays.tsv.gz (52 MB) |
Public domain |
| ChEMBL pChEMBL < 5 (quality-filtered) | ~527K records → ~100-200K unique pairs | SQLite via chembl_downloader (4.6 GB, 1h setup) |
CC BY-SA 3.0 |
| ChEMBL activity_comment "Not Active" | ~763K (literature-curated) | SQL query on same SQLite dump | CC BY-SA 3.0 |
| BindingDB (Kd/Ki > 10 uM) | ~30K+ | Bulk TSV download + filter | CC BY |
| DAVIS complete matrix (pKd ≤ 5) | ~27K | TDC Python download | Public/academic |
NOT bundled (license issues):
- HCDT 2.0 (CC BY-NC-ND) — Use as validation reference only; we use 10 uM threshold (not 100 uM) to differentiate
- InertDB (CC BY-NC) — Optional download script for users
PubChem FTP extraction pipeline (< 1 day):
1. bioassays.tsv.gz → filter confirmatory AIDs with target annotations → ~260K AIDs
2. bioactivities.tsv.gz (stream) → filter AID ∈ confirmatory, Outcome=Inactive → ~61M records
3. Prioritize MLPCN/MLSCN assays (~4,500 AIDs, genuine HTS dose-response) for Silver tier
4. Map SID→CID via Sid2CidSMILES.gz, targets via Aid2GeneidAccessionUniProt.gz
- Download PubChem FTP files (bioactivities.tsv.gz + bioassays.tsv.gz + mapping files)
- Download ChEMBL v36 SQLite via chembl_downloader
- Download BindingDB bulk TSV
- Build PubChem FTP extraction script (streaming with chunksize=100K — 12GB uncompressed)
- Build ChEMBL extraction SQL: inactive (activity_comment + pChEMBL < 5) AND active (pChEMBL ≥ 6) for positive data
- Build BindingDB extraction script (filter Kd/Ki > 10 uM, human targets)
- Integrate DAVIS matrix from TDC (both actives pKd ≥ 7 and inactives pKd ≤ 5)
- Run compound/target standardization on all extracted data (multiprocessing for RDKit)
- Run cross-DB deduplication + overlap analysis (vs DAVIS, TDC, DUD-E, LIT-PCBA)
- Assign confidence tiers (gold/silver/bronze/copper — lowercase, matching DDL CHECK constraint)
- Extract ChEMBL positives: 883K → 863K after 21K overlap removal (pChEMBL ≥ 6, shared targets only)
- Positive-negative pairing: M1 balanced (1.73M, 1:1) + M1 realistic (9.49M, 1:10). Zero compound-target overlap verified.
- Borderline exclusion: pChEMBL 4.5–5.5 removed from both pools
- Spot-check top 100 most-duplicated compounds (manual QC checkpoint)
- Run data leakage check: cold split leaks = 0, cross-source overlaps documented
Week 3-5: Benchmark Construction (ML + LLM)
ML Track:
- Implement 3 must-have splits (Random, Cold-Compound, Cold-Target) + DDB for Exp 4
- Implement ML evaluation metrics: LogAUC[0.001,0.1], BEDROC, EF@1%, EF@5%, AUPRC, MCC, AUROC
- (Should have) Add Cold-Both, Temporal, Scaffold splits (all 6 implemented)
LLM Track: ✅ INFRASTRUCTURE COMPLETE (2026-03-12)
- Design prompt templates for L1, L2, L4 (priority tasks) →
llm_prompts.py - Construct L1 dataset: 2,000 MCQ from NegBioDB entries →
build_l1_dataset.py - Construct L2 dataset: 116 candidates (semi-automated) →
build_l2_dataset.py - Construct L4 dataset: 500 tested/untested pairs →
build_l4_dataset.py - Implement automated evaluation scripts →
llm_eval.py(L1: accuracy/F1, L2: entity F1, L4: classification F1) - Build compound name cache →
compound_names.parquet(144,633 names from ChEMBL) - Construct L3 dataset: 50 pilot reasoning examples →
build_l3_dataset.py - LLM client (vLLM + Gemini) →
llm_client.py - SLURM templates + batch submission →
run_llm_local.slurm,run_llm_gemini.slurm,submit_llm_all.sh - Results aggregation →
collect_llm_results.py(Table 2) - 54 new tests (29 eval + 25 dataset), 329 total pass
- L2 gold annotation: 15–20h human review needed for
l2_gold.jsonl
Shared:
- Generate Croissant machine-readable metadata (mandatory for submission)
- Validate Croissant with
mlcroissantlibrary. Gate:mlcroissant.Dataset('metadata.json')runs without errors - Write Datasheet for Datasets (Gebru et al. template)
Week 5-7: Baseline Experiments (ML + LLM)
ML Baselines:
| Model | Type | Priority | Runs (3 splits) | Status |
|---|---|---|---|---|
| DeepDTA | Sequence CNN | Must have | 3 | ✅ Implemented |
| GraphDTA | Graph neural network | Must have | 3 | ✅ Implemented |
| DrugBAN | Bilinear attention | Must have | 3 | ✅ Implemented |
| Random Forest | Traditional ML | Should have | 3 | Planned |
| XGBoost | Traditional ML | Should have | 3 | Planned |
| DTI-LM | Language model-based | Nice to have | 3 | Planned |
| EviDTI | Evidential/uncertainty | Nice to have | 3 | Planned |
Must-have ML: 9 baseline runs (3 models × 3 splits) + 6 Exp 1 (2 random conditions) + 3 Exp 4 (DDB split) = 18 total (~36-72 GPU-hours, 3-4 days)
Status (2026-03-13): All 18/18 ML baseline runs COMPLETE on Cayuga HPC. Results in
results/baselines/. 3 timed-out DrugBAN jobs recovered viaeval_checkpoint.py. Key findings: degree-matched negatives inflate LogAUC by +0.112 avg; cold-target LogAUC drops to 0.15–0.33; DDB ≈ random (≤0.010 diff).
LLM Baselines (all free):
| Model | Access | Priority |
|---|---|---|
| Gemini 2.5 Flash | Free API (250 RPD) | Must have |
| Llama 3.3 70B | Ollama local | Must have |
| Mistral 7B | Ollama local | Must have |
| Phi-3.5 3.8B | Ollama local | Should have |
| Qwen2.5 7B | Ollama local | Should have |
Must-have LLM: 3 models × 3 tasks (L1,L2,L4) × 2 configs (zero-shot, 3-shot) = 18 eval runs (all automated)
Flagship models (post-stabilization):
- GPT-4/4.1, Claude Sonnet/Opus, Gemini Pro — added to leaderboard later
Must-have experiments (minimum for paper):
- Exp 1: NegBioDB vs. random negatives ✅ COMPLETE — degree-matched avg +0.112 over negbiodb → benchmark inflation confirmed
- Exp 4: Node degree bias ✅ COMPLETE — DDB ≈ random (≤0.010 diff) → degree balancing alone not harder
- Exp 9: LLM vs. ML comparison (L1 vs. M1 on matched test set — reuses baseline results; awaiting LLM runs)
- Exp 10: LLM extraction quality (L2 entity F1 — awaiting LLM runs)
Should-have experiments (strengthen paper, no extra training):
- Exp 5: Cross-database consistency (analysis only, no training)
- Exp 7: Target class coverage analysis (analysis only)
- Exp 11: Prompt strategy comparison (add CoT config to LLM baselines)
- L3 task + Exp 12: LLM-as-Judge reliability (1,530 judge calls = 6 days)
Nice-to-have experiments (defer to camera-ready):
- Exp 2: Confidence tier discrimination
- Exp 3: Assay context dependency (with assay format stratification)
- Exp 6: Temporal generalization
- Exp 8: LIT-PCBA recapitulation
Week 8-10: Paper Writing
- Write benchmark paper (9 pages + unlimited appendix)
- Create key figures (see
paper/scripts/generate_figures.py) - Paper structure (9 pages): Intro (1.5) → DB Design (1.5) → Benchmark (1.5) → Experiments (3) → Discussion (1.5)
- Appendix contents: Full schema DDL, all metric tables, L2 annotation details, few-shot examples, Datasheet
- Python download script:
pip install negbiodbor simple wget script - Host dataset (HuggingFace primary + Zenodo DOI for archival)
- Author ethical statement
- Dockerfile for full pipeline reproducibility: Python 3.11, rdkit, torch, chembl_downloader, pyarrow, mlcroissant. Must reproduce full pipeline from raw data → final benchmark export
Week 10-11: Review & Submit
- Internal review and polish
- Submit abstract (~May 1)
- Submit full paper (~May 15)
- Post ArXiv preprint (same day or before submission)
Phase 1-CT: Clinical Trial Failure Domain
Initiated: 2026-03-17 | Pipeline code + data loading complete, benchmark design complete
Step CT-1: Infrastructure ✅ COMPLETE
- CT schema design (2 migrations: 001 initial + 002 expert review fixes)
- 5 pipeline modules: etl_aact, etl_classify, drug_resolver, etl_outcomes, ct_db
- 138 tests passing
- Data download scripts for all 4 sources
Step CT-2: Data Loading ✅ COMPLETE
- AACT ETL: 216,987 trials, 476K trial-interventions, 372K trial-conditions
- Failure classification (3-tier): 132,925 results (bronze 60K / silver 28K / gold 23K / copper 20K)
- Open Targets: 32,782 intervention-target mappings
- Pair aggregation: 102,850 intervention-condition pairs
Step CT-3: Enrichment & Resolution ✅ COMPLETE
- Outcome enrichment: +66 AACT p-values, +31,969 Shi & Du SAE records
- Drug resolution Steps 1-2: ChEMBL exact (18K) + PubChem API
- Drug resolution Step 3: Fuzzy matching — 15,616 resolved
- Drug resolution Step 4: Manual overrides — 291 resolved (88 entries used)
- Pair aggregation refresh (post-resolution) — 102,850 pairs
- Post-run coverage analysis — 36,361/176,741 (20.6%) ChEMBL, 27,534 SMILES, 66,393 targets
Step CT-4: Analysis & Benchmark Design ✅ COMPLETE
- Data quality analysis script (
scripts_ct/analyze_ct_data.py) — 16 queries, JSON+MD output - Data quality report (
results/ct/ct_data_quality.md) - ML benchmark design
- 3 tasks: CT-M1 (binary), CT-M2 (7-way category), CT-M3 (phase transition, deferred)
- 6 split strategies, 3 models (XGBoost, MLP, GNN+Tabular)
- 3 experiments: negative source, generalization, temporal
- LLM benchmark design
- 4 levels: CT-L1 (5-way MCQ), CT-L2 (extraction), CT-L3 (reasoning), CT-L4 (discrimination)
- 5 models, anti-contamination analysis
Step CT-5: ML Export & Splits ✅ COMPLETE
- CT export module (
src/negbiodb_ct/ct_export.py) - CTO success trials extraction (CT-M1 positive class)
- Feature engineering (drug FP + mol properties + condition one-hot + trial design)
- 6 split strategies implementation
Step CT-6: ML Baseline Experiments ✅ COMPLETE (108/108 runs)
- XGBoost baseline (CT-M1 + CT-M2)
- MLP baseline
- GNN+Tabular baseline
- Key finding: CT-M1 trivially separable on NegBioDB negatives (AUROC=1.0); M2 XGBoost macro-F1=0.51
Step CT-7: LLM Benchmark Execution ✅ COMPLETE (80/80 runs)
- CT-L1/L2/L3/L4 dataset construction
- CT prompt templates + evaluation functions
- Inference runs on Cayuga HPC (5 models × 4 levels × 4 configs)
- Key finding: CT L4 MCC 0.48–0.56 — highest discrimination across domains
Phase 1b: Post-Submission Expansion (Months 3-6)
Data Expansion (if not at 10K+ for submission)
- Complete PubChem BioAssay extraction (full confirmatory set)
- LLM text mining pipeline activation (PubMed abstracts)
- Supplementary materials table extraction (pilot)
Benchmark Refinement
- Add remaining ML and LLM baseline models
- Complete all 12 validation experiments (8 ML + 4 LLM)
- Complete LLM tasks L5, L6 datasets
- Add flagship LLM evaluations (GPT-4, Claude)
- Build public leaderboard (simple GitHub-based, separate ML and LLM tracks)
Phase 2: Community & Platform (Months 6-18)
2.1 Platform Development
- Web interface (search, browse, download)
- Python library:
pip install negbiodb - REST API with tiered access
- Community submission portal with controlled vocabularies
- Leaderboard system
2.2 Community Building
- GitHub repository with documentation and tutorials
- Partner with SGC and Target 2035/AIRCHECK for data access
- Engage with DREAM challenge community
- Tutorial at relevant workshop
- Researcher incentive design (citation credit, DOI per submission)
Schema Design
Common Layer
NegativeResult {
id: UUID
compound_id: InChIKey + ChEMBL ID + PubChem CID
target_id: UniProt ID + ChEMBL Target ID
// Core negative result
result_type: ENUM [hard_negative, conditional_negative, methodological_negative,
hypothesis_negative, dose_time_negative]
confidence_tier: ENUM [gold, silver, bronze, copper]
// Quantitative evidence
activity_value: FLOAT (IC50, Kd, Ki, EC50)
activity_unit: STRING
activity_type: STRING
pchembl_value: FLOAT
inactivity_threshold: FLOAT
max_concentration_tested: FLOAT
// Assay context (BAO-based)
assay_type: BAO term
assay_format: ENUM [biochemical, cell-based, in_vivo]
assay_technology: STRING
detection_method: STRING
cell_line: STRING (if cell-based)
organism: STRING
// Quality metrics
z_factor: FLOAT
ssmd: FLOAT
num_replicates: INT
screen_type: ENUM [primary_single_point, confirmatory_dose_response,
counter_screen, orthogonal_assay]
// Provenance
source_db: STRING (PubChem, ChEMBL, literature, community)
source_id: STRING (assay ID, paper DOI)
extraction_method: ENUM [database_direct, text_mining, llm_extracted,
community_submitted]
curator_validated: BOOLEAN
// Target context (DTO-based)
target_type: DTO term
target_family: STRING (kinase, GPCR, ion_channel, etc.)
target_development_level: ENUM [Tclin, Tchem, Tbio, Tdark]
// Metadata
created_at: TIMESTAMP
updated_at: TIMESTAMP
related_positive_results: [UUID] (links to known actives for same target)
}
Biology/DTI Domain Layer
DTIContext {
negative_result_id: UUID (FK)
binding_site: STRING (orthosteric, allosteric, unknown)
selectivity_data: BOOLEAN (part of selectivity panel?)
species_tested: STRING
counterpart_species_result: STRING (active in other species?)
cell_permeability_issue: BOOLEAN
compound_solubility: FLOAT
compound_stability: STRING
}
Benchmark Design (NegBioBench) — Dual ML + LLM Track
Track A: Traditional ML Tasks
| Task | Input | Output | Primary Metric |
|---|---|---|---|
| M1: DTI Binary Prediction | (compound SMILES, target sequence) | Active / Inactive | LogAUC[0.001,0.1], AUPRC |
| M2: Negative Confidence Prediction | (SMILES, sequence, assay features) | gold/silver/bronze/copper | Weighted F1, MCC |
| M3: Activity Value Regression | (SMILES, sequence) | pIC50 / pKd | RMSE, R², Spearman ρ |
ML Baselines: DeepDTA, GraphDTA, DrugBAN, RF, XGBoost, DTI-LM, EviDTI
Track B: LLM Tasks
| Task | Input | Output | Metric | Eval Method |
|---|---|---|---|---|
| L1: Negative DTI Classification | Natural language description | Active/Inactive/Inconclusive/Conditional (MCQ) | Accuracy, F1, MCC | Automated |
| L2: Negative Result Extraction | Paper abstract | Structured JSON (compound, target, outcome) | Schema compliance, Entity F1, STED | Automated |
| L3: Inactivity Reasoning | Confirmed negative + context | Scientific explanation | 4-dim rubric (accuracy, reasoning, completeness, specificity) | LLM-as-Judge + human sample |
| L4: Tested-vs-Untested Discrimination | Compound-target pairs | Tested/Untested + evidence | Accuracy, F1, evidence quality | Automated + spot-check |
| L5: Assay Context Reasoning | Negative result + condition changes | Prediction + reasoning per scenario | Prediction accuracy, reasoning quality | LLM-as-Judge |
| L6: Evidence Quality Assessment | Negative result + metadata | Confidence tier + justification | Tier F1, justification quality | Automated + LLM-judge |
LLM Baselines (Phase 1 — Free): Gemini 2.5 Flash, Llama 3.3, Mistral 7B, Phi-3.5, Qwen2.5 LLM Baselines (Phase 2 — Flagship): GPT-4, Claude Sonnet/Opus, Gemini Pro LLM-as-Judge: Gemini 2.5 Flash free tier (validated against human annotations)
Track C: Cross-Track (Future)
| Task | Description |
|---|---|
| C1: Ensemble Prediction | Combine ML model scores + LLM reasoning — does LLM improve ML? |
Splitting Strategies (7 total, for Track A)
- Random (stratified 70/10/20)
- Cold compound (Butina clustering on Murcko scaffolds)
- Cold target (by UniProt accession)
- Cold both (compound + target unseen)
- Temporal (train < 2020, val 2020-2022, test > 2022)
- Scaffold (Murcko scaffold cluster-based)
- DDB — Degree Distribution Balanced (addresses node degree bias)
Evaluation Metrics (Track A)
| Metric | Type | Role |
|---|---|---|
| LogAUC[0.001,0.1] | Enrichment | Primary ranking metric |
| BEDROC (α=20) | Enrichment | Early enrichment |
| EF@1%, EF@5% | Enrichment | Top-ranked performance |
| AUPRC | Ranking | Secondary ranking metric |
| MCC | Classification | Balanced classification |
| AUROC | Ranking | Backward compatibility only (not for ranking) |
LLM Evaluation Configuration
- Full benchmark (5 configs): zero-shot, 3-shot, 5-shot, CoT, CoT+3-shot
- Must-have (2 configs): zero-shot, 3-shot only (see research/08 §3)
- Should-have (add CoT): 3 configs total for Exp 11 (prompt strategy comparison)
- 3 runs per evaluation, report mean ± std
- Temperature = 0, prompts version-controlled
- Anti-contamination: temporal holdout + paraphrased variants + contamination detection
Phase 3: Scale & Sustainability (Months 18-36)
3.1 Data Expansion
- Expand to 100K+ curated negative DTIs
- Full LLM-based literature mining pipeline (PubMed/PMC)
- Supplementary materials table extraction (Table Transformer)
- Integrate Target 2035 AIRCHECK data as it becomes available
- Begin Gene Function (KO/KD) negative data collection
3.2 Benchmark Evolution (NegBioBench v1.0)
- Track A expansion: multi-modal integration (protein structures, assay images)
- Track B expansion: additional tasks — Failure Diagnosis, Experimental Design Critique, Literature Contradiction Detection
- Track C: Cross-track ensemble evaluation (ML + LLM combined prediction)
- Specialized bio-LLM evaluations (LlaSMol, BioMedGPT, DrugChat)
- Regular leaderboard updates (both ML and LLM tracks)
Phase 4: Domain Expansion (Months 36+)
DTI (Phase 1 — COMPLETE)
│
├── Clinical Trial Failure (Phase 1-CT — COMPLETE ✅)
│ └── 132,925 failure results loaded, benchmarks designed
│
├── Gene Function (CRISPR KO/KD negatives)
│ └── Leverage CRISPR screen data, DepMap
│
├── Chemistry Domain Layer
│ └── Failed reactions, yield = 0 data
│
└── Materials Science Domain Layer
└── HTEM DB integration, failed synthesis conditions
Key Milestones (Revised)
| Milestone | Target Date | Deliverable | Status |
|---|---|---|---|
| Schema v1.0 finalized | Week 2 (Mar 2026) | SQLite schema + standardization pipeline | ✅ Done |
| Data extraction complete | Week 3-4 (Mar 2026) | 30.5M negative results (far exceeded 10K target) | ✅ Done |
| ML export & splits | Week 3 (Mar 2026) | 6 split strategies + M1 benchmark datasets | ✅ Done |
| ML evaluation metrics | Week 3 (Mar 2026) | 7 metrics, 329 tests | ✅ Done |
| ML baseline infrastructure | Week 4 (Mar 2026) | 3 models + SLURM harness | ✅ Done |
| ML baseline experiments | Week 5 (Mar 2026) | 18/18 runs complete, key findings confirmed | ✅ Done |
| LLM benchmark infrastructure | Week 5 (Mar 2026) | L1–L4 datasets, prompts, eval, SLURM templates | ✅ Done |
| LLM benchmark execution | Week 5-6 (Mar 2026) | 81/81 runs complete (9 models × 4 tasks + configs) | ✅ Done |
| Python library v0.1 | Month 8 | pip install negbiodb |
|
| Web platform launch | Month 12 | Public access + leaderboard | |
| 100K+ entries | Month 24 | Scale milestone |