NegBioDB / PROJECT_OVERVIEW.md
jang1563's picture
NegBioDB final: 4 domains, fully audited
6d1bbc7
# NegBioDB: Negative Results Database & Dual ML/LLM Benchmark
> Biology-first, science-extensible negative results database and dual ML+LLM benchmark
*Last updated: 2026-03-30*
---
## Project Vision
Approximately 90% of scientific experiments produce null or inconclusive results, yet the vast majority remain unpublished. This systematic gap fundamentally distorts AI/ML model training and evaluation.
**Goal:** Systematically collect and structure experimentally confirmed negative results across biomedical domains, and build benchmarks that quantify the impact of publication bias on AI/ML models.
## Why This Matters
1. **Publication Bias**: 85% of published papers report only positive results
2. **AI Model Bias**: Models trained without negative data produce excessive false positives
3. **Economic Waste**: Duplicated experiments, failed drug discovery pipelines (billions of dollars)
4. **Proven Impact**: Models trained with negative data are more accurate (Organic Letters 2023, bioRxiv 2024)
---
## Architecture
```
Four Biomedical Domains
┌────────────────────────────────────────────────────────────┐
│ NegBioDB │
│ DTI CT PPI GE │
│ (30.5M neg) (133K neg) (2.2M neg) (28.8M neg) │
│ ChEMBL+ AACT+ IntAct+ DepMap │
│ PubChem+ CTO+ HuRI+ CRISPR+RNAi │
│ BindingDB+ OpenTargets+ hu.MAP+ │
│ DAVIS Shi&Du STRING │
└────────────────────────────────────────────────────────────┘
│ │
┌──────┴──────┐ ┌─────┴──────┐
│ ML Benchmark │ │LLM Benchmark│
│ 3 models × │ │ 5 models × │
│ 5 splits × │ │ 4 levels × │
│ 2 neg types │ │ 4 configs │
└─────────────┘ └────────────┘
```
## Key Technical Decisions
| Decision | Choice | Rationale |
|----------|--------|-----------|
| License | CC BY-SA 4.0 | Compatible with ChEMBL CC BY-SA 3.0 (viral clause) |
| Storage | SQLite per domain | Portable, zero-infrastructure, reproducible |
| Export | Parquet with split columns | Standard ML format; lazy-loading friendly |
| ML metrics | LogAUC + 6 others | LogAUC[0.001,0.1] measures early enrichment, not just AUROC |
| LLM evaluation | 4 levels (L1–L4) | Progressive difficulty: MCQ → extraction → reasoning → discrimination |
---
## Domain Status Summary (as of 2026-03-30)
| Domain | DB Size | Negatives | ML Runs | LLM Runs | Status |
|--------|---------|-----------|---------|----------|--------|
| **DTI** | ~21 GB | 30,459,583 | 24/24 ✅ | 81/81 ✅ | Complete |
| **CT** | ~500 MB | 132,925 | 108/108 ✅ | 80/80 ✅ | Complete |
| **PPI** | 849 MB | 2,229,670 | 54/54 ✅ | 80/80 ✅ | Complete |
| **GE** | ~16 GB | 28,759,256 | Seed 42 ✅ | 4/5 models ✅ | Near complete |
---
## DTI Domain (Drug-Target Interaction)
Four sources: ChEMBL v36, PubChem BioAssay, BindingDB, DAVIS
### Database
- **30,459,583** negative results
- Source tiers: gold 818,611 / silver 198 / bronze 28,845,632
- 5 split strategies: random / cold_compound / cold_target / scaffold / temporal
### Key Results
- **ML:** Degree-matched negatives inflate LogAUC by +0.112 on average. Cold-target splits catastrophic (LogAUC 0.15–0.33) while AUROC stays deceptively high (0.76–0.89).
- **LLM L4:** All models near-random (MCC ≤ 0.18). DTI binding decisions are too nuanced for LLMs without domain context.
- **LLM L1:** Gemini achieves perfect accuracy (1.000) on 3-shot MCQ — artifact of format recognition.
---
## CT Domain (Clinical Trial Failure)
Four sources: AACT (ClinicalTrials.gov), CTO, Open Targets, Shi & Du 2024
### Database
- **132,925** failure results from 216,987 trials
- Tiers: gold 23,570 / silver 28,505 / bronze 60,223 / copper 20,627
- 8 failure categories: safety > efficacy > enrollment > strategic > regulatory > design > other
- Drug resolution: 4-step pipeline (ChEMBL exact → PubChem API → fuzzy JaroWinkler → manual CSV)
### Benchmark Design
- **ML:** CT-M1 binary failure prediction; CT-M2 7-way failure category (most challenging)
- **LLM:** L1 5-way MCQ (1,500 items), L2 failure report extraction (500), L3 reasoning (200), L4 discrimination (500)
### Key Results
- **CT-M1:** NegBioDB negatives trivially separable (AUROC=1.0). Control negatives reveal real difficulty (0.76–0.84).
- **CT-M2:** XGBoost best (macro-F1=0.51). Scaffold/temporal splits hardest (0.19).
- **LLM L4:** Gemini MCC=0.56 — highest across all domains. Meaningful discrimination possible for trial failure.
- **LLM L3:** Ceiling effect — GPT-4o-mini judge too lenient (4.4–5.0/5.0).
---
## PPI Domain (Protein-Protein Interaction)
Four sources: IntAct, HuRI, hu.MAP 3.0, STRING v12.0
### Database
- **2,229,670** negative results; 61,728 positive pairs (HuRI Y2H)
- 18,412 proteins; 4 split strategies: random / cold_protein / cold_both / degree_balanced
### Key Results
- **ML:** MLPFeatures (hand-crafted) dominates cold splits (AUROC 0.95 cold_both); PIPR collapses to 0.41 (below random).
- **LLM L1:** 3-shot near-perfect (0.997–1.000) is an artifact of example format leakage.
- **LLM L3:** zero-shot >> 3-shot (4.3–4.7 vs 3.1–3.7); gold reasoning examples degrade structural reasoning.
- **LLM L4:** MCC 0.33–0.44 with confirmed temporal contamination (pre-2015 acc ~0.6–0.8, post-2020 acc ~0.07–0.25).
---
## GE Domain (Gene Essentiality / DepMap)
Two sources: DepMap CRISPR (Chronos scores) and RNAi (DEMETER2)
### Database
- **28,759,256** negative results (genes with no essentiality signal)
- Final tiers: Gold 753,878 / Silver 18,608,686 / Bronze 9,396,692
- 19,554 genes × 2,132 cell lines; 22,549,910 aggregated pairs
- 5 split strategies: random / cold_gene / cold_cell_line / cold_both / degree_balanced
### Benchmark Design
- **ML:** XGBoost and MLPFeatures on gene expression + lineage features (gene-cell pair prediction)
- **LLM:** L1 4-way essentiality MCQ (1,200 items), L2 essentiality data extraction (500), L3 reasoning (200), L4 discrimination (475)
### Key Results (partial — Llama pending)
- **LLM L3:** zero-shot >> 3-shot (overall mean 4.5 vs 2.5) — same pattern as PPI.
- **LLM L4:** Expected intermediate MCC (DepMap is widely studied; likely contamination present).
- **ML:** Seed 42 complete; final aggregated results pending seeds 43/44.
---
## Dual Benchmark Framework
### LLM Benchmark Levels
| Level | Task | Difficulty | Automation |
|-------|------|-----------|------------|
| L1 | Multiple-choice classification | Easy | Fully automated (exact match) |
| L2 | Structured field extraction | Medium | Automated (JSON schema check + field F1) |
| L3 | Free-text reasoning quality | Hard | LLM-as-judge (Gemini 2.5-Flash, 4 rubric dimensions) |
| L4 | Real vs synthetic discrimination | Hard | Automated (MCC on binary decision) |
### LLM Models Evaluated
| Model | Provider | Type |
|-------|----------|------|
| Claude Haiku 4.5 | Anthropic API | Small API model |
| Gemini 2.5-Flash | Google API | Small API model |
| GPT-4o-mini | OpenAI API | Small API model |
| Qwen2.5-7B-Instruct | HuggingFace / vLLM | Open-weight local |
| Llama-3.1-8B-Instruct | HuggingFace / vLLM | Open-weight local |
### Cross-Domain LLM L4 Summary
```
DTI (≤0.18) < PPI (0.33–0.44) < CT (0.48–0.56)
Increasing task complexity
and LLM accessible signal
```
---
## Timeline
| Milestone | Date |
|-----------|------|
| Project initiated | 2026-03-02 |
| DTI domain complete (ML + LLM) | 2026-03-13 |
| CT domain initiated | 2026-03-17 |
| CT domain complete (ML + LLM) | 2026-03-20 |
| PPI domain complete (ML + LLM) | 2026-03-23 |
| GE domain ETL + ML export | 2026-03-23 |
| GE LLM (4/5 models) | 2026-03-24 |
| Public release (GitHub + HuggingFace) | 2026-03-30 |