Datasets:
Formats:
parquet
Languages:
English
Size:
10M - 100M
Tags:
biology
chemistry
drug-discovery
clinical-trials
protein-protein-interaction
gene-essentiality
License:
File size: 8,341 Bytes
6d1bbc7 | 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 | # NegBioDB: Negative Results Database & Dual ML/LLM Benchmark
> Biology-first, science-extensible negative results database and dual ML+LLM benchmark
*Last updated: 2026-03-30*
---
## Project Vision
Approximately 90% of scientific experiments produce null or inconclusive results, yet the vast majority remain unpublished. This systematic gap fundamentally distorts AI/ML model training and evaluation.
**Goal:** Systematically collect and structure experimentally confirmed negative results across biomedical domains, and build benchmarks that quantify the impact of publication bias on AI/ML models.
## Why This Matters
1. **Publication Bias**: 85% of published papers report only positive results
2. **AI Model Bias**: Models trained without negative data produce excessive false positives
3. **Economic Waste**: Duplicated experiments, failed drug discovery pipelines (billions of dollars)
4. **Proven Impact**: Models trained with negative data are more accurate (Organic Letters 2023, bioRxiv 2024)
---
## Architecture
```
Four Biomedical Domains
┌────────────────────────────────────────────────────────────┐
│ NegBioDB │
│ DTI CT PPI GE │
│ (30.5M neg) (133K neg) (2.2M neg) (28.8M neg) │
│ ChEMBL+ AACT+ IntAct+ DepMap │
│ PubChem+ CTO+ HuRI+ CRISPR+RNAi │
│ BindingDB+ OpenTargets+ hu.MAP+ │
│ DAVIS Shi&Du STRING │
└────────────────────────────────────────────────────────────┘
│ │
┌──────┴──────┐ ┌─────┴──────┐
│ ML Benchmark │ │LLM Benchmark│
│ 3 models × │ │ 5 models × │
│ 5 splits × │ │ 4 levels × │
│ 2 neg types │ │ 4 configs │
└─────────────┘ └────────────┘
```
## Key Technical Decisions
| Decision | Choice | Rationale |
|----------|--------|-----------|
| License | CC BY-SA 4.0 | Compatible with ChEMBL CC BY-SA 3.0 (viral clause) |
| Storage | SQLite per domain | Portable, zero-infrastructure, reproducible |
| Export | Parquet with split columns | Standard ML format; lazy-loading friendly |
| ML metrics | LogAUC + 6 others | LogAUC[0.001,0.1] measures early enrichment, not just AUROC |
| LLM evaluation | 4 levels (L1–L4) | Progressive difficulty: MCQ → extraction → reasoning → discrimination |
---
## Domain Status Summary (as of 2026-03-30)
| Domain | DB Size | Negatives | ML Runs | LLM Runs | Status |
|--------|---------|-----------|---------|----------|--------|
| **DTI** | ~21 GB | 30,459,583 | 24/24 ✅ | 81/81 ✅ | Complete |
| **CT** | ~500 MB | 132,925 | 108/108 ✅ | 80/80 ✅ | Complete |
| **PPI** | 849 MB | 2,229,670 | 54/54 ✅ | 80/80 ✅ | Complete |
| **GE** | ~16 GB | 28,759,256 | Seed 42 ✅ | 4/5 models ✅ | Near complete |
---
## DTI Domain (Drug-Target Interaction)
Four sources: ChEMBL v36, PubChem BioAssay, BindingDB, DAVIS
### Database
- **30,459,583** negative results
- Source tiers: gold 818,611 / silver 198 / bronze 28,845,632
- 5 split strategies: random / cold_compound / cold_target / scaffold / temporal
### Key Results
- **ML:** Degree-matched negatives inflate LogAUC by +0.112 on average. Cold-target splits catastrophic (LogAUC 0.15–0.33) while AUROC stays deceptively high (0.76–0.89).
- **LLM L4:** All models near-random (MCC ≤ 0.18). DTI binding decisions are too nuanced for LLMs without domain context.
- **LLM L1:** Gemini achieves perfect accuracy (1.000) on 3-shot MCQ — artifact of format recognition.
---
## CT Domain (Clinical Trial Failure)
Four sources: AACT (ClinicalTrials.gov), CTO, Open Targets, Shi & Du 2024
### Database
- **132,925** failure results from 216,987 trials
- Tiers: gold 23,570 / silver 28,505 / bronze 60,223 / copper 20,627
- 8 failure categories: safety > efficacy > enrollment > strategic > regulatory > design > other
- Drug resolution: 4-step pipeline (ChEMBL exact → PubChem API → fuzzy JaroWinkler → manual CSV)
### Benchmark Design
- **ML:** CT-M1 binary failure prediction; CT-M2 7-way failure category (most challenging)
- **LLM:** L1 5-way MCQ (1,500 items), L2 failure report extraction (500), L3 reasoning (200), L4 discrimination (500)
### Key Results
- **CT-M1:** NegBioDB negatives trivially separable (AUROC=1.0). Control negatives reveal real difficulty (0.76–0.84).
- **CT-M2:** XGBoost best (macro-F1=0.51). Scaffold/temporal splits hardest (0.19).
- **LLM L4:** Gemini MCC=0.56 — highest across all domains. Meaningful discrimination possible for trial failure.
- **LLM L3:** Ceiling effect — GPT-4o-mini judge too lenient (4.4–5.0/5.0).
---
## PPI Domain (Protein-Protein Interaction)
Four sources: IntAct, HuRI, hu.MAP 3.0, STRING v12.0
### Database
- **2,229,670** negative results; 61,728 positive pairs (HuRI Y2H)
- 18,412 proteins; 4 split strategies: random / cold_protein / cold_both / degree_balanced
### Key Results
- **ML:** MLPFeatures (hand-crafted) dominates cold splits (AUROC 0.95 cold_both); PIPR collapses to 0.41 (below random).
- **LLM L1:** 3-shot near-perfect (0.997–1.000) is an artifact of example format leakage.
- **LLM L3:** zero-shot >> 3-shot (4.3–4.7 vs 3.1–3.7); gold reasoning examples degrade structural reasoning.
- **LLM L4:** MCC 0.33–0.44 with confirmed temporal contamination (pre-2015 acc ~0.6–0.8, post-2020 acc ~0.07–0.25).
---
## GE Domain (Gene Essentiality / DepMap)
Two sources: DepMap CRISPR (Chronos scores) and RNAi (DEMETER2)
### Database
- **28,759,256** negative results (genes with no essentiality signal)
- Final tiers: Gold 753,878 / Silver 18,608,686 / Bronze 9,396,692
- 19,554 genes × 2,132 cell lines; 22,549,910 aggregated pairs
- 5 split strategies: random / cold_gene / cold_cell_line / cold_both / degree_balanced
### Benchmark Design
- **ML:** XGBoost and MLPFeatures on gene expression + lineage features (gene-cell pair prediction)
- **LLM:** L1 4-way essentiality MCQ (1,200 items), L2 essentiality data extraction (500), L3 reasoning (200), L4 discrimination (475)
### Key Results (partial — Llama pending)
- **LLM L3:** zero-shot >> 3-shot (overall mean 4.5 vs 2.5) — same pattern as PPI.
- **LLM L4:** Expected intermediate MCC (DepMap is widely studied; likely contamination present).
- **ML:** Seed 42 complete; final aggregated results pending seeds 43/44.
---
## Dual Benchmark Framework
### LLM Benchmark Levels
| Level | Task | Difficulty | Automation |
|-------|------|-----------|------------|
| L1 | Multiple-choice classification | Easy | Fully automated (exact match) |
| L2 | Structured field extraction | Medium | Automated (JSON schema check + field F1) |
| L3 | Free-text reasoning quality | Hard | LLM-as-judge (Gemini 2.5-Flash, 4 rubric dimensions) |
| L4 | Real vs synthetic discrimination | Hard | Automated (MCC on binary decision) |
### LLM Models Evaluated
| Model | Provider | Type |
|-------|----------|------|
| Claude Haiku 4.5 | Anthropic API | Small API model |
| Gemini 2.5-Flash | Google API | Small API model |
| GPT-4o-mini | OpenAI API | Small API model |
| Qwen2.5-7B-Instruct | HuggingFace / vLLM | Open-weight local |
| Llama-3.1-8B-Instruct | HuggingFace / vLLM | Open-weight local |
### Cross-Domain LLM L4 Summary
```
DTI (≤0.18) < PPI (0.33–0.44) < CT (0.48–0.56)
↑
Increasing task complexity
and LLM accessible signal
```
---
## Timeline
| Milestone | Date |
|-----------|------|
| Project initiated | 2026-03-02 |
| DTI domain complete (ML + LLM) | 2026-03-13 |
| CT domain initiated | 2026-03-17 |
| CT domain complete (ML + LLM) | 2026-03-20 |
| PPI domain complete (ML + LLM) | 2026-03-23 |
| GE domain ETL + ML export | 2026-03-23 |
| GE LLM (4/5 models) | 2026-03-24 |
| Public release (GitHub + HuggingFace) | 2026-03-30 |
|