Datasets:
Formats:
parquet
Languages:
English
Size:
10M - 100M
Tags:
biology
chemistry
drug-discovery
clinical-trials
protein-protein-interaction
gene-essentiality
License:
File size: 5,153 Bytes
6d1bbc7 | 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 | #!/usr/bin/env python3
"""Re-evaluate existing PPI LLM predictions with updated metrics.
Reads predictions.jsonl from run directories, recomputes metrics using
the latest evaluation code, and writes updated results.json (backing up
the original).
Usage:
PYTHONPATH=src python scripts_ppi/reeval_ppi_llm.py --task ppi-l2
PYTHONPATH=src python scripts_ppi/reeval_ppi_llm.py --task ppi-l4
PYTHONPATH=src python scripts_ppi/reeval_ppi_llm.py --run-dir results/ppi_llm/ppi-l2_gpt-4o-mini_3-shot_fs0
"""
from __future__ import annotations
import argparse
import json
import re
import shutil
import sys
from pathlib import Path
PROJECT_ROOT = Path(__file__).resolve().parent.parent
RESULTS_DIR = PROJECT_ROOT / "results" / "ppi_llm"
EXPORTS_DIR = PROJECT_ROOT / "exports" / "ppi_llm"
TASK_DATASET = {
"ppi-l1": "ppi_l1_dataset.jsonl",
"ppi-l2": "ppi_l2_dataset.jsonl",
"ppi-l3": "ppi_l3_dataset.jsonl",
"ppi-l4": "ppi_l4_dataset.jsonl",
}
def load_jsonl(path: Path) -> list[dict]:
records = []
with open(path) as f:
for line in f:
records.append(json.loads(line))
return records
def reeval_run(run_dir: Path, gold_records: dict[str, dict], task: str) -> bool:
"""Re-evaluate a single run directory. Returns True if successful."""
pred_path = run_dir / "predictions.jsonl"
if not pred_path.exists():
print(f" SKIP {run_dir.name}: no predictions.jsonl")
return False
# Load predictions
preds = load_jsonl(pred_path)
if not preds:
print(f" SKIP {run_dir.name}: empty predictions")
return False
# Match predictions to gold records by question_id
pred_texts = []
gold_list = []
for p in preds:
qid = p.get("question_id", "")
if qid not in gold_records:
continue
pred_texts.append(str(p.get("prediction", "")))
gold_list.append(gold_records[qid])
if not pred_texts:
print(f" SKIP {run_dir.name}: no matching question_ids")
return False
# Import and compute metrics
from negbiodb_ppi.llm_eval import compute_all_ppi_llm_metrics
try:
metrics = compute_all_ppi_llm_metrics(task, pred_texts, gold_list)
except Exception as e:
print(f" ERROR {run_dir.name}: {e}")
return False
# Backup original results
results_path = run_dir / "results.json"
if results_path.exists():
backup_path = run_dir / "results_original.json"
if not backup_path.exists():
shutil.copy2(results_path, backup_path)
# Write updated results
with open(results_path, "w") as f:
json.dump(metrics, f, indent=2)
print(f" OK {run_dir.name}: {len(pred_texts)} predictions re-evaluated")
return True
def main(argv: list[str] | None = None) -> int:
parser = argparse.ArgumentParser(description="Re-evaluate PPI LLM predictions")
parser.add_argument("--task", type=str, help="Task filter (e.g. ppi-l2, ppi-l4)")
parser.add_argument("--run-dir", type=Path, help="Single run directory to re-evaluate")
parser.add_argument("--results-dir", type=Path, default=RESULTS_DIR)
parser.add_argument("--exports-dir", type=Path, default=EXPORTS_DIR)
args = parser.parse_args(argv)
# Determine which run directories to process
if args.run_dir:
run_dirs = [args.run_dir]
# Infer task from directory name
task_match = re.match(r"(ppi-l\d)", args.run_dir.name)
tasks = {task_match.group(1)} if task_match else set()
else:
run_dirs = sorted(d for d in args.results_dir.iterdir() if d.is_dir() and not d.name.endswith("_judged"))
if args.task:
run_dirs = [d for d in run_dirs if d.name.startswith(args.task + "_")]
tasks = {args.task}
else:
tasks = set()
for d in run_dirs:
m = re.match(r"(ppi-l\d)", d.name)
if m:
tasks.add(m.group(1))
# Load gold datasets for needed tasks
gold_by_task: dict[str, dict[str, dict]] = {}
for task in tasks:
dataset_file = args.exports_dir / TASK_DATASET.get(task, "")
if not dataset_file.exists():
print(f"WARNING: Gold dataset not found: {dataset_file}")
continue
records = load_jsonl(dataset_file)
# Filter to test + val splits (what the benchmark runs evaluate)
gold_by_task[task] = {
r["question_id"]: r for r in records if r.get("split") in ("test", "val")
}
print(f"Loaded {len(gold_by_task[task])} gold records for {task}")
# Process runs
n_ok = 0
n_skip = 0
for run_dir in run_dirs:
task_match = re.match(r"(ppi-l\d)", run_dir.name)
if not task_match:
continue
task = task_match.group(1)
if task not in gold_by_task:
n_skip += 1
continue
if reeval_run(run_dir, gold_by_task[task], task):
n_ok += 1
else:
n_skip += 1
print(f"\nDone: {n_ok} re-evaluated, {n_skip} skipped")
return 0
if __name__ == "__main__":
sys.exit(main())
|