Datasets:
Formats:
parquet
Languages:
English
Size:
10M - 100M
Tags:
biology
chemistry
drug-discovery
clinical-trials
protein-protein-interaction
gene-essentiality
License:
| #!/usr/bin/env python3 | |
| """Compute inter-rater agreement between primary and second judge for L3. | |
| Reads: | |
| - Primary judge scores: results/llm/l3_{run}_judged/judge_scores.jsonl | |
| - Second judge scores: results/llm/judge_validation/{judge}_scores.jsonl | |
| - (Optional) Human scores: results/llm/judge_validation/human_scores.jsonl | |
| Computes: | |
| - Cohen's quadratic weighted kappa per dimension | |
| - Pearson correlation per dimension | |
| - Overall agreement summary | |
| Usage: | |
| python scripts/compute_judge_agreement.py | |
| python scripts/compute_judge_agreement.py --include-human | |
| Output: | |
| results/llm/judge_validation/agreement_report.json | |
| """ | |
| import argparse | |
| import json | |
| from pathlib import Path | |
| import numpy as np | |
| from sklearn.metrics import cohen_kappa_score | |
| PROJECT_ROOT = Path(__file__).resolve().parent.parent | |
| RESULTS_DIR = PROJECT_ROOT / "results" / "llm" | |
| VALIDATION_DIR = RESULTS_DIR / "judge_validation" | |
| DIMS = ["accuracy", "reasoning", "completeness", "specificity"] | |
| def load_primary_scores(results_dir: Path) -> dict[tuple[str, str], dict]: | |
| """Load primary judge scores indexed by (run_name, question_id).""" | |
| scores = {} | |
| for judged_dir in sorted(results_dir.iterdir()): | |
| if not judged_dir.is_dir() or not judged_dir.name.endswith("_judged"): | |
| continue | |
| if not judged_dir.name.startswith("l3_"): | |
| continue | |
| run_name = judged_dir.name.replace("_judged", "") | |
| scores_path = judged_dir / "judge_scores.jsonl" | |
| if not scores_path.exists(): | |
| continue | |
| with open(scores_path) as f: | |
| for line in f: | |
| rec = json.loads(line) | |
| if rec.get("scores") is not None: | |
| scores[(run_name, rec["question_id"])] = rec["scores"] | |
| return scores | |
| def load_second_scores(scores_path: Path) -> dict[tuple[str, str], dict]: | |
| """Load second judge scores indexed by (source_run, question_id).""" | |
| scores = {} | |
| if not scores_path.exists(): | |
| return scores | |
| with open(scores_path) as f: | |
| for line in f: | |
| rec = json.loads(line) | |
| if rec.get("scores") is not None: | |
| key = (rec["source_run"], rec["question_id"]) | |
| scores[key] = rec["scores"] | |
| return scores | |
| def load_human_scores(scores_path: Path) -> dict[tuple[str, str], dict]: | |
| """Load human scores indexed by (source_run, question_id).""" | |
| return load_second_scores(scores_path) # Same format | |
| def compute_agreement( | |
| scores_a: dict[tuple[str, str], dict], | |
| scores_b: dict[tuple[str, str], dict], | |
| label_a: str, | |
| label_b: str, | |
| ) -> dict: | |
| """Compute agreement metrics between two sets of scores.""" | |
| # Find common items | |
| common_keys = set(scores_a.keys()) & set(scores_b.keys()) | |
| if not common_keys: | |
| return { | |
| "n_common": 0, | |
| "error": "No common items found", | |
| "label_a": label_a, | |
| "label_b": label_b, | |
| } | |
| result = { | |
| "n_common": len(common_keys), | |
| "label_a": label_a, | |
| "label_b": label_b, | |
| "dimensions": {}, | |
| } | |
| overall_a = [] | |
| overall_b = [] | |
| for dim in DIMS: | |
| vals_a = [] | |
| vals_b = [] | |
| for key in sorted(common_keys): | |
| a_val = scores_a[key].get(dim) | |
| b_val = scores_b[key].get(dim) | |
| if a_val is not None and b_val is not None: | |
| vals_a.append(a_val) | |
| vals_b.append(b_val) | |
| if len(vals_a) < 2: | |
| result["dimensions"][dim] = { | |
| "n": len(vals_a), | |
| "error": "Insufficient data", | |
| } | |
| continue | |
| arr_a = np.array(vals_a) | |
| arr_b = np.array(vals_b) | |
| # Cohen's quadratic weighted kappa | |
| # Round to integers for kappa (scores are 1-5) | |
| int_a = np.clip(np.round(arr_a).astype(int), 1, 5) | |
| int_b = np.clip(np.round(arr_b).astype(int), 1, 5) | |
| try: | |
| kappa = cohen_kappa_score(int_a, int_b, weights="quadratic") | |
| except ValueError: | |
| kappa = None | |
| # Pearson correlation | |
| if np.std(arr_a) > 0 and np.std(arr_b) > 0: | |
| pearson = float(np.corrcoef(arr_a, arr_b)[0, 1]) | |
| else: | |
| pearson = None | |
| # Mean absolute difference | |
| mad = float(np.mean(np.abs(arr_a - arr_b))) | |
| result["dimensions"][dim] = { | |
| "n": len(vals_a), | |
| "kappa": round(kappa, 4) if kappa is not None else None, | |
| "pearson": round(pearson, 4) if pearson is not None else None, | |
| "mean_abs_diff": round(mad, 4), | |
| "mean_a": round(float(np.mean(arr_a)), 3), | |
| "mean_b": round(float(np.mean(arr_b)), 3), | |
| } | |
| overall_a.extend(vals_a) | |
| overall_b.extend(vals_b) | |
| # Overall (all dimensions pooled) | |
| if len(overall_a) >= 2: | |
| arr_a = np.array(overall_a) | |
| arr_b = np.array(overall_b) | |
| int_a = np.clip(np.round(arr_a).astype(int), 1, 5) | |
| int_b = np.clip(np.round(arr_b).astype(int), 1, 5) | |
| try: | |
| kappa = cohen_kappa_score(int_a, int_b, weights="quadratic") | |
| except ValueError: | |
| kappa = None | |
| if np.std(arr_a) > 0 and np.std(arr_b) > 0: | |
| pearson = float(np.corrcoef(arr_a, arr_b)[0, 1]) | |
| else: | |
| pearson = None | |
| result["overall"] = { | |
| "n": len(overall_a), | |
| "kappa": round(kappa, 4) if kappa is not None else None, | |
| "pearson": round(pearson, 4) if pearson is not None else None, | |
| "mean_abs_diff": round(float(np.mean(np.abs(arr_a - arr_b))), 4), | |
| } | |
| return result | |
| def main(): | |
| parser = argparse.ArgumentParser(description="Compute L3 judge agreement") | |
| parser.add_argument( | |
| "--results-dir", type=Path, default=RESULTS_DIR, | |
| ) | |
| parser.add_argument( | |
| "--second-judge-file", type=str, default=None, | |
| help="Second judge scores file name (default: auto-detect)", | |
| ) | |
| parser.add_argument( | |
| "--include-human", action="store_true", | |
| help="Include human scores in agreement analysis", | |
| ) | |
| args = parser.parse_args() | |
| val_dir = args.results_dir / "judge_validation" | |
| if not val_dir.exists(): | |
| print("No judge_validation directory found.") | |
| return | |
| # Load primary (Gemini Flash-Lite) scores | |
| print("Loading primary judge scores...") | |
| primary = load_primary_scores(args.results_dir) | |
| print(f" {len(primary)} primary scores loaded") | |
| # Find second judge scores | |
| if args.second_judge_file: | |
| second_path = val_dir / args.second_judge_file | |
| else: | |
| # Auto-detect: find *_scores.jsonl (not human_scores.jsonl) | |
| candidates = [ | |
| p for p in val_dir.glob("*_scores.jsonl") | |
| if p.name != "human_scores.jsonl" | |
| ] | |
| if not candidates: | |
| print("No second judge scores found.") | |
| return | |
| second_path = candidates[0] | |
| print(f"Loading second judge scores from: {second_path.name}") | |
| second = load_second_scores(second_path) | |
| print(f" {len(second)} second judge scores loaded") | |
| # Compute primary vs second agreement | |
| report = {} | |
| judge_name = second_path.stem.replace("_scores", "") | |
| print(f"\n=== Primary (Flash-Lite) vs Second ({judge_name}) ===") | |
| agreement = compute_agreement(primary, second, "flash-lite", judge_name) | |
| report["primary_vs_second"] = agreement | |
| print(f" Common items: {agreement['n_common']}") | |
| if "dimensions" in agreement: | |
| for dim in DIMS: | |
| d = agreement["dimensions"].get(dim, {}) | |
| kappa = d.get("kappa") | |
| pearson = d.get("pearson") | |
| kappa_str = f"{kappa:.3f}" if kappa is not None else "N/A" | |
| pearson_str = f"{pearson:.3f}" if pearson is not None else "N/A" | |
| print(f" {dim:15s} kappa={kappa_str} r={pearson_str} " | |
| f"MAD={d.get('mean_abs_diff', 'N/A')}") | |
| if "overall" in agreement: | |
| o = agreement["overall"] | |
| print(f" {'OVERALL':15s} kappa={o['kappa']:.3f} r={o['pearson']:.3f} " | |
| f"MAD={o['mean_abs_diff']:.3f}") | |
| # Human scores (optional) | |
| if args.include_human: | |
| human_path = val_dir / "human_scores.jsonl" | |
| if human_path.exists(): | |
| print(f"\nLoading human scores...") | |
| human = load_human_scores(human_path) | |
| print(f" {len(human)} human scores loaded") | |
| # Primary vs human | |
| print("\n=== Primary (Flash-Lite) vs Human ===") | |
| ph_agreement = compute_agreement(primary, human, "flash-lite", "human") | |
| report["primary_vs_human"] = ph_agreement | |
| if "dimensions" in ph_agreement: | |
| for dim in DIMS: | |
| d = ph_agreement["dimensions"].get(dim, {}) | |
| kappa = d.get("kappa") | |
| kappa_str = f"{kappa:.3f}" if kappa is not None else "N/A" | |
| print(f" {dim:15s} kappa={kappa_str}") | |
| # Second vs human | |
| print(f"\n=== Second ({judge_name}) vs Human ===") | |
| sh_agreement = compute_agreement(second, human, judge_name, "human") | |
| report["second_vs_human"] = sh_agreement | |
| if "dimensions" in sh_agreement: | |
| for dim in DIMS: | |
| d = sh_agreement["dimensions"].get(dim, {}) | |
| kappa = d.get("kappa") | |
| kappa_str = f"{kappa:.3f}" if kappa is not None else "N/A" | |
| print(f" {dim:15s} kappa={kappa_str}") | |
| else: | |
| print(f"\n Human scores not found at {human_path}") | |
| # Save report | |
| report_path = val_dir / "agreement_report.json" | |
| with open(report_path, "w") as f: | |
| json.dump(report, f, indent=2) | |
| print(f"\nReport saved: {report_path}") | |
| # Interpretation | |
| if "overall" in agreement and agreement["overall"].get("kappa") is not None: | |
| kappa = agreement["overall"]["kappa"] | |
| if kappa >= 0.8: | |
| interp = "almost perfect" | |
| elif kappa >= 0.6: | |
| interp = "substantial" | |
| elif kappa >= 0.4: | |
| interp = "moderate" | |
| elif kappa >= 0.2: | |
| interp = "fair" | |
| else: | |
| interp = "slight/poor" | |
| print(f"\nOverall kappa = {kappa:.3f} ({interp} agreement)") | |
| if kappa < 0.6: | |
| print(" Note: kappa < 0.6 — consider adding caveat in paper") | |
| if __name__ == "__main__": | |
| main() | |