trentdoney's picture
Upload src/evaluator.py
d9003b7 verified
"""BrainCore Memory Benchmark — main evaluation harness."""
import argparse
import importlib
import json
import time
from pathlib import Path
from typing import Any
from metrics import (
aggregate,
exact_match,
semantic_placeholder_score,
temporal_order_score,
contradiction_resolution_score,
latency_ms,
)
def load_dataset(path: str) -> list[dict]:
sessions = []
with open(path, "r", encoding="utf-8") as f:
for line in f:
line = line.strip()
if line:
sessions.append(json.loads(line))
return sessions
def evaluate_session(session: dict, adapter: Any, top_k: int = 1) -> list[dict]:
"""Ingest memories then run every query. Returns per-query result dicts."""
adapter.ingest(session.get("memories", []))
storage = adapter.storage_bytes()
results = []
for q in session.get("queries", []):
t0 = time.perf_counter()
retrieved = adapter.retrieve(q["query_text"], top_k=top_k)
t1 = time.perf_counter()
pred_text = " ".join(m.get("text", "") for m in retrieved)
ref_text = q.get("expected_answer", "")
em = exact_match(pred_text, ref_text)
sem = semantic_placeholder_score(pred_text, ref_text)
qtype = q.get("query_type", "retrieval")
to_score = temporal_order_score(retrieved, q.get("required_memory_ids", []))
# Contradiction: if the dataset marks the latest revised memory, use it.
latest_id = q.get("latest_memory_id") if qtype == "contradiction" else None
cr_score = contradiction_resolution_score(retrieved, latest_id)
# For pure retrieval/temporal without explicit contradiction, cr_score is neutral.
if qtype != "contradiction":
cr_score = 1.0
results.append({
"session_id": session["session_id"],
"query_id": q["query_id"],
"query_type": qtype,
"exact_match": em,
"semantic_placeholder_score": sem,
"temporal_order_score": to_score,
"contradiction_resolution_score": cr_score,
"latency_ms": latency_ms(t0, t1),
"storage_bytes": storage,
"pred": pred_text,
"ref": ref_text,
})
return results
def main():
parser = argparse.ArgumentParser(description="BrainCore Memory Benchmark Harness")
parser.add_argument("--adapter", required=True, help="Dotted path to adapter module (must expose `build() -> MemoryAdapter`)")
parser.add_argument("--dataset", required=True, help="Path to JSONL dataset")
parser.add_argument("--output", required=True, help="Path to write JSON results")
parser.add_argument("--top_k", type=int, default=1, help="Number of memories to retrieve")
args = parser.parse_args()
# Dynamically import adapter module.
module = importlib.import_module(args.adapter)
adapter = module.build()
sessions = load_dataset(args.dataset)
all_results = []
for session in sessions:
all_results.extend(evaluate_session(session, adapter, top_k=args.top_k))
summary = aggregate(all_results)
payload = {
"config": {
"adapter": args.adapter,
"dataset": args.dataset,
"top_k": args.top_k,
},
"summary": summary,
"per_query": all_results,
}
Path(args.output).parent.mkdir(parents=True, exist_ok=True)
with open(args.output, "w", encoding="utf-8") as f:
json.dump(payload, f, indent=2, ensure_ascii=False)
print("BrainCore Memory Benchmark — Summary")
print("=" * 40)
for k, v in summary.items():
print(f" {k:40s}: {v}")
print(f"\nFull results written to {args.output}")
if __name__ == "__main__":
main()