"""Evaluate one generated answer with proxy factual/style/worldview scores.""" from __future__ import annotations import argparse import json from fic_agent.config import RuntimeConfig from fic_agent.eval.judge import score_response_proxy, score_response_llm def _load_json(path: str): with open(path, "r", encoding="utf-8") as f: return json.load(f) def _build_compact_report(result: dict) -> dict: mode = str(result.get("mode", "")).strip() scores = result.get("scores") if isinstance(result.get("scores"), dict) else {} issues_obj = result.get("issues") if isinstance(result.get("issues"), dict) else {} critical = [str(x).strip() for x in issues_obj.get("critical", []) if str(x).strip()] major = [str(x).strip() for x in issues_obj.get("major", []) if str(x).strip()] minor = [str(x).strip() for x in issues_obj.get("minor", []) if str(x).strip()] if mode == "proxy": return { "mode": mode, "scores": scores, "key_conclusion": "Proxy-only heuristic scores (fast check, not final LLM judgment).", } same_character = result.get("same_character") confidence_100 = result.get("confidence_100") scorecard = result.get("scorecard") if isinstance(result.get("scorecard"), dict) else {} penalties = result.get("penalties") if isinstance(result.get("penalties"), dict) else {} overall_100 = scorecard.get("overall_100") if overall_100 is None: overall_100 = scores.get("overall") usefulness_100 = scores.get("usefulness") if usefulness_100 is None and isinstance(scorecard.get("response_usefulness"), dict): usefulness_module = scorecard.get("response_usefulness", {}).get("module_score") if usefulness_module is not None: try: usefulness_100 = round((float(usefulness_module) / 5.0) * 100.0, 2) except Exception: usefulness_100 = None if critical: verdict = "High-risk answer: critical consistency issues detected." elif major: verdict = "Usable with caution: major issues remain." elif same_character == "Yes": verdict = "Good result: role consistency and overall quality are acceptable." else: verdict = "Role consistency is insufficient." return { "mode": mode or "llm", "scores": scores, "overall_100": overall_100, "usefulness_100": usefulness_100, "same_character": same_character, "confidence_100": confidence_100, "issues": { "critical": critical, "major": major, "minor": minor[:3], }, "penalty": { "formula": penalties.get("formula"), "additive_deduction": penalties.get("additive_deduction"), "multiplier": penalties.get("multiplier"), "overall_deduction": penalties.get("overall_deduction"), }, "key_conclusion": verdict, } def main() -> None: parser = argparse.ArgumentParser(description="Evaluate generated answer") parser.add_argument("--result-json", required=True, help="Path produced by run_meta_qa --save-json") parser.add_argument("--character", default=None, help="Character override") parser.add_argument("--processed-dir", default="data/processed", help="Processed directory") parser.add_argument("--mode", choices=["proxy", "llm"], default="llm", help="Scoring mode") parser.add_argument("--model", default=None, help="Judge model override for LLM mode") parser.add_argument("--rounds", type=int, default=3, help="Judge rounds for LLM mode") parser.add_argument("--temperature", type=float, default=0.2, help="Judge temperature for LLM mode") parser.add_argument("--top-n", type=int, default=6, help="Evidence items per lane shown to LLM judge") parser.add_argument("--full-report", action="store_true", help="Keep full detailed report instead of compact summary") parser.add_argument("--save-json", default=None, help="Optional path to save full score report") args = parser.parse_args() obj = _load_json(args.result_json) query = obj.get("query", "") response = obj.get("answer", "") evidence = obj.get("evidence", {}) character = args.character or obj.get("character") if args.mode == "proxy": scores = score_response_proxy( response=response, evidence=evidence, character=character, processed_dir=args.processed_dir, ) result = {"mode": "proxy", "scores": scores} else: cfg = RuntimeConfig() result = score_response_llm( query=query, response=response, evidence=evidence, cfg=cfg, character=character, model=args.model, rounds=args.rounds, temperature=args.temperature, top_n=args.top_n, ) output = result if args.full_report else _build_compact_report(result) print(json.dumps(output, ensure_ascii=False, indent=2)) if args.save_json: with open(args.save_json, "w", encoding="utf-8") as f: json.dump(output, f, ensure_ascii=False, indent=2) if __name__ == "__main__": main()