| | """Meta-cognitive QA loop: retrieve -> assess -> probe -> retrieve -> compose.""" |
| |
|
| | from __future__ import annotations |
| |
|
| | from dataclasses import dataclass, field |
| | import json |
| | from typing import Any, Dict, List, Optional, Sequence |
| |
|
| | from ..config import RuntimeConfig |
| | from ..retrieval.retriever import decompose_query, tri_retrieve |
| | from ..utils.retry import retry_call |
| | from .compose import compose_response |
| | from .token_usage import new_token_usage, record_token_usage |
| |
|
| |
|
| | @dataclass |
| | class LoopStep: |
| | iteration: int |
| | probe: str |
| | sufficient: bool |
| | confidence: float |
| | missing: List[str] = field(default_factory=list) |
| | next_probe: Optional[str] = None |
| | evidence_counts: Dict[str, int] = field(default_factory=dict) |
| |
|
| |
|
| | @dataclass |
| | class MetaLoopResult: |
| | answer: str |
| | evidence: Dict[str, object] |
| | trace: List[LoopStep] |
| | token_usage: Dict[str, Any] |
| |
|
| |
|
| | def _safe_text(v: object) -> str: |
| | return str(v or "").strip() |
| |
|
| |
|
| | def _dedup_merge_rows(existing: List[Dict], new_rows: List[Dict], lane: str) -> List[Dict]: |
| | merged = list(existing) |
| | seen = { |
| | ( |
| | lane, |
| | _safe_text(r.get("id")), |
| | _safe_text(r.get("text")), |
| | _safe_text(r.get("speaker")), |
| | _safe_text(r.get("entity")), |
| | ) |
| | for r in existing |
| | } |
| | for r in new_rows: |
| | key = ( |
| | lane, |
| | _safe_text(r.get("id")), |
| | _safe_text(r.get("text")), |
| | _safe_text(r.get("speaker")), |
| | _safe_text(r.get("entity")), |
| | ) |
| | if key in seen: |
| | continue |
| | seen.add(key) |
| | merged.append(r) |
| | merged.sort(key=lambda x: float(x.get("score", 0.0)), reverse=True) |
| | return merged |
| |
|
| |
|
| | def _merge_evidence_pool(pool: Dict[str, object], delta: Dict[str, object]) -> Dict[str, object]: |
| | out = dict(pool) |
| | for lane in ("facts", "persona", "worldview"): |
| | out[lane] = _dedup_merge_rows( |
| | existing=list(out.get(lane, [])), |
| | new_rows=list(delta.get(lane, [])), |
| | lane=lane, |
| | ) |
| | out["query_plan"] = delta.get("query_plan") or out.get("query_plan") or {} |
| | out["budgets"] = delta.get("budgets") or out.get("budgets") or {} |
| | return out |
| |
|
| |
|
| | def _sample_rows(rows: List[Dict], top_n: int) -> List[Dict]: |
| | sampled = [] |
| | for r in rows[:top_n]: |
| | sampled.append( |
| | { |
| | "rank": r.get("rank"), |
| | "score": r.get("score"), |
| | "text": _safe_text(r.get("text"))[:260], |
| | "speaker": r.get("speaker"), |
| | "type": r.get("type"), |
| | "entity": r.get("entity"), |
| | } |
| | ) |
| | return sampled |
| |
|
| |
|
| | def _parse_json_object(text: str) -> Optional[Dict]: |
| | if not text: |
| | return None |
| | try: |
| | obj = json.loads(text) |
| | if isinstance(obj, dict): |
| | return obj |
| | except json.JSONDecodeError: |
| | pass |
| | s = text.find("{") |
| | e = text.rfind("}") |
| | if s != -1 and e != -1 and e > s: |
| | try: |
| | obj = json.loads(text[s : e + 1]) |
| | if isinstance(obj, dict): |
| | return obj |
| | except json.JSONDecodeError: |
| | return None |
| | return None |
| |
|
| |
|
| | def _heuristic_assess(pool: Dict[str, object], cfg: RuntimeConfig) -> Dict: |
| | return _heuristic_assess_with_lanes(pool=pool, cfg=cfg, active_lanes=None) |
| |
|
| |
|
| | def _heuristic_assess_with_lanes( |
| | pool: Dict[str, object], |
| | cfg: RuntimeConfig, |
| | active_lanes: Optional[Sequence[str]] = None, |
| | ) -> Dict: |
| | facts = list(pool.get("facts", [])) |
| | persona = list(pool.get("persona", [])) |
| | worldview = list(pool.get("worldview", [])) |
| | enabled = {str(x).strip().lower() for x in (active_lanes or ["facts", "persona", "worldview"])} |
| | fact_max = max((float(r.get("score", 0.0)) for r in facts), default=0.0) |
| | checks: List[bool] = [] |
| | if "facts" in enabled: |
| | checks.append(len(facts) >= cfg.tri_k_min and fact_max >= cfg.tri_tau_low) |
| | if "worldview" in enabled: |
| | checks.append(len(worldview) >= 1) |
| | if "persona" in enabled: |
| | checks.append(len(persona) >= 1) |
| | sufficient = all(checks) if checks else True |
| | missing: List[str] = [] |
| | if "facts" in enabled and len(facts) < cfg.tri_k_min: |
| | missing.append("more veridical facts") |
| | if "worldview" in enabled and len(worldview) < 1: |
| | missing.append("world rules/setting evidence") |
| | if "persona" in enabled and len(persona) < 1: |
| | missing.append("character speaking evidence") |
| | next_probe = None |
| | if missing: |
| | next_probe = "Focus on: " + "; ".join(missing) |
| | return { |
| | "sufficient": sufficient, |
| | "confidence": 0.75 if sufficient else 0.35, |
| | "missing": missing, |
| | "next_probe": next_probe, |
| | "need_narrative": False, |
| | } |
| |
|
| |
|
| | def _llm_assess( |
| | query: str, |
| | character: Optional[str], |
| | pool: Dict[str, object], |
| | cfg: RuntimeConfig, |
| | token_usage: Optional[Dict[str, Any]] = None, |
| | active_lanes: Optional[Sequence[str]] = None, |
| | ) -> Optional[Dict]: |
| | if not cfg.llm_api_key: |
| | return None |
| | try: |
| | from openai import OpenAI |
| | except Exception: |
| | return None |
| |
|
| | client = OpenAI(base_url=cfg.llm_base_url, api_key=cfg.llm_api_key) |
| | model = cfg.meta_judge_model or cfg.llm_model |
| |
|
| | enabled = [str(x).strip().lower() for x in (active_lanes or ["facts", "persona", "worldview"]) if str(x).strip()] |
| | payload = { |
| | "query": query, |
| | "character": character, |
| | "active_lanes": enabled, |
| | "facts": _sample_rows(list(pool.get("facts", [])), cfg.meta_trace_top_n), |
| | "persona": _sample_rows(list(pool.get("persona", [])), cfg.meta_trace_top_n), |
| | "worldview": _sample_rows(list(pool.get("worldview", [])), cfg.meta_trace_top_n), |
| | } |
| | system_prompt = ( |
| | "You are an evidence sufficiency judge for a fiction QA agent.\n" |
| | "Return JSON only with fields:\n" |
| | "sufficient (bool), confidence (0-1), need_narrative (bool),\n" |
| | "missing (array of short strings), next_probe (string).\n" |
| | "Only judge sufficiency for active_lanes in payload; ignore disabled lanes.\n" |
| | "If evidence is weak, provide a concrete next_probe query for retrieval.\n" |
| | "Keep next_probe under 20 words." |
| | ) |
| | user_prompt = "Assess if evidence is sufficient.\n\n" + json.dumps(payload, ensure_ascii=False) |
| | try: |
| | def _call(): |
| | return client.chat.completions.create( |
| | model=model, |
| | messages=[ |
| | {"role": "system", "content": system_prompt}, |
| | {"role": "user", "content": user_prompt}, |
| | ], |
| | temperature=cfg.meta_judge_temperature, |
| | max_tokens=cfg.meta_new_probe_max_tokens, |
| | ) |
| |
|
| | def _on_retry(attempt: int, err: Exception, delay: float) -> None: |
| | print( |
| | f"[meta_assess][retry] attempt={attempt + 1}/{max(1, int(cfg.api_retry_attempts))} sleep={delay:.1f}s err={err}", |
| | flush=True, |
| | ) |
| |
|
| | resp = retry_call( |
| | _call, |
| | max_attempts=max(1, int(cfg.api_retry_attempts)), |
| | base_delay_sec=float(cfg.api_retry_base_delay_sec), |
| | max_delay_sec=float(cfg.api_retry_max_delay_sec), |
| | jitter_sec=float(cfg.api_retry_jitter_sec), |
| | on_retry=_on_retry, |
| | ) |
| | if token_usage is not None: |
| | record_token_usage( |
| | token_usage, |
| | response=resp, |
| | stage="meta_assess", |
| | model=model, |
| | ) |
| | content = _safe_text(resp.choices[0].message.content) |
| | obj = _parse_json_object(content) |
| | if not obj: |
| | return None |
| | return { |
| | "sufficient": bool(obj.get("sufficient", False)), |
| | "confidence": float(obj.get("confidence", 0.0)), |
| | "need_narrative": bool(obj.get("need_narrative", False)), |
| | "missing": [str(x) for x in obj.get("missing", []) if str(x).strip()], |
| | "next_probe": _safe_text(obj.get("next_probe")) or None, |
| | } |
| | except Exception: |
| | return None |
| |
|
| |
|
| | def _build_next_probe( |
| | query: str, |
| | assess: Dict, |
| | previous_probes: List[str], |
| | ) -> Optional[str]: |
| | probe = _safe_text(assess.get("next_probe")) |
| | if probe: |
| | if probe.lower() not in {p.lower() for p in previous_probes}: |
| | return probe |
| | missing = [str(x).strip() for x in assess.get("missing", []) if str(x).strip()] |
| | if missing: |
| | fallback = f"{query}. Focus on {', '.join(missing[:2])}." |
| | if fallback.lower() not in {p.lower() for p in previous_probes}: |
| | return fallback |
| | return None |
| |
|
| |
|
| | def run_meta_cognitive_qa( |
| | query: str, |
| | cfg: RuntimeConfig, |
| | character: Optional[str] = None, |
| | style_correct: bool = False, |
| | max_iterations: Optional[int] = None, |
| | active_lanes: Optional[Sequence[str]] = None, |
| | ) -> MetaLoopResult: |
| | max_iter = int(max_iterations or cfg.meta_max_iterations) |
| | pool: Dict[str, object] = {"facts": [], "persona": [], "worldview": [], "query_plan": {}} |
| | trace: List[LoopStep] = [] |
| | probes: List[str] = [query] |
| | token_usage: Dict[str, Any] = new_token_usage() |
| |
|
| | enabled_lanes = [str(x).strip().lower() for x in (active_lanes or ["facts", "persona", "worldview"]) if str(x).strip()] |
| | if not enabled_lanes: |
| | enabled_lanes = ["facts", "persona", "worldview"] |
| |
|
| | for i in range(1, max_iter + 1): |
| | probe = probes[-1] |
| | retrieval = tri_retrieve(query=probe, cfg=cfg, character=character, active_lanes=enabled_lanes) |
| | pool = _merge_evidence_pool(pool, retrieval) |
| |
|
| | assess = _llm_assess( |
| | query=query, |
| | character=character, |
| | pool=pool, |
| | cfg=cfg, |
| | token_usage=token_usage, |
| | active_lanes=enabled_lanes, |
| | ) |
| | if assess is None: |
| | assess = _heuristic_assess_with_lanes(pool=pool, cfg=cfg, active_lanes=enabled_lanes) |
| |
|
| | step = LoopStep( |
| | iteration=i, |
| | probe=probe, |
| | sufficient=bool(assess.get("sufficient", False)), |
| | confidence=float(assess.get("confidence", 0.0)), |
| | missing=[str(x) for x in assess.get("missing", [])], |
| | next_probe=_safe_text(assess.get("next_probe")) or None, |
| | evidence_counts={ |
| | "facts": len(list(pool.get("facts", []))), |
| | "persona": len(list(pool.get("persona", []))), |
| | "worldview": len(list(pool.get("worldview", []))), |
| | }, |
| | ) |
| | trace.append(step) |
| | if step.sufficient: |
| | break |
| |
|
| | next_probe = _build_next_probe(query=query, assess=assess, previous_probes=probes) |
| | if not next_probe: |
| | break |
| | probes.append(next_probe) |
| |
|
| | |
| | pool["query_plan"] = decompose_query(query=query, character=character) |
| | pool["active_lanes"] = enabled_lanes |
| | answer = compose_response( |
| | query=query, |
| | evidence=pool, |
| | cfg=cfg, |
| | character=character, |
| | style_correct=style_correct, |
| | token_usage=token_usage, |
| | ) |
| | return MetaLoopResult(answer=answer, evidence=pool, trace=trace, token_usage=token_usage) |
| |
|