"""Build vector indexes for facts, persona, and worldview.""" from __future__ import annotations import argparse import json from pathlib import Path from fic_agent.config import RuntimeConfig from fic_agent.retrieval.retriever import build_index_for_texts def _load_jsonl(path: str): rows = [] with open(path, "r", encoding="utf-8") as f: for line in f: line = line.strip() if not line: continue rows.append(json.loads(line)) return rows def main() -> None: parser = argparse.ArgumentParser(description="Build faiss indexes") parser.add_argument("--processed-dir", default=None, help="Processed data directory") args = parser.parse_args() cfg = RuntimeConfig() processed_dir = Path(args.processed_dir or cfg.data_processed_dir) # Facts chunks = _load_jsonl(str(processed_dir / "chunks.jsonl")) fact_texts = [c["text"] for c in chunks] fact_meta = [{"id": c["chunk_id"], "text": c["text"], "chapter_id": c["chapter_id"]} for c in chunks] build_index_for_texts(fact_texts, fact_meta, cfg, "facts") # Persona (all dialogues) dialogues = _load_jsonl(str(processed_dir / "dialogues.jsonl")) persona_texts = [d["utterance"] for d in dialogues] persona_meta = [ { "id": f"dlg-{i}", "text": d["utterance"], "speaker": d.get("speaker"), "chunk_id": d.get("chunk_id"), } for i, d in enumerate(dialogues) ] build_index_for_texts(persona_texts, persona_meta, cfg, "persona") # Worldview worldview = _load_jsonl(str(processed_dir / "worldview_notes.jsonl")) worldview_texts = [] worldview_meta = [] seen = set() for i, w in enumerate(worldview): text = str(w.get("text", "")).strip() if not text: continue norm = " ".join(text.split()).lower() if norm in seen: continue seen.add(norm) worldview_texts.append(text) worldview_meta.append( { "id": f"wv-{i}", "text": text, "type": w.get("type"), "entity": w.get("entity"), "source_chunk": w.get("source_chunk"), } ) build_index_for_texts(worldview_texts, worldview_meta, cfg, "worldview") print("Indexes built in", cfg.data_index_dir) if __name__ == "__main__": main()