""" Memory Builder Stage 1: Semantic Structured Compression (Section 3.1) & Stage 2: Online Semantic Synthesis (Section 3.2) Implements: - Implicit semantic density gating: Φ_gate(W) → {m_k} (filters low-density windows) - Sliding window processing for dialogue segmentation - Generates compact memory units with resolved coreferences and absolute timestamps """ from typing import List, Optional from models.memory_entry import MemoryEntry, Dialogue from utils.llm_client import LLMClient from database.vector_store import VectorStore import config import json import asyncio import concurrent.futures from functools import partial class MemoryBuilder: """ Memory Builder - Semantic Structured Compression (Section 3.1) Core Functions: 1. Sliding window segmentation 2. Implicit semantic density gating: Φ_gate(W) → {m_k} 3. Multi-view indexing: I(m_k) = {s_k, l_k, r_k} 4. Intra-session consolidation during write (Section 3.2): by generating enough memory entries to ensure ALL information is captured """ def __init__( self, llm_client: LLMClient, vector_store: VectorStore, window_size: int = None, overlap_size: Optional[int] = None, enable_parallel_processing: bool = True, max_parallel_workers: int = 3, fail_on_extraction_error: Optional[bool] = None, ): self.llm_client = llm_client self.vector_store = vector_store self.window_size = window_size or config.WINDOW_SIZE self.overlap_size = overlap_size if overlap_size is not None else getattr(config, 'OVERLAP_SIZE', 0) # step_size is how far the window advances each iteration; overlap retains # the last overlap_size dialogues so the next window has continuity context self.step_size = max(1, self.window_size - self.overlap_size) # Use config values as default if not explicitly provided self.enable_parallel_processing = enable_parallel_processing if enable_parallel_processing is not None else getattr(config, 'ENABLE_PARALLEL_PROCESSING', True) self.max_parallel_workers = max_parallel_workers if max_parallel_workers is not None else getattr(config, 'MAX_PARALLEL_WORKERS', 4) # Dialogue buffer self.dialogue_buffer: List[Dialogue] = [] self.processed_count = 0 self.fail_on_extraction_error = ( fail_on_extraction_error if fail_on_extraction_error is not None else getattr(config, "FAIL_ON_EXTRACTION_ERROR", False) ) # Previous window entries (for context) self.previous_entries: List[MemoryEntry] = [] def add_dialogue(self, dialogue: Dialogue, auto_process: bool = True): """ Add a dialogue to the buffer """ self.dialogue_buffer.append(dialogue) # Auto process if auto_process and len(self.dialogue_buffer) >= self.window_size: self.process_window() def add_dialogues(self, dialogues: List[Dialogue], auto_process: bool = True): """ Batch add dialogues with optional parallel processing """ if self.enable_parallel_processing and len(dialogues) > self.window_size * 2: # Use parallel processing for large batches self.add_dialogues_parallel(dialogues) else: # Use sequential processing for smaller batches for dialogue in dialogues: self.add_dialogue(dialogue, auto_process=False) # Process complete windows if auto_process: while len(self.dialogue_buffer) >= self.window_size: self.process_window() def add_dialogues_parallel(self, dialogues: List[Dialogue]): """ Add dialogues using parallel processing for better performance """ # Snapshot pre-existing buffer items so the fallback can restore them # if the buffer is cleared mid-way through parallel processing pre_existing = list(self.dialogue_buffer) windows_to_process = [] try: # Add all dialogues to buffer first self.dialogue_buffer.extend(dialogues) # Group into windows using step_size so that each window retains # overlap_size dialogues of context from the previous window pos = 0 while pos + self.window_size <= len(self.dialogue_buffer): window = self.dialogue_buffer[pos:pos + self.window_size] windows_to_process.append(window) pos += self.step_size # Add remaining dialogues as a smaller batch (no need to process separately) remaining = self.dialogue_buffer[pos:] if remaining: windows_to_process.append(remaining) self.dialogue_buffer = [] # Clear buffer since we're processing all if windows_to_process: print(f"\n[Parallel Processing] Processing {len(windows_to_process)} batches in parallel with {self.max_parallel_workers} workers") print(f"Batch sizes: {[len(w) for w in windows_to_process]}") # Process all windows/batches in parallel (including remaining dialogues) self._process_windows_parallel(windows_to_process) except Exception as e: print(f"[Parallel Processing] Failed: {e}. Falling back to sequential processing...") # Fallback: overlapping windows cannot be re-stacked naively. # If the buffer was cleared (exception after line 107), restore the full # original state: pre-existing items that were already in the buffer # PLUS the new dialogues we were asked to process. # If the buffer was NOT cleared (exception before line 107), it already # contains pre_existing + dialogues, so leave it as-is. if not self.dialogue_buffer: self.dialogue_buffer = pre_existing + list(dialogues) # process_window() uses step_size, so overlap is handled correctly here while len(self.dialogue_buffer) >= self.window_size: self.process_window() def process_window(self): """ Process current window dialogues - Core logic """ if not self.dialogue_buffer: return # Extract window; advance by step_size to retain overlap_size dialogues # at the tail so the next window has continuity context window = self.dialogue_buffer[:self.window_size] self.dialogue_buffer = self.dialogue_buffer[self.step_size:] print(f"\nProcessing window: {len(window)} dialogues (processed {self.processed_count} so far)") # Call LLM to generate memory entries entries = self._generate_memory_entries(window) self.processed_count += len(window) # Store to database if entries: self.vector_store.add_entries(entries) self.previous_entries = entries # Save as context print(f"Generated {len(entries)} memory entries") def process_remaining(self): """ Process remaining dialogues (fallback method, normally handled in parallel) """ if self.dialogue_buffer: print(f"\nProcessing remaining dialogues: {len(self.dialogue_buffer)} (fallback mode)") entries = self._generate_memory_entries(self.dialogue_buffer) self.processed_count += len(self.dialogue_buffer) if entries: self.vector_store.add_entries(entries) self.dialogue_buffer = [] print(f"Generated {len(entries)} memory entries") def _generate_memory_entries(self, dialogues: List[Dialogue]) -> List[MemoryEntry]: """ Implicit Semantic Density Gating (Section 3.1) Φ_gate(W) → {m_k}, generates compact memory units from dialogue window """ # Build dialogue text dialogue_text = "\n".join([str(d) for d in dialogues]) dialogue_ids = [d.dialogue_id for d in dialogues] # Build context context = "" if self.previous_entries: context = "\n[Previous Window Memory Entries (for reference to avoid duplication)]\n" for entry in self.previous_entries[:3]: # Only show first 3 context += f"- {entry.lossless_restatement}\n" # Build prompt prompt = self._build_extraction_prompt(dialogue_text, dialogue_ids, context) # Call LLM messages = [ { "role": "system", "content": "You are a professional information extraction assistant, skilled at extracting structured, unambiguous information from conversations. You must output valid JSON format." }, { "role": "user", "content": prompt } ] # Retry up to 3 times if parsing fails max_retries = 3 last_exception = None for attempt in range(max_retries): try: # Use JSON format if configured response_format = None if hasattr(config, 'USE_JSON_FORMAT') and config.USE_JSON_FORMAT: response_format = {"type": "json_object"} response = self.llm_client.chat_completion( messages, temperature=0.1, response_format=response_format, max_tokens=2600, ) # Parse response entries = self._parse_llm_response(response, dialogue_ids) return entries except Exception as e: last_exception = e if attempt < max_retries - 1: print(f"Attempt {attempt + 1}/{max_retries} failed to parse LLM response: {e}") print(f"Retrying...") else: print(f"All {max_retries} attempts failed to parse LLM response: {e}") print(f"Raw response: {response[:500] if 'response' in locals() else 'No response'}") if self.fail_on_extraction_error: raise RuntimeError( f"Failed to generate memory entries for dialogue ids {dialogue_ids}" ) from last_exception return [] def _build_extraction_prompt( self, dialogue_text: str, dialogue_ids: List[int], context: str ) -> str: """ Build LLM extraction prompt """ return f""" Your task is to extract all valuable information from the following dialogues and convert them into structured memory entries. {context} [Current Window Dialogues] {dialogue_text} [Requirements] 1. **Complete Coverage With Concise Synthesis**: Preserve important information from the dialogues while merging tightly related facts into semantically dense entries instead of many small entries 2. **Force Disambiguation**: Absolutely PROHIBIT using pronouns (he, she, it, they, this, that) and relative time (yesterday, today, last week, tomorrow) 3. **Lossless Information**: Each entry's lossless_restatement must be a complete, independent, understandable sentence 4. **Precise Extraction**: - keywords: Core keywords (names, places, entities, topic words) - timestamp: Absolute time in ISO 8601 format (if explicit time mentioned in dialogue) - location: Specific location name (if mentioned) - persons: All person names mentioned - entities: Companies, products, organizations, etc. - topic: The topic of this information [Output Format] Return a JSON object with an "entries" array, where each element is a memory entry: ```json {{ "entries": [ {{ "lossless_restatement": "Complete unambiguous restatement (must include all subjects, objects, time, location, etc.)", "keywords": ["keyword1", "keyword2", ...], "timestamp": "YYYY-MM-DDTHH:MM:SS or null", "location": "location name or null", "persons": ["name1", "name2", ...], "entities": ["entity1", "entity2", ...], "topic": "topic phrase" }}, ... ] }} ``` [Example] Dialogues: [2025-11-15T14:30:00] Alice: Bob, let's meet at Starbucks tomorrow at 2pm to discuss the new product [2025-11-15T14:31:00] Bob: Okay, I'll prepare the materials Output: ```json {{ "entries": [ {{ "lossless_restatement": "Alice suggested at 2025-11-15T14:30:00 to meet with Bob at Starbucks on 2025-11-16T14:00:00 to discuss the new product.", "keywords": ["Alice", "Bob", "Starbucks", "new product", "meeting"], "timestamp": "2025-11-16T14:00:00", "location": "Starbucks", "persons": ["Alice", "Bob"], "entities": ["new product"], "topic": "Product discussion meeting arrangement" }}, {{ "lossless_restatement": "Bob agreed to attend the meeting and committed to prepare relevant materials.", "keywords": ["Bob", "prepare materials", "agree"], "timestamp": null, "location": null, "persons": ["Bob"], "entities": [], "topic": "Meeting preparation confirmation" }} ] }} ``` Now process the above dialogues. Return ONLY the JSON object, no other explanations. """ def _parse_llm_response( self, response: str, dialogue_ids: List[int] ) -> List[MemoryEntry]: """ Parse LLM response to MemoryEntry list """ # Extract JSON data = self.llm_client.extract_json(response) if isinstance(data, dict): data = data.get("entries") if not isinstance(data, list): raise ValueError(f"Expected JSON array or object with 'entries' but got: {type(data)}") entries = [] for item in data: # Create MemoryEntry entry = MemoryEntry( lossless_restatement=item["lossless_restatement"], keywords=item.get("keywords", []), timestamp=item.get("timestamp"), location=item.get("location"), persons=item.get("persons", []), entities=item.get("entities", []), topic=item.get("topic") ) entries.append(entry) return entries def _process_windows_parallel(self, windows: List[List[Dialogue]]): """ Process multiple windows in parallel using ThreadPoolExecutor """ all_entries = [] # Use ThreadPoolExecutor for parallel processing with concurrent.futures.ThreadPoolExecutor(max_workers=self.max_parallel_workers) as executor: # Submit all window processing tasks future_to_window = {} for i, window in enumerate(windows): dialogue_ids = [d.dialogue_id for d in window] future = executor.submit(self._generate_memory_entries_worker, window, dialogue_ids, i+1) future_to_window[future] = (window, i+1) # Collect results as they complete for future in concurrent.futures.as_completed(future_to_window): window, window_num = future_to_window[future] try: entries = future.result() all_entries.extend(entries) print(f"[Parallel Processing] Window {window_num} completed: {len(entries)} entries") except Exception as e: print(f"[Parallel Processing] Window {window_num} failed: {e}") # Store all entries to database in batch if all_entries: print(f"\n[Parallel Processing] Storing {len(all_entries)} entries to database...") self.vector_store.add_entries(all_entries) self.processed_count += sum(len(window) for window in windows) # Update previous entries (use last window's entries for context) if all_entries: self.previous_entries = all_entries[-10:] # Keep last 10 entries for context print(f"[Parallel Processing] Completed processing {len(windows)} windows") def _generate_memory_entries_worker(self, window: List[Dialogue], dialogue_ids: List[int], window_num: int) -> List[MemoryEntry]: """ Worker function for parallel processing of a single batch (full window or remaining dialogues) """ batch_size = len(window) batch_type = "full window" if batch_size == self.window_size else f"remaining batch" print(f"[Worker {window_num}] Processing {batch_type} with {batch_size} dialogues") # Build dialogue text dialogue_text = "\n".join([str(d) for d in window]) # Build context (shared across all workers - this is fine for parallel processing) context = "" if self.previous_entries: context = "\n[Previous Window Memory Entries (for reference to avoid duplication)]\n" for entry in self.previous_entries[:3]: # Only show first 3 context += f"- {entry.lossless_restatement}\n" # Build prompt prompt = self._build_extraction_prompt(dialogue_text, dialogue_ids, context) # Call LLM messages = [ { "role": "system", "content": "You are a professional information extraction assistant, skilled at extracting structured, unambiguous information from conversations. You must output valid JSON format." }, { "role": "user", "content": prompt } ] # Retry up to 3 times if parsing fails max_retries = 3 for attempt in range(max_retries): try: # Use JSON format if configured response_format = None if hasattr(config, 'USE_JSON_FORMAT') and config.USE_JSON_FORMAT: response_format = {"type": "json_object"} response = self.llm_client.chat_completion( messages, temperature=0.1, response_format=response_format, max_tokens=2600, ) # Parse response entries = self._parse_llm_response(response, dialogue_ids) print(f"[Worker {window_num}] Generated {len(entries)} entries") return entries except Exception as e: if attempt < max_retries - 1: print(f"[Worker {window_num}] Attempt {attempt + 1}/{max_retries} failed: {e}. Retrying...") else: print(f"[Worker {window_num}] All {max_retries} attempts failed: {e}") return []