| """ | |
| Configuration file - System parameters and LLM settings | |
| IMPORTANT: | |
| 1. Copy this file to config.py | |
| 2. Configure your settings below | |
| 3. Never commit config.py to version control (it contains your API key) | |
| """ | |
| # ============================================================================ | |
| # LLM Configuration | |
| # ============================================================================ | |
| # OpenAI API Key (required) | |
| # Get your key from: https://platform.openai.com/api-keys | |
| OPENAI_API_KEY = "your-api-key-here" | |
| # Custom OpenAI Base URL (optional) | |
| # Set to None to use default OpenAI endpoint | |
| # Examples: | |
| # - Qwen/Alibaba: "https://dashscope.aliyuncs.com/compatible-mode/v1" | |
| # - Azure OpenAI: "https://YOUR-RESOURCE.openai.azure.com/openai/deployments/YOUR-DEPLOYMENT" | |
| # - Local server: "http://localhost:8000/v1" | |
| # - OpenAI (default): None | |
| OPENAI_BASE_URL = None | |
| # LLM Model name | |
| # Examples: "gpt-4.1-mini", "gpt-4.1", "qwen3-max", "qwen-plus-2025-07-28" | |
| LLM_MODEL = "gpt-4.1-mini" | |
| # Embedding model (local, no API needed) | |
| EMBEDDING_MODEL = "Qwen/Qwen3-Embedding-0.6B" | |
| EMBEDDING_DIMENSION = 1024 # For Qwen3: up to 1024, supports 32-1024 | |
| EMBEDDING_CONTEXT_LENGTH = 32768 # Qwen3 supports 32k context | |
| # ============================================================================ | |
| # Advanced LLM Features | |
| # ============================================================================ | |
| # Enable deep thinking mode (for Qwen and compatible models) | |
| # Adds extra_body={"enable_thinking": True} to API calls | |
| # Set to False for OpenAI models (they don't support this) | |
| ENABLE_THINKING = False | |
| # Enable streaming responses (outputs content as it's generated) | |
| USE_STREAMING = True | |
| # Enable JSON format mode (ensures LLM outputs valid JSON) | |
| # Adds response_format={"type": "json_object"} to API calls | |
| # Helps prevent parsing failures from extra text like ```json | |
| USE_JSON_FORMAT = False | |
| # ============================================================================ | |
| # Memory Building Parameters | |
| # ============================================================================ | |
| # Number of dialogues per window (for locomo; for other dataset, please finetune it) | |
| WINDOW_SIZE = 40 | |
| # Window overlap size (for context continuity) | |
| OVERLAP_SIZE = 2 | |
| # ============================================================================ | |
| # Retrieval Parameters (can be adjusted to balance between token usage and performance) | |
| # ============================================================================ | |
| # Max entries returned by semantic search (vector similarity) | |
| SEMANTIC_TOP_K = 25 | |
| # Max entries returned by keyword search (BM25 matching) | |
| KEYWORD_TOP_K = 5 | |
| # Max entries returned by structured search (metadata filtering) | |
| STRUCTURED_TOP_K = 5 | |
| # ============================================================================ | |
| # Database Configuration | |
| # ============================================================================ | |
| # Path to LanceDB storage | |
| LANCEDB_PATH = "./lancedb_data" | |
| # Memory table name | |
| MEMORY_TABLE_NAME = "memory_entries" | |
| # ============================================================================ | |
| # Parallel Processing Configuration | |
| # ============================================================================ | |
| # Memory Building Parallel Processing | |
| ENABLE_PARALLEL_PROCESSING = True | |
| MAX_PARALLEL_WORKERS = 16 # Number of parallel workers for memory building | |
| # Retrieval Parallel Processing | |
| ENABLE_PARALLEL_RETRIEVAL = True | |
| MAX_RETRIEVAL_WORKERS = 8 # Number of parallel workers for retrieval queries | |
| # Planning and Reflection Configuration | |
| ENABLE_PLANNING = True | |
| ENABLE_REFLECTION = True | |
| MAX_REFLECTION_ROUNDS = 2 | |
| # ============================================================================ | |
| # LLM-as-Judge Configuration (not used yet) | |
| # ============================================================================ | |
| # Judge LLM API Key (optional - if None, uses OPENAI_API_KEY) | |
| JUDGE_API_KEY = "your api-key here" | |
| # Judge LLM Base URL (optional - if None, uses OPENAI_BASE_URL) | |
| # Example: Use cheaper endpoint for evaluation | |
| JUDGE_BASE_URL = "https://api.openai.com/v1/" | |
| # Judge LLM Model (optional - if None, uses LLM_MODEL) | |
| JUDGE_MODEL = "gpt-4.1-mini" | |
| # Judge specific settings | |
| JUDGE_ENABLE_THINKING = False # Usually false for evaluation tasks | |
| JUDGE_USE_STREAMING = False # Usually false for evaluation | |
| JUDGE_TEMPERATURE = 0.3 | |
| # Example configurations: | |
| # 1. Use cheaper model for judge evaluation: | |
| # JUDGE_MODEL = "gpt-4.1-mini" | |
| # | |
| # 2. Use different API provider for judge: | |
| # JUDGE_API_KEY = "your-judge-api-key" | |
| # JUDGE_BASE_URL = "https://api.different-provider.com/v1" | |
| # JUDGE_MODEL = "different-provider-model" | |
| # | |
| # 3. Use Qwen for judge (if available): | |
| # JUDGE_BASE_URL = "https://dashscope.aliyuncs.com/compatible-mode/v1" | |
| # JUDGE_MODEL = "qwen-plus-2025-09-11" | |