| # Azure OpenAI Configuration | |
| # Get these from https://portal.azure.com → Your Azure OpenAI Resource | |
| AZURE_OPENAI_ENDPOINT=https://your-resource.openai.azure.com/ | |
| AZURE_OPENAI_API_KEY=your-api-key-here | |
| AZURE_OPENAI_DEPLOYMENT_NAME=gpt-4o-mini | |
| AZURE_OPENAI_API_VERSION=2024-05-01-preview | |
| # ⚠️ CRITICAL: Embedding model deployment name | |
| # This MUST match an existing deployment in your Azure OpenAI resource | |
| # Common deployment names (check Azure Portal → Model deployments): | |
| # - text-embedding-3-small (recommended, most cost-effective) | |
| # - text-embedding-3-large (higher quality, more expensive) | |
| # - text-embedding-ada-002 (legacy, widely compatible) | |
| # | |
| # HOW TO VERIFY: | |
| # 1. Run: python scripts/validate_azure_embeddings.py | |
| # 2. Or check Azure Portal → Your Resource → Model deployments | |
| # | |
| # ⚠️ If this deployment doesn't exist, you'll get a 404 error! | |
| AZURE_OPENAI_EMBEDDING_DEPLOYMENT_NAME=text-embedding-3-small | |
| # Optional: Cost Pricing Overrides (per 1M tokens) | |
| # These override the prices in config/pricing.json for all models | |
| # Useful for testing or when using custom pricing | |
| # PRICING_INPUT_PER_1M=0.08 | |
| # PRICING_OUTPUT_PER_1M=0.32 | |
| # PRICING_EMBEDDING_PER_1M=0.02 | |
| # MCP (Model Context Protocol) Configuration | |
| # Set to 'true' to use MCP for arXiv access (default: FastMCP) | |
| # Set to 'false' to use direct arXiv API | |
| USE_MCP_ARXIV=false | |
| # Set to 'true' to force legacy MCP instead of FastMCP | |
| # Only applies when USE_MCP_ARXIV=true | |
| USE_LEGACY_MCP=false | |
| # Path where MCP server stores downloaded papers | |
| # Should match the storage path configured in your MCP server | |
| MCP_ARXIV_STORAGE_PATH=./data/mcp_papers/ | |
| # FastMCP Configuration | |
| # Port for FastMCP server (auto-started when USE_MCP_ARXIV=true) | |
| FASTMCP_SERVER_PORT=5555 | |
| # LangFuse Observability Configuration | |
| # Enable/disable LangFuse tracing (default: true) | |
| LANGFUSE_ENABLED=true | |
| # LangFuse Cloud API Keys (get from https://cloud.langfuse.com) | |
| LANGFUSE_PUBLIC_KEY=pk-lf | |
| LANGFUSE_SECRET_KEY=sk-lf- | |
| # LangFuse Host URL (default: https://cloud.langfuse.com) | |
| # For self-hosted: LANGFUSE_HOST=http://localhost:3000 | |
| LANGFUSE_HOST=https://cloud.langfuse.com | |
| # Optional: LangFuse Tracing Settings | |
| # Trace all LLM calls automatically (default: true) | |
| LANGFUSE_TRACE_ALL_LLM=true | |
| # Trace RAG retrieval operations (default: true) | |
| LANGFUSE_TRACE_RAG=true | |
| # Flush observations after N items (default: 15) | |
| LANGFUSE_FLUSH_AT=15 | |
| # Flush interval in seconds (default: 10) | |
| LANGFUSE_FLUSH_INTERVAL=10 | |