File size: 2,470 Bytes
aca8ab4 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 |
# Azure OpenAI Configuration
# Get these from https://portal.azure.com → Your Azure OpenAI Resource
AZURE_OPENAI_ENDPOINT=https://your-resource.openai.azure.com/
AZURE_OPENAI_API_KEY=your-api-key-here
AZURE_OPENAI_DEPLOYMENT_NAME=gpt-4o-mini
AZURE_OPENAI_API_VERSION=2024-05-01-preview
# ⚠️ CRITICAL: Embedding model deployment name
# This MUST match an existing deployment in your Azure OpenAI resource
# Common deployment names (check Azure Portal → Model deployments):
# - text-embedding-3-small (recommended, most cost-effective)
# - text-embedding-3-large (higher quality, more expensive)
# - text-embedding-ada-002 (legacy, widely compatible)
#
# HOW TO VERIFY:
# 1. Run: python scripts/validate_azure_embeddings.py
# 2. Or check Azure Portal → Your Resource → Model deployments
#
# ⚠️ If this deployment doesn't exist, you'll get a 404 error!
AZURE_OPENAI_EMBEDDING_DEPLOYMENT_NAME=text-embedding-3-small
# Optional: Cost Pricing Overrides (per 1M tokens)
# These override the prices in config/pricing.json for all models
# Useful for testing or when using custom pricing
# PRICING_INPUT_PER_1M=0.08
# PRICING_OUTPUT_PER_1M=0.32
# PRICING_EMBEDDING_PER_1M=0.02
# MCP (Model Context Protocol) Configuration
# Set to 'true' to use MCP for arXiv access (default: FastMCP)
# Set to 'false' to use direct arXiv API
USE_MCP_ARXIV=false
# Set to 'true' to force legacy MCP instead of FastMCP
# Only applies when USE_MCP_ARXIV=true
USE_LEGACY_MCP=false
# Path where MCP server stores downloaded papers
# Should match the storage path configured in your MCP server
MCP_ARXIV_STORAGE_PATH=./data/mcp_papers/
# FastMCP Configuration
# Port for FastMCP server (auto-started when USE_MCP_ARXIV=true)
FASTMCP_SERVER_PORT=5555
# LangFuse Observability Configuration
# Enable/disable LangFuse tracing (default: true)
LANGFUSE_ENABLED=true
# LangFuse Cloud API Keys (get from https://cloud.langfuse.com)
LANGFUSE_PUBLIC_KEY=pk-lf
LANGFUSE_SECRET_KEY=sk-lf-
# LangFuse Host URL (default: https://cloud.langfuse.com)
# For self-hosted: LANGFUSE_HOST=http://localhost:3000
LANGFUSE_HOST=https://cloud.langfuse.com
# Optional: LangFuse Tracing Settings
# Trace all LLM calls automatically (default: true)
LANGFUSE_TRACE_ALL_LLM=true
# Trace RAG retrieval operations (default: true)
LANGFUSE_TRACE_RAG=true
# Flush observations after N items (default: 15)
LANGFUSE_FLUSH_AT=15
# Flush interval in seconds (default: 10)
LANGFUSE_FLUSH_INTERVAL=10
|