SPARKNET / .streamlit /secrets.toml.example
MHamdan's picture
the update
4718630
# ============================================================================
# SPARKNET - Streamlit Secrets Configuration
# ============================================================================
# Copy this file to secrets.toml (DO NOT commit secrets.toml!)
# For Streamlit Cloud: Add these via the Streamlit Cloud dashboard
#
# VISTA/Horizon EU Project - Technology Transfer Office Automation
# ============================================================================
# ============================================================================
# Authentication (Required)
# ============================================================================
[auth]
# Single user mode
password = "your-secure-password"
# ============================================================================
# Backend Server (Optional - for GPU processing)
# ============================================================================
# If you have a GPU server (e.g., Lytos), configure the backend URL here.
# The backend provides GPU-accelerated OCR, embeddings, and RAG processing.
# See DEPLOYMENT.md for setup instructions.
# BACKEND_URL = "https://your-gpu-server.com:8000"
# Or for local testing:
# BACKEND_URL = "http://localhost:8000"
# Multi-user mode (uncomment to use):
# [auth.users]
# admin = "admin-password-here"
# viewer = "viewer-password-here"
# analyst = "analyst-password-here"
# ============================================================================
# LLM Provider API Keys
# ============================================================================
# Add only the providers you want to use - system auto-selects best available
# Priority: Groq > Gemini > OpenRouter > GitHub > Together > Mistral > HuggingFace > Offline
# Groq - Fastest inference, 14,400 requests/day free
# Get key: https://console.groq.com/keys
GROQ_API_KEY = ""
# Google Gemini/AI Studio - 15 requests/min free
# Get key: https://aistudio.google.com/apikey
GOOGLE_API_KEY = ""
# OpenRouter - Access to many free models with single API
# Get key: https://openrouter.ai/keys
OPENROUTER_API_KEY = ""
# GitHub Models - Free GPT-4o, Llama 3.1 access
# Get token: https://github.com/settings/tokens (enable 'models' scope)
GITHUB_TOKEN = ""
# HuggingFace - Thousands of free models, embeddings
# Get token: https://huggingface.co/settings/tokens
HF_TOKEN = ""
# Together AI - $25 free credits
# Get key: https://www.together.ai/
TOGETHER_API_KEY = ""
# Mistral AI - Free experiment plan
# Get key: https://console.mistral.ai/
MISTRAL_API_KEY = ""
# ============================================================================
# Premium/Paid Providers (Optional)
# ============================================================================
# OpenAI - GPT-4, embeddings (paid)
# Get key: https://platform.openai.com/api-keys
OPENAI_API_KEY = ""
# Anthropic Claude (paid)
# Get key: https://console.anthropic.com/
ANTHROPIC_API_KEY = ""
# ============================================================================
# Database Configuration (Optional - for production)
# ============================================================================
[database]
# PostgreSQL connection (uncomment for production)
# url = "postgresql://user:password@host:5432/sparknet"
# ChromaDB persistence directory
chroma_persist_dir = "./data/chroma"
# ============================================================================
# Security Configuration
# ============================================================================
[security]
# Secret key for session management (generate with: python -c "import secrets; print(secrets.token_hex(32))")
secret_key = ""
# Enable audit logging
audit_logging = false
# ============================================================================
# GDPR & Data Privacy
# ============================================================================
# IMPORTANT: For EU/VISTA deployments, configure these settings
[privacy]
# Data retention in days (0 = indefinite)
data_retention_days = 0
# Enable PII detection and masking
pii_detection = false
# Enable data anonymization for exports
anonymize_exports = false
# ============================================================================
# Feature Flags
# ============================================================================
[features]
# Enable experimental scenarios
experimental_scenarios = false
# Enable GPU acceleration (requires CUDA)
gpu_enabled = true
# Enable response caching
caching_enabled = true
cache_ttl_seconds = 3600
# ============================================================================
# Private Deployment Notes
# ============================================================================
# For enterprise/private deployments:
#
# 1. LOCAL INFERENCE (Maximum Privacy):
# - Use Ollama for 100% on-premise inference
# - No data leaves your network
# - Set OLLAMA_HOST = "http://localhost:11434"
#
# 2. HYBRID DEPLOYMENT:
# - Use local Ollama for sensitive documents
# - Use cloud LLMs for non-sensitive queries
# - Configure document classification rules
#
# 3. CLOUD DEPLOYMENT (Streamlit Cloud):
# - Use secrets management (this file)
# - Enable audit logging
# - Configure data retention policies
# - Review GDPR compliance checklist
#
# See DEPLOYMENT.md for detailed instructions
# ============================================================================