# ── HuggingFace Spaces compatible Dockerfile ────────────────────────────── # Port MUST be 7860 for HF Spaces. # Runs as user 1000 (HF requirement). FROM python:3.10-slim # System deps RUN apt-get update && apt-get install -y --no-install-recommends \ build-essential curl git \ && rm -rf /var/lib/apt/lists/* # Create non-root user (required by HF Spaces) RUN useradd -m -u 1000 appuser WORKDIR /app # Install Python deps first (better layer caching) COPY requirements_api.txt . RUN pip install --no-cache-dir -r requirements_api.txt # Copy application files COPY . . # Pin HuggingFace cache inside /app so the sentence-transformer model is # downloaded once during `docker build` and baked into the image layer. ENV HF_HOME=/app/.hf_cache ENV TRANSFORMERS_CACHE=/app/.hf_cache/transformers ENV SENTENCE_TRANSFORMERS_HOME=/app/.hf_cache/sentence_transformers # Set ownership (includes .hf_cache written by the build step below) RUN chown -R appuser:appuser /app USER appuser # Pre-build FAISS index + download the embedding model into /app/.hf_cache RUN python src/build_faiss.py # ── Offline mode ──────────────────────────────────────────────────────────── # Model is now cached in the image. Tell all HF libraries to NEVER call the # network at runtime — prevents "Could not resolve host: huggingface.co" errors. ENV TRANSFORMERS_OFFLINE=1 ENV HF_DATASETS_OFFLINE=1 ENV HF_HUB_OFFLINE=1 # Environment defaults (override via HF Space secrets) ENV GROQ_API_KEY_1="" ENV GROQ_API_KEY_2="" ENV GROQ_API_KEY_3="" ENV GROQ_MODEL="llama-3.3-70b-versatile" ENV ALLOWED_ORIGINS="*" ENV PORT=7860 # Expose HF Spaces port EXPOSE 7860 CMD ["python", "api_server_fastapi.py"]