| # ββ HuggingFace Spaces compatible Dockerfile ββββββββββββββββββββββββββββββ | |
| # Port MUST be 7860 for HF Spaces. | |
| # Runs as user 1000 (HF requirement). | |
| FROM python:3.10-slim | |
| # System deps | |
| RUN apt-get update && apt-get install -y --no-install-recommends \ | |
| build-essential curl git \ | |
| && rm -rf /var/lib/apt/lists/* | |
| # Create non-root user (required by HF Spaces) | |
| RUN useradd -m -u 1000 appuser | |
| WORKDIR /app | |
| # Install Python deps first (better layer caching) | |
| COPY requirements_api.txt . | |
| RUN pip install --no-cache-dir -r requirements_api.txt | |
| # Copy application files | |
| COPY . . | |
| # Pin HuggingFace cache inside /app so the sentence-transformer model is | |
| # downloaded once during `docker build` and baked into the image layer. | |
| ENV HF_HOME=/app/.hf_cache | |
| ENV TRANSFORMERS_CACHE=/app/.hf_cache/transformers | |
| ENV SENTENCE_TRANSFORMERS_HOME=/app/.hf_cache/sentence_transformers | |
| # Set ownership (includes .hf_cache written by the build step below) | |
| RUN chown -R appuser:appuser /app | |
| USER appuser | |
| # Pre-build FAISS index + download the embedding model into /app/.hf_cache | |
| RUN python src/build_faiss.py | |
| # ββ Offline mode ββββββββββββββββββββββββββββββββββββββββββββββββββββββββββββ | |
| # Model is now cached in the image. Tell all HF libraries to NEVER call the | |
| # network at runtime β prevents "Could not resolve host: huggingface.co" errors. | |
| ENV TRANSFORMERS_OFFLINE=1 | |
| ENV HF_DATASETS_OFFLINE=1 | |
| ENV HF_HUB_OFFLINE=1 | |
| # Environment defaults (override via HF Space secrets) | |
| ENV GROQ_API_KEY_1="" | |
| ENV GROQ_API_KEY_2="" | |
| ENV GROQ_API_KEY_3="" | |
| ENV GROQ_MODEL="llama-3.3-70b-versatile" | |
| ENV ALLOWED_ORIGINS="*" | |
| ENV PORT=7860 | |
| # Expose HF Spaces port | |
| EXPOSE 7860 | |
| CMD ["python", "api_server_fastapi.py"] | |