File size: 4,177 Bytes
7e09b15 | 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 | # ============================================================
# CCPA Compliance Analyzer
# Base: NVIDIA CUDA 12.1 + Ubuntu 22.04 for GPU support
# Falls back gracefully to CPU if no GPU is available
# ============================================================
FROM nvidia/cuda:12.1.0-base-ubuntu22.04
# ── Build args ───────────────────────────────────────────────
ARG MODEL_NAME=llama3.2:3b
ARG DEBIAN_FRONTEND=noninteractive
# ── System dependencies ──────────────────────────────────────
RUN apt-get update && apt-get install -y \
python3.11 \
python3.11-dev \
python3-pip \
curl \
ca-certificates \
&& rm -rf /var/lib/apt/lists/*
# Make python3.11 the default python
RUN update-alternatives --install /usr/bin/python3 python3 /usr/bin/python3.11 1 && \
update-alternatives --install /usr/bin/python python /usr/bin/python3.11 1
# ── Install Ollama ───────────────────────────────────────────
RUN curl -fsSL https://ollama.com/install.sh | sh
# ── Working directory ────────────────────────────────────────
WORKDIR /app
# ── Python dependencies (cached layer) ──────────────────────
COPY requirements.txt .
RUN pip install --no-cache-dir -r requirements.txt
# ── Application code ─────────────────────────────────────────
COPY app.py .
COPY ccpa_knowledge.py .
# ── Pre-download model weights into the image ────────────────
# Ollama needs to be running to pull; we start it, pull, then stop.
# OLLAMA_MODELS sets a consistent path for the weights.
ENV OLLAMA_MODELS=/root/.ollama/models
RUN ollama serve & \
SERVER_PID=$! && \
echo "Waiting for Ollama daemon..." && \
timeout 60 sh -c 'until curl -sf http://localhost:11434/api/tags; do sleep 2; done' && \
echo "Pulling model: ${MODEL_NAME}" && \
ollama pull ${MODEL_NAME} && \
echo "Model pull complete." && \
kill $SERVER_PID && \
wait $SERVER_PID 2>/dev/null || true
# ── Startup script ───────────────────────────────────────────
COPY start.sh .
RUN chmod +x start.sh
# ── Runtime environment ──────────────────────────────────────
ENV MODEL_NAME=${MODEL_NAME}
ENV OLLAMA_HOST=http://localhost:11434
ENV PYTHONUNBUFFERED=1
# ── Port ─────────────────────────────────────────────────────
EXPOSE 8000
# ── Health check ─────────────────────────────────────────────
HEALTHCHECK --interval=15s --timeout=10s --start-period=120s --retries=5 \
CMD curl -sf http://localhost:8000/health || exit 1
# ── Entrypoint ───────────────────────────────────────────────
CMD ["./start.sh"]
# Use a suitable base image (e.g., Ubuntu, Alpine, Node.js)
FROM alpine:latest
# Install unzip since Docker does not natively support extracting .zip files with ADD
# and clean up package cache in a single RUN command to keep the image small
RUN apk update && apk add unzip && rm -rf /var/cache/apk/*
# Set the working directory inside the container
WORKDIR /app
# Copy the local zip file into the container
COPY app.zip .
# Unzip the file and then remove the zip file to keep the final image clean
RUN unzip app.zip && rm app.zip
# Define the command to run when the container starts (replace with your application's start command)
CMD ["your_app_start_command"]
|