Spaces:
Running
Running
| FROM python:3.10-slim | |
| WORKDIR /app | |
| # 1. Install wget (essential for the model) | |
| RUN apt-get update && apt-get install -y wget && rm -rf /var/lib/apt/lists/* | |
| # 2. Install llama-cpp-python FORCING the binary (No compilation allowed) | |
| # We use version 0.2.76 as it is highly stable for MPT models | |
| RUN pip install --no-cache-dir fastapi uvicorn | |
| RUN pip install llama-cpp-python==0.2.76 \ | |
| --extra-index-url https://abetlen.github.io/llama-cpp-python/whl/cpu \ | |
| --prefer-binary | |
| # 3. Download the model silently | |
| RUN wget -q -O model.gguf \ | |
| "https://huggingface.co/maddes8cht/mosaicml-mpt-7b-chat-gguf/resolve/main/mosaicml-mpt-7b-chat-Q2_K.gguf?download=true" | |
| # 4. Copy your script | |
| COPY app.py . | |
| EXPOSE 7860 | |
| CMD ["python", "app.py"] |